Symbol: mem
arch/alpha/include/asm/atomic.h
219
: [mem] "m"(*v), [a] "rI"(a), [u] "rI"((long)u)
arch/alpha/include/asm/atomic.h
242
: [mem] "m"(*v), [a] "rI"(a), [u] "rI"(u)
arch/alpha/include/asm/atomic.h
264
: [mem] "m"(*v)
arch/alpha/include/uapi/asm/compiler.h
94
#define __kernel_ldbu(mem) (mem)
arch/alpha/include/uapi/asm/compiler.h
95
#define __kernel_ldwu(mem) (mem)
arch/alpha/include/uapi/asm/compiler.h
96
#define __kernel_stb(val,mem) ((mem) = (val))
arch/alpha/include/uapi/asm/compiler.h
97
#define __kernel_stw(val,mem) ((mem) = (val))
arch/alpha/kernel/core_marvel.c
972
marvel_agp_bind_memory(alpha_agp_info *agp, off_t pg_start, struct agp_memory *mem)
arch/alpha/kernel/core_marvel.c
976
mem->page_count, mem->pages);
arch/alpha/kernel/core_marvel.c
980
marvel_agp_unbind_memory(alpha_agp_info *agp, off_t pg_start, struct agp_memory *mem)
arch/alpha/kernel/core_marvel.c
984
mem->page_count);
arch/alpha/kernel/core_mcpcia.c
292
struct resource *io, *mem, *hae_mem;
arch/alpha/kernel/core_mcpcia.c
299
mem = alloc_resource();
arch/alpha/kernel/core_mcpcia.c
316
mem->start = MCPCIA_DENSE(mid) - MCPCIA_MEM_BIAS;
arch/alpha/kernel/core_mcpcia.c
317
mem->end = mem->start + 0xffffffff;
arch/alpha/kernel/core_mcpcia.c
318
mem->name = pci_mem_names[h];
arch/alpha/kernel/core_mcpcia.c
319
mem->flags = IORESOURCE_MEM;
arch/alpha/kernel/core_mcpcia.c
321
hae_mem->start = mem->start;
arch/alpha/kernel/core_mcpcia.c
322
hae_mem->end = mem->start + MCPCIA_MEM_MASK;
arch/alpha/kernel/core_mcpcia.c
328
if (request_resource(&iomem_resource, mem) < 0)
arch/alpha/kernel/core_mcpcia.c
330
if (request_resource(mem, hae_mem) < 0)
arch/alpha/kernel/core_titan.c
683
titan_agp_bind_memory(alpha_agp_info *agp, off_t pg_start, struct agp_memory *mem)
arch/alpha/kernel/core_titan.c
687
mem->page_count, mem->pages);
arch/alpha/kernel/core_titan.c
691
titan_agp_unbind_memory(alpha_agp_info *agp, off_t pg_start, struct agp_memory *mem)
arch/alpha/kernel/core_titan.c
695
mem->page_count);
arch/alpha/kernel/pci_iommu.c
53
unsigned long mem = max_low_pfn << PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
54
if (mem < max)
arch/alpha/kernel/pci_iommu.c
55
max = roundup_pow_of_two(mem);
arch/arc/include/asm/entry-compact.h
213
.macro PROLOG_FREEUP_REG reg, mem
arch/arc/include/asm/entry-compact.h
217
.macro PROLOG_RESTORE_REG reg, mem
arch/arc/kernel/unwind.c
373
core_text = &module->mem[MOD_TEXT];
arch/arc/kernel/unwind.c
374
init_text = &module->mem[MOD_INIT_TEXT];
arch/arc/net/bpf_jit_core.c
195
static inline void maybe_free(struct jit_context *ctx, void **mem)
arch/arc/net/bpf_jit_core.c
197
if (*mem) {
arch/arc/net/bpf_jit_core.c
199
kfree(*mem);
arch/arc/net/bpf_jit_core.c
200
*mem = NULL;
arch/arm/boot/compressed/atags_to_fdt.c
175
if (!atag->u.mem.size)
arch/arm/boot/compressed/atags_to_fdt.c
186
cpu_to_fdt64(atag->u.mem.start);
arch/arm/boot/compressed/atags_to_fdt.c
188
cpu_to_fdt64(atag->u.mem.size);
arch/arm/boot/compressed/atags_to_fdt.c
191
cpu_to_fdt32(atag->u.mem.start);
arch/arm/boot/compressed/atags_to_fdt.c
193
cpu_to_fdt32(atag->u.mem.size);
arch/arm/common/locomo.c
356
__locomo_probe(struct device *me, struct resource *mem, int irq)
arch/arm/common/locomo.c
372
lchip->phys = mem->start;
arch/arm/common/locomo.c
380
lchip->base = ioremap(mem->start, PAGE_SIZE);
arch/arm/common/locomo.c
488
struct resource *mem;
arch/arm/common/locomo.c
491
mem = platform_get_resource(dev, IORESOURCE_MEM, 0);
arch/arm/common/locomo.c
492
if (!mem)
arch/arm/common/locomo.c
498
return __locomo_probe(&dev->dev, mem, irq);
arch/arm/common/sa1111.c
1118
struct resource *mem;
arch/arm/common/sa1111.c
1121
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
arch/arm/common/sa1111.c
1122
if (!mem)
arch/arm/common/sa1111.c
1128
return __sa1111_probe(&pdev->dev, mem, irq);
arch/arm/common/sa1111.c
792
static int __sa1111_probe(struct device *me, struct resource *mem, int irq)
arch/arm/common/sa1111.c
821
sachip->phys = mem->start;
arch/arm/common/sa1111.c
828
sachip->base = ioremap(mem->start, PAGE_SIZE * 2);
arch/arm/common/sa1111.c
903
sa1111_init_one_child(sachip, mem, &sa1111_devices[i]);
arch/arm/common/scoop.c
182
struct resource *mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
arch/arm/common/scoop.c
185
if (!mem)
arch/arm/common/scoop.c
195
devptr->base = ioremap(mem->start, resource_size(mem));
arch/arm/common/scoop.c
204
printk("Sharp Scoop Device found at 0x%08x -> 0x%8p\n",(unsigned int)mem->start, devptr->base);
arch/arm/include/uapi/asm/setup.h
151
struct tag_mem32 mem;
arch/arm/kernel/atags_compat.c
86
tag->u.mem.size = size;
arch/arm/kernel/atags_compat.c
87
tag->u.mem.start = start;
arch/arm/kernel/atags_parse.c
183
default_tags.mem.start = PHYS_OFFSET;
arch/arm/kernel/atags_parse.c
43
struct tag_mem32 mem;
arch/arm/kernel/atags_parse.c
67
return arm_add_memory(tag->u.mem.start, tag->u.mem.size);
arch/arm/kernel/kgdb.c
51
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/arm/kernel/kgdb.c
57
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/arm/kernel/kgdb.c
60
memset(mem, 0, dbg_reg_def[regno].size);
arch/arm/kernel/kgdb.c
64
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/arm/kernel/kgdb.c
70
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/arm/kernel/machine_kexec.c
58
if (!memblock_is_region_memory(idmap_to_phys(current_segment->mem),
arch/arm/kernel/machine_kexec.c
67
image->arch.kernel_r2 = current_segment->mem;
arch/arm/mach-omap2/cm_common.c
317
struct omap_domain_base *mem = NULL;
arch/arm/mach-omap2/cm_common.c
329
mem = &cm_base;
arch/arm/mach-omap2/cm_common.c
332
mem = &cm2_base;
arch/arm/mach-omap2/cm_common.c
334
data->mem = ioremap(res.start, resource_size(&res));
arch/arm/mach-omap2/cm_common.c
336
if (mem) {
arch/arm/mach-omap2/cm_common.c
337
mem->pa = res.start + data->offset;
arch/arm/mach-omap2/cm_common.c
338
mem->va = data->mem + data->offset;
arch/arm/mach-omap2/cm_common.c
339
mem->offset = data->offset;
arch/arm/mach-omap2/cm_common.c
371
ret = omap2_clk_provider_init(np, data->index, NULL, data->mem);
arch/arm/mach-omap2/control.c
662
void __iomem *mem;
arch/arm/mach-omap2/control.c
705
void __iomem *mem;
arch/arm/mach-omap2/control.c
710
mem = of_iomap(np, 0);
arch/arm/mach-omap2/control.c
711
if (!mem) {
arch/arm/mach-omap2/control.c
717
omap2_ctrl_base = mem;
arch/arm/mach-omap2/control.c
721
data->mem = mem;
arch/arm/mach-omap2/control.c
772
data->mem);
arch/arm/mach-omap2/prcm-common.h
544
void __iomem *mem;
arch/arm/mach-omap2/prm_common.c
717
data->mem = ioremap(res.start, resource_size(&res));
arch/arm/mach-omap2/prm_common.c
720
prm_base.va = data->mem + data->offset;
arch/arm/mach-omap2/prm_common.c
760
ret = omap2_clk_provider_init(np, data->index, NULL, data->mem);
arch/arm/mach-orion5x/common.c
371
(!t->u.mem.size || t->u.mem.size & ~PAGE_MASK ||
arch/arm/mach-orion5x/common.c
372
t->u.mem.start & ~PAGE_MASK)) {
arch/arm/mach-orion5x/common.c
375
t->u.mem.size / 1024, t->u.mem.start);
arch/arm/mach-rpc/include/mach/uncompress.h
130
nr_pages += (t->u.mem.size / PAGE_SIZE);
arch/arm/mm/pmsa-v7.c
25
static struct region __initdata mem[MPU_MAX_REGIONS];
arch/arm/mm/pmsa-v7.c
293
memset(mem, 0, sizeof(mem));
arch/arm/mm/pmsa-v7.c
294
num = allocate_region(mem_start, specified_mem_size, mem_max_regions, mem);
arch/arm/mm/pmsa-v7.c
297
unsigned long subreg = mem[i].size / PMSAv7_NR_SUBREGS;
arch/arm/mm/pmsa-v7.c
299
total_mem_size += mem[i].size - subreg * hweight_long(mem[i].subreg);
arch/arm/mm/pmsa-v7.c
302
&mem[i].base, &mem[i].size, PMSAv7_NR_SUBREGS, &mem[i].subreg);
arch/arm/mm/pmsa-v7.c
453
for (i = 0; i < ARRAY_SIZE(mem); i++) {
arch/arm/mm/pmsa-v7.c
454
if (!mem[i].size)
arch/arm/mm/pmsa-v7.c
457
err |= mpu_setup_region(region++, mem[i].base, ilog2(mem[i].size),
arch/arm/mm/pmsa-v7.c
459
mem[i].subreg, false);
arch/arm/mm/pmsa-v8.c
248
add_range(mem, ARRAY_SIZE(mem), 0, memblock.memory.regions[0].base,
arch/arm/mm/pmsa-v8.c
255
subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END));
arch/arm/mm/pmsa-v8.c
260
subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom));
arch/arm/mm/pmsa-v8.c
266
subtract_range(mem, ARRAY_SIZE(mem), vectors_base, vectors_base + 2 * PAGE_SIZE);
arch/arm/mm/pmsa-v8.c
270
for (i = 0; i < ARRAY_SIZE(mem); i++)
arch/arm/mm/pmsa-v8.c
271
subtract_range(io, ARRAY_SIZE(io), mem[i].start, mem[i].end);
arch/arm/mm/pmsa-v8.c
292
for (i = 0; i < ARRAY_SIZE(mem); i++) {
arch/arm/mm/pmsa-v8.c
293
if (!mem[i].end)
arch/arm/mm/pmsa-v8.c
296
err |= pmsav8_setup_ram(region++, mem[i].start, mem[i].end);
arch/arm/mm/pmsa-v8.c
79
static struct range __initdata mem[MPU_MAX_REGIONS];
arch/arm/probes/kprobes/test-core.c
1278
static void print_memory(u32 *mem, size_t size)
arch/arm/probes/kprobes/test-core.c
1282
pr_err("%08x %08x %08x %08x\n", mem[i], mem[i+1],
arch/arm/probes/kprobes/test-core.c
1283
mem[i+2], mem[i+3]);
arch/arm/probes/kprobes/test-core.c
1433
u32 *mem = 0;
arch/arm/probes/kprobes/test-core.c
1441
mem = (u32 *)result_regs.ARM_sp;
arch/arm/probes/kprobes/test-core.c
1442
mem_size = expected_memory_size(mem);
arch/arm/probes/kprobes/test-core.c
1443
if (memcmp(expected_memory, mem, mem_size)) {
arch/arm/probes/kprobes/test-core.c
1459
if (mem) {
arch/arm/probes/kprobes/test-core.c
1463
print_memory(mem, mem_size);
arch/arm/probes/kprobes/test-core.c
1498
u32 *mem = (u32 *)result_regs.ARM_sp;
arch/arm/probes/kprobes/test-core.c
1500
memcpy(expected_memory, mem, expected_memory_size(mem));
arch/arm64/kernel/kexec_image.c
100
ret = load_other_segments(image, kernel_segment->mem,
arch/arm64/kernel/kexec_image.c
111
kbuf.buf_min = kernel_segment->mem + kernel_segment->memsz;
arch/arm64/kernel/kexec_image.c
112
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/arm64/kernel/kexec_image.c
121
kernel_segment->mem += text_offset;
arch/arm64/kernel/kexec_image.c
123
image->start = kernel_segment->mem;
arch/arm64/kernel/kexec_image.c
126
kernel_segment->mem, kbuf.bufsz,
arch/arm64/kernel/kexec_image.c
82
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/arm64/kernel/kgdb.c
104
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/arm64/kernel/kgdb.c
110
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/arm64/kernel/kgdb.c
113
memset(mem, 0, dbg_reg_def[regno].size);
arch/arm64/kernel/kgdb.c
117
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/arm64/kernel/kgdb.c
123
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/arm64/kernel/machine_kexec.c
281
if (addr >= kexec_crash_image->segment[i].mem &&
arch/arm64/kernel/machine_kexec.c
282
addr < (kexec_crash_image->segment[i].mem +
arch/arm64/kernel/machine_kexec.c
77
kimage->segment[i].mem,
arch/arm64/kernel/machine_kexec.c
78
kimage->segment[i].mem + kimage->segment[i].memsz,
arch/arm64/kernel/machine_kexec.c
83
(unsigned long)phys_to_virt(kimage->segment[i].mem),
arch/arm64/kernel/machine_kexec.c
84
(unsigned long)phys_to_virt(kimage->segment[i].mem) +
arch/arm64/kernel/machine_kexec_file.c
120
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/arm64/kernel/machine_kexec_file.c
132
image->elf_load_addr = kbuf.mem;
arch/arm64/kernel/machine_kexec_file.c
144
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/arm64/kernel/machine_kexec_file.c
155
initrd_load_addr = kbuf.mem;
arch/arm64/kernel/machine_kexec_file.c
175
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/arm64/kernel/machine_kexec_file.c
186
image->arch.dtb_mem = kbuf.mem;
arch/arm64/kernel/machine_kexec_file.c
189
kbuf.mem, kbuf.bufsz, kbuf.memsz);
arch/arm64/kernel/setup.c
251
struct resource *mem = &standard_resources[i];
arch/arm64/kernel/setup.c
252
phys_addr_t r_start, r_end, mem_size = resource_size(mem);
arch/arm64/kernel/setup.c
254
if (!memblock_is_region_reserved(mem->start, mem_size))
arch/arm64/kernel/setup.c
260
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/arm64/kernel/setup.c
261
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/arm64/kernel/setup.c
263
if (start > mem->end || end < mem->start)
arch/arm64/kernel/setup.c
266
reserve_region_with_split(mem, start, end, "reserved");
arch/hexagon/kernel/kgdb.c
78
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/hexagon/kernel/kgdb.c
83
*((unsigned long *) mem) = *((unsigned long *) ((void *)regs +
arch/hexagon/kernel/kgdb.c
89
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/hexagon/kernel/kgdb.c
95
*((unsigned long *) mem);
arch/loongarch/include/asm/barrier.h
123
: [mem] "+ZB" (*(u32 *)&p), [tmp] "=&r" (__tmp) \
arch/loongarch/include/asm/barrier.h
130
: [mem] "+ZB" (*(u64 *)&p), [tmp] "=&r" (__tmp) \
arch/loongarch/kernel/kexec_efi.c
102
image->start = kernel_segment->mem + h->kernel_entry - text_offset;
arch/loongarch/kernel/kexec_efi.c
105
kernel_segment->mem, kbuf.bufsz, kernel_segment->memsz);
arch/loongarch/kernel/kexec_efi.c
63
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/loongarch/kernel/kexec_efi.c
79
ret = load_other_segments(image, kernel_segment->mem,
arch/loongarch/kernel/kexec_efi.c
90
kbuf.buf_min = kernel_segment->mem + kernel_segment->memsz;
arch/loongarch/kernel/kexec_efi.c
91
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/loongarch/kernel/kexec_elf.c
45
kbuf->mem = KEXEC_BUF_MEM_UNKNOWN;
arch/loongarch/kernel/kexec_elf.c
86
ret = load_other_segments(image, kernel_segment->mem, kernel_segment->memsz,
arch/loongarch/kernel/kexec_elf.c
92
image->start = kernel_segment->mem + __pa(ehdr.e_entry) - text_offset;
arch/loongarch/kernel/kexec_elf.c
95
kernel_segment->mem, kbuf.bufsz, kernel_segment->memsz);
arch/loongarch/kernel/kgdb.c
108
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/loongarch/kernel/kgdb.c
123
memcpy(mem, (void *)regs + reg_offset, reg_size);
arch/loongarch/kernel/kgdb.c
135
memcpy(mem, (void *)&current->thread.fpu.fcsr, reg_size);
arch/loongarch/kernel/kgdb.c
138
memcpy(mem, (void *)&current->thread.fpu.fcc + reg_offset, reg_size);
arch/loongarch/kernel/kgdb.c
141
memcpy(mem, (void *)&current->thread.fpu.fpr[reg_offset], reg_size);
arch/loongarch/kernel/kgdb.c
151
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/loongarch/kernel/kgdb.c
166
memcpy((void *)regs + reg_offset, mem, reg_size);
arch/loongarch/kernel/kgdb.c
176
memcpy((void *)&current->thread.fpu.fcsr, mem, reg_size);
arch/loongarch/kernel/kgdb.c
179
memcpy((void *)&current->thread.fpu.fcc + reg_offset, mem, reg_size);
arch/loongarch/kernel/kgdb.c
182
memcpy((void *)&current->thread.fpu.fpr[reg_offset], mem, reg_size);
arch/loongarch/kernel/machine_kexec_file.c
174
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/loongarch/kernel/machine_kexec_file.c
186
image->elf_load_addr = kbuf.mem;
arch/loongarch/kernel/machine_kexec_file.c
204
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/loongarch/kernel/machine_kexec_file.c
214
initrd_load_addr = kbuf.mem;
arch/loongarch/kernel/setup.c
546
struct resource *mem = &standard_resources[i];
arch/loongarch/kernel/setup.c
547
phys_addr_t r_start, r_end, mem_size = resource_size(mem);
arch/loongarch/kernel/setup.c
549
if (!memblock_is_region_reserved(mem->start, mem_size))
arch/loongarch/kernel/setup.c
555
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/loongarch/kernel/setup.c
556
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/loongarch/kernel/setup.c
558
if (start > mem->end || end < mem->start)
arch/loongarch/kernel/setup.c
561
reserve_region_with_split(mem, start, end, "Reserved");
arch/m68k/include/asm/syscalls.h
11
int d4, int d5, unsigned long __user *mem);
arch/m68k/kernel/setup_mm.c
481
unsigned long mem;
arch/m68k/kernel/setup_mm.c
490
for (mem = 0, i = 0; i < m68k_num_memory; i++)
arch/m68k/kernel/setup_mm.c
491
mem += m68k_memory[i].size;
arch/m68k/kernel/setup_mm.c
492
seq_printf(m, "System Memory:\t%ldK\n", mem >> 10);
arch/m68k/kernel/sys_m68k.c
464
unsigned long __user * mem)
arch/m68k/kernel/sys_m68k.c
478
pgd = pgd_offset(mm, (unsigned long)mem);
arch/m68k/kernel/sys_m68k.c
481
p4d = p4d_offset(pgd, (unsigned long)mem);
arch/m68k/kernel/sys_m68k.c
484
pud = pud_offset(p4d, (unsigned long)mem);
arch/m68k/kernel/sys_m68k.c
487
pmd = pmd_offset(pud, (unsigned long)mem);
arch/m68k/kernel/sys_m68k.c
490
pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl);
arch/m68k/kernel/sys_m68k.c
503
__get_user(mem_value, mem);
arch/m68k/kernel/sys_m68k.c
505
__put_user(newval, mem);
arch/m68k/kernel/sys_m68k.c
523
if (do_page_fault(fp, (unsigned long)mem, 3))
arch/m68k/kernel/sys_m68k.c
547
unsigned long __user * mem)
arch/m68k/kernel/sys_m68k.c
554
mem_value = *mem;
arch/m68k/kernel/sys_m68k.c
556
*mem = newval;
arch/mips/bcm47xx/prom.c
101
if (c->cputype == CPU_74K && (mem == (128 << 20)))
arch/mips/bcm47xx/prom.c
102
mem -= 0x1000;
arch/mips/bcm47xx/prom.c
103
memblock_add(0, mem);
arch/mips/bcm47xx/prom.c
60
unsigned long mem;
arch/mips/bcm47xx/prom.c
84
for (mem = 1 << 20; mem < max; mem += 1 << 20) {
arch/mips/bcm47xx/prom.c
86
if (off + mem >= max) {
arch/mips/bcm47xx/prom.c
87
mem = max;
arch/mips/bcm47xx/prom.c
91
if (!memcmp((void *)prom_init, (void *)prom_init + mem, 32))
arch/mips/bcm47xx/prom.c
94
lowmem = mem;
arch/mips/boot/tools/relocs.c
457
void *mem = realloc(r->offset, newsize * sizeof(r->offset[0]));
arch/mips/boot/tools/relocs.c
459
if (!mem)
arch/mips/boot/tools/relocs.c
462
r->offset = mem;
arch/mips/cavium-octeon/executive/cvmx-boot-vector.c
134
static void cvmx_boot_vector_init(void *mem)
arch/mips/cavium-octeon/executive/cvmx-boot-vector.c
139
memset(mem, 0, VECTOR_TABLE_SIZE);
arch/mips/cavium-octeon/executive/cvmx-boot-vector.c
140
kseg0_mem = cvmx_ptr_to_phys(mem) | 0x8000000000000000ull;
arch/mips/cavium-octeon/setup.c
177
image->segment[i].mem + offt;
arch/mips/cavium-octeon/setup.c
232
kimage_ptr->segment[i].mem - PAGE_SIZE,
arch/mips/cavium-octeon/setup.c
928
static __init void memory_exclude_page(u64 addr, u64 *mem, u64 *size)
arch/mips/cavium-octeon/setup.c
930
if (addr > *mem && addr < *mem + *size) {
arch/mips/cavium-octeon/setup.c
931
u64 inc = addr - *mem;
arch/mips/cavium-octeon/setup.c
932
memblock_add(*mem, inc);
arch/mips/cavium-octeon/setup.c
933
*mem += inc;
arch/mips/cavium-octeon/setup.c
937
if (addr == *mem && *size > PAGE_SIZE) {
arch/mips/cavium-octeon/setup.c
938
*mem += PAGE_SIZE;
arch/mips/include/asm/bitops.h
26
#define __bit_op(mem, insn, inputs...) do { \
arch/mips/include/asm/bitops.h
38
: "=&r"(__temp), "+" GCC_OFF_SMALL_ASM()(mem) \
arch/mips/include/asm/bitops.h
43
#define __test_bit_op(mem, ll_dst, insn, inputs...) ({ \
arch/mips/include/asm/bitops.h
56
"+" GCC_OFF_SMALL_ASM()(mem) \
arch/mips/include/asm/io.h
183
volatile void __iomem *mem) \
arch/mips/include/asm/io.h
193
__mem = (void *)__swizzle_addr_##bwlq((unsigned long)(mem)); \
arch/mips/include/asm/io.h
222
static inline type pfx##read##bwlq(const volatile void __iomem *mem) \
arch/mips/include/asm/io.h
227
__mem = (void *)__swizzle_addr_##bwlq((unsigned long)(mem)); \
arch/mips/include/asm/io.h
378
static inline void writes##bwlq(volatile void __iomem *mem, \
arch/mips/include/asm/io.h
384
__mem_write##bwlq(*__addr, mem); \
arch/mips/include/asm/io.h
389
static inline void reads##bwlq(const volatile void __iomem *mem, \
arch/mips/include/asm/io.h
395
*__addr = __mem_read##bwlq(mem); \
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
29
# Read the cavium mem control register
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
34
dmtc0 v0, CP0_CVMMEMCTL_REG # Write the cavium mem control register
arch/mips/include/asm/mach-jazz/floppy.h
106
unsigned long mem;
arch/mips/include/asm/mach-jazz/floppy.h
108
mem = __get_dma_pages(GFP_KERNEL, get_order(size));
arch/mips/include/asm/mach-jazz/floppy.h
109
if(!mem)
arch/mips/include/asm/mach-jazz/floppy.h
111
vdma_alloc(CPHYSADDR(mem), size); /* XXX error checking */
arch/mips/include/asm/mach-jazz/floppy.h
113
return mem;
arch/mips/include/asm/octeon/cvmx-rnm-defs.h
43
uint64_t mem:1;
arch/mips/include/asm/octeon/cvmx-rnm-defs.h
45
uint64_t mem:1;
arch/mips/kernel/kgdb.c
130
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/mips/kernel/kgdb.c
138
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/mips/kernel/kgdb.c
146
memcpy((void *)&current->thread.fpu.fcr31, mem,
arch/mips/kernel/kgdb.c
154
memcpy((void *)&current->thread.fpu.fpr[fp_reg], mem,
arch/mips/kernel/kgdb.c
163
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/mips/kernel/kgdb.c
172
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/mips/kernel/kgdb.c
181
memcpy(mem, (void *)&current->thread.fpu.fcr31,
arch/mips/kernel/kgdb.c
186
memset(mem, 0, dbg_reg_def[regno].size);
arch/mips/kernel/kgdb.c
190
memcpy(mem, (void *)&current->thread.fpu.fpr[fp_reg],
arch/mips/kernel/machine_kexec.c
47
kimage->segment[i].mem,
arch/mips/kernel/machine_kexec.c
48
kimage->segment[i].mem + kimage->segment[i].memsz,
arch/mips/kernel/machine_kexec.c
79
phys_to_virt((unsigned long)kimage->segment[i].mem);
arch/mips/kernel/vpe.c
205
mod_mem = &mod->mem[MOD_TEXT];
arch/mips/kernel/vpe.c
644
v->load_addr = alloc_progmem(mod.mem[MOD_TEXT].size);
arch/mips/pci/pci-lantiq.c
84
u32 mem, bar11mask;
arch/mips/pci/pci-lantiq.c
87
mem = get_num_physpages() * PAGE_SIZE;
arch/mips/pci/pci-lantiq.c
88
bar11mask = (0x0ffffff0 & ~((1 << (fls(mem) - 1)) - 1)) | 8;
arch/mips/pci/pci-xtalk-bridge.c
653
pci_add_resource_offset(&host->windows, &bd->mem, bd->mem_offset);
arch/mips/pci/pcie-octeon.c
143
pcie_addr.mem.upper = 0;
arch/mips/pci/pcie-octeon.c
144
pcie_addr.mem.io = 1;
arch/mips/pci/pcie-octeon.c
145
pcie_addr.mem.did = 3;
arch/mips/pci/pcie-octeon.c
146
pcie_addr.mem.subdid = 3 + pcie_port;
arch/mips/pci/pcie-octeon.c
92
} mem;
arch/mips/sgi-ip27/ip27-xtalk.c
89
bd->mem.name = "Bridge PCI MEM";
arch/mips/sgi-ip27/ip27-xtalk.c
90
bd->mem.start = offset + (widget << SWIN_SIZE_BITS) + BRIDGE_DEVIO0;
arch/mips/sgi-ip27/ip27-xtalk.c
91
bd->mem.end = offset + (widget << SWIN_SIZE_BITS) + SWIN_SIZE - 1;
arch/mips/sgi-ip27/ip27-xtalk.c
92
bd->mem.flags = IORESOURCE_MEM;
arch/mips/sgi-ip30/ip30-xtalk.c
100
bd->mem.end = IP30_SWIN_BASE(widget) + IP30_SWIN_SIZE - 1;
arch/mips/sgi-ip30/ip30-xtalk.c
101
bd->mem.flags = IORESOURCE_MEM;
arch/mips/sgi-ip30/ip30-xtalk.c
98
bd->mem.name = "Bridge PCI MEM";
arch/mips/sgi-ip30/ip30-xtalk.c
99
bd->mem.start = IP30_SWIN_BASE(widget) + BRIDGE_DEVIO0;
arch/nios2/kernel/kgdb.c
71
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/nios2/kernel/kgdb.c
77
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/nios2/kernel/kgdb.c
80
memset(mem, 0, dbg_reg_def[regno].size);
arch/nios2/kernel/kgdb.c
85
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/nios2/kernel/kgdb.c
91
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/parisc/kernel/kexec.c
27
kimage->segment[n].mem,
arch/parisc/kernel/kexec.c
28
kimage->segment[n].mem + kimage->segment[n].memsz,
arch/parisc/kernel/kexec_file.c
39
image->segment[i].mem = __pa(image->segment[i].mem);
arch/parisc/kernel/kexec_file.c
49
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/parisc/kernel/kexec_file.c
54
kexec_dprintk("Loaded initrd at 0x%lx\n", kbuf.mem);
arch/parisc/kernel/kexec_file.c
55
image->arch.initrd_start = kbuf.mem;
arch/parisc/kernel/kexec_file.c
56
image->arch.initrd_end = kbuf.mem + initrd_len;
arch/parisc/kernel/kexec_file.c
66
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/parisc/kernel/kexec_file.c
71
kexec_dprintk("Loaded cmdline at 0x%lx\n", kbuf.mem);
arch/parisc/kernel/kexec_file.c
72
image->arch.cmdline = kbuf.mem;
arch/parisc/kernel/module.c
319
mod_mem = &me->mem[MOD_TEXT];
arch/parisc/kernel/module.c
345
got = me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
363
Elf_Fdesc *fdesc = me->mem[MOD_TEXT].base + me->arch.fdesc_offset;
arch/parisc/kernel/module.c
381
fdesc->gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
813
gp = (Elf_Addr)me->mem[MOD_TEXT].base + me->arch.got_offset;
arch/parisc/kernel/module.c
951
unsigned long start_opd = (Elf64_Addr)mod->mem[MOD_TEXT].base +
arch/powerpc/boot/cuboot-pq2.c
122
struct pci_range *mem = NULL, *mmio = NULL,
arch/powerpc/boot/cuboot-pq2.c
162
mem = &pci_ranges_buf[i];
arch/powerpc/boot/cuboot-pq2.c
169
if (!mem || !mmio || !io)
arch/powerpc/boot/cuboot-pq2.c
171
if (mem->size[1] != mmio->size[1])
arch/powerpc/boot/cuboot-pq2.c
173
if (mem->size[1] & (mem->size[1] - 1))
arch/powerpc/boot/cuboot-pq2.c
178
if (mem->phys_addr + mem->size[1] == mmio->phys_addr)
arch/powerpc/boot/cuboot-pq2.c
179
mem_base = mem;
arch/powerpc/boot/cuboot-pq2.c
180
else if (mmio->phys_addr + mmio->size[1] == mem->phys_addr)
arch/powerpc/boot/cuboot-pq2.c
186
out_be32(&pci_regs[2][0], ~(mem->size[1] + mmio->size[1] - 1));
arch/powerpc/boot/cuboot-pq2.c
191
out_le32(&pci_regs[0][0], mem->pci_addr[1] >> 12);
arch/powerpc/boot/cuboot-pq2.c
192
out_le32(&pci_regs[0][2], mem->phys_addr >> 12);
arch/powerpc/boot/cuboot-pq2.c
193
out_le32(&pci_regs[0][4], (~(mem->size[1] - 1) >> 12) | 0xa0000000);
arch/powerpc/boot/wii.c
101
void *mem;
arch/powerpc/boot/wii.c
107
mem = finddevice("/memory");
arch/powerpc/boot/wii.c
108
if (!mem)
arch/powerpc/boot/wii.c
112
len = getprop(mem, "reg", reg, sizeof(reg));
arch/powerpc/boot/wii.c
128
setprop(mem, "reg", reg, sizeof(reg));
arch/powerpc/include/asm/mmu_context.h
32
struct mm_iommu_table_group_mem_t *mem);
arch/powerpc/include/asm/mmu_context.h
38
extern long mm_iommu_ua_to_hpa(struct mm_iommu_table_group_mem_t *mem,
arch/powerpc/include/asm/mmu_context.h
42
extern long mm_iommu_mapped_inc(struct mm_iommu_table_group_mem_t *mem);
arch/powerpc/include/asm/mmu_context.h
43
extern void mm_iommu_mapped_dec(struct mm_iommu_table_group_mem_t *mem);
arch/powerpc/kernel/kgdb.c
319
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/powerpc/kernel/kgdb.c
327
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/powerpc/kernel/kgdb.c
334
memcpy(mem, &current->thread.evr[regno-32],
arch/powerpc/kernel/kgdb.c
338
memset(mem, 0, dbg_reg_def[regno].size);
arch/powerpc/kernel/kgdb.c
345
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/powerpc/kernel/kgdb.c
353
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/powerpc/kernel/kgdb.c
359
memcpy(&current->thread.evr[regno-32], mem,
arch/powerpc/kernel/module_32.c
324
module->arch.tramp = do_plt_call(module->mem[MOD_TEXT].base,
arch/powerpc/kernel/module_32.c
331
module->arch.tramp_regs = do_plt_call(module->mem[MOD_TEXT].base,
arch/powerpc/kexec/core_64.c
135
flush_icache_range((unsigned long)__va(ranges[i].mem),
arch/powerpc/kexec/core_64.c
136
(unsigned long)__va(ranges[i].mem + ranges[i].memsz));
arch/powerpc/kexec/core_64.c
52
if (image->segment[i].mem < __pa(_end))
arch/powerpc/kexec/core_64.c
66
begin = image->segment[i].mem;
arch/powerpc/kexec/crash.c
433
void *ptr, *mem, *elfbuf = NULL;
arch/powerpc/kexec/crash.c
437
mem = (void *) ksegment->mem;
arch/powerpc/kexec/crash.c
477
ptr = __va(mem);
arch/powerpc/kexec/crash.c
504
unsigned long mem;
arch/powerpc/kexec/crash.c
509
mem = image->segment[i].mem;
arch/powerpc/kexec/crash.c
510
ptr = __va(mem);
arch/powerpc/kexec/crash.c
537
fdt = __va((void *)image->segment[fdt_index].mem);
arch/powerpc/kexec/elf_64.c
102
initrd_load_addr = kbuf.mem;
arch/powerpc/kexec/elf_64.c
131
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/powerpc/kexec/elf_64.c
139
fdt_load_addr = kbuf.mem;
arch/powerpc/kexec/elf_64.c
46
.mem = KEXEC_BUF_MEM_UNKNOWN };
arch/powerpc/kexec/elf_64.c
72
kexec_dprintk("Loaded purgatory at 0x%lx\n", pbuf.mem);
arch/powerpc/kexec/elf_64.c
98
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/powerpc/kexec/file_load_64.c
362
kbuf->mem = KEXEC_BUF_MEM_UNKNOWN;
arch/powerpc/kexec/file_load_64.c
373
image->arch.backup_start = kbuf->mem;
arch/powerpc/kexec/file_load_64.c
451
kbuf->mem = KEXEC_BUF_MEM_UNKNOWN;
arch/powerpc/kexec/file_load_64.c
467
image->elf_load_addr = kbuf->mem;
arch/powerpc/kexec/file_load_64.c
501
kexec_dprintk("Loaded the backup region at 0x%lx\n", kbuf->mem);
arch/powerpc/kvm/book3s_64_vio.c
389
struct mm_iommu_table_group_mem_t *mem;
arch/powerpc/kvm/book3s_64_vio.c
392
mem = mm_iommu_lookup(stt->kvm->mm, ua, 1ULL << shift);
arch/powerpc/kvm/book3s_64_vio.c
393
if (!mem || mm_iommu_ua_to_hpa(mem, ua, shift, &hpa)) {
arch/powerpc/kvm/book3s_64_vio.c
451
struct mm_iommu_table_group_mem_t *mem = NULL;
arch/powerpc/kvm/book3s_64_vio.c
458
mem = mm_iommu_lookup(kvm->mm, be64_to_cpu(*pua), pgsize);
arch/powerpc/kvm/book3s_64_vio.c
459
if (!mem)
arch/powerpc/kvm/book3s_64_vio.c
462
mm_iommu_mapped_dec(mem);
arch/powerpc/kvm/book3s_64_vio.c
516
struct mm_iommu_table_group_mem_t *mem;
arch/powerpc/kvm/book3s_64_vio.c
522
mem = mm_iommu_lookup(kvm->mm, ua, 1ULL << tbl->it_page_shift);
arch/powerpc/kvm/book3s_64_vio.c
523
if (!mem)
arch/powerpc/kvm/book3s_64_vio.c
527
if (WARN_ON_ONCE(mm_iommu_ua_to_hpa(mem, ua, tbl->it_page_shift, &hpa)))
arch/powerpc/kvm/book3s_64_vio.c
530
if (mm_iommu_mapped_inc(mem))
arch/powerpc/kvm/book3s_64_vio.c
535
mm_iommu_mapped_dec(mem);
arch/powerpc/lib/sstep.c
1010
u8 mem[32];
arch/powerpc/lib/sstep.c
1047
emulate_vsx_store(op, buf, mem, cross_endian);
arch/powerpc/lib/sstep.c
1048
return copy_mem_out(mem, ea, size, regs);
arch/powerpc/lib/sstep.c
784
const void *mem, bool rev)
arch/powerpc/lib/sstep.c
802
memcpy(reg, mem, size);
arch/powerpc/lib/sstep.c
812
memcpy(&reg->b[i], mem, read_size);
arch/powerpc/lib/sstep.c
827
unsigned long v = *(unsigned long *)(mem + 8);
arch/powerpc/lib/sstep.c
835
wp = mem;
arch/powerpc/lib/sstep.c
850
hp = mem;
arch/powerpc/lib/sstep.c
858
bp = mem;
arch/powerpc/lib/sstep.c
868
void *mem, bool rev)
arch/powerpc/lib/sstep.c
891
memcpy(mem, buf32, size);
arch/powerpc/lib/sstep.c
893
memcpy(mem, reg, size);
arch/powerpc/lib/sstep.c
908
memcpy(mem, reg, size);
arch/powerpc/lib/sstep.c
921
memcpy(mem, &reg->b[i], write_size);
arch/powerpc/lib/sstep.c
923
memcpy(mem + 8, &reg->d[IS_BE], 8);
arch/powerpc/lib/sstep.c
925
do_byte_reverse(mem, write_size);
arch/powerpc/lib/sstep.c
927
do_byte_reverse(mem + 8, 8);
arch/powerpc/lib/sstep.c
932
wp = mem;
arch/powerpc/lib/sstep.c
940
hp = mem;
arch/powerpc/lib/sstep.c
948
bp = mem;
arch/powerpc/lib/sstep.c
963
u8 mem[32];
arch/powerpc/lib/sstep.c
967
if (!address_ok(regs, ea, size) || copy_mem_in(mem, ea, size, regs))
arch/powerpc/lib/sstep.c
971
emulate_vsx_load(op, buf, mem, cross_endian);
arch/powerpc/mm/book3s64/iommu_api.c
108
mem->hpages + entry);
arch/powerpc/mm/book3s64/iommu_api.c
125
atomic64_set(&mem->mapped, 1);
arch/powerpc/mm/book3s64/iommu_api.c
126
mem->used = 1;
arch/powerpc/mm/book3s64/iommu_api.c
127
mem->ua = ua;
arch/powerpc/mm/book3s64/iommu_api.c
128
mem->entries = entries;
arch/powerpc/mm/book3s64/iommu_api.c
144
if (mem->dev_hpa == MM_IOMMU_TABLE_INVALID_HPA) {
arch/powerpc/mm/book3s64/iommu_api.c
152
struct page *page = mem->hpages[i];
arch/powerpc/mm/book3s64/iommu_api.c
154
if ((mem->pageshift > PAGE_SHIFT) && PageHuge(page))
arch/powerpc/mm/book3s64/iommu_api.c
156
mem->pageshift = min(mem->pageshift, pageshift);
arch/powerpc/mm/book3s64/iommu_api.c
161
mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
165
list_add_rcu(&mem->next, &mm->context.iommu_group_mem_list);
arch/powerpc/mm/book3s64/iommu_api.c
169
*pmem = mem;
arch/powerpc/mm/book3s64/iommu_api.c
175
unpin_user_pages(mem->hpages, pinned);
arch/powerpc/mm/book3s64/iommu_api.c
177
vfree(mem->hpas);
arch/powerpc/mm/book3s64/iommu_api.c
178
kfree(mem);
arch/powerpc/mm/book3s64/iommu_api.c
202
static void mm_iommu_unpin(struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
207
if (!mem->hpas)
arch/powerpc/mm/book3s64/iommu_api.c
210
for (i = 0; i < mem->entries; ++i) {
arch/powerpc/mm/book3s64/iommu_api.c
211
if (!mem->hpas[i])
arch/powerpc/mm/book3s64/iommu_api.c
214
page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT);
arch/powerpc/mm/book3s64/iommu_api.c
218
if (mem->hpas[i] & MM_IOMMU_TABLE_GROUP_PAGE_DIRTY)
arch/powerpc/mm/book3s64/iommu_api.c
223
mem->hpas[i] = 0;
arch/powerpc/mm/book3s64/iommu_api.c
227
static void mm_iommu_do_free(struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
230
mm_iommu_unpin(mem);
arch/powerpc/mm/book3s64/iommu_api.c
231
vfree(mem->hpas);
arch/powerpc/mm/book3s64/iommu_api.c
232
kfree(mem);
arch/powerpc/mm/book3s64/iommu_api.c
237
struct mm_iommu_table_group_mem_t *mem = container_of(head,
arch/powerpc/mm/book3s64/iommu_api.c
240
mm_iommu_do_free(mem);
arch/powerpc/mm/book3s64/iommu_api.c
243
static void mm_iommu_release(struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
245
list_del_rcu(&mem->next);
arch/powerpc/mm/book3s64/iommu_api.c
246
call_rcu(&mem->rcu, mm_iommu_free);
arch/powerpc/mm/book3s64/iommu_api.c
249
long mm_iommu_put(struct mm_struct *mm, struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
256
if (mem->used == 0) {
arch/powerpc/mm/book3s64/iommu_api.c
261
--mem->used;
arch/powerpc/mm/book3s64/iommu_api.c
263
if (mem->used)
arch/powerpc/mm/book3s64/iommu_api.c
267
if (atomic64_cmpxchg(&mem->mapped, 1, 0) != 1) {
arch/powerpc/mm/book3s64/iommu_api.c
268
++mem->used;
arch/powerpc/mm/book3s64/iommu_api.c
273
if (mem->dev_hpa == MM_IOMMU_TABLE_INVALID_HPA)
arch/powerpc/mm/book3s64/iommu_api.c
274
unlock_entries = mem->entries;
arch/powerpc/mm/book3s64/iommu_api.c
277
mm_iommu_release(mem);
arch/powerpc/mm/book3s64/iommu_api.c
291
struct mm_iommu_table_group_mem_t *mem, *ret = NULL;
arch/powerpc/mm/book3s64/iommu_api.c
294
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
arch/powerpc/mm/book3s64/iommu_api.c
295
if ((mem->ua <= ua) &&
arch/powerpc/mm/book3s64/iommu_api.c
296
(ua + size <= mem->ua +
arch/powerpc/mm/book3s64/iommu_api.c
297
(mem->entries << PAGE_SHIFT))) {
arch/powerpc/mm/book3s64/iommu_api.c
298
ret = mem;
arch/powerpc/mm/book3s64/iommu_api.c
311
struct mm_iommu_table_group_mem_t *mem, *ret = NULL;
arch/powerpc/mm/book3s64/iommu_api.c
315
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next,
arch/powerpc/mm/book3s64/iommu_api.c
317
if ((mem->ua == ua) && (mem->entries == entries)) {
arch/powerpc/mm/book3s64/iommu_api.c
318
ret = mem;
arch/powerpc/mm/book3s64/iommu_api.c
319
++mem->used;
arch/powerpc/mm/book3s64/iommu_api.c
330
long mm_iommu_ua_to_hpa(struct mm_iommu_table_group_mem_t *mem,
arch/powerpc/mm/book3s64/iommu_api.c
333
const long entry = (ua - mem->ua) >> PAGE_SHIFT;
arch/powerpc/mm/book3s64/iommu_api.c
336
if (entry >= mem->entries)
arch/powerpc/mm/book3s64/iommu_api.c
339
if (pageshift > mem->pageshift)
arch/powerpc/mm/book3s64/iommu_api.c
342
if (!mem->hpas) {
arch/powerpc/mm/book3s64/iommu_api.c
343
*hpa = mem->dev_hpa + (ua - mem->ua);
arch/powerpc/mm/book3s64/iommu_api.c
347
va = &mem->hpas[entry];
arch/powerpc/mm/book3s64/iommu_api.c
357
struct mm_iommu_table_group_mem_t *mem;
arch/powerpc/mm/book3s64/iommu_api.c
361
list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
arch/powerpc/mm/book3s64/iommu_api.c
362
if (mem->dev_hpa == MM_IOMMU_TABLE_INVALID_HPA)
arch/powerpc/mm/book3s64/iommu_api.c
365
end = mem->dev_hpa + (mem->entries << PAGE_SHIFT);
arch/powerpc/mm/book3s64/iommu_api.c
366
if ((mem->dev_hpa <= hpa) && (hpa < end)) {
arch/powerpc/mm/book3s64/iommu_api.c
383
long mm_iommu_mapped_inc(struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
385
if (atomic64_inc_not_zero(&mem->mapped))
arch/powerpc/mm/book3s64/iommu_api.c
393
void mm_iommu_mapped_dec(struct mm_iommu_table_group_mem_t *mem)
arch/powerpc/mm/book3s64/iommu_api.c
395
atomic64_add_unless(&mem->mapped, -1, 1);
arch/powerpc/mm/book3s64/iommu_api.c
60
struct mm_iommu_table_group_mem_t *mem, *mem2;
arch/powerpc/mm/book3s64/iommu_api.c
73
mem = kzalloc_obj(*mem);
arch/powerpc/mm/book3s64/iommu_api.c
74
if (!mem) {
arch/powerpc/mm/book3s64/iommu_api.c
80
mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT));
arch/powerpc/mm/book3s64/iommu_api.c
81
mem->dev_hpa = dev_hpa;
arch/powerpc/mm/book3s64/iommu_api.c
84
mem->dev_hpa = MM_IOMMU_TABLE_INVALID_HPA;
arch/powerpc/mm/book3s64/iommu_api.c
91
mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT));
arch/powerpc/mm/book3s64/iommu_api.c
92
mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0])));
arch/powerpc/mm/book3s64/iommu_api.c
93
if (!mem->hpas) {
arch/powerpc/mm/book3s64/iommu_api.c
94
kfree(mem);
arch/powerpc/mm/numa.c
111
mem = memparse(p, &p);
arch/powerpc/mm/numa.c
112
if (!mem)
arch/powerpc/mm/numa.c
115
if (mem < curr_boundary)
arch/powerpc/mm/numa.c
118
curr_boundary = mem;
arch/powerpc/mm/numa.c
120
if ((end_pfn << PAGE_SHIFT) > mem) {
arch/powerpc/mm/numa.c
92
unsigned long long mem;
arch/powerpc/perf/imc-pmu.c
1277
static int trace_imc_prepare_sample(struct trace_imc_data *mem,
arch/powerpc/perf/imc-pmu.c
1284
if (be64_to_cpu(READ_ONCE(mem->tb1)) > *prev_tb)
arch/powerpc/perf/imc-pmu.c
1285
*prev_tb = be64_to_cpu(READ_ONCE(mem->tb1));
arch/powerpc/perf/imc-pmu.c
1289
if ((be64_to_cpu(READ_ONCE(mem->tb1)) & IMC_TRACE_RECORD_TB1_MASK) !=
arch/powerpc/perf/imc-pmu.c
1290
be64_to_cpu(READ_ONCE(mem->tb2)))
arch/powerpc/perf/imc-pmu.c
1294
data->ip = be64_to_cpu(READ_ONCE(mem->ip));
arch/powerpc/perf/imc-pmu.c
1302
switch (IMC_TRACE_RECORD_VAL_HVPR(be64_to_cpu(READ_ONCE(mem->val)))) {
arch/powerpc/perf/imc-pmu.c
1332
struct trace_imc_data *mem;
arch/powerpc/perf/imc-pmu.c
1336
mem = (struct trace_imc_data *)get_trace_imc_event_base_addr();
arch/powerpc/perf/imc-pmu.c
1338
i++, mem++) {
arch/powerpc/perf/imc-pmu.c
1342
ret = trace_imc_prepare_sample(mem, &data, &prev_tb, &header, event);
arch/powerpc/perf/power10-pmu.c
127
GENERIC_EVENT_ATTR(mem-loads, MEM_LOADS);
arch/powerpc/perf/power10-pmu.c
128
GENERIC_EVENT_ATTR(mem-stores, MEM_STORES);
arch/powerpc/perf/power9-pmu.c
174
GENERIC_EVENT_ATTR(mem-loads, MEM_LOADS);
arch/powerpc/perf/power9-pmu.c
175
GENERIC_EVENT_ATTR(mem-stores, MEM_STORES);
arch/powerpc/platforms/44x/hsta_msi.c
125
struct resource *mem;
arch/powerpc/platforms/44x/hsta_msi.c
129
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
arch/powerpc/platforms/44x/hsta_msi.c
130
if (!mem) {
arch/powerpc/platforms/44x/hsta_msi.c
142
ppc4xx_hsta_msi.address = mem->start;
arch/powerpc/platforms/44x/hsta_msi.c
143
ppc4xx_hsta_msi.data = ioremap(mem->start, resource_size(mem));
arch/powerpc/platforms/powermac/bootx_init.c
110
#define dt_push_token(token, mem) \
arch/powerpc/platforms/powermac/bootx_init.c
112
*(mem) = ALIGN(*(mem),4); \
arch/powerpc/platforms/powermac/bootx_init.c
113
*((u32 *)*(mem)) = token; \
arch/powerpc/platforms/powermac/bootx_init.c
114
*(mem) += 4; \
arch/powerpc/platforms/powernv/memtrace.c
176
ent->mem = ioremap(ent->start, ent->size);
arch/powerpc/platforms/powernv/memtrace.c
178
if (!ent->mem) {
arch/powerpc/platforms/powernv/memtrace.c
233
if (ent->mem) {
arch/powerpc/platforms/powernv/memtrace.c
234
iounmap(ent->mem);
arch/powerpc/platforms/powernv/memtrace.c
235
ent->mem = 0;
arch/powerpc/platforms/powernv/memtrace.c
25
void *mem;
arch/powerpc/platforms/powernv/memtrace.c
45
return simple_read_from_buffer(ubuf, count, ppos, ent->mem, ent->size);
arch/powerpc/platforms/pseries/ibmebus.c
73
void *mem;
arch/powerpc/platforms/pseries/ibmebus.c
75
mem = kmalloc(size, flag);
arch/powerpc/platforms/pseries/ibmebus.c
76
*dma_handle = (dma_addr_t)mem;
arch/powerpc/platforms/pseries/ibmebus.c
78
return mem;
arch/powerpc/sysdev/fsl_pci.c
203
u64 mem, sz, paddr_hi = 0;
arch/powerpc/sysdev/fsl_pci.c
321
mem = memblock_end_of_DRAM();
arch/powerpc/sysdev/fsl_pci.c
322
pr_info("%s: end of DRAM %llx\n", __func__, mem);
arch/powerpc/sysdev/fsl_pci.c
340
if ((address >= mem) && (address < (mem + PAGE_SIZE))) {
arch/powerpc/sysdev/fsl_pci.c
342
mem += PAGE_SIZE;
arch/powerpc/sysdev/fsl_pci.c
350
sz = min(mem, paddr_lo);
arch/powerpc/sysdev/fsl_pci.c
356
if ((1ull << mem_log) != mem) {
arch/powerpc/sysdev/fsl_pci.c
358
if ((1ull << mem_log) > mem)
arch/powerpc/sysdev/fsl_pci.c
381
if (sz != mem) {
arch/powerpc/sysdev/fsl_pci.c
382
mem_log = ilog2(mem);
arch/powerpc/sysdev/fsl_pci.c
385
if ((1ull << mem_log) != mem)
arch/powerpc/sysdev/fsl_pci.c
444
if (hose->dma_window_size < mem) {
arch/riscv/kernel/kexec_elf.c
103
*new_pbase = kbuf.mem;
arch/riscv/kernel/kexec_elf.c
104
image->start = ehdr->e_entry - lowest_vaddr + kbuf.mem;
arch/riscv/kernel/kexec_elf.c
48
kbuf.mem = phdr->p_paddr - old_pbase + new_pbase;
arch/riscv/kernel/kexec_elf.c
96
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/kexec_image.c
71
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/kexec_image.c
81
image->start = kbuf.mem;
arch/riscv/kernel/kexec_image.c
84
kbuf.mem, kbuf.bufsz, kbuf.memsz);
arch/riscv/kernel/kexec_image.c
86
ret = load_extra_segments(image, kbuf.mem, kbuf.memsz,
arch/riscv/kernel/kgdb.c
207
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/riscv/kernel/kgdb.c
213
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/riscv/kernel/kgdb.c
216
memset(mem, 0, dbg_reg_def[regno].size);
arch/riscv/kernel/kgdb.c
220
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/riscv/kernel/kgdb.c
226
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/riscv/kernel/machine_kexec.c
52
internal->fdt_addr = (unsigned long) image->segment[i].mem;
arch/riscv/kernel/machine_kexec_file.c
284
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/machine_kexec_file.c
295
image->elf_load_addr = kbuf.mem;
arch/riscv/kernel/machine_kexec_file.c
316
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/machine_kexec_file.c
322
kexec_dprintk("Loaded purgatory at 0x%lx\n", kbuf.mem);
arch/riscv/kernel/machine_kexec_file.c
337
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/machine_kexec_file.c
341
initrd_pbase = kbuf.mem;
arch/riscv/kernel/machine_kexec_file.c
358
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/riscv/kernel/machine_kexec_file.c
367
kexec_dprintk("Loaded device tree at 0x%lx\n", kbuf.mem);
arch/riscv/kernel/setup.c
239
struct resource *mem = &standard_resources[i];
arch/riscv/kernel/setup.c
240
phys_addr_t r_start, r_end, mem_size = resource_size(mem);
arch/riscv/kernel/setup.c
242
if (!memblock_is_region_reserved(mem->start, mem_size))
arch/riscv/kernel/setup.c
248
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/riscv/kernel/setup.c
249
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/riscv/kernel/setup.c
251
if (start > mem->end || end < mem->start)
arch/riscv/kernel/setup.c
254
reserve_region_with_split(mem, start, end, "Reserved");
arch/s390/include/asm/sclp.h
177
int sclp_early_get_memsize(unsigned long *mem);
arch/s390/kernel/early.c
51
decompressor_handled_param(mem);
arch/s390/kernel/ipl.c
2402
comp->entry.addr = kbuf->mem;
arch/s390/kernel/kexec_elf.c
42
buf.mem = ALIGN(phdr->p_paddr, phdr->p_align);
arch/s390/kernel/kexec_elf.c
45
buf.mem += crashk_res.start;
arch/s390/kernel/kexec_elf.c
52
data->kernel_mem = buf.mem;
arch/s390/kernel/kexec_image.c
26
buf.mem = 0;
arch/s390/kernel/kexec_image.c
29
buf.mem += crashk_res.start;
arch/s390/kernel/kexec_image.c
34
data->kernel_mem = buf.mem;
arch/s390/kernel/machine_kexec_file.c
138
buf.mem = data->memsz;
arch/s390/kernel/machine_kexec_file.c
141
buf.mem += crashk_res.start;
arch/s390/kernel/machine_kexec_file.c
164
buf.mem = data->memsz;
arch/s390/kernel/machine_kexec_file.c
167
buf.mem += crashk_res.start;
arch/s390/kernel/machine_kexec_file.c
195
buf.mem = data->memsz;
arch/s390/kernel/machine_kexec_file.c
230
*lc_ipl_parmblock_ptr = (__u32)buf.mem;
arch/s390/kernel/machine_kexec_file.c
234
buf.mem += crashk_res.start;
arch/s390/kernel/module.c
164
mod_mem = &me->mem[MOD_TEXT];
arch/s390/kernel/module.c
296
Elf_Addr *gotent = me->mem[MOD_TEXT].base +
arch/s390/kernel/module.c
321
val += (Elf_Addr)me->mem[MOD_TEXT].base +
arch/s390/kernel/module.c
338
plt_base = me->mem[MOD_TEXT].base + me->arch.plt_offset;
arch/s390/kernel/module.c
368
val = (Elf_Addr) me->mem[MOD_TEXT].base +
arch/s390/kernel/module.c
390
((Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset);
arch/s390/kernel/module.c
400
val = (Elf_Addr) me->mem[MOD_TEXT].base + me->arch.got_offset +
arch/s390/kernel/module.c
506
ij = me->mem[MOD_TEXT].base + me->arch.plt_offset +
arch/s390/kvm/priv.c
822
static void handle_stsi_3_2_2(struct kvm_vcpu *vcpu, struct sysinfo_3_2_2 *mem)
arch/s390/kvm/priv.c
830
if (stsi(mem, 3, 2, 2))
arch/s390/kvm/priv.c
831
mem->count = 0;
arch/s390/kvm/priv.c
832
if (mem->count < 8)
arch/s390/kvm/priv.c
833
mem->count++;
arch/s390/kvm/priv.c
834
for (n = mem->count - 1; n > 0 ; n--)
arch/s390/kvm/priv.c
835
memcpy(&mem->vm[n], &mem->vm[n - 1], sizeof(mem->vm[0]));
arch/s390/kvm/priv.c
837
memset(&mem->vm[0], 0, sizeof(mem->vm[0]));
arch/s390/kvm/priv.c
838
mem->vm[0].cpus_total = cpus;
arch/s390/kvm/priv.c
839
mem->vm[0].cpus_configured = cpus;
arch/s390/kvm/priv.c
840
mem->vm[0].cpus_standby = 0;
arch/s390/kvm/priv.c
841
mem->vm[0].cpus_reserved = 0;
arch/s390/kvm/priv.c
842
mem->vm[0].caf = 1000;
arch/s390/kvm/priv.c
843
memcpy(mem->vm[0].name, "KVMguest", 8);
arch/s390/kvm/priv.c
844
ASCEBC(mem->vm[0].name, 8);
arch/s390/kvm/priv.c
845
memcpy(mem->vm[0].cpi, "KVM/Linux ", 16);
arch/s390/kvm/priv.c
846
ASCEBC(mem->vm[0].cpi, 16);
arch/s390/kvm/priv.c
865
unsigned long mem = 0;
arch/s390/kvm/priv.c
907
mem = get_zeroed_page(GFP_KERNEL_ACCOUNT);
arch/s390/kvm/priv.c
908
if (!mem)
arch/s390/kvm/priv.c
910
if (stsi((void *) mem, fc, sel1, sel2))
arch/s390/kvm/priv.c
916
mem = get_zeroed_page(GFP_KERNEL_ACCOUNT);
arch/s390/kvm/priv.c
917
if (!mem)
arch/s390/kvm/priv.c
919
handle_stsi_3_2_2(vcpu, (void *) mem);
arch/s390/kvm/priv.c
927
memcpy(sida_addr(vcpu->arch.sie_block), (void *)mem, PAGE_SIZE);
arch/s390/kvm/priv.c
930
rc = write_guest(vcpu, operand2, ar, (void *)mem, PAGE_SIZE);
arch/s390/kvm/priv.c
941
free_page(mem);
arch/s390/kvm/priv.c
948
free_page(mem);
arch/s390/tools/relocs.c
255
void *mem = realloc(r->offset, newsize * sizeof(r->offset[0]));
arch/s390/tools/relocs.c
257
if (!mem)
arch/s390/tools/relocs.c
260
r->offset = mem;
arch/sh/include/asm/io.h
102
static inline void pfx##reads##bwlq(const volatile void __iomem *mem, \
arch/sh/include/asm/io.h
108
*__addr = __raw_read##bwlq(mem); \
arch/sh/include/asm/io.h
91
pfx##writes##bwlq(volatile void __iomem *mem, const void *addr, \
arch/sh/include/asm/io.h
97
__raw_write##bwlq(*__addr, mem); \
arch/sh/kernel/kgdb.c
193
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/sh/kernel/kgdb.c
199
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/sh/kernel/kgdb.c
205
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/sh/kernel/kgdb.c
211
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/sh/kernel/kgdb.c
216
__asm__ __volatile__ ("stc vbr, %0" : "=r" (mem));
arch/sh/kernel/machine_kexec.c
57
(unsigned int)image->segment[i].mem,
arch/sh/kernel/machine_kexec.c
58
(unsigned int)image->segment[i].mem +
arch/um/include/asm/uaccess.h
23
extern unsigned long __clear_user(void __user *mem, unsigned long len);
arch/um/include/shared/skas/stub-data.h
50
} mem;
arch/um/kernel/skas/stub.c
47
fd = fd_map[sc->mem.fd];
arch/um/kernel/skas/stub.c
49
fd = sc->mem.fd;
arch/um/kernel/skas/stub.c
52
sc->mem.addr, sc->mem.length,
arch/um/kernel/skas/stub.c
53
sc->mem.prot,
arch/um/kernel/skas/stub.c
55
fd, sc->mem.offset);
arch/um/kernel/skas/stub.c
56
if (res != sc->mem.addr) {
arch/um/kernel/skas/stub.c
64
sc->mem.addr, sc->mem.length);
arch/um/kernel/skas/uaccess.c
203
unsigned long __clear_user(void __user *mem, unsigned long len)
arch/um/kernel/skas/uaccess.c
205
return buffer_op((unsigned long) mem, len, 1, clear_chunk, NULL);
arch/um/os-Linux/skas/mem.c
189
sc->mem.addr + sc->mem.length == virt)
arch/um/os-Linux/skas/mem.c
241
if (sc && sc->mem.prot == prot &&
arch/um/os-Linux/skas/mem.c
242
sc->mem.offset == MMAP_OFFSET(offset - sc->mem.length)) {
arch/um/os-Linux/skas/mem.c
243
int prev_fd = sc->mem.fd;
arch/um/os-Linux/skas/mem.c
246
prev_fd = mm_idp->syscall_fd_map[sc->mem.fd];
arch/um/os-Linux/skas/mem.c
249
sc->mem.length += len;
arch/um/os-Linux/skas/mem.c
258
sc->mem.addr = virt;
arch/um/os-Linux/skas/mem.c
259
sc->mem.length = len;
arch/um/os-Linux/skas/mem.c
260
sc->mem.prot = prot;
arch/um/os-Linux/skas/mem.c
261
sc->mem.fd = phys_fd;
arch/um/os-Linux/skas/mem.c
262
sc->mem.offset = MMAP_OFFSET(offset);
arch/um/os-Linux/skas/mem.c
274
sc->mem.length += len;
arch/um/os-Linux/skas/mem.c
280
sc->mem.addr = addr;
arch/um/os-Linux/skas/mem.c
281
sc->mem.length = len;
arch/x86/coco/sev/core.c
138
void __iomem *mem;
arch/x86/coco/sev/core.c
141
mem = ioremap_encrypted(sev_secrets_pa, PAGE_SIZE);
arch/x86/coco/sev/core.c
142
if (!mem) {
arch/x86/coco/sev/core.c
147
secrets = (__force struct snp_secrets_page *)mem;
arch/x86/coco/sev/core.c
150
iounmap(mem);
arch/x86/coco/sev/core.c
1590
void __iomem *mem;
arch/x86/coco/sev/core.c
1598
mem = ioremap_encrypted(sev_secrets_pa, PAGE_SIZE);
arch/x86/coco/sev/core.c
1599
if (!mem)
arch/x86/coco/sev/core.c
1602
mdesc->secrets = (__force struct snp_secrets_page *)mem;
arch/x86/coco/sev/core.c
1618
iounmap(mem);
arch/x86/coco/sev/core.c
2030
void *mem;
arch/x86/coco/sev/core.c
2035
mem = early_memremap_encrypted(sev_secrets_pa, PAGE_SIZE);
arch/x86/coco/sev/core.c
2036
if (!mem) {
arch/x86/coco/sev/core.c
2041
secrets = (__force struct snp_secrets_page *)mem;
arch/x86/coco/sev/core.c
2054
early_memunmap(mem, PAGE_SIZE);
arch/x86/events/intel/core.c
2385
EVENT_ATTR_STR(mem-loads, mem_ld_grt, "event=0xd0,umask=0x5,ldlat=3");
arch/x86/events/intel/core.c
2386
EVENT_ATTR_STR(mem-stores, mem_st_grt, "event=0xd0,umask=0x6");
arch/x86/events/intel/core.c
517
EVENT_ATTR_STR(mem-loads, mem_ld_nhm, "event=0x0b,umask=0x10,ldlat=3");
arch/x86/events/intel/core.c
518
EVENT_ATTR_STR(mem-loads, mem_ld_snb, "event=0xcd,umask=0x1,ldlat=3");
arch/x86/events/intel/core.c
519
EVENT_ATTR_STR(mem-stores, mem_st_snb, "event=0xcd,umask=0x2");
arch/x86/events/intel/core.c
564
EVENT_ATTR_STR(topdown-mem-bound, td_mem_bound, "event=0x00,umask=0x87");
arch/x86/events/intel/core.c
6619
EVENT_ATTR_STR(mem-loads, mem_ld_hsw, "event=0xcd,umask=0x1,ldlat=3");
arch/x86/events/intel/core.c
6620
EVENT_ATTR_STR(mem-stores, mem_st_hsw, "event=0xd0,umask=0x82")
arch/x86/events/intel/core.c
6708
EVENT_ATTR_STR(mem-stores, mem_st_spr, "event=0xcd,umask=0x2");
arch/x86/events/intel/core.c
6709
EVENT_ATTR_STR(mem-loads-aux, mem_ld_aux, "event=0x03,umask=0x82");
arch/x86/events/intel/core.c
7030
EVENT_ATTR_STR_HYBRID(topdown-mem-bound, td_mem_bound_adl, "event=0x00,umask=0x87", hybrid_big);
arch/x86/events/intel/core.c
7094
EVENT_ATTR_STR_HYBRID(mem-loads, mem_ld_adl, "event=0xd0,umask=0x5,ldlat=3;event=0xcd,umask=0x1,ldlat=3", hybrid_big_small);
arch/x86/events/intel/core.c
7095
EVENT_ATTR_STR_HYBRID(mem-stores, mem_st_adl, "event=0xd0,umask=0x6;event=0xcd,umask=0x2", hybrid_big_small);
arch/x86/events/intel/core.c
7096
EVENT_ATTR_STR_HYBRID(mem-loads-aux, mem_ld_aux_adl, "event=0x03,umask=0x82", hybrid_big);
arch/x86/events/intel/core.c
7111
EVENT_ATTR_STR_HYBRID(mem-loads,
arch/x86/events/intel/core.c
7115
EVENT_ATTR_STR_HYBRID(mem-stores,
arch/x86/include/asm/efi.h
402
void *buf, struct efi_mem_range *mem);
arch/x86/include/asm/realmode.h
89
static inline void set_real_mode_mem(phys_addr_t mem)
arch/x86/include/asm/realmode.h
91
real_mode_header = (struct real_mode_header *) __va(mem);
arch/x86/include/asm/uaccess_32.h
36
unsigned long __must_check clear_user(void __user *mem, unsigned long len);
arch/x86/include/asm/uaccess_32.h
37
unsigned long __must_check __clear_user(void __user *mem, unsigned long len);
arch/x86/include/asm/uv/uv_geo.h
67
struct geo_mem_s mem;
arch/x86/include/asm/word-at-a-time.h
79
: [mem] "m" (*(unsigned long *)addr));
arch/x86/kernel/alternative.c
2377
: [mem] "m" (__alt_reloc_selftest_addr)
arch/x86/kernel/apic/io_apic.c
2506
char *mem;
arch/x86/kernel/apic/io_apic.c
2515
mem = memblock_alloc_or_panic(n, SMP_CACHE_BYTES);
arch/x86/kernel/apic/io_apic.c
2516
res = (void *)mem;
arch/x86/kernel/apic/io_apic.c
2518
mem += sizeof(struct resource) * nr_ioapics;
arch/x86/kernel/apic/io_apic.c
2521
res[i].name = mem;
arch/x86/kernel/apic/io_apic.c
2523
snprintf(mem, IOAPIC_RESOURCE_NAME_SIZE, "IOAPIC %u", i);
arch/x86/kernel/apic/io_apic.c
2524
mem += IOAPIC_RESOURCE_NAME_SIZE;
arch/x86/kernel/callthunks.c
327
.base = (unsigned long)mod->mem[MOD_TEXT].base,
arch/x86/kernel/callthunks.c
328
.end = (unsigned long)mod->mem[MOD_TEXT].base + mod->mem[MOD_TEXT].size,
arch/x86/kernel/crash.c
453
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/x86/kernel/crash.c
457
image->elf_load_addr = kbuf.mem;
arch/x86/kernel/crash.c
515
unsigned long mem, memsz;
arch/x86/kernel/crash.c
541
mem = image->segment[image->elfcorehdr_index].mem;
arch/x86/kernel/crash.c
552
old_elfcorehdr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
arch/x86/kernel/kexec-bzimage64.c
294
kho->scratch_addr = image->kho.scratch->mem;
arch/x86/kernel/kexec-bzimage64.c
553
kexec_dprintk("Loaded purgatory at 0x%lx\n", pbuf.mem);
arch/x86/kernel/kexec-bzimage64.c
609
bootparam_load_addr = kbuf.mem;
arch/x86/kernel/kexec-bzimage64.c
622
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/x86/kernel/kexec-bzimage64.c
626
kernel_load_addr = kbuf.mem;
arch/x86/kernel/kexec-bzimage64.c
637
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
arch/x86/kernel/kexec-bzimage64.c
641
initrd_load_addr = kbuf.mem;
arch/x86/kernel/kgdb.c
101
memcpy((void *)regs + dbg_reg_def[regno].offset, mem,
arch/x86/kernel/kgdb.c
106
char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs)
arch/x86/kernel/kgdb.c
109
memcpy(mem, &regs->orig_ax, sizeof(regs->orig_ax));
arch/x86/kernel/kgdb.c
116
memcpy(mem, (void *)regs + dbg_reg_def[regno].offset,
arch/x86/kernel/kgdb.c
123
*(unsigned long *)mem = 0xFFFF;
arch/x86/kernel/kgdb.c
91
int dbg_set_reg(int regno, void *mem, struct pt_regs *regs)
arch/x86/kernel/machine_kexec_64.c
280
mstart = image->segment[i].mem;
arch/x86/kernel/module.c
320
void *text = me->mem[MOD_TEXT].base;
arch/x86/kernel/module.c
321
void *text_end = text + me->mem[MOD_TEXT].size;
arch/x86/kvm/emulate.c
1243
op->addr.mem.ea = modrm_ea;
arch/x86/kvm/emulate.c
1245
ctxt->memop.addr.mem.ea = (u32)ctxt->memop.addr.mem.ea;
arch/x86/kvm/emulate.c
1259
op->addr.mem.ea = insn_fetch(u16, ctxt);
arch/x86/kvm/emulate.c
1262
op->addr.mem.ea = insn_fetch(u32, ctxt);
arch/x86/kvm/emulate.c
1265
op->addr.mem.ea = insn_fetch(u64, ctxt);
arch/x86/kvm/emulate.c
1286
ctxt->dst.addr.mem.ea = address_mask(ctxt,
arch/x86/kvm/emulate.c
1287
ctxt->dst.addr.mem.ea + (sv >> 3));
arch/x86/kvm/emulate.c
1768
op->addr.mem,
arch/x86/kvm/emulate.c
1774
op->addr.mem,
arch/x86/kvm/emulate.c
1779
op->addr.mem,
arch/x86/kvm/emulate.c
3002
op->addr.mem.ea = register_address(ctxt, reg);
arch/x86/kvm/emulate.c
3400
rc = __linearize(ctxt, ctxt->src.addr.mem, &max_size, 1, ctxt->mode,
arch/x86/kvm/emulate.c
3452
return segmented_write_std(ctxt, ctxt->dst.addr.mem,
arch/x86/kvm/emulate.c
3473
rc = read_descriptor(ctxt, ctxt->src.addr.mem,
arch/x86/kvm/emulate.c
3727
return segmented_write_std(ctxt, ctxt->memop.addr.mem, &fx_state,
arch/x86/kvm/emulate.c
3762
rc = segmented_read_std(ctxt, ctxt->memop.addr.mem, &fx_state, size);
arch/x86/kvm/emulate.c
4571
op->addr.mem.ea = ctxt->_eip;
arch/x86/kvm/emulate.c
4653
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4655
op->addr.mem.seg = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4707
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4709
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
4716
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4720
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
4725
op->addr.mem.ea = ctxt->_eip;
arch/x86/kvm/emulate.c
5171
ctxt->memop.addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
5193
ctxt->memopp->addr.mem.ea = address_mask(ctxt,
arch/x86/kvm/emulate.c
5194
ctxt->memopp->addr.mem.ea + ctxt->_eip);
arch/x86/kvm/emulate.c
5378
rc = segmented_read(ctxt, ctxt->src.addr.mem,
arch/x86/kvm/emulate.c
5386
rc = segmented_read(ctxt, ctxt->src2.addr.mem,
arch/x86/kvm/emulate.c
5398
rc = segmented_read(ctxt, ctxt->dst.addr.mem,
arch/x86/kvm/emulate.c
5443
ctxt->dst.val = ctxt->src.addr.mem.ea;
arch/x86/kvm/kvm_emulate.h
264
} mem;
arch/x86/lib/iomem.c
89
volatile char __iomem *mem = a;
arch/x86/lib/iomem.c
93
writeb(b, &mem[i]);
arch/x86/platform/efi/memmap.c
163
struct efi_mem_range *mem)
arch/x86/platform/efi/memmap.c
171
m_start = mem->range.start;
arch/x86/platform/efi/memmap.c
172
m_end = mem->range.end;
arch/x86/platform/efi/memmap.c
173
m_attr = mem->attribute;
arch/x86/platform/olpc/olpc_dt.c
126
static u8 *mem;
arch/x86/platform/olpc/olpc_dt.c
143
mem = res;
arch/x86/platform/olpc/olpc_dt.c
148
res = mem;
arch/x86/platform/olpc/olpc_dt.c
149
mem += size;
arch/x86/realmode/init.c
49
phys_addr_t mem;
arch/x86/realmode/init.c
58
mem = memblock_phys_alloc_range(size, PAGE_SIZE, 0, 1<<20);
arch/x86/realmode/init.c
59
if (!mem)
arch/x86/realmode/init.c
62
set_real_mode_mem(mem);
arch/x86/tools/relocs.c
685
void *mem = realloc(r->offset, newsize * sizeof(r->offset[0]));
arch/x86/tools/relocs.c
687
if (!mem)
arch/x86/tools/relocs.c
690
r->offset = mem;
arch/xtensa/include/asm/atomic.h
131
[mem] "+m" (*v) \
arch/xtensa/include/asm/atomic.h
151
[mem] "+m" (*v) \
arch/xtensa/include/asm/atomic.h
172
[mem] "+m" (*v) \
arch/xtensa/include/asm/atomic.h
194
: [result] "=&a" (vval), [mem] "+m" (*v) \
arch/xtensa/include/asm/atomic.h
212
: [result] "=&a" (vval), [mem] "+m" (*v) \
arch/xtensa/include/asm/atomic.h
233
[mem] "+m" (*v) \
arch/xtensa/include/asm/bitops.h
159
[mem] "+m" (*p) \
arch/xtensa/include/asm/bitops.h
180
[mem] "+m" (*p) \
arch/xtensa/include/asm/cmpxchg.h
154
[mem] "+m" (*m)
arch/xtensa/include/asm/cmpxchg.h
167
: [tmp] "=&a" (tmp), [mem] "+m" (*m)
arch/xtensa/include/asm/cmpxchg.h
48
: [new] "+a" (new), [mem] "+m" (*p)
arch/xtensa/include/asm/cmpxchg.h
63
: [old] "=&a" (old), [mem] "+m" (*p)
arch/xtensa/include/asm/futex.h
68
[mem] "+m" (*(uaddr)) \
arch/xtensa/include/asm/uaccess.h
146
:[err] "+r"(err_), [tmp] "=r"(cb), [mem] "=m"(*(addr_)) \
arch/xtensa/include/asm/uaccess.h
217
:[mem] "m"(*(addr_)), [efault] "i"(-EFAULT)); \
block/blk-mq-dma.c
82
iter->addr = pci_p2pdma_bus_addr_map(iter->p2pdma.mem, vec->paddr);
crypto/api.c
510
char *mem;
crypto/api.c
515
mem = kzalloc_node(total, gfp, node);
crypto/api.c
516
if (mem == NULL)
crypto/api.c
519
tfm = (struct crypto_tfm *)(mem + tfmsize);
crypto/api.c
524
return mem;
crypto/api.c
532
char *mem;
crypto/api.c
535
mem = crypto_alloc_tfmmem(alg, frontend, node, GFP_KERNEL);
crypto/api.c
536
if (IS_ERR(mem))
crypto/api.c
539
tfm = (struct crypto_tfm *)(mem + frontend->tfmsize);
crypto/api.c
556
kfree(mem);
crypto/api.c
557
mem = ERR_PTR(err);
crypto/api.c
559
return mem;
crypto/api.c
568
char *mem;
crypto/api.c
570
mem = ERR_PTR(-ESTALE);
crypto/api.c
574
mem = crypto_alloc_tfmmem(alg, frontend, otfm->node, GFP_ATOMIC);
crypto/api.c
575
if (IS_ERR(mem)) {
crypto/api.c
580
tfm = (struct crypto_tfm *)(mem + frontend->tfmsize);
crypto/api.c
585
return mem;
crypto/api.c
671
void crypto_destroy_tfm(void *mem, struct crypto_tfm *tfm)
crypto/api.c
675
if (IS_ERR_OR_NULL(mem))
crypto/api.c
686
kfree_sensitive(mem);
crypto/jitterentropy.c
475
if (NULL == ec || NULL == ec->mem)
crypto/jitterentropy.c
487
unsigned char *tmpval = ec->mem + ec->memlocation;
crypto/jitterentropy.c
671
entropy_collector->mem = jent_kvzalloc(JENT_MEMORY_SIZE);
crypto/jitterentropy.c
672
if (!entropy_collector->mem) {
crypto/jitterentropy.c
702
jent_kvzfree(entropy_collector->mem, JENT_MEMORY_SIZE);
crypto/jitterentropy.c
703
entropy_collector->mem = NULL;
crypto/jitterentropy.c
82
unsigned char *mem; /* Memory access location with size of
drivers/accel/amdxdna/aie2_ctx.c
1002
if (abo->mem.map_invalid) {
drivers/accel/amdxdna/aie2_ctx.c
589
i, abo->mem.dev_addr, abo->mem.size);
drivers/accel/amdxdna/aie2_ctx.c
630
heap->mem.userptr, heap->mem.size);
drivers/accel/amdxdna/aie2_ctx.c
71
heap->mem.userptr, heap->mem.size);
drivers/accel/amdxdna/aie2_ctx.c
896
list_for_each_entry(mapp, &abo->mem.umap_list, node) {
drivers/accel/amdxdna/aie2_ctx.c
904
abo->mem.map_invalid = false;
drivers/accel/amdxdna/aie2_message.c
1013
size = cmdbuf_abo->mem.size - offset;
drivers/accel/amdxdna/aie2_message.c
1014
ret = aie2_cmdlist_fill_slot(cmdbuf_abo->mem.kva + offset,
drivers/accel/amdxdna/aie2_message.c
1027
EXEC_MSG_OPS(xdna)->init_chain_req(&req, cmdbuf_abo->mem.dev_addr,
drivers/accel/amdxdna/aie2_message.c
1029
drm_clflush_virt_range(cmdbuf_abo->mem.kva, offset);
drivers/accel/amdxdna/aie2_message.c
1058
size = cmdbuf_abo->mem.size;
drivers/accel/amdxdna/aie2_message.c
1059
ret = aie2_cmdlist_fill_slot(cmdbuf_abo->mem.kva, cmd_abo, &size, &op);
drivers/accel/amdxdna/aie2_message.c
1067
EXEC_MSG_OPS(xdna)->init_chain_req(&req, cmdbuf_abo->mem.dev_addr,
drivers/accel/amdxdna/aie2_message.c
1069
drm_clflush_virt_range(cmdbuf_abo->mem.kva, size);
drivers/accel/amdxdna/aie2_message.c
1096
req.size = abo->mem.size;
drivers/accel/amdxdna/aie2_message.c
1135
req.size = abo->mem.size;
drivers/accel/amdxdna/aie2_message.c
175
metadata->mem.row_count = resp.info.mem_rows;
drivers/accel/amdxdna/aie2_message.c
176
metadata->mem.row_start = resp.info.mem_row_start;
drivers/accel/amdxdna/aie2_message.c
177
metadata->mem.dma_channel_count = resp.info.mem_dma_channels;
drivers/accel/amdxdna/aie2_message.c
178
metadata->mem.lock_count = resp.info.mem_locks;
drivers/accel/amdxdna/aie2_message.c
179
metadata->mem.event_reg_count = resp.info.mem_events;
drivers/accel/amdxdna/aie2_message.c
535
abo->mem.dev_addr >> shift);
drivers/accel/amdxdna/aie2_message.c
538
abo->mem.dev_addr, req.cfgs[i]);
drivers/accel/amdxdna/aie2_pci.c
699
meta->mem.row_count = ndev->metadata.mem.row_count;
drivers/accel/amdxdna/aie2_pci.c
700
meta->mem.row_start = ndev->metadata.mem.row_start;
drivers/accel/amdxdna/aie2_pci.c
701
meta->mem.dma_channel_count = ndev->metadata.mem.dma_channel_count;
drivers/accel/amdxdna/aie2_pci.c
702
meta->mem.lock_count = ndev->metadata.mem.lock_count;
drivers/accel/amdxdna/aie2_pci.c
703
meta->mem.event_reg_count = ndev->metadata.mem.event_reg_count;
drivers/accel/amdxdna/aie2_pci.h
107
struct aie_tile_metadata mem;
drivers/accel/amdxdna/amdxdna_ctx.c
111
abo->mem.size)) {
drivers/accel/amdxdna/amdxdna_ctx.c
122
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.c
144
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.c
156
cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.c
159
memset(cmd->data, 0xff, abo->mem.size - sizeof(*cmd));
drivers/accel/amdxdna/amdxdna_ctx.c
98
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.h
146
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.h
154
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_ctx.h
163
struct amdxdna_cmd *cmd = abo->mem.kva;
drivers/accel/amdxdna/amdxdna_gem.c
118
abo->mem.map_invalid = true;
drivers/accel/amdxdna/amdxdna_gem.c
148
list_for_each_entry(mapp, &abo->mem.umap_list, node) {
drivers/accel/amdxdna/amdxdna_gem.c
235
if (abo->mem.userptr == AMDXDNA_INVALID_ADDR)
drivers/accel/amdxdna/amdxdna_gem.c
236
abo->mem.userptr = addr;
drivers/accel/amdxdna/amdxdna_gem.c
242
list_add_tail(&mapp->node, &abo->mem.umap_list);
drivers/accel/amdxdna/amdxdna_gem.c
259
XDNA_DBG(xdna, "BO type %d xdna_addr 0x%llx", abo->type, abo->mem.dev_addr);
drivers/accel/amdxdna/amdxdna_gem.c
31
struct amdxdna_mem *mem = &abo->mem;
drivers/accel/amdxdna/amdxdna_gem.c
411
if (!abo->mem.kva)
drivers/accel/amdxdna/amdxdna_gem.c
414
iosys_map_set_vaddr(&map, abo->mem.kva);
drivers/accel/amdxdna/amdxdna_gem.c
45
if (heap->mem.userptr == AMDXDNA_INVALID_ADDR) {
drivers/accel/amdxdna/amdxdna_gem.c
455
XDNA_DBG(xdna, "BO type %d xdna_addr 0x%llx", abo->type, abo->mem.dev_addr);
drivers/accel/amdxdna/amdxdna_gem.c
507
abo->mem.userptr = AMDXDNA_INVALID_ADDR;
drivers/accel/amdxdna/amdxdna_gem.c
508
abo->mem.dev_addr = AMDXDNA_INVALID_ADDR;
drivers/accel/amdxdna/amdxdna_gem.c
509
abo->mem.size = size;
drivers/accel/amdxdna/amdxdna_gem.c
51
if (mem->size == 0 || mem->size > heap->mem.size) {
drivers/accel/amdxdna/amdxdna_gem.c
510
INIT_LIST_HEAD(&abo->mem.umap_list);
drivers/accel/amdxdna/amdxdna_gem.c
53
mem->size, heap->mem.size);
drivers/accel/amdxdna/amdxdna_gem.c
60
mem->size, align,
drivers/accel/amdxdna/amdxdna_gem.c
67
mem->dev_addr = abo->mm_node.start;
drivers/accel/amdxdna/amdxdna_gem.c
68
offset = mem->dev_addr - heap->mem.dev_addr;
drivers/accel/amdxdna/amdxdna_gem.c
686
abo->mem.dev_addr = client->xdna->dev_info->dev_mem_base;
drivers/accel/amdxdna/amdxdna_gem.c
687
drm_mm_init(&abo->mm, abo->mem.dev_addr, abo->mem.size);
drivers/accel/amdxdna/amdxdna_gem.c
689
ret = amdxdna_gem_obj_vmap(abo, &abo->mem.kva);
drivers/accel/amdxdna/amdxdna_gem.c
69
mem->userptr = heap->mem.userptr + offset;
drivers/accel/amdxdna/amdxdna_gem.c
70
mem->kva = heap->mem.kva + offset;
drivers/accel/amdxdna/amdxdna_gem.c
801
args->handle, args->type, abo->mem.userptr,
drivers/accel/amdxdna/amdxdna_gem.c
802
abo->mem.dev_addr, abo->mem.size);
drivers/accel/amdxdna/amdxdna_gem.c
868
if (bo_type != AMDXDNA_BO_CMD || abo->mem.kva)
drivers/accel/amdxdna/amdxdna_gem.c
871
if (abo->mem.size > SZ_32K) {
drivers/accel/amdxdna/amdxdna_gem.c
872
XDNA_ERR(xdna, "Cmd bo is too big %ld", abo->mem.size);
drivers/accel/amdxdna/amdxdna_gem.c
876
ret = amdxdna_gem_obj_vmap(abo, &abo->mem.kva);
drivers/accel/amdxdna/amdxdna_gem.c
907
args->vaddr = abo->mem.userptr;
drivers/accel/amdxdna/amdxdna_gem.c
908
args->xdna_addr = abo->mem.dev_addr;
drivers/accel/amdxdna/amdxdna_gem.c
953
else if (abo->mem.kva)
drivers/accel/amdxdna/amdxdna_gem.c
954
drm_clflush_virt_range(abo->mem.kva + args->offset, args->size);
drivers/accel/amdxdna/amdxdna_gem.h
41
struct amdxdna_mem mem;
drivers/accel/amdxdna/amdxdna_gem.h
68
return abo->mem.dev_addr - abo->client->dev_heap->mem.dev_addr;
drivers/accel/ethosu/ethosu_gem.c
683
struct drm_gem_dma_object *mem;
drivers/accel/ethosu/ethosu_gem.c
686
mem = drm_gem_dma_create(ddev, size);
drivers/accel/ethosu/ethosu_gem.c
687
if (IS_ERR(mem))
drivers/accel/ethosu/ethosu_gem.c
688
return PTR_ERR(mem);
drivers/accel/ethosu/ethosu_gem.c
690
bo = to_ethosu_bo(&mem->base);
drivers/accel/ethosu/ethosu_gem.c
703
ret = drm_gem_handle_create(file, &mem->base, handle);
drivers/accel/ethosu/ethosu_gem.c
707
drm_gem_object_put(&mem->base);
drivers/accel/ethosu/ethosu_gem.c
75
struct drm_gem_dma_object *mem;
drivers/accel/ethosu/ethosu_gem.c
79
mem = drm_gem_dma_create(ddev, *size);
drivers/accel/ethosu/ethosu_gem.c
80
if (IS_ERR(mem))
drivers/accel/ethosu/ethosu_gem.c
81
return PTR_ERR(mem);
drivers/accel/ethosu/ethosu_gem.c
83
bo = to_ethosu_bo(&mem->base);
drivers/accel/ethosu/ethosu_gem.c
90
ret = drm_gem_handle_create(file, &mem->base, handle);
drivers/accel/ethosu/ethosu_gem.c
95
drm_gem_object_put(&mem->base);
drivers/accel/ivpu/ivpu_fw.c
388
fw->mem = ivpu_bo_create_runtime(vdev, fw->runtime_addr, fw->runtime_size,
drivers/accel/ivpu/ivpu_fw.c
390
if (!fw->mem) {
drivers/accel/ivpu/ivpu_fw.c
441
ivpu_bo_free(fw->mem);
drivers/accel/ivpu/ivpu_fw.c
460
ivpu_bo_free(fw->mem);
drivers/accel/ivpu/ivpu_fw.c
466
fw->mem = NULL;
drivers/accel/ivpu/ivpu_fw.c
509
memset(ivpu_bo_vaddr(fw->mem), 0, fw->image_load_offset);
drivers/accel/ivpu/ivpu_fw.c
510
memcpy(ivpu_bo_vaddr(fw->mem) + fw->image_load_offset,
drivers/accel/ivpu/ivpu_fw.c
514
u8 *start = ivpu_bo_vaddr(fw->mem) + image_end_offset;
drivers/accel/ivpu/ivpu_fw.c
515
u64 size = ivpu_bo_size(fw->mem) - image_end_offset;
drivers/accel/ivpu/ivpu_fw.h
25
struct ivpu_bo *mem;
drivers/accel/ivpu/ivpu_job.c
105
cmdq->mem = ivpu_bo_create_global(vdev, SZ_4K, DRM_IVPU_BO_WC | DRM_IVPU_BO_MAPPABLE);
drivers/accel/ivpu/ivpu_job.c
106
if (!cmdq->mem)
drivers/accel/ivpu/ivpu_job.c
124
size_t size = ivpu_bo_size(cmdq->mem) - sizeof(struct vpu_job_queue_header);
drivers/accel/ivpu/ivpu_job.c
156
ivpu_bo_free(cmdq->mem);
drivers/accel/ivpu/ivpu_job.c
183
cmdq->jobq = (struct vpu_job_queue *)ivpu_bo_vaddr(cmdq->mem);
drivers/accel/ivpu/ivpu_job.c
204
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
drivers/accel/ivpu/ivpu_job.c
230
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
drivers/accel/ivpu/ivpu_job.c
233
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
drivers/accel/ivpu/ivpu_job.h
34
struct ivpu_bo *mem;
drivers/acpi/acpi_extlog.c
229
struct cper_sec_mem_err *mem = acpi_hest_get_payload(gdata);
drivers/acpi/acpi_extlog.c
231
if (gdata->error_data_length >= sizeof(*mem))
drivers/acpi/acpi_extlog.c
232
trace_extlog_mem_event(mem, err_seq, fru_id, fru_text,
drivers/acpi/acpi_memhotplug.c
146
static int acpi_bind_memblk(struct memory_block *mem, void *arg)
drivers/acpi/acpi_memhotplug.c
148
return acpi_bind_one(&mem->dev, arg);
drivers/acpi/acpi_memhotplug.c
158
static int acpi_unbind_memblk(struct memory_block *mem, void *arg)
drivers/acpi/acpi_memhotplug.c
160
acpi_unbind_one(&mem->dev);
drivers/acpi/acpi_memhotplug.c
75
if ((info->caching == address64.info.mem.caching) &&
drivers/acpi/acpi_memhotplug.c
76
(info->write_protect == address64.info.mem.write_protect) &&
drivers/acpi/acpi_memhotplug.c
88
new->caching = address64.info.mem.caching;
drivers/acpi/acpi_memhotplug.c
89
new->write_protect = address64.info.mem.write_protect;
drivers/acpi/acpica/rsaddr.c
214
{ACPI_RSC_1BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.write_protect),
drivers/acpi/acpica/rsaddr.c
218
{ACPI_RSC_2BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.caching),
drivers/acpi/acpica/rsaddr.c
222
{ACPI_RSC_2BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.range_type),
drivers/acpi/acpica/rsaddr.c
226
{ACPI_RSC_1BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.translation),
drivers/acpi/acpica/rsdumpinfo.c
545
{ACPI_RSD_1BITFLAG, ACPI_RSD_OFFSET(address.info.mem.write_protect),
drivers/acpi/acpica/rsdumpinfo.c
547
{ACPI_RSD_2BITFLAG, ACPI_RSD_OFFSET(address.info.mem.caching),
drivers/acpi/acpica/rsdumpinfo.c
549
{ACPI_RSD_2BITFLAG, ACPI_RSD_OFFSET(address.info.mem.range_type),
drivers/acpi/acpica/rsdumpinfo.c
551
{ACPI_RSD_1BITFLAG, ACPI_RSD_OFFSET(address.info.mem.translation),
drivers/acpi/resource.c
212
bool wp = addr->info.mem.write_protect;
drivers/acpi/resource.c
255
if (addr->info.mem.caching == ACPI_PREFETCHABLE_MEMORY)
drivers/acpi/viot.c
344
struct resource *mem;
drivers/acpi/viot.c
347
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/acpi/viot.c
348
if (!mem)
drivers/acpi/viot.c
352
if (ep->address == mem->start)
drivers/ata/acard-ahci.c
278
void *mem;
drivers/ata/acard-ahci.c
309
mem = dmam_alloc_coherent(dev, dma_sz, &mem_dma, GFP_KERNEL);
drivers/ata/acard-ahci.c
310
if (!mem)
drivers/ata/acard-ahci.c
317
pp->cmd_slot = mem;
drivers/ata/acard-ahci.c
320
mem += AHCI_CMD_SLOT_SZ;
drivers/ata/acard-ahci.c
326
pp->rx_fis = mem;
drivers/ata/acard-ahci.c
329
mem += rx_fis_sz;
drivers/ata/acard-ahci.c
336
pp->cmd_tbl = mem;
drivers/ata/libahci.c
2487
void *mem;
drivers/ata/libahci.c
2528
mem = dmam_alloc_coherent(dev, dma_sz, &mem_dma, GFP_KERNEL);
drivers/ata/libahci.c
2529
if (!mem)
drivers/ata/libahci.c
2536
pp->cmd_slot = mem;
drivers/ata/libahci.c
2539
mem += AHCI_CMD_SLOT_SZ;
drivers/ata/libahci.c
2545
pp->rx_fis = mem;
drivers/ata/libahci.c
2548
mem += rx_fis_sz;
drivers/ata/libahci.c
2555
pp->cmd_tbl = mem;
drivers/ata/sata_fsl.c
703
void *mem;
drivers/ata/sata_fsl.c
713
mem = dma_alloc_coherent(dev, SATA_FSL_PORT_PRIV_DMA_SZ, &mem_dma,
drivers/ata/sata_fsl.c
715
if (!mem) {
drivers/ata/sata_fsl.c
720
pp->cmdslot = mem;
drivers/ata/sata_fsl.c
723
mem += SATA_FSL_CMD_SLOT_SIZE;
drivers/ata/sata_fsl.c
726
pp->cmdentry = mem;
drivers/ata/sata_highbank.c
458
struct resource *mem;
drivers/ata/sata_highbank.c
466
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/ata/sata_highbank.c
467
if (!mem) {
drivers/ata/sata_highbank.c
492
hpriv->mmio = devm_ioremap(dev, mem->start, resource_size(mem));
drivers/ata/sata_highbank.c
494
dev_err(dev, "can't map %pR\n", mem);
drivers/ata/sata_highbank.c
540
ata_port_desc(ap, "mmio %pR", mem);
drivers/ata/sata_nv.c
1101
void *mem;
drivers/ata/sata_nv.c
1139
mem = dmam_alloc_coherent(dev, NV_ADMA_PORT_PRIV_DMA_SZ,
drivers/ata/sata_nv.c
1141
if (!mem)
drivers/ata/sata_nv.c
1149
pp->cpb = mem;
drivers/ata/sata_nv.c
1155
mem += NV_ADMA_MAX_CPBS * NV_ADMA_CPB_SZ;
drivers/ata/sata_nv.c
1161
pp->aprd = mem;
drivers/atm/eni.c
1294
void __iomem *mem;
drivers/atm/eni.c
1318
mem = NULL; /* for gcc */
drivers/atm/eni.c
1321
mem = eni_alloc_mem(eni_dev,&size);
drivers/atm/eni.c
1322
if (!mem) return -ENOBUFS;
drivers/atm/eni.c
1325
eni_free_mem(eni_dev,mem,size);
drivers/atm/eni.c
1330
tx->send = mem;
drivers/atm/eni.c
1353
eni_free_mem(eni_dev,mem,size);
drivers/atm/eni.c
1762
eni_dev->mem = i;
drivers/atm/eni.c
1763
memset_io(eni_dev->ram,0,eni_dev->mem);
drivers/atm/eni.c
1765
printk("mem=%dkB (",eni_dev->mem >> 10);
drivers/atm/eni.c
1846
buffer_mem = eni_dev->mem - (buf - eni_dev->ram);
drivers/atm/eni.c
2094
eni_dev->mem >> 10,eni_dev->tx_bw);
drivers/atm/eni.h
114
int mem; /* RAM on board (in bytes) */
drivers/atm/iphase.c
1018
RAM_BASE*((iadev->mem)/(128 * 1024))
drivers/atm/iphase.c
1020
IPHASE5575_FRAG_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))
drivers/atm/iphase.c
1022
IPHASE5575_REASS_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))
drivers/atm/iphase.c
2377
iadev->mem = iadev->pci_map_size /2;
drivers/atm/iphase.h
1023
int mem;
drivers/base/memory.c
1010
struct memory_block *mem;
drivers/base/memory.c
1018
mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
1019
if (!mem)
drivers/base/memory.c
1022
ret = func(mem, arg);
drivers/base/memory.c
1023
put_device(&mem->dev);
drivers/base/memory.c
1037
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
1040
return cb_data->func(mem, cb_data->arg);
drivers/base/memory.c
1231
struct memory_block *mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
1233
if (mem)
drivers/base/memory.c
1234
atomic_long_inc(&mem->nr_hwpoison);
drivers/base/memory.c
1240
struct memory_block *mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
1242
if (mem)
drivers/base/memory.c
1243
atomic_long_sub(i, &mem->nr_hwpoison);
drivers/base/memory.c
1246
static unsigned long memblk_nr_poison(struct memory_block *mem)
drivers/base/memory.c
1248
return atomic_long_read(&mem->nr_hwpoison);
drivers/base/memory.c
161
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
163
return sysfs_emit(buf, "%08lx\n", memory_block_id(mem->start_section_nr));
drivers/base/memory.c
182
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
189
switch (mem->state) {
drivers/base/memory.c
201
return sysfs_emit(buf, "ERROR-UNKNOWN-%d\n", mem->state);
drivers/base/memory.c
213
static unsigned long memblk_nr_poison(struct memory_block *mem);
drivers/base/memory.c
215
static inline unsigned long memblk_nr_poison(struct memory_block *mem)
drivers/base/memory.c
224
static int memory_block_online(struct memory_block *mem)
drivers/base/memory.c
226
unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
drivers/base/memory.c
232
if (memblk_nr_poison(mem))
drivers/base/memory.c
235
zone = zone_for_pfn_range(mem->online_type, mem->nid, mem->group,
drivers/base/memory.c
245
if (mem->altmap)
drivers/base/memory.c
246
nr_vmemmap_pages = mem->altmap->free;
drivers/base/memory.c
256
nr_pages - nr_vmemmap_pages, zone, mem->group);
drivers/base/memory.c
268
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
271
mem->zone = zone;
drivers/base/memory.c
280
static int memory_block_offline(struct memory_block *mem)
drivers/base/memory.c
282
unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
drivers/base/memory.c
287
if (!mem->zone)
drivers/base/memory.c
294
if (mem->altmap)
drivers/base/memory.c
295
nr_vmemmap_pages = mem->altmap->free;
drivers/base/memory.c
299
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
303
nr_pages - nr_vmemmap_pages, mem->zone, mem->group);
drivers/base/memory.c
308
mem->group, nr_vmemmap_pages);
drivers/base/memory.c
315
mem->zone = NULL;
drivers/base/memory.c
326
memory_block_action(struct memory_block *mem, unsigned long action)
drivers/base/memory.c
332
ret = memory_block_online(mem);
drivers/base/memory.c
335
ret = memory_block_offline(mem);
drivers/base/memory.c
339
"%ld\n", __func__, mem->start_section_nr, action, action);
drivers/base/memory.c
346
static int memory_block_change_state(struct memory_block *mem,
drivers/base/memory.c
351
if (mem->state != from_state_req)
drivers/base/memory.c
355
mem->state = MEM_GOING_OFFLINE;
drivers/base/memory.c
357
ret = memory_block_action(mem, to_state);
drivers/base/memory.c
358
mem->state = ret ? from_state_req : to_state;
drivers/base/memory.c
366
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
369
if (mem->state == MEM_ONLINE)
drivers/base/memory.c
376
if (mem->online_type == MMOP_OFFLINE)
drivers/base/memory.c
377
mem->online_type = MMOP_ONLINE;
drivers/base/memory.c
379
ret = memory_block_change_state(mem, MEM_ONLINE, MEM_OFFLINE);
drivers/base/memory.c
380
mem->online_type = MMOP_OFFLINE;
drivers/base/memory.c
387
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
389
if (mem->state == MEM_OFFLINE)
drivers/base/memory.c
392
return memory_block_change_state(mem, MEM_OFFLINE, MEM_ONLINE);
drivers/base/memory.c
399
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
414
mem->online_type = online_type;
drivers/base/memory.c
415
ret = device_online(&mem->dev);
drivers/base/memory.c
418
ret = device_offline(&mem->dev);
drivers/base/memory.c
444
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
445
unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
drivers/base/memory.c
469
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
470
unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
drivers/base/memory.c
472
struct memory_group *group = mem->group;
drivers/base/memory.c
474
int nid = mem->nid;
drivers/base/memory.c
481
if (mem->state == MEM_ONLINE) {
drivers/base/memory.c
487
mem->zone ? mem->zone->name : "none");
drivers/base/memory.c
654
struct memory_block *mem;
drivers/base/memory.c
656
mem = xa_load(&memory_blocks, block_id);
drivers/base/memory.c
657
if (mem)
drivers/base/memory.c
658
get_device(&mem->dev);
drivers/base/memory.c
659
return mem;
drivers/base/memory.c
715
static struct zone *early_node_zone_for_memory_block(struct memory_block *mem,
drivers/base/memory.c
718
const unsigned long start_pfn = section_nr_to_pfn(mem->start_section_nr);
drivers/base/memory.c
764
void memory_block_add_nid_early(struct memory_block *mem, int nid)
drivers/base/memory.c
766
if (mem->nid != nid) {
drivers/base/memory.c
776
if (mem->nid == NUMA_NO_NODE)
drivers/base/memory.c
777
mem->zone = early_node_zone_for_memory_block(mem, nid);
drivers/base/memory.c
779
mem->zone = NULL;
drivers/base/memory.c
786
mem->nid = nid;
drivers/base/memory.c
795
struct memory_block *mem;
drivers/base/memory.c
798
mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
799
if (mem) {
drivers/base/memory.c
800
put_device(&mem->dev);
drivers/base/memory.c
803
mem = kzalloc_obj(*mem);
drivers/base/memory.c
804
if (!mem)
drivers/base/memory.c
807
mem->start_section_nr = block_id * sections_per_block;
drivers/base/memory.c
808
mem->state = state;
drivers/base/memory.c
809
mem->nid = nid;
drivers/base/memory.c
810
mem->altmap = altmap;
drivers/base/memory.c
811
INIT_LIST_HEAD(&mem->group_next);
drivers/base/memory.c
821
mem->zone = early_node_zone_for_memory_block(mem, NUMA_NO_NODE);
drivers/base/memory.c
824
ret = __add_memory_block(mem);
drivers/base/memory.c
829
mem->group = group;
drivers/base/memory.c
830
list_add(&mem->group_next, &group->memory_blocks);
drivers/base/memory.c
866
struct memory_block *mem;
drivers/base/memory.c
883
mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
884
if (WARN_ON_ONCE(!mem))
drivers/base/memory.c
886
remove_memory_block(mem);
drivers/base/memory.c
903
struct memory_block *mem;
drivers/base/memory.c
911
mem = find_memory_block_by_id(block_id);
drivers/base/memory.c
912
if (WARN_ON_ONCE(!mem))
drivers/base/memory.c
914
num_poisoned_pages_sub(-1UL, memblk_nr_poison(mem));
drivers/base/memory.c
915
unregister_memory_block_under_nodes(mem);
drivers/base/memory.c
916
remove_memory_block(mem);
drivers/base/memory.c
94
struct memory_block *mem = to_memory_block(dev);
drivers/base/memory.c
96
WARN_ON(mem->altmap);
drivers/base/memory.c
97
kfree(mem);
drivers/base/node.c
843
struct memory_block *mem;
drivers/base/node.c
845
mem = find_memory_block_by_id(block_id);
drivers/base/node.c
846
if (!mem)
drivers/base/node.c
849
memory_block_add_nid_early(mem, nid);
drivers/base/node.c
850
do_register_memory_block_under_node(nid, mem);
drivers/base/node.c
851
put_device(&mem->dev);
drivers/block/zram/backend_lz4.c
34
vfree(zctx->mem);
drivers/block/zram/backend_lz4.c
50
zctx->mem = vmalloc(LZ4_MEM_COMPRESS);
drivers/block/zram/backend_lz4.c
51
if (!zctx->mem)
drivers/block/zram/backend_lz4.c
79
zctx->mem);
drivers/block/zram/backend_lz4.c
9
void *mem;
drivers/block/zram/backend_lz4hc.c
36
vfree(zctx->mem);
drivers/block/zram/backend_lz4hc.c
50
zctx->mem = vmalloc(LZ4HC_MEM_COMPRESS);
drivers/block/zram/backend_lz4hc.c
51
if (!zctx->mem)
drivers/block/zram/backend_lz4hc.c
79
zctx->mem);
drivers/block/zram/backend_lz4hc.c
9
void *mem;
drivers/block/zram/zram_drv.c
2039
void *mem;
drivers/block/zram/zram_drv.c
2041
mem = kmap_local_page(page);
drivers/block/zram/zram_drv.c
2042
zram_fill_page(mem, PAGE_SIZE, get_slot_handle(zram, index));
drivers/block/zram/zram_drv.c
2043
kunmap_local(mem);
drivers/block/zram/zram_drv.c
2246
void *mem;
drivers/block/zram/zram_drv.c
2251
mem = kmap_local_page(page);
drivers/block/zram/zram_drv.c
2252
same_filled = page_same_filled(mem, &element);
drivers/block/zram/zram_drv.c
2253
kunmap_local(mem);
drivers/block/zram/zram_drv.c
2258
mem = kmap_local_page(page);
drivers/block/zram/zram_drv.c
2260
mem, &comp_len);
drivers/block/zram/zram_drv.c
2261
kunmap_local(mem);
drivers/bus/mvebu-mbus.c
1219
struct resource *mem,
drivers/bus/mvebu-mbus.c
1229
memset(mem, 0, sizeof(struct resource));
drivers/bus/mvebu-mbus.c
1230
mem->end = -1;
drivers/bus/mvebu-mbus.c
1236
mem->start = reg[0];
drivers/bus/mvebu-mbus.c
1237
mem->end = mem->start + reg[1] - 1;
drivers/bus/mvebu-mbus.c
1238
mem->flags = IORESOURCE_MEM;
drivers/cache/hisi_soc_hha.c
126
struct resource *mem;
drivers/cache/hisi_soc_hha.c
138
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/cache/hisi_soc_hha.c
139
if (!mem) {
drivers/cache/hisi_soc_hha.c
144
soc_hha->base = ioremap(mem->start, resource_size(mem));
drivers/char/agp/agp.h
192
int agp_generic_insert_memory(struct agp_memory *mem, off_t pg_start, int type);
drivers/char/agp/agp.h
193
int agp_generic_remove_memory(struct agp_memory *mem, off_t pg_start, int type);
drivers/char/agp/agp.h
216
void agp_alloc_page_array(size_t size, struct agp_memory *mem);
drivers/char/agp/agp.h
217
static inline void agp_free_page_array(struct agp_memory *mem)
drivers/char/agp/agp.h
219
kvfree(mem->pages);
drivers/char/agp/agp.h
225
void agp3_generic_tlbflush(struct agp_memory *mem);
drivers/char/agp/ali-agp.c
43
static void ali_tlbflush(struct agp_memory *mem)
drivers/char/agp/alpha-agp.c
101
alpha_core_agp_tlbflush(mem);
drivers/char/agp/alpha-agp.c
106
static int alpha_core_agp_remove_memory(struct agp_memory *mem, off_t pg_start,
drivers/char/agp/alpha-agp.c
112
status = agp->ops->unbind(agp, pg_start, mem);
drivers/char/agp/alpha-agp.c
113
alpha_core_agp_tlbflush(mem);
drivers/char/agp/alpha-agp.c
65
static void alpha_core_agp_tlbflush(struct agp_memory *mem)
drivers/char/agp/alpha-agp.c
84
static int alpha_core_agp_insert_memory(struct agp_memory *mem, off_t pg_start,
drivers/char/agp/alpha-agp.c
91
if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES)
drivers/char/agp/alpha-agp.c
96
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/alpha-agp.c
99
status = agp->ops->bind(agp, pg_start, mem);
drivers/char/agp/amd-k7-agp.c
284
static int amd_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/amd-k7-agp.c
292
if (type != mem->type ||
drivers/char/agp/amd-k7-agp.c
296
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd-k7-agp.c
300
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/amd-k7-agp.c
308
if (!mem->is_flushed) {
drivers/char/agp/amd-k7-agp.c
310
mem->is_flushed = true;
drivers/char/agp/amd-k7-agp.c
313
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/amd-k7-agp.c
317
page_to_phys(mem->pages[i]),
drivers/char/agp/amd-k7-agp.c
318
mem->type),
drivers/char/agp/amd-k7-agp.c
322
amd_irongate_tlbflush(mem);
drivers/char/agp/amd-k7-agp.c
326
static int amd_remove_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/amd-k7-agp.c
332
if (type != mem->type ||
drivers/char/agp/amd-k7-agp.c
336
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/amd-k7-agp.c
343
amd_irongate_tlbflush(mem);
drivers/char/agp/amd64-agp.c
45
static int amd64_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/amd64-agp.c
50
struct agp_bridge_data *bridge = mem->bridge;
drivers/char/agp/amd64-agp.c
55
if (type != mem->type)
drivers/char/agp/amd64-agp.c
64
if (((unsigned long)pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd64-agp.c
70
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/amd64-agp.c
76
if (!mem->is_flushed) {
drivers/char/agp/amd64-agp.c
78
mem->is_flushed = true;
drivers/char/agp/amd64-agp.c
81
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/amd64-agp.c
83
page_to_phys(mem->pages[i]),
drivers/char/agp/amd64-agp.c
94
amd64_tlbflush(mem);
drivers/char/agp/ati-agp.c
175
static void ati_tlbflush(struct agp_memory * mem)
drivers/char/agp/ati-agp.c
258
static int ati_insert_memory(struct agp_memory * mem,
drivers/char/agp/ati-agp.c
268
mask_type = agp_generic_type_to_mask_type(mem->bridge, type);
drivers/char/agp/ati-agp.c
269
if (mask_type != 0 || type != mem->type)
drivers/char/agp/ati-agp.c
272
if (mem->page_count == 0)
drivers/char/agp/ati-agp.c
275
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/ati-agp.c
279
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/ati-agp.c
287
if (!mem->is_flushed) {
drivers/char/agp/ati-agp.c
290
mem->is_flushed = true;
drivers/char/agp/ati-agp.c
293
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/ati-agp.c
297
page_to_phys(mem->pages[i]),
drivers/char/agp/ati-agp.c
298
mem->type),
drivers/char/agp/ati-agp.c
302
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/ati-agp.c
306
static int ati_remove_memory(struct agp_memory * mem, off_t pg_start,
drivers/char/agp/ati-agp.c
314
mask_type = agp_generic_type_to_mask_type(mem->bridge, type);
drivers/char/agp/ati-agp.c
315
if (mask_type != 0 || type != mem->type)
drivers/char/agp/ati-agp.c
318
if (mem->page_count == 0)
drivers/char/agp/ati-agp.c
321
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/ati-agp.c
328
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/backend.c
103
static const struct { int mem, agp; } maxes_table[] = {
drivers/char/agp/backend.c
126
while ((memory > maxes_table[index].mem) && (index < 8))
drivers/char/agp/backend.c
130
( (memory - maxes_table[index - 1].mem) *
drivers/char/agp/backend.c
132
(maxes_table[index].mem - maxes_table[index - 1].mem);
drivers/char/agp/efficeon-agp.c
109
static void efficeon_tlbflush(struct agp_memory * mem)
drivers/char/agp/efficeon-agp.c
236
static int efficeon_insert_memory(struct agp_memory * mem, off_t pg_start, int type)
drivers/char/agp/efficeon-agp.c
238
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
246
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/efficeon-agp.c
248
if (type != 0 || mem->type != 0)
drivers/char/agp/efficeon-agp.c
251
if (!mem->is_flushed) {
drivers/char/agp/efficeon-agp.c
253
mem->is_flushed = true;
drivers/char/agp/efficeon-agp.c
259
unsigned long insert = efficeon_mask_memory(mem->pages[i]);
drivers/char/agp/efficeon-agp.c
281
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/efficeon-agp.c
285
static int efficeon_remove_memory(struct agp_memory * mem, off_t pg_start, int type)
drivers/char/agp/efficeon-agp.c
287
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
293
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/efficeon-agp.c
295
if (type != 0 || mem->type != 0)
drivers/char/agp/efficeon-agp.c
307
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/generic.c
1025
int agp_generic_insert_memory(struct agp_memory * mem, off_t pg_start, int type)
drivers/char/agp/generic.c
1034
bridge = mem->bridge;
drivers/char/agp/generic.c
1038
if (mem->page_count == 0)
drivers/char/agp/generic.c
1067
if (type != mem->type)
drivers/char/agp/generic.c
1076
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
1077
((pg_start + mem->page_count) < pg_start))
drivers/char/agp/generic.c
1082
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/generic.c
1088
if (!mem->is_flushed) {
drivers/char/agp/generic.c
1090
mem->is_flushed = true;
drivers/char/agp/generic.c
1093
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/generic.c
1095
page_to_phys(mem->pages[i]),
drivers/char/agp/generic.c
1101
bridge->driver->tlb_flush(mem);
drivers/char/agp/generic.c
1107
int agp_generic_remove_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/generic.c
1113
bridge = mem->bridge;
drivers/char/agp/generic.c
1117
if (mem->page_count == 0)
drivers/char/agp/generic.c
1120
if (type != mem->type)
drivers/char/agp/generic.c
1124
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
1125
((pg_start + mem->page_count) < pg_start))
drivers/char/agp/generic.c
1135
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/generic.c
1140
bridge->driver->tlb_flush(mem);
drivers/char/agp/generic.c
1187
int agp_generic_alloc_pages(struct agp_bridge_data *bridge, struct agp_memory *mem, size_t num_pages)
drivers/char/agp/generic.c
1204
mem->pages[i] = page;
drivers/char/agp/generic.c
1205
mem->page_count++;
drivers/char/agp/generic.c
1209
set_pages_array_uc(mem->pages, num_pages);
drivers/char/agp/generic.c
1233
void agp_generic_destroy_pages(struct agp_memory *mem)
drivers/char/agp/generic.c
1238
if (!mem)
drivers/char/agp/generic.c
1242
set_pages_array_wb(mem->pages, mem->page_count);
drivers/char/agp/generic.c
1245
for (i = 0; i < mem->page_count; i++) {
drivers/char/agp/generic.c
1246
page = mem->pages[i];
drivers/char/agp/generic.c
1254
mem->pages[i] = NULL;
drivers/char/agp/generic.c
1363
void agp3_generic_tlbflush(struct agp_memory *mem)
drivers/char/agp/generic.c
90
void agp_alloc_page_array(size_t size, struct agp_memory *mem)
drivers/char/agp/generic.c
92
mem->pages = kvmalloc(size, GFP_KERNEL);
drivers/char/agp/intel-agp.c
186
static void intel_820_tlbflush(struct agp_memory *mem)
drivers/char/agp/intel-agp.c
75
static void intel_tlbflush(struct agp_memory *mem)
drivers/char/agp/intel-agp.c
82
static void intel_8xx_tlbflush(struct agp_memory *mem)
drivers/char/agp/intel-gtt.c
1002
intel_gmch_gtt_clear_range(pg_start, mem->page_count);
drivers/char/agp/intel-gtt.c
1005
intel_gtt_unmap_memory(mem->sg_list, mem->num_sg);
drivers/char/agp/intel-gtt.c
1006
mem->sg_list = NULL;
drivers/char/agp/intel-gtt.c
1007
mem->num_sg = 0;
drivers/char/agp/intel-gtt.c
129
DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count);
drivers/char/agp/intel-gtt.c
214
static int i810_insert_dcache_entries(struct agp_memory *mem, off_t pg_start,
drivers/char/agp/intel-gtt.c
219
if ((pg_start + mem->page_count)
drivers/char/agp/intel-gtt.c
223
if (!mem->is_flushed)
drivers/char/agp/intel-gtt.c
226
for (i = pg_start; i < (pg_start + mem->page_count); i++) {
drivers/char/agp/intel-gtt.c
931
static int intel_fake_agp_insert_entries(struct agp_memory *mem,
drivers/char/agp/intel-gtt.c
944
return i810_insert_dcache_entries(mem, pg_start, type);
drivers/char/agp/intel-gtt.c
946
if (mem->page_count == 0)
drivers/char/agp/intel-gtt.c
949
if (pg_start + mem->page_count > intel_private.gtt_total_entries)
drivers/char/agp/intel-gtt.c
952
if (type != mem->type)
drivers/char/agp/intel-gtt.c
958
if (!mem->is_flushed)
drivers/char/agp/intel-gtt.c
964
ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st);
drivers/char/agp/intel-gtt.c
969
mem->sg_list = st.sgl;
drivers/char/agp/intel-gtt.c
970
mem->num_sg = st.nents;
drivers/char/agp/intel-gtt.c
972
intel_gmch_gtt_insert_pages(pg_start, mem->page_count, mem->pages,
drivers/char/agp/intel-gtt.c
978
mem->is_flushed = true;
drivers/char/agp/intel-gtt.c
996
static int intel_fake_agp_remove_entries(struct agp_memory *mem,
drivers/char/agp/intel-gtt.c
999
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
202
static int nvidia_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/nvidia-agp.c
207
mask_type = agp_generic_type_to_mask_type(mem->bridge, type);
drivers/char/agp/nvidia-agp.c
208
if (mask_type != 0 || type != mem->type)
drivers/char/agp/nvidia-agp.c
211
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
214
if ((pg_start + mem->page_count) >
drivers/char/agp/nvidia-agp.c
218
for (j = pg_start; j < (pg_start + mem->page_count); j++) {
drivers/char/agp/nvidia-agp.c
223
if (!mem->is_flushed) {
drivers/char/agp/nvidia-agp.c
225
mem->is_flushed = true;
drivers/char/agp/nvidia-agp.c
227
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/nvidia-agp.c
229
page_to_phys(mem->pages[i]), mask_type),
drivers/char/agp/nvidia-agp.c
236
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/nvidia-agp.c
241
static int nvidia_remove_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/nvidia-agp.c
247
mask_type = agp_generic_type_to_mask_type(mem->bridge, type);
drivers/char/agp/nvidia-agp.c
248
if (mask_type != 0 || type != mem->type)
drivers/char/agp/nvidia-agp.c
251
if (mem->page_count == 0)
drivers/char/agp/nvidia-agp.c
254
for (i = pg_start; i < (mem->page_count + pg_start); i++)
drivers/char/agp/nvidia-agp.c
257
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/nvidia-agp.c
262
static void nvidia_tlbflush(struct agp_memory *mem)
drivers/char/agp/parisc-agp.c
124
parisc_agp_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/parisc-agp.c
131
if (type != mem->type ||
drivers/char/agp/parisc-agp.c
137
io_pg_count = info->io_pages_per_kpage * mem->page_count;
drivers/char/agp/parisc-agp.c
149
if (!mem->is_flushed) {
drivers/char/agp/parisc-agp.c
151
mem->is_flushed = true;
drivers/char/agp/parisc-agp.c
154
for (i = 0, j = io_pg_start; i < mem->page_count; i++) {
drivers/char/agp/parisc-agp.c
157
paddr = page_to_phys(mem->pages[i]);
drivers/char/agp/parisc-agp.c
168
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/parisc-agp.c
174
parisc_agp_remove_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/parisc-agp.c
179
if (type != mem->type ||
drivers/char/agp/parisc-agp.c
185
io_pg_count = info->io_pages_per_kpage * mem->page_count;
drivers/char/agp/parisc-agp.c
190
agp_bridge->driver->tlb_flush(mem);
drivers/char/agp/parisc-agp.c
89
parisc_agp_tlbflush(struct agp_memory *mem)
drivers/char/agp/sis-agp.c
46
static void sis_tlbflush(struct agp_memory *mem)
drivers/char/agp/sworks-agp.c
316
static int serverworks_insert_memory(struct agp_memory *mem,
drivers/char/agp/sworks-agp.c
325
if (type != 0 || mem->type != 0) {
drivers/char/agp/sworks-agp.c
328
if ((pg_start + mem->page_count) > num_entries) {
drivers/char/agp/sworks-agp.c
333
while (j < (pg_start + mem->page_count)) {
drivers/char/agp/sworks-agp.c
341
if (!mem->is_flushed) {
drivers/char/agp/sworks-agp.c
343
mem->is_flushed = true;
drivers/char/agp/sworks-agp.c
346
for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
drivers/char/agp/sworks-agp.c
350
page_to_phys(mem->pages[i]), mem->type),
drivers/char/agp/sworks-agp.c
353
serverworks_tlbflush(mem);
drivers/char/agp/sworks-agp.c
357
static int serverworks_remove_memory(struct agp_memory *mem, off_t pg_start,
drivers/char/agp/sworks-agp.c
364
if (type != 0 || mem->type != 0) {
drivers/char/agp/sworks-agp.c
369
serverworks_tlbflush(mem);
drivers/char/agp/sworks-agp.c
371
for (i = pg_start; i < (mem->page_count + pg_start); i++) {
drivers/char/agp/sworks-agp.c
377
serverworks_tlbflush(mem);
drivers/char/agp/uninorth-agp.c
150
static int uninorth_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/uninorth-agp.c
157
if (type != mem->type)
drivers/char/agp/uninorth-agp.c
166
if (mem->page_count == 0)
drivers/char/agp/uninorth-agp.c
172
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/uninorth-agp.c
176
for (i = 0; i < mem->page_count; ++i) {
drivers/char/agp/uninorth-agp.c
185
for (i = 0; i < mem->page_count; i++) {
drivers/char/agp/uninorth-agp.c
187
gp[i] = (page_to_phys(mem->pages[i]) >> PAGE_SHIFT) | 0x80000000UL;
drivers/char/agp/uninorth-agp.c
189
gp[i] = cpu_to_le32((page_to_phys(mem->pages[i]) & 0xFFFFF000UL) |
drivers/char/agp/uninorth-agp.c
191
flush_dcache_range((unsigned long)__va(page_to_phys(mem->pages[i])),
drivers/char/agp/uninorth-agp.c
192
(unsigned long)__va(page_to_phys(mem->pages[i]))+0x1000);
drivers/char/agp/uninorth-agp.c
195
uninorth_tlbflush(mem);
drivers/char/agp/uninorth-agp.c
200
static int uninorth_remove_memory(struct agp_memory *mem, off_t pg_start, int type)
drivers/char/agp/uninorth-agp.c
206
if (type != mem->type)
drivers/char/agp/uninorth-agp.c
215
if (mem->page_count == 0)
drivers/char/agp/uninorth-agp.c
219
for (i = 0; i < mem->page_count; ++i) {
drivers/char/agp/uninorth-agp.c
223
uninorth_tlbflush(mem);
drivers/char/agp/uninorth-agp.c
74
static void uninorth_tlbflush(struct agp_memory *mem)
drivers/char/agp/uninorth-agp.c
84
if (!mem && uninorth_rev <= 0x30) {
drivers/char/agp/via-agp.c
160
static void via_tlbflush_agp3(struct agp_memory *mem)
drivers/char/agp/via-agp.c
80
static void via_tlbflush(struct agp_memory *mem)
drivers/char/applicom.c
247
if (!mem || !irq) {
drivers/char/applicom.c
260
RamIO = ioremap(mem + (LEN_RAM_IO * i), LEN_RAM_IO);
drivers/char/applicom.c
267
if (!(boardno = ac_register_board((unsigned long)mem+ (LEN_RAM_IO*i),
drivers/char/applicom.c
273
printk(KERN_NOTICE "Applicom ISA card found at mem 0x%lx, irq %d\n", mem + (LEN_RAM_IO*i), irq);
drivers/char/applicom.c
292
"at mem 0x%lx\n", mem);
drivers/char/applicom.c
792
printk(KERN_INFO "Segment of board ........... %X\n", (int) mem);
drivers/char/applicom.c
93
static unsigned long mem; /* physical segment of board */
drivers/char/applicom.c
97
module_param_hw(mem, ulong, iomem, 0);
drivers/char/applicom.c
98
MODULE_PARM_DESC(mem, "Shared Memory Address of Applicom board");
drivers/char/hw_random/bcm74110-rng.c
34
static inline int bcm74110_rng_fifo_count(void __iomem *mem)
drivers/char/hw_random/bcm74110-rng.c
36
return readl_relaxed(mem) & HOST_FIFO_COUNT_MASK;
drivers/char/hw_random/exynos-trng.c
151
writel_relaxed(val, trng->mem + EXYNOS_TRNG_CLKDIV);
drivers/char/hw_random/exynos-trng.c
155
writel_relaxed(val, trng->mem + EXYNOS_TRNG_CTRL);
drivers/char/hw_random/exynos-trng.c
161
writel_relaxed(0, trng->mem + EXYNOS_TRNG_POST_CTRL);
drivers/char/hw_random/exynos-trng.c
212
trng->mem = devm_platform_ioremap_resource(pdev, 0);
drivers/char/hw_random/exynos-trng.c
213
if (IS_ERR(trng->mem))
drivers/char/hw_random/exynos-trng.c
214
return PTR_ERR(trng->mem);
drivers/char/hw_random/exynos-trng.c
75
void __iomem *mem;
drivers/char/hw_random/exynos-trng.c
89
writel_relaxed(max * 8, trng->mem + EXYNOS_TRNG_FIFO_CTRL);
drivers/char/hw_random/exynos-trng.c
90
val = readl_poll_timeout(trng->mem + EXYNOS_TRNG_FIFO_CTRL, val,
drivers/char/hw_random/exynos-trng.c
95
memcpy_fromio(data, trng->mem + EXYNOS_TRNG_FIFO_0, max);
drivers/char/hw_random/geode-rng.c
120
mem = ioremap(rng_base, 0x58);
drivers/char/hw_random/geode-rng.c
121
if (!mem)
drivers/char/hw_random/geode-rng.c
125
priv->membase = mem;
drivers/char/hw_random/geode-rng.c
138
iounmap(mem);
drivers/char/hw_random/geode-rng.c
62
void __iomem *mem = priv->membase;
drivers/char/hw_random/geode-rng.c
64
*data = readl(mem + GEODE_RNG_DATA_REG);
drivers/char/hw_random/geode-rng.c
72
void __iomem *mem = priv->membase;
drivers/char/hw_random/geode-rng.c
76
data = !!(readl(mem + GEODE_RNG_STATUS_REG));
drivers/char/hw_random/geode-rng.c
97
void __iomem *mem;
drivers/char/hw_random/intel-rng.c
155
static inline u8 hwstatus_get(void __iomem *mem)
drivers/char/hw_random/intel-rng.c
157
return readb(mem + INTEL_RNG_HW_STATUS);
drivers/char/hw_random/intel-rng.c
160
static inline u8 hwstatus_set(void __iomem *mem,
drivers/char/hw_random/intel-rng.c
163
writeb(hw_status, mem + INTEL_RNG_HW_STATUS);
drivers/char/hw_random/intel-rng.c
164
return hwstatus_get(mem);
drivers/char/hw_random/intel-rng.c
169
void __iomem *mem = (void __iomem *)rng->priv;
drivers/char/hw_random/intel-rng.c
173
data = !!(readb(mem + INTEL_RNG_STATUS) &
drivers/char/hw_random/intel-rng.c
184
void __iomem *mem = (void __iomem *)rng->priv;
drivers/char/hw_random/intel-rng.c
186
*data = readb(mem + INTEL_RNG_DATA);
drivers/char/hw_random/intel-rng.c
193
void __iomem *mem = (void __iomem *)rng->priv;
drivers/char/hw_random/intel-rng.c
197
hw_status = hwstatus_get(mem);
drivers/char/hw_random/intel-rng.c
200
hw_status = hwstatus_set(mem, hw_status | INTEL_RNG_ENABLED);
drivers/char/hw_random/intel-rng.c
212
void __iomem *mem = (void __iomem *)rng->priv;
drivers/char/hw_random/intel-rng.c
215
hw_status = hwstatus_get(mem);
drivers/char/hw_random/intel-rng.c
217
hwstatus_set(mem, hw_status & ~INTEL_RNG_ENABLED);
drivers/char/hw_random/intel-rng.c
233
void __iomem *mem;
drivers/char/hw_random/intel-rng.c
258
writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem);
drivers/char/hw_random/intel-rng.c
259
writeb(INTEL_FWH_READ_ID_CMD, intel_rng_hw->mem);
drivers/char/hw_random/intel-rng.c
260
mfc = readb(intel_rng_hw->mem + INTEL_FWH_MANUFACTURER_CODE_ADDRESS);
drivers/char/hw_random/intel-rng.c
261
dvc = readb(intel_rng_hw->mem + INTEL_FWH_DEVICE_CODE_ADDRESS);
drivers/char/hw_random/intel-rng.c
262
writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem);
drivers/char/hw_random/intel-rng.c
320
intel_rng_hw->mem = ioremap(INTEL_FWH_ADDR, INTEL_FWH_ADDR_LEN);
drivers/char/hw_random/intel-rng.c
321
if (intel_rng_hw->mem == NULL)
drivers/char/hw_random/intel-rng.c
333
void __iomem *mem;
drivers/char/hw_random/intel-rng.c
374
iounmap(intel_rng_hw->mem);
drivers/char/hw_random/intel-rng.c
381
mem = ioremap(INTEL_RNG_ADDR, INTEL_RNG_ADDR_LEN);
drivers/char/hw_random/intel-rng.c
382
if (!mem)
drivers/char/hw_random/intel-rng.c
384
intel_rng.priv = (unsigned long)mem;
drivers/char/hw_random/intel-rng.c
388
hw_status = hwstatus_get(mem);
drivers/char/hw_random/intel-rng.c
390
iounmap(mem);
drivers/char/hw_random/intel-rng.c
399
iounmap(mem);
drivers/char/hw_random/intel-rng.c
408
void __iomem *mem = (void __iomem *)intel_rng.priv;
drivers/char/hw_random/intel-rng.c
411
iounmap(mem);
drivers/char/hw_random/mxc-rnga.c
106
ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
107
__raw_writel(ctrl & ~RNGA_CONTROL_SLEEP, mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
110
osc = __raw_readl(mxc_rng->mem + RNGA_STATUS);
drivers/char/hw_random/mxc-rnga.c
117
ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
118
__raw_writel(ctrl | RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
128
ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
131
__raw_writel(ctrl & ~RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
156
mxc_rng->mem = devm_platform_ioremap_resource(pdev, 0);
drivers/char/hw_random/mxc-rnga.c
157
if (IS_ERR(mxc_rng->mem)) {
drivers/char/hw_random/mxc-rnga.c
158
err = PTR_ERR(mxc_rng->mem);
drivers/char/hw_random/mxc-rnga.c
57
void __iomem *mem;
drivers/char/hw_random/mxc-rnga.c
68
int level = (__raw_readl(mxc_rng->mem + RNGA_STATUS) &
drivers/char/hw_random/mxc-rnga.c
84
*data = __raw_readl(mxc_rng->mem + RNGA_OUTPUT_FIFO);
drivers/char/hw_random/mxc-rnga.c
87
err = __raw_readl(mxc_rng->mem + RNGA_STATUS) & RNGA_STATUS_ERROR_INT;
drivers/char/hw_random/mxc-rnga.c
92
ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/mxc-rnga.c
94
mxc_rng->mem + RNGA_CONTROL);
drivers/char/hw_random/xiphera-trng.c
102
if (readl(trng->mem + STATUS_REG) != TRNG_SUCCESSFUL_STARTUP) {
drivers/char/hw_random/xiphera-trng.c
104
if (readl(trng->mem + STATUS_REG) == TRNG_FAILED_STARTUP) {
drivers/char/hw_random/xiphera-trng.c
112
writel(HOST_TO_TRNG_ACK_ZEROIZE, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
31
void __iomem *mem;
drivers/char/hw_random/xiphera-trng.c
42
if (readl(trng->mem + STATUS_REG) == TRNG_NEW_RAND_AVAILABLE) {
drivers/char/hw_random/xiphera-trng.c
43
*(u32 *)buf = readl(trng->mem + RAND_REG);
drivers/char/hw_random/xiphera-trng.c
48
writel(HOST_TO_TRNG_READ, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
49
writel(HOST_TO_TRNG_ENABLE, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
70
trng->mem = devm_platform_ioremap_resource(pdev, 0);
drivers/char/hw_random/xiphera-trng.c
71
if (IS_ERR(trng->mem))
drivers/char/hw_random/xiphera-trng.c
72
return PTR_ERR(trng->mem);
drivers/char/hw_random/xiphera-trng.c
78
writel(HOST_TO_TRNG_RESET, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
81
if (readl(trng->mem + STATUS_REG) != TRNG_ACK_RESET) {
drivers/char/hw_random/xiphera-trng.c
87
if (readl(trng->mem + STATUS_REG) != TRNG_ACK_RESET) {
drivers/char/hw_random/xiphera-trng.c
97
writel(HOST_TO_TRNG_RELEASE_RESET, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
98
writel(HOST_TO_TRNG_ENABLE, trng->mem + CONTROL_REG);
drivers/char/hw_random/xiphera-trng.c
99
writel(HOST_TO_TRNG_ZEROIZE, trng->mem + CONTROL_REG);
drivers/char/xillybus/xillyusb.c
256
rc = (*copier)(fifo->mem[writebuf] + writepos, data + done, n);
drivers/char/xillybus/xillyusb.c
319
rc = (*copier)(data + done, fifo->mem[readbuf] + readpos, n);
drivers/char/xillybus/xillyusb.c
396
fifo->mem = kmalloc_array(fifo->bufnum, sizeof(void *), GFP_KERNEL);
drivers/char/xillybus/xillyusb.c
398
if (!fifo->mem)
drivers/char/xillybus/xillyusb.c
402
fifo->mem[i] = (void *)
drivers/char/xillybus/xillyusb.c
405
if (!fifo->mem[i])
drivers/char/xillybus/xillyusb.c
420
free_pages((unsigned long)fifo->mem[i], buf_order);
drivers/char/xillybus/xillyusb.c
422
kfree(fifo->mem);
drivers/char/xillybus/xillyusb.c
423
fifo->mem = NULL;
drivers/char/xillybus/xillyusb.c
437
if (!fifo->mem)
drivers/char/xillybus/xillyusb.c
441
free_pages((unsigned long)fifo->mem[i], fifo->buf_order);
drivers/char/xillybus/xillyusb.c
443
kfree(fifo->mem);
drivers/char/xillybus/xillyusb.c
84
char **mem;
drivers/clk/tegra/clk-dfll.c
1942
struct resource *mem;
drivers/clk/tegra/clk-dfll.c
1998
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/clk/tegra/clk-dfll.c
1999
if (!mem) {
drivers/clk/tegra/clk-dfll.c
2004
td->base = devm_ioremap(td->dev, mem->start, resource_size(mem));
drivers/clk/tegra/clk-dfll.c
2010
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
drivers/clk/tegra/clk-dfll.c
2011
if (!mem) {
drivers/clk/tegra/clk-dfll.c
2016
td->i2c_base = devm_ioremap(td->dev, mem->start, resource_size(mem));
drivers/clk/tegra/clk-dfll.c
2022
mem = platform_get_resource(pdev, IORESOURCE_MEM, 2);
drivers/clk/tegra/clk-dfll.c
2023
if (!mem) {
drivers/clk/tegra/clk-dfll.c
2028
td->i2c_controller_base = devm_ioremap(td->dev, mem->start,
drivers/clk/tegra/clk-dfll.c
2029
resource_size(mem));
drivers/clk/tegra/clk-dfll.c
2036
mem = platform_get_resource(pdev, IORESOURCE_MEM, 3);
drivers/clk/tegra/clk-dfll.c
2037
if (!mem) {
drivers/clk/tegra/clk-dfll.c
2042
td->lut_base = devm_ioremap(td->dev, mem->start, resource_size(mem));
drivers/clk/ti/clk.c
38
void __iomem *mem;
drivers/clk/ti/clk.c
399
struct regmap *syscon, void __iomem *mem)
drivers/clk/ti/clk.c
419
io->mem = mem;
drivers/clk/ti/clk.c
433
void __init omap2_clk_legacy_provider_init(int index, void __iomem *mem)
drivers/clk/ti/clk.c
439
io->mem = mem;
drivers/clk/ti/clk.c
52
writel_relaxed(val, io->mem + reg->offset);
drivers/clk/ti/clk.c
74
_clk_rmw(val, mask, io->mem + reg->offset);
drivers/clk/ti/clk.c
88
val = readl_relaxed(io->mem + reg->offset);
drivers/clocksource/sh_cmt.c
959
struct resource *mem;
drivers/clocksource/sh_cmt.c
961
mem = platform_get_resource(cmt->pdev, IORESOURCE_MEM, 0);
drivers/clocksource/sh_cmt.c
962
if (!mem) {
drivers/clocksource/sh_cmt.c
967
cmt->mapbase = ioremap(mem->start, resource_size(mem));
drivers/crypto/ccp/sev-dev.c
2238
void *id_blob, *mem;
drivers/crypto/ccp/sev-dev.c
2253
mem = kzalloc(data_size + user_size, GFP_KERNEL);
drivers/crypto/ccp/sev-dev.c
2254
if (!mem)
drivers/crypto/ccp/sev-dev.c
2257
data = mem;
drivers/crypto/ccp/sev-dev.c
2258
id_blob = mem + data_size;
drivers/crypto/ccp/sev-dev.c
2269
kfree(mem);
drivers/crypto/exynos-rng.c
165
memcpy_fromio(dst, rng->mem + EXYNOS_RNG_OUT_BASE, *read);
drivers/crypto/exynos-rng.c
291
rng->mem = devm_platform_ioremap_resource(pdev, 0);
drivers/crypto/exynos-rng.c
292
if (IS_ERR(rng->mem))
drivers/crypto/exynos-rng.c
293
return PTR_ERR(rng->mem);
drivers/crypto/exynos-rng.c
73
void __iomem *mem;
drivers/crypto/exynos-rng.c
89
return readl_relaxed(rng->mem + offset);
drivers/crypto/exynos-rng.c
94
writel_relaxed(val, rng->mem + offset);
drivers/dma/acpi-dma.c
50
resource_size_t mem = 0, irq = 0;
drivers/dma/acpi-dma.c
63
mem = rentry->res->start;
drivers/dma/acpi-dma.c
71
if (mem == 0 && irq == 0)
drivers/dma/acpi-dma.c
77
if (si->mmio_base_low != lower_32_bits(mem) ||
drivers/dma/acpi-dma.c
78
si->mmio_base_high != upper_32_bits(mem))
drivers/dma/at_hdmac.c
1301
u32 mem;
drivers/dma/at_hdmac.c
1310
mem = sg_dma_address(sg);
drivers/dma/at_hdmac.c
1318
if (unlikely(mem & 3 || len & 3))
drivers/dma/at_hdmac.c
1321
lli->saddr = mem;
drivers/dma/at_hdmac.c
1350
u32 mem;
drivers/dma/at_hdmac.c
1359
mem = sg_dma_address(sg);
drivers/dma/at_hdmac.c
1367
if (unlikely(mem & 3 || len & 3))
drivers/dma/at_hdmac.c
1371
lli->daddr = mem;
drivers/dma/at_xdmac.c
793
u32 len, mem, dwidth, fixed_dwidth;
drivers/dma/at_xdmac.c
796
mem = sg_dma_address(sg);
drivers/dma/at_xdmac.c
802
__func__, i, len, mem);
drivers/dma/at_xdmac.c
816
desc->lld.mbr_da = mem;
drivers/dma/at_xdmac.c
818
desc->lld.mbr_sa = mem;
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
855
dma_addr_t mem;
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
863
mem = sg_dma_address(sgl);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
866
axi_block_len = calculate_block_len(chan, mem, len, direction);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
880
mem = sg_dma_address(sg);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
887
status = dw_axi_dma_set_hw_desc(chan, hw_desc, mem, segment_len);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
893
mem += segment_len;
drivers/dma/dw-edma/dw-edma-v0-core.c
285
struct dw_edma_v0_lli *lli = chunk->ll_region.vaddr.mem + ofs;
drivers/dma/dw-edma/dw-edma-v0-core.c
307
struct dw_edma_v0_llp *llp = chunk->ll_region.vaddr.mem + ofs;
drivers/dma/dw-edma/dw-hdma-v0-core.c
161
struct dw_hdma_v0_lli *lli = chunk->ll_region.vaddr.mem + ofs;
drivers/dma/dw-edma/dw-hdma-v0-core.c
183
struct dw_hdma_v0_llp *llp = chunk->ll_region.vaddr.mem + ofs;
drivers/dma/dw/core.c
656
u32 len, mem;
drivers/dma/dw/core.c
659
mem = sg_dma_address(sg);
drivers/dma/dw/core.c
662
mem_width = __ffs(sconfig->src_addr_width | mem | len);
drivers/dma/dw/core.c
671
lli_write(desc, sar, mem);
drivers/dma/dw/core.c
685
mem += dlen;
drivers/dma/dw/core.c
706
u32 len, mem;
drivers/dma/dw/core.c
709
mem = sg_dma_address(sg);
drivers/dma/dw/core.c
720
lli_write(desc, dar, mem);
drivers/dma/dw/core.c
722
mem_width = __ffs(sconfig->dst_addr_width | mem);
drivers/dma/dw/core.c
734
mem += dlen;
drivers/dma/sh/rcar-dmac.c
741
if (desc->hwdescs.mem) {
drivers/dma/sh/rcar-dmac.c
743
desc->hwdescs.mem, desc->hwdescs.dma);
drivers/dma/sh/rcar-dmac.c
744
desc->hwdescs.mem = NULL;
drivers/dma/sh/rcar-dmac.c
751
desc->hwdescs.mem = dma_alloc_coherent(chan->chan.device->dev, size,
drivers/dma/sh/rcar-dmac.c
753
if (!desc->hwdescs.mem)
drivers/dma/sh/rcar-dmac.c
767
hwdesc = desc->hwdescs.mem;
drivers/dma/sh/rcar-dmac.c
85
struct rcar_dmac_hw_desc *mem;
drivers/dma/st_fdma.h
185
readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
drivers/dma/st_fdma.h
190
writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
drivers/dma/st_fdma.h
197
writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
drivers/dma/st_fdma.h
207
readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
drivers/dma/st_fdma.h
212
writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
drivers/dma/tegra186-gpc-dma.c
1083
dma_addr_t mem;
drivers/dma/tegra186-gpc-dma.c
1085
mem = sg_dma_address(sg);
drivers/dma/tegra186-gpc-dma.c
1088
if ((len & 3) || (mem & 3) || len > max_dma_count) {
drivers/dma/tegra186-gpc-dma.c
1099
sg_req[i].ch_regs.src_ptr = mem;
drivers/dma/tegra186-gpc-dma.c
1102
FIELD_PREP(TEGRA_GPCDMA_HIGH_ADDR_SRC_PTR, (mem >> 32));
drivers/dma/tegra186-gpc-dma.c
1105
sg_req[i].ch_regs.dst_ptr = mem;
drivers/dma/tegra186-gpc-dma.c
1107
FIELD_PREP(TEGRA_GPCDMA_HIGH_ADDR_DST_PTR, (mem >> 32));
drivers/dma/tegra186-gpc-dma.c
1136
dma_addr_t mem = buf_addr;
drivers/dma/tegra186-gpc-dma.c
1220
sg_req[i].ch_regs.src_ptr = mem;
drivers/dma/tegra186-gpc-dma.c
1223
FIELD_PREP(TEGRA_GPCDMA_HIGH_ADDR_SRC_PTR, (mem >> 32));
drivers/dma/tegra186-gpc-dma.c
1226
sg_req[i].ch_regs.dst_ptr = mem;
drivers/dma/tegra186-gpc-dma.c
1228
FIELD_PREP(TEGRA_GPCDMA_HIGH_ADDR_DST_PTR, (mem >> 32));
drivers/dma/tegra186-gpc-dma.c
1240
mem += len;
drivers/dma/tegra20-apb-dma.c
1102
u32 len, mem;
drivers/dma/tegra20-apb-dma.c
1104
mem = sg_dma_address(sg);
drivers/dma/tegra20-apb-dma.c
1107
if ((len & 3) || (mem & 3) ||
drivers/dma/tegra20-apb-dma.c
1126
sg_req->ch_regs.ahb_ptr = mem;
drivers/dma/tegra20-apb-dma.c
1172
dma_addr_t mem = buf_addr;
drivers/dma/tegra20-apb-dma.c
1261
sg_req->ch_regs.ahb_ptr = mem;
drivers/dma/tegra20-apb-dma.c
1273
mem += len;
drivers/dma/ti/cppi41.c
278
static void cppi_writel(u32 val, void *__iomem *mem)
drivers/dma/ti/cppi41.c
280
__raw_writel(val, mem);
drivers/dma/ti/cppi41.c
283
static u32 cppi_readl(void *__iomem *mem)
drivers/dma/ti/cppi41.c
285
return __raw_readl(mem);
drivers/dma/ti/edma.c
2293
struct resource *mem;
drivers/dma/ti/edma.c
2332
mem = platform_get_resource_byname(pdev, IORESOURCE_MEM, "edma3_cc");
drivers/dma/ti/edma.c
2333
if (!mem) {
drivers/dma/ti/edma.c
2335
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/dma/ti/edma.c
2336
if (!mem) {
drivers/dma/ti/edma.c
2341
ecc->base = devm_ioremap_resource(dev, mem);
drivers/dma/txx9dmac.c
829
dma_addr_t mem;
drivers/dma/txx9dmac.c
838
mem = sg_dma_address(sg);
drivers/dma/txx9dmac.c
842
desc->hwdesc.SAR = mem;
drivers/dma/txx9dmac.c
846
desc->hwdesc.DAR = mem;
drivers/dma/txx9dmac.c
851
desc->hwdesc32.SAR = mem;
drivers/dma/txx9dmac.c
855
desc->hwdesc32.DAR = mem;
drivers/edac/ghes_edac.c
245
static int print_mem_error_other_detail(const struct cper_sec_mem_err *mem, char *msg,
drivers/edac/ghes_edac.c
258
if (!(mem->validation_bits & CPER_MEM_VALID_ERROR_STATUS))
drivers/edac/ghes_edac.c
261
n += scnprintf(msg + n, len - n, "status(0x%016llx): ", mem->error_status);
drivers/edac/ghes_edac.c
262
n += scnprintf(msg + n, len - n, "%s ", cper_mem_err_status_str(mem->error_status));
drivers/edac/thunderx_edac.c
201
struct page *mem;
drivers/edac/thunderx_edac.c
338
addr = (uintptr_t)page_address(lmc->mem);
drivers/edac/thunderx_edac.c
339
phys = (uintptr_t)page_to_phys(lmc->mem);
drivers/edac/thunderx_edac.c
412
lmc->mem = alloc_pages_node(lmc->node, GFP_KERNEL, 0);
drivers/edac/thunderx_edac.c
413
if (!lmc->mem)
drivers/edac/thunderx_edac.c
418
__free_pages(lmc->mem, 0);
drivers/edac/thunderx_edac.c
422
addr = page_address(lmc->mem);
drivers/edac/thunderx_edac.c
438
__free_pages(lmc->mem, 0);
drivers/firmware/arm_scpi.c
389
struct legacy_scpi_shared_mem __iomem *mem =
drivers/firmware/arm_scpi.c
395
match->status = ioread32(&mem->status);
drivers/firmware/arm_scpi.c
396
memcpy_fromio(match->rx_buf, mem->payload, len);
drivers/firmware/arm_scpi.c
398
struct scpi_shared_mem __iomem *mem = ch->rx_payload;
drivers/firmware/arm_scpi.c
402
match->status = ioread32(&mem->status);
drivers/firmware/arm_scpi.c
403
memcpy_fromio(match->rx_buf, mem->payload, len);
drivers/firmware/arm_scpi.c
416
struct scpi_shared_mem __iomem *mem = ch->rx_payload;
drivers/firmware/arm_scpi.c
420
cmd = ioread32(&mem->command);
drivers/firmware/arm_scpi.c
430
struct scpi_shared_mem __iomem *mem = ch->tx_payload;
drivers/firmware/arm_scpi.c
436
memcpy_toio(mem->payload, t->tx_buf, t->tx_len);
drivers/firmware/arm_scpi.c
449
iowrite32(t->cmd, &mem->command);
drivers/firmware/cirrus/cs_dsp.c
1489
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
1591
mem = cs_dsp_find_region(dsp, type);
drivers/firmware/cirrus/cs_dsp.c
1592
if (!mem) {
drivers/firmware/cirrus/cs_dsp.c
1599
reg = dsp->ops->region_to_reg(mem, offset);
drivers/firmware/cirrus/cs_dsp.c
1709
const struct cs_dsp_region *mem,
drivers/firmware/cirrus/cs_dsp.c
1728
reg = dsp->ops->region_to_reg(mem, pos + len);
drivers/firmware/cirrus/cs_dsp.c
1748
reg = dsp->ops->region_to_reg(mem, pos);
drivers/firmware/cirrus/cs_dsp.c
1866
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
1871
mem = cs_dsp_find_region(dsp, WMFW_ADSP1_DM);
drivers/firmware/cirrus/cs_dsp.c
1872
if (WARN_ON(!mem))
drivers/firmware/cirrus/cs_dsp.c
1875
ret = regmap_raw_read(dsp->regmap, mem->base, &adsp1_id,
drivers/firmware/cirrus/cs_dsp.c
1903
adsp1_alg = cs_dsp_read_algs(dsp, n_algs, mem, pos, len);
drivers/firmware/cirrus/cs_dsp.c
1971
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
1976
mem = cs_dsp_find_region(dsp, WMFW_ADSP2_XM);
drivers/firmware/cirrus/cs_dsp.c
1977
if (WARN_ON(!mem))
drivers/firmware/cirrus/cs_dsp.c
1980
ret = regmap_raw_read(dsp->regmap, mem->base, &adsp2_id,
drivers/firmware/cirrus/cs_dsp.c
2014
adsp2_alg = cs_dsp_read_algs(dsp, n_algs, mem, pos, len);
drivers/firmware/cirrus/cs_dsp.c
2117
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
2122
mem = cs_dsp_find_region(dsp, WMFW_ADSP2_XM);
drivers/firmware/cirrus/cs_dsp.c
2123
if (WARN_ON(!mem))
drivers/firmware/cirrus/cs_dsp.c
2126
ret = regmap_raw_read(dsp->regmap, mem->base, &halo_id,
drivers/firmware/cirrus/cs_dsp.c
2147
halo_alg = cs_dsp_read_algs(dsp, n_algs, mem, pos, len);
drivers/firmware/cirrus/cs_dsp.c
2181
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
2273
mem = cs_dsp_find_region(dsp, type);
drivers/firmware/cirrus/cs_dsp.c
2274
if (!mem) {
drivers/firmware/cirrus/cs_dsp.c
2278
reg = dsp->ops->region_to_reg(mem, 0);
drivers/firmware/cirrus/cs_dsp.c
2305
mem = cs_dsp_find_region(dsp, type);
drivers/firmware/cirrus/cs_dsp.c
2306
if (!mem) {
drivers/firmware/cirrus/cs_dsp.c
2325
reg = dsp->ops->region_to_reg(mem, reg);
drivers/firmware/cirrus/cs_dsp.c
3083
struct cs_dsp_region const *mem = cs_dsp_find_region(dsp, mem_type);
drivers/firmware/cirrus/cs_dsp.c
3089
if (!mem)
drivers/firmware/cirrus/cs_dsp.c
3092
reg = dsp->ops->region_to_reg(mem, mem_addr);
drivers/firmware/cirrus/cs_dsp.c
3138
struct cs_dsp_region const *mem = cs_dsp_find_region(dsp, mem_type);
drivers/firmware/cirrus/cs_dsp.c
3144
if (!mem)
drivers/firmware/cirrus/cs_dsp.c
3147
reg = dsp->ops->region_to_reg(mem, mem_addr);
drivers/firmware/cirrus/cs_dsp.c
331
unsigned int (*region_to_reg)(struct cs_dsp_region const *mem,
drivers/firmware/cirrus/cs_dsp.c
575
if (dsp->mem[i].type == type)
drivers/firmware/cirrus/cs_dsp.c
576
return &dsp->mem[i];
drivers/firmware/cirrus/cs_dsp.c
581
static unsigned int cs_dsp_region_to_reg(struct cs_dsp_region const *mem,
drivers/firmware/cirrus/cs_dsp.c
584
switch (mem->type) {
drivers/firmware/cirrus/cs_dsp.c
586
return mem->base + (offset * 3);
drivers/firmware/cirrus/cs_dsp.c
591
return mem->base + (offset * 2);
drivers/firmware/cirrus/cs_dsp.c
598
static unsigned int cs_dsp_halo_region_to_reg(struct cs_dsp_region const *mem,
drivers/firmware/cirrus/cs_dsp.c
601
switch (mem->type) {
drivers/firmware/cirrus/cs_dsp.c
604
return mem->base + (offset * 4);
drivers/firmware/cirrus/cs_dsp.c
607
return (mem->base + (offset * 3)) & ~0x3;
drivers/firmware/cirrus/cs_dsp.c
609
return mem->base + (offset * 5);
drivers/firmware/cirrus/cs_dsp.c
671
const struct cs_dsp_region *mem;
drivers/firmware/cirrus/cs_dsp.c
673
mem = cs_dsp_find_region(dsp, alg_region->type);
drivers/firmware/cirrus/cs_dsp.c
674
if (!mem) {
drivers/firmware/cirrus/cs_dsp.c
680
*reg = dsp->ops->region_to_reg(mem, ctl->alg_region.base + ctl->offset + off);
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
101
else if (dsp->mem == cs_dsp_mock_adsp2_16bit_dsp1_regions)
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
107
if (dsp->mem[i].type == mem_type)
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
126
const struct cs_dsp_region *region = priv->dsp->mem;
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
97
if (dsp->mem == cs_dsp_mock_halo_dsp1_regions)
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
99
else if (dsp->mem == cs_dsp_mock_adsp2_32bit_dsp1_regions)
drivers/firmware/cirrus/test/cs_dsp_test_bin.c
2263
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_bin.c
2293
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_bin.c
2312
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
503
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
523
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
542
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_callbacks.c
530
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_callbacks.c
550
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_callbacks.c
579
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_cache.c
2769
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_cache.c
2789
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_cache.c
2818
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_parse.c
1456
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_parse.c
1476
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_parse.c
1505
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_rw.c
2137
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_rw.c
2157
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_control_rw.c
2186
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw.c
1887
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw.c
1907
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw.c
1941
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
1101
dsp->mem = cs_dsp_mock_halo_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
1121
dsp->mem = cs_dsp_mock_adsp2_32bit_dsp1_regions;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
1155
dsp->mem = cs_dsp_mock_adsp2_16bit_dsp1_regions;
drivers/firmware/efi/cper.c
301
int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg)
drivers/firmware/efi/cper.c
310
if (mem->validation_bits & CPER_MEM_VALID_NODE)
drivers/firmware/efi/cper.c
311
n += scnprintf(msg + n, len - n, "node:%d ", mem->node);
drivers/firmware/efi/cper.c
312
if (mem->validation_bits & CPER_MEM_VALID_CARD)
drivers/firmware/efi/cper.c
313
n += scnprintf(msg + n, len - n, "card:%d ", mem->card);
drivers/firmware/efi/cper.c
314
if (mem->validation_bits & CPER_MEM_VALID_MODULE)
drivers/firmware/efi/cper.c
315
n += scnprintf(msg + n, len - n, "module:%d ", mem->module);
drivers/firmware/efi/cper.c
316
if (mem->validation_bits & CPER_MEM_VALID_RANK_NUMBER)
drivers/firmware/efi/cper.c
317
n += scnprintf(msg + n, len - n, "rank:%d ", mem->rank);
drivers/firmware/efi/cper.c
318
if (mem->validation_bits & CPER_MEM_VALID_BANK)
drivers/firmware/efi/cper.c
319
n += scnprintf(msg + n, len - n, "bank:%d ", mem->bank);
drivers/firmware/efi/cper.c
320
if (mem->validation_bits & CPER_MEM_VALID_BANK_GROUP)
drivers/firmware/efi/cper.c
322
mem->bank >> CPER_MEM_BANK_GROUP_SHIFT);
drivers/firmware/efi/cper.c
323
if (mem->validation_bits & CPER_MEM_VALID_BANK_ADDRESS)
drivers/firmware/efi/cper.c
325
mem->bank & CPER_MEM_BANK_ADDRESS_MASK);
drivers/firmware/efi/cper.c
326
if (mem->validation_bits & CPER_MEM_VALID_DEVICE)
drivers/firmware/efi/cper.c
327
n += scnprintf(msg + n, len - n, "device:%d ", mem->device);
drivers/firmware/efi/cper.c
328
if (mem->validation_bits & (CPER_MEM_VALID_ROW | CPER_MEM_VALID_ROW_EXT)) {
drivers/firmware/efi/cper.c
329
u32 row = mem->row;
drivers/firmware/efi/cper.c
331
row |= cper_get_mem_extension(mem->validation_bits, mem->extended);
drivers/firmware/efi/cper.c
334
if (mem->validation_bits & CPER_MEM_VALID_COLUMN)
drivers/firmware/efi/cper.c
335
n += scnprintf(msg + n, len - n, "column:%d ", mem->column);
drivers/firmware/efi/cper.c
336
if (mem->validation_bits & CPER_MEM_VALID_BIT_POSITION)
drivers/firmware/efi/cper.c
338
mem->bit_pos);
drivers/firmware/efi/cper.c
339
if (mem->validation_bits & CPER_MEM_VALID_REQUESTOR_ID)
drivers/firmware/efi/cper.c
341
mem->requestor_id);
drivers/firmware/efi/cper.c
342
if (mem->validation_bits & CPER_MEM_VALID_RESPONDER_ID)
drivers/firmware/efi/cper.c
344
mem->responder_id);
drivers/firmware/efi/cper.c
345
if (mem->validation_bits & CPER_MEM_VALID_TARGET_ID)
drivers/firmware/efi/cper.c
347
mem->target_id);
drivers/firmware/efi/cper.c
348
if (mem->validation_bits & CPER_MEM_VALID_CHIP_ID)
drivers/firmware/efi/cper.c
350
mem->extended >> CPER_MEM_CHIP_ID_SHIFT);
drivers/firmware/efi/cper.c
356
int cper_dimm_err_location(struct cper_mem_err_compact *mem, char *msg)
drivers/firmware/efi/cper.c
361
if (!msg || !(mem->validation_bits & CPER_MEM_VALID_MODULE_HANDLE))
drivers/firmware/efi/cper.c
365
dmi_memdev_name(mem->mem_dev_handle, &bank, &device);
drivers/firmware/efi/cper.c
371
mem->mem_dev_handle);
drivers/firmware/efi/cper.c
377
void cper_mem_err_pack(const struct cper_sec_mem_err *mem,
drivers/firmware/efi/cper.c
380
cmem->validation_bits = mem->validation_bits;
drivers/firmware/efi/cper.c
381
cmem->node = mem->node;
drivers/firmware/efi/cper.c
382
cmem->card = mem->card;
drivers/firmware/efi/cper.c
383
cmem->module = mem->module;
drivers/firmware/efi/cper.c
384
cmem->bank = mem->bank;
drivers/firmware/efi/cper.c
385
cmem->device = mem->device;
drivers/firmware/efi/cper.c
386
cmem->row = mem->row;
drivers/firmware/efi/cper.c
387
cmem->column = mem->column;
drivers/firmware/efi/cper.c
388
cmem->bit_pos = mem->bit_pos;
drivers/firmware/efi/cper.c
389
cmem->requestor_id = mem->requestor_id;
drivers/firmware/efi/cper.c
390
cmem->responder_id = mem->responder_id;
drivers/firmware/efi/cper.c
391
cmem->target_id = mem->target_id;
drivers/firmware/efi/cper.c
392
cmem->extended = mem->extended;
drivers/firmware/efi/cper.c
393
cmem->rank = mem->rank;
drivers/firmware/efi/cper.c
394
cmem->mem_array_handle = mem->mem_array_handle;
drivers/firmware/efi/cper.c
395
cmem->mem_dev_handle = mem->mem_dev_handle;
drivers/firmware/efi/cper.c
414
static void cper_print_mem(const char *pfx, const struct cper_sec_mem_err *mem,
drivers/firmware/efi/cper.c
422
(mem->validation_bits & ~(CPER_MEM_VALID_RANK_NUMBER - 1))) {
drivers/firmware/efi/cper.c
426
if (mem->validation_bits & CPER_MEM_VALID_ERROR_STATUS)
drivers/firmware/efi/cper.c
428
pfx, cper_mem_err_status_str(mem->error_status),
drivers/firmware/efi/cper.c
429
mem->error_status);
drivers/firmware/efi/cper.c
430
if (mem->validation_bits & CPER_MEM_VALID_PA)
drivers/firmware/efi/cper.c
432
pfx, mem->physical_addr);
drivers/firmware/efi/cper.c
433
if (mem->validation_bits & CPER_MEM_VALID_PA_MASK)
drivers/firmware/efi/cper.c
435
pfx, mem->physical_addr_mask);
drivers/firmware/efi/cper.c
436
cper_mem_err_pack(mem, &cmem);
drivers/firmware/efi/cper.c
439
if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE) {
drivers/firmware/efi/cper.c
440
u8 etype = mem->error_type;
drivers/firmware/efi/libstub/efistub.h
771
efi_pci_io_protocol_access_t mem;
drivers/firmware/efi/libstub/efistub.h
794
efi_pci_io_protocol_access_32_t mem;
drivers/firmware/google/cbmem.c
56
static const BIN_ATTR_ADMIN_RW(mem, 0);
drivers/gpio/gpio-elkhartlake.c
39
priv->reg_base = devm_ioremap_resource(dev, &data->mem);
drivers/gpio/gpio-menz127.c
155
men_z127_gpio->mem = mcb_request_mem(mdev, dev_name(dev));
drivers/gpio/gpio-menz127.c
156
if (IS_ERR(men_z127_gpio->mem))
drivers/gpio/gpio-menz127.c
157
return dev_err_probe(dev, PTR_ERR(men_z127_gpio->mem),
drivers/gpio/gpio-menz127.c
161
men_z127_gpio->mem);
drivers/gpio/gpio-menz127.c
165
men_z127_gpio->reg_base = devm_ioremap(dev, men_z127_gpio->mem->start,
drivers/gpio/gpio-menz127.c
166
resource_size(men_z127_gpio->mem));
drivers/gpio/gpio-menz127.c
41
struct resource *mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
207
unsigned long cur_seq, struct kgd_mem *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
231
unsigned long cur_seq, struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
251
int amdgpu_amdkfd_add_gws_to_process(void *info, void *gws, struct kgd_mem **mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
252
int amdgpu_amdkfd_remove_gws_from_process(void *info, void *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
315
void *drm_priv, struct kgd_mem **mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
318
struct amdgpu_device *adev, struct kgd_mem *mem, void *drm_priv,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
321
struct kgd_mem *mem, void *drm_priv);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
323
struct amdgpu_device *adev, struct kgd_mem *mem, void *drm_priv);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
324
int amdgpu_amdkfd_gpuvm_dmaunmap_mem(struct kgd_mem *mem, void *drm_priv);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
326
struct amdgpu_device *adev, struct kgd_mem *mem, bool intr);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
327
int amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
329
void amdgpu_amdkfd_gpuvm_unmap_gtt_bo_from_kernel(struct kgd_mem *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
339
struct kgd_mem **mem, uint64_t *size,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
341
int amdgpu_amdkfd_gpuvm_export_dmabuf(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
354
bool amdgpu_amdkfd_bo_mapped_to_dev(void *drm_priv, struct kgd_mem *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1028
static void add_kgd_mem_to_kfd_bo_list(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1034
list_add_tail(&mem->validate_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1037
list_add_tail(&mem->validate_list, &process_info->kfd_bo_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1041
static void remove_kgd_mem_from_kfd_bo_list(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1045
list_del(&mem->validate_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1061
static int init_user_pages(struct kgd_mem *mem, uint64_t user_addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1064
struct amdkfd_process_info *process_info = mem->process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1065
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1093
mem->invalid++;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1123
amdgpu_bo_placement_from_domain(bo, mem->domain);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
115
uint64_t mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1165
static int reserve_bo_and_vm(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1169
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1175
ctx->sync = &mem->sync;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1206
static int reserve_bo_and_cond_vms(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
121
mem = si.totalram - si.totalhigh;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1211
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1214
ctx->sync = &mem->sync;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1219
list_for_each_entry(entry, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
122
mem *= si.mem_unit;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
125
kfd_mem_limit.max_system_mem_limit = mem - (mem >> 6);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1269
static int unmap_bo_from_gpuvm(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1295
static int update_gpuvm_pte(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1303
ret = kfd_mem_dmamap_attachment(mem, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1317
static int map_bo_to_gpuvm(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1337
ret = update_gpuvm_pte(mem, entry, sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1346
unmap_bo_from_gpuvm(mem, entry, sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1347
kfd_mem_dmaunmap_attachment(mem, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1711
void *drm_priv, struct kgd_mem **mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1780
*mem = kzalloc_obj(struct kgd_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1781
if (!*mem) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1785
INIT_LIST_HEAD(&(*mem)->attachments);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1786
mutex_init(&(*mem)->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1787
(*mem)->aql_queue = !!(flags & KFD_IOC_ALLOC_MEM_FLAGS_AQL_QUEUE_MEM);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1793
if ((*mem)->aql_queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1797
(*mem)->alloc_flags = flags;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1799
amdgpu_sync_create(&(*mem)->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1809
va, (*mem)->aql_queue ? size << 1 : size,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1824
ret = drm_gem_handle_create(adev->kfd.client.file, gobj, &(*mem)->gem_handle);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1832
bo->kfd_bo = *mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1833
(*mem)->bo = bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1837
(*mem)->va = va;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1838
(*mem)->domain = domain;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1839
(*mem)->mapped_to_gpu_memory = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1840
(*mem)->process_info = avm->process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1842
add_kgd_mem_to_kfd_bo_list(*mem, avm->process_info, user_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1846
ret = init_user_pages(*mem, user_addr, criu_resume);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1877
remove_kgd_mem_from_kfd_bo_list(*mem, avm->process_info);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1878
drm_gem_handle_delete(adev->kfd.client.file, (*mem)->gem_handle);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1887
amdgpu_sync_free(&(*mem)->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1888
mutex_destroy(&(*mem)->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1892
kfree(*mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1902
struct amdgpu_device *adev, struct kgd_mem *mem, void *drm_priv,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1905
struct amdkfd_process_info *process_info = mem->process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1906
unsigned long bo_size = mem->bo->tbo.base.size;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1907
bool use_release_notifier = (mem->bo->kfd_bo == mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1914
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1917
if (mem->alloc_flags &
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1920
amdgpu_amdkfd_gpuvm_unpin_bo(mem->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1923
mapped_to_gpu_memory = mem->mapped_to_gpu_memory;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1924
is_imported = mem->is_imported;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1925
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1932
mem->va, bo_size);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1938
if (!list_empty(&mem->validate_list))
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1939
list_del_init(&mem->validate_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1942
ret = reserve_bo_and_cond_vms(mem, NULL, BO_VM_ALL, &ctx);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1947
if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1948
amdgpu_hmm_unregister(mem->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1949
amdgpu_hmm_range_free(mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1950
mem->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1953
amdgpu_amdkfd_remove_eviction_fence(mem->bo,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1955
pr_debug("Release VA 0x%llx - 0x%llx\n", mem->va,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1956
mem->va + bo_size * (1 + mem->aql_queue));
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1959
list_for_each_entry_safe(entry, tmp, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1960
kfd_mem_dmaunmap_attachment(mem, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1967
amdgpu_sync_free(&mem->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1972
if (mem->bo->tbo.sg) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1973
sg_free_table(mem->bo->tbo.sg);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1974
kfree(mem->bo->tbo.sg);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1983
mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_VRAM)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1990
drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1991
drm_gem_handle_delete(adev->kfd.client.file, mem->gem_handle);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1992
if (mem->dmabuf) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1993
dma_buf_put(mem->dmabuf);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1994
mem->dmabuf = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1996
mutex_destroy(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2002
drm_gem_object_put(&mem->bo->tbo.base);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2010
kfree(mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2016
struct amdgpu_device *adev, struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2028
bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2038
mutex_lock(&mem->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2045
mutex_lock(&mem->process_info->notifier_lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2046
is_invalid_userptr = !!mem->invalid;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2047
mutex_unlock(&mem->process_info->notifier_lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2050
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2052
domain = mem->domain;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2056
mem->va,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2057
mem->va + bo_size * (1 + mem->aql_queue),
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2060
if (!kfd_mem_is_attached(avm, mem)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2061
ret = kfd_mem_attach(adev, mem, avm, mem->aql_queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2066
ret = reserve_bo_and_vm(mem, avm, &ctx);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2083
list_for_each_entry(entry, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2090
ret = map_bo_to_gpuvm(mem, entry, ctx.sync,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2104
mem->mapped_to_gpu_memory++;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2106
mem->mapped_to_gpu_memory);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2116
mutex_unlock(&mem->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2117
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2121
int amdgpu_amdkfd_gpuvm_dmaunmap_mem(struct kgd_mem *mem, void *drm_priv)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2129
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2131
ret = amdgpu_bo_reserve(mem->bo, true);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2135
list_for_each_entry(entry, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2142
kfd_mem_dmaunmap_attachment(mem, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2145
amdgpu_bo_unreserve(mem->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2147
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2153
struct amdgpu_device *adev, struct kgd_mem *mem, void *drm_priv)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2156
unsigned long bo_size = mem->bo->tbo.base.size;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2161
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2163
ret = reserve_bo_and_cond_vms(mem, avm, BO_VM_MAPPED, &ctx);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2177
mem->va,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2178
mem->va + bo_size * (1 + mem->aql_queue),
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2181
list_for_each_entry(entry, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2188
ret = unmap_bo_from_gpuvm(mem, entry, ctx.sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2194
mem->mapped_to_gpu_memory--;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2196
mem->mapped_to_gpu_memory);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2202
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2207
struct amdgpu_device *adev, struct kgd_mem *mem, bool intr)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2214
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2215
amdgpu_sync_clone(&mem->sync, &sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2216
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2284
int amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2288
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2295
mutex_lock(&mem->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2316
bo, mem->process_info->eviction_fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2323
mutex_unlock(&mem->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2331
mutex_unlock(&mem->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2344
void amdgpu_amdkfd_gpuvm_unmap_gtt_bo_from_kernel(struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2346
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2355
struct kfd_vm_fault_info *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2358
*mem = *adev->gmc.vm_fault_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2368
struct kgd_mem **mem, uint64_t *size,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2381
*mem = kzalloc_obj(struct kgd_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2382
if (!*mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2395
INIT_LIST_HEAD(&(*mem)->attachments);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2396
mutex_init(&(*mem)->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2398
(*mem)->alloc_flags =
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2405
(*mem)->dmabuf = dma_buf;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2406
(*mem)->bo = bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2407
(*mem)->va = va;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2408
(*mem)->domain = (bo->preferred_domains & AMDGPU_GEM_DOMAIN_VRAM) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2412
(*mem)->mapped_to_gpu_memory = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2413
(*mem)->process_info = avm->process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2414
add_kgd_mem_to_kfd_bo_list(*mem, avm->process_info, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2415
amdgpu_sync_create(&(*mem)->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2416
(*mem)->is_imported = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2421
ret = amdgpu_amdkfd_bo_validate_and_fence(bo, (*mem)->domain,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2430
remove_kgd_mem_from_kfd_bo_list(*mem, avm->process_info);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2433
kfree(*mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2439
struct kgd_mem **mem, uint64_t *size,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2456
ret = import_obj_create(adev, obj->dma_buf, obj, va, drm_priv, mem, size,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2461
(*mem)->gem_handle = handle;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2472
int amdgpu_amdkfd_gpuvm_export_dmabuf(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2477
mutex_lock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2478
ret = kfd_mem_export_dmabuf(mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2482
get_dma_buf(mem->dmabuf);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2483
*dma_buf = mem->dmabuf;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2485
mutex_unlock(&mem->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2500
unsigned long cur_seq, struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2502
struct amdkfd_process_info *process_info = mem->process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2514
mem->invalid++;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2542
struct kgd_mem *mem, *tmp_mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2551
list_for_each_entry_safe(mem, tmp_mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2554
if (mem->invalid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2555
list_move_tail(&mem->validate_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2559
list_for_each_entry(mem, &process_info->userptr_inval_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2561
invalid = mem->invalid;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2568
bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2570
amdgpu_hmm_range_free(mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2571
mem->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2594
mem->range = amdgpu_hmm_range_alloc(NULL);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2595
if (unlikely(!mem->range))
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2598
ret = amdgpu_ttm_tt_get_user_pages(bo, mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2600
amdgpu_hmm_range_free(mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2601
mem->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2621
pid_nr(process_info->pid), mem->va);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2626
kfd_signal_vm_fault_event_with_userptr(p, mem->va);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2634
amdgpu_ttm_tt_set_user_pages(bo->tbo.ttm, mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2641
if (mem->invalid != invalid) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2646
if (mem->range)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2647
mem->invalid = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2668
struct kgd_mem *mem, *tmp_mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2687
list_for_each_entry(mem, &process_info->userptr_inval_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2691
gobj = &mem->bo->tbo.base;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2704
list_for_each_entry_safe(mem, tmp_mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2709
bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2713
amdgpu_bo_placement_from_domain(bo, mem->domain);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2727
list_for_each_entry(attachment, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2731
kfd_mem_dmaunmap_attachment(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2732
ret = update_gpuvm_pte(mem, attachment, &sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2737
mem->invalid++;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2761
struct kgd_mem *mem, *tmp_mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2764
list_for_each_entry_safe(mem, tmp_mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2770
if (!mem->range)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2774
valid = amdgpu_hmm_range_valid(mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2775
amdgpu_hmm_range_free(mem->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2777
mem->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2779
WARN(!mem->invalid, "Invalid BO not marked invalid");
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2784
if (mem->invalid) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2790
list_move_tail(&mem->validate_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2920
struct kgd_mem *mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2947
list_for_each_entry(mem, &process_info->kfd_bo_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2951
gobj = &mem->bo->tbo.base;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2964
list_for_each_entry(mem, &process_info->kfd_bo_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2967
struct amdgpu_bo *bo = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
2968
uint32_t domain = mem->domain;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3008
list_for_each_entry(mem, &process_info->kfd_bo_list,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3012
list_for_each_entry(attachment, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3016
kfd_mem_dmaunmap_attachment(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3017
ret = update_gpuvm_pte(mem, attachment, &sync_obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3101
list_for_each_entry(mem, &process_info->kfd_bo_list, validate_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3102
if (mem->bo->tbo.pin_count)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3105
dma_resv_add_fence(mem->bo->tbo.base.resv,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3127
int amdgpu_amdkfd_add_gws_to_process(void *info, void *gws, struct kgd_mem **mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3136
*mem = kzalloc_obj(struct kgd_mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3137
if (!*mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3140
mutex_init(&(*mem)->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3141
INIT_LIST_HEAD(&(*mem)->attachments);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3142
(*mem)->bo = amdgpu_bo_ref(gws_bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3143
(*mem)->domain = AMDGPU_GEM_DOMAIN_GWS;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3144
(*mem)->process_info = process_info;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3145
add_kgd_mem_to_kfd_bo_list(*mem, process_info, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3146
amdgpu_sync_create(&(*mem)->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3150
mutex_lock(&(*mem)->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3173
mutex_unlock(&(*mem)->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3181
mutex_unlock(&(*mem)->process_info->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3182
amdgpu_sync_free(&(*mem)->sync);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3183
remove_kgd_mem_from_kfd_bo_list(*mem, process_info);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3185
mutex_destroy(&(*mem)->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3186
kfree(*mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3187
*mem = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3191
int amdgpu_amdkfd_remove_gws_from_process(void *info, void *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3195
struct kgd_mem *kgd_mem = (struct kgd_mem *)mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3215
kfree(mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3239
bool amdgpu_amdkfd_bo_mapped_to_dev(void *drm_priv, struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
3244
list_for_each_entry(entry, &mem->attachments, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
332
struct kgd_mem *mem, struct amdgpu_bo **bo_out)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
338
ret = amdgpu_bo_reserve(mem->bo, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
342
if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
343
flags |= mem->bo->flags & (AMDGPU_GEM_CREATE_COHERENT |
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
346
ret = amdgpu_gem_object_create(adev, mem->bo->tbo.base.size, 1,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
348
ttm_bo_type_sg, mem->bo->tbo.base.resv, &gem_obj, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
350
amdgpu_bo_unreserve(mem->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
358
(*bo_out)->parent = amdgpu_bo_ref(mem->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
514
struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
519
if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
521
if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
560
kfd_mem_dmamap_userptr(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
564
mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
569
struct ttm_tt *src_ttm = mem->bo->tbo.ttm;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
647
kfd_mem_dmamap_sg_bo(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
660
mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
666
dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
668
dma_addr = mem->bo->tbo.sg->sgl->dma_address;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
669
pr_debug("%d BO size: %d\n", mmio, mem->bo->tbo.sg->sgl->length);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
672
mem->bo->tbo.sg->sgl->length, dir, DMA_ATTR_SKIP_CPU_SYNC);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
678
ttm->sg = create_sg_table(dma_addr, mem->bo->tbo.sg->sgl->length);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
696
dma_unmap_resource(adev->dev, dma_addr, mem->bo->tbo.sg->sgl->length,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
702
kfd_mem_dmamap_attachment(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
709
return kfd_mem_dmamap_userptr(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
713
return kfd_mem_dmamap_sg_bo(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
721
kfd_mem_dmaunmap_userptr(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
725
mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
770
kfd_mem_dmaunmap_sg_bo(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
78
struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
787
dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
798
kfd_mem_dmaunmap_attachment(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
805
kfd_mem_dmaunmap_userptr(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
811
kfd_mem_dmaunmap_sg_bo(mem, attachment);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
818
static int kfd_mem_export_dmabuf(struct kgd_mem *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
82
list_for_each_entry(entry, &mem->attachments, list)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
820
if (!mem->dmabuf) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
824
bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
826
mem->gem_handle,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
827
mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
831
mem->dmabuf = dmabuf;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
838
kfd_mem_attach_dmabuf(struct amdgpu_device *adev, struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
844
ret = kfd_mem_export_dmabuf(mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
848
gobj = amdgpu_gem_prime_import(adev_to_drm(adev), mem->dmabuf);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
871
static int kfd_mem_attach(struct amdgpu_device *adev, struct kgd_mem *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
874
struct amdgpu_device *bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
875
unsigned long bo_size = mem->bo->tbo.base.size;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
876
uint64_t va = mem->va;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
898
((mem->domain == AMDGPU_GEM_DOMAIN_VRAM) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
899
(mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
900
(mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP))) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
901
if (mem->domain == AMDGPU_GEM_DOMAIN_VRAM)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
917
if ((adev == bo_adev && !(mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP)) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
918
(amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && reuse_dmamap(adev, bo_adev)) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
919
(mem->domain == AMDGPU_GEM_DOMAIN_GTT && reuse_dmamap(adev, bo_adev)) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
926
bo[i] = mem->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
933
} else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
936
ret = create_dmamap_sg_bo(adev, mem, &bo[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
940
} else if (mem->bo->tbo.type == ttm_bo_type_sg) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
941
WARN_ONCE(!(mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
942
mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP),
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
945
ret = create_dmamap_sg_bo(adev, mem, &bo[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
949
} else if (mem->domain == AMDGPU_GEM_DOMAIN_GTT ||
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
950
mem->domain == AMDGPU_GEM_DOMAIN_VRAM) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
952
ret = kfd_mem_attach_dmabuf(adev, mem, &bo[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
988
attachment[i]->pte_flags = get_pte_flags(adev, vm, mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
990
list_add(&attachment[i]->list, &mem->attachments);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
839
struct drm_amdgpu_memory_info mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
845
memset(&mem, 0, sizeof(mem));
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
846
mem.vram.total_heap_size = adev->gmc.real_vram_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
847
mem.vram.usable_heap_size = adev->gmc.real_vram_size -
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
850
mem.vram.heap_usage = ttm_resource_manager_used(&adev->mman.vram_mgr.manager) ?
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
852
mem.vram.max_allocation = mem.vram.usable_heap_size * 3 / 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
854
mem.cpu_accessible_vram.total_heap_size =
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
856
mem.cpu_accessible_vram.usable_heap_size =
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
859
mem.vram.usable_heap_size);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
860
mem.cpu_accessible_vram.heap_usage =
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
862
mem.cpu_accessible_vram.max_allocation =
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
863
mem.cpu_accessible_vram.usable_heap_size * 3 / 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
865
mem.gtt.total_heap_size = gtt_man->size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
866
mem.gtt.usable_heap_size = mem.gtt.total_heap_size -
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
868
mem.gtt.heap_usage = ttm_resource_manager_usage(gtt_man);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
869
mem.gtt.max_allocation = mem.gtt.usable_heap_size * 3 / 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
871
return copy_to_user(out, &mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
872
min((size_t)size, sizeof(mem)))
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
809
uint32_t *mem = (uint32_t *)(adev->mes.event_log_cpu_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
812
mem, adev->mes.event_log_size, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1403
uint64_t amdgpu_ttm_tt_pde_flags(struct ttm_tt *ttm, struct ttm_resource *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1407
if (mem && mem->mem_type != TTM_PL_SYSTEM)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1410
if (mem && (mem->mem_type == TTM_PL_TT ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1411
mem->mem_type == AMDGPU_PL_DOORBELL ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1412
mem->mem_type == AMDGPU_PL_PREEMPT ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1413
mem->mem_type == AMDGPU_PL_MMIO_REMAP)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1420
if (mem && mem->mem_type == TTM_PL_VRAM &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1421
mem->bus.caching == ttm_cached)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1437
struct ttm_resource *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1439
uint64_t flags = amdgpu_ttm_tt_pde_flags(ttm, mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
193
struct ttm_resource *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
209
if (WARN_ON(mem->mem_type == AMDGPU_PL_PREEMPT))
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
213
if (!tmz && mem->start != AMDGPU_BO_INVALID_OFFSET) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
214
*addr = amdgpu_ttm_domain_start(adev, mem->mem_type) +
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
255
flags = amdgpu_ttm_tt_pte_flags(adev, bo->ttm, mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
261
if (mem->mem_type == TTM_PL_TT) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
313
amdgpu_res_first(src->mem, src->offset, size, &src_mm);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
314
amdgpu_res_first(dst->mem, dst->offset, size, &dst_mm);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
326
r = amdgpu_ttm_map_buffer(entity, src->bo, src->mem, &src_mm,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
331
r = amdgpu_ttm_map_buffer(entity, dst->bo, dst->mem, &dst_mm,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
344
(dst->mem->mem_type == TTM_PL_VRAM)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
395
src.mem = old_mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
396
dst.mem = new_mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
480
struct ttm_resource *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
482
if (!amdgpu_res_cpu_visible(adev, mem))
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
486
if (mem->mem_type == TTM_PL_VRAM &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
487
!(mem->placement & TTM_PL_FLAG_CONTIGUOUS))
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
614
struct ttm_resource *mem)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
618
switch (mem->mem_type) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
626
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
629
mem->placement & TTM_PL_FLAG_CONTIGUOUS)
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
630
mem->bus.addr = (u8 *)adev->mman.aper_base_kaddr +
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
631
mem->bus.offset;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
633
mem->bus.offset += adev->gmc.aper_base;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
634
mem->bus.is_iomem = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
637
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
638
mem->bus.offset += adev->doorbell.base;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
639
mem->bus.is_iomem = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
640
mem->bus.caching = ttm_uncached;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
643
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
644
mem->bus.offset += adev->rmmio_remap.bus_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
645
mem->bus.is_iomem = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
646
mem->bus.caching = ttm_uncached;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
113
struct ttm_resource *mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
141
bool amdgpu_gtt_mgr_has_gart_addr(struct ttm_resource *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
154
struct ttm_resource *mem,
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
221
uint64_t amdgpu_ttm_tt_pde_flags(struct ttm_tt *ttm, struct ttm_resource *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
223
struct ttm_resource *mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1270
struct ttm_resource *mem;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1280
mem = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1295
mem = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1311
mem = bo->tbo.resource;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1312
if (mem && (mem->mem_type == TTM_PL_TT ||
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1313
mem->mem_type == AMDGPU_PL_PREEMPT))
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1326
flags = amdgpu_ttm_tt_pte_flags(adev, bo->tbo.ttm, mem);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1373
mapping->offset, vram_base, mem,
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1560
uint32_t bank, way, mem;
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1564
mem = instance % blk->num_mem_blocks;
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1573
bank, vml2_way_str[way], mem, sec_cnt, ded_cnt);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1583
bank, utcl2_router_str[mem], sec_cnt, ded_cnt);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1595
bank, way, mem, sec_cnt, ded_cnt);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1601
bank, way, mem, sec_cnt, ded_cnt);
drivers/gpu/drm/amd/amdgpu/vpe_6_1_fw_if.h
205
#define VPE_CMD_POLL_REGMEM_HEADER_MEM(mem) \
drivers/gpu/drm/amd/amdgpu/vpe_6_1_fw_if.h
206
(((mem) & VPE_CMD_POLL_REGMEM_HEADER_MEM_MASK) << VPE_CMD_POLL_REGMEM_HEADER_MEM__SHIFT)
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1063
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1162
pdd->drm_priv, (struct kgd_mem **) &mem, &offset,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1168
idr_handle = kfd_process_device_create_obj_handle(pdd, mem);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1198
amdgpu_amdkfd_gpuvm_free_memory_of_gpu(dev->adev, (struct kgd_mem *)mem,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1212
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1234
mem = kfd_process_device_translate_handle(
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1236
if (!mem) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1242
(struct kgd_mem *)mem, pdd->drm_priv, &size);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1264
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1306
mem = kfd_process_device_translate_handle(pdd,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1308
if (!mem) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1329
peer_pdd->dev->adev, (struct kgd_mem *)mem,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1340
((struct kgd_mem *)mem)->domain);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1346
err = amdgpu_amdkfd_gpuvm_sync_memory(dev->adev, (struct kgd_mem *) mem, true);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1382
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1416
mem = kfd_process_device_translate_handle(pdd,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1418
if (!mem) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1430
peer_pdd->dev->adev, (struct kgd_mem *)mem, peer_pdd->drm_priv);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1441
(struct kgd_mem *) mem, true);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1457
err = amdgpu_amdkfd_gpuvm_dmaunmap_mem(mem, peer_pdd->drm_priv);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1583
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1601
(struct kgd_mem **)&mem, &size,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1606
idr_handle = kfd_process_device_create_obj_handle(pdd, mem);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1619
amdgpu_amdkfd_gpuvm_free_memory_of_gpu(pdd->dev->adev, (struct kgd_mem *)mem,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1633
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1648
mem = kfd_process_device_translate_handle(pdd,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1650
if (!mem) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1655
ret = amdgpu_amdkfd_gpuvm_export_dmabuf(mem, &dmabuf);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1858
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1861
idr_for_each_entry(&pdd->alloc_idr, mem, id) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1862
struct kgd_mem *kgd_mem = (struct kgd_mem *)mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1871
static int criu_get_prime_handle(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1878
ret = amdgpu_amdkfd_gpuvm_export_dmabuf(mem, &dmabuf);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1928
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1951
idr_for_each_entry(&pdd->alloc_idr, mem, id) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1956
kgd_mem = (struct kgd_mem *)mem;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1092
static int kfd_parse_subtype_mem(struct crat_subtype_memory *mem,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1103
mem->proximity_domain);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1105
if (mem->proximity_domain == dev->proximity_domain) {
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1109
if (mem->visibility_type == 0)
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1114
heap_type = mem->visibility_type;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1118
if (mem->flags & CRAT_MEM_FLAGS_HOT_PLUGGABLE)
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1120
if (mem->flags & CRAT_MEM_FLAGS_NON_VOLATILE)
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1124
((uint64_t)mem->length_high << 32) +
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1125
mem->length_low;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1126
width = mem->width;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1311
struct crat_subtype_memory *mem;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1322
mem = (struct crat_subtype_memory *)sub_type_hdr;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1323
ret = kfd_parse_subtype_mem(mem, device_list);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2904
&(mem_obj->mem), &(mem_obj->gpu_addr),
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2915
amdgpu_amdkfd_free_kernel_mem(dev->adev, &mqd->mem);
drivers/gpu/drm/amd/amdkfd/kfd_events.c
360
void *mem, *kern_addr;
drivers/gpu/drm/amd/amdkfd/kfd_events.c
380
mem = kfd_process_device_translate_handle(pdd,
drivers/gpu/drm/amd/amdkfd/kfd_events.c
382
if (!mem) {
drivers/gpu/drm/amd/amdkfd/kfd_events.c
387
err = amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel(mem, &kern_addr, &size);
drivers/gpu/drm/amd/amdkfd/kfd_events.c
396
amdgpu_amdkfd_gpuvm_unmap_gtt_bo_from_kernel(mem);
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
229
if (mqd_mem_obj->mem) {
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
230
amdgpu_amdkfd_free_kernel_mem(mm->dev->adev, &mqd_mem_obj->mem);
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
285
mqd_mem_obj->mem = (virtual_xcc_id == 0) ?
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
286
dev->dqm->hiq_sdma_mqd.mem : NULL;
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
58
mqd_mem_obj->mem = dev->dqm->hiq_sdma_mqd.mem;
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
84
mqd_mem_obj->mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.mem
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
96
WARN_ON(!mqd_mem_obj->mem);
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v12_1.c
441
xcc_mqd_mem_obj->mem = (offset == 0) ?
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v12_1.c
442
mqd_mem_obj->mem : NULL;
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
166
&(mqd_mem_obj->mem),
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
692
xcc_mqd_mem_obj->mem = (offset == 0) ?
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v9.c
693
mqd_mem_obj->mem : NULL;
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1110
void *mem);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
255
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1043
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1051
idr_for_each_entry(&pdd->alloc_idr, mem, id) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1059
peer_pdd->dev->adev, mem, peer_pdd->drm_priv);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1062
amdgpu_amdkfd_gpuvm_free_memory_of_gpu(pdd->dev->adev, mem,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1076
void *mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1088
mem = kfd_process_device_translate_handle(
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1090
if (!mem)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1093
amdgpu_amdkfd_gpuvm_unmap_gtt_bo_from_kernel(mem);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1461
struct kgd_mem *mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1470
KFD_CWSR_TBA_TMA_SIZE, flags, &mem, &kaddr);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1474
qpd->cwsr_mem = mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1874
void *mem)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1876
return idr_alloc(&pdd->alloc_idr, mem, 0, 0, GFP_KERNEL);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
717
static void kfd_process_free_gpuvm(struct kgd_mem *mem,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
723
amdgpu_amdkfd_gpuvm_unmap_gtt_bo_from_kernel(mem);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
727
amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu(dev->adev, mem, pdd->drm_priv);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
728
amdgpu_amdkfd_gpuvm_free_memory_of_gpu(dev->adev, mem, pdd->drm_priv,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
740
uint32_t flags, struct kgd_mem **mem, void **kptr)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
746
pdd->drm_priv, mem, NULL,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
751
err = amdgpu_amdkfd_gpuvm_map_memory_to_gpu(kdev->adev, *mem,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
756
err = amdgpu_amdkfd_gpuvm_sync_memory(kdev->adev, *mem, true);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
764
(struct kgd_mem *)*mem, kptr, NULL);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
774
amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu(kdev->adev, *mem, pdd->drm_priv);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
777
amdgpu_amdkfd_gpuvm_free_memory_of_gpu(kdev->adev, *mem, pdd->drm_priv,
drivers/gpu/drm/amd/amdkfd/kfd_process.c
780
*mem = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
798
struct kgd_mem *mem;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
807
&mem, &kaddr);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
811
qpd->ib_mem = mem;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
111
struct kgd_mem *mem = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
144
gws, &mem);
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
150
pqn->q->gws = mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
114
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1148
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1164
list_for_each_entry(mem, &dev->mem_props, list)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1165
mem->gpu = dev->gpu;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1191
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1206
list_for_each_entry(mem, &dev->mem_props, list)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1207
mem->mem_clk_max = local_mem_info.mem_clk_max;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
123
mem = container_of(dev->mem_props.next,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
125
list_del(&mem->list);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
126
kfree(mem);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
297
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
302
mem = container_of(attr, struct kfd_mem_properties, attr);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
303
if (mem->gpu && kfd_devcgroup_check_permission(mem->gpu))
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
305
sysfs_show_32bit_prop(buffer, offs, "heap_type", mem->heap_type);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
307
mem->size_in_bytes);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
308
sysfs_show_32bit_prop(buffer, offs, "flags", mem->flags);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
309
sysfs_show_32bit_prop(buffer, offs, "width", mem->width);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
311
mem->mem_clk_max);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
571
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
611
list_for_each_entry(mem, &dev->mem_props, list)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
612
if (mem->kobj) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
613
kfd_remove_sysfs_file(mem->kobj, &mem->attr);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
614
mem->kobj = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
647
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
713
list_for_each_entry(mem, &dev->mem_props, list) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
714
mem->kobj = kzalloc_obj(struct kobject);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
715
if (!mem->kobj)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
717
ret = kobject_init_and_add(mem->kobj, &mem_type,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
720
kobject_put(mem->kobj);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
724
mem->attr.name = "properties";
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
725
mem->attr.mode = KFD_SYSFS_FILE_MODE;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
726
sysfs_attr_init(&mem->attr);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
727
ret = sysfs_create_file(mem->kobj, &mem->attr);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
965
struct kfd_mem_properties *mem;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
974
list_for_each_entry(mem, &kdev->mem_props, list) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
976
mem->width = memdev->total_width;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
978
mem->mem_clk_max = memdev->speed;
drivers/gpu/drm/armada/armada_drv.c
118
drm_mm_init(&priv->linear, mem->start, resource_size(mem));
drivers/gpu/drm/armada/armada_drv.c
63
struct resource *mem = NULL;
drivers/gpu/drm/armada/armada_drv.c
74
mem = r;
drivers/gpu/drm/armada/armada_drv.c
79
if (!mem)
drivers/gpu/drm/armada/armada_drv.c
82
if (!devm_request_mem_region(dev, mem->start, resource_size(mem),
drivers/gpu/drm/drm_format_helper.c
111
kfree(state->tmp.mem);
drivers/gpu/drm/drm_format_helper.c
112
state->tmp.mem = NULL;
drivers/gpu/drm/drm_format_helper.c
35
state->tmp.mem = NULL;
drivers/gpu/drm/drm_format_helper.c
56
state->tmp.mem = NULL;
drivers/gpu/drm/drm_format_helper.c
79
void *mem;
drivers/gpu/drm/drm_format_helper.c
86
mem = krealloc(state->tmp.mem, new_size, flags);
drivers/gpu/drm/drm_format_helper.c
87
if (!mem)
drivers/gpu/drm/drm_format_helper.c
90
state->tmp.mem = mem;
drivers/gpu/drm/drm_format_helper.c
94
return state->tmp.mem;
drivers/gpu/drm/drm_gem_vram_helper.c
791
struct ttm_resource *mem)
drivers/gpu/drm/drm_gem_vram_helper.c
795
switch (mem->mem_type) {
drivers/gpu/drm/drm_gem_vram_helper.c
799
mem->bus.offset = (mem->start << PAGE_SHIFT) + vmm->vram_base;
drivers/gpu/drm/drm_gem_vram_helper.c
800
mem->bus.is_iomem = true;
drivers/gpu/drm/drm_gem_vram_helper.c
801
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/hyperv/hyperv_drm.h
27
struct resource *mem;
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
101
vmbus_free_mmio(hv->mem->start, hv->fb_size);
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
159
vmbus_free_mmio(hv->mem->start, hv->fb_size);
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
178
vmbus_free_mmio(hv->mem->start, hv->fb_size);
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
78
ret = vmbus_allocate_mmio(&hv->mem, hdev, 0, -1, hv->fb_size, 0x100000,
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
90
hv->vram = ioremap_cache(hv->mem->start, hv->fb_size);
drivers/gpu/drm/hyperv/hyperv_drm_drv.c
97
hv->fb_base = hv->mem->start;
drivers/gpu/drm/i915/display/intel_display_types.h
758
struct intel_memory_region *mem;
drivers/gpu/drm/i915/display/intel_fbdev_fb.c
67
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/display/intel_fbdev_fb.c
71
(unsigned long)(mem->io.start +
drivers/gpu/drm/i915/display/intel_fbdev_fb.c
73
mem->region.start);
drivers/gpu/drm/i915/gem/i915_gem_region.c
107
i915_gem_object_create_region(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_region.c
112
return __i915_gem_object_create_region(mem, I915_BO_INVALID_OFFSET,
drivers/gpu/drm/i915/gem/i915_gem_region.c
117
i915_gem_object_create_region_at(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_region.c
124
if (GEM_WARN_ON(!IS_ALIGNED(size, mem->min_page_size)) ||
drivers/gpu/drm/i915/gem/i915_gem_region.c
125
GEM_WARN_ON(!IS_ALIGNED(offset, mem->min_page_size)))
drivers/gpu/drm/i915/gem/i915_gem_region.c
128
if (range_overflows(offset, size, resource_size(&mem->region)))
drivers/gpu/drm/i915/gem/i915_gem_region.c
132
offset + size > resource_size(&mem->io) &&
drivers/gpu/drm/i915/gem/i915_gem_region.c
133
!i915_ggtt_has_aperture(to_gt(mem->i915)->ggtt))
drivers/gpu/drm/i915/gem/i915_gem_region.c
136
return __i915_gem_object_create_region(mem, offset, size, 0,
drivers/gpu/drm/i915/gem/i915_gem_region.c
14
struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_region.c
16
obj->mm.region = mem;
drivers/gpu/drm/i915/gem/i915_gem_region.c
18
mutex_lock(&mem->objects.lock);
drivers/gpu/drm/i915/gem/i915_gem_region.c
19
list_add(&obj->mm.region_link, &mem->objects.list);
drivers/gpu/drm/i915/gem/i915_gem_region.c
20
mutex_unlock(&mem->objects.lock);
drivers/gpu/drm/i915/gem/i915_gem_region.c
25
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/gem/i915_gem_region.c
27
mutex_lock(&mem->objects.lock);
drivers/gpu/drm/i915/gem/i915_gem_region.c
29
mutex_unlock(&mem->objects.lock);
drivers/gpu/drm/i915/gem/i915_gem_region.c
33
__i915_gem_object_create_region(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_region.c
56
if (!mem)
drivers/gpu/drm/i915/gem/i915_gem_region.c
59
default_page_size = mem->min_page_size;
drivers/gpu/drm/i915/gem/i915_gem_region.c
91
if (default_page_size < mem->min_page_size)
drivers/gpu/drm/i915/gem/i915_gem_region.c
94
err = mem->ops->init_object(mem, obj, offset, size, page_size, flags);
drivers/gpu/drm/i915/gem/i915_gem_region.h
51
struct intel_memory_region *mem);
drivers/gpu/drm/i915/gem/i915_gem_region.h
55
i915_gem_object_create_region(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_region.h
60
i915_gem_object_create_region_at(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
216
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
235
ret = shmem_sg_alloc_table(i915, st, obj->base.size, mem, mapping,
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
548
static int shmem_object_init(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
556
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
608
i915_gem_object_init_memory_region(obj, mem);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
656
static int init_shmem(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
658
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
685
intel_memory_region_set_name(mem, "system");
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1000
return mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1007
struct intel_memory_region *mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1009
mem = intel_memory_region_create(i915,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1014
if (IS_ERR(mem))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1015
return mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1017
intel_memory_region_set_name(mem, "stolen-system");
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1019
mem->private = true;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
1021
return mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
536
static int i915_gem_init_stolen(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
538
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
556
if (adjust_stolen(i915, &mem->region))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
559
if (request_smem_stolen(i915, &mem->region))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
562
i915->dsm.stolen = mem->region;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
568
mem->region.end = i915->dsm.reserved.start - 1;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
569
mem->io = DEFINE_RES_MEM(mem->io.start,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
570
min(resource_size(&mem->io),
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
571
resource_size(&mem->region)));
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
573
i915->dsm.usable_size = resource_size(&mem->region);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
725
static int __i915_gem_object_create_stolen(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
740
drm_gem_private_object_init(&mem->i915->drm, &obj->base, stolen->size);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
745
cache_level = HAS_LLC(mem->i915) ? I915_CACHE_LLC : I915_CACHE_NONE;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
751
i915_gem_object_init_memory_region(obj, mem);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
761
static int _i915_gem_object_stolen_init(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
768
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
782
if (mem->type == INTEL_MEMORY_STOLEN_LOCAL && !resource_size(&mem->io) &&
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
802
mem->min_page_size);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
807
ret = __i915_gem_object_create_stolen(mem, obj, stolen);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
827
static int init_stolen_smem(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
835
err = i915_gem_init_stolen(mem);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
837
drm_dbg(&mem->i915->drm, "Skip stolen region: failed to setup\n");
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
842
static int release_stolen_smem(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
844
i915_gem_cleanup_stolen(mem->i915);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
854
static int init_stolen_lmem(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
858
if (GEM_WARN_ON(resource_size(&mem->region) == 0))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
861
err = i915_gem_init_stolen(mem);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
863
drm_dbg(&mem->i915->drm, "Skip stolen region: failed to setup\n");
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
867
if (resource_size(&mem->io) &&
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
868
!io_mapping_init_wc(&mem->iomap, mem->io.start, resource_size(&mem->io)))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
874
i915_gem_cleanup_stolen(mem->i915);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
878
static int release_stolen_lmem(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
880
if (resource_size(&mem->io))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
881
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
882
i915_gem_cleanup_stolen(mem->i915);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
922
struct intel_memory_region *mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
988
mem = intel_memory_region_create(i915, dsm_base, dsm_size,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
993
if (IS_ERR(mem))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
994
return mem;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
996
intel_memory_region_set_name(mem, "stolen-local");
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
998
mem->private = true;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1284
int __i915_gem_ttm_object_init(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1292
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1306
obj->mm.region = mem;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1353
i915_gem_object_init_memory_region(obj, mem);
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
656
static int i915_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource *mem)
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
658
struct drm_i915_gem_object *obj = i915_ttm_to_gem(mem->bo);
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
661
if (i915_ttm_is_ghost_object(mem->bo))
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
674
if (!i915_ttm_cpu_maps_iomem(mem))
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
677
if (!i915_ttm_resource_mappable(mem))
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
680
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
681
mem->bus.is_iomem = true;
drivers/gpu/drm/i915/gem/i915_gem_ttm.h
102
return mem && mem->mem_type != I915_PL_SYSTEM;
drivers/gpu/drm/i915/gem/i915_gem_ttm.h
55
int __i915_gem_ttm_object_init(struct intel_memory_region *mem,
drivers/gpu/drm/i915/gem/i915_gem_ttm.h
88
static inline bool i915_ttm_gtt_binds_lmem(struct ttm_resource *mem)
drivers/gpu/drm/i915/gem/i915_gem_ttm.h
90
return mem->mem_type != I915_PL_SYSTEM;
drivers/gpu/drm/i915/gem/i915_gem_ttm.h
99
static inline bool i915_ttm_cpu_maps_iomem(struct ttm_resource *mem)
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
503
struct intel_memory_region *mem;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
509
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
510
if (IS_ERR(mem)) {
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
512
return PTR_ERR(mem);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
521
obj = i915_gem_object_create_region(mem,
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
572
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/gt/intel_gt.c
103
mem->id = id;
drivers/gpu/drm/i915/gt/intel_gt.c
104
mem->instance = instance;
drivers/gpu/drm/i915/gt/intel_gt.c
106
intel_memory_region_set_name(mem, "local%u", mem->instance);
drivers/gpu/drm/i915/gt/intel_gt.c
110
i915->mm.regions[id] = mem;
drivers/gpu/drm/i915/gt/intel_gt.c
89
struct intel_memory_region *mem;
drivers/gpu/drm/i915/gt/intel_gt.c
92
mem = intel_gt_setup_lmem(gt);
drivers/gpu/drm/i915/gt/intel_gt.c
93
if (IS_ERR(mem)) {
drivers/gpu/drm/i915/gt/intel_gt.c
94
err = PTR_ERR(mem);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
117
region_lmem_release(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gt/intel_region_lmem.c
121
ret = intel_region_ttm_fini(mem);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
122
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
128
region_lmem_init(struct intel_memory_region *mem)
drivers/gpu/drm/i915/gt/intel_region_lmem.c
132
if (!io_mapping_init_wc(&mem->iomap,
drivers/gpu/drm/i915/gt/intel_region_lmem.c
133
mem->io.start,
drivers/gpu/drm/i915/gt/intel_region_lmem.c
134
resource_size(&mem->io)))
drivers/gpu/drm/i915/gt/intel_region_lmem.c
137
ret = intel_region_ttm_init(mem);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
144
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
171
struct intel_memory_region *mem)
drivers/gpu/drm/i915/gt/intel_region_lmem.c
180
ret = intel_memory_region_reserve(mem, reserve_start, reserve_size);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
192
struct intel_memory_region *mem;
drivers/gpu/drm/i915/gt/intel_region_lmem.c
246
mem = intel_memory_region_create(i915,
drivers/gpu/drm/i915/gt/intel_region_lmem.c
255
if (IS_ERR(mem))
drivers/gpu/drm/i915/gt/intel_region_lmem.c
256
return mem;
drivers/gpu/drm/i915/gt/intel_region_lmem.c
258
err = reserve_lowmem_region(uncore, mem);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
266
return mem;
drivers/gpu/drm/i915/gt/intel_region_lmem.c
269
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/gvt/firmware.c
137
const void *mem;
drivers/gpu/drm/i915/gvt/firmware.c
144
mem = fw->data + crc32_start;
drivers/gpu/drm/i915/gvt/firmware.c
154
VERIFY("crc32", h->crc32, crc32_le(0, mem, fw->size - crc32_start));
drivers/gpu/drm/i915/gvt/firmware.c
158
mem = (fw->data + h->cfg_space_offset);
drivers/gpu/drm/i915/gvt/firmware.c
160
id = *(u16 *)(mem + PCI_VENDOR_ID);
drivers/gpu/drm/i915/gvt/firmware.c
163
id = *(u16 *)(mem + PCI_DEVICE_ID);
drivers/gpu/drm/i915/gvt/firmware.c
166
id = *(u8 *)(mem + PCI_REVISION_ID);
drivers/gpu/drm/i915/gvt/firmware.c
193
void *mem;
drivers/gpu/drm/i915/gvt/firmware.c
200
mem = kmalloc(info->cfg_space_size, GFP_KERNEL);
drivers/gpu/drm/i915/gvt/firmware.c
201
if (!mem) {
drivers/gpu/drm/i915/gvt/firmware.c
206
firmware->cfg_space = mem;
drivers/gpu/drm/i915/gvt/firmware.c
208
mem = vmalloc(info->mmio_size);
drivers/gpu/drm/i915/gvt/firmware.c
209
if (!mem) {
drivers/gpu/drm/i915/gvt/firmware.c
215
firmware->mmio = mem;
drivers/gpu/drm/i915/gvt/gtt.c
1359
ops->get_entry(oos_page->mem, &old, index, false, 0, vgpu);
drivers/gpu/drm/i915/gvt/gtt.c
1375
ops->set_entry(oos_page->mem, &new, index, false, 0, vgpu);
drivers/gpu/drm/i915/gvt/gtt.c
1410
oos_page->mem, I915_GTT_PAGE_SIZE);
drivers/gpu/drm/i915/gvt/gtt.c
1679
ops->set_entry(spt->guest_page.oos_page->mem, &we, index,
drivers/gpu/drm/i915/gvt/gtt.c
2489
free_page((unsigned long)oos_page->mem);
drivers/gpu/drm/i915/gvt/gtt.c
2510
oos_page->mem = (void *)__get_free_pages(GFP_KERNEL, 0);
drivers/gpu/drm/i915/gvt/gtt.c
2511
if (!oos_page->mem) {
drivers/gpu/drm/i915/gvt/gtt.h
235
void *mem;
drivers/gpu/drm/i915/i915_gpu_error.c
1208
struct intel_memory_region *mem = vma_res->mr;
drivers/gpu/drm/i915/i915_gpu_error.c
1212
dma_addr_t offset = dma - mem->region.start;
drivers/gpu/drm/i915/i915_gpu_error.c
1215
if (offset + PAGE_SIZE > resource_size(&mem->io)) {
drivers/gpu/drm/i915/i915_gpu_error.c
1220
s = io_mapping_map_wc(&mem->iomap, offset, PAGE_SIZE);
drivers/gpu/drm/i915/i915_initial_plane.c
105
mem = plane_config->mem;
drivers/gpu/drm/i915/i915_initial_plane.c
109
mem->min_page_size);
drivers/gpu/drm/i915/i915_initial_plane.c
118
mem == i915->mm.stolen_region &&
drivers/gpu/drm/i915/i915_initial_plane.c
124
obj = i915_gem_object_create_region_at(mem, phys_base, size,
drivers/gpu/drm/i915/i915_initial_plane.c
129
mem->region.name);
drivers/gpu/drm/i915/i915_initial_plane.c
39
struct intel_memory_region *mem;
drivers/gpu/drm/i915/i915_initial_plane.c
46
mem = intel_memory_region_by_type(i915, mem_type);
drivers/gpu/drm/i915/i915_initial_plane.c
47
if (!mem) {
drivers/gpu/drm/i915/i915_initial_plane.c
63
if (intel_memory_type_is_local(mem->type) != is_local) {
drivers/gpu/drm/i915/i915_initial_plane.c
65
mem->region.name);
drivers/gpu/drm/i915/i915_initial_plane.c
69
if (dma_addr < mem->region.start || dma_addr > mem->region.end) {
drivers/gpu/drm/i915/i915_initial_plane.c
72
&dma_addr, mem->region.name, &mem->region.start, &mem->region.end);
drivers/gpu/drm/i915/i915_initial_plane.c
79
plane_config->phys_base = dma_addr - mem->region.start;
drivers/gpu/drm/i915/i915_initial_plane.c
80
plane_config->mem = mem;
drivers/gpu/drm/i915/i915_initial_plane.c
89
struct intel_memory_region *mem;
drivers/gpu/drm/i915/intel_gvt.c
114
void *mem;
drivers/gpu/drm/i915/intel_gvt.c
117
mem = kzalloc(PCI_CFG_SPACE_EXP_SIZE, GFP_KERNEL);
drivers/gpu/drm/i915/intel_gvt.c
118
if (!mem)
drivers/gpu/drm/i915/intel_gvt.c
121
vgpu->initial_cfg_space = mem;
drivers/gpu/drm/i915/intel_gvt.c
124
pci_read_config_dword(pdev, i, mem + i);
drivers/gpu/drm/i915/intel_gvt.c
126
mem = vzalloc(2 * SZ_1M);
drivers/gpu/drm/i915/intel_gvt.c
127
if (!mem) {
drivers/gpu/drm/i915/intel_gvt.c
132
vgpu->initial_mmio = mem;
drivers/gpu/drm/i915/intel_memory_region.c
106
if (resource_size(&mem->io) < PAGE_SIZE)
drivers/gpu/drm/i915/intel_memory_region.c
109
last = resource_size(&mem->io) - PAGE_SIZE;
drivers/gpu/drm/i915/intel_memory_region.c
124
err = iopagetest(mem, page, caller);
drivers/gpu/drm/i915/intel_memory_region.c
129
err = iopagetest(mem, 0, caller);
drivers/gpu/drm/i915/intel_memory_region.c
133
err = iopagetest(mem, last, caller);
drivers/gpu/drm/i915/intel_memory_region.c
137
err = iopagetest(mem, random_page(last), caller);
drivers/gpu/drm/i915/intel_memory_region.c
194
int intel_memory_region_reserve(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.c
198
struct ttm_resource_manager *man = mem->region_private;
drivers/gpu/drm/i915/intel_memory_region.c
200
GEM_BUG_ON(mem->is_range_manager);
drivers/gpu/drm/i915/intel_memory_region.c
216
static int intel_memory_region_memtest(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.c
219
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/intel_memory_region.c
222
if (!mem->io.start)
drivers/gpu/drm/i915/intel_memory_region.c
226
err = iomemtest(mem, i915->params.memtest, caller);
drivers/gpu/drm/i915/intel_memory_region.c
258
struct intel_memory_region *mem;
drivers/gpu/drm/i915/intel_memory_region.c
261
mem = kzalloc_obj(*mem);
drivers/gpu/drm/i915/intel_memory_region.c
262
if (!mem)
drivers/gpu/drm/i915/intel_memory_region.c
265
mem->i915 = i915;
drivers/gpu/drm/i915/intel_memory_region.c
266
mem->region = DEFINE_RES_MEM(start, size);
drivers/gpu/drm/i915/intel_memory_region.c
267
mem->io = DEFINE_RES_MEM(io_start, io_size);
drivers/gpu/drm/i915/intel_memory_region.c
268
mem->min_page_size = min_page_size;
drivers/gpu/drm/i915/intel_memory_region.c
269
mem->ops = ops;
drivers/gpu/drm/i915/intel_memory_region.c
270
mem->total = size;
drivers/gpu/drm/i915/intel_memory_region.c
271
mem->type = type;
drivers/gpu/drm/i915/intel_memory_region.c
272
mem->instance = instance;
drivers/gpu/drm/i915/intel_memory_region.c
274
snprintf(mem->uabi_name, sizeof(mem->uabi_name), "%s%u",
drivers/gpu/drm/i915/intel_memory_region.c
277
mutex_init(&mem->objects.lock);
drivers/gpu/drm/i915/intel_memory_region.c
278
INIT_LIST_HEAD(&mem->objects.list);
drivers/gpu/drm/i915/intel_memory_region.c
281
err = ops->init(mem);
drivers/gpu/drm/i915/intel_memory_region.c
286
err = intel_memory_region_memtest(mem, (void *)_RET_IP_);
drivers/gpu/drm/i915/intel_memory_region.c
290
return mem;
drivers/gpu/drm/i915/intel_memory_region.c
293
if (mem->ops->release)
drivers/gpu/drm/i915/intel_memory_region.c
294
mem->ops->release(mem);
drivers/gpu/drm/i915/intel_memory_region.c
296
kfree(mem);
drivers/gpu/drm/i915/intel_memory_region.c
300
void intel_memory_region_set_name(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.c
306
vsnprintf(mem->name, sizeof(mem->name), fmt, ap);
drivers/gpu/drm/i915/intel_memory_region.c
324
void intel_memory_region_destroy(struct intel_memory_region *mem)
drivers/gpu/drm/i915/intel_memory_region.c
328
if (mem->ops->release)
drivers/gpu/drm/i915/intel_memory_region.c
329
ret = mem->ops->release(mem);
drivers/gpu/drm/i915/intel_memory_region.c
331
GEM_WARN_ON(!list_empty_careful(&mem->objects.list));
drivers/gpu/drm/i915/intel_memory_region.c
332
mutex_destroy(&mem->objects.lock);
drivers/gpu/drm/i915/intel_memory_region.c
334
kfree(mem);
drivers/gpu/drm/i915/intel_memory_region.c
344
struct intel_memory_region *mem = ERR_PTR(-ENODEV);
drivers/gpu/drm/i915/intel_memory_region.c
355
mem = i915_gem_ttm_system_setup(i915, type,
drivers/gpu/drm/i915/intel_memory_region.c
358
mem = i915_gem_shmem_setup(i915, type,
drivers/gpu/drm/i915/intel_memory_region.c
362
mem = i915_gem_stolen_lmem_setup(i915, type, instance);
drivers/gpu/drm/i915/intel_memory_region.c
363
if (!IS_ERR(mem))
drivers/gpu/drm/i915/intel_memory_region.c
364
i915->mm.stolen_region = mem;
drivers/gpu/drm/i915/intel_memory_region.c
367
mem = i915_gem_stolen_smem_setup(i915, type, instance);
drivers/gpu/drm/i915/intel_memory_region.c
368
if (!IS_ERR(mem))
drivers/gpu/drm/i915/intel_memory_region.c
369
i915->mm.stolen_region = mem;
drivers/gpu/drm/i915/intel_memory_region.c
37
static int __iopagetest(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.c
375
if (IS_ERR(mem)) {
drivers/gpu/drm/i915/intel_memory_region.c
376
err = PTR_ERR(mem);
drivers/gpu/drm/i915/intel_memory_region.c
383
if (mem) { /* Skip on non-fatal errors */
drivers/gpu/drm/i915/intel_memory_region.c
384
mem->id = i;
drivers/gpu/drm/i915/intel_memory_region.c
385
i915->mm.regions[i] = mem;
drivers/gpu/drm/i915/intel_memory_region.c
390
struct intel_memory_region *mem = i915->mm.regions[i];
drivers/gpu/drm/i915/intel_memory_region.c
393
if (!mem)
drivers/gpu/drm/i915/intel_memory_region.c
396
region_size = resource_size(&mem->region) >> 20;
drivers/gpu/drm/i915/intel_memory_region.c
397
io_size = resource_size(&mem->io) >> 20;
drivers/gpu/drm/i915/intel_memory_region.c
399
if (resource_size(&mem->io))
drivers/gpu/drm/i915/intel_memory_region.c
401
mem->id, mem->name, region_size, &mem->region, io_size, &mem->io);
drivers/gpu/drm/i915/intel_memory_region.c
404
mem->id, mem->name, region_size, &mem->region);
drivers/gpu/drm/i915/intel_memory_region.c
52
dev_err(mem->i915->drm.dev,
drivers/gpu/drm/i915/intel_memory_region.c
54
&mem->region, &mem->io.start, &offset, caller,
drivers/gpu/drm/i915/intel_memory_region.c
62
static int iopagetest(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.c
71
va = ioremap_wc(mem->io.start + offset, PAGE_SIZE);
drivers/gpu/drm/i915/intel_memory_region.c
73
dev_err(mem->i915->drm.dev,
drivers/gpu/drm/i915/intel_memory_region.c
75
&mem->io.start, &offset, caller);
drivers/gpu/drm/i915/intel_memory_region.c
80
err = __iopagetest(mem, va, PAGE_SIZE, val[i], offset, caller);
drivers/gpu/drm/i915/intel_memory_region.c
84
err = __iopagetest(mem, va, PAGE_SIZE, ~val[i], offset, caller);
drivers/gpu/drm/i915/intel_memory_region.c
99
static int iomemtest(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.h
105
void intel_memory_region_destroy(struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_memory_region.h
115
intel_memory_region_set_name(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.h
118
int intel_memory_region_reserve(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_memory_region.h
48
int (*init)(struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_memory_region.h
49
int (*release)(struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_memory_region.h
51
int (*init_object)(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.c
108
int intel_region_ttm_fini(struct intel_memory_region *mem)
drivers/gpu/drm/i915/intel_region_ttm.c
110
struct ttm_resource_manager *man = mem->region_private;
drivers/gpu/drm/i915/intel_region_ttm.c
124
i915_gem_flush_free_objects(mem->i915);
drivers/gpu/drm/i915/intel_region_ttm.c
126
mutex_lock(&mem->objects.lock);
drivers/gpu/drm/i915/intel_region_ttm.c
127
if (list_empty(&mem->objects.list))
drivers/gpu/drm/i915/intel_region_ttm.c
129
mutex_unlock(&mem->objects.lock);
drivers/gpu/drm/i915/intel_region_ttm.c
134
drain_workqueue(mem->i915->bdev.wq);
drivers/gpu/drm/i915/intel_region_ttm.c
141
ret = i915_ttm_buddy_man_fini(&mem->i915->bdev,
drivers/gpu/drm/i915/intel_region_ttm.c
142
intel_region_to_ttm_type(mem));
drivers/gpu/drm/i915/intel_region_ttm.c
144
mem->region_private = NULL;
drivers/gpu/drm/i915/intel_region_ttm.c
163
intel_region_ttm_resource_to_rsgt(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.c
167
if (mem->is_range_manager) {
drivers/gpu/drm/i915/intel_region_ttm.c
172
mem->region.start,
drivers/gpu/drm/i915/intel_region_ttm.c
175
return i915_rsgt_from_buddy_resource(res, mem->region.start,
drivers/gpu/drm/i915/intel_region_ttm.c
198
intel_region_ttm_resource_alloc(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.c
203
struct ttm_resource_manager *man = mem->region_private;
drivers/gpu/drm/i915/intel_region_ttm.c
222
} else if (resource_size(&mem->io) && resource_size(&mem->io) < mem->total) {
drivers/gpu/drm/i915/intel_region_ttm.c
227
if (WARN_ON(overflows_type(resource_size(&mem->io) >> PAGE_SHIFT, place.lpfn))) {
drivers/gpu/drm/i915/intel_region_ttm.c
231
place.lpfn = resource_size(&mem->io) >> PAGE_SHIFT;
drivers/gpu/drm/i915/intel_region_ttm.c
236
mock_bo.bdev = &mem->i915->bdev;
drivers/gpu/drm/i915/intel_region_ttm.c
255
void intel_region_ttm_resource_free(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.c
258
struct ttm_resource_manager *man = mem->region_private;
drivers/gpu/drm/i915/intel_region_ttm.c
262
mock_bo.bdev = &mem->i915->bdev;
drivers/gpu/drm/i915/intel_region_ttm.c
54
int intel_region_to_ttm_type(const struct intel_memory_region *mem)
drivers/gpu/drm/i915/intel_region_ttm.c
58
GEM_BUG_ON(mem->type != INTEL_MEMORY_LOCAL &&
drivers/gpu/drm/i915/intel_region_ttm.c
59
mem->type != INTEL_MEMORY_MOCK &&
drivers/gpu/drm/i915/intel_region_ttm.c
60
mem->type != INTEL_MEMORY_SYSTEM);
drivers/gpu/drm/i915/intel_region_ttm.c
62
if (mem->type == INTEL_MEMORY_SYSTEM)
drivers/gpu/drm/i915/intel_region_ttm.c
65
type = mem->instance + TTM_PL_PRIV;
drivers/gpu/drm/i915/intel_region_ttm.c
82
int intel_region_ttm_init(struct intel_memory_region *mem)
drivers/gpu/drm/i915/intel_region_ttm.c
84
struct ttm_device *bdev = &mem->i915->bdev;
drivers/gpu/drm/i915/intel_region_ttm.c
85
int mem_type = intel_region_to_ttm_type(mem);
drivers/gpu/drm/i915/intel_region_ttm.c
89
resource_size(&mem->region),
drivers/gpu/drm/i915/intel_region_ttm.c
90
resource_size(&mem->io),
drivers/gpu/drm/i915/intel_region_ttm.c
91
mem->min_page_size, PAGE_SIZE);
drivers/gpu/drm/i915/intel_region_ttm.c
95
mem->region_private = ttm_manager_type(bdev, mem_type);
drivers/gpu/drm/i915/intel_region_ttm.h
21
int intel_region_ttm_init(struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_region_ttm.h
23
int intel_region_ttm_fini(struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_region_ttm.h
26
intel_region_ttm_resource_to_rsgt(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.h
30
void intel_region_ttm_resource_free(struct intel_memory_region *mem,
drivers/gpu/drm/i915/intel_region_ttm.h
33
int intel_region_to_ttm_type(const struct intel_memory_region *mem);
drivers/gpu/drm/i915/intel_region_ttm.h
39
intel_region_ttm_resource_alloc(struct intel_memory_region *mem,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
101
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
107
igt_object_create(struct intel_memory_region *mem,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
115
obj = i915_gem_object_create_region(mem, size, 0, flags);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1351
struct intel_memory_region *mem;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1359
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1360
if (IS_ERR(mem)) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1362
err = PTR_ERR(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1366
err = i915_subtests(tests, mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1368
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
158
struct intel_memory_region *mem = arg;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
159
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
160
resource_size_t avail = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
174
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
175
if (IS_ERR(mem)) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
177
err = PTR_ERR(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
194
err = intel_memory_region_reserve(mem, start + offset, size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
211
obj = igt_object_create(mem, &objects, size, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
229
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
230
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
238
struct intel_memory_region *mem = arg;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
249
total = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
252
obj = igt_object_create(mem, &objects, PAGE_SIZE,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
266
obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
284
obj = igt_object_create(mem, &objects, target,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
32
static void close_objects(struct intel_memory_region *mem,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
320
obj = igt_object_create(mem, list, target,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
328
close_objects(mem, &holes);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
334
obj = igt_object_create(mem, &objects, target, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
35
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
350
obj = igt_object_create(mem, &objects, target,
drivers/gpu/drm/i915/selftests/intel_memory_region.c
364
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
370
struct intel_memory_region *mem = arg;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
371
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
387
mem = mock_region_create(i915, 0, size, PAGE_SIZE, 0, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
388
if (IS_ERR(mem))
drivers/gpu/drm/i915/selftests/intel_memory_region.c
389
return PTR_ERR(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
391
obj = igt_object_create(mem, &objects, size, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
414
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
416
obj = igt_object_create(mem, &objects, roundup_pow_of_two(size),
drivers/gpu/drm/i915/selftests/intel_memory_region.c
425
obj = igt_object_create(mem, &objects, size, I915_BO_ALLOC_CONTIGUOUS);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
434
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
436
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
446
struct intel_memory_region *mem = arg;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
447
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
475
mem = mock_region_create(i915, 0, size, ps, 0, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
476
if (IS_ERR(mem))
drivers/gpu/drm/i915/selftests/intel_memory_region.c
477
return PTR_ERR(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
479
obj = igt_object_create(mem, &objects, size, 0);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
519
close_objects(mem, &objects);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
521
intel_memory_region_destroy(mem);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
56
struct intel_memory_region *mem = arg;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
57
resource_size_t total = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
73
obj = i915_gem_object_create_region(mem, size, 0, 0);
drivers/gpu/drm/i915/selftests/mock_region.c
60
static int mock_object_init(struct intel_memory_region *mem,
drivers/gpu/drm/i915/selftests/mock_region.c
68
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/mock_region.c
70
if (size > resource_size(&mem->region))
drivers/gpu/drm/i915/selftests/mock_region.c
82
i915_gem_object_init_memory_region(obj, mem);
drivers/gpu/drm/i915/selftests/mock_region.c
87
static int mock_region_fini(struct intel_memory_region *mem)
drivers/gpu/drm/i915/selftests/mock_region.c
89
struct drm_i915_private *i915 = mem->i915;
drivers/gpu/drm/i915/selftests/mock_region.c
90
int instance = mem->instance;
drivers/gpu/drm/i915/selftests/mock_region.c
93
ret = intel_region_ttm_fini(mem);
drivers/gpu/drm/imagination/pvr_fw.c
196
struct pvr_fw_mem *fw_mem = &pvr_dev->fw_dev.mem;
drivers/gpu/drm/imagination/pvr_fw.c
294
&fw_dev->mem.fwif_connection_ctl_obj);
drivers/gpu/drm/imagination/pvr_fw.c
309
pvr_fw_object_unmap_and_destroy(fw_dev->mem.fwif_connection_ctl_obj);
drivers/gpu/drm/imagination/pvr_fw.c
318
struct pvr_fw_mem *fw_mem = &fw_dev->mem;
drivers/gpu/drm/imagination/pvr_fw.c
346
struct pvr_fw_mem *fw_mem = &pvr_dev->fw_dev.mem;
drivers/gpu/drm/imagination/pvr_fw.c
366
struct pvr_fw_mem *fw_mem = &fw_dev->mem;
drivers/gpu/drm/imagination/pvr_fw.c
459
struct pvr_fw_mem *fw_mem = &fw_dev->mem;
drivers/gpu/drm/imagination/pvr_fw.c
604
struct pvr_fw_mem *fw_mem = &fw_dev->mem;
drivers/gpu/drm/imagination/pvr_fw.c
634
struct pvr_fw_mem *fw_mem = &pvr_dev->fw_dev.mem;
drivers/gpu/drm/imagination/pvr_fw.c
812
struct pvr_fw_mem *fw_mem = &pvr_dev->fw_dev.mem;
drivers/gpu/drm/imagination/pvr_fw.c
843
struct pvr_fw_mem *fw_mem = &pvr_dev->fw_dev.mem;
drivers/gpu/drm/imagination/pvr_fw.h
296
struct pvr_fw_mem mem;
drivers/gpu/drm/imagination/pvr_fw_meta.c
388
u64 seg_out_addr = get_fw_obj_gpu_addr(pvr_dev->fw_dev.mem.data_obj);
drivers/gpu/drm/imagination/pvr_fw_meta.c
484
if (fw_dev->mem.core_code_obj) {
drivers/gpu/drm/imagination/pvr_fw_meta.c
487
pvr_fw_object_get_fw_addr(fw_dev->mem.core_code_obj, &core_code_fw_addr);
drivers/gpu/drm/imagination/pvr_fw_mips.c
58
WARN_ON(pvr_gem_get_dma_addr(fw_dev->mem.code_obj->gem, boot_code_entry->alloc_offset,
drivers/gpu/drm/imagination/pvr_fw_mips.c
60
WARN_ON(pvr_gem_get_dma_addr(fw_dev->mem.data_obj->gem, boot_data_entry->alloc_offset,
drivers/gpu/drm/imagination/pvr_fw_mips.c
62
WARN_ON(pvr_gem_get_dma_addr(fw_dev->mem.code_obj->gem,
drivers/gpu/drm/imagination/pvr_fw_mips.c
73
WARN_ON(pvr_fw_object_get_dma_addr(fw_dev->mem.data_obj, stack_entry->alloc_offset,
drivers/gpu/drm/imagination/pvr_fw_riscv.c
31
u64 code_addr = pvr_fw_obj_get_gpu_addr(pvr_dev->fw_dev.mem.code_obj);
drivers/gpu/drm/imagination/pvr_fw_riscv.c
32
u64 data_addr = pvr_fw_obj_get_gpu_addr(pvr_dev->fw_dev.mem.data_obj);
drivers/gpu/drm/imagination/pvr_fw_riscv.c
72
struct pvr_fw_mem *fw_mem = &fw_dev->mem;
drivers/gpu/drm/imagination/pvr_mmu.c
148
pvr_fw_object_get_fw_addr(pvr_dev->fw_dev.mem.mmucache_sync_obj,
drivers/gpu/drm/kmb/kmb_drv.c
100
void __iomem *mem;
drivers/gpu/drm/kmb/kmb_drv.c
107
mem = devm_ioremap_resource(drm->dev, res);
drivers/gpu/drm/kmb/kmb_drv.c
108
if (IS_ERR(mem))
drivers/gpu/drm/kmb/kmb_drv.c
110
return mem;
drivers/gpu/drm/loongson/lsdc_ttm.c
221
struct ttm_resource *mem)
drivers/gpu/drm/loongson/lsdc_ttm.c
225
switch (mem->mem_type) {
drivers/gpu/drm/loongson/lsdc_ttm.c
231
mem->bus.offset = (mem->start << PAGE_SHIFT) + ldev->vram_base;
drivers/gpu/drm/loongson/lsdc_ttm.c
232
mem->bus.is_iomem = true;
drivers/gpu/drm/loongson/lsdc_ttm.c
233
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/mediatek/mtk_hdmi_ddc.c
273
struct resource *mem;
drivers/gpu/drm/mediatek/mtk_hdmi_ddc.c
285
ddc->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/gpu/drm/mediatek/mtk_hdmi_ddc.c
311
dev_dbg(dev, "physical adr: %pa, end: %pa\n", &mem->start,
drivers/gpu/drm/mediatek/mtk_hdmi_ddc.c
312
&mem->end);
drivers/gpu/drm/mgag200/mgag200_drv.c
51
resource_size_t mgag200_probe_vram(void __iomem *mem, resource_size_t size)
drivers/gpu/drm/mgag200/mgag200_drv.c
60
orig = ioread16(mem);
drivers/gpu/drm/mgag200/mgag200_drv.c
61
iowrite16(0, mem);
drivers/gpu/drm/mgag200/mgag200_drv.c
66
orig1 = ioread8(mem + offset);
drivers/gpu/drm/mgag200/mgag200_drv.c
67
orig2 = ioread8(mem + offset + 0x100);
drivers/gpu/drm/mgag200/mgag200_drv.c
69
iowrite16(0xaa55, mem + offset);
drivers/gpu/drm/mgag200/mgag200_drv.c
70
iowrite16(0xaa55, mem + offset + 0x100);
drivers/gpu/drm/mgag200/mgag200_drv.c
72
test1 = ioread16(mem + offset);
drivers/gpu/drm/mgag200/mgag200_drv.c
73
test2 = ioread16(mem);
drivers/gpu/drm/mgag200/mgag200_drv.c
75
iowrite16(orig1, mem + offset);
drivers/gpu/drm/mgag200/mgag200_drv.c
76
iowrite16(orig2, mem + offset + 0x100);
drivers/gpu/drm/mgag200/mgag200_drv.c
85
iowrite16(orig, mem);
drivers/gpu/drm/mgag200/mgag200_drv.h
323
resource_size_t mgag200_probe_vram(void __iomem *mem, resource_size_t size);
drivers/gpu/drm/nouveau/dispnv50/crc.c
125
memset_io(ctx->mem.object.map.ptr, 0, ctx->mem.object.map.size);
drivers/gpu/drm/nouveau/dispnv50/crc.c
508
ret = nvif_mem_ctor_map(mmu, "kmsCrcNtfy", NVIF_MEM_VRAM, len, &ctx->mem);
drivers/gpu/drm/nouveau/dispnv50/crc.c
522
.start = ctx->mem.addr,
drivers/gpu/drm/nouveau/dispnv50/crc.c
523
.limit = ctx->mem.addr
drivers/gpu/drm/nouveau/dispnv50/crc.c
524
+ ctx->mem.size - 1,
drivers/gpu/drm/nouveau/dispnv50/crc.c
533
nvif_mem_dtor(&ctx->mem);
drivers/gpu/drm/nouveau/dispnv50/crc.c
541
nvif_mem_dtor(&ctx->mem);
drivers/gpu/drm/nouveau/dispnv50/crc.h
42
struct nvif_mem mem;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
101
struct crc907d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
92
struct crc907d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
70
struct crcc37d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
85
struct crcc37d_notifier __iomem *notifier = ctx->mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
25
const u32 crc_hi = upper_32_bits(ctx->mem.addr);
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
26
const u32 crc_lo = lower_32_bits(ctx->mem.addr);
drivers/gpu/drm/nouveau/dispnv50/disp.c
130
nvif_mem_dtor(&dmac->push.mem);
drivers/gpu/drm/nouveau/dispnv50/disp.c
138
dmac->cur = push->cur - (u32 __iomem *)dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
143
if (dmac->push.mem.type & NVIF_MEM_VRAM) {
drivers/gpu/drm/nouveau/dispnv50/disp.c
201
dmac->cur = push->cur - (u32 __iomem *)dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
207
push->cur = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
220
push->bgn = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
255
ret = nvif_mem_ctor_map(&drm->mmu, "kmsChanPush", type, 0x1000, &dmac->push.mem);
drivers/gpu/drm/nouveau/dispnv50/disp.c
261
dmac->push.bgn = dmac->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/disp.c
272
args->pushbuf = nvif_handle(&dmac->push.mem.object);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
314
head507d_olut_load(struct drm_color_lut *in, int size, void __iomem *mem)
drivers/gpu/drm/nouveau/dispnv50/head507d.c
316
for (; size--; in++, mem += 8) {
drivers/gpu/drm/nouveau/dispnv50/head507d.c
317
writew(drm_color_lut_extract(in-> red, 11) << 3, mem + 0);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
318
writew(drm_color_lut_extract(in->green, 11) << 3, mem + 2);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
319
writew(drm_color_lut_extract(in-> blue, 11) << 3, mem + 4);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
325
writew(readw(mem - 8), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
326
writew(readw(mem - 6), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/head507d.c
327
writew(readw(mem - 4), mem + 4);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
286
head907d_olut_load(struct drm_color_lut *in, int size, void __iomem *mem)
drivers/gpu/drm/nouveau/dispnv50/head907d.c
288
for (; size--; in++, mem += 8) {
drivers/gpu/drm/nouveau/dispnv50/head907d.c
289
writew(drm_color_lut_extract(in-> red, 14) + 0x6000, mem + 0);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
290
writew(drm_color_lut_extract(in->green, 14) + 0x6000, mem + 2);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
291
writew(drm_color_lut_extract(in-> blue, 14) + 0x6000, mem + 4);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
297
writew(readw(mem - 8), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
298
writew(readw(mem - 6), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/head907d.c
299
writew(readw(mem - 4), mem + 4);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
133
headc57d_olut_load_8(struct drm_color_lut *in, int size, void __iomem *mem)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
135
memset_io(mem, 0x00, 0x20); /* VSS header. */
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
136
mem += 0x20;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
150
for (i = 0; i < 4; i++, mem += 8) {
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
151
writew(r + ri * i, mem + 0);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
152
writew(g + gi * i, mem + 2);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
153
writew(b + bi * i, mem + 4);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
160
writew(readw(mem - 8), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
161
writew(readw(mem - 6), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
162
writew(readw(mem - 4), mem + 4);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
166
headc57d_olut_load(struct drm_color_lut *in, int size, void __iomem *mem)
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
168
memset_io(mem, 0x00, 0x20); /* VSS header. */
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
169
mem += 0x20;
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
171
for (; size--; in++, mem += 0x08) {
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
172
writew(drm_color_lut_extract(in-> red, 16), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
173
writew(drm_color_lut_extract(in->green, 16), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
174
writew(drm_color_lut_extract(in-> blue, 16), mem + 4);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
180
writew(readw(mem - 8), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
181
writew(readw(mem - 6), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/headc57d.c
182
writew(readw(mem - 4), mem + 4);
drivers/gpu/drm/nouveau/dispnv50/lut.c
36
void __iomem *mem = lut->mem[buffer].object.map.ptr;
drivers/gpu/drm/nouveau/dispnv50/lut.c
37
const u32 addr = lut->mem[buffer].addr;
drivers/gpu/drm/nouveau/dispnv50/lut.c
48
load(in, 1024, mem);
drivers/gpu/drm/nouveau/dispnv50/lut.c
52
load(in, drm_color_lut_size(blob), mem);
drivers/gpu/drm/nouveau/dispnv50/lut.c
62
for (i = 0; i < ARRAY_SIZE(lut->mem); i++)
drivers/gpu/drm/nouveau/dispnv50/lut.c
63
nvif_mem_dtor(&lut->mem[i]);
drivers/gpu/drm/nouveau/dispnv50/lut.c
72
for (i = 0; i < ARRAY_SIZE(lut->mem); i++) {
drivers/gpu/drm/nouveau/dispnv50/lut.c
74
size * 8, &lut->mem[i]);
drivers/gpu/drm/nouveau/dispnv50/lut.h
9
struct nvif_mem mem[2];
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
158
wndwc57e_ilut_load(struct drm_color_lut *in, int size, void __iomem *mem)
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
160
memset_io(mem, 0x00, 0x20); /* VSS header. */
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
161
mem += 0x20;
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
163
for (; size--; in++, mem += 0x08) {
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
167
writew(r, mem + 0);
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
168
writew(g, mem + 2);
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
169
writew(b, mem + 4);
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
175
writew(readw(mem - 8), mem + 0);
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
176
writew(readw(mem - 6), mem + 2);
drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c
177
writew(readw(mem - 4), mem + 4);
drivers/gpu/drm/nouveau/include/nvif/mmu.h
12
s32 mem;
drivers/gpu/drm/nouveau/include/nvif/push.h
33
struct nvif_mem mem;
drivers/gpu/drm/nouveau/include/nvif/push.h
75
u32 __o = _ppp->cur - (u32 *)_ppp->mem.object.map.ptr; \
drivers/gpu/drm/nouveau/include/nvif/push.h
76
NVIF_DEBUG(&_ppp->mem.object, "%08x: "f, __o * 4, ##a); \
drivers/gpu/drm/nouveau/include/nvif/vmm.h
53
void *argv, u32 argc, struct nvif_mem *mem, u64 offset);
drivers/gpu/drm/nouveau/include/nvkm/core/falcon.h
13
nvkm_falcon_mem(enum nvkm_falcon_mem mem)
drivers/gpu/drm/nouveau/include/nvkm/core/falcon.h
15
switch (mem) {
drivers/gpu/drm/nouveau/include/nvkm/core/firmware.h
29
} mem;
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
24
struct nvkm_memory *mem;
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
79
struct nvkm_memory *mem;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
160
struct nvkm_gsp_mem mem;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
94
struct nvkm_mm_node *mem;
drivers/gpu/drm/nouveau/nouveau_bo.c
1069
struct nouveau_mem *mem = new_reg ? nouveau_mem(new_reg) : NULL;
drivers/gpu/drm/nouveau/nouveau_bo.c
1080
if (mem && new_reg->mem_type != TTM_PL_SYSTEM &&
drivers/gpu/drm/nouveau/nouveau_bo.c
1081
mem->mem.page == nvbo->page) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1083
nouveau_vma_map(vma, mem);
drivers/gpu/drm/nouveau/nouveau_bo.c
1085
nouveau_uvmm_bo_map_all(nvbo, mem);
drivers/gpu/drm/nouveau/nouveau_bo.c
1236
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_bo.c
1238
if (drm->client.mem->oclass >= NVIF_CLASS_MEM_NV50) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1241
if (mem->kind)
drivers/gpu/drm/nouveau/nouveau_bo.c
1242
nvif_object_unmap_handle(&mem->mem.object);
drivers/gpu/drm/nouveau/nouveau_bo.c
1245
nvif_object_unmap_handle(&mem->mem.object);
drivers/gpu/drm/nouveau/nouveau_bo.c
1258
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_bo.c
1278
if (drm->client.mem->oclass < NVIF_CLASS_MEM_NV50 ||
drivers/gpu/drm/nouveau/nouveau_bo.c
1279
!mem->kind) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1297
if (drm->client.mem->oclass >= NVIF_CLASS_MEM_NV50) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1305
switch (mem->mem.object.oclass) {
drivers/gpu/drm/nouveau/nouveau_bo.c
1309
args.nv50.kind = mem->kind;
drivers/gpu/drm/nouveau/nouveau_bo.c
1310
args.nv50.comp = mem->comp;
drivers/gpu/drm/nouveau/nouveau_bo.c
1316
args.gf100.kind = mem->kind;
drivers/gpu/drm/nouveau/nouveau_bo.c
1324
ret = nvif_object_map_handle(&mem->mem.object,
drivers/gpu/drm/nouveau/nouveau_bo.c
803
u16 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem);
drivers/gpu/drm/nouveau/nouveau_bo.c
805
mem += index;
drivers/gpu/drm/nouveau/nouveau_bo.c
808
iowrite16_native(val, (void __force __iomem *)mem);
drivers/gpu/drm/nouveau/nouveau_bo.c
810
*mem = val;
drivers/gpu/drm/nouveau/nouveau_bo.c
817
u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem);
drivers/gpu/drm/nouveau/nouveau_bo.c
819
mem += index;
drivers/gpu/drm/nouveau/nouveau_bo.c
822
return ioread32_native((void __force __iomem *)mem);
drivers/gpu/drm/nouveau/nouveau_bo.c
824
return *mem;
drivers/gpu/drm/nouveau/nouveau_bo.c
831
u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem);
drivers/gpu/drm/nouveau/nouveau_bo.c
833
mem += index;
drivers/gpu/drm/nouveau/nouveau_bo.c
836
iowrite32_native(val, (void __force __iomem *)mem);
drivers/gpu/drm/nouveau/nouveau_bo.c
838
*mem = val;
drivers/gpu/drm/nouveau/nouveau_bo.c
912
ret = nvif_vmm_get(vmm, LAZY, false, old_mem->mem.page, 0,
drivers/gpu/drm/nouveau/nouveau_bo.c
913
old_mem->mem.size, &old_mem->vma[0]);
drivers/gpu/drm/nouveau/nouveau_bo.c
917
ret = nvif_vmm_get(vmm, LAZY, false, new_mem->mem.page, 0,
drivers/gpu/drm/nouveau/nouveau_bo.c
918
new_mem->mem.size, &old_mem->vma[1]);
drivers/gpu/drm/nouveau/nouveau_bo5039.c
42
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_bo5039.c
45
u64 src_offset = mem->vma[0].addr;
drivers/gpu/drm/nouveau/nouveau_bo5039.c
46
u64 dst_offset = mem->vma[1].addr;
drivers/gpu/drm/nouveau/nouveau_bo5039.c
47
int src_tiled = !!mem->kind;
drivers/gpu/drm/nouveau/nouveau_bo74c1.c
39
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_bo74c1.c
48
0x0308, upper_32_bits(mem->vma[0].addr),
drivers/gpu/drm/nouveau/nouveau_bo74c1.c
49
0x030c, lower_32_bits(mem->vma[0].addr),
drivers/gpu/drm/nouveau/nouveau_bo74c1.c
50
0x0310, upper_32_bits(mem->vma[1].addr),
drivers/gpu/drm/nouveau/nouveau_bo74c1.c
51
0x0314, lower_32_bits(mem->vma[1].addr),
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
43
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
45
u64 src_offset = mem->vma[0].addr;
drivers/gpu/drm/nouveau/nouveau_bo85b5.c
46
u64 dst_offset = mem->vma[1].addr;
drivers/gpu/drm/nouveau/nouveau_bo9039.c
42
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_bo9039.c
43
u64 src_offset = mem->vma[0].addr;
drivers/gpu/drm/nouveau/nouveau_bo9039.c
44
u64 dst_offset = mem->vma[1].addr;
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
36
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
38
u64 src_offset = mem->vma[0].addr;
drivers/gpu/drm/nouveau/nouveau_bo90b5.c
39
u64 dst_offset = mem->vma[1].addr;
drivers/gpu/drm/nouveau/nouveau_boa0b5.c
41
struct nouveau_mem *mem = nouveau_mem(old_reg);
drivers/gpu/drm/nouveau/nouveau_boa0b5.c
50
NVVAL(NVA0B5, OFFSET_IN_UPPER, UPPER, upper_32_bits(mem->vma[0].addr)),
drivers/gpu/drm/nouveau/nouveau_boa0b5.c
52
OFFSET_IN_LOWER, lower_32_bits(mem->vma[0].addr),
drivers/gpu/drm/nouveau/nouveau_boa0b5.c
55
NVVAL(NVA0B5, OFFSET_OUT_UPPER, UPPER, upper_32_bits(mem->vma[1].addr)),
drivers/gpu/drm/nouveau/nouveau_boa0b5.c
57
OFFSET_OUT_LOWER, lower_32_bits(mem->vma[1].addr),
drivers/gpu/drm/nouveau/nouveau_chan.c
133
chan->chan.push.bgn = chan->chan.push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nouveau_chan.c
171
chan->chan.push.mem.object.parent = cli->base.object.parent;
drivers/gpu/drm/nouveau/nouveau_chan.c
172
chan->chan.push.mem.object.client = &cli->base;
drivers/gpu/drm/nouveau/nouveau_chan.c
173
chan->chan.push.mem.object.name = "chanPush";
drivers/gpu/drm/nouveau/nouveau_chan.c
174
chan->chan.push.mem.object.map.ptr = chan->push.buffer->kmap.virtual;
drivers/gpu/drm/nouveau/nouveau_drm.c
304
cli->mem = &mems[ret];
drivers/gpu/drm/nouveau/nouveau_drv.h
107
const struct nvif_mclass *mem;
drivers/gpu/drm/nouveau/nouveau_mem.c
103
if (mem->kind && !(mmu->type[type].type & NVIF_MEM_KIND))
drivers/gpu/drm/nouveau/nouveau_mem.c
104
mem->comp = mem->kind = 0;
drivers/gpu/drm/nouveau/nouveau_mem.c
105
if (mem->comp && !(mmu->type[type].type & NVIF_MEM_COMP)) {
drivers/gpu/drm/nouveau/nouveau_mem.c
107
mem->kind = mmu->kind[mem->kind];
drivers/gpu/drm/nouveau/nouveau_mem.c
108
mem->comp = 0;
drivers/gpu/drm/nouveau/nouveau_mem.c
117
ret = nvif_mem_ctor_type(mmu, "ttmHostMem", mmu->mem, type, PAGE_SHIFT,
drivers/gpu/drm/nouveau/nouveau_mem.c
119
&args, sizeof(args), &mem->mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
127
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_mem.c
128
struct nouveau_drm *drm = mem->drm;
drivers/gpu/drm/nouveau/nouveau_mem.c
134
switch (mmu->mem) {
drivers/gpu/drm/nouveau/nouveau_mem.c
136
ret = nvif_mem_ctor_type(mmu, "ttmVram", mmu->mem,
drivers/gpu/drm/nouveau/nouveau_mem.c
141
&mem->mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
144
ret = nvif_mem_ctor_type(mmu, "ttmVram", mmu->mem,
drivers/gpu/drm/nouveau/nouveau_mem.c
147
.bankswz = mmu->kind[mem->kind] == 2,
drivers/gpu/drm/nouveau/nouveau_mem.c
150
&mem->mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
159
reg->start = mem->mem.addr >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_mem.c
166
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_mem.c
168
nouveau_mem_fini(mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
170
kfree(mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
177
struct nouveau_mem *mem;
drivers/gpu/drm/nouveau/nouveau_mem.c
179
if (!(mem = kzalloc_obj(*mem)))
drivers/gpu/drm/nouveau/nouveau_mem.c
182
mem->drm = drm;
drivers/gpu/drm/nouveau/nouveau_mem.c
183
mem->kind = kind;
drivers/gpu/drm/nouveau/nouveau_mem.c
184
mem->comp = comp;
drivers/gpu/drm/nouveau/nouveau_mem.c
186
*res = &mem->base;
drivers/gpu/drm/nouveau/nouveau_mem.c
37
nouveau_mem_map(struct nouveau_mem *mem,
drivers/gpu/drm/nouveau/nouveau_mem.c
53
args.nv50.kind = mem->kind;
drivers/gpu/drm/nouveau/nouveau_mem.c
54
args.nv50.comp = mem->comp;
drivers/gpu/drm/nouveau/nouveau_mem.c
61
if (mem->mem.type & NVIF_MEM_VRAM)
drivers/gpu/drm/nouveau/nouveau_mem.c
67
args.gf100.kind = mem->kind;
drivers/gpu/drm/nouveau/nouveau_mem.c
75
return nvif_vmm_map(vmm, vma->addr, mem->mem.size, &args, argc, &mem->mem, 0);
drivers/gpu/drm/nouveau/nouveau_mem.c
79
nouveau_mem_fini(struct nouveau_mem *mem)
drivers/gpu/drm/nouveau/nouveau_mem.c
81
nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[1]);
drivers/gpu/drm/nouveau/nouveau_mem.c
82
nvif_vmm_put(&mem->drm->client.vmm.vmm, &mem->vma[0]);
drivers/gpu/drm/nouveau/nouveau_mem.c
83
mutex_lock(&mem->drm->client_mutex);
drivers/gpu/drm/nouveau/nouveau_mem.c
84
nvif_mem_dtor(&mem->mem);
drivers/gpu/drm/nouveau/nouveau_mem.c
85
mutex_unlock(&mem->drm->client_mutex);
drivers/gpu/drm/nouveau/nouveau_mem.c
91
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_mem.c
92
struct nouveau_drm *drm = mem->drm;
drivers/gpu/drm/nouveau/nouveau_mem.c
99
type = drm->ttm.type_ncoh[!!mem->kind];
drivers/gpu/drm/nouveau/nouveau_mem.h
14
struct nvif_mem mem;
drivers/gpu/drm/nouveau/nouveau_mem.h
39
nouveau_mem_map_fixed(struct nouveau_mem *mem,
drivers/gpu/drm/nouveau/nouveau_sgdma.c
16
struct nouveau_mem *mem;
drivers/gpu/drm/nouveau/nouveau_sgdma.c
35
struct nouveau_mem *mem = nouveau_mem(reg);
drivers/gpu/drm/nouveau/nouveau_sgdma.c
38
if (nvbe->mem)
drivers/gpu/drm/nouveau/nouveau_sgdma.c
46
ret = nouveau_mem_map(mem, &drm->client.vmm.vmm, &mem->vma[0]);
drivers/gpu/drm/nouveau/nouveau_sgdma.c
48
nouveau_mem_fini(mem);
drivers/gpu/drm/nouveau/nouveau_sgdma.c
53
nvbe->mem = mem;
drivers/gpu/drm/nouveau/nouveau_sgdma.c
61
if (nvbe->mem) {
drivers/gpu/drm/nouveau/nouveau_sgdma.c
62
nouveau_mem_fini(nvbe->mem);
drivers/gpu/drm/nouveau/nouveau_sgdma.c
63
nvbe->mem = NULL;
drivers/gpu/drm/nouveau/nouveau_ttm.c
132
struct nouveau_mem *mem;
drivers/gpu/drm/nouveau/nouveau_ttm.c
139
mem = nouveau_mem(*res);
drivers/gpu/drm/nouveau/nouveau_ttm.c
142
(long)(*res)->size, &mem->vma[0]);
drivers/gpu/drm/nouveau/nouveau_ttm.c
148
(*res)->start = mem->vma[0].addr >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
139
struct nouveau_mem *mem)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
152
if (mem->mem.type & NVIF_MEM_VRAM)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
168
&mem->mem, bo_offset);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1827
nouveau_uvmm_bo_map_all(struct nouveau_bo *nvbo, struct nouveau_mem *mem)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1839
nouveau_uvma_map(uvma, mem);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
192
struct nouveau_mem *mem)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
201
mem);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
483
struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
494
if ((mem->mem.type & NVIF_MEM_VRAM) && !vmm->page[i].vram)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
496
if ((mem->mem.type & NVIF_MEM_HOST) &&
drivers/gpu/drm/nouveau/nouveau_uvmm.h
80
void nouveau_uvmm_bo_map_all(struct nouveau_bo *nvbov, struct nouveau_mem *mem);
drivers/gpu/drm/nouveau/nouveau_vmm.c
100
mem->mem.page == nvbo->page) {
drivers/gpu/drm/nouveau/nouveau_vmm.c
101
ret = nvif_vmm_get(&vmm->vmm, LAZY, false, mem->mem.page, 0,
drivers/gpu/drm/nouveau/nouveau_vmm.c
102
mem->mem.size, &tmp);
drivers/gpu/drm/nouveau/nouveau_vmm.c
107
ret = nouveau_vma_map(vma, mem);
drivers/gpu/drm/nouveau/nouveau_vmm.c
109
ret = nvif_vmm_get(&vmm->vmm, PTES, false, mem->mem.page, 0,
drivers/gpu/drm/nouveau/nouveau_vmm.c
110
mem->mem.size, &tmp);
drivers/gpu/drm/nouveau/nouveau_vmm.c
31
if (vma->mem) {
drivers/gpu/drm/nouveau/nouveau_vmm.c
33
vma->mem = NULL;
drivers/gpu/drm/nouveau/nouveau_vmm.c
38
nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem)
drivers/gpu/drm/nouveau/nouveau_vmm.c
41
int ret = nouveau_mem_map(mem, &vma->vmm->vmm, &tmp);
drivers/gpu/drm/nouveau/nouveau_vmm.c
44
vma->mem = mem;
drivers/gpu/drm/nouveau/nouveau_vmm.c
80
struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource);
drivers/gpu/drm/nouveau/nouveau_vmm.c
95
vma->mem = NULL;
drivers/gpu/drm/nouveau/nouveau_vmm.h
13
struct nouveau_mem *mem;
drivers/gpu/drm/nouveau/nvif/chan.c
107
chan->push.mem.object.map.ptr = push;
drivers/gpu/drm/nouveau/nvif/chan.c
11
u32 put = push->bgn - (u32 *)chan->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/chan.c
119
u32 cur = push->cur - (u32 *)push->mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/chan.c
155
push->bgn = (u32 *)push->mem.object.map.ptr + cur;
drivers/gpu/drm/nouveau/nvif/chan.c
38
const u32 *map = chan->push.mem.object.map.ptr;
drivers/gpu/drm/nouveau/nvif/mem.c
29
struct nvif_mem *mem)
drivers/gpu/drm/nouveau/nvif/mem.c
31
int ret = nvif_mem_ctor(mmu, name, mmu->mem, NVIF_MEM_MAPPABLE | type,
drivers/gpu/drm/nouveau/nvif/mem.c
32
0, size, NULL, 0, mem);
drivers/gpu/drm/nouveau/nvif/mem.c
34
ret = nvif_object_map(&mem->object, NULL, 0);
drivers/gpu/drm/nouveau/nvif/mem.c
36
nvif_mem_dtor(mem);
drivers/gpu/drm/nouveau/nvif/mem.c
42
nvif_mem_dtor(struct nvif_mem *mem)
drivers/gpu/drm/nouveau/nvif/mem.c
44
nvif_object_dtor(&mem->object);
drivers/gpu/drm/nouveau/nvif/mem.c
50
struct nvif_mem *mem)
drivers/gpu/drm/nouveau/nvif/mem.c
56
mem->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/mem.c
73
args, sizeof(*args) + argc, &mem->object);
drivers/gpu/drm/nouveau/nvif/mem.c
75
mem->type = mmu->type[type].type;
drivers/gpu/drm/nouveau/nvif/mem.c
76
mem->page = args->page;
drivers/gpu/drm/nouveau/nvif/mem.c
77
mem->addr = args->addr;
drivers/gpu/drm/nouveau/nvif/mem.c
78
mem->size = args->size;
drivers/gpu/drm/nouveau/nvif/mem.c
89
u8 page, u64 size, void *argv, u32 argc, struct nvif_mem *mem)
drivers/gpu/drm/nouveau/nvif/mem.c
93
mem->object.client = NULL;
drivers/gpu/drm/nouveau/nvif/mem.c
98
size, argv, argc, mem);
drivers/gpu/drm/nouveau/nvif/mmu.c
70
mmu->mem = mems[ret].oclass;
drivers/gpu/drm/nouveau/nvif/vmm.c
140
void *argv, u32 argc, struct nvif_mem *mem, u64 offset)
drivers/gpu/drm/nouveau/nvif/vmm.c
148
.memory = nvif_handle(&mem->object),
drivers/gpu/drm/nouveau/nvif/vmm.c
37
struct nvif_mem *mem, u64 offset)
drivers/gpu/drm/nouveau/nvif/vmm.c
53
args->memory = nvif_handle(&mem->object);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
113
#define nvkm_firmware_mem(p) container_of((p), struct nvkm_firmware, mem.memory)
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
121
case NVKM_FIRMWARE_IMG_DMA: return &fw->mem.sgl;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
122
case NVKM_FIRMWARE_IMG_SGT: return fw->mem.sgt.sgl;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
137
.memory = &fw->mem.memory,
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
197
struct nvkm_memory *memory = &fw->mem.memory;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
208
dma_free_noncoherent(fw->device->dev, sg_dma_len(&fw->mem.sgl),
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
213
dma_unmap_sgtable(fw->device->dev, &fw->mem.sgt, DMA_TO_DEVICE, 0);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
214
sg_free_table(&fw->mem.sgt);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
251
sg_init_one(&fw->mem.sgl, fw->img, len);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
252
sg_dma_address(&fw->mem.sgl) = fw->phys;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
253
sg_dma_len(&fw->mem.sgl) = len;
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
266
ret = sg_alloc_table(&fw->mem.sgt, pages, GFP_KERNEL);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
272
for_each_sgtable_sg(&fw->mem.sgt, sgl, i) {
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
285
ret = dma_map_sgtable(fw->device->dev, &fw->mem.sgt,
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
290
sg_free_table(&fw->mem.sgt);
drivers/gpu/drm/nouveau/nvkm/core/firmware.c
307
nvkm_memory_ctor(&nvkm_firmware_mem, &fw->mem.memory);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
311
&fifo->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
315
ret = nvkm_vmm_get(bar1, 12, nvkm_memory_size(fifo->userd.mem), &fifo->userd.bar1);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
319
ret = nvkm_memory_map(fifo->userd.mem, 0, bar1, fifo->userd.bar1, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c
342
nvkm_memory_unref(&fifo->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
282
nvkm_memory_unref(&chan->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
449
ret = nvkm_memory_kmap(userd, &chan->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c
457
chan->userd.mem = nvkm_memory_ref(fifo->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
120
nvkm_kmap(chan->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
121
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x040, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
122
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x044, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
123
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x048, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
124
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x04c, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
125
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x050, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
126
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x058, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
127
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x05c, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
128
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x060, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
129
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x088, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
130
nvkm_wo32(chan->userd.mem, chan->userd.base + 0x08c, 0x00000000);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
131
nvkm_done(chan->userd.mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
87
const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
84
const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
184
const u64 user = nvkm_memory_addr(chan->userd.mem) + chan->userd.base;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
43
const u64 userd = nvkm_memory_addr(chan->userd.mem) + chan->userd.base;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
260
if (unlikely(!runl->mem)) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
262
maxsize * 2 * runl->func->size, 0, false, &runl->mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
268
if (runl->offset + segment >= nvkm_memory_size(runl->mem)) {
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c
281
return runl->mem;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
322
nvkm_memory_unref(&runl->mem);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
88
struct nvkm_memory *mem;
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.c
661
nv40_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem)
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.c
666
.data = mem,
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c
256
nv50_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem)
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv50.c
261
.data = mem,
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
106
sg_dma_len(&fw->fw.mem.sgl),
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
138
ret = nvkm_vmm_get(fw->vmm, 12, nvkm_memory_size(&fw->fw.mem.memory), &fw->vma);
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
144
ret = nvkm_memory_map(&fw->fw.mem.memory, 0, fw->vmm, fw->vma, NULL, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
115
ret = nvkm_gpuobj_new(device, 0x20000, 0, false, NULL, &bar->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
119
ret = nvkm_gpuobj_new(device, bar->pgd_addr, 0, false, bar->mem,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
124
ret = nvkm_gpuobj_new(device, 0x4000, 0, false, bar->mem, &bar->pgd);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
147
ret = nvkm_vmm_join(bar->bar2_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
151
ret = nvkm_gpuobj_new(device, 24, 16, false, bar->mem, &bar->bar2);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
183
ret = nvkm_vmm_join(bar->bar1_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
187
ret = nvkm_gpuobj_new(device, 24, 16, false, bar->mem, &bar->bar1);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
207
if (bar->mem) {
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
209
nvkm_vmm_part(bar->bar1_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
212
nvkm_vmm_part(bar->bar2_vmm, bar->mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
216
nvkm_gpuobj_del(&bar->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
89
nvkm_wr32(device, 0x001704, 0x00000000 | bar->mem->addr >> 12);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
90
nvkm_wr32(device, 0x001704, 0x40000000 | bar->mem->addr >> 12);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.h
10
struct nvkm_gpuobj *mem;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
137
nvkm_memory_unref(&fault->buffer[i]->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c
90
&buffer->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp100.c
54
return nvkm_memory_bar2(buffer->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gp10b.c
32
return nvkm_memory_addr(buffer->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
36
struct nvkm_memory *mem = buffer->mem;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
42
nvkm_kmap(mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
45
const u32 instlo = nvkm_ro32(mem, base + 0x00);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
46
const u32 insthi = nvkm_ro32(mem, base + 0x04);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
47
const u32 addrlo = nvkm_ro32(mem, base + 0x08);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
48
const u32 addrhi = nvkm_ro32(mem, base + 0x0c);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
49
const u32 timelo = nvkm_ro32(mem, base + 0x10);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
50
const u32 timehi = nvkm_ro32(mem, base + 0x14);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
51
const u32 info0 = nvkm_ro32(mem, base + 0x18);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
52
const u32 info1 = nvkm_ro32(mem, base + 0x1c);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c
72
nvkm_done(mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h
17
struct nvkm_memory *mem;
drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c
54
*size = nvkm_memory_size(buffer->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ram.c
52
.mem = vram->mn,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv1a.c
30
u32 mem, mib;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv1a.c
47
pci_read_config_dword(bridge, 0x7c, &mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv1a.c
48
mib = ((mem >> 6) & 31) + 1;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv1a.c
50
pci_read_config_dword(bridge, 0x84, &mem);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramnv1a.c
51
mib = ((mem >> 4) & 127) + 1;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.h
18
struct nvkm_memory *mem[R515_GR_MAX_CTXBUFS];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.h
45
struct nvkm_memory *mem[R515_GR_MAX_CTXBUFS];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
185
nvkm_memory_addr(chan->userd.mem) + chan->userd.base,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
159
ret = r535_gr_promote_ctx(gr, false, grc->vmm, grc->mem, grc->vma, &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
47
nvkm_memory_unref(&grc->mem[i]);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1152
&gsp->shm.mem);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1156
gsp->shm.ptes.ptr = gsp->shm.mem.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1161
gsp->shm.ptes.ptr[i] = gsp->shm.mem.addr + (i << GSP_PAGE_SHIFT);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1189
args->messageQueueInitArguments.sharedMemPhysAddr = gsp->shm.mem.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1192
(u8 *)gsp->shm.cmdq.ptr - (u8 *)gsp->shm.mem.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1194
(u8 *)gsp->shm.msgq.ptr - (u8 *)gsp->shm.mem.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2115
nvkm_gsp_mem_dtor(&gsp->shm.mem);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2163
ret = nvkm_gsp_radix3_sg(gsp, &gsp->fw.mem.sgt, gsp->fw.len, &gsp->radix3);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
251
nvkm_gsp_mem_dtor(struct nvkm_gsp_mem *mem)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
253
if (mem->data) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
258
memset(mem->data, 0xFF, mem->size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
260
dma_free_coherent(mem->dev, mem->size, mem->data, mem->addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
261
put_device(mem->dev);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
263
memset(mem, 0, sizeof(*mem));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
280
nvkm_gsp_mem_ctor(struct nvkm_gsp *gsp, size_t size, struct nvkm_gsp_mem *mem)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
282
mem->data = dma_alloc_coherent(gsp->subdev.device->dev, size, &mem->addr, GFP_KERNEL);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
283
if (WARN_ON(!mem->data))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
286
mem->size = size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
287
mem->dev = get_device(gsp->subdev.device->dev);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
180
rm->mem = hw->mem;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
142
ret = r535_gr_promote_ctx(gr, false, gr->scrubber.vmm, gr->scrubber.ctxbuf.mem,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
85
nvkm_memory_unref(&gr->scrubber.ctxbuf.mem[i]);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
188
args->messageQueueInitArguments.sharedMemPhysAddr = gsp->shm.mem.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
191
(u8 *)gsp->shm.cmdq.ptr - (u8 *)gsp->shm.mem.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
193
(u8 *)gsp->shm.msgq.ptr - (u8 *)gsp->shm.mem.data;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
288
.mem = node->mn,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/g84.c
31
.mem = {{ -1, 0, NVIF_CLASS_MEM_NV50}, nv50_mem_new, nv50_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gf100.c
80
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gh100.c
14
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gk104.c
31
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gk20a.c
31
.mem = {{ -1, -1, NVIF_CLASS_MEM_GF100}, .umap = gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gm200.c
76
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gm200.c
86
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gm20b.c
33
.mem = {{ -1, -1, NVIF_CLASS_MEM_GF100}, .umap = gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gm20b.c
43
.mem = {{ -1, -1, NVIF_CLASS_MEM_GF100}, .umap = gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gp100.c
33
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gp10b.c
33
.mem = {{ -1, -1, NVIF_CLASS_MEM_GF100}, .umap = gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/gv100.c
33
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mcp77.c
31
.mem = {{ -1, 0, NVIF_CLASS_MEM_NV50}, nv50_mem_new, nv50_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
113
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
115
.memory = &mem->memory,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
117
.sgl = mem->sgl,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
135
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
136
if (mem->mem) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
137
*pmap = vmap(mem->mem, mem->pages, VM_MAP, PAGE_KERNEL);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
154
struct nvkm_mem *mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
166
if (!(mem = kzalloc_obj(*mem)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
168
mem->target = target;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
169
mem->mmu = mmu;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
170
*pmemory = &mem->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
174
nvkm_memory_ctor(&nvkm_mem_dma, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
175
mem->dma = args->v0.dma;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
177
nvkm_memory_ctor(&nvkm_mem_sgl, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
178
mem->sgl = args->v0.sgl;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
183
mem->pages = size >> PAGE_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
187
kfree(mem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
191
nvkm_memory_ctor(&nvkm_mem_dma, &mem->memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
194
if (!(mem->mem = kvmalloc_objs(*mem->mem, size)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
196
if (!(mem->dma = kvmalloc_objs(*mem->dma, size)))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
204
for (mem->pages = 0; size; size--, mem->pages++) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
209
mem->dma[mem->pages] = dma_map_page(mmu->subdev.device->dev,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
212
if (dma_mapping_error(dev, mem->dma[mem->pages])) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
217
mem->mem[mem->pages] = p;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
231
ret = mmu->func->mem.vram(mmu, type, page, size,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
35
struct page **mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
57
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
58
if (mem->pages == 1 && mem->mem)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
59
return mem->dma[0];
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
73
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
75
.memory = &mem->memory,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
77
.dma = mem->dma,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
85
struct nvkm_mem *mem = nvkm_mem(memory);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
86
if (mem->mem) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
87
while (mem->pages--) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
88
dma_unmap_page(mem->mmu->subdev.device->dev,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
89
mem->dma[mem->pages], PAGE_SIZE,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
91
__free_page(mem->mem[mem->pages]);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
93
kvfree(mem->dma);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
94
kvfree(mem->mem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/mem.c
96
return mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv04.c
33
.mem = {{ -1, -1, NVIF_CLASS_MEM_NV04}, nv04_mem_new, nv04_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv41.c
45
.mem = {{ -1, -1, NVIF_CLASS_MEM_NV04}, nv04_mem_new, nv04_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c
60
.mem = {{ -1, -1, NVIF_CLASS_MEM_NV04}, nv04_mem_new, nv04_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv50.c
68
.mem = {{ -1, 0, NVIF_CLASS_MEM_NV50}, nv50_mem_new, nv50_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h
31
} mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/tu102.c
48
.mem = {{ -1, 0, NVIF_CLASS_MEM_GF100}, gf100_mem_new, gf100_mem_map },
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c
109
int ret = mmu->func->mem.umap(mmu, umem->memory, argv, argc,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
37
if (mmu->func->mem.user.oclass) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c
39
oclass->base = mmu->func->mem.user;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1552
if (map->mem) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1553
for (; map->off; map->mem = map->mem->next) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1554
u64 size = (u64)map->mem->length << NVKM_RAM_MM_SHIFT;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1559
func = map->page->desc->func->mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1640
const bool mem = next->memory != NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1649
(next->memory != NULL) == mem &&
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1660
!mem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
351
((u64)MAP->mem->offset << NVKM_RAM_MM_SHIFT), \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
352
((u64)MAP->mem->length << NVKM_RAM_MM_SHIFT), \
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
353
(MAP->mem = MAP->mem->next))
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
75
nvkm_vmm_pte_func mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
100
.mem = gf100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
175
.mem = gh100_vmm_pd0_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
71
.mem = gh100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
90
.mem = gh100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
36
.mem = gf100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
39
.mem = gf100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
49
.mem = gf100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
190
.mem = gp100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
211
.mem = gp100_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
374
.mem = gp100_vmm_pd0_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
125
struct nvkm_memory *mem;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
135
mem = vmm->pd->pt[0]->memory;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
136
nvkm_kmap(mem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
137
nvkm_wo32(mem, 0x00000, 0x0002103d); /* PCI, RW, PT, !LN */
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
138
nvkm_wo32(mem, 0x00004, vmm->limit - 1);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c
139
nvkm_done(mem);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
100
.mem = nv50_vmm_pgt_mem,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
286
if (map->mem && map->mem->type != kindm[kind]) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
288
kindm[kind], map->mem->type);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
775
struct resource *mem;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
801
mem = platform_get_resource(dev, IORESOURCE_MEM, 0);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
802
if (!mem) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
807
omap_dmm->phys_base = mem->start;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
808
omap_dmm->base = ioremap(mem->start, SZ_2K);
drivers/gpu/drm/panthor/panthor_fw.c
1268
panthor_kernel_bo_destroy(section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
179
struct panthor_kernel_bo *mem;
drivers/gpu/drm/panthor/panthor_fw.c
447
bool was_mapped = !!section->mem->kmap;
drivers/gpu/drm/panthor/panthor_fw.c
454
ret = panthor_kernel_bo_vmap(section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
458
memcpy(section->mem->kmap, section->data.buf, section->data.size);
drivers/gpu/drm/panthor/panthor_fw.c
460
memset(section->mem->kmap + section->data.size, 0,
drivers/gpu/drm/panthor/panthor_fw.c
461
panthor_kernel_bo_size(section->mem) - section->data.size);
drivers/gpu/drm/panthor/panthor_fw.c
465
panthor_kernel_bo_vunmap(section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
491
struct panthor_kernel_bo *mem;
drivers/gpu/drm/panthor/panthor_fw.c
494
mem = panthor_kernel_bo_create(ptdev, ptdev->fw->vm, SZ_8K,
drivers/gpu/drm/panthor/panthor_fw.c
500
if (IS_ERR(mem))
drivers/gpu/drm/panthor/panthor_fw.c
501
return mem;
drivers/gpu/drm/panthor/panthor_fw.c
503
ret = panthor_kernel_bo_vmap(mem);
drivers/gpu/drm/panthor/panthor_fw.c
505
panthor_kernel_bo_destroy(mem);
drivers/gpu/drm/panthor/panthor_fw.c
509
memset(mem->kmap, 0, panthor_kernel_bo_size(mem));
drivers/gpu/drm/panthor/panthor_fw.c
510
*input = mem->kmap;
drivers/gpu/drm/panthor/panthor_fw.c
511
*output = mem->kmap + SZ_4K;
drivers/gpu/drm/panthor/panthor_fw.c
512
*input_fw_va = panthor_kernel_bo_gpuva(mem);
drivers/gpu/drm/panthor/panthor_fw.c
515
return mem;
drivers/gpu/drm/panthor/panthor_fw.c
649
section->mem = panthor_kernel_bo_create(ptdev, panthor_fw_vm(ptdev),
drivers/gpu/drm/panthor/panthor_fw.c
653
if (IS_ERR(section->mem))
drivers/gpu/drm/panthor/panthor_fw.c
654
return PTR_ERR(section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
656
if (drm_WARN_ON(&ptdev->base, section->mem->va_node.start != hdr.va.start))
drivers/gpu/drm/panthor/panthor_fw.c
660
ret = panthor_kernel_bo_vmap(section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
667
bo = to_panthor_bo(section->mem->obj);
drivers/gpu/drm/panthor/panthor_fw.c
732
sgt = drm_gem_shmem_get_pages_sgt(&to_panthor_bo(section->mem->obj)->base);
drivers/gpu/drm/panthor/panthor_fw.c
858
u64 shared_mem_start = panthor_kernel_bo_gpuva(ptdev->fw->shared_section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
860
panthor_kernel_bo_size(ptdev->fw->shared_section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
864
return ptdev->fw->shared_section->mem->kmap + (mcu_va - shared_mem_start);
drivers/gpu/drm/panthor/panthor_fw.c
873
u64 shared_section_sz = panthor_kernel_bo_size(ptdev->fw->shared_section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
885
cs_iface->control = ptdev->fw->shared_section->mem->kmap + iface_offset;
drivers/gpu/drm/panthor/panthor_fw.c
928
u64 shared_section_sz = panthor_kernel_bo_size(ptdev->fw->shared_section->mem);
drivers/gpu/drm/panthor/panthor_fw.c
936
csg_iface->control = ptdev->fw->shared_section->mem->kmap + iface_offset;
drivers/gpu/drm/panthor/panthor_fw.c
984
if (!ptdev->fw->shared_section->mem->kmap)
drivers/gpu/drm/panthor/panthor_fw.c
988
glb_iface->control = ptdev->fw->shared_section->mem->kmap;
drivers/gpu/drm/panthor/panthor_sched.c
3544
queue->iface.mem = panthor_fw_alloc_queue_iface_mem(group->ptdev,
drivers/gpu/drm/panthor/panthor_sched.c
3549
if (IS_ERR(queue->iface.mem)) {
drivers/gpu/drm/panthor/panthor_sched.c
3550
ret = PTR_ERR(queue->iface.mem);
drivers/gpu/drm/panthor/panthor_sched.c
3627
group->fdinfo.kbo_sizes += queue->iface.mem->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
399
struct panthor_kernel_bo *mem;
drivers/gpu/drm/panthor/panthor_sched.c
932
panthor_kernel_bo_destroy(queue->iface.mem);
drivers/gpu/drm/qxl/qxl_cmd.c
391
create->mem = qxl_bo_physical_address(qdev, bo, 0);
drivers/gpu/drm/qxl/qxl_cmd.c
393
DRM_DEBUG_DRIVER("mem = %llx, from %p\n", create->mem, bo->kptr);
drivers/gpu/drm/qxl/qxl_dev.h
321
QXLPHYSICAL mem;
drivers/gpu/drm/qxl/qxl_drv.h
327
struct ttm_resource *mem);
drivers/gpu/drm/qxl/qxl_ttm.c
73
struct ttm_resource *mem)
drivers/gpu/drm/qxl/qxl_ttm.c
77
switch (mem->mem_type) {
drivers/gpu/drm/qxl/qxl_ttm.c
82
mem->bus.is_iomem = true;
drivers/gpu/drm/qxl/qxl_ttm.c
83
mem->bus.offset = (mem->start << PAGE_SHIFT) + qdev->vram_base;
drivers/gpu/drm/qxl/qxl_ttm.c
84
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/qxl/qxl_ttm.c
87
mem->bus.is_iomem = true;
drivers/gpu/drm/qxl/qxl_ttm.c
88
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/qxl/qxl_ttm.c
90
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/radeon/radeon.h
2856
struct ttm_resource *mem);
drivers/gpu/drm/radeon/radeon_ttm.c
256
static int radeon_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource *mem)
drivers/gpu/drm/radeon/radeon_ttm.c
259
size_t bus_size = (size_t)mem->size;
drivers/gpu/drm/radeon/radeon_ttm.c
261
switch (mem->mem_type) {
drivers/gpu/drm/radeon/radeon_ttm.c
269
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/radeon/radeon_ttm.c
271
mem->bus.is_iomem = !rdev->agp->cant_use_aperture;
drivers/gpu/drm/radeon/radeon_ttm.c
272
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/radeon/radeon_ttm.c
277
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_ttm.c
279
if ((mem->bus.offset + bus_size) > rdev->mc.visible_vram_size)
drivers/gpu/drm/radeon/radeon_ttm.c
281
mem->bus.offset += rdev->mc.aper_base;
drivers/gpu/drm/radeon/radeon_ttm.c
282
mem->bus.is_iomem = true;
drivers/gpu/drm/radeon/radeon_ttm.c
283
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/radeon/radeon_ttm.c
289
mem->bus.addr = ioremap_wc(mem->bus.offset, bus_size);
drivers/gpu/drm/radeon/radeon_ttm.c
290
if (!mem->bus.addr)
drivers/gpu/drm/radeon/radeon_ttm.c
299
mem->bus.offset = (mem->bus.offset & 0x0ffffffffUL) +
drivers/gpu/drm/radeon/radeon_vm.c
910
struct ttm_resource *mem)
drivers/gpu/drm/radeon/radeon_vm.c
926
if (mem) {
drivers/gpu/drm/radeon/radeon_vm.c
944
if (mem) {
drivers/gpu/drm/radeon/radeon_vm.c
945
addr = (u64)mem->start << PAGE_SHIFT;
drivers/gpu/drm/radeon/radeon_vm.c
946
if (mem->mem_type != TTM_PL_SYSTEM)
drivers/gpu/drm/radeon/radeon_vm.c
949
if (mem->mem_type == TTM_PL_TT) {
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
246
cfg.mem[i] = sg_dma_address(state->sg_tables[i].sgl)
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
237
cfg->mem[i] = sg_dma_address(rjob->sg_tables[i].sgl)
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
147
cfg.mem[i] = gem->dma_addr + fb->offsets[i];
drivers/gpu/drm/sysfb/efidrm.c
153
struct resource *mem = NULL;
drivers/gpu/drm/sysfb/efidrm.c
227
mem = devm_request_mem_region(&pdev->dev, res->start, vsize, drv->name);
drivers/gpu/drm/sysfb/efidrm.c
228
if (!mem) {
drivers/gpu/drm/sysfb/efidrm.c
235
mem = res;
drivers/gpu/drm/sysfb/efidrm.c
241
void __iomem *screen_base = devm_ioremap_wc(&pdev->dev, mem->start,
drivers/gpu/drm/sysfb/efidrm.c
242
resource_size(mem));
drivers/gpu/drm/sysfb/efidrm.c
248
void __iomem *screen_base = devm_ioremap(&pdev->dev, mem->start,
drivers/gpu/drm/sysfb/efidrm.c
249
resource_size(mem));
drivers/gpu/drm/sysfb/efidrm.c
255
void *screen_base = devm_memremap(&pdev->dev, mem->start,
drivers/gpu/drm/sysfb/efidrm.c
256
resource_size(mem), MEMREMAP_WT);
drivers/gpu/drm/sysfb/efidrm.c
262
void *screen_base = devm_memremap(&pdev->dev, mem->start,
drivers/gpu/drm/sysfb/efidrm.c
263
resource_size(mem), MEMREMAP_WB);
drivers/gpu/drm/sysfb/ofdrm.c
430
void __iomem *mem;
drivers/gpu/drm/sysfb/ofdrm.c
448
mem = devm_ioremap(dev->dev, address + offset, size);
drivers/gpu/drm/sysfb/ofdrm.c
449
if (!mem)
drivers/gpu/drm/sysfb/ofdrm.c
452
return mem;
drivers/gpu/drm/sysfb/ofdrm.c
830
struct resource *res, *mem;
drivers/gpu/drm/sysfb/ofdrm.c
957
mem = devm_request_mem_region(&pdev->dev, fb_pgbase, fb_pgsize, drv->name);
drivers/gpu/drm/sysfb/ofdrm.c
958
if (!mem) {
drivers/gpu/drm/sysfb/ofdrm.c
963
screen_base = devm_ioremap(&pdev->dev, mem->start, resource_size(mem));
drivers/gpu/drm/sysfb/simpledrm.c
605
struct resource *res, *mem = NULL;
drivers/gpu/drm/sysfb/simpledrm.c
661
mem = simplefb_get_memory_of(dev, of_node);
drivers/gpu/drm/sysfb/simpledrm.c
662
if (IS_ERR(mem))
drivers/gpu/drm/sysfb/simpledrm.c
663
return ERR_CAST(mem);
drivers/gpu/drm/sysfb/simpledrm.c
692
if (mem) {
drivers/gpu/drm/sysfb/simpledrm.c
695
ret = devm_aperture_acquire_for_platform_device(pdev, mem->start,
drivers/gpu/drm/sysfb/simpledrm.c
696
resource_size(mem));
drivers/gpu/drm/sysfb/simpledrm.c
698
drm_err(dev, "could not acquire memory range %pr: %d\n", mem, ret);
drivers/gpu/drm/sysfb/simpledrm.c
702
drm_dbg(dev, "using system memory framebuffer at %pr\n", mem);
drivers/gpu/drm/sysfb/simpledrm.c
704
screen_base = devm_memremap(dev->dev, mem->start, resource_size(mem), MEMREMAP_WC);
drivers/gpu/drm/sysfb/simpledrm.c
725
mem = devm_request_mem_region(&pdev->dev, res->start, resource_size(res),
drivers/gpu/drm/sysfb/simpledrm.c
727
if (!mem) {
drivers/gpu/drm/sysfb/simpledrm.c
734
mem = res;
drivers/gpu/drm/sysfb/simpledrm.c
737
screen_base = devm_ioremap_wc(&pdev->dev, mem->start, resource_size(mem));
drivers/gpu/drm/sysfb/vesadrm.c
403
struct resource *mem = NULL;
drivers/gpu/drm/sysfb/vesadrm.c
497
mem = devm_request_mem_region(&pdev->dev, res->start, vsize, drv->name);
drivers/gpu/drm/sysfb/vesadrm.c
498
if (!mem) {
drivers/gpu/drm/sysfb/vesadrm.c
505
mem = res;
drivers/gpu/drm/sysfb/vesadrm.c
508
screen_base = devm_ioremap_wc(&pdev->dev, mem->start, resource_size(mem));
drivers/gpu/drm/tiny/bochs.c
233
unsigned long addr, size, mem, ioaddr, iosize;
drivers/gpu/drm/tiny/bochs.c
263
mem = bochs_dispi_read(bochs, VBE_DISPI_INDEX_VIDEO_MEMORY_64K)
drivers/gpu/drm/tiny/bochs.c
276
if (size != mem) {
drivers/gpu/drm/tiny/bochs.c
278
size, mem);
drivers/gpu/drm/tiny/bochs.c
279
size = min(size, mem);
drivers/gpu/drm/ttm/ttm_agp_backend.c
110
return (agp_be->mem != NULL);
drivers/gpu/drm/ttm/ttm_agp_backend.c
118
if (agp_be->mem)
drivers/gpu/drm/ttm/ttm_agp_backend.c
135
agp_be->mem = NULL;
drivers/gpu/drm/ttm/ttm_agp_backend.c
47
struct agp_memory *mem;
drivers/gpu/drm/ttm/ttm_agp_backend.c
55
struct agp_memory *mem;
drivers/gpu/drm/ttm/ttm_agp_backend.c
59
if (agp_be->mem)
drivers/gpu/drm/ttm/ttm_agp_backend.c
62
mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY);
drivers/gpu/drm/ttm/ttm_agp_backend.c
63
if (unlikely(mem == NULL))
drivers/gpu/drm/ttm/ttm_agp_backend.c
66
mem->page_count = 0;
drivers/gpu/drm/ttm/ttm_agp_backend.c
73
mem->pages[mem->page_count++] = page;
drivers/gpu/drm/ttm/ttm_agp_backend.c
75
agp_be->mem = mem;
drivers/gpu/drm/ttm/ttm_agp_backend.c
77
mem->is_flushed = 1;
drivers/gpu/drm/ttm/ttm_agp_backend.c
78
mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY;
drivers/gpu/drm/ttm/ttm_agp_backend.c
80
ret = agp_bind_memory(mem, bo_mem->start);
drivers/gpu/drm/ttm/ttm_agp_backend.c
92
if (agp_be->mem) {
drivers/gpu/drm/ttm/ttm_agp_backend.c
93
if (agp_be->mem->is_bound) {
drivers/gpu/drm/ttm/ttm_agp_backend.c
94
agp_unbind_memory(agp_be->mem);
drivers/gpu/drm/ttm/ttm_agp_backend.c
97
agp_free_memory(agp_be->mem);
drivers/gpu/drm/ttm/ttm_agp_backend.c
98
agp_be->mem = NULL;
drivers/gpu/drm/ttm/ttm_bo.c
121
struct ttm_resource *mem, bool evict,
drivers/gpu/drm/ttm/ttm_bo.c
130
new_use_tt = ttm_manager_type(bdev, mem->mem_type)->use_tt;
drivers/gpu/drm/ttm/ttm_bo.c
146
if (mem->mem_type != TTM_PL_SYSTEM) {
drivers/gpu/drm/ttm/ttm_bo.c
157
ret = bdev->funcs->move(bo, evict, ctx, mem, hop);
drivers/gpu/drm/ttm/ttm_bo_util.c
319
struct ttm_resource *mem = bo->resource;
drivers/gpu/drm/ttm/ttm_bo_util.c
321
if (mem->bus.addr) {
drivers/gpu/drm/ttm/ttm_bo_util.c
323
map->virtual = ((u8 *)mem->bus.addr) + offset;
drivers/gpu/drm/ttm/ttm_bo_util.c
325
resource_size_t res = mem->bus.offset + offset;
drivers/gpu/drm/ttm/ttm_bo_util.c
328
if (mem->bus.caching == ttm_write_combined)
drivers/gpu/drm/ttm/ttm_bo_util.c
331
else if (mem->bus.caching == ttm_cached)
drivers/gpu/drm/ttm/ttm_bo_util.c
345
struct ttm_resource *mem = bo->resource;
drivers/gpu/drm/ttm/ttm_bo_util.c
349
ttm_manager_type(bo->bdev, mem->mem_type);
drivers/gpu/drm/ttm/ttm_bo_util.c
374
prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
drivers/gpu/drm/ttm/ttm_bo_util.c
50
struct ttm_resource *mem)
drivers/gpu/drm/ttm/ttm_bo_util.c
500
struct ttm_resource *mem = bo->resource;
drivers/gpu/drm/ttm/ttm_bo_util.c
505
ret = ttm_mem_io_reserve(bo->bdev, mem);
drivers/gpu/drm/ttm/ttm_bo_util.c
509
if (mem->bus.is_iomem) {
drivers/gpu/drm/ttm/ttm_bo_util.c
512
if (mem->bus.addr)
drivers/gpu/drm/ttm/ttm_bo_util.c
513
vaddr_iomem = (void __iomem *)mem->bus.addr;
drivers/gpu/drm/ttm/ttm_bo_util.c
514
else if (mem->bus.caching == ttm_write_combined)
drivers/gpu/drm/ttm/ttm_bo_util.c
515
vaddr_iomem = ioremap_wc(mem->bus.offset,
drivers/gpu/drm/ttm/ttm_bo_util.c
518
else if (mem->bus.caching == ttm_cached)
drivers/gpu/drm/ttm/ttm_bo_util.c
519
vaddr_iomem = ioremap_cache(mem->bus.offset,
drivers/gpu/drm/ttm/ttm_bo_util.c
52
if (mem->bus.offset || mem->bus.addr)
drivers/gpu/drm/ttm/ttm_bo_util.c
523
vaddr_iomem = ioremap(mem->bus.offset, bo->base.size);
drivers/gpu/drm/ttm/ttm_bo_util.c
544
prot = ttm_io_prot(bo, mem, PAGE_KERNEL);
drivers/gpu/drm/ttm/ttm_bo_util.c
55
mem->bus.is_iomem = false;
drivers/gpu/drm/ttm/ttm_bo_util.c
566
struct ttm_resource *mem = bo->resource;
drivers/gpu/drm/ttm/ttm_bo_util.c
575
else if (!mem->bus.addr)
drivers/gpu/drm/ttm/ttm_bo_util.c
579
ttm_mem_io_free(bo->bdev, mem);
drivers/gpu/drm/ttm/ttm_bo_util.c
59
return bdev->funcs->io_mem_reserve(bdev, mem);
drivers/gpu/drm/ttm/ttm_bo_util.c
63
struct ttm_resource *mem)
drivers/gpu/drm/ttm/ttm_bo_util.c
65
if (!mem)
drivers/gpu/drm/ttm/ttm_bo_util.c
68
if (!mem->bus.offset && !mem->bus.addr)
drivers/gpu/drm/ttm/ttm_bo_util.c
72
bdev->funcs->io_mem_free(bdev, mem);
drivers/gpu/drm/ttm/ttm_bo_util.c
74
mem->bus.offset = 0;
drivers/gpu/drm/ttm/ttm_bo_util.c
75
mem->bus.addr = NULL;
drivers/gpu/drm/ttm/ttm_resource.c
833
struct ttm_resource *mem)
drivers/gpu/drm/ttm/ttm_resource.c
837
ret = ttm_mem_io_reserve(bdev, mem);
drivers/gpu/drm/ttm/ttm_resource.c
840
if (!mem->bus.is_iomem) {
drivers/gpu/drm/ttm/ttm_resource.c
845
if (mem->bus.addr) {
drivers/gpu/drm/ttm/ttm_resource.c
846
iosys_map_set_vaddr(&iter_io->dmap, mem->bus.addr);
drivers/gpu/drm/ttm/ttm_resource.c
851
if (mem->bus.caching == ttm_write_combined)
drivers/gpu/drm/ttm/ttm_resource.c
853
ioremap_wc(mem->bus.offset,
drivers/gpu/drm/ttm/ttm_resource.c
854
mem->size));
drivers/gpu/drm/ttm/ttm_resource.c
855
else if (mem->bus.caching == ttm_cached)
drivers/gpu/drm/ttm/ttm_resource.c
857
memremap(mem->bus.offset, mem->size,
drivers/gpu/drm/ttm/ttm_resource.c
865
ioremap(mem->bus.offset,
drivers/gpu/drm/ttm/ttm_resource.c
866
mem->size));
drivers/gpu/drm/ttm/ttm_resource.c
878
ttm_mem_io_free(bdev, mem);
drivers/gpu/drm/ttm/ttm_resource.c
895
struct ttm_resource *mem)
drivers/gpu/drm/ttm/ttm_resource.c
904
ttm_mem_io_free(bdev, mem);
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
716
struct ttm_resource *mem)
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
725
if (mem->mem_type == TTM_PL_VRAM || bo->resource->mem_type == TTM_PL_VRAM)
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
733
if (mem->mem_type != VMW_PL_MOB && bo->resource->mem_type == VMW_PL_MOB)
drivers/gpu/drm/vmwgfx/vmwgfx_bo.h
141
struct ttm_resource *mem);
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
438
static int vmw_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource *mem)
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
442
switch (mem->mem_type) {
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
449
mem->bus.offset = (mem->start << PAGE_SHIFT) +
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
451
mem->bus.is_iomem = true;
drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c
452
mem->bus.caching = ttm_cached;
drivers/gpu/drm/xe/display/xe_fb_pin.c
229
struct xe_ggtt *ggtt = tile0->mem.ggtt;
drivers/gpu/drm/xe/display/xe_fb_pin.c
26
struct xe_ggtt *ggtt = xe_device_get_root_tile(xe)->mem.ggtt;
drivers/gpu/drm/xe/display/xe_fb_pin.c
290
struct xe_vram_region *vram = xe_device_get_root_tile(xe)->mem.vram;
drivers/gpu/drm/xe/display/xe_fb_pin.c
59
struct xe_ggtt *ggtt = xe_device_get_root_tile(xe)->mem.ggtt;
drivers/gpu/drm/xe/display/xe_fb_pin.c
89
struct xe_ggtt *ggtt = tile0->mem.ggtt;
drivers/gpu/drm/xe/display/xe_initial_plane.c
60
u64 pte = xe_ggtt_read_pte(tile0->mem.ggtt, base);
drivers/gpu/drm/xe/display/xe_initial_plane.c
75
if (phys_base >= xe_vram_region_usable_size(tile0->mem.vram)) {
drivers/gpu/drm/xe/display/xe_panic.c
55
struct xe_vram_region *vram = xe_bo_device(bo)->mem.vram;
drivers/gpu/drm/xe/tests/xe_guc_buf_kunit.c
39
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/tests/xe_guc_buf_kunit.c
65
ggtt = xe_device_get_root_tile(test->priv)->mem.ggtt;
drivers/gpu/drm/xe/xe_assert.h
150
xe_vram_region_actual_physical_size(__tile->mem.vram), 1, \
drivers/gpu/drm/xe/xe_bb.c
46
bb->bo = xe_sa_bo_new(!usm ? tile->mem.kernel_bb_pool : gt->usm.bb_pool,
drivers/gpu/drm/xe/xe_bb.c
80
bb_pool = xe->sriov.vf.ccs.contexts[ctx_id].mem.ccs_bb_pool;
drivers/gpu/drm/xe/xe_bo.c
1718
xe_ggtt_remove_bo(tile->mem.ggtt, bo);
drivers/gpu/drm/xe/xe_bo.c
2368
err = xe_ggtt_insert_bo_at(t->mem.ggtt, bo,
drivers/gpu/drm/xe/xe_bo.c
2372
err = xe_ggtt_insert_bo(t->mem.ggtt, bo, exec);
drivers/gpu/drm/xe/xe_bo.c
238
return xe->tiles[tile_id].mem.kernel_vram->placement;
drivers/gpu/drm/xe/xe_bo.c
240
return xe->tiles[tile_id].mem.vram->placement;
drivers/gpu/drm/xe/xe_bo.c
3612
xe_ggtt_might_lock(tile->mem.ggtt);
drivers/gpu/drm/xe/xe_bo.c
444
xe_shrinker_mod_pages(xe->mem.shrinker, 0, tt->num_pages);
drivers/gpu/drm/xe/xe_bo.c
446
xe_shrinker_mod_pages(xe->mem.shrinker, tt->num_pages, 0);
drivers/gpu/drm/xe/xe_bo.c
454
xe_shrinker_mod_pages(xe->mem.shrinker, 0, -(long)tt->num_pages);
drivers/gpu/drm/xe/xe_bo.c
456
xe_shrinker_mod_pages(xe->mem.shrinker, -(long)tt->num_pages, 0);
drivers/gpu/drm/xe/xe_bo.c
604
static bool xe_ttm_resource_visible(struct ttm_resource *mem)
drivers/gpu/drm/xe/xe_bo.c
607
to_xe_ttm_vram_mgr_resource(mem);
drivers/gpu/drm/xe/xe_bo.c
609
return vres->used_visible_size == mem->size;
drivers/gpu/drm/xe/xe_bo.c
630
struct ttm_resource *mem)
drivers/gpu/drm/xe/xe_bo.c
634
switch (mem->mem_type) {
drivers/gpu/drm/xe/xe_bo.c
640
struct xe_vram_region *vram = res_to_mem_region(mem);
drivers/gpu/drm/xe/xe_bo.c
642
if (!xe_ttm_resource_visible(mem))
drivers/gpu/drm/xe/xe_bo.c
645
mem->bus.offset = mem->start << PAGE_SHIFT;
drivers/gpu/drm/xe/xe_bo.c
648
mem->placement & TTM_PL_FLAG_CONTIGUOUS)
drivers/gpu/drm/xe/xe_bo.c
649
mem->bus.addr = (u8 __force *)vram->mapping +
drivers/gpu/drm/xe/xe_bo.c
650
mem->bus.offset;
drivers/gpu/drm/xe/xe_bo.c
652
mem->bus.offset += vram->io_start;
drivers/gpu/drm/xe/xe_bo.c
653
mem->bus.is_iomem = true;
drivers/gpu/drm/xe/xe_bo.c
656
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/xe/xe_bo.c
660
return xe_ttm_stolen_io_mem_reserve(xe, mem);
drivers/gpu/drm/xe/xe_bo.h
31
XE_BO_FLAG_VRAM((tile)->mem.vram) : \
drivers/gpu/drm/xe/xe_bo_evict.c
208
xe_ggtt_map_bo_unlocked(tile->mem.ggtt, bo);
drivers/gpu/drm/xe/xe_device.c
687
xe->mem.vram = vram;
drivers/gpu/drm/xe/xe_device.c
878
err = xe_ggtt_init_early(tile->mem.ggtt);
drivers/gpu/drm/xe/xe_device_types.h
215
} mem;
drivers/gpu/drm/xe/xe_device_types.h
443
} mem;
drivers/gpu/drm/xe/xe_gt.c
542
err = xe_ggtt_init(gt_to_tile(gt)->mem.ggtt);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1822
return xe_ttm_vram_get_avail(&tile->mem.vram->ttm.manager);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
3169
struct xe_ggtt *ggtt = gt_to_tile(gt)->mem.ggtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
394
u64 ggtt_start = xe_ggtt_start(gt_to_tile(gt)->mem.ggtt);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
395
u64 ggtt_size = xe_ggtt_size(gt_to_tile(gt)->mem.ggtt);
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
510
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
715
struct xe_ggtt *ggtt = gt_to_tile(gt)->mem.ggtt;
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
491
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/xe_migrate.c
1151
bb_pool = ctx->mem.ccs_bb_pool;
drivers/gpu/drm/xe/xe_migrate.c
1241
bb_pool = ctx->mem.ccs_bb_pool;
drivers/gpu/drm/xe/xe_migrate.c
135
(xe->mem.vram), SZ_1G);
drivers/gpu/drm/xe/xe_migrate.c
137
addr -= xe_vram_region_dpa_base(xe->mem.vram);
drivers/gpu/drm/xe/xe_migrate.c
144
struct xe_vram_region *vram = xe->mem.vram;
drivers/gpu/drm/xe/xe_migrate.c
197
struct xe_bo *bo, *batch = tile->mem.kernel_bb_pool->bo;
drivers/gpu/drm/xe/xe_migrate.c
317
resource_size_t actual_phy_size = xe_vram_region_actual_physical_size(xe->mem.vram);
drivers/gpu/drm/xe/xe_page_reclaim.c
71
prl_sa = __xe_sa_bo_new(tile->mem.reclaim_pool,
drivers/gpu/drm/xe/xe_pagefault.c
106
err = xe_pagefault_begin(&exec, vma, tile->mem.vram,
drivers/gpu/drm/xe/xe_query.c
335
if (xe->mem.vram)
drivers/gpu/drm/xe/xe_query.c
411
BIT(gt_to_tile(gt)->mem.vram->id) << 1;
drivers/gpu/drm/xe/xe_shrinker.c
303
xe->mem.shrinker = shrinker;
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
173
ctx->mem.ccs_bb_pool = sa_manager;
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
180
u64 addr = xe_sa_manager_gpu_addr(ctx->mem.ccs_bb_pool);
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
391
u64 addr = xe_sa_manager_gpu_addr(ctx->mem.ccs_bb_pool);
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
482
bb_pool = xe->sriov.vf.ccs.contexts[ctx_id].mem.ccs_bb_pool;
drivers/gpu/drm/xe/xe_sriov_vf_ccs_types.h
37
} mem;
drivers/gpu/drm/xe/xe_svm.c
1912
tile->mem.vram->dpagemap_cache = cache;
drivers/gpu/drm/xe/xe_tile.c
125
tile->mem.vram = vram;
drivers/gpu/drm/xe/xe_tile.c
132
if (!tile->mem.kernel_vram)
drivers/gpu/drm/xe/xe_tile.c
133
tile->mem.kernel_vram = tile->mem.vram;
drivers/gpu/drm/xe/xe_tile.c
191
if (IS_DGFX(xe) && !ttm_resource_manager_used(&tile->mem.vram->ttm.manager)) {
drivers/gpu/drm/xe/xe_tile.c
192
err = xe_ttm_vram_mgr_init(xe, tile->mem.vram);
drivers/gpu/drm/xe/xe_tile.c
196
xe->info.mem_region_mask |= BIT(tile->mem.vram->id) << 1;
drivers/gpu/drm/xe/xe_tile.c
210
tile->mem.kernel_bb_pool = xe_sa_bo_manager_init(tile, SZ_1M, 16);
drivers/gpu/drm/xe/xe_tile.c
211
if (IS_ERR(tile->mem.kernel_bb_pool))
drivers/gpu/drm/xe/xe_tile.c
212
return PTR_ERR(tile->mem.kernel_bb_pool);
drivers/gpu/drm/xe/xe_tile.c
215
tile->mem.reclaim_pool = xe_sa_bo_manager_init(tile, SZ_1M, XE_PAGE_RECLAIM_LIST_MAX_SIZE);
drivers/gpu/drm/xe/xe_tile.c
216
if (IS_ERR(tile->mem.reclaim_pool))
drivers/gpu/drm/xe/xe_tile.c
217
return PTR_ERR(tile->mem.reclaim_pool);
drivers/gpu/drm/xe/xe_tile.c
94
tile->mem.ggtt = xe_ggtt_alloc(tile);
drivers/gpu/drm/xe/xe_tile.c
95
if (!tile->mem.ggtt)
drivers/gpu/drm/xe/xe_tile.h
36
return tile->mem.vram;
drivers/gpu/drm/xe/xe_tile_debugfs.c
111
if (tile->mem.vram)
drivers/gpu/drm/xe/xe_tile_debugfs.c
112
ttm_resource_manager_create_debugfs(&tile->mem.vram->ttm.manager, tile->debugfs,
drivers/gpu/drm/xe/xe_tile_debugfs.c
92
return xe_ggtt_dump(tile->mem.ggtt, p);
drivers/gpu/drm/xe/xe_tile_debugfs.c
97
drm_suballoc_dump_debug_info(&tile->mem.kernel_bb_pool->base, p,
drivers/gpu/drm/xe/xe_tile_debugfs.c
98
xe_sa_manager_gpu_addr(tile->mem.kernel_bb_pool));
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
102
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
126
mutex_lock(&tile->mem.ggtt->lock);
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
128
mutex_unlock(&tile->mem.ggtt->lock);
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
19
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
244
struct xe_ggtt *ggtt = tile->mem.ggtt;
drivers/gpu/drm/xe/xe_tile_sriov_vf.c
53
lockdep_assert_held(&tile->mem.ggtt->lock);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
281
struct ttm_resource *mem)
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
288
xe_res_first(mem, 0, 4096, &cur);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
289
mem->bus.offset = cur.start;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
291
drm_WARN_ON(&xe->drm, !(mem->placement & TTM_PL_FLAG_CONTIGUOUS));
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
293
if (mem->placement & TTM_PL_FLAG_CONTIGUOUS && mgr->mapping)
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
294
mem->bus.addr = (u8 __force *)mgr->mapping + mem->bus.offset;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
296
mem->bus.offset += mgr->io_base;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
297
mem->bus.is_iomem = true;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
298
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
305
struct ttm_resource *mem)
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
308
struct xe_bo *bo = ttm_to_xe_bo(mem->bo);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
317
mem->bus.offset = xe_bo_ggtt_addr(bo) + mgr->io_base;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
319
mem->bus.is_iomem = true;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
320
mem->bus.caching = ttm_write_combined;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
330
int xe_ttm_stolen_io_mem_reserve(struct xe_device *xe, struct ttm_resource *mem)
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
339
return __xe_ttm_stolen_io_mem_reserve_stolen(xe, mgr, mem);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
341
return __xe_ttm_stolen_io_mem_reserve_bar2(xe, mgr, mem);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
85
struct xe_vram_region *tile_vram = xe_device_get_root_tile(xe)->mem.vram;
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.c
93
tile_offset = tile_io_start - xe_vram_region_io_start(xe->mem.vram);
drivers/gpu/drm/xe/xe_ttm_stolen_mgr.h
16
int xe_ttm_stolen_io_mem_reserve(struct xe_device *xe, struct ttm_resource *mem);
drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
105
struct ttm_resource_manager *man = &xe->mem.sys_mgr;
drivers/gpu/drm/xe/xe_ttm_sys_mgr.c
90
struct ttm_resource_manager *man = &xe->mem.sys_mgr;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
402
phys_addr_t phys = cursor.start + xe_vram_region_io_start(tile->mem.vram);
drivers/gpu/drm/xe/xe_vram.c
182
xe->mem.vram->mapping = NULL;
drivers/gpu/drm/xe/xe_vram.c
185
tile->mem.vram->mapping = NULL;
drivers/gpu/drm/xe/xe_vram.c
186
if (tile->mem.kernel_vram)
drivers/gpu/drm/xe/xe_vram.c
187
tile->mem.kernel_vram->mapping = NULL;
drivers/gpu/drm/xe/xe_vram.c
292
err = vram_region_init(xe, tile->mem.vram, &lmem_bar, tile_offset, usable_size,
drivers/gpu/drm/xe/xe_vram.c
302
remain_io_size -= min_t(u64, tile->mem.vram->actual_physical_size, remain_io_size);
drivers/gpu/drm/xe/xe_vram.c
305
err = vram_region_init(xe, xe->mem.vram, &lmem_bar, 0, available_size, total_size,
drivers/gpu/host1x/hw/channel_hw.c
25
void *mem = NULL;
drivers/gpu/host1x/hw/channel_hw.c
28
mem = host1x_bo_mmap(bo);
drivers/gpu/host1x/hw/channel_hw.c
30
if (mem) {
drivers/gpu/host1x/hw/channel_hw.c
43
mem);
drivers/gpu/host1x/hw/channel_hw.c
46
host1x_bo_munmap(bo, mem);
drivers/gpu/host1x/job.c
34
void *mem;
drivers/gpu/host1x/job.c
51
mem = job = kzalloc(total, GFP_KERNEL);
drivers/gpu/host1x/job.c
61
mem += sizeof(struct host1x_job);
drivers/gpu/host1x/job.c
62
job->relocs = num_relocs ? mem : NULL;
drivers/gpu/host1x/job.c
63
mem += num_relocs * sizeof(struct host1x_reloc);
drivers/gpu/host1x/job.c
64
job->unpins = num_unpins ? mem : NULL;
drivers/gpu/host1x/job.c
65
mem += num_unpins * sizeof(struct host1x_job_unpin_data);
drivers/gpu/host1x/job.c
66
job->cmds = num_cmdbufs ? mem : NULL;
drivers/gpu/host1x/job.c
67
mem += num_cmdbufs * sizeof(struct host1x_job_cmd);
drivers/gpu/host1x/job.c
68
job->addr_phys = num_unpins ? mem : NULL;
drivers/hsi/controllers/omap_ssi_core.c
322
struct resource *mem;
drivers/hsi/controllers/omap_ssi_core.c
326
mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name);
drivers/hsi/controllers/omap_ssi_core.c
327
base = devm_ioremap_resource(&ssi->device, mem);
drivers/hsi/controllers/omap_ssi_core.c
334
*phy = mem->start;
drivers/hsi/controllers/omap_ssi_port.c
1079
struct resource *mem;
drivers/hsi/controllers/omap_ssi_port.c
1083
mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name);
drivers/hsi/controllers/omap_ssi_port.c
1084
if (!mem) {
drivers/hsi/controllers/omap_ssi_port.c
1088
ioarea = devm_request_mem_region(&port->device, mem->start,
drivers/hsi/controllers/omap_ssi_port.c
1089
resource_size(mem), dev_name(&pd->dev));
drivers/hsi/controllers/omap_ssi_port.c
1092
mem->name);
drivers/hsi/controllers/omap_ssi_port.c
1095
base = devm_ioremap(&port->device, mem->start, resource_size(mem));
drivers/hsi/controllers/omap_ssi_port.c
1097
dev_err(&pd->dev, "%s IO remap failed\n", mem->name);
drivers/hsi/controllers/omap_ssi_port.c
1103
*phy = mem->start;
drivers/hv/hv_balloon.c
639
struct memory_notify *mem = (struct memory_notify *)v;
drivers/hv/hv_balloon.c
650
pfn_count = hv_page_offline_check(mem->start_pfn,
drivers/hv/hv_balloon.c
651
mem->nr_pages);
drivers/hv/hv_common.c
479
void *mem;
drivers/hv/hv_common.c
492
mem = kmalloc_array(pgcount, HV_HYP_PAGE_SIZE, flags);
drivers/hv/hv_common.c
493
if (!mem)
drivers/hv/hv_common.c
498
*outputarg = (char *)mem + HV_HYP_PAGE_SIZE;
drivers/hv/hv_common.c
503
ret = set_memory_decrypted((unsigned long)mem, pgcount);
drivers/hv/hv_common.c
509
memset(mem, 0x00, pgcount * HV_HYP_PAGE_SIZE);
drivers/hv/hv_common.c
526
*inputarg = mem;
drivers/hv/mshv_root_main.c
1180
struct mshv_user_mem_region *mem,
drivers/hv/mshv_root_main.c
1185
u64 nr_pages = HVPFN_DOWN(mem->size);
drivers/hv/mshv_root_main.c
1190
if (mem->guest_pfn + nr_pages <= rg->start_gfn ||
drivers/hv/mshv_root_main.c
1191
rg->start_gfn + rg->nr_pages <= mem->guest_pfn)
drivers/hv/mshv_root_main.c
1198
rg = mshv_region_create(mem->guest_pfn, nr_pages,
drivers/hv/mshv_root_main.c
1199
mem->userspace_addr, mem->flags);
drivers/hv/mshv_root_main.c
1300
struct mshv_user_mem_region *mem)
drivers/hv/mshv_root_main.c
1308
if (mem->flags & BIT(MSHV_SET_MEM_BIT_UNMAP) ||
drivers/hv/mshv_root_main.c
1309
!access_ok((const void __user *)mem->userspace_addr, mem->size))
drivers/hv/mshv_root_main.c
1313
vma = vma_lookup(current->mm, mem->userspace_addr);
drivers/hv/mshv_root_main.c
1321
ret = mshv_partition_create_region(partition, mem, &region,
drivers/hv/mshv_root_main.c
1366
struct mshv_user_mem_region *mem)
drivers/hv/mshv_root_main.c
1370
if (!(mem->flags & BIT(MSHV_SET_MEM_BIT_UNMAP)))
drivers/hv/mshv_root_main.c
1375
region = mshv_partition_region_by_gfn(partition, mem->guest_pfn);
drivers/hv/mshv_root_main.c
1382
if (region->start_uaddr != mem->userspace_addr ||
drivers/hv/mshv_root_main.c
1383
region->start_gfn != mem->guest_pfn ||
drivers/hv/mshv_root_main.c
1384
region->nr_pages != HVPFN_DOWN(mem->size)) {
drivers/hv/mshv_root_main.c
1402
struct mshv_user_mem_region mem;
drivers/hv/mshv_root_main.c
1404
if (copy_from_user(&mem, user_mem, sizeof(mem)))
drivers/hv/mshv_root_main.c
1407
if (!mem.size ||
drivers/hv/mshv_root_main.c
1408
!PAGE_ALIGNED(mem.size) ||
drivers/hv/mshv_root_main.c
1409
!PAGE_ALIGNED(mem.userspace_addr) ||
drivers/hv/mshv_root_main.c
1410
(mem.flags & ~MSHV_SET_MEM_FLAGS_MASK) ||
drivers/hv/mshv_root_main.c
1411
mshv_field_nonzero(mem, rsvd))
drivers/hv/mshv_root_main.c
1414
if (mem.flags & BIT(MSHV_SET_MEM_BIT_UNMAP))
drivers/hv/mshv_root_main.c
1415
return mshv_unmap_user_memory(partition, &mem);
drivers/hv/mshv_root_main.c
1417
return mshv_map_user_memory(partition, &mem);
drivers/hwtracing/coresight/coresight-tpda.c
305
switch (tpda_attr->mem) {
drivers/hwtracing/coresight/coresight-tpda.c
335
switch (tpda_attr->mem) {
drivers/hwtracing/coresight/coresight-tpda.h
102
#define tpda_trig_sysfs_rw(name, mem) \
drivers/hwtracing/coresight/coresight-tpda.h
107
mem, \
drivers/hwtracing/coresight/coresight-tpda.h
99
enum tpda_cr_mem mem;
drivers/hwtracing/coresight/coresight-tpdm.c
130
switch (tpdm_attr->mem) {
drivers/hwtracing/coresight/coresight-tpdm.c
48
switch (tpdm_attr->mem) {
drivers/hwtracing/coresight/coresight-tpdm.c
837
if (tpdm_attr->mem == DSB_PATT)
drivers/hwtracing/coresight/coresight-tpdm.c
840
else if (tpdm_attr->mem == CMB_PATT)
drivers/hwtracing/coresight/coresight-tpdm.c
864
if (tpdm_attr->mem == DSB_PATT)
drivers/hwtracing/coresight/coresight-tpdm.c
866
else if (tpdm_attr->mem == CMB_PATT)
drivers/hwtracing/coresight/coresight-tpdm.h
143
#define tpdm_simple_dataset_ro(name, mem, idx) \
drivers/hwtracing/coresight/coresight-tpdm.h
147
mem, \
drivers/hwtracing/coresight/coresight-tpdm.h
152
#define tpdm_simple_dataset_rw(name, mem, idx) \
drivers/hwtracing/coresight/coresight-tpdm.h
157
mem, \
drivers/hwtracing/coresight/coresight-tpdm.h
162
#define tpdm_patt_enable_ts(name, mem) \
drivers/hwtracing/coresight/coresight-tpdm.h
167
mem, \
drivers/hwtracing/coresight/coresight-tpdm.h
343
enum dataset_mem mem;
drivers/i2c/busses/i2c-at91-core.c
199
struct resource *mem;
drivers/i2c/busses/i2c-at91-core.c
209
dev->base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/i2c/busses/i2c-at91-core.c
212
phy_addr = mem->start;
drivers/iio/adc/fsl-imx25-gcq.c
300
void __iomem *mem;
drivers/iio/adc/fsl-imx25-gcq.c
310
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/iio/adc/fsl-imx25-gcq.c
311
if (IS_ERR(mem))
drivers/iio/adc/fsl-imx25-gcq.c
312
return PTR_ERR(mem);
drivers/iio/adc/fsl-imx25-gcq.c
314
priv->regs = devm_regmap_init_mmio(dev, mem, &mx25_gcq_regconfig);
drivers/iio/adc/men_z188_adc.c
105
struct resource *mem;
drivers/iio/adc/men_z188_adc.c
119
mem = mcb_request_mem(dev, "z188-adc");
drivers/iio/adc/men_z188_adc.c
120
if (IS_ERR(mem))
drivers/iio/adc/men_z188_adc.c
121
return PTR_ERR(mem);
drivers/iio/adc/men_z188_adc.c
123
adc->base = ioremap(mem->start, resource_size(mem));
drivers/iio/adc/men_z188_adc.c
129
adc->mem = mem;
drivers/iio/adc/men_z188_adc.c
141
mcb_release_mem(mem);
drivers/iio/adc/men_z188_adc.c
152
mcb_release_mem(adc->mem);
drivers/iio/adc/men_z188_adc.c
25
struct resource *mem;
drivers/iio/adc/nxp-sar-adc.c
897
struct resource *mem;
drivers/iio/adc/nxp-sar-adc.c
907
info->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/iio/adc/nxp-sar-adc.c
912
info->regs_phys = mem->start;
drivers/infiniband/core/rw.c
417
void *mem;
drivers/infiniband/core/rw.c
423
mem = kzalloc(wrs_size, GFP_KERNEL);
drivers/infiniband/core/rw.c
424
if (!mem)
drivers/infiniband/core/rw.c
427
ctx->map.sges = sge = mem;
drivers/infiniband/core/rw.c
428
ctx->map.wrs = mem + wrs_offset;
drivers/infiniband/hw/erdma/erdma_qp.c
567
mr->mem.mtt_nents);
drivers/infiniband/hw/erdma/erdma_qp.c
569
if (mr->mem.mtt_nents <= ERDMA_MAX_INLINE_MTT_ENTRIES) {
drivers/infiniband/hw/erdma/erdma_qp.c
574
mr->mem.mtt->buf, MTT_SIZE(mr->mem.mtt_nents));
drivers/infiniband/hw/erdma/erdma_qp.c
576
MTT_SIZE(mr->mem.mtt_nents);
drivers/infiniband/hw/erdma/erdma_verbs.c
1179
mr->mem.page_size = PAGE_SIZE; /* update it later. */
drivers/infiniband/hw/erdma/erdma_verbs.c
1180
mr->mem.page_cnt = max_num_sg;
drivers/infiniband/hw/erdma/erdma_verbs.c
1181
mr->mem.mtt = erdma_create_mtt(dev, MTT_SIZE(max_num_sg), true);
drivers/infiniband/hw/erdma/erdma_verbs.c
1182
if (IS_ERR(mr->mem.mtt)) {
drivers/infiniband/hw/erdma/erdma_verbs.c
1183
ret = PTR_ERR(mr->mem.mtt);
drivers/infiniband/hw/erdma/erdma_verbs.c
1194
erdma_destroy_mtt(dev, mr->mem.mtt);
drivers/infiniband/hw/erdma/erdma_verbs.c
1210
if (mr->mem.mtt_nents >= mr->mem.page_cnt)
drivers/infiniband/hw/erdma/erdma_verbs.c
1213
mr->mem.mtt->buf[mr->mem.mtt_nents] = addr;
drivers/infiniband/hw/erdma/erdma_verbs.c
1214
mr->mem.mtt_nents++;
drivers/infiniband/hw/erdma/erdma_verbs.c
1225
mr->mem.mtt_nents = 0;
drivers/infiniband/hw/erdma/erdma_verbs.c
1252
ret = get_mtt_entries(dev, &mr->mem, start, len, access, virt,
drivers/infiniband/hw/erdma/erdma_verbs.c
1263
mr->mem.va = virt;
drivers/infiniband/hw/erdma/erdma_verbs.c
1264
mr->mem.len = len;
drivers/infiniband/hw/erdma/erdma_verbs.c
1280
put_mtt_entries(dev, &mr->mem);
drivers/infiniband/hw/erdma/erdma_verbs.c
1310
put_mtt_entries(dev, &mr->mem);
drivers/infiniband/hw/erdma/erdma_verbs.c
147
mr->mem.page_cnt > ERDMA_MAX_INLINE_MTT_ENTRIES) {
drivers/infiniband/hw/erdma/erdma_verbs.c
148
if (mr->mem.mtt->continuous) {
drivers/infiniband/hw/erdma/erdma_verbs.c
149
req.phy_addr[0] = mr->mem.mtt->buf_dma;
drivers/infiniband/hw/erdma/erdma_verbs.c
152
req.phy_addr[0] = mr->mem.mtt->dma_addrs[0];
drivers/infiniband/hw/erdma/erdma_verbs.c
153
mtt_level = mr->mem.mtt->level;
drivers/infiniband/hw/erdma/erdma_verbs.c
156
memcpy(req.phy_addr, mr->mem.mtt->buf,
drivers/infiniband/hw/erdma/erdma_verbs.c
157
MTT_SIZE(mr->mem.page_cnt));
drivers/infiniband/hw/erdma/erdma_verbs.c
167
ilog2(mr->mem.page_size)) |
drivers/infiniband/hw/erdma/erdma_verbs.c
169
FIELD_PREP(ERDMA_CMD_REGMR_MTT_CNT_MASK, mr->mem.page_cnt);
drivers/infiniband/hw/erdma/erdma_verbs.c
175
req.start_va = mr->mem.va;
drivers/infiniband/hw/erdma/erdma_verbs.c
176
req.size = mr->mem.len;
drivers/infiniband/hw/erdma/erdma_verbs.c
179
if (!mr->mem.mtt->continuous && mr->mem.mtt->level > 1) {
drivers/infiniband/hw/erdma/erdma_verbs.c
183
req.size_h = upper_32_bits(mr->mem.len);
drivers/infiniband/hw/erdma/erdma_verbs.c
184
req.mtt_cnt_h = mr->mem.page_cnt >> 20;
drivers/infiniband/hw/erdma/erdma_verbs.c
196
struct erdma_mem *mem;
drivers/infiniband/hw/erdma/erdma_verbs.c
22
static void assemble_qbuf_mtt_for_cmd(struct erdma_mem *mem, u32 *cfg,
drivers/infiniband/hw/erdma/erdma_verbs.c
220
mem = &cq->user_cq.qbuf_mem;
drivers/infiniband/hw/erdma/erdma_verbs.c
223
ilog2(mem->page_size) - ERDMA_HW_PAGE_SHIFT);
drivers/infiniband/hw/erdma/erdma_verbs.c
224
if (mem->mtt_nents == 1) {
drivers/infiniband/hw/erdma/erdma_verbs.c
225
req.qbuf_addr_l = lower_32_bits(mem->mtt->buf[0]);
drivers/infiniband/hw/erdma/erdma_verbs.c
226
req.qbuf_addr_h = upper_32_bits(mem->mtt->buf[0]);
drivers/infiniband/hw/erdma/erdma_verbs.c
231
req.qbuf_addr_l = lower_32_bits(mem->mtt->buf_dma);
drivers/infiniband/hw/erdma/erdma_verbs.c
232
req.qbuf_addr_h = upper_32_bits(mem->mtt->buf_dma);
drivers/infiniband/hw/erdma/erdma_verbs.c
238
mem->mtt_nents);
drivers/infiniband/hw/erdma/erdma_verbs.c
240
req.first_page_offset = mem->page_offset;
drivers/infiniband/hw/erdma/erdma_verbs.c
25
struct erdma_mtt *mtt = mem->mtt;
drivers/infiniband/hw/erdma/erdma_verbs.c
27
if (mem->mtt_nents > ERDMA_MAX_INLINE_MTT_ENTRIES) {
drivers/infiniband/hw/erdma/erdma_verbs.c
33
memcpy(addr1, mtt->buf + 1, MTT_SIZE(mem->mtt_nents - 1));
drivers/infiniband/hw/erdma/erdma_verbs.c
584
static void erdma_fill_bottom_mtt(struct erdma_dev *dev, struct erdma_mem *mem)
drivers/infiniband/hw/erdma/erdma_verbs.c
586
struct erdma_mtt *mtt = mem->mtt;
drivers/infiniband/hw/erdma/erdma_verbs.c
593
rdma_umem_for_each_dma_block(mem->umem, &biter, mem->page_size)
drivers/infiniband/hw/erdma/erdma_verbs.c
826
static int get_mtt_entries(struct erdma_dev *dev, struct erdma_mem *mem,
drivers/infiniband/hw/erdma/erdma_verbs.c
832
mem->umem = ib_umem_get(&dev->ibdev, start, len, access);
drivers/infiniband/hw/erdma/erdma_verbs.c
833
if (IS_ERR(mem->umem)) {
drivers/infiniband/hw/erdma/erdma_verbs.c
834
ret = PTR_ERR(mem->umem);
drivers/infiniband/hw/erdma/erdma_verbs.c
835
mem->umem = NULL;
drivers/infiniband/hw/erdma/erdma_verbs.c
839
mem->va = virt;
drivers/infiniband/hw/erdma/erdma_verbs.c
840
mem->len = len;
drivers/infiniband/hw/erdma/erdma_verbs.c
841
mem->page_size = ib_umem_find_best_pgsz(mem->umem, req_page_size, virt);
drivers/infiniband/hw/erdma/erdma_verbs.c
842
mem->page_offset = start & (mem->page_size - 1);
drivers/infiniband/hw/erdma/erdma_verbs.c
843
mem->mtt_nents = ib_umem_num_dma_blocks(mem->umem, mem->page_size);
drivers/infiniband/hw/erdma/erdma_verbs.c
844
mem->page_cnt = mem->mtt_nents;
drivers/infiniband/hw/erdma/erdma_verbs.c
845
mem->mtt = erdma_create_mtt(dev, MTT_SIZE(mem->page_cnt),
drivers/infiniband/hw/erdma/erdma_verbs.c
847
if (IS_ERR(mem->mtt)) {
drivers/infiniband/hw/erdma/erdma_verbs.c
848
ret = PTR_ERR(mem->mtt);
drivers/infiniband/hw/erdma/erdma_verbs.c
852
erdma_fill_bottom_mtt(dev, mem);
drivers/infiniband/hw/erdma/erdma_verbs.c
857
if (mem->umem) {
drivers/infiniband/hw/erdma/erdma_verbs.c
858
ib_umem_release(mem->umem);
drivers/infiniband/hw/erdma/erdma_verbs.c
859
mem->umem = NULL;
drivers/infiniband/hw/erdma/erdma_verbs.c
865
static void put_mtt_entries(struct erdma_dev *dev, struct erdma_mem *mem)
drivers/infiniband/hw/erdma/erdma_verbs.c
867
if (mem->mtt)
drivers/infiniband/hw/erdma/erdma_verbs.c
868
erdma_destroy_mtt(dev, mem->mtt);
drivers/infiniband/hw/erdma/erdma_verbs.c
870
if (mem->umem) {
drivers/infiniband/hw/erdma/erdma_verbs.c
871
ib_umem_release(mem->umem);
drivers/infiniband/hw/erdma/erdma_verbs.c
872
mem->umem = NULL;
drivers/infiniband/hw/erdma/erdma_verbs.h
126
struct erdma_mem mem;
drivers/infiniband/hw/irdma/cm.c
3157
16, 8, rbuf->mem.va, rbuf->totallen, false);
drivers/infiniband/hw/irdma/cm.c
3168
ethh = rbuf->mem.va;
drivers/infiniband/hw/irdma/cm.c
328
buf = sqbuf->mem.va;
drivers/infiniband/hw/irdma/cm.c
397
16, 8, sqbuf->mem.va, sqbuf->totallen, false);
drivers/infiniband/hw/irdma/cm.c
436
buf = sqbuf->mem.va;
drivers/infiniband/hw/irdma/cm.c
585
16, 8, sqbuf->mem.va, sqbuf->totallen, false);
drivers/infiniband/hw/irdma/hmc.c
336
struct irdma_dma_mem *mem;
drivers/infiniband/hw/irdma/hmc.c
348
mem = (sd_entry->entry_type == IRDMA_SD_TYPE_PAGED) ?
drivers/infiniband/hw/irdma/hmc.c
352
if (!mem || !mem->va) {
drivers/infiniband/hw/irdma/hmc.c
355
dma_free_coherent(dev->hw->device, mem->size, mem->va,
drivers/infiniband/hw/irdma/hmc.c
356
mem->pa);
drivers/infiniband/hw/irdma/hmc.c
357
mem->va = NULL;
drivers/infiniband/hw/irdma/hmc.c
557
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/hmc.c
558
struct irdma_dma_mem *page = &mem;
drivers/infiniband/hw/irdma/hmc.c
632
struct irdma_dma_mem *mem;
drivers/infiniband/hw/irdma/hmc.c
658
mem = &pd_entry->bp.addr;
drivers/infiniband/hw/irdma/hmc.c
659
if (!mem || !mem->va)
drivers/infiniband/hw/irdma/hmc.c
662
dma_free_coherent(dev->hw->device, mem->size, mem->va,
drivers/infiniband/hw/irdma/hmc.c
663
mem->pa);
drivers/infiniband/hw/irdma/hmc.c
664
mem->va = NULL;
drivers/infiniband/hw/irdma/hw.c
1025
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/hw.c
1059
status = irdma_obj_aligned_mem(rf, &mem, sizeof(struct irdma_cqp_ctx),
drivers/infiniband/hw/irdma/hw.c
1064
dev->cqp->host_ctx_pa = mem.pa;
drivers/infiniband/hw/irdma/hw.c
1065
dev->cqp->host_ctx = mem.va;
drivers/infiniband/hw/irdma/hw.c
1071
cqp_init_info.host_ctx_pa = mem.pa;
drivers/infiniband/hw/irdma/hw.c
1072
cqp_init_info.host_ctx = mem.va;
drivers/infiniband/hw/irdma/hw.c
1319
iwceq->mem.size = ALIGN(sizeof(struct irdma_ceqe) * ceq_size,
drivers/infiniband/hw/irdma/hw.c
1321
iwceq->mem.va = dma_alloc_coherent(dev->hw->device, iwceq->mem.size,
drivers/infiniband/hw/irdma/hw.c
1322
&iwceq->mem.pa, GFP_KERNEL);
drivers/infiniband/hw/irdma/hw.c
1323
if (!iwceq->mem.va)
drivers/infiniband/hw/irdma/hw.c
1327
info.ceqe_base = iwceq->mem.va;
drivers/infiniband/hw/irdma/hw.c
1328
info.ceqe_pa = iwceq->mem.pa;
drivers/infiniband/hw/irdma/hw.c
1343
dma_free_coherent(dev->hw->device, iwceq->mem.size,
drivers/infiniband/hw/irdma/hw.c
1344
iwceq->mem.va, iwceq->mem.pa);
drivers/infiniband/hw/irdma/hw.c
1345
iwceq->mem.va = NULL;
drivers/infiniband/hw/irdma/hw.c
1466
aeq->mem.size = sizeof(struct irdma_sc_aeqe) * size;
drivers/infiniband/hw/irdma/hw.c
1467
aeq->mem.va = vzalloc(aeq->mem.size);
drivers/infiniband/hw/irdma/hw.c
1469
if (!aeq->mem.va)
drivers/infiniband/hw/irdma/hw.c
1472
pg_cnt = DIV_ROUND_UP(aeq->mem.size, PAGE_SIZE);
drivers/infiniband/hw/irdma/hw.c
1475
vfree(aeq->mem.va);
drivers/infiniband/hw/irdma/hw.c
1480
status = irdma_map_vm_page_list(&rf->hw, aeq->mem.va, pg_arr, pg_cnt);
drivers/infiniband/hw/irdma/hw.c
1483
vfree(aeq->mem.va);
drivers/infiniband/hw/irdma/hw.c
1514
aeq->mem.size = ALIGN(sizeof(struct irdma_sc_aeqe) * aeq_size,
drivers/infiniband/hw/irdma/hw.c
1516
aeq->mem.va = dma_alloc_coherent(dev->hw->device, aeq->mem.size,
drivers/infiniband/hw/irdma/hw.c
1517
&aeq->mem.pa,
drivers/infiniband/hw/irdma/hw.c
1519
if (aeq->mem.va)
drivers/infiniband/hw/irdma/hw.c
1535
info.aeqe_base = aeq->mem.va;
drivers/infiniband/hw/irdma/hw.c
1536
info.aeq_elem_pa = aeq->mem.pa;
drivers/infiniband/hw/irdma/hw.c
1554
dma_free_coherent(dev->hw->device, aeq->mem.size, aeq->mem.va,
drivers/infiniband/hw/irdma/hw.c
1555
aeq->mem.pa);
drivers/infiniband/hw/irdma/hw.c
1556
aeq->mem.va = NULL;
drivers/infiniband/hw/irdma/hw.c
1731
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/hw.c
1747
status = irdma_obj_aligned_mem(rf, &mem, IRDMA_QUERY_FPM_BUF_SIZE,
drivers/infiniband/hw/irdma/hw.c
1752
info.fpm_query_buf_pa = mem.pa;
drivers/infiniband/hw/irdma/hw.c
1753
info.fpm_query_buf = mem.va;
drivers/infiniband/hw/irdma/hw.c
1755
status = irdma_obj_aligned_mem(rf, &mem, IRDMA_COMMIT_FPM_BUF_SIZE,
drivers/infiniband/hw/irdma/hw.c
1760
info.fpm_commit_buf_pa = mem.pa;
drivers/infiniband/hw/irdma/hw.c
1761
info.fpm_commit_buf = mem.va;
drivers/infiniband/hw/irdma/hw.c
703
u32 pg_cnt = DIV_ROUND_UP(aeq->mem.size, PAGE_SIZE);
drivers/infiniband/hw/irdma/hw.c
708
vfree(aeq->mem.va);
drivers/infiniband/hw/irdma/hw.c
743
dma_free_coherent(dev->hw->device, aeq->mem.size, aeq->mem.va,
drivers/infiniband/hw/irdma/hw.c
744
aeq->mem.pa);
drivers/infiniband/hw/irdma/hw.c
745
aeq->mem.va = NULL;
drivers/infiniband/hw/irdma/hw.c
776
dma_free_coherent(dev->hw->device, iwceq->mem.size, iwceq->mem.va,
drivers/infiniband/hw/irdma/hw.c
777
iwceq->mem.pa);
drivers/infiniband/hw/irdma/hw.c
778
iwceq->mem.va = NULL;
drivers/infiniband/hw/irdma/hw.c
835
dma_free_coherent(rf->sc_dev.hw->device, iwceq->mem.size,
drivers/infiniband/hw/irdma/hw.c
836
iwceq->mem.va, iwceq->mem.pa);
drivers/infiniband/hw/irdma/hw.c
837
iwceq->mem.va = NULL;
drivers/infiniband/hw/irdma/main.h
199
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/main.h
209
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/osdep.h
59
struct irdma_dma_mem *mem);
drivers/infiniband/hw/irdma/pble.c
141
struct irdma_dma_mem mem;
drivers/infiniband/hw/irdma/pble.c
162
mem.pa = (u64)chunk->dmainfo.dmaaddrs[i];
drivers/infiniband/hw/irdma/pble.c
163
mem.size = 4096;
drivers/infiniband/hw/irdma/pble.c
164
mem.va = addr;
drivers/infiniband/hw/irdma/pble.c
168
pd_idx++, &mem);
drivers/infiniband/hw/irdma/protos.h
90
struct irdma_dma_mem *mem);
drivers/infiniband/hw/irdma/puda.c
100
FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_LEN, buf->mem.size));
drivers/infiniband/hw/irdma/puda.c
103
FIELD_PREP(IRDMAQPSQ_FRAG_LEN, buf->mem.size) |
drivers/infiniband/hw/irdma/puda.c
1120
dma_sync_single_for_device(qp->dev->hw->device, buf->mem.pa,
drivers/infiniband/hw/irdma/puda.c
1121
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
1196
void *mem1 = (u8 *)buf->mem.va + buf_offset;
drivers/infiniband/hw/irdma/puda.c
1197
void *mem2 = (u8 *)txbuf->mem.va + txbuf_offset;
drivers/infiniband/hw/irdma/puda.c
1269
txbuf->data = (u8 *)txbuf->mem.va + txoffset;
drivers/infiniband/hw/irdma/puda.c
1273
txbuf->data = (u8 *)txbuf->mem.va + buf->hdrlen;
drivers/infiniband/hw/irdma/puda.c
1275
bufoffset = (u16)(buf->data - (u8 *)buf->mem.va);
drivers/infiniband/hw/irdma/puda.c
1297
bufoffset = (u16)(buf->data - (u8 *)buf->mem.va);
drivers/infiniband/hw/irdma/puda.c
1401
txbuf->mem.va, txbuf->totallen, false);
drivers/infiniband/hw/irdma/puda.c
1444
ioffset = (u16)(buf->data - (u8 *)buf->mem.va);
drivers/infiniband/hw/irdma/puda.c
1499
16, 8, txbuf->mem.va, txbuf->totallen,
drivers/infiniband/hw/irdma/puda.c
151
buf->mem.size = len;
drivers/infiniband/hw/irdma/puda.c
152
buf->mem.va = kzalloc(buf->mem.size, GFP_KERNEL);
drivers/infiniband/hw/irdma/puda.c
153
if (!buf->mem.va)
drivers/infiniband/hw/irdma/puda.c
155
buf->mem.pa = dma_map_single(dev->hw->device, buf->mem.va,
drivers/infiniband/hw/irdma/puda.c
156
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
157
if (dma_mapping_error(dev->hw->device, buf->mem.pa)) {
drivers/infiniband/hw/irdma/puda.c
158
kfree(buf->mem.va);
drivers/infiniband/hw/irdma/puda.c
1599
buf->mem.va, buf->totallen, false);
drivers/infiniband/hw/irdma/puda.c
180
dma_unmap_single(dev->hw->device, buf->mem.pa, buf->mem.size,
drivers/infiniband/hw/irdma/puda.c
182
kfree(buf->mem.va);
drivers/infiniband/hw/irdma/puda.c
366
dma_sync_single_for_cpu(dev->hw->device, buf->mem.pa,
drivers/infiniband/hw/irdma/puda.c
367
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
397
dma_sync_single_for_cpu(dev->hw->device, buf->mem.pa,
drivers/infiniband/hw/irdma/puda.c
398
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
532
info.paddr = buf->mem.pa;
drivers/infiniband/hw/irdma/puda.c
545
dma_sync_single_for_cpu(rsrc->dev->hw->device, buf->mem.pa,
drivers/infiniband/hw/irdma/puda.c
546
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
650
struct irdma_dma_mem *mem;
drivers/infiniband/hw/irdma/puda.c
662
mem = &rsrc->qpmem;
drivers/infiniband/hw/irdma/puda.c
663
memset(mem->va, 0, rsrc->qpmem.size);
drivers/infiniband/hw/irdma/puda.c
670
qp->sq_pa = mem->pa;
drivers/infiniband/hw/irdma/puda.c
673
ukqp->sq_base = mem->va;
drivers/infiniband/hw/irdma/puda.c
771
struct irdma_dma_mem *mem;
drivers/infiniband/hw/irdma/puda.c
784
mem = &rsrc->cqmem;
drivers/infiniband/hw/irdma/puda.c
789
info.cq_base_pa = mem->pa;
drivers/infiniband/hw/irdma/puda.c
790
info.shadow_area_pa = mem->pa + cqsize;
drivers/infiniband/hw/irdma/puda.c
791
init_info->cq_base = mem->va;
drivers/infiniband/hw/irdma/puda.c
792
init_info->shadow_area = (__le64 *)((u8 *)mem->va + cqsize);
drivers/infiniband/hw/irdma/puda.c
87
dma_sync_single_for_device(rsrc->dev->hw->device, buf->mem.pa,
drivers/infiniband/hw/irdma/puda.c
88
buf->mem.size, DMA_BIDIRECTIONAL);
drivers/infiniband/hw/irdma/puda.c
97
set_64bit_val(wqe, 0, buf->mem.pa);
drivers/infiniband/hw/irdma/puda.h
58
struct irdma_dma_mem mem; /* DMA memory for the buffer */
drivers/infiniband/hw/irdma/utils.c
1503
u8 *addr = buf->mem.va;
drivers/infiniband/hw/irdma/utils.c
1528
addr = buf->mem.va;
drivers/infiniband/hw/irdma/utils.c
1547
u8 *mem = buf->mem.va;
drivers/infiniband/hw/irdma/utils.c
1548
struct ethhdr *ethh = buf->mem.va;
drivers/infiniband/hw/irdma/utils.c
1559
buf->iph = mem + buf->maclen;
drivers/infiniband/hw/irdma/utils.c
1599
u8 *mem;
drivers/infiniband/hw/irdma/utils.c
1604
mem = buf->mem.va;
drivers/infiniband/hw/irdma/utils.c
1611
buf->iph = mem + IRDMA_IPV4_PAD;
drivers/infiniband/hw/irdma/utils.c
1613
buf->iph = mem;
drivers/infiniband/hw/irdma/utils.c
1615
buf->tcph = mem + IRDMA_TCP_OFFSET;
drivers/infiniband/hw/irdma/verbs.c
2299
struct irdma_dma_mem *mem = &iwsrq->kmem;
drivers/infiniband/hw/irdma/verbs.c
2305
mem->size = ALIGN(size, 256);
drivers/infiniband/hw/irdma/verbs.c
2306
mem->va = dma_alloc_coherent(iwdev->rf->hw.device, mem->size,
drivers/infiniband/hw/irdma/verbs.c
2307
&mem->pa, GFP_KERNEL);
drivers/infiniband/hw/irdma/verbs.c
2308
if (!mem->va)
drivers/infiniband/hw/irdma/verbs.c
2311
ukinfo->srq = mem->va;
drivers/infiniband/hw/irdma/verbs.c
2313
ukinfo->shadow_area = mem->va + ring_size;
drivers/infiniband/hw/irdma/verbs.c
2315
info->srq_pa = mem->pa;
drivers/infiniband/hw/irdma/verbs.c
697
struct irdma_dma_mem *mem = &iwqp->kqp.dma_mem;
drivers/infiniband/hw/irdma/verbs.c
732
mem->size = ALIGN(size, 256);
drivers/infiniband/hw/irdma/verbs.c
733
mem->va = dma_alloc_coherent(iwdev->rf->hw.device, mem->size,
drivers/infiniband/hw/irdma/verbs.c
734
&mem->pa, GFP_KERNEL);
drivers/infiniband/hw/irdma/verbs.c
735
if (!mem->va) {
drivers/infiniband/hw/irdma/verbs.c
743
ukinfo->sq = mem->va;
drivers/infiniband/hw/irdma/verbs.c
744
info->sq_pa = mem->pa;
drivers/infiniband/hw/mlx5/wr.c
362
struct ib_sig_domain *mem = &sig_attrs->mem;
drivers/infiniband/hw/mlx5/wr.c
374
switch (sig_attrs->mem.sig_type) {
drivers/infiniband/hw/mlx5/wr.c
378
basic->mem.bs_selector = bs_selector(mem->sig.dif.pi_interval);
drivers/infiniband/hw/mlx5/wr.c
380
mlx5_fill_inl_bsf(mem, &bsf->m_inl);
drivers/infiniband/hw/mlx5/wr.c
391
if (mem->sig.dif.pi_interval == wire->sig.dif.pi_interval &&
drivers/infiniband/hw/mlx5/wr.c
392
mem->sig_type == wire->sig_type) {
drivers/infiniband/hw/mlx5/wr.c
395
if (mem->sig.dif.bg_type == wire->sig.dif.bg_type)
drivers/infiniband/hw/mlx5/wr.c
397
if (mem->sig.dif.app_tag == wire->sig.dif.app_tag)
drivers/infiniband/hw/mlx5/wr.c
399
if (mem->sig.dif.ref_tag == wire->sig.dif.ref_tag)
drivers/infiniband/hw/mlx5/wr.c
480
u16 block_size = sig_attrs->mem.sig.dif.pi_interval;
drivers/infiniband/hw/mlx5/wr.c
487
prot_size = prot_field_size(sig_attrs->mem.sig_type);
drivers/infiniband/hw/mlx5/wr.c
583
if (sig_attrs->mem.sig_type != IB_SIG_TYPE_NONE)
drivers/infiniband/hw/mlx5/wr.c
914
&sig_attrs->mem, mr->sig->psv_memory.psv_idx,
drivers/infiniband/hw/mthca/mthca_memfree.c
107
static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask)
drivers/infiniband/hw/mthca/mthca_memfree.c
119
sg_set_page(mem, page, PAGE_SIZE << order, 0);
drivers/infiniband/hw/mthca/mthca_memfree.c
123
static int mthca_alloc_icm_coherent(struct device *dev, struct scatterlist *mem,
drivers/infiniband/hw/mthca/mthca_memfree.c
126
void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, &sg_dma_address(mem),
drivers/infiniband/hw/mthca/mthca_memfree.c
131
sg_set_buf(mem, buf, PAGE_SIZE << order);
drivers/infiniband/hw/mthca/mthca_memfree.c
132
BUG_ON(mem->offset);
drivers/infiniband/hw/mthca/mthca_memfree.c
133
sg_dma_len(mem) = PAGE_SIZE << order;
drivers/infiniband/hw/mthca/mthca_memfree.c
164
sg_init_table(chunk->mem, MTHCA_ICM_CHUNK_LEN);
drivers/infiniband/hw/mthca/mthca_memfree.c
175
&chunk->mem[chunk->npages],
drivers/infiniband/hw/mthca/mthca_memfree.c
178
ret = mthca_alloc_icm_pages(&chunk->mem[chunk->npages],
drivers/infiniband/hw/mthca/mthca_memfree.c
188
dma_map_sg(&dev->pdev->dev, chunk->mem,
drivers/infiniband/hw/mthca/mthca_memfree.c
208
chunk->nsg = dma_map_sg(&dev->pdev->dev, chunk->mem,
drivers/infiniband/hw/mthca/mthca_memfree.c
300
if (sg_dma_len(&chunk->mem[i]) > dma_offset)
drivers/infiniband/hw/mthca/mthca_memfree.c
301
*dma_handle = sg_dma_address(&chunk->mem[i]) +
drivers/infiniband/hw/mthca/mthca_memfree.c
303
dma_offset -= sg_dma_len(&chunk->mem[i]);
drivers/infiniband/hw/mthca/mthca_memfree.c
308
if (chunk->mem[i].length > offset) {
drivers/infiniband/hw/mthca/mthca_memfree.c
309
page = sg_page(&chunk->mem[i]);
drivers/infiniband/hw/mthca/mthca_memfree.c
312
offset -= chunk->mem[i].length;
drivers/infiniband/hw/mthca/mthca_memfree.c
480
sg_set_page(&db_tab->page[i].mem, pages[0], MTHCA_ICM_PAGE_SIZE,
drivers/infiniband/hw/mthca/mthca_memfree.c
483
ret = dma_map_sg(&dev->pdev->dev, &db_tab->page[i].mem, 1,
drivers/infiniband/hw/mthca/mthca_memfree.c
490
ret = mthca_MAP_ICM_page(dev, sg_dma_address(&db_tab->page[i].mem),
drivers/infiniband/hw/mthca/mthca_memfree.c
493
dma_unmap_sg(&dev->pdev->dev, &db_tab->page[i].mem, 1,
drivers/infiniband/hw/mthca/mthca_memfree.c
495
unpin_user_page(sg_page(&db_tab->page[i].mem));
drivers/infiniband/hw/mthca/mthca_memfree.c
543
sg_init_table(&db_tab->page[i].mem, 1);
drivers/infiniband/hw/mthca/mthca_memfree.c
560
dma_unmap_sg(&dev->pdev->dev, &db_tab->page[i].mem, 1,
drivers/infiniband/hw/mthca/mthca_memfree.c
562
unpin_user_page(sg_page(&db_tab->page[i].mem));
drivers/infiniband/hw/mthca/mthca_memfree.c
59
struct scatterlist mem;
drivers/infiniband/hw/mthca/mthca_memfree.c
69
dma_unmap_sg(&dev->pdev->dev, chunk->mem, chunk->npages,
drivers/infiniband/hw/mthca/mthca_memfree.c
73
__free_pages(sg_page(&chunk->mem[i]),
drivers/infiniband/hw/mthca/mthca_memfree.c
74
get_order(chunk->mem[i].length));
drivers/infiniband/hw/mthca/mthca_memfree.c
82
dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length,
drivers/infiniband/hw/mthca/mthca_memfree.c
83
lowmem_page_address(sg_page(&chunk->mem[i])),
drivers/infiniband/hw/mthca/mthca_memfree.c
84
sg_dma_address(&chunk->mem[i]));
drivers/infiniband/hw/mthca/mthca_memfree.h
130
return sg_dma_address(&iter->chunk->mem[iter->page_idx]);
drivers/infiniband/hw/mthca/mthca_memfree.h
135
return sg_dma_len(&iter->chunk->mem[iter->page_idx]);
drivers/infiniband/hw/mthca/mthca_memfree.h
55
struct scatterlist mem[MTHCA_ICM_CHUNK_LEN];
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
69
struct stats_mem *mem = &dev->stats_mem;
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
73
mem->size = max_t(u32, sizeof(struct ocrdma_rdma_stats_req),
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
76
mem->va = dma_alloc_coherent(&dev->nic_info.pdev->dev, mem->size,
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
77
&mem->pa, GFP_KERNEL);
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
78
if (!mem->va) {
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
84
mem->debugfs_mem = kzalloc(OCRDMA_MAX_DBGFS_MEM, GFP_KERNEL);
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
85
if (!mem->debugfs_mem)
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
93
struct stats_mem *mem = &dev->stats_mem;
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
95
if (mem->va)
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
96
dma_free_coherent(&dev->nic_info.pdev->dev, mem->size,
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
97
mem->va, mem->pa);
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
98
mem->va = NULL;
drivers/infiniband/hw/ocrdma/ocrdma_stats.c
99
kfree(mem->debugfs_mem);
drivers/infiniband/sw/siw/siw.h
163
struct siw_mem *mem;
drivers/infiniband/sw/siw/siw.h
192
struct siw_mem *mem[SIW_MAX_SGE]; /* per sge's resolved mem */
drivers/infiniband/sw/siw/siw.h
474
#define rx_mem(rctx) ((rctx)->wqe_active.mem[0])
drivers/infiniband/sw/siw/siw.h
716
#define siw_dbg_mem(mem, fmt, ...) \
drivers/infiniband/sw/siw/siw.h
717
ibdev_dbg(&mem->sdev->base_dev, \
drivers/infiniband/sw/siw/siw.h
718
"MEM[0x%08x] %s: " fmt, mem->stag, __func__, ##__VA_ARGS__)
drivers/infiniband/sw/siw/siw_mem.c
103
found = xa_erase(&mem->sdev->mem_xa, mem->stag >> 8);
drivers/infiniband/sw/siw/siw_mem.c
104
WARN_ON(found != mem);
drivers/infiniband/sw/siw/siw_mem.c
105
siw_mem_put(mem);
drivers/infiniband/sw/siw/siw_mem.c
110
struct siw_mem *mem = container_of(ref, struct siw_mem, ref);
drivers/infiniband/sw/siw/siw_mem.c
112
siw_dbg_mem(mem, "free mem, pbl: %s\n", mem->is_pbl ? "y" : "n");
drivers/infiniband/sw/siw/siw_mem.c
114
if (!mem->is_mw && mem->mem_obj) {
drivers/infiniband/sw/siw/siw_mem.c
115
if (mem->is_pbl == 0)
drivers/infiniband/sw/siw/siw_mem.c
116
siw_umem_release(mem->umem);
drivers/infiniband/sw/siw/siw_mem.c
118
kfree(mem->pbl);
drivers/infiniband/sw/siw/siw_mem.c
120
kfree(mem);
drivers/infiniband/sw/siw/siw_mem.c
136
int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr,
drivers/infiniband/sw/siw/siw_mem.c
139
if (!mem->stag_valid) {
drivers/infiniband/sw/siw/siw_mem.c
140
siw_dbg_pd(pd, "STag 0x%08x invalid\n", mem->stag);
drivers/infiniband/sw/siw/siw_mem.c
143
if (mem->pd != pd) {
drivers/infiniband/sw/siw/siw_mem.c
144
siw_dbg_pd(pd, "STag 0x%08x: PD mismatch\n", mem->stag);
drivers/infiniband/sw/siw/siw_mem.c
150
if ((mem->perms & perms) < perms) {
drivers/infiniband/sw/siw/siw_mem.c
152
mem->perms, perms);
drivers/infiniband/sw/siw/siw_mem.c
158
if (addr < mem->va || addr + len > mem->va + mem->len) {
drivers/infiniband/sw/siw/siw_mem.c
164
(void *)(uintptr_t)mem->va,
drivers/infiniband/sw/siw/siw_mem.c
165
(void *)(uintptr_t)(mem->va + mem->len),
drivers/infiniband/sw/siw/siw_mem.c
166
mem->stag);
drivers/infiniband/sw/siw/siw_mem.c
190
int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[],
drivers/infiniband/sw/siw/siw_mem.c
201
if (*mem == NULL) {
drivers/infiniband/sw/siw/siw_mem.c
208
*mem = new;
drivers/infiniband/sw/siw/siw_mem.c
211
if (unlikely((*mem)->stag != sge->lkey)) {
drivers/infiniband/sw/siw/siw_mem.c
212
siw_dbg_mem((*mem), "STag mismatch: 0x%08x\n", sge->lkey);
drivers/infiniband/sw/siw/siw_mem.c
216
rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len);
drivers/infiniband/sw/siw/siw_mem.c
224
*mem = NULL;
drivers/infiniband/sw/siw/siw_mem.c
240
siw_unref_mem_sgl(wqe->mem, wqe->sqe.num_sge);
drivers/infiniband/sw/siw/siw_mem.c
244
siw_unref_mem_sgl(wqe->mem, wqe->rqe.num_sge);
drivers/infiniband/sw/siw/siw_mem.c
248
siw_unref_mem_sgl(wqe->mem, 1);
drivers/infiniband/sw/siw/siw_mem.c
263
struct siw_mem *mem = siw_mem_id2obj(sdev, stag >> 8);
drivers/infiniband/sw/siw/siw_mem.c
266
if (unlikely(!mem)) {
drivers/infiniband/sw/siw/siw_mem.c
270
if (unlikely(mem->pd != pd)) {
drivers/infiniband/sw/siw/siw_mem.c
279
mem->stag_valid = 0;
drivers/infiniband/sw/siw/siw_mem.c
283
siw_mem_put(mem);
drivers/infiniband/sw/siw/siw_mem.c
29
struct siw_mem *mem;
drivers/infiniband/sw/siw/siw_mem.c
32
mem = xa_load(&sdev->mem_xa, stag_index);
drivers/infiniband/sw/siw/siw_mem.c
33
if (likely(mem && kref_get_unless_zero(&mem->ref))) {
drivers/infiniband/sw/siw/siw_mem.c
35
return mem;
drivers/infiniband/sw/siw/siw_mem.c
61
struct siw_mem *mem = kzalloc_obj(*mem);
drivers/infiniband/sw/siw/siw_mem.c
65
if (!mem)
drivers/infiniband/sw/siw/siw_mem.c
68
mem->mem_obj = mem_obj;
drivers/infiniband/sw/siw/siw_mem.c
69
mem->stag_valid = 0;
drivers/infiniband/sw/siw/siw_mem.c
70
mem->sdev = sdev;
drivers/infiniband/sw/siw/siw_mem.c
71
mem->va = start;
drivers/infiniband/sw/siw/siw_mem.c
72
mem->len = len;
drivers/infiniband/sw/siw/siw_mem.c
73
mem->pd = pd;
drivers/infiniband/sw/siw/siw_mem.c
74
mem->perms = rights & IWARP_ACCESS_MASK;
drivers/infiniband/sw/siw/siw_mem.c
75
kref_init(&mem->ref);
drivers/infiniband/sw/siw/siw_mem.c
80
if (xa_alloc_cyclic(&sdev->mem_xa, &id, mem, limit, &next,
drivers/infiniband/sw/siw/siw_mem.c
82
kfree(mem);
drivers/infiniband/sw/siw/siw_mem.c
86
mr->mem = mem;
drivers/infiniband/sw/siw/siw_mem.c
88
mem->stag = id << 8;
drivers/infiniband/sw/siw/siw_mem.c
89
mr->base_mr.lkey = mr->base_mr.rkey = mem->stag;
drivers/infiniband/sw/siw/siw_mem.c
96
struct siw_mem *mem = mr->mem, *found;
drivers/infiniband/sw/siw/siw_mem.c
98
mem->stag_valid = 0;
drivers/infiniband/sw/siw/siw_mem.h
16
int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr,
drivers/infiniband/sw/siw/siw_mem.h
19
struct siw_mem *mem[], enum ib_access_flags perms,
drivers/infiniband/sw/siw/siw_mem.h
27
static inline void siw_mem_put(struct siw_mem *mem)
drivers/infiniband/sw/siw/siw_mem.h
29
kref_put(&mem->ref, siw_free_mem);
drivers/infiniband/sw/siw/siw_mem.h
32
static inline void siw_unref_mem_sgl(struct siw_mem **mem, unsigned int num_sge)
drivers/infiniband/sw/siw/siw_mem.h
35
if (*mem == NULL)
drivers/infiniband/sw/siw/siw_mem.h
38
siw_mem_put(*mem);
drivers/infiniband/sw/siw/siw_mem.h
39
*mem = NULL;
drivers/infiniband/sw/siw/siw_mem.h
40
mem++;
drivers/infiniband/sw/siw/siw_qp.c
248
memset(wqe->mem, 0, sizeof(*wqe->mem) * SIW_MAX_SGE);
drivers/infiniband/sw/siw/siw_qp.c
862
memset(wqe->mem, 0, sizeof(*wqe->mem) * SIW_MAX_SGE);
drivers/infiniband/sw/siw/siw_qp.c
969
memset(wqe->mem, 0, sizeof(*wqe->mem) * SIW_MAX_SGE);
drivers/infiniband/sw/siw/siw_qp_rx.c
128
struct siw_mem *mem, u64 addr, int len)
drivers/infiniband/sw/siw/siw_qp_rx.c
130
struct siw_pbl *pbl = mem->pbl;
drivers/infiniband/sw/siw/siw_qp_rx.c
131
u64 offset = addr - mem->va;
drivers/infiniband/sw/siw/siw_qp_rx.c
371
wqe->mem[i] = NULL;
drivers/infiniband/sw/siw/siw_qp_rx.c
470
struct siw_mem **mem, *mem_p;
drivers/infiniband/sw/siw/siw_qp_rx.c
484
mem = &wqe->mem[frx->sge_idx];
drivers/infiniband/sw/siw/siw_qp_rx.c
491
rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE,
drivers/infiniband/sw/siw/siw_qp_rx.c
501
mem_p = *mem;
drivers/infiniband/sw/siw/siw_qp_rx.c
549
struct siw_mem *mem;
drivers/infiniband/sw/siw/siw_qp_rx.c
582
mem = rx_mem(frx);
drivers/infiniband/sw/siw/siw_qp_rx.c
588
if (unlikely(mem->stag != srx->ddp_stag)) {
drivers/infiniband/sw/siw/siw_qp_rx.c
594
rv = siw_check_mem(qp->pd, mem, srx->ddp_to + srx->fpdu_part_rcvd,
drivers/infiniband/sw/siw/siw_qp_rx.c
606
rv = siw_rx_data(mem, srx, &frx->pbl_idx,
drivers/infiniband/sw/siw/siw_qp_rx.c
689
tx_work->mem[0] = NULL;
drivers/infiniband/sw/siw/siw_qp_rx.c
763
wqe->mem[0] = NULL;
drivers/infiniband/sw/siw/siw_qp_rx.c
787
struct siw_mem **mem, *mem_p;
drivers/infiniband/sw/siw/siw_qp_rx.c
824
mem = &wqe->mem[0];
drivers/infiniband/sw/siw/siw_qp_rx.c
826
if (!(*mem)) {
drivers/infiniband/sw/siw/siw_qp_rx.c
830
rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0,
drivers/infiniband/sw/siw/siw_qp_rx.c
845
mem_p = *mem;
drivers/infiniband/sw/siw/siw_qp_tx.c
25
static struct page *siw_get_pblpage(struct siw_mem *mem, u64 addr, int *idx)
drivers/infiniband/sw/siw/siw_qp_tx.c
27
struct siw_pbl *pbl = mem->pbl;
drivers/infiniband/sw/siw/siw_qp_tx.c
28
u64 offset = addr - mem->va;
drivers/infiniband/sw/siw/siw_qp_tx.c
37
static struct page *siw_get_page(struct siw_mem *mem, struct siw_sge *sge,
drivers/infiniband/sw/siw/siw_qp_tx.c
40
if (!mem->is_pbl)
drivers/infiniband/sw/siw/siw_qp_tx.c
41
return siw_get_upage(mem->umem, sge->laddr + offset);
drivers/infiniband/sw/siw/siw_qp_tx.c
43
return siw_get_pblpage(mem, sge->laddr + offset, pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
472
struct siw_mem *mem;
drivers/infiniband/sw/siw/siw_qp_tx.c
475
mem = wqe->mem[sge_idx];
drivers/infiniband/sw/siw/siw_qp_tx.c
476
is_kva = mem->mem_obj == NULL ? 1 : 0;
drivers/infiniband/sw/siw/siw_qp_tx.c
505
p = siw_get_page(mem, sge, sge_off, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
64
struct siw_mem *mem = wqe->mem[0];
drivers/infiniband/sw/siw/siw_qp_tx.c
66
if (!mem->mem_obj) {
drivers/infiniband/sw/siw/siw_qp_tx.c
777
int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0,
drivers/infiniband/sw/siw/siw_qp_tx.c
79
p = siw_get_page(mem, sge, 0, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
93
p = siw_get_page(mem, sge, part, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
935
struct siw_mem *mem;
drivers/infiniband/sw/siw/siw_qp_tx.c
950
mem = siw_mem_id2obj(sdev, sqe->rkey >> 8);
drivers/infiniband/sw/siw/siw_qp_tx.c
951
if (unlikely(!mem)) {
drivers/infiniband/sw/siw/siw_qp_tx.c
956
if (unlikely(mem->pd != pd)) {
drivers/infiniband/sw/siw/siw_qp_tx.c
961
if (unlikely(mem->stag_valid)) {
drivers/infiniband/sw/siw/siw_qp_tx.c
967
mem->stag = sqe->rkey;
drivers/infiniband/sw/siw/siw_qp_tx.c
968
mem->perms = sqe->access;
drivers/infiniband/sw/siw/siw_qp_tx.c
970
siw_dbg_mem(mem, "STag 0x%08x now valid\n", sqe->rkey);
drivers/infiniband/sw/siw/siw_qp_tx.c
971
mem->va = base_mr->iova;
drivers/infiniband/sw/siw/siw_qp_tx.c
972
mem->stag_valid = 1;
drivers/infiniband/sw/siw/siw_qp_tx.c
974
siw_mem_put(mem);
drivers/infiniband/sw/siw/siw_verbs.c
1307
siw_dbg_mem(mr->mem, "deregister MR\n");
drivers/infiniband/sw/siw/siw_verbs.c
1374
struct siw_mem *mem = mr->mem;
drivers/infiniband/sw/siw/siw_verbs.c
1386
mem->stag |= ureq.stag_key;
drivers/infiniband/sw/siw/siw_verbs.c
1387
uresp.stag = mem->stag;
drivers/infiniband/sw/siw/siw_verbs.c
1397
mr->mem->stag_valid = 1;
drivers/infiniband/sw/siw/siw_verbs.c
1404
if (mr->mem)
drivers/infiniband/sw/siw/siw_verbs.c
1453
mr->mem->is_pbl = 1;
drivers/infiniband/sw/siw/siw_verbs.c
1455
siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
drivers/infiniband/sw/siw/siw_verbs.c
1465
if (mr->mem)
drivers/infiniband/sw/siw/siw_verbs.c
1485
struct siw_mem *mem = mr->mem;
drivers/infiniband/sw/siw/siw_verbs.c
1486
struct siw_pbl *pbl = mem->pbl;
drivers/infiniband/sw/siw/siw_verbs.c
1492
siw_dbg_mem(mem, "no PBL allocated\n");
drivers/infiniband/sw/siw/siw_verbs.c
1498
siw_dbg_mem(mem, "too many SGE's: %d > %d\n",
drivers/infiniband/sw/siw/siw_verbs.c
1504
siw_dbg_mem(mem, "empty SGE\n");
drivers/infiniband/sw/siw/siw_verbs.c
1526
siw_dbg_mem(mem,
drivers/infiniband/sw/siw/siw_verbs.c
1533
mem->len = base_mr->length;
drivers/infiniband/sw/siw/siw_verbs.c
1534
mem->va = base_mr->iova;
drivers/infiniband/sw/siw/siw_verbs.c
1535
siw_dbg_mem(mem,
drivers/infiniband/sw/siw/siw_verbs.c
1537
mem->len, (void *)(uintptr_t)mem->va, num_sle,
drivers/infiniband/sw/siw/siw_verbs.c
1568
mr->mem->stag_valid = 1;
drivers/infiniband/sw/siw/siw_verbs.c
1570
siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
drivers/infiniband/ulp/iser/iser_memory.c
123
static int iser_reg_dma(struct iser_device *device, struct iser_data_buf *mem,
drivers/infiniband/ulp/iser/iser_memory.c
126
struct scatterlist *sg = mem->sg;
drivers/infiniband/ulp/iser/iser_memory.c
199
sig_attrs->mem.sig_type = IB_SIG_TYPE_NONE;
drivers/infiniband/ulp/iser/iser_memory.c
206
iser_set_dif_domain(sc, &sig_attrs->mem);
drivers/infiniband/ulp/iser/iser_memory.c
207
sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ?
drivers/infiniband/ulp/iser/iser_memory.c
214
iser_set_dif_domain(sc, &sig_attrs->mem);
drivers/infiniband/ulp/iser/iser_memory.c
215
sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ?
drivers/infiniband/ulp/iser/iser_memory.c
248
struct iser_data_buf *mem,
drivers/infiniband/ulp/iser/iser_memory.c
272
ret = ib_map_mr_sg_pi(mr, mem->sg, mem->dma_nents, NULL,
drivers/infiniband/ulp/iser/iser_memory.c
276
mem->dma_nents + sig_mem->dma_nents);
drivers/infiniband/ulp/iser/iser_memory.c
306
struct iser_data_buf *mem,
drivers/infiniband/ulp/iser/iser_memory.c
321
n = ib_map_mr_sg(mr, mem->sg, mem->dma_nents, NULL, SZ_4K);
drivers/infiniband/ulp/iser/iser_memory.c
322
if (unlikely(n != mem->dma_nents)) {
drivers/infiniband/ulp/iser/iser_memory.c
324
n, mem->dma_nents);
drivers/infiniband/ulp/iser/iser_memory.c
358
struct iser_data_buf *mem = &task->data[dir];
drivers/infiniband/ulp/iser/iser_memory.c
364
use_dma_key = mem->dma_nents == 1 && (all_imm || !iser_always_reg) &&
drivers/infiniband/ulp/iser/iser_memory.c
367
return iser_reg_dma(device, mem, reg);
drivers/infiniband/ulp/iser/iser_memory.c
371
err = iser_fast_reg_mr(task, mem, &desc->rsc, reg);
drivers/infiniband/ulp/iser/iser_memory.c
375
err = iser_reg_sig_mr(task, mem, &task->prot[dir],
drivers/infiniband/ulp/isert/ib_isert.c
2005
sig_attrs->mem.sig_type = IB_SIG_TYPE_NONE;
drivers/infiniband/ulp/isert/ib_isert.c
2011
isert_set_dif_domain(se_cmd, &sig_attrs->mem);
drivers/infiniband/ulp/isert/ib_isert.c
2016
isert_set_dif_domain(se_cmd, &sig_attrs->mem);
drivers/input/evdev.c
891
unsigned long *mem;
drivers/input/evdev.c
893
mem = bitmap_alloc(maxbit, GFP_KERNEL);
drivers/input/evdev.c
894
if (!mem)
drivers/input/evdev.c
900
bitmap_copy(mem, bits, maxbit);
drivers/input/evdev.c
908
ret = bits_to_user(mem, maxbit, maxlen, p, compat);
drivers/input/evdev.c
912
bitmap_free(mem);
drivers/input/misc/da7280.c
774
unsigned int i, mem[DA7280_SNP_MEM_SIZE];
drivers/input/misc/da7280.c
883
&mem[0], DA7280_SNP_MEM_SIZE);
drivers/input/misc/da7280.c
888
if (mem[i] <= 0xff) {
drivers/input/misc/da7280.c
889
haptics->snp_mem[i] = (u8)mem[i];
drivers/input/misc/da7280.c
893
i, mem[i]);
drivers/input/touchscreen/atmel_mxt_ts.c
1486
*(cfg->mem + byte_offset) = val;
drivers/input/touchscreen/atmel_mxt_ts.c
1512
size, cfg->mem + byte_offset);
drivers/input/touchscreen/atmel_mxt_ts.c
1645
cfg.mem = kzalloc(cfg.mem_size, GFP_KERNEL);
drivers/input/touchscreen/atmel_mxt_ts.c
1646
if (!cfg.mem) {
drivers/input/touchscreen/atmel_mxt_ts.c
1664
calculated_crc = mxt_calculate_crc(cfg.mem,
drivers/input/touchscreen/atmel_mxt_ts.c
1693
kfree(cfg.mem);
drivers/input/touchscreen/atmel_mxt_ts.c
277
u8 *mem;
drivers/input/touchscreen/fsl-imx25-tcq.c
505
void __iomem *mem;
drivers/input/touchscreen/fsl-imx25-tcq.c
513
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/input/touchscreen/fsl-imx25-tcq.c
514
if (IS_ERR(mem))
drivers/input/touchscreen/fsl-imx25-tcq.c
515
return PTR_ERR(mem);
drivers/input/touchscreen/fsl-imx25-tcq.c
521
priv->regs = devm_regmap_init_mmio(dev, mem, &mx25_tcq_regconfig);
drivers/iommu/dma-iommu.c
1444
p2pdma_state.mem, sg_phys(s));
drivers/iommu/iommufd/pages.c
727
void *mem;
drivers/iommu/iommufd/pages.c
729
mem = kmap_local_page(page);
drivers/iommu/iommufd/pages.c
731
memcpy(mem + offset, data, length);
drivers/iommu/iommufd/pages.c
734
memcpy(data, mem + offset, length);
drivers/iommu/iommufd/pages.c
736
kunmap_local(mem);
drivers/iommu/virtio-iommu.c
483
struct virtio_iommu_probe_resv_mem *mem,
drivers/iommu/virtio-iommu.c
492
start = start64 = le64_to_cpu(mem->start);
drivers/iommu/virtio-iommu.c
493
end = end64 = le64_to_cpu(mem->end);
drivers/iommu/virtio-iommu.c
500
if (len < sizeof(*mem))
drivers/iommu/virtio-iommu.c
503
switch (mem->subtype) {
drivers/iommu/virtio-iommu.c
506
mem->subtype);
drivers/irqchip/irq-tb10x.c
104
if (!request_mem_region(mem.start, resource_size(&mem),
drivers/irqchip/irq-tb10x.c
110
reg_base = ioremap(mem.start, resource_size(&mem));
drivers/irqchip/irq-tb10x.c
173
release_mem_region(mem.start, resource_size(&mem));
drivers/irqchip/irq-tb10x.c
93
struct resource mem;
drivers/irqchip/irq-tb10x.c
98
if (of_address_to_resource(ictl, 0, &mem)) {
drivers/leds/leds-bcm6328.c
117
mode = led->mem + BCM6328_REG_MODE_HI;
drivers/leds/leds-bcm6328.c
119
mode = led->mem + BCM6328_REG_MODE_LO;
drivers/leds/leds-bcm6328.c
212
val = bcm6328_led_read(led->mem + BCM6328_REG_INIT);
drivers/leds/leds-bcm6328.c
215
bcm6328_led_write(led->mem + BCM6328_REG_INIT, val);
drivers/leds/leds-bcm6328.c
236
val = bcm6328_led_read(led->mem + BCM6328_REG_INIT);
drivers/leds/leds-bcm6328.c
239
bcm6328_led_write(led->mem + BCM6328_REG_INIT, val);
drivers/leds/leds-bcm6328.c
256
void __iomem *mem, spinlock_t *lock)
drivers/leds/leds-bcm6328.c
262
val = bcm6328_led_read(mem + BCM6328_REG_HWDIS);
drivers/leds/leds-bcm6328.c
264
bcm6328_led_write(mem + BCM6328_REG_HWDIS, val);
drivers/leds/leds-bcm6328.c
278
addr = mem + BCM6328_REG_LNKACTSEL_LO;
drivers/leds/leds-bcm6328.c
280
addr = mem + BCM6328_REG_LNKACTSEL_HI;
drivers/leds/leds-bcm6328.c
305
addr = mem + BCM6328_REG_LNKACTSEL_LO;
drivers/leds/leds-bcm6328.c
307
addr = mem + BCM6328_REG_LNKACTSEL_HI;
drivers/leds/leds-bcm6328.c
328
void __iomem *mem, spinlock_t *lock,
drivers/leds/leds-bcm6328.c
343
led->mem = mem;
drivers/leds/leds-bcm6328.c
361
mode = mem + BCM6328_REG_MODE_HI;
drivers/leds/leds-bcm6328.c
363
mode = mem + BCM6328_REG_MODE_LO;
drivers/leds/leds-bcm6328.c
395
void __iomem *mem;
drivers/leds/leds-bcm6328.c
399
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/leds/leds-bcm6328.c
400
if (IS_ERR(mem))
drivers/leds/leds-bcm6328.c
401
return PTR_ERR(mem);
drivers/leds/leds-bcm6328.c
419
bcm6328_led_write(mem + BCM6328_REG_HWDIS, ~0);
drivers/leds/leds-bcm6328.c
420
bcm6328_led_write(mem + BCM6328_REG_LNKACTSEL_HI, 0);
drivers/leds/leds-bcm6328.c
421
bcm6328_led_write(mem + BCM6328_REG_LNKACTSEL_LO, 0);
drivers/leds/leds-bcm6328.c
423
val = bcm6328_led_read(mem + BCM6328_REG_INIT);
drivers/leds/leds-bcm6328.c
435
bcm6328_led_write(mem + BCM6328_REG_INIT, val);
drivers/leds/leds-bcm6328.c
451
rc = bcm6328_hwled(dev, child, reg, mem, lock);
drivers/leds/leds-bcm6328.c
453
rc = bcm6328_led(dev, child, reg, mem, lock,
drivers/leds/leds-bcm6328.c
70
void __iomem *mem;
drivers/leds/leds-bcm6358.c
108
led->mem = mem;
drivers/leds/leds-bcm6358.c
122
val = bcm6358_led_read(led->mem + BCM6358_REG_MODE);
drivers/leds/leds-bcm6358.c
150
void __iomem *mem;
drivers/leds/leds-bcm6358.c
155
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/leds/leds-bcm6358.c
156
if (IS_ERR(mem))
drivers/leds/leds-bcm6358.c
157
return PTR_ERR(mem);
drivers/leds/leds-bcm6358.c
165
val = bcm6358_led_busy(mem);
drivers/leds/leds-bcm6358.c
184
bcm6358_led_write(mem + BCM6358_REG_CTRL, val);
drivers/leds/leds-bcm6358.c
199
rc = bcm6358_led(dev, child, reg, mem, lock);
drivers/leds/leds-bcm6358.c
40
void __iomem *mem;
drivers/leds/leds-bcm6358.c
64
static unsigned long bcm6358_led_busy(void __iomem *mem)
drivers/leds/leds-bcm6358.c
68
while ((val = bcm6358_led_read(mem + BCM6358_REG_CTRL)) &
drivers/leds/leds-bcm6358.c
83
bcm6358_led_busy(led->mem);
drivers/leds/leds-bcm6358.c
84
val = bcm6358_led_read(led->mem + BCM6358_REG_MODE);
drivers/leds/leds-bcm6358.c
90
bcm6358_led_write(led->mem + BCM6358_REG_MODE, val);
drivers/leds/leds-bcm6358.c
95
void __iomem *mem, spinlock_t *lock)
drivers/leds/leds-sun50i-a100.c
395
struct resource *mem;
drivers/leds/leds-sun50i-a100.c
441
priv->base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/leds/leds-sun50i-a100.c
472
dma_cfg.dst_addr = mem->start + LEDC_DATA_REG;
drivers/mcb/mcb-core.c
424
return &dev->mem;
drivers/mcb/mcb-core.c
442
struct resource *mem;
drivers/mcb/mcb-core.c
448
size = resource_size(&dev->mem);
drivers/mcb/mcb-core.c
450
mem = request_mem_region(dev->mem.start, size, name);
drivers/mcb/mcb-core.c
451
if (!mem)
drivers/mcb/mcb-core.c
454
return mem;
drivers/mcb/mcb-core.c
464
void mcb_release_mem(struct resource *mem)
drivers/mcb/mcb-core.c
468
size = resource_size(mem);
drivers/mcb/mcb-core.c
469
release_mem_region(mem->start, size);
drivers/mcb/mcb-lpc.c
18
struct resource *mem;
drivers/mcb/mcb-lpc.c
32
priv->mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/mcb/mcb-lpc.c
33
if (!priv->mem) {
drivers/mcb/mcb-lpc.c
38
res = devm_request_mem_region(&pdev->dev, priv->mem->start,
drivers/mcb/mcb-lpc.c
39
resource_size(priv->mem),
drivers/mcb/mcb-lpc.c
46
priv->base = devm_ioremap(&pdev->dev, priv->mem->start,
drivers/mcb/mcb-lpc.c
47
resource_size(priv->mem));
drivers/mcb/mcb-lpc.c
59
ret = chameleon_parse_cells(priv->bus, priv->mem->start, priv->base);
drivers/mcb/mcb-lpc.c
69
devm_release_mem_region(&pdev->dev, priv->mem->start, resource_size(priv->mem));
drivers/mcb/mcb-lpc.c
72
res = devm_request_mem_region(&pdev->dev, priv->mem->start,
drivers/mcb/mcb-lpc.c
81
priv->base = devm_ioremap(&pdev->dev, priv->mem->start, table_size);
drivers/mcb/mcb-parse.c
92
mdev->mem.start = dev_mapbase + offset;
drivers/mcb/mcb-parse.c
94
mdev->mem.end = mdev->mem.start + size - 1;
drivers/mcb/mcb-parse.c
95
mdev->mem.flags = IORESOURCE_MEM;
drivers/md/dm-bufio.c
1315
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-bufio.c
1316
io_req.mem.ptr.addr = (char *)b->data + offset;
drivers/md/dm-bufio.c
1318
io_req.mem.type = DM_IO_VMA;
drivers/md/dm-bufio.c
1319
io_req.mem.ptr.vma = (char *)b->data + offset;
drivers/md/dm-bufio.c
2210
.mem.type = DM_IO_KMEM,
drivers/md/dm-bufio.c
2211
.mem.ptr.addr = NULL,
drivers/md/dm-bufio.c
2234
.mem.type = DM_IO_KMEM,
drivers/md/dm-bufio.c
2235
.mem.ptr.addr = NULL,
drivers/md/dm-bufio.c
2835
__u64 mem;
drivers/md/dm-bufio.c
2843
mem = (__u64)mult_frac(totalram_pages() - totalhigh_pages(),
drivers/md/dm-bufio.c
2846
if (mem > ULONG_MAX)
drivers/md/dm-bufio.c
2847
mem = ULONG_MAX;
drivers/md/dm-bufio.c
2850
if (mem > mult_frac(VMALLOC_TOTAL, DM_BUFIO_VMALLOC_PERCENT, 100))
drivers/md/dm-bufio.c
2851
mem = mult_frac(VMALLOC_TOTAL, DM_BUFIO_VMALLOC_PERCENT, 100);
drivers/md/dm-bufio.c
2854
dm_bufio_default_cache_size = mem;
drivers/md/dm-integrity.c
1066
io_req.mem.type = DM_IO_PAGE_LIST;
drivers/md/dm-integrity.c
1068
io_req.mem.ptr.pl = &ic->journal_io[pl_index];
drivers/md/dm-integrity.c
1070
io_req.mem.ptr.pl = &ic->journal[pl_index];
drivers/md/dm-integrity.c
1071
io_req.mem.offset = pl_offset;
drivers/md/dm-integrity.c
1190
io_req.mem.type = DM_IO_PAGE_LIST;
drivers/md/dm-integrity.c
1191
io_req.mem.ptr.pl = &ic->journal[pl_index];
drivers/md/dm-integrity.c
1192
io_req.mem.offset = pl_offset;
drivers/md/dm-integrity.c
1517
fr.io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-integrity.c
1518
fr.io_req.mem.ptr.addr = NULL;
drivers/md/dm-integrity.c
1820
char *mem;
drivers/md/dm-integrity.c
1827
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-integrity.c
1828
io_req.mem.ptr.addr = buffer;
drivers/md/dm-integrity.c
1865
mem = bvec_kmap_local(&bv);
drivers/md/dm-integrity.c
1866
memcpy(mem + pos, buffer + buffer_offset, ic->sectors_per_block << SECTOR_SHIFT);
drivers/md/dm-integrity.c
1867
kunmap_local(mem);
drivers/md/dm-integrity.c
1946
char *mem, *checksums_ptr;
drivers/md/dm-integrity.c
1949
mem = integrity_kmap(ic, bv_copy.bv_page);
drivers/md/dm-integrity.c
1953
integrity_sector_checksum(ic, &dio->ahash_req, sector, mem, bv_copy.bv_offset + pos, checksums_ptr);
drivers/md/dm-integrity.c
1959
integrity_kunmap(ic, mem);
drivers/md/dm-integrity.c
2147
char *mem;
drivers/md/dm-integrity.c
2154
mem = kmap_local_page(bv.bv_page);
drivers/md/dm-integrity.c
2168
kunmap_local(mem);
drivers/md/dm-integrity.c
2176
mem_ptr = mem + bv.bv_offset;
drivers/md/dm-integrity.c
2216
memcpy(js, mem + bv.bv_offset, ic->sectors_per_block << SECTOR_SHIFT);
drivers/md/dm-integrity.c
2253
kunmap_local(mem);
drivers/md/dm-integrity.c
2607
const char *mem = integrity_kmap(ic, bv.bv_page);
drivers/md/dm-integrity.c
2610
integrity_sector_checksum(ic, &dio->ahash_req, dio->bio_details.bi_iter.bi_sector, mem, bv.bv_offset, dio->integrity_payload + pos);
drivers/md/dm-integrity.c
2611
integrity_kunmap(ic, mem);
drivers/md/dm-integrity.c
2657
char *mem;
drivers/md/dm-integrity.c
2704
mem = bvec_kmap_local(&bv);
drivers/md/dm-integrity.c
2705
memcpy(mem, outgoing_data, ic->sectors_per_block << SECTOR_SHIFT);
drivers/md/dm-integrity.c
2706
kunmap_local(mem);
drivers/md/dm-integrity.c
2722
char *mem = integrity_kmap(ic, bv.bv_page);
drivers/md/dm-integrity.c
2723
integrity_sector_checksum(ic, &dio->ahash_req, dio->bio_details.bi_iter.bi_sector, mem, bv.bv_offset, digest);
drivers/md/dm-integrity.c
2726
integrity_kunmap(ic, mem);
drivers/md/dm-integrity.c
2732
integrity_kunmap(ic, mem);
drivers/md/dm-integrity.c
3214
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-integrity.c
3215
io_req.mem.ptr.addr = recalc_buffer;
drivers/md/dm-integrity.c
548
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-integrity.c
549
io_req.mem.ptr.addr = ic->sb;
drivers/md/dm-io.c
472
switch (io_req->mem.type) {
drivers/md/dm-io.c
474
list_dp_init(dp, io_req->mem.ptr.pl, io_req->mem.offset);
drivers/md/dm-io.c
478
bio_dp_init(dp, io_req->mem.ptr.bio);
drivers/md/dm-io.c
482
flush_kernel_vmap_range(io_req->mem.ptr.vma, size);
drivers/md/dm-io.c
484
dp->vma_invalidate_address = io_req->mem.ptr.vma;
drivers/md/dm-io.c
487
vm_dp_init(dp, io_req->mem.ptr.vma);
drivers/md/dm-io.c
491
km_dp_init(dp, io_req->mem.ptr.addr);
drivers/md/dm-kcopyd.c
560
.mem.type = DM_IO_PAGE_LIST,
drivers/md/dm-kcopyd.c
561
.mem.ptr.pl = job->pages,
drivers/md/dm-kcopyd.c
562
.mem.offset = 0,
drivers/md/dm-log.c
461
lc->io_req.mem.type = DM_IO_VMA;
drivers/md/dm-log.c
479
lc->io_req.mem.ptr.vma = lc->disk_header;
drivers/md/dm-raid1.c
268
.mem.type = DM_IO_KMEM,
drivers/md/dm-raid1.c
269
.mem.ptr.addr = NULL,
drivers/md/dm-raid1.c
547
.mem.type = DM_IO_BIO,
drivers/md/dm-raid1.c
548
.mem.ptr.bio = bio,
drivers/md/dm-raid1.c
661
.mem.type = DM_IO_BIO,
drivers/md/dm-raid1.c
662
.mem.ptr.bio = bio,
drivers/md/dm-raid1.c
670
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-raid1.c
671
io_req.mem.ptr.addr = NULL;
drivers/md/dm-snap-persistent.c
242
.mem.type = DM_IO_VMA,
drivers/md/dm-snap-persistent.c
243
.mem.ptr.vma = area,
drivers/md/dm-snap.c
679
struct kmem_cache *mem)
drivers/md/dm-snap.c
692
kmem_cache_free(mem, ex);
drivers/md/dm-snap.c
865
unsigned long mem = 2 * 1024 * 1024;
drivers/md/dm-snap.c
867
mem /= sizeof(struct dm_hlist_head);
drivers/md/dm-snap.c
869
return mem;
drivers/md/dm-vdo/dedupe.c
2197
.memory_size = geometry.index_config.mem,
drivers/md/dm-vdo/encodings.c
252
u32 unused, mem;
drivers/md/dm-vdo/encodings.c
285
decode_u32_le(buffer, offset, &mem);
drivers/md/dm-vdo/encodings.c
290
.mem = mem,
drivers/md/dm-vdo/encodings.h
71
u32 mem;
drivers/md/dm-vdo/message-stats.c
435
static void write_index_memory(u32 mem, char **buf, unsigned int *maxlen)
drivers/md/dm-vdo/message-stats.c
440
if (mem == (u32)UDS_MEMORY_CONFIG_256MB)
drivers/md/dm-vdo/message-stats.c
442
else if (mem == (u32)UDS_MEMORY_CONFIG_512MB)
drivers/md/dm-vdo/message-stats.c
444
else if (mem == (u32)UDS_MEMORY_CONFIG_768MB)
drivers/md/dm-vdo/message-stats.c
447
write_u32(prefix, mem, ", ", buf, maxlen);
drivers/md/dm-vdo/message-stats.c
455
write_index_memory(config->mem, buf, maxlen);
drivers/md/dm-verity-target.c
389
io_req.mem.type = DM_IO_KMEM;
drivers/md/dm-verity-target.c
390
io_req.mem.ptr.addr = buffer;
drivers/md/dm-writecache.c
526
req.mem.type = DM_IO_VMA;
drivers/md/dm-writecache.c
527
req.mem.ptr.vma = (char *)wc->memory_map + (size_t)i * BITMAP_GRANULARITY;
drivers/md/dm-writecache.c
564
req.mem.type = DM_IO_VMA;
drivers/md/dm-writecache.c
565
req.mem.ptr.vma = (char *)wc->memory_map;
drivers/md/dm-writecache.c
593
req.mem.type = DM_IO_KMEM;
drivers/md/dm-writecache.c
594
req.mem.ptr.addr = NULL;
drivers/md/dm-writecache.c
987
req.mem.type = DM_IO_VMA;
drivers/md/dm-writecache.c
988
req.mem.ptr.vma = (char *)wc->memory_map;
drivers/media/common/saa7146/saa7146_core.c
169
void *mem = vmalloc_32(length);
drivers/media/common/saa7146/saa7146_core.c
172
if (NULL == mem)
drivers/media/common/saa7146/saa7146_core.c
175
if (!(pt->slist = vmalloc_to_sg(mem, pages)))
drivers/media/common/saa7146/saa7146_core.c
189
return mem;
drivers/media/common/saa7146/saa7146_core.c
199
vfree(mem);
drivers/media/common/saa7146/saa7146_core.c
204
void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt)
drivers/media/common/saa7146/saa7146_core.c
210
vfree(mem);
drivers/media/common/saa7146/saa7146_core.c
368
dev->mem = ioremap(pci_resource_start(pci, 0),
drivers/media/common/saa7146/saa7146_core.c
370
if (!dev->mem) {
drivers/media/common/saa7146/saa7146_core.c
423
dev->mem, dev->revision, pci->irq,
drivers/media/common/saa7146/saa7146_core.c
475
iounmap(dev->mem);
drivers/media/common/saa7146/saa7146_core.c
515
iounmap(dev->mem);
drivers/media/common/saa7146/saa7146_i2c.c
32
int mem = 0, op_count = 0;
drivers/media/common/saa7146/saa7146_i2c.c
36
mem += m[i].len + 1;
drivers/media/common/saa7146/saa7146_i2c.c
41
mem = 1 + ((mem-1) / 3);
drivers/media/common/saa7146/saa7146_i2c.c
46
if ((4 * mem) > SAA7146_I2C_MEM) {
drivers/media/common/saa7146/saa7146_i2c.c
52
memset(op,0,sizeof(__le32)*mem);
drivers/media/common/saa7146/saa7146_i2c.c
83
return mem;
drivers/media/dvb-core/dmxdev.c
160
void *mem;
drivers/media/dvb-core/dmxdev.c
166
mem = vmalloc(DVR_BUFFER_SIZE);
drivers/media/dvb-core/dmxdev.c
167
if (!mem) {
drivers/media/dvb-core/dmxdev.c
171
dmxdev->dvr_buffer.data = mem;
drivers/media/dvb-core/dmxdev.c
225
void *mem = dmxdev->dvr_buffer.data;
drivers/media/dvb-core/dmxdev.c
231
vfree(mem);
drivers/media/dvb-core/dmxdev.c
671
void *mem;
drivers/media/dvb-core/dmxdev.c
681
mem = vmalloc(filter->buffer.size);
drivers/media/dvb-core/dmxdev.c
682
if (!mem)
drivers/media/dvb-core/dmxdev.c
685
filter->buffer.data = mem;
drivers/media/dvb-core/dmxdev.c
860
void *mem = dmxdevfilter->buffer.data;
drivers/media/dvb-core/dmxdev.c
865
vfree(mem);
drivers/media/dvb-frontends/sp2.c
125
ret = ci_op_cam(s->priv, read, addr, data, &mem);
drivers/media/dvb-frontends/sp2.c
137
(read) ? mem : data);
drivers/media/dvb-frontends/sp2.c
140
return mem;
drivers/media/dvb-frontends/sp2.c
97
int mem, ret;
drivers/media/dvb-frontends/sp887x.c
143
const unsigned char *mem = fw->data + 10;
drivers/media/dvb-frontends/sp887x.c
178
memcpy(&buf[2], mem + i, c);
drivers/media/dvb-frontends/tda1004x.c
307
const unsigned char *mem, unsigned int len,
drivers/media/dvb-frontends/tda1004x.c
328
memcpy(buf + 1, mem + pos, tx_size);
drivers/media/pci/cx18/cx18-firmware.c
87
static int load_cpu_fw_direct(const char *fn, u8 __iomem *mem, struct cx18 *cx)
drivers/media/pci/cx18/cx18-firmware.c
92
u32 __iomem *dst = (u32 __iomem *)mem;
drivers/media/pci/cx23885/altera-ci.c
265
int mem = 0;
drivers/media/pci/cx23885/altera-ci.c
280
mem = netup_fpga_op_rw(inter, NETUP_CI_DATA, val, read);
drivers/media/pci/cx23885/altera-ci.c
287
(read) ? mem : val);
drivers/media/pci/cx23885/altera-ci.c
289
return mem;
drivers/media/pci/cx23885/cimax2.c
150
int mem;
drivers/media/pci/cx23885/cimax2.c
154
mem = cx_read(MC417_RWD);
drivers/media/pci/cx23885/cimax2.c
155
if ((mem & NETUP_ACK) == 0)
drivers/media/pci/cx23885/cimax2.c
164
return mem & 0xff;
drivers/media/pci/cx23885/cimax2.c
175
int mem;
drivers/media/pci/cx23885/cimax2.c
218
mem = netup_ci_get_mem(dev);
drivers/media/pci/cx23885/cimax2.c
223
if (mem < 0)
drivers/media/pci/cx23885/cimax2.c
229
(read) ? mem : data);
drivers/media/pci/cx23885/cimax2.c
232
return mem;
drivers/media/pci/cx23885/cx23885-dvb.c
662
u8 data, int *mem)
drivers/media/pci/cx23885/cx23885-dvb.c
716
*mem = tmp & 0xff;
drivers/media/pci/cx23885/cx23885-dvb.c
721
if (*mem < 0) {
drivers/media/pci/cx23885/cx23885-dvb.c
936
uint32_t mem = 0;
drivers/media/pci/cx23885/cx23885-dvb.c
938
mem = cx_read(MC417_RWD);
drivers/media/pci/cx23885/cx23885-dvb.c
943
mem &= ~ALT_DATA;
drivers/media/pci/cx23885/cx23885-dvb.c
944
mem |= (data & ALT_DATA);
drivers/media/pci/cx23885/cx23885-dvb.c
948
mem |= ALT_AD_RG;
drivers/media/pci/cx23885/cx23885-dvb.c
950
mem &= ~ALT_AD_RG;
drivers/media/pci/cx23885/cx23885-dvb.c
952
mem &= ~ALT_CS;
drivers/media/pci/cx23885/cx23885-dvb.c
954
mem = (mem & ~ALT_RD) | ALT_WR;
drivers/media/pci/cx23885/cx23885-dvb.c
956
mem = (mem & ~ALT_WR) | ALT_RD;
drivers/media/pci/cx23885/cx23885-dvb.c
958
cx_write(MC417_RWD, mem); /* start RW cycle */
drivers/media/pci/cx23885/cx23885-dvb.c
961
mem = cx_read(MC417_RWD);
drivers/media/pci/cx23885/cx23885-dvb.c
962
if ((mem & ALT_RDY) == 0)
drivers/media/pci/cx23885/cx23885-dvb.c
971
return mem & ALT_DATA;
drivers/media/pci/ddbridge/ddbridge-core.c
128
u64 mem;
drivers/media/pci/ddbridge/ddbridge-core.c
133
mem = dma->pbuf[i];
drivers/media/pci/ddbridge/ddbridge-core.c
134
ddbwritel(dev, mem & 0xffffffff, dma->bufregs + i * 8);
drivers/media/pci/ddbridge/ddbridge-core.c
135
ddbwritel(dev, mem >> 32, dma->bufregs + i * 8 + 4);
drivers/media/pci/ddbridge/ddbridge-core.c
165
u64 mem;
drivers/media/pci/ddbridge/ddbridge-core.c
170
mem = ddma->pbuf[i];
drivers/media/pci/ddbridge/ddbridge-core.c
171
ddbwritel(dev, mem & 0xffffffff, base + i * 8);
drivers/media/pci/ddbridge/ddbridge-core.c
172
ddbwritel(dev, mem >> 32, base + i * 8 + 4);
drivers/media/pci/ivtv/ivtv-firmware.c
166
static volatile struct ivtv_mailbox __iomem *ivtv_search_mailbox(const volatile u8 __iomem *mem, u32 size)
drivers/media/pci/ivtv/ivtv-firmware.c
173
if (readl(mem + i) == 0x12345678 &&
drivers/media/pci/ivtv/ivtv-firmware.c
174
readl(mem + i + 4) == 0x34567812 &&
drivers/media/pci/ivtv/ivtv-firmware.c
175
readl(mem + i + 8) == 0x56781234 &&
drivers/media/pci/ivtv/ivtv-firmware.c
176
readl(mem + i + 12) == 0x78123456) {
drivers/media/pci/ivtv/ivtv-firmware.c
177
return (volatile struct ivtv_mailbox __iomem *)(mem + i + 16);
drivers/media/pci/ivtv/ivtv-firmware.c
40
static int load_fw_direct(const char *fn, volatile u8 __iomem *mem, struct ivtv *itv, long size)
drivers/media/pci/ivtv/ivtv-firmware.c
48
volatile u32 __iomem *dst = (volatile u32 __iomem *)mem;
drivers/media/pci/smipcie/smipcie-main.c
778
u64 mem;
drivers/media/pci/smipcie/smipcie-main.c
784
mem = port->dma_addr[0];
drivers/media/pci/smipcie/smipcie-main.c
785
dmaMemPtrLow = mem & 0xffffffff;
drivers/media/pci/smipcie/smipcie-main.c
786
dmaMemPtrHi = mem >> 32;
drivers/media/pci/smipcie/smipcie-main.c
799
mem = port->dma_addr[1];
drivers/media/pci/smipcie/smipcie-main.c
800
dmaMemPtrLow = mem & 0xffffffff;
drivers/media/pci/smipcie/smipcie-main.c
801
dmaMemPtrHi = mem >> 32;
drivers/media/pci/ttpci/budget-core.c
178
u8 *mem = (u8 *) (budget->grabbing);
drivers/media/pci/ttpci/budget-core.c
200
dvb_dmx_swfilter_packets(&budget->demux, mem + olddma, count / 188);
drivers/media/pci/ttpci/budget-core.c
203
dvb_dmx_swfilter_packets(&budget->demux, mem + olddma, count / 188);
drivers/media/pci/ttpci/budget-core.c
205
dvb_dmx_swfilter_packets(&budget->demux, mem, newdma / 188);
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
128
#define _CFG_ADDR(plat, cfg, mem) (&((cfg)->config_##plat.mem))
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
129
#define CFG_ADDR(plat, cfg, mem) \
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
130
(IS_ERR_OR_NULL(cfg) ? NULL : _CFG_ADDR(plat, cfg, mem))
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
132
#define _CFG_GET(plat, cfg, mem) ((cfg)->config_##plat.mem)
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
133
#define CFG_GET(plat, cfg, mem) \
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
134
(IS_ERR_OR_NULL(cfg) ? 0 : _CFG_GET(plat, cfg, mem))
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
136
#define _CFG_COMP(plat, comp, mem) ((comp)->comp_##plat.mem)
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
137
#define CFG_COMP(plat, comp, mem) \
drivers/media/platform/mediatek/mdp3/mtk-img-ipi.h
138
(IS_ERR_OR_NULL(comp) ? 0 : _CFG_COMP(plat, comp, mem))
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
101
if (mem->size)
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
102
mtk_v4l2_err(plat_dev, "Failed to free %zu bytes", mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
106
mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va,
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
107
(unsigned long)mem->dma_addr, mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
109
dma_free_coherent(&plat_dev->dev, mem->size, mem->va, mem->dma_addr);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
110
mem->va = NULL;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
111
mem->dma_addr = 0;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
112
mem->size = 0;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
48
int mtk_vcodec_mem_alloc(void *priv, struct mtk_vcodec_mem *mem)
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
66
mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr,
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
68
if (!mem->va) {
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
70
__func__, mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
74
mtk_v4l2_debug(plat_dev, 3, "[%d] - va = %p dma = 0x%lx size = 0x%zx", id, mem->va,
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
75
(unsigned long)mem->dma_addr, mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
81
void mtk_vcodec_mem_free(void *priv, struct mtk_vcodec_mem *mem)
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
99
if (!mem->va) {
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.h
67
int mtk_vcodec_mem_alloc(void *priv, struct mtk_vcodec_mem *mem);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.h
68
void mtk_vcodec_mem_free(void *priv, struct mtk_vcodec_mem *mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
156
struct mtk_vcodec_mem *mem = NULL;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
159
mem = &inst->pred_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
160
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
161
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
168
struct mtk_vcodec_mem *mem = NULL;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
172
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
173
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
174
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
175
mem->size = buf_sz;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
176
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
181
inst->vsi->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
190
struct mtk_vcodec_mem *mem = NULL;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
194
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
195
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
196
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
174
struct mtk_vcodec_mem *mem = &inst->pred_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
177
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
178
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
186
struct mtk_vcodec_mem *mem = NULL;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
191
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
192
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
193
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
194
mem->size = buf_sz;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
195
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
200
inst->vsi_ctx.mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
209
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
213
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
214
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
215
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1015
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1064
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1065
inst->vsi_ctx_ext.mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1111
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1168
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1169
inst->vsi_ctx.mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
405
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
410
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
411
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
412
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
413
mem->size = buf_sz;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
414
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
427
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
430
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
431
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
432
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
491
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
498
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
499
inst->vsi_ext->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
500
inst->vsi_ext->mv_buf_dma[i].size = mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
522
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
567
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
568
inst->vsi_core_ext->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
569
inst->vsi_core_ext->mv_buf_dma[i].size = mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
635
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
672
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
673
inst->vsi_core->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
871
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
930
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
931
inst->vsi->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
648
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
653
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
654
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
655
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
656
mem->size = buf_sz;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
657
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
670
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
673
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
674
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
675
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
735
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
761
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
762
inst->vsi->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
763
inst->vsi->mv_buf_dma[i].size = mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
800
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
845
mem = &inst->mv_buf[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
846
inst->vsi_core->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
847
inst->vsi_core->mv_buf_dma[i].size = mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
911
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
916
mem = &inst->wrap_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
917
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
918
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
365
struct mtk_vcodec_mem *mem = &inst->working_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
367
mem->size = VP8_WORKING_BUF_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
368
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
374
inst->vsi->dec.working_buf_dma = (uint64_t)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
380
struct mtk_vcodec_mem *mem = &inst->working_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
382
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
383
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
148
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
150
mem = &inst->seg_id_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
151
mem->size = VP8_SEG_ID_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
152
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
157
inst->vsi->dec.seg_id_buf_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
159
mem = &inst->wrap_y_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
160
mem->size = VP8_PP_WRAPY_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
161
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
166
inst->vsi->dec.wrap_y_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
168
mem = &inst->wrap_c_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
169
mem->size = VP8_PP_WRAPC_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
170
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
175
inst->vsi->dec.wrap_c_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
177
mem = &inst->vld_wrapper_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
178
mem->size = VP8_VLD_PRED_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
179
err = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
184
inst->vsi->dec.vld_wrapper_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
191
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
193
mem = &inst->seg_id_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
194
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
195
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
198
mem = &inst->wrap_y_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
199
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
200
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
203
mem = &inst->wrap_c_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
204
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
205
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
208
mem = &inst->vld_wrapper_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
209
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
210
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
204
struct mtk_vcodec_mem mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
365
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
390
mem = &inst->mv_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
391
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
392
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
394
mem->size = ((vsi->buf_w / 64) *
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
396
result = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
398
mem->size = 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
403
vsi->mv_buf.va = (unsigned long)mem->va;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
404
vsi->mv_buf.pa = (unsigned long)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
405
vsi->mv_buf.sz = (unsigned int)mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
408
mem = &inst->seg_id_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
409
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
410
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
412
mem->size = VP9_SEG_ID_SZ;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
413
result = mtk_vcodec_mem_alloc(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
415
mem->size = 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
420
vsi->seg_id_buf.va = (unsigned long)mem->va;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
421
vsi->seg_id_buf.pa = (unsigned long)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
422
vsi->seg_id_buf.sz = (unsigned int)mem->size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
551
struct mtk_vcodec_mem mem = { };
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
554
mem.size = sizeof(struct vdec_vp9_inst);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
555
result = mtk_vcodec_mem_alloc(ctx, &mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
559
inst = mem.va;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
560
inst->mem = mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
567
struct mtk_vcodec_mem mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
569
mem = inst->mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
570
if (mem.va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
571
mtk_vcodec_mem_free(inst->ctx, &mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
756
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
763
mem = &inst->mv_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
764
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
765
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
767
mem = &inst->seg_id_buf;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
768
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
769
mtk_vcodec_mem_free(inst->ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
205
struct mtk_vcodec_mem *mem;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
208
mem = &msg_queue->wdma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
209
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
210
mtk_vcodec_mem_free(ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
214
mem = &lat_buf->wdma_err_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
215
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
216
mtk_vcodec_mem_free(ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
218
mem = &lat_buf->slice_bc_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
219
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
220
mtk_vcodec_mem_free(ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
222
mem = &lat_buf->rd_mv_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
223
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
224
mtk_vcodec_mem_free(ctx, mem);
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
226
mem = &lat_buf->tile_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
227
if (mem->va)
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
228
mtk_vcodec_mem_free(ctx, mem);
drivers/media/platform/qcom/venus/hfi_cmds.c
84
res->mem = addr;
drivers/media/platform/qcom/venus/hfi_helper.h
1076
u32 mem;
drivers/media/platform/qcom/venus/hfi_venus.c
360
static void venus_free(struct venus_hfi_device *hdev, struct mem_desc *mem)
drivers/media/platform/qcom/venus/hfi_venus.c
364
dma_free_attrs(dev, mem->size, mem->kva, mem->da, mem->attrs);
drivers/media/platform/renesas/vsp1/vsp1_dl.c
135
void *mem;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
159
void *mem;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
283
pool->mem = dma_alloc_wc(vsp1->bus_master, pool->size, &pool->dma,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
285
if (!pool->mem) {
drivers/media/platform/renesas/vsp1/vsp1_dl.c
301
dlb->entries = pool->mem + i * dlb_size;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
320
if (pool->mem)
drivers/media/platform/renesas/vsp1/vsp1_dl.c
321
dma_free_wc(pool->vsp1->bus_master, pool->size, pool->mem,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
457
pool->mem = dma_alloc_wc(vsp1->bus_master, pool->size, &pool->dma,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
459
if (!pool->mem) {
drivers/media/platform/renesas/vsp1/vsp1_dl.c
480
cmd->cmds = pool->mem + cmd_offset;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
483
cmd->data = pool->mem + data_offset;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
531
if (pool->mem)
drivers/media/platform/renesas/vsp1/vsp1_dl.c
532
dma_free_wc(pool->vsp1->bus_master, pool->size, pool->mem,
drivers/media/platform/renesas/vsp1/vsp1_drm.c
859
&cfg->pixelformat, cfg->pitch, &cfg->mem[0], &cfg->mem[1],
drivers/media/platform/renesas/vsp1/vsp1_drm.c
860
&cfg->mem[2], cfg->zpos);
drivers/media/platform/renesas/vsp1/vsp1_drm.c
873
rpf->mem.addr[0] = cfg->mem[0];
drivers/media/platform/renesas/vsp1/vsp1_drm.c
874
rpf->mem.addr[1] = cfg->mem[1];
drivers/media/platform/renesas/vsp1/vsp1_drm.c
875
rpf->mem.addr[2] = cfg->mem[2];
drivers/media/platform/renesas/vsp1/vsp1_drm.c
918
pipe->output->mem.addr[0] = wb_cfg->mem[0];
drivers/media/platform/renesas/vsp1/vsp1_drm.c
919
pipe->output->mem.addr[1] = wb_cfg->mem[1];
drivers/media/platform/renesas/vsp1/vsp1_drm.c
920
pipe->output->mem.addr[2] = wb_cfg->mem[2];
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
283
auto_fld->top_y0 = rpf->mem.addr[0];
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
284
auto_fld->top_c0 = rpf->mem.addr[1];
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
285
auto_fld->top_c1 = rpf->mem.addr[2];
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
290
auto_fld->bottom_y0 = rpf->mem.addr[0] + offset_y;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
291
auto_fld->bottom_c0 = rpf->mem.addr[1] + offset_c;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
292
auto_fld->bottom_c1 = rpf->mem.addr[2] + offset_c;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
319
struct vsp1_rwpf_memory mem = rpf->mem;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
346
mem.addr[0] += crop.top * format->plane_fmt[0].bytesperline
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
355
mem.addr[1] += offset;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
356
mem.addr[2] += offset;
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
365
swap(mem.addr[1], mem.addr[2]);
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
374
vsp1_rpf_write(rpf, dlb, VI6_RPF_SRCM_ADDR_Y, mem.addr[0]);
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
375
vsp1_rpf_write(rpf, dlb, VI6_RPF_SRCM_ADDR_C0, mem.addr[1]);
drivers/media/platform/renesas/vsp1/vsp1_rpf.c
376
vsp1_rpf_write(rpf, dlb, VI6_RPF_SRCM_ADDR_C1, mem.addr[2]);
drivers/media/platform/renesas/vsp1/vsp1_rwpf.h
63
struct vsp1_rwpf_memory mem;
drivers/media/platform/renesas/vsp1/vsp1_video.c
253
video->rwpf->mem = buf->mem;
drivers/media/platform/renesas/vsp1/vsp1_video.c
646
buf->mem.addr[i] = vb2_dma_contig_plane_dma_addr(vb, i);
drivers/media/platform/renesas/vsp1/vsp1_video.c
653
buf->mem.addr[i] = 0;
drivers/media/platform/renesas/vsp1/vsp1_video.c
677
video->rwpf->mem = buf->mem;
drivers/media/platform/renesas/vsp1/vsp1_video.h
22
struct vsp1_rwpf_memory mem;
drivers/media/platform/renesas/vsp1/vsp1_vspx.c
153
rpf0->mem.addr[0] = addr;
drivers/media/platform/renesas/vsp1/vsp1_vspx.c
460
ret = vsp1_vspx_pipeline_configure(vsp1, job->config.mem,
drivers/media/platform/renesas/vsp1/vsp1_vspx.c
481
ret = vsp1_vspx_pipeline_configure(vsp1, job->img.mem,
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
402
struct vsp1_rwpf_memory mem = wpf->mem;
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
464
mem.addr[i] += offset / vsub
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
467
mem.addr[i] += offset / hsub
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
483
mem.addr[0] += (height - 1)
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
489
mem.addr[1] += offset;
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
490
mem.addr[2] += offset;
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
504
mem.addr[i] += hoffset / hsub
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
515
swap(mem.addr[1], mem.addr[2]);
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
517
vsp1_wpf_write(wpf, dlb, VI6_WPF_DSTM_ADDR_Y, mem.addr[0]);
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
518
vsp1_wpf_write(wpf, dlb, VI6_WPF_DSTM_ADDR_C0, mem.addr[1]);
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
519
vsp1_wpf_write(wpf, dlb, VI6_WPF_DSTM_ADDR_C1, mem.addr[2]);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
43
module_param_named(mem, mfc_mem_size, charp, 0644);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
44
MODULE_PARM_DESC(mem, "Preallocated memory size for the firmware and context buffers");
drivers/media/platform/ti/omap3isp/isp.c
2230
struct resource *mem;
drivers/media/platform/ti/omap3isp/isp.c
2294
devm_platform_get_and_ioremap_resource(pdev, i, &mem);
drivers/media/platform/ti/omap3isp/isp.c
2350
mem->start + isp_res_maps[m].offset[OMAP3_ISP_IOMEM_HIST];
drivers/media/usb/au0828/au0828-vbi.c
62
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/au0828/au0828-video.c
406
(*buf)->vb_buf = (*buf)->mem;
drivers/media/usb/au0828/au0828-video.c
481
(*buf)->vb_buf = (*buf)->mem;
drivers/media/usb/au0828/au0828-video.c
723
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/au0828/au0828.h
160
void *mem;
drivers/media/usb/dvb-usb-v2/dvbsky.c
346
u8 data, int *mem)
drivers/media/usb/dvb-usb-v2/dvbsky.c
369
*mem = respond[1];
drivers/media/usb/em28xx/em28xx-vbi.c
73
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/em28xx/em28xx-video.c
1218
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/em28xx/em28xx-video.c
602
buf->vb_buf = buf->mem;
drivers/media/usb/em28xx/em28xx.h
280
void *mem;
drivers/media/usb/go7007/go7007-fw.c
368
u16 mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
390
code[off + 1] = __cpu_to_le16(0x8000 | mem);
drivers/media/usb/go7007/go7007-fw.c
393
if (mem + chunk > 0x4000)
drivers/media/usb/go7007/go7007-fw.c
394
chunk = 0x4000 - mem;
drivers/media/usb/go7007/go7007-fw.c
401
mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
405
mem += 28;
drivers/media/usb/go7007/go7007-fw.c
635
u16 mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
679
code[off + 1] = __cpu_to_le16(0x8000 | mem);
drivers/media/usb/go7007/go7007-fw.c
682
if (mem + chunk > 0x4000)
drivers/media/usb/go7007/go7007-fw.c
683
chunk = 0x4000 - mem;
drivers/media/usb/go7007/go7007-fw.c
690
if (mem + chunk == 0x4000) {
drivers/media/usb/go7007/go7007-fw.c
691
mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
697
mem += 28;
drivers/media/usb/go7007/go7007-fw.c
821
u16 mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
847
code[off + 1] = __cpu_to_le16(0x8000 | mem);
drivers/media/usb/go7007/go7007-fw.c
850
if (mem + chunk > 0x4000)
drivers/media/usb/go7007/go7007-fw.c
851
chunk = 0x4000 - mem;
drivers/media/usb/go7007/go7007-fw.c
858
if (mem + chunk == 0x4000) {
drivers/media/usb/go7007/go7007-fw.c
859
mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
865
mem += 28;
drivers/media/usb/go7007/go7007-fw.c
871
mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
890
code[off + 1] = __cpu_to_le16(0x8000 | mem);
drivers/media/usb/go7007/go7007-fw.c
893
if (mem + chunk > 0x4000)
drivers/media/usb/go7007/go7007-fw.c
894
chunk = 0x4000 - mem;
drivers/media/usb/go7007/go7007-fw.c
901
if (mem + chunk == 0x4000) {
drivers/media/usb/go7007/go7007-fw.c
902
mem = 0x3e00;
drivers/media/usb/go7007/go7007-fw.c
908
mem += 28;
drivers/media/usb/hdpvr/hdpvr-video.c
141
u8 *mem;
drivers/media/usb/hdpvr/hdpvr-video.c
162
mem = usb_alloc_coherent(dev->udev, dev->bulk_in_size, GFP_KERNEL,
drivers/media/usb/hdpvr/hdpvr-video.c
164
if (!mem) {
drivers/media/usb/hdpvr/hdpvr-video.c
173
mem, dev->bulk_in_size,
drivers/media/usb/stk1160/stk1160-v4l.c
697
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/stk1160/stk1160-video.c
105
u8 *dst = buf->mem;
drivers/media/usb/stk1160/stk1160-video.c
139
offset = dst - (u8 *)buf->mem;
drivers/media/usb/stk1160/stk1160-video.c
160
(unsigned long)buf->mem + buf->length) {
drivers/media/usb/stk1160/stk1160-video.c
184
offset = dst - (u8 *)buf->mem;
drivers/media/usb/stk1160/stk1160-video.c
204
(unsigned long)buf->mem + buf->length) {
drivers/media/usb/stk1160/stk1160.h
75
void *mem;
drivers/media/usb/uvc/uvc_isight.c
46
u8 *mem;
drivers/media/usb/uvc/uvc_isight.c
86
mem = buf->mem + buf->bytesused;
drivers/media/usb/uvc/uvc_isight.c
88
memcpy(mem, data, nbytes);
drivers/media/usb/uvc/uvc_queue.c
126
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/media/usb/uvc/uvc_video.c
1330
op->dst = buf->mem + buf->bytesused;
drivers/media/usb/uvc/uvc_video.c
1386
void *mem;
drivers/media/usb/uvc/uvc_video.c
1389
mem = buf->mem + queue->buf_used;
drivers/media/usb/uvc/uvc_video.c
1393
memcpy(data, mem, nbytes);
drivers/media/usb/uvc/uvc_video.c
1417
const u8 *mem, unsigned int length)
drivers/media/usb/uvc/uvc_video.c
1430
has_pts = mem[1] & UVC_STREAM_PTS;
drivers/media/usb/uvc/uvc_video.c
1431
has_scr = mem[1] & UVC_STREAM_SCR;
drivers/media/usb/uvc/uvc_video.c
1435
scr = mem + 6;
drivers/media/usb/uvc/uvc_video.c
1437
scr = mem + 2;
drivers/media/usb/uvc/uvc_video.c
1456
meta = (struct uvc_meta_buf *)((u8 *)meta_buf->mem + meta_buf->bytesused);
drivers/media/usb/uvc/uvc_video.c
1467
meta->length = mem[0];
drivers/media/usb/uvc/uvc_video.c
1468
meta->flags = mem[1];
drivers/media/usb/uvc/uvc_video.c
1469
memcpy(meta->buf, &mem[2], length - 2);
drivers/media/usb/uvc/uvc_video.c
1527
u8 *mem;
drivers/media/usb/uvc/uvc_video.c
1542
mem = urb->transfer_buffer + urb->iso_frame_desc[i].offset;
drivers/media/usb/uvc/uvc_video.c
1544
ret = uvc_video_decode_start(stream, buf, mem,
drivers/media/usb/uvc/uvc_video.c
1553
uvc_video_decode_meta(stream, meta_buf, mem, ret);
drivers/media/usb/uvc/uvc_video.c
1556
uvc_video_decode_data(uvc_urb, buf, mem + ret,
drivers/media/usb/uvc/uvc_video.c
1560
uvc_video_decode_end(stream, buf, mem,
drivers/media/usb/uvc/uvc_video.c
1573
u8 *mem;
drivers/media/usb/uvc/uvc_video.c
1583
mem = urb->transfer_buffer;
drivers/media/usb/uvc/uvc_video.c
1593
ret = uvc_video_decode_start(stream, buf, mem, len);
drivers/media/usb/uvc/uvc_video.c
1602
memcpy(stream->bulk.header, mem, ret);
drivers/media/usb/uvc/uvc_video.c
1605
uvc_video_decode_meta(stream, meta_buf, mem, ret);
drivers/media/usb/uvc/uvc_video.c
1607
mem += ret;
drivers/media/usb/uvc/uvc_video.c
1620
uvc_video_decode_data(uvc_urb, buf, mem, len);
drivers/media/usb/uvc/uvc_video.c
1647
u8 *mem = urb->transfer_buffer;
drivers/media/usb/uvc/uvc_video.c
1657
ret = uvc_video_encode_header(stream, buf, mem, len);
drivers/media/usb/uvc/uvc_video.c
1660
mem += ret;
drivers/media/usb/uvc/uvc_video.c
1665
ret = uvc_video_encode_data(stream, buf, mem, len);
drivers/media/usb/uvc/uvcvideo.h
321
void *mem;
drivers/memory/brcmstb_dpfe.c
595
static int __write_firmware(u32 __iomem *mem, const u32 *fw,
drivers/memory/brcmstb_dpfe.c
605
writel_relaxed(0, mem + i);
drivers/memory/brcmstb_dpfe.c
610
writel_relaxed(be32_to_cpu(fw[i]), mem + i);
drivers/memory/brcmstb_dpfe.c
613
writel_relaxed(le32_to_cpu(fw[i]), mem + i);
drivers/memory/omap-gpmc.c
1031
res = &gpmc->mem;
drivers/memory/omap-gpmc.c
204
struct resource mem;
drivers/memory/omap-gpmc.c
947
struct resource *res = &gpmc->mem;
drivers/memory/omap-gpmc.c
963
struct resource *res = &gpmc->mem;
drivers/memory/omap-gpmc.c
978
struct resource *res = &gpmc->mem;
drivers/message/fusion/mptbase.c
1638
u8 __iomem *mem;
drivers/message/fusion/mptbase.c
1714
mem = NULL;
drivers/message/fusion/mptbase.c
1717
mem = ioremap(mem_phys, msize);
drivers/message/fusion/mptbase.c
1718
if (mem == NULL) {
drivers/message/fusion/mptbase.c
1724
ioc->memmap = mem;
drivers/message/fusion/mptbase.c
1726
ioc->name, mem, (unsigned long long)mem_phys));
drivers/message/fusion/mptbase.c
1729
ioc->chip = (SYSIF_REGS __iomem *)mem;
drivers/message/fusion/mptbase.c
4318
u8 *mem;
drivers/message/fusion/mptbase.c
4327
mem = kmalloc(sz, GFP_ATOMIC);
drivers/message/fusion/mptbase.c
4328
if (mem == NULL)
drivers/message/fusion/mptbase.c
4331
ioc->ReqToChain = (int *) mem;
drivers/message/fusion/mptbase.c
4333
ioc->name, mem, sz));
drivers/message/fusion/mptbase.c
4334
mem = kmalloc(sz, GFP_ATOMIC);
drivers/message/fusion/mptbase.c
4335
if (mem == NULL)
drivers/message/fusion/mptbase.c
4338
ioc->RequestNB = (int *) mem;
drivers/message/fusion/mptbase.c
4340
ioc->name, mem, sz));
drivers/message/fusion/mptbase.c
4401
mem = kmalloc(sz, GFP_ATOMIC);
drivers/message/fusion/mptbase.c
4402
if (mem == NULL)
drivers/message/fusion/mptbase.c
4405
ioc->ChainToChain = (int *) mem;
drivers/message/fusion/mptbase.c
4407
ioc->name, mem, sz));
drivers/message/fusion/mptbase.c
4409
mem = (u8 *) ioc->ChainToChain;
drivers/message/fusion/mptbase.c
4411
memset(mem, 0xFF, sz);
drivers/message/fusion/mptbase.c
4432
u8 *mem;
drivers/message/fusion/mptbase.c
4490
mem = dma_alloc_coherent(&ioc->pcidev->dev, total_size,
drivers/message/fusion/mptbase.c
4492
if (mem == NULL) {
drivers/message/fusion/mptbase.c
4499
ioc->name, mem, (void *)(ulong)alloc_dma, total_size, total_size));
drivers/message/fusion/mptbase.c
4501
memset(mem, 0, total_size);
drivers/message/fusion/mptbase.c
4503
ioc->alloc = mem;
drivers/message/fusion/mptbase.c
4506
ioc->reply_frames = (MPT_FRAME_HDR *) mem;
drivers/message/fusion/mptbase.c
4513
mem += reply_sz;
drivers/message/fusion/mptbase.c
4517
ioc->req_frames = (MPT_FRAME_HDR *) mem;
drivers/message/fusion/mptbase.c
4521
ioc->name, mem, (void *)(ulong)alloc_dma));
drivers/message/fusion/mptbase.c
4527
mem += ioc->req_sz;
drivers/message/fusion/mptbase.c
4530
ioc->ChainBuffer = mem;
drivers/message/fusion/mptbase.c
4543
mem = (u8 *)ioc->ChainBuffer;
drivers/message/fusion/mptbase.c
4545
mf = (MPT_FRAME_HDR *) mem;
drivers/message/fusion/mptbase.c
4547
mem += ioc->req_sz;
drivers/message/fusion/mptbase.c
4553
mem = (u8 *) ioc->req_frames;
drivers/message/fusion/mptbase.c
4558
mf = (MPT_FRAME_HDR *) mem;
drivers/message/fusion/mptbase.c
4563
mem += ioc->req_sz;
drivers/message/fusion/mptbase.c
5372
u8 *mem;
drivers/message/fusion/mptbase.c
5374
mem = kmalloc(sz, GFP_ATOMIC);
drivers/message/fusion/mptbase.c
5375
if (mem == NULL)
drivers/message/fusion/mptbase.c
5378
ioc->spi_data.nvram = (int *) mem;
drivers/message/fusion/mptbase.c
5963
u8 *mem;
drivers/message/fusion/mptbase.c
6009
mem = kmemdup(pIoc2, iocpage2sz, GFP_KERNEL);
drivers/message/fusion/mptbase.c
6010
if (!mem) {
drivers/message/fusion/mptbase.c
6015
ioc->raid_data.pIocPg2 = (IOCPage2_t *) mem;
drivers/message/fusion/mptbase.c
6034
u8 *mem;
drivers/message/fusion/mptbase.c
6078
mem = kmalloc(iocpage3sz, GFP_KERNEL);
drivers/message/fusion/mptbase.c
6079
if (mem) {
drivers/message/fusion/mptbase.c
6080
memcpy(mem, (u8 *)pIoc3, iocpage3sz);
drivers/message/fusion/mptbase.c
6081
ioc->raid_data.pIocPg3 = (IOCPage3_t *) mem;
drivers/mfd/cs40l50-core.c
214
cs40l50->dsp.mem = cs40l50_dsp_regions;
drivers/mfd/intel-lpss-acpi.c
190
info->mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/mfd/intel-lpss-pci.c
66
info->mem = pci_resource_n(pdev, 0);
drivers/mfd/intel-lpss.c
254
resource_size_t addr = lpss->info->mem->start;
drivers/mfd/intel-lpss.c
396
if (!info || !info->mem)
drivers/mfd/intel-lpss.c
406
lpss->priv = devm_ioremap_uc(dev, info->mem->start + LPSS_PRIV_OFFSET,
drivers/mfd/intel-lpss.c
442
1, info->mem, info->irq, NULL);
drivers/mfd/intel-lpss.c
449
1, info->mem, info->irq, NULL);
drivers/mfd/intel-lpss.h
33
struct resource *mem;
drivers/mfd/intel_pmc_bxt.c
321
scu_data->mem.flags = res->flags;
drivers/mfd/intel_pmc_bxt.c
322
scu_data->mem.start = res->start;
drivers/mfd/intel_pmc_bxt.c
323
scu_data->mem.end = res->start + PLAT_RESOURCE_IPC_SIZE - 1;
drivers/mfd/lpc_ich.c
1253
struct resource *mem = info->resources[i];
drivers/mfd/lpc_ich.c
1257
mem->start = base.start + offset;
drivers/mfd/lpc_ich.c
1258
mem->end = base.start + offset + INTEL_GPIO_RESOURCE_SIZE - 1;
drivers/mfd/lpc_ich.c
1259
mem->flags = base.flags;
drivers/mfd/vexpress-sysreg.c
102
struct resource *mem;
drivers/mfd/vexpress-sysreg.c
106
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/mfd/vexpress-sysreg.c
107
if (!mem)
drivers/mfd/vexpress-sysreg.c
110
base = devm_ioremap(&pdev->dev, mem->start, resource_size(mem));
drivers/mfd/vexpress-sysreg.c
141
ARRAY_SIZE(vexpress_sysreg_cells), mem, 0, NULL);
drivers/misc/eeprom/at25.c
482
static int at25_probe(struct spi_mem *mem)
drivers/misc/eeprom/at25.c
484
struct spi_device *spi = mem->spi;
drivers/misc/eeprom/at25.c
494
at25->spimem = mem;
drivers/misc/mei/gsc-me.c
35
static void mei_gsc_set_ext_op_mem(const struct mei_me_hw *hw, struct resource *mem)
drivers/misc/mei/gsc-me.c
37
u32 low = lower_32_bits(mem->start);
drivers/misc/mei/gsc-me.c
38
u32 hi = upper_32_bits(mem->start);
drivers/misc/mei/gsc-me.c
39
u32 limit = (resource_size(mem) / SZ_4K) | GSC_EXT_OP_MEM_VALID;
drivers/mmc/core/mmc_test.c
1384
err = mmc_test_map_sg_max_scatter(t->mem, sz, t->sg,
drivers/mmc/core/mmc_test.c
1388
err = mmc_test_map_sg(t->mem, sz, t->sg, 1, t->max_segs,
drivers/mmc/core/mmc_test.c
1396
err = mmc_test_map_sg_max_scatter(t->mem, sz, t->sg_areq,
drivers/mmc/core/mmc_test.c
1400
err = mmc_test_map_sg(t->mem, sz, t->sg_areq, 1, t->max_segs,
drivers/mmc/core/mmc_test.c
1520
mmc_test_free_mem(t->mem);
drivers/mmc/core/mmc_test.c
1568
t->mem = mmc_test_alloc_mem(min_sz, t->max_tfr, t->max_segs,
drivers/mmc/core/mmc_test.c
1570
if (!t->mem)
drivers/mmc/core/mmc_test.c
311
static void mmc_test_free_mem(struct mmc_test_mem *mem)
drivers/mmc/core/mmc_test.c
313
if (!mem)
drivers/mmc/core/mmc_test.c
315
while (mem->cnt--)
drivers/mmc/core/mmc_test.c
316
__free_pages(mem->arr[mem->cnt].page,
drivers/mmc/core/mmc_test.c
317
mem->arr[mem->cnt].order);
drivers/mmc/core/mmc_test.c
318
kfree(mem->arr);
drivers/mmc/core/mmc_test.c
319
kfree(mem);
drivers/mmc/core/mmc_test.c
338
struct mmc_test_mem *mem;
drivers/mmc/core/mmc_test.c
351
mem = kzalloc_obj(*mem);
drivers/mmc/core/mmc_test.c
352
if (!mem)
drivers/mmc/core/mmc_test.c
355
mem->arr = kzalloc_objs(*mem->arr, max_segs);
drivers/mmc/core/mmc_test.c
356
if (!mem->arr)
drivers/mmc/core/mmc_test.c
377
mem->arr[mem->cnt].page = page;
drivers/mmc/core/mmc_test.c
378
mem->arr[mem->cnt].order = order;
drivers/mmc/core/mmc_test.c
379
mem->cnt += 1;
drivers/mmc/core/mmc_test.c
384
if (mem->cnt >= max_segs) {
drivers/mmc/core/mmc_test.c
391
return mem;
drivers/mmc/core/mmc_test.c
394
mmc_test_free_mem(mem);
drivers/mmc/core/mmc_test.c
402
static int mmc_test_map_sg(struct mmc_test_mem *mem, unsigned long size,
drivers/mmc/core/mmc_test.c
417
for (i = 0; i < mem->cnt; i++) {
drivers/mmc/core/mmc_test.c
418
unsigned long len = PAGE_SIZE << mem->arr[i].order;
drivers/mmc/core/mmc_test.c
432
sg_set_page(sg, mem->arr[i].page, len, 0);
drivers/mmc/core/mmc_test.c
453
static int mmc_test_map_sg_max_scatter(struct mmc_test_mem *mem,
drivers/mmc/core/mmc_test.c
461
unsigned int i = mem->cnt, cnt;
drivers/mmc/core/mmc_test.c
469
base = page_address(mem->arr[--i].page);
drivers/mmc/core/mmc_test.c
470
cnt = 1 << mem->arr[i].order;
drivers/mmc/core/mmc_test.c
492
i = mem->cnt;
drivers/mmc/core/mmc_test.c
83
struct mmc_test_mem *mem;
drivers/mmc/host/davinci_mmc.c
1188
struct resource *r, *mem = NULL;
drivers/mmc/host/davinci_mmc.c
1201
mem = devm_request_mem_region(&pdev->dev, r->start, mem_size,
drivers/mmc/host/davinci_mmc.c
1203
if (!mem)
drivers/mmc/host/davinci_mmc.c
1213
host->mem_res = mem;
drivers/mmc/host/davinci_mmc.c
1214
host->base = devm_ioremap(&pdev->dev, mem->start, mem_size);
drivers/mmc/host/davinci_mmc.c
1314
rename_region(mem, mmc_hostname(mmc));
drivers/mtd/nand/raw/lpc32xx_mlc.c
392
static int lpc32xx_xmit_dma(struct mtd_info *mtd, void *mem, int len,
drivers/mtd/nand/raw/lpc32xx_mlc.c
401
sg_init_one(&host->sgl, mem, len);
drivers/mtd/nand/raw/lpc32xx_slc.c
426
void *mem, int len, enum dma_transfer_direction dir)
drivers/mtd/nand/raw/lpc32xx_slc.c
448
sg_init_one(&host->sgl, mem, len);
drivers/mtd/nand/spi/core.c
1950
static int spinand_probe(struct spi_mem *mem)
drivers/mtd/nand/spi/core.c
1956
spinand = devm_kzalloc(&mem->spi->dev, sizeof(*spinand),
drivers/mtd/nand/spi/core.c
1961
spinand->spimem = mem;
drivers/mtd/nand/spi/core.c
1962
spi_mem_set_drvdata(mem, spinand);
drivers/mtd/nand/spi/core.c
1963
spinand_set_of_node(spinand, mem->spi->dev.of_node);
drivers/mtd/nand/spi/core.c
1966
mtd->dev.parent = &mem->spi->dev;
drivers/mtd/nand/spi/core.c
1984
static int spinand_remove(struct spi_mem *mem)
drivers/mtd/nand/spi/core.c
1990
spinand = spi_mem_get_drvdata(mem);
drivers/net/can/c_can/c_can_platform.c
262
struct resource *mem;
drivers/net/can/c_can/c_can_platform.c
280
addr = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/net/can/c_can/c_can_platform.c
293
switch (mem->flags & IORESOURCE_MEM_TYPE_MASK) {
drivers/net/can/cc770/cc770_isa.c
172
idx, port[idx], mem[idx], irq[idx]);
drivers/net/can/cc770/cc770_isa.c
173
if (mem[idx]) {
drivers/net/can/cc770/cc770_isa.c
174
if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) {
drivers/net/can/cc770/cc770_isa.c
178
base = ioremap(mem[idx], iosize);
drivers/net/can/cc770/cc770_isa.c
202
if (mem[idx]) {
drivers/net/can/cc770/cc770_isa.c
204
dev->base_addr = mem[idx];
drivers/net/can/cc770/cc770_isa.c
277
if (mem[idx])
drivers/net/can/cc770/cc770_isa.c
280
if (mem[idx])
drivers/net/can/cc770/cc770_isa.c
281
release_mem_region(mem[idx], iosize);
drivers/net/can/cc770/cc770_isa.c
296
if (mem[idx]) {
drivers/net/can/cc770/cc770_isa.c
298
release_mem_region(mem[idx], CC770_IOSIZE);
drivers/net/can/cc770/cc770_isa.c
321
if ((port[idx] || mem[idx]) && irq[idx]) {
drivers/net/can/cc770/cc770_isa.c
335
idx, port[idx], mem[idx], irq[idx]);
drivers/net/can/cc770/cc770_isa.c
336
} else if (idx == 0 || port[idx] || mem[idx]) {
drivers/net/can/cc770/cc770_isa.c
69
static unsigned long mem[MAXDEV];
drivers/net/can/cc770/cc770_isa.c
80
module_param_hw_array(mem, ulong, iomem, NULL, 0444);
drivers/net/can/cc770/cc770_isa.c
81
MODULE_PARM_DESC(mem, "I/O memory address");
drivers/net/can/cc770/cc770_platform.c
153
struct resource *mem;
drivers/net/can/cc770/cc770_platform.c
158
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/can/cc770/cc770_platform.c
160
if (!mem || irq <= 0)
drivers/net/can/cc770/cc770_platform.c
163
mem_size = resource_size(mem);
drivers/net/can/cc770/cc770_platform.c
164
if (!request_mem_region(mem->start, mem_size, pdev->name))
drivers/net/can/cc770/cc770_platform.c
167
base = ioremap(mem->start, mem_size);
drivers/net/can/cc770/cc770_platform.c
218
release_mem_region(mem->start, mem_size);
drivers/net/can/cc770/cc770_platform.c
227
struct resource *mem;
drivers/net/can/cc770/cc770_platform.c
233
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/can/cc770/cc770_platform.c
234
release_mem_region(mem->start, resource_size(mem));
drivers/net/can/esd/esdacc.c
144
void acc_init_bm_ptr(struct acc_ov *ov, struct acc_core *cores, const void *mem)
drivers/net/can/esd/esdacc.c
162
ov->bmfifo.messages = mem;
drivers/net/can/esd/esdacc.c
163
ov->bmfifo.irq_cnt = mem + (ov->total_cores + 1U) * ACC_CORE_DMABUF_SIZE;
drivers/net/can/esd/esdacc.c
168
core->bmfifo.messages = mem + (u + 1U) * ACC_CORE_DMABUF_SIZE;
drivers/net/can/esd/esdacc.h
350
const void *mem);
drivers/net/can/sja1000/sja1000_isa.c
126
idx, port[idx], mem[idx], irq[idx]);
drivers/net/can/sja1000/sja1000_isa.c
128
if (mem[idx]) {
drivers/net/can/sja1000/sja1000_isa.c
129
if (!request_mem_region(mem[idx], iosize, DRV_NAME)) {
drivers/net/can/sja1000/sja1000_isa.c
133
base = ioremap(mem[idx], iosize);
drivers/net/can/sja1000/sja1000_isa.c
157
if (mem[idx]) {
drivers/net/can/sja1000/sja1000_isa.c
159
dev->base_addr = mem[idx];
drivers/net/can/sja1000/sja1000_isa.c
215
if (mem[idx])
drivers/net/can/sja1000/sja1000_isa.c
218
if (mem[idx])
drivers/net/can/sja1000/sja1000_isa.c
219
release_mem_region(mem[idx], iosize);
drivers/net/can/sja1000/sja1000_isa.c
234
if (mem[idx]) {
drivers/net/can/sja1000/sja1000_isa.c
236
release_mem_region(mem[idx], SJA1000_IOSIZE);
drivers/net/can/sja1000/sja1000_isa.c
259
if ((port[idx] || mem[idx]) && irq[idx]) {
drivers/net/can/sja1000/sja1000_isa.c
273
DRV_NAME, idx, port[idx], mem[idx], irq[idx]);
drivers/net/can/sja1000/sja1000_isa.c
274
} else if (idx == 0 || port[idx] || mem[idx]) {
drivers/net/can/sja1000/sja1000_isa.c
32
static unsigned long mem[MAXDEV];
drivers/net/can/sja1000/sja1000_isa.c
43
module_param_hw_array(mem, ulong, iomem, NULL, 0444);
drivers/net/can/sja1000/sja1000_isa.c
44
MODULE_PARM_DESC(mem, "I/O memory address");
drivers/net/can/softing/softing_fw.c
116
mem = *pmem;
drivers/net/can/softing/softing_fw.c
117
*ptype = le16_to_cpup((void *)&mem[0]);
drivers/net/can/softing/softing_fw.c
118
*paddr = le32_to_cpup((void *)&mem[2]);
drivers/net/can/softing/softing_fw.c
119
*plen = le16_to_cpup((void *)&mem[6]);
drivers/net/can/softing/softing_fw.c
120
*pdat = &mem[8];
drivers/net/can/softing/softing_fw.c
122
end = &mem[8 + *plen];
drivers/net/can/softing/softing_fw.c
124
for (checksum[1] = 0; mem < end; ++mem)
drivers/net/can/softing/softing_fw.c
125
checksum[1] += *mem;
drivers/net/can/softing/softing_fw.c
138
const uint8_t *mem, *end, *dat;
drivers/net/can/softing/softing_fw.c
153
mem = fw->data;
drivers/net/can/softing/softing_fw.c
154
end = &mem[fw->size];
drivers/net/can/softing/softing_fw.c
156
ret = fw_parse(&mem, &type, &addr, &len, &dat);
drivers/net/can/softing/softing_fw.c
166
while (mem < end) {
drivers/net/can/softing/softing_fw.c
167
ret = fw_parse(&mem, &type, &addr, &len, &dat);
drivers/net/can/softing/softing_fw.c
221
const uint8_t *mem, *end, *dat;
drivers/net/can/softing/softing_fw.c
237
mem = fw->data;
drivers/net/can/softing/softing_fw.c
238
end = &mem[fw->size];
drivers/net/can/softing/softing_fw.c
240
ret = fw_parse(&mem, &type, &addr, &len, &dat);
drivers/net/can/softing/softing_fw.c
255
while (mem < end) {
drivers/net/can/softing/softing_fw.c
256
ret = fw_parse(&mem, &type, &addr, &len, &dat);
drivers/net/can/softing/softing_fw.c
98
const uint8_t *mem;
drivers/net/dsa/b53/b53_mmap.c
368
void __iomem *mem;
drivers/net/dsa/b53/b53_mmap.c
370
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/net/dsa/b53/b53_mmap.c
371
if (IS_ERR(mem))
drivers/net/dsa/b53/b53_mmap.c
372
return PTR_ERR(mem);
drivers/net/dsa/b53/b53_mmap.c
379
pdata->regs = mem;
drivers/net/dsa/sja1105/sja1105_static_config.c
1020
int i, mem = 0;
drivers/net/dsa/sja1105/sja1105_static_config.c
1025
mem += l2_fwd_params->part_spc[i];
drivers/net/dsa/sja1105/sja1105_static_config.c
1030
mem += vl_fwd_params->partspc[i];
drivers/net/dsa/sja1105/sja1105_static_config.c
1036
if (mem > max_mem)
drivers/net/ethernet/8390/8390.h
82
void __iomem *mem;
drivers/net/ethernet/8390/ax88796.c
120
void __iomem *ioaddr = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
190
void __iomem *nic_base = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
234
void __iomem *nic_base = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
271
void __iomem *nic_base = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
404
ei_outb(reg_gpoc, ei_local->mem + EI_SHIFT(0x17));
drivers/net/ethernet/8390/ax88796.c
467
ax->addr_memr = ei_local->mem + AX_MEMR;
drivers/net/ethernet/8390/ax88796.c
615
u8 reg = ei_inb(ei_local->mem + AX_MEMR);
drivers/net/ethernet/8390/ax88796.c
626
u8 reg = ei_inb(ei_local->mem + AX_MEMR);
drivers/net/ethernet/8390/ax88796.c
637
ei_outb(reg, ei_local->mem + AX_MEMR);
drivers/net/ethernet/8390/ax88796.c
662
void __iomem *ioaddr = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
685
void __iomem *ioaddr = ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
741
ei_outb(ax->plat->dcr_val, ei_local->mem + EN0_DCFG);
drivers/net/ethernet/8390/ax88796.c
754
ei_local->mem + E8390_CMD); /* 0x61 */
drivers/net/ethernet/8390/ax88796.c
819
struct resource *mem;
drivers/net/ethernet/8390/ax88796.c
823
iounmap(ei_local->mem);
drivers/net/ethernet/8390/ax88796.c
824
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/8390/ax88796.c
825
release_mem_region(mem->start, resource_size(mem));
drivers/net/ethernet/8390/ax88796.c
829
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
drivers/net/ethernet/8390/ax88796.c
830
release_mem_region(mem->start, resource_size(mem));
drivers/net/ethernet/8390/ax88796.c
849
struct resource *irq, *mem, *mem2;
drivers/net/ethernet/8390/ax88796.c
881
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/8390/ax88796.c
882
if (!mem) {
drivers/net/ethernet/8390/ax88796.c
888
mem_size = resource_size(mem);
drivers/net/ethernet/8390/ax88796.c
902
if (!request_mem_region(mem->start, mem_size, pdev->name)) {
drivers/net/ethernet/8390/ax88796.c
908
ei_local->mem = ioremap(mem->start, mem_size);
drivers/net/ethernet/8390/ax88796.c
909
dev->base_addr = (unsigned long)ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
911
if (ei_local->mem == NULL) {
drivers/net/ethernet/8390/ax88796.c
912
dev_err(&pdev->dev, "Cannot ioremap area %pR\n", mem);
drivers/net/ethernet/8390/ax88796.c
941
ei_local->reg_offset[0x1f] = ax->map2 - ei_local->mem;
drivers/net/ethernet/8390/ax88796.c
959
iounmap(ei_local->mem);
drivers/net/ethernet/8390/ax88796.c
962
release_mem_region(mem->start, mem_size);
drivers/net/ethernet/8390/mcf8390.c
406
struct resource *mem;
drivers/net/ethernet/8390/mcf8390.c
414
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/8390/mcf8390.c
415
if (mem == NULL) {
drivers/net/ethernet/8390/mcf8390.c
419
msize = resource_size(mem);
drivers/net/ethernet/8390/mcf8390.c
420
if (!request_mem_region(mem->start, msize, pdev->name))
drivers/net/ethernet/8390/mcf8390.c
425
release_mem_region(mem->start, msize);
drivers/net/ethernet/8390/mcf8390.c
433
dev->base_addr = mem->start;
drivers/net/ethernet/8390/mcf8390.c
437
release_mem_region(mem->start, msize);
drivers/net/ethernet/8390/mcf8390.c
447
struct resource *mem;
drivers/net/ethernet/8390/mcf8390.c
450
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/8390/mcf8390.c
451
release_mem_region(mem->start, resource_size(mem));
drivers/net/ethernet/8390/pcnet_cs.c
1369
void __iomem *xfer_start = ei_status.mem + (TX_PAGES<<8)
drivers/net/ethernet/8390/pcnet_cs.c
1383
void __iomem *base = ei_status.mem;
drivers/net/ethernet/8390/pcnet_cs.c
1404
void __iomem *shmem = ei_status.mem + (start_page << 8);
drivers/net/ethernet/8390/pcnet_cs.c
1460
ei_status.mem = info->base + offset;
drivers/net/ethernet/8390/pcnet_cs.c
1462
dev->mem_start = (u_long)ei_status.mem;
drivers/net/ethernet/8390/smc-ultra.c
299
ei_status.mem = ioremap(dev->mem_start, (ei_status.stop_page - START_PG)*256);
drivers/net/ethernet/8390/smc-ultra.c
300
if (!ei_status.mem) {
drivers/net/ethernet/8390/smc-ultra.c
445
void __iomem *hdr_start = ei_status.mem + ((ring_page - START_PG)<<8);
drivers/net/ethernet/8390/smc-ultra.c
465
void __iomem *xfer_start = ei_status.mem + ring_offset - (START_PG<<8);
drivers/net/ethernet/8390/smc-ultra.c
475
memcpy_fromio(skb->data + semi_count, ei_status.mem + TX_PAGES * 256, count);
drivers/net/ethernet/8390/smc-ultra.c
487
void __iomem *shmem = ei_status.mem + ((start_page - START_PG)<<8);
drivers/net/ethernet/8390/smc-ultra.c
613
iounmap(ei_status.mem);
drivers/net/ethernet/8390/wd.c
338
ei_status.mem = ioremap(dev->mem_start, ei_status.priv);
drivers/net/ethernet/8390/wd.c
339
if (!ei_status.mem) {
drivers/net/ethernet/8390/wd.c
366
iounmap(ei_status.mem);
drivers/net/ethernet/8390/wd.c
416
void __iomem *hdr_start = ei_status.mem + ((ring_page - WD_START_PG)<<8);
drivers/net/ethernet/8390/wd.c
443
void __iomem *xfer_start = ei_status.mem + offset;
drivers/net/ethernet/8390/wd.c
450
memcpy_fromio(skb->data + semi_count, ei_status.mem + TX_PAGES * 256, count);
drivers/net/ethernet/8390/wd.c
466
void __iomem *shmem = ei_status.mem + ((start_page - WD_START_PG)<<8);
drivers/net/ethernet/8390/wd.c
504
static int mem[MAX_WD_CARDS];
drivers/net/ethernet/8390/wd.c
509
module_param_hw_array(mem, int, iomem, NULL, 0);
drivers/net/ethernet/8390/wd.c
514
MODULE_PARM_DESC(mem, "memory base address(es)(ignored for PureData boards)");
drivers/net/ethernet/8390/wd.c
538
dev->mem_start = mem[this_dev];
drivers/net/ethernet/8390/wd.c
558
iounmap(ei_status.mem);
drivers/net/ethernet/8390/xsurf100.c
115
ei_outw(*(uint16_t *)src, ei_local->mem + NE_DATAPORT);
drivers/net/ethernet/8390/xsurf100.c
119
ei_outb(*(uint8_t *)src, ei_local->mem + NE_DATAPORT);
drivers/net/ethernet/8390/xsurf100.c
141
*(uint16_t *)dst = ei_inw(ei_local->mem + NE_DATAPORT);
drivers/net/ethernet/8390/xsurf100.c
145
*(uint8_t *)dst = ei_inb(ei_local->mem + NE_DATAPORT);
drivers/net/ethernet/8390/xsurf100.c
158
void __iomem *nic_base = ei_local->mem;
drivers/net/ethernet/8390/xsurf100.c
187
void __iomem *nic_base = ei_local->mem;
drivers/net/ethernet/amd/atarilance.c
200
/* base+0xf */ volatile unsigned char mem;
drivers/net/ethernet/amd/atarilance.c
222
struct lance_memory *mem;
drivers/net/ethernet/amd/atarilance.c
234
#define MEM lp->mem
drivers/net/ethernet/amd/atarilance.c
542
i = IO->mem;
drivers/net/ethernet/amd/atarilance.c
600
i = IO->mem;
drivers/net/ethernet/amd/sun3lance.c
153
struct lance_memory *mem;
drivers/net/ethernet/amd/sun3lance.c
163
#define MEM lp->mem
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
100
if (!mem)
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
101
mem = dma_alloc_coherent(dev, size, dma, GFP_KERNEL);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
103
return mem;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
81
void *mem;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
83
mem = kzalloc_node(size, GFP_KERNEL, node);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
84
if (!mem)
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
85
mem = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
87
return mem;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
93
void *mem;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
97
mem = dma_alloc_coherent(dev, size, dma, GFP_KERNEL);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
54
void *mem;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
56
mem = kzalloc_node(size, GFP_KERNEL, node);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
57
if (!mem)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
58
mem = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
60
return mem;
drivers/net/ethernet/broadcom/cnic.c
1102
uinfo->mem[0].addr = pci_resource_start(dev->pcidev, 0);
drivers/net/ethernet/broadcom/cnic.c
1103
uinfo->mem[0].internal_addr = dev->regview;
drivers/net/ethernet/broadcom/cnic.c
1104
uinfo->mem[0].memtype = UIO_MEM_PHYS;
drivers/net/ethernet/broadcom/cnic.c
1107
uinfo->mem[0].size = MB_GET_CID_ADDR(TX_TSS_CID +
drivers/net/ethernet/broadcom/cnic.c
1109
uinfo->mem[1].addr = (unsigned long) cp->status_blk.gen &
drivers/net/ethernet/broadcom/cnic.c
1111
uinfo->mem[1].dma_addr = cp->status_blk_map;
drivers/net/ethernet/broadcom/cnic.c
1113
uinfo->mem[1].size = PAGE_ALIGN(BNX2_SBLK_MSIX_ALIGN_SIZE * 9);
drivers/net/ethernet/broadcom/cnic.c
1115
uinfo->mem[1].size = PAGE_ALIGN(BNX2_SBLK_MSIX_ALIGN_SIZE);
drivers/net/ethernet/broadcom/cnic.c
1119
uinfo->mem[0].size = pci_resource_len(dev->pcidev, 0);
drivers/net/ethernet/broadcom/cnic.c
1121
uinfo->mem[1].addr = (unsigned long) cp->bnx2x_def_status_blk &
drivers/net/ethernet/broadcom/cnic.c
1123
uinfo->mem[1].dma_addr = cp->status_blk_map;
drivers/net/ethernet/broadcom/cnic.c
1124
uinfo->mem[1].size = PAGE_ALIGN(sizeof(*cp->bnx2x_def_status_blk));
drivers/net/ethernet/broadcom/cnic.c
1129
uinfo->mem[1].dma_device = &dev->pcidev->dev;
drivers/net/ethernet/broadcom/cnic.c
1130
uinfo->mem[1].memtype = UIO_MEM_DMA_COHERENT;
drivers/net/ethernet/broadcom/cnic.c
1132
uinfo->mem[2].addr = (unsigned long) udev->l2_ring;
drivers/net/ethernet/broadcom/cnic.c
1133
uinfo->mem[2].dma_addr = udev->l2_ring_map;
drivers/net/ethernet/broadcom/cnic.c
1134
uinfo->mem[2].size = PAGE_ALIGN(udev->l2_ring_size);
drivers/net/ethernet/broadcom/cnic.c
1135
uinfo->mem[2].dma_device = &dev->pcidev->dev;
drivers/net/ethernet/broadcom/cnic.c
1136
uinfo->mem[2].memtype = UIO_MEM_DMA_COHERENT;
drivers/net/ethernet/broadcom/cnic.c
1138
uinfo->mem[3].addr = (unsigned long) udev->l2_buf;
drivers/net/ethernet/broadcom/cnic.c
1139
uinfo->mem[3].dma_addr = udev->l2_buf_map;
drivers/net/ethernet/broadcom/cnic.c
1140
uinfo->mem[3].size = PAGE_ALIGN(udev->l2_buf_size);
drivers/net/ethernet/broadcom/cnic.c
1141
uinfo->mem[3].dma_device = &dev->pcidev->dev;
drivers/net/ethernet/broadcom/cnic.c
1142
uinfo->mem[3].memtype = UIO_MEM_DMA_COHERENT;
drivers/net/ethernet/cadence/macb_main.c
4465
static int macb_probe_queues(struct device *dev, void __iomem *mem, bool native_io)
drivers/net/ethernet/cadence/macb_main.c
4471
if (hw_is_gem(mem, native_io)) {
drivers/net/ethernet/cadence/macb_main.c
4473
queue_mask |= __raw_readl(mem + GEM_DCFG6) & 0xFF;
drivers/net/ethernet/cadence/macb_main.c
4475
queue_mask |= readl_relaxed(mem + GEM_DCFG6) & 0xFF;
drivers/net/ethernet/cadence/macb_main.c
5548
void __iomem *mem;
drivers/net/ethernet/cadence/macb_main.c
5554
mem = devm_platform_get_and_ioremap_resource(pdev, 0, &regs);
drivers/net/ethernet/cadence/macb_main.c
5555
if (IS_ERR(mem))
drivers/net/ethernet/cadence/macb_main.c
5556
return PTR_ERR(mem);
drivers/net/ethernet/cadence/macb_main.c
5571
native_io = hw_is_native_io(mem);
drivers/net/ethernet/cadence/macb_main.c
5573
num_queues = macb_probe_queues(&pdev->dev, mem, native_io);
drivers/net/ethernet/cadence/macb_main.c
5592
bp->regs = mem;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2445
struct mc7 *mem;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2461
mem = &adapter->cm;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2463
mem = &adapter->pmrx;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2465
mem = &adapter->pmtx;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2488
t3_mc7_bd_read(mem, t.addr / 8, chunk / 8,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
311
struct mc7 *mem;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
316
mem = &adapter->cm;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
318
mem = &adapter->pmrx;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
320
mem = &adapter->pmtx;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
325
t3_mc7_bd_read(mem, t->addr / 8, t->len / 8,
drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h
115
struct cudbg_mem_desc mem[ARRAY_SIZE(cudbg_region) + 3];
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1200
if (meminfo->mem[i].idx >= ARRAY_SIZE(cudbg_region))
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1203
if (!(meminfo->mem[i].limit))
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1204
meminfo->mem[i].limit =
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1206
meminfo->mem[i + 1].base - 1 : ~0;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1208
if (meminfo->mem[i].idx == idx) {
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1210
if (meminfo->mem[i].base < meminfo->avail[mc].base &&
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1211
meminfo->mem[i].limit < meminfo->avail[mc].base)
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1214
if (meminfo->mem[i].base > meminfo->avail[mc].limit)
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
1217
memcpy(mem_desc, &meminfo->mem[i],
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
547
memset(meminfo_buff->mem, 0,
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
549
md = meminfo_buff->mem;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
551
for (i = 0; i < ARRAY_SIZE(meminfo_buff->mem); i++) {
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
552
meminfo_buff->mem[i].limit = 0;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
553
meminfo_buff->mem[i].idx = i;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
732
n = md - meminfo_buff->mem;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
735
sort(meminfo_buff->mem, n, sizeof(struct cudbg_mem_desc),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3255
unsigned int mem;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3260
mem = (uintptr_t)file->private_data & 0x7;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3261
adap = file->private_data - mem;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3273
unsigned int mem = (uintptr_t)file->private_data & 0x7;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3274
struct adapter *adap = file->private_data - mem;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3290
ret = t4_memory_rw(adap, 0, mem, pos, count, data, T4_MEMORY_READ);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3485
if (meminfo.mem[i].idx >= ARRAY_SIZE(cudbg_region))
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3487
if (!meminfo.mem[i].limit)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3488
meminfo.mem[i].limit =
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3490
meminfo.mem[i + 1].base - 1 : ~0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3491
mem_region_show(seq, cudbg_region[meminfo.mem[i].idx],
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3492
meminfo.mem[i].base, meminfo.mem[i].limit);
drivers/net/ethernet/cirrus/ep93xx_eth.c
744
struct resource *mem;
drivers/net/ethernet/cirrus/ep93xx_eth.c
760
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/cirrus/ep93xx_eth.c
761
release_mem_region(mem->start, resource_size(mem));
drivers/net/ethernet/cirrus/ep93xx_eth.c
771
struct resource *mem;
drivers/net/ethernet/cirrus/ep93xx_eth.c
782
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/cirrus/ep93xx_eth.c
784
if (!mem || irq < 0)
drivers/net/ethernet/cirrus/ep93xx_eth.c
787
base_addr = ioremap(mem->start, resource_size(mem));
drivers/net/ethernet/cirrus/ep93xx_eth.c
819
ep->res = request_mem_region(mem->start, resource_size(mem),
drivers/net/ethernet/emulex/benet/be_cmds.c
1981
struct be_dma_mem *mem = &adapter->rx_filter;
drivers/net/ethernet/emulex/benet/be_cmds.c
1982
struct be_cmd_req_rx_filter *req = mem->va;
drivers/net/ethernet/emulex/benet/be_cmds.c
1995
wrb, mem);
drivers/net/ethernet/emulex/benet/be_cmds.c
798
struct be_dma_mem *mem)
drivers/net/ethernet/emulex/benet/be_cmds.c
808
if (mem) {
drivers/net/ethernet/emulex/benet/be_cmds.c
812
sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma));
drivers/net/ethernet/emulex/benet/be_cmds.c
813
sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF);
drivers/net/ethernet/emulex/benet/be_cmds.c
814
sge->len = cpu_to_le32(mem->size);
drivers/net/ethernet/emulex/benet/be_cmds.c
821
struct be_dma_mem *mem)
drivers/net/ethernet/emulex/benet/be_cmds.c
823
int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
drivers/net/ethernet/emulex/benet/be_cmds.c
824
u64 dma = (u64)mem->dma;
drivers/net/ethernet/emulex/benet/be_main.c
146
struct be_dma_mem *mem = &q->dma_mem;
drivers/net/ethernet/emulex/benet/be_main.c
148
if (mem->va) {
drivers/net/ethernet/emulex/benet/be_main.c
149
dma_free_coherent(&adapter->pdev->dev, mem->size, mem->va,
drivers/net/ethernet/emulex/benet/be_main.c
150
mem->dma);
drivers/net/ethernet/emulex/benet/be_main.c
151
mem->va = NULL;
drivers/net/ethernet/emulex/benet/be_main.c
158
struct be_dma_mem *mem = &q->dma_mem;
drivers/net/ethernet/emulex/benet/be_main.c
163
mem->size = len * entry_size;
drivers/net/ethernet/emulex/benet/be_main.c
164
mem->va = dma_alloc_coherent(&adapter->pdev->dev, mem->size,
drivers/net/ethernet/emulex/benet/be_main.c
165
&mem->dma, GFP_KERNEL);
drivers/net/ethernet/emulex/benet/be_main.c
166
if (!mem->va)
drivers/net/ethernet/emulex/benet/be_main.c
5609
struct be_dma_mem *mem = &adapter->mbox_mem_alloced;
drivers/net/ethernet/emulex/benet/be_main.c
5612
if (mem->va)
drivers/net/ethernet/emulex/benet/be_main.c
5613
dma_free_coherent(dev, mem->size, mem->va, mem->dma);
drivers/net/ethernet/emulex/benet/be_main.c
5615
mem = &adapter->rx_filter;
drivers/net/ethernet/emulex/benet/be_main.c
5616
if (mem->va)
drivers/net/ethernet/emulex/benet/be_main.c
5617
dma_free_coherent(dev, mem->size, mem->va, mem->dma);
drivers/net/ethernet/emulex/benet/be_main.c
5619
mem = &adapter->stats_cmd;
drivers/net/ethernet/emulex/benet/be_main.c
5620
if (mem->va)
drivers/net/ethernet/emulex/benet/be_main.c
5621
dma_free_coherent(dev, mem->size, mem->va, mem->dma);
drivers/net/ethernet/ethoc.c
1029
struct resource *mem = NULL;
drivers/net/ethernet/ethoc.c
1067
mem = devm_request_mem_region(&pdev->dev, res->start,
drivers/net/ethernet/ethoc.c
1069
if (!mem) {
drivers/net/ethernet/ethoc.c
1075
netdev->mem_start = mem->start;
drivers/net/ethernet/ethoc.c
1076
netdev->mem_end = mem->end;
drivers/net/ethernet/ethoc.c
1101
netdev->mem_start, resource_size(mem));
drivers/net/ethernet/fealnx.c
1079
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1116
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1142
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1191
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1363
iowrite32(0, np->mem + TXPDR);
drivers/net/ethernet/fealnx.c
1421
np->mem + RXLBA);
drivers/net/ethernet/fealnx.c
1431
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1611
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1746
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1775
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
1892
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
413
void __iomem *mem;
drivers/net/ethernet/fealnx.c
537
np->mem = ioaddr;
drivers/net/ethernet/fealnx.c
690
pci_iounmap(pdev, np->mem);
drivers/net/ethernet/fealnx.c
750
void __iomem *miiport = np->mem + MANAGEMENT;
drivers/net/ethernet/fealnx.c
789
void __iomem *miiport = np->mem + MANAGEMENT;
drivers/net/ethernet/fealnx.c
821
void __iomem *ioaddr = np->mem;
drivers/net/ethernet/fealnx.c
928
if (ioread32(np->mem + BMCRSR) & LinkIsUp2) {
drivers/net/ethernet/fealnx.c
951
if (ioread32(np->mem + TCRRCR) & CR_R_FD)
drivers/net/ethernet/fealnx.c
955
if (ioread32(np->mem + TCRRCR) & CR_R_PS10)
drivers/net/ethernet/freescale/fec_mpc52xx.c
814
struct resource mem;
drivers/net/ethernet/freescale/fec_mpc52xx.c
831
rv = of_address_to_resource(np, 0, &mem);
drivers/net/ethernet/freescale/fec_mpc52xx.c
836
if (resource_size(&mem) < sizeof(struct mpc52xx_fec)) {
drivers/net/ethernet/freescale/fec_mpc52xx.c
838
(unsigned long)resource_size(&mem),
drivers/net/ethernet/freescale/fec_mpc52xx.c
844
if (!request_mem_region(mem.start, sizeof(struct mpc52xx_fec),
drivers/net/ethernet/freescale/fec_mpc52xx.c
854
ndev->base_addr = mem.start;
drivers/net/ethernet/freescale/fec_mpc52xx.c
860
priv->fec = ioremap(mem.start, sizeof(struct mpc52xx_fec));
drivers/net/ethernet/freescale/fec_mpc52xx.c
970
release_mem_region(mem.start, sizeof(struct mpc52xx_fec));
drivers/net/ethernet/freescale/fs_enet/fs_enet.h
168
void __iomem *mem; /* FCC DPRAM */
drivers/net/ethernet/freescale/fs_enet/mac-fcc.c
102
fep->fcc.mem = (void __iomem *)cpm2_immr;
drivers/net/ethernet/freescale/fs_enet/mac-fcc.c
285
memset_io(fep->fcc.mem + fpi->dpram_offset + 64, 0x88, 32);
drivers/net/ethernet/huawei/hinic/hinic_common.c
20
u32 *mem = data;
drivers/net/ethernet/huawei/hinic/hinic_common.c
26
*mem = cpu_to_be32(*mem);
drivers/net/ethernet/huawei/hinic/hinic_common.c
27
mem++;
drivers/net/ethernet/huawei/hinic/hinic_common.c
38
u32 *mem = data;
drivers/net/ethernet/huawei/hinic/hinic_common.c
44
*mem = be32_to_cpu(*mem);
drivers/net/ethernet/huawei/hinic/hinic_common.c
45
mem++;
drivers/net/ethernet/intel/e100.c
1388
offsetof(struct mem, dump_buf));
drivers/net/ethernet/intel/e100.c
1548
nic->dma_addr + offsetof(struct mem, stats))))
drivers/net/ethernet/intel/e100.c
1603
struct stats *s = &nic->mem->stats;
drivers/net/ethernet/intel/e100.c
2443
return (1 + E100_PHY_REGS) * sizeof(u32) + sizeof(nic->mem->dump_buf);
drivers/net/ethernet/intel/e100.c
2464
memset(nic->mem->dump_buf, 0, sizeof(nic->mem->dump_buf));
drivers/net/ethernet/intel/e100.c
2467
memcpy(&buff[1 + E100_PHY_REGS], nic->mem->dump_buf,
drivers/net/ethernet/intel/e100.c
2468
sizeof(nic->mem->dump_buf));
drivers/net/ethernet/intel/e100.c
2767
nic->mem = dma_alloc_coherent(&nic->pdev->dev, sizeof(struct mem),
drivers/net/ethernet/intel/e100.c
2769
return nic->mem ? 0 : -ENOMEM;
drivers/net/ethernet/intel/e100.c
2774
if (nic->mem) {
drivers/net/ethernet/intel/e100.c
2775
dma_free_coherent(&nic->pdev->dev, sizeof(struct mem),
drivers/net/ethernet/intel/e100.c
2776
nic->mem, nic->dma_addr);
drivers/net/ethernet/intel/e100.c
2777
nic->mem = NULL;
drivers/net/ethernet/intel/e100.c
578
struct mem *mem;
drivers/net/ethernet/intel/e100.c
649
u32 dma_addr = nic->dma_addr + offsetof(struct mem, selftest);
drivers/net/ethernet/intel/e100.c
654
nic->mem->selftest.signature = 0;
drivers/net/ethernet/intel/e100.c
655
nic->mem->selftest.result = 0xFFFFFFFF;
drivers/net/ethernet/intel/e100.c
666
if (nic->mem->selftest.result != 0) {
drivers/net/ethernet/intel/e100.c
669
nic->mem->selftest.result);
drivers/net/ethernet/intel/e100.c
672
if (nic->mem->selftest.signature == 0) {
drivers/net/ethernet/intel/i40e/i40e_alloc.h
25
struct i40e_dma_mem *mem,
drivers/net/ethernet/intel/i40e/i40e_alloc.h
28
struct i40e_dma_mem *mem);
drivers/net/ethernet/intel/i40e/i40e_alloc.h
30
struct i40e_virt_mem *mem,
drivers/net/ethernet/intel/i40e/i40e_alloc.h
33
struct i40e_virt_mem *mem);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1207
struct i40e_virt_mem mem;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1215
ret = i40e_allocate_virt_mem(hw, &mem, I40E_LLDPDU_SIZE);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1224
lldpmib = (u8 *)mem.va;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1228
i40e_free_virt_mem(hw, &mem);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1826
u16 mem;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1838
mem = le16_to_cpu(raw_mem);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1842
if (mem & I40E_PTR_TYPE)
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1843
address = (0x7FFF & mem) * 4096;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1845
address = (0x7FFF & mem) * 2;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1857
mem = le16_to_cpu(raw_mem);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1858
offset = mem + word_offset;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1885
u32 mem;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1894
ret = i40e_aq_read_nvm(hw, I40E_SR_NVM_CONTROL_WORD, 0, sizeof(mem),
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1895
&mem, true, NULL);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
1904
if (mem & I40E_SR_NVM_MAP_STRUCTURE_TYPE) {
drivers/net/ethernet/intel/i40e/i40e_dcb.c
534
struct i40e_virt_mem mem;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
539
ret = i40e_allocate_virt_mem(hw, &mem, I40E_LLDPDU_SIZE);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
543
lldpmib = (u8 *)mem.va;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
554
i40e_free_virt_mem(hw, &mem);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
910
struct i40e_virt_mem mem;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
918
ret = i40e_allocate_virt_mem(hw, &mem, I40E_LLDPDU_SIZE);
drivers/net/ethernet/intel/i40e/i40e_dcb.c
922
lldpmib = (u8 *)mem.va;
drivers/net/ethernet/intel/i40e/i40e_dcb.c
937
i40e_free_virt_mem(hw, &mem);
drivers/net/ethernet/intel/i40e/i40e_hmc.c
109
struct i40e_dma_mem mem;
drivers/net/ethernet/intel/i40e/i40e_hmc.c
110
struct i40e_dma_mem *page = &mem;
drivers/net/ethernet/intel/i40e/i40e_hmc.c
25
struct i40e_dma_mem mem;
drivers/net/ethernet/intel/i40e/i40e_hmc.c
49
ret_code = i40e_allocate_dma_mem(hw, &mem, alloc_len,
drivers/net/ethernet/intel/i40e/i40e_hmc.c
63
sd_entry->u.pd_table.pd_page_addr = mem;
drivers/net/ethernet/intel/i40e/i40e_hmc.c
65
sd_entry->u.bp.addr = mem;
drivers/net/ethernet/intel/i40e/i40e_hmc.c
80
i40e_free_dma_mem(hw, &mem);
drivers/net/ethernet/intel/i40e/i40e_main.c
157
int i40e_allocate_dma_mem(struct i40e_hw *hw, struct i40e_dma_mem *mem,
drivers/net/ethernet/intel/i40e/i40e_main.c
162
mem->size = ALIGN(size, alignment);
drivers/net/ethernet/intel/i40e/i40e_main.c
163
mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa,
drivers/net/ethernet/intel/i40e/i40e_main.c
165
if (!mem->va)
drivers/net/ethernet/intel/i40e/i40e_main.c
176
int i40e_free_dma_mem(struct i40e_hw *hw, struct i40e_dma_mem *mem)
drivers/net/ethernet/intel/i40e/i40e_main.c
180
dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa);
drivers/net/ethernet/intel/i40e/i40e_main.c
181
mem->va = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
182
mem->pa = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
183
mem->size = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
194
int i40e_allocate_virt_mem(struct i40e_hw *hw, struct i40e_virt_mem *mem,
drivers/net/ethernet/intel/i40e/i40e_main.c
197
mem->size = size;
drivers/net/ethernet/intel/i40e/i40e_main.c
198
mem->va = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/intel/i40e/i40e_main.c
200
if (!mem->va)
drivers/net/ethernet/intel/i40e/i40e_main.c
211
int i40e_free_virt_mem(struct i40e_hw *hw, struct i40e_virt_mem *mem)
drivers/net/ethernet/intel/i40e/i40e_main.c
214
kfree(mem->va);
drivers/net/ethernet/intel/i40e/i40e_main.c
215
mem->va = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
216
mem->size = 0;
drivers/net/ethernet/intel/iavf/iavf_alloc.h
24
struct iavf_dma_mem *mem,
drivers/net/ethernet/intel/iavf/iavf_alloc.h
28
struct iavf_dma_mem *mem);
drivers/net/ethernet/intel/iavf/iavf_alloc.h
30
struct iavf_virt_mem *mem, u32 size);
drivers/net/ethernet/intel/iavf/iavf_alloc.h
31
void iavf_free_virt_mem(struct iavf_hw *hw, struct iavf_virt_mem *mem);
drivers/net/ethernet/intel/iavf/iavf_main.c
196
struct iavf_dma_mem *mem,
drivers/net/ethernet/intel/iavf/iavf_main.c
201
if (!mem)
drivers/net/ethernet/intel/iavf/iavf_main.c
204
mem->size = ALIGN(size, alignment);
drivers/net/ethernet/intel/iavf/iavf_main.c
205
mem->va = dma_alloc_coherent(&adapter->pdev->dev, mem->size,
drivers/net/ethernet/intel/iavf/iavf_main.c
206
(dma_addr_t *)&mem->pa, GFP_KERNEL);
drivers/net/ethernet/intel/iavf/iavf_main.c
207
if (mem->va)
drivers/net/ethernet/intel/iavf/iavf_main.c
218
enum iavf_status iavf_free_dma_mem(struct iavf_hw *hw, struct iavf_dma_mem *mem)
drivers/net/ethernet/intel/iavf/iavf_main.c
222
if (!mem || !mem->va)
drivers/net/ethernet/intel/iavf/iavf_main.c
224
dma_free_coherent(&adapter->pdev->dev, mem->size,
drivers/net/ethernet/intel/iavf/iavf_main.c
225
mem->va, (dma_addr_t)mem->pa);
drivers/net/ethernet/intel/iavf/iavf_main.c
236
struct iavf_virt_mem *mem, u32 size)
drivers/net/ethernet/intel/iavf/iavf_main.c
238
if (!mem)
drivers/net/ethernet/intel/iavf/iavf_main.c
241
mem->size = size;
drivers/net/ethernet/intel/iavf/iavf_main.c
242
mem->va = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/intel/iavf/iavf_main.c
244
if (mem->va)
drivers/net/ethernet/intel/iavf/iavf_main.c
255
void iavf_free_virt_mem(struct iavf_hw *hw, struct iavf_virt_mem *mem)
drivers/net/ethernet/intel/iavf/iavf_main.c
257
kfree(mem->va);
drivers/net/ethernet/intel/idpf/idpf_controlq.h
141
void *idpf_alloc_dma_mem(struct idpf_hw *hw, struct idpf_dma_mem *mem,
drivers/net/ethernet/intel/idpf/idpf_controlq.h
143
void idpf_free_dma_mem(struct idpf_hw *hw, struct idpf_dma_mem *mem);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2579
void *idpf_alloc_dma_mem(struct idpf_hw *hw, struct idpf_dma_mem *mem, u64 size)
drivers/net/ethernet/intel/idpf/idpf_lib.c
2588
mem->va = dma_alloc_attrs(&adapter->pdev->dev, sz, &mem->pa,
drivers/net/ethernet/intel/idpf/idpf_lib.c
2590
mem->size = sz;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2592
return mem->va;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2600
void idpf_free_dma_mem(struct idpf_hw *hw, struct idpf_dma_mem *mem)
drivers/net/ethernet/intel/idpf/idpf_lib.c
2604
dma_free_attrs(&adapter->pdev->dev, mem->size,
drivers/net/ethernet/intel/idpf/idpf_lib.c
2605
mem->va, mem->pa, DMA_ATTR_FORCE_CONTIGUOUS);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2606
mem->size = 0;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2607
mem->va = NULL;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2608
mem->pa = 0;
drivers/net/ethernet/intel/libeth/xdp.c
304
if (xdp->base.rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL)
drivers/net/ethernet/intel/libie/fwlog.c
102
ring->data = mem;
drivers/net/ethernet/intel/libie/fwlog.c
103
mem += LIBIE_AQ_MAX_BUF_LEN;
drivers/net/ethernet/intel/libie/fwlog.c
91
u8 *mem;
drivers/net/ethernet/intel/libie/fwlog.c
94
mem = vzalloc(nr_bytes);
drivers/net/ethernet/intel/libie/fwlog.c
95
if (!mem)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
779
struct nix_sqe_mem_s *mem;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
781
mem = (struct nix_sqe_mem_s *)(sq->sqe_base + *offset);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
782
mem->subdc = NIX_SUBDC_MEM;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
783
mem->alg = alg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
784
mem->wmem = 1; /* wait for the memory operation */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
785
mem->addr = iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
788
mem->start_offset = ptp_offset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
789
mem->udp_csum_crt = !!udp_csum_crt;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
790
mem->base_ns = base_ns;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
791
mem->step_type = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
794
*offset += sizeof(*mem);
drivers/net/ethernet/marvell/skge.c
2520
u64 base = skge->dma + (e->desc - skge->mem);
drivers/net/ethernet/marvell/skge.c
2555
skge->mem = dma_alloc_coherent(&hw->pdev->dev, skge->mem_size,
drivers/net/ethernet/marvell/skge.c
2557
if (!skge->mem)
drivers/net/ethernet/marvell/skge.c
2568
err = skge_ring_alloc(&skge->rx_ring, skge->mem, skge->dma);
drivers/net/ethernet/marvell/skge.c
2576
err = skge_ring_alloc(&skge->tx_ring, skge->mem + rx_size,
drivers/net/ethernet/marvell/skge.c
2634
dma_free_coherent(&hw->pdev->dev, skge->mem_size, skge->mem,
drivers/net/ethernet/marvell/skge.c
2636
skge->mem = NULL;
drivers/net/ethernet/marvell/skge.c
2656
if (!skge->mem)
drivers/net/ethernet/marvell/skge.c
2725
dma_free_coherent(&hw->pdev->dev, skge->mem_size, skge->mem,
drivers/net/ethernet/marvell/skge.c
2727
skge->mem = NULL;
drivers/net/ethernet/marvell/skge.h
2466
void *mem; /* PCI memory for rings */
drivers/net/ethernet/mellanox/mlx4/icm.c
110
sg_set_page(mem, page, PAGE_SIZE << order, 0);
drivers/net/ethernet/mellanox/mlx4/icm.c
98
static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
80
if (xdp->rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL) {
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5598
static void mlx5e_queue_mem_free(struct net_device *dev, void *mem)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5600
struct mlx5_qmgmt_data *data = (struct mlx5_qmgmt_data *)mem;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/internal.h
44
static inline bool is_mem_zero(const u8 *mem, size_t size)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/internal.h
51
return (*mem == 0) && memcmp(mem, mem + 1, size - 1) == 0;
drivers/net/ethernet/microchip/sparx5/sparx5_port.c
287
char *mem = "";
drivers/net/ethernet/microchip/sparx5/sparx5_port.c
305
mem = resource == 0 ?
drivers/net/ethernet/microchip/sparx5/sparx5_port.c
318
portno, mem);
drivers/net/ethernet/netronome/nfp/bpf/main.c
340
u8 __iomem *mem, *start;
drivers/net/ethernet/netronome/nfp/bpf/main.c
342
mem = nfp_rtsym_map(app->pf->rtbl, "_abi_bpf_capabilities", "bpf.cap",
drivers/net/ethernet/netronome/nfp/bpf/main.c
344
if (IS_ERR(mem))
drivers/net/ethernet/netronome/nfp/bpf/main.c
345
return PTR_ERR(mem) == -ENOENT ? 0 : PTR_ERR(mem);
drivers/net/ethernet/netronome/nfp/bpf/main.c
347
start = mem;
drivers/net/ethernet/netronome/nfp/bpf/main.c
348
while (mem - start + 8 <= nfp_cpp_area_size(area)) {
drivers/net/ethernet/netronome/nfp/bpf/main.c
352
type = readl(mem);
drivers/net/ethernet/netronome/nfp/bpf/main.c
353
length = readl(mem + 4);
drivers/net/ethernet/netronome/nfp/bpf/main.c
354
value = mem + 8;
drivers/net/ethernet/netronome/nfp/bpf/main.c
356
mem += 8 + length;
drivers/net/ethernet/netronome/nfp/bpf/main.c
357
if (mem - start > nfp_cpp_area_size(area))
drivers/net/ethernet/netronome/nfp/bpf/main.c
402
if (mem - start != nfp_cpp_area_size(area)) {
drivers/net/ethernet/netronome/nfp/bpf/main.c
404
mem - start, nfp_cpp_area_size(area));
drivers/net/ethernet/netronome/nfp/bpf/main.c
413
nfp_err(cpp, "invalid BPF capabilities at offset:%zd\n", mem - start);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1000
mem = nn->dp.ctrl_bar + nn->tlv_caps.vnic_stats_off;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1003
id_word = readq(mem + i * 2);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1029
u8 __iomem *mem;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1032
mem = nn->dp.ctrl_bar + nn->tlv_caps.vnic_stats_off;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1033
mem += roundup(2 * nn->tlv_caps.vnic_stats_cnt, 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1035
*data++ = readq(mem + i * 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1037
mem = nn->dp.ctrl_bar;
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1039
*data++ = readq(mem + NFP_NET_CFG_RXR_STATS(i));
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1040
*data++ = readq(mem + NFP_NET_CFG_RXR_STATS(i) + 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1041
*data++ = readq(mem + NFP_NET_CFG_TXR_STATS(i));
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
1042
*data++ = readq(mem + NFP_NET_CFG_TXR_STATS(i) + 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
972
nfp_vnic_get_hw_stats(u64 *data, u8 __iomem *mem, unsigned int num_vecs)
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
977
*data++ = readq(mem + nfp_net_et_stats[i].off);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
980
*data++ = readq(mem + NFP_NET_CFG_RXR_STATS(i));
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
981
*data++ = readq(mem + NFP_NET_CFG_RXR_STATS(i) + 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
982
*data++ = readq(mem + NFP_NET_CFG_TXR_STATS(i));
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
983
*data++ = readq(mem + NFP_NET_CFG_TXR_STATS(i) + 8);
drivers/net/ethernet/netronome/nfp/nfp_net_ethtool.c
997
u8 __iomem *mem;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
440
u8 __iomem *mem;
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
444
mem = nfp_pf_map_rtsym(pf, "net.bar0", "_pf%d_net_bar0",
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
446
if (IS_ERR(mem)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
448
return PTR_ERR(mem);
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
489
mem = nfp_cpp_map_area(pf->cpp, "net.qc", cpp_id,
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
492
if (IS_ERR(mem)) {
drivers/net/ethernet/netronome/nfp/nfp_net_main.c
494
err = PTR_ERR(mem);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
62
u8 __iomem *mem = port->eth_stats;
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
64
stats->tx_packets = readq(mem + NFP_MAC_STATS_TX_FRAMES_TRANSMITTED_OK);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
65
stats->tx_bytes = readq(mem + NFP_MAC_STATS_TX_OUT_OCTETS);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
66
stats->tx_dropped = readq(mem + NFP_MAC_STATS_TX_OUT_ERRORS);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
68
stats->rx_packets = readq(mem + NFP_MAC_STATS_RX_FRAMES_RECEIVED_OK);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
69
stats->rx_bytes = readq(mem + NFP_MAC_STATS_RX_IN_OCTETS);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
70
stats->rx_dropped = readq(mem + NFP_MAC_STATS_RX_IN_ERRORS);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
527
u8 __iomem *mem;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
548
mem = nfp_cpp_map_area(rtbl->cpp, id, cpp_id, addr, sym->size, area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
549
if (IS_ERR(mem)) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
551
name, PTR_ERR(mem));
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
552
return mem;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c
555
return mem;
drivers/net/ethernet/qlogic/qed/qed_debug.c
2724
const struct dbg_dump_mem *mem =
drivers/net/ethernet/qlogic/qed/qed_debug.c
2734
mem_group_id = GET_FIELD(mem->dword0,
drivers/net/ethernet/qlogic/qed/qed_debug.c
2747
mem_addr = GET_FIELD(mem->dword0, DBG_DUMP_MEM_ADDRESS);
drivers/net/ethernet/qlogic/qed/qed_debug.c
2748
mem_len = GET_FIELD(mem->dword1, DBG_DUMP_MEM_LENGTH);
drivers/net/ethernet/qlogic/qed/qed_debug.c
2749
mem_wide_bus = GET_FIELD(mem->dword1,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
1020
struct __mem *rom = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
234
struct __mem mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
563
struct __mem *rom = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
661
struct __mem *mem, __le32 *buffer,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
667
reg_read = mem->size;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
668
addr = mem->addr;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
709
return mem->size;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
721
struct __mem *mem)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
734
ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
744
ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
768
struct __mem *mem,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
809
temp = mem->dma_desc_cmd & 0xff0f;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
817
while (read_size < mem->size) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
818
if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
821
size = mem->size - read_size;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
823
dma_descr->src_addr_low = mem->addr + read_size;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
828
*ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
832
mem->desc_card_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
836
*ret = qlcnic_start_pex_dma(adapter, mem);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
858
struct __mem *mem = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
863
data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
873
data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret);
drivers/net/ethernet/realtek/8139cp.c
1116
void *mem;
drivers/net/ethernet/realtek/8139cp.c
1119
mem = dma_alloc_coherent(d, CP_RING_BYTES, &cp->ring_dma, GFP_KERNEL);
drivers/net/ethernet/realtek/8139cp.c
1120
if (!mem)
drivers/net/ethernet/realtek/8139cp.c
1123
cp->rx_ring = mem;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5166
bool zc = !!(xdp->rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL);
drivers/net/ethernet/sun/sunbmac.h
273
#define bib_offset(mem, elem) \
drivers/net/ethernet/sun/sunbmac.h
274
((__u32)((unsigned long)(&(((struct bmac_init_block *)0)->mem[elem]))))
drivers/net/ethernet/sun/sunhme.h
380
#define hblock_offset(mem, elem) \
drivers/net/ethernet/sun/sunhme.h
381
((__u32)((unsigned long)(&(((struct hmeal_init_block *)0)->mem[elem]))))
drivers/net/ethernet/sun/sunqe.h
309
#define qib_offset(mem, elem) \
drivers/net/ethernet/sun/sunqe.h
310
((__u32)((unsigned long)(&(((struct qe_init_block *)0)->mem[elem]))))
drivers/net/ethernet/sun/sunqe.h
331
#define qebuf_offset(mem, elem) \
drivers/net/ethernet/sun/sunqe.h
332
((__u32)((unsigned long)(&(((struct sunqe_buffers *)0)->mem[elem][0]))))
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1272
ram = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1291
ram = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1307
ram = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1321
ram = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1371
sram = prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
173
void __iomem *sram_base = prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1920
&prueth->mem[i]);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1940
prueth->mem[PRUETH_MEM_OCMC].va =
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1943
if (!prueth->mem[PRUETH_MEM_OCMC].va) {
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1948
prueth->mem[PRUETH_MEM_OCMC].pa = gen_pool_virt_to_phys
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1950
prueth->mem[PRUETH_MEM_OCMC].va);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1951
prueth->mem[PRUETH_MEM_OCMC].size = prueth->ocmc_ram_size;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1953
&prueth->mem[PRUETH_MEM_OCMC].pa,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1954
prueth->mem[PRUETH_MEM_OCMC].va,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
1955
prueth->mem[PRUETH_MEM_OCMC].size);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2047
(unsigned long)prueth->mem[PRUETH_MEM_OCMC].va,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2052
if (prueth->mem[i].va)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2053
pruss_release_mem_region(pruss, &prueth->mem[i]);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2095
(unsigned long)prueth->mem[PRUETH_MEM_OCMC].va,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2099
if (prueth->mem[i].va)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
2101
&prueth->mem[i]);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
285
memset_io(prueth->mem[region].va, 0, prueth->mem[region].size);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
326
ram = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
343
ocmcaddr = (u32)prueth->mem[PRUETH_MEM_OCMC].pa;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
344
sram = prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
349
dram_base = prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
434
writeb(port_status, prueth->mem[region].va +
drivers/net/ethernet/ti/icssm/icssm_prueth.c
47
writel_relaxed(val, prueth->mem[region].va + reg);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
528
dram = prueth->mem[PRUETH_MEM_DRAM1].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
530
dram = emac->prueth->mem[emac->dram].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
580
ocmc_ram = (__force void *)emac->prueth->mem[PRUETH_MEM_OCMC].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
612
sram = prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
703
ocmc_ram = (__force void *)emac->prueth->mem[PRUETH_MEM_OCMC].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
774
shared_ram = emac->prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
924
void __iomem *sram = emac->prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth.h
245
struct pruss_mem_region mem[PRUETH_MEM_MAX];
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
217
sram_base = prueth->mem[PRUETH_MEM_SHARED_RAM].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
854
void __iomem *dram1_base = prueth->mem[PRUETH_MEM_DRAM1].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
897
dram1_base = prueth->mem[PRUETH_MEM_DRAM1].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
906
dram_mac = prueth->mem[PRUETH_MEM_DRAM0].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
914
dram_mac = prueth->mem[PRUETH_MEM_DRAM1].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
925
dram_base = prueth->mem[PRUETH_MEM_DRAM1].va;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
987
ocmcaddr = (u32)prueth->mem[PRUETH_MEM_OCMC].pa;
drivers/net/ethernet/wiznet/w5100.c
1042
struct resource *mem;
drivers/net/ethernet/wiznet/w5100.c
1049
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/net/ethernet/wiznet/w5100.c
1050
if (!mem)
drivers/net/ethernet/wiznet/w5100.c
1052
if (resource_size(mem) < W5100_BUS_DIRECT_SIZE)
drivers/net/ethernet/wiznet/w5300.c
531
struct resource *mem;
drivers/net/ethernet/wiznet/w5300.c
542
priv->base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/net/ethernet/wiznet/w5300.c
546
mem_size = resource_size(mem);
drivers/net/ethernet/wiznet/w5300.c
584
netdev_info(ndev, "at 0x%llx irq %d\n", (u64)mem->start, irq);
drivers/net/ethernet/xscale/ixp4xx_eth.c
864
void *mem;
drivers/net/ethernet/xscale/ixp4xx_eth.c
884
mem = skb->data;
drivers/net/ethernet/xscale/ixp4xx_eth.c
888
if (!(mem = kmalloc(bytes, GFP_ATOMIC))) {
drivers/net/ethernet/xscale/ixp4xx_eth.c
893
memcpy_swab32(mem, (u32 *)((uintptr_t)skb->data & ~3), bytes / 4);
drivers/net/ethernet/xscale/ixp4xx_eth.c
896
phys = dma_map_single(&dev->dev, mem, bytes, DMA_TO_DEVICE);
drivers/net/ethernet/xscale/ixp4xx_eth.c
900
kfree(mem);
drivers/net/ethernet/xscale/ixp4xx_eth.c
913
port->tx_buff_tab[n] = mem;
drivers/net/fddi/defxx.c
366
writel(data, bp->base.mem + offset);
drivers/net/fddi/defxx.c
3725
iounmap(bp->base.mem);
drivers/net/fddi/defxx.c
389
*data = readl(bp->base.mem + offset);
drivers/net/fddi/defxx.c
621
bp->base.mem = ioremap(bar_start[0], bar_len[0]);
drivers/net/fddi/defxx.c
622
if (!bp->base.mem) {
drivers/net/fddi/defxx.c
664
iounmap(bp->base.mem);
drivers/net/fddi/defxx.h
1776
void __iomem *mem;
drivers/net/fddi/skfp/skfddi.c
208
void __iomem *mem;
drivers/net/fddi/skfp/skfddi.c
233
mem = ioremap(pci_resource_start(pdev, 0), 0x4000);
drivers/net/fddi/skfp/skfddi.c
241
mem = ioport_map(pci_resource_start(pdev, 1), FP_IO_LEN);
drivers/net/fddi/skfp/skfddi.c
243
if (!mem) {
drivers/net/fddi/skfp/skfddi.c
272
smc->hw.iop = mem;
drivers/net/fddi/skfp/skfddi.c
276
dev->base_addr = (unsigned long)mem;
drivers/net/fddi/skfp/skfddi.c
309
iounmap(mem);
drivers/net/fddi/skfp/skfddi.c
311
ioport_unmap(mem);
drivers/net/fjes/fjes_hw.c
115
void *mem;
drivers/net/fjes/fjes_hw.c
117
mem = vzalloc(EP_BUFFER_SIZE);
drivers/net/fjes/fjes_hw.c
118
if (!mem)
drivers/net/fjes/fjes_hw.c
121
epbh->buffer = mem;
drivers/net/fjes/fjes_hw.c
124
epbh->info = (union ep_buffer_info *)mem;
drivers/net/fjes/fjes_hw.c
125
epbh->ring = (u8 *)(mem + sizeof(union ep_buffer_info));
drivers/net/ipa/ipa.h
101
const struct ipa_mem *mem;
drivers/net/ipa/ipa_cmd.c
172
bool ipa_cmd_table_init_valid(struct ipa *ipa, const struct ipa_mem *mem,
drivers/net/ipa/ipa_cmd.c
193
if (mem->offset > offset_max ||
drivers/net/ipa/ipa_cmd.c
194
ipa->mem_offset > offset_max - mem->offset) {
drivers/net/ipa/ipa_cmd.c
197
ipa->mem_offset, mem->offset, offset_max);
drivers/net/ipa/ipa_cmd.c
209
const struct ipa_mem *mem;
drivers/net/ipa/ipa_cmd.c
226
mem = ipa_mem_find(ipa, IPA_MEM_MODEM_HEADER);
drivers/net/ipa/ipa_cmd.c
227
offset = mem->offset;
drivers/net/ipa/ipa_cmd.c
228
size = mem->size;
drivers/net/ipa/ipa_cmd.c
240
mem = ipa_mem_find(ipa, IPA_MEM_AP_HEADER);
drivers/net/ipa/ipa_cmd.c
241
if (mem)
drivers/net/ipa/ipa_cmd.c
242
size += mem->size;
drivers/net/ipa/ipa_cmd.h
53
bool ipa_cmd_table_init_valid(struct ipa *ipa, const struct ipa_mem *mem,
drivers/net/ipa/ipa_mem.c
100
mem = ipa_mem_find(ipa, IPA_MEM_MODEM_HEADER);
drivers/net/ipa/ipa_mem.c
101
offset = mem->offset;
drivers/net/ipa/ipa_mem.c
102
size = mem->size;
drivers/net/ipa/ipa_mem.c
103
mem = ipa_mem_find(ipa, IPA_MEM_AP_HEADER);
drivers/net/ipa/ipa_mem.c
104
if (mem)
drivers/net/ipa/ipa_mem.c
105
size += mem->size;
drivers/net/ipa/ipa_mem.c
116
mem = ipa_mem_find(ipa, IPA_MEM_MODEM_PROC_CTX);
drivers/net/ipa/ipa_mem.c
117
offset = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_mem.c
220
static bool ipa_mem_valid_one(struct ipa *ipa, const struct ipa_mem *mem)
drivers/net/ipa/ipa_mem.c
222
enum ipa_mem_id mem_id = mem->id;
drivers/net/ipa/ipa_mem.c
232
if (!mem->size && !mem->canary_count) {
drivers/net/ipa/ipa_mem.c
239
if (mem->size % size_multiple)
drivers/net/ipa/ipa_mem.c
242
else if (mem->offset % 8)
drivers/net/ipa/ipa_mem.c
244
else if (mem->offset < mem->canary_count * sizeof(__le32))
drivers/net/ipa/ipa_mem.c
246
mem_id, mem->canary_count);
drivers/net/ipa/ipa_mem.c
247
else if (mem_id == IPA_MEM_END_MARKER && mem->size)
drivers/net/ipa/ipa_mem.c
270
const struct ipa_mem *mem = &mem_data->local[i];
drivers/net/ipa/ipa_mem.c
272
if (__test_and_set_bit(mem->id, regions)) {
drivers/net/ipa/ipa_mem.c
273
dev_err(dev, "duplicate memory region %u\n", mem->id);
drivers/net/ipa/ipa_mem.c
278
if (!ipa_mem_valid_one(ipa, mem))
drivers/net/ipa/ipa_mem.c
300
const struct ipa_mem *mem = &ipa->mem[i];
drivers/net/ipa/ipa_mem.c
302
if (mem->offset + mem->size <= limit)
drivers/net/ipa/ipa_mem.c
306
mem->id, limit);
drivers/net/ipa/ipa_mem.c
323
const struct ipa_mem *mem;
drivers/net/ipa/ipa_mem.c
36
const struct ipa_mem *mem = &ipa->mem[i];
drivers/net/ipa/ipa_mem.c
367
u16 canary_count = ipa->mem[i].canary_count;
drivers/net/ipa/ipa_mem.c
374
canary = ipa->mem_virt + ipa->mem_offset + ipa->mem[i].offset;
drivers/net/ipa/ipa_mem.c
38
if (mem->id == mem_id)
drivers/net/ipa/ipa_mem.c
381
mem = ipa_mem_find(ipa, IPA_MEM_UC_EVENT_RING);
drivers/net/ipa/ipa_mem.c
382
if (mem && mem->offset % 1024) {
drivers/net/ipa/ipa_mem.c
39
return mem;
drivers/net/ipa/ipa_mem.c
50
const struct ipa_mem *mem = ipa_mem_find(ipa, mem_id);
drivers/net/ipa/ipa_mem.c
53
if (!mem->size)
drivers/net/ipa/ipa_mem.c
56
ipa_cmd_dma_shared_mem_add(trans, mem->offset, mem->size, addr, true);
drivers/net/ipa/ipa_mem.c
629
ipa->mem = mem_data->local;
drivers/net/ipa/ipa_mem.c
80
const struct ipa_mem *mem;
drivers/net/ipa/ipa_qmi.c
285
const struct ipa_mem *mem;
drivers/net/ipa/ipa_qmi.c
298
mem = ipa_mem_find(ipa, IPA_MEM_MODEM_HEADER);
drivers/net/ipa/ipa_qmi.c
299
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
301
req.hdr_tbl_info.start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
302
req.hdr_tbl_info.end = req.hdr_tbl_info.start + mem->size - 1;
drivers/net/ipa/ipa_qmi.c
305
mem = ipa_mem_find(ipa, IPA_MEM_V4_ROUTE);
drivers/net/ipa/ipa_qmi.c
307
req.v4_route_tbl_info.start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
310
mem = ipa_mem_find(ipa, IPA_MEM_V6_ROUTE);
drivers/net/ipa/ipa_qmi.c
312
req.v6_route_tbl_info.start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
315
mem = ipa_mem_find(ipa, IPA_MEM_V4_FILTER);
drivers/net/ipa/ipa_qmi.c
317
req.v4_filter_tbl_start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
319
mem = ipa_mem_find(ipa, IPA_MEM_V6_FILTER);
drivers/net/ipa/ipa_qmi.c
321
req.v6_filter_tbl_start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
323
mem = ipa_mem_find(ipa, IPA_MEM_MODEM);
drivers/net/ipa/ipa_qmi.c
324
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
326
req.modem_mem_info.start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
327
req.modem_mem_info.size = mem->size;
drivers/net/ipa/ipa_qmi.c
336
mem = ipa_mem_find(ipa, IPA_MEM_MODEM_PROC_CTX);
drivers/net/ipa/ipa_qmi.c
337
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
340
ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
342
req.hdr_proc_ctx_tbl_info.start + mem->size - 1;
drivers/net/ipa/ipa_qmi.c
347
mem = ipa_mem_find(ipa, IPA_MEM_V4_ROUTE_HASHED);
drivers/net/ipa/ipa_qmi.c
348
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
351
ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
355
mem = ipa_mem_find(ipa, IPA_MEM_V6_ROUTE_HASHED);
drivers/net/ipa/ipa_qmi.c
356
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
359
ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
363
mem = ipa_mem_find(ipa, IPA_MEM_V4_FILTER_HASHED);
drivers/net/ipa/ipa_qmi.c
364
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
366
req.v4_hash_filter_tbl_start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
369
mem = ipa_mem_find(ipa, IPA_MEM_V6_FILTER_HASHED);
drivers/net/ipa/ipa_qmi.c
370
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
372
req.v6_hash_filter_tbl_start = ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
377
mem = ipa_mem_find(ipa, IPA_MEM_STATS_QUOTA_MODEM);
drivers/net/ipa/ipa_qmi.c
378
if (mem->size) {
drivers/net/ipa/ipa_qmi.c
381
ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
383
req.hw_stats_quota_size = ipa->mem_offset + mem->size;
drivers/net/ipa/ipa_qmi.c
387
mem = ipa_mem_find(ipa, IPA_MEM_STATS_DROP);
drivers/net/ipa/ipa_qmi.c
388
if (mem && mem->size) {
drivers/net/ipa/ipa_qmi.c
391
ipa->mem_offset + mem->offset;
drivers/net/ipa/ipa_qmi.c
393
req.hw_stats_drop_size = ipa->mem_offset + mem->size;
drivers/net/ipa/ipa_table.c
209
const struct ipa_mem *mem;
drivers/net/ipa/ipa_table.c
215
mem = ipa_table_mem(ipa, filter, hashed, ipv6);
drivers/net/ipa/ipa_table.c
216
if (!mem || !mem->size)
drivers/net/ipa/ipa_table.c
222
offset = mem->offset + first * sizeof(__le64);
drivers/net/ipa/ipa_table.c
391
const struct ipa_mem *mem;
drivers/net/ipa/ipa_table.c
408
mem = ipa_table_mem(ipa, filter, false, ipv6);
drivers/net/ipa/ipa_table.c
425
count = mem->size / sizeof(__le64);
drivers/net/ipa/ipa_table.c
434
ipa_cmd_table_init_add(trans, opcode, size, mem->offset, addr,
drivers/net/ipa/ipa_table.c
440
zero_offset = mem->offset + size;
drivers/net/ipa/ipa_table.c
441
zero_size = mem->size - size;
drivers/net/ipa/ipa_uc.c
122
const struct ipa_mem *mem = ipa_mem_find(ipa, IPA_MEM_UC_SHARED);
drivers/net/ipa/ipa_uc.c
123
u32 offset = ipa->mem_offset + mem->offset;
drivers/net/veth.c
1132
rq->xdp_mem = rq->xdp_rxq.mem;
drivers/net/veth.c
1159
rq->xdp_rxq.mem = rq->xdp_mem;
drivers/net/veth.c
666
xdp->rxq->mem.type = frame->mem_type;
drivers/net/veth.c
678
xdp->rxq->mem.type = frame->mem_type;
drivers/net/veth.c
827
xdp->rxq->mem = rq->xdp_mem;
drivers/net/veth.c
839
xdp->rxq->mem = rq->xdp_mem;
drivers/net/wan/farsync.c
1237
card->mem + BUF_OFFSET(rxBuffer[pi][rxp][0]),
drivers/net/wan/farsync.c
1331
memcpy_toio(card->mem +
drivers/net/wan/farsync.c
1964
memcpy_toio(card->mem + wrthdr.offset, buf, wrthdr.size);
drivers/net/wan/farsync.c
2383
card->mem = ioremap(card->phys_mem, FST_MEMSIZE);
drivers/net/wan/farsync.c
2384
if (!card->mem) {
drivers/net/wan/farsync.c
2395
dbg(DBG_PCI, "kernel mem %p, ctlmem %p\n", card->mem, card->ctlmem);
drivers/net/wan/farsync.c
2524
iounmap(card->mem);
drivers/net/wan/farsync.c
2557
iounmap(card->mem);
drivers/net/wan/farsync.c
446
char __iomem *mem; /* Card memory mapped to kernel space */
drivers/net/wan/farsync.c
492
#define FST_RDB(C, E) (readb((C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
493
#define FST_RDW(C, E) (readw((C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
494
#define FST_RDL(C, E) (readl((C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
496
#define FST_WRB(C, E, B) (writeb((B), (C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
497
#define FST_WRW(C, E, W) (writew((W), (C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
498
#define FST_WRL(C, E, L) (writel((L), (C)->mem + WIN_OFFSET(E)))
drivers/net/wan/farsync.c
711
(void)readb(card->mem);
drivers/net/wan/farsync.c
858
fst_rx_dma(struct fst_card_info *card, dma_addr_t dma, u32 mem, int len)
drivers/net/wan/farsync.c
863
dbg(DBG_RX, "In fst_rx_dma %x %x %d\n", (u32)dma, mem, len);
drivers/net/wan/farsync.c
868
outl(mem, card->pci_conf + DMALADR0); /* from here */
drivers/net/wan/farsync.c
881
fst_tx_dma(struct fst_card_info *card, dma_addr_t dma, u32 mem, int len)
drivers/net/wan/farsync.c
886
dbg(DBG_TX, "In fst_tx_dma %x %x %d\n", (u32)dma, mem, len);
drivers/net/wan/farsync.c
891
outl(mem, card->pci_conf + DMALADR1); /* to here */
drivers/net/wan/ixp4xx_hss.c
842
void *mem;
drivers/net/wan/ixp4xx_hss.c
862
mem = skb->data;
drivers/net/wan/ixp4xx_hss.c
866
mem = kmalloc(bytes, GFP_ATOMIC);
drivers/net/wan/ixp4xx_hss.c
867
if (!mem) {
drivers/net/wan/ixp4xx_hss.c
872
memcpy_swab32(mem, (u32 *)((uintptr_t)skb->data & ~3), bytes / 4);
drivers/net/wan/ixp4xx_hss.c
876
phys = dma_map_single(&dev->dev, mem, bytes, DMA_TO_DEVICE);
drivers/net/wan/ixp4xx_hss.c
881
kfree(mem);
drivers/net/wan/ixp4xx_hss.c
894
port->tx_buff_tab[n] = mem;
drivers/net/wan/wanxl.c
557
u8 __iomem *mem; /* memory virtual base addr */
drivers/net/wan/wanxl.c
707
mem = ioremap(mem_phy, PDM_OFFSET + sizeof(firmware));
drivers/net/wan/wanxl.c
708
if (!mem) {
drivers/net/wan/wanxl.c
715
writel(ntohl(*(__be32 *)(firmware + i)), mem + PDM_OFFSET + i);
drivers/net/wan/wanxl.c
720
(void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
drivers/net/wan/wanxl.c
721
writel(card->status_address, mem + PDM_OFFSET + 20);
drivers/net/wan/wanxl.c
722
writel(PDM_OFFSET, mem);
drivers/net/wan/wanxl.c
723
iounmap(mem);
drivers/net/wireguard/selftest/allowedips.c
459
#define insert(version, mem, ipa, ipb, ipc, ipd, cidr) \
drivers/net/wireguard/selftest/allowedips.c
461
cidr, mem, &mutex)
drivers/net/wireguard/selftest/allowedips.c
463
#define remove(version, mem, ipa, ipb, ipc, ipd, cidr) \
drivers/net/wireguard/selftest/allowedips.c
465
cidr, mem, &mutex)
drivers/net/wireguard/selftest/allowedips.c
475
#define test(version, mem, ipa, ipb, ipc, ipd) do { \
drivers/net/wireguard/selftest/allowedips.c
477
ip##version(ipa, ipb, ipc, ipd)) == (mem); \
drivers/net/wireguard/selftest/allowedips.c
481
#define test_negative(version, mem, ipa, ipb, ipc, ipd) do { \
drivers/net/wireguard/selftest/allowedips.c
483
ip##version(ipa, ipb, ipc, ipd)) != (mem); \
drivers/net/wireless/ath/ath10k/ahb.c
37
iowrite32(value, ar_ahb->mem + offset);
drivers/net/wireless/ath/ath10k/ahb.c
44
return ioread32(ar_ahb->mem + offset);
drivers/net/wireless/ath/ath10k/ahb.c
445
ar_ahb->mem = devm_platform_get_and_ioremap_resource(pdev, 0, &res);
drivers/net/wireless/ath/ath10k/ahb.c
446
if (IS_ERR(ar_ahb->mem)) {
drivers/net/wireless/ath/ath10k/ahb.c
448
ret = PTR_ERR(ar_ahb->mem);
drivers/net/wireless/ath/ath10k/ahb.c
501
ar_ahb->mem, ar_ahb->mem_len,
drivers/net/wireless/ath/ath10k/ahb.c
517
devm_iounmap(&pdev->dev, ar_ahb->mem);
drivers/net/wireless/ath/ath10k/ahb.c
520
ar_ahb->mem = NULL;
drivers/net/wireless/ath/ath10k/ahb.c
531
if (ar_ahb->mem)
drivers/net/wireless/ath/ath10k/ahb.c
532
devm_iounmap(dev, ar_ahb->mem);
drivers/net/wireless/ath/ath10k/ahb.c
540
ar_ahb->mem = NULL;
drivers/net/wireless/ath/ath10k/ahb.c
763
ar_pci->mem = ar_ahb->mem;
drivers/net/wireless/ath/ath10k/ahb.h
14
void __iomem *mem;
drivers/net/wireless/ath/ath10k/pci.c
1592
base_addr = ioread32(ar_pci->mem + QCA99X0_PCIE_BAR0_START_REG);
drivers/net/wireless/ath/ath10k/pci.c
1596
iowrite32(base_addr + i, ar_pci->mem + QCA99X0_CPU_MEM_ADDR_REG);
drivers/net/wireless/ath/ath10k/pci.c
1597
*(u32 *)(buf + i) = ioread32(ar_pci->mem + QCA99X0_CPU_MEM_DATA_REG);
drivers/net/wireless/ath/ath10k/pci.c
1620
*(u32 *)(buf + i) = ioread32(ar_pci->mem + region->start + i);
drivers/net/wireless/ath/ath10k/pci.c
3409
ar_pci->mem = pci_iomap(pdev, BAR_NUM, 0);
drivers/net/wireless/ath/ath10k/pci.c
3410
if (!ar_pci->mem) {
drivers/net/wireless/ath/ath10k/pci.c
3416
ath10k_dbg(ar, ATH10K_DBG_BOOT, "boot pci_mem 0x%p\n", ar_pci->mem);
drivers/net/wireless/ath/ath10k/pci.c
3433
pci_iounmap(pdev, ar_pci->mem);
drivers/net/wireless/ath/ath10k/pci.c
434
u32 val = ioread32(ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS +
drivers/net/wireless/ath/ath10k/pci.c
450
ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS +
drivers/net/wireless/ath/ath10k/pci.c
464
ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS +
drivers/net/wireless/ath/ath10k/pci.c
505
ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS +
drivers/net/wireless/ath/ath10k/pci.c
526
ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS +
drivers/net/wireless/ath/ath10k/pci.c
649
iowrite32(value, ar_pci->mem + offset);
drivers/net/wireless/ath/ath10k/pci.c
672
val = ioread32(ar_pci->mem + offset);
drivers/net/wireless/ath/ath10k/pci.h
112
void __iomem *mem;
drivers/net/wireless/ath/ath10k/sdio.c
1734
void *mem;
drivers/net/wireless/ath/ath10k/sdio.c
1736
mem = kzalloc(buf_len, GFP_KERNEL);
drivers/net/wireless/ath/ath10k/sdio.c
1737
if (!mem)
drivers/net/wireless/ath/ath10k/sdio.c
1748
ret = ath10k_sdio_read(ar, MBOX_WINDOW_DATA_ADDRESS, mem, buf_len);
drivers/net/wireless/ath/ath10k/sdio.c
1755
memcpy(buf, mem, buf_len);
drivers/net/wireless/ath/ath10k/sdio.c
1758
kfree(mem);
drivers/net/wireless/ath/ath10k/snoc.c
1328
ar_snoc->mem = devm_ioremap(&pdev->dev, ar_snoc->mem_pa,
drivers/net/wireless/ath/ath10k/snoc.c
1330
if (!ar_snoc->mem) {
drivers/net/wireless/ath/ath10k/snoc.c
480
iowrite32(value, ar_snoc->mem + offset);
drivers/net/wireless/ath/ath10k/snoc.c
488
val = ioread32(ar_snoc->mem + offset);
drivers/net/wireless/ath/ath10k/snoc.h
70
void __iomem *mem;
drivers/net/wireless/ath/ath10k/wmi.h
4739
struct wmi_pdev_stats_mem mem;
drivers/net/wireless/ath/ath10k/wmi.h
4749
struct wmi_pdev_stats_mem mem;
drivers/net/wireless/ath/ath11k/ahb.c
174
iowrite32(value, ab->mem + window_start +
drivers/net/wireless/ath/ath11k/ahb.c
186
val = ioread32(ab->mem + window_start +
drivers/net/wireless/ath/ath11k/ahb.c
201
return ioread32(ab->mem + offset);
drivers/net/wireless/ath/ath11k/ahb.c
206
iowrite32(value, ab->mem + offset);
drivers/net/wireless/ath/ath11k/ahb.c
901
void __iomem *mem;
drivers/net/wireless/ath/ath11k/ahb.c
906
mem = devm_platform_get_and_ioremap_resource(pdev, 0, &mem_res);
drivers/net/wireless/ath/ath11k/ahb.c
907
if (IS_ERR(mem)) {
drivers/net/wireless/ath/ath11k/ahb.c
909
return PTR_ERR(mem);
drivers/net/wireless/ath/ath11k/ahb.c
912
ab->mem = mem;
drivers/net/wireless/ath/ath11k/ahb.c
953
ab->mem_ce = ab->mem;
drivers/net/wireless/ath/ath11k/ce.c
1058
void ath11k_ce_byte_swap(void *mem, u32 len)
drivers/net/wireless/ath/ath11k/ce.c
1063
if (!mem)
drivers/net/wireless/ath/ath11k/ce.c
1067
*(u32 *)mem = swab32(*(u32 *)mem);
drivers/net/wireless/ath/ath11k/ce.c
1068
mem += 4;
drivers/net/wireless/ath/ath11k/ce.h
28
void ath11k_ce_byte_swap(void *mem, u32 len);
drivers/net/wireless/ath/ath11k/core.h
944
void __iomem *mem;
drivers/net/wireless/ath/ath11k/hal.c
1063
(u32 *)((unsigned long)ab->mem + reg_base);
drivers/net/wireless/ath/ath11k/hal.c
1070
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath11k/hal.c
1096
(u32 *)((unsigned long)ab->mem + reg_base +
drivers/net/wireless/ath/ath11k/hal.c
1105
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath11k/hal.c
1140
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath11k/hal.c
1143
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath11k/hal.c
888
(unsigned long)ab->mem,
drivers/net/wireless/ath/ath11k/hal.c
898
(unsigned long)ab->mem,
drivers/net/wireless/ath/ath11k/hw.h
84
#define ATH11K_CE_OFFSET(ab) (ab->mem_ce - ab->mem)
drivers/net/wireless/ath/ath11k/mhi.c
346
mhi_ctrl->regs = ab->mem;
drivers/net/wireless/ath/ath11k/pci.c
102
iowrite32(value, ab->mem + window_start +
drivers/net/wireless/ath/ath11k/pci.c
106
iowrite32(value, ab->mem + window_start +
drivers/net/wireless/ath/ath11k/pci.c
121
val = ioread32(ab->mem + window_start +
drivers/net/wireless/ath/ath11k/pci.c
125
val = ioread32(ab->mem + window_start +
drivers/net/wireless/ath/ath11k/pci.c
177
ab_pci->ab->mem + ATH11K_PCI_WINDOW_REG_ADDRESS);
drivers/net/wireless/ath/ath11k/pci.c
187
ab->mem + ATH11K_PCI_WINDOW_REG_ADDRESS);
drivers/net/wireless/ath/ath11k/pci.c
188
ioread32(ab->mem + ATH11K_PCI_WINDOW_REG_ADDRESS);
drivers/net/wireless/ath/ath11k/pci.c
573
ab->mem = pci_iomap(pdev, ATH11K_PCI_BAR_NUM, 0);
drivers/net/wireless/ath/ath11k/pci.c
574
if (!ab->mem) {
drivers/net/wireless/ath/ath11k/pci.c
580
ab->mem_ce = ab->mem;
drivers/net/wireless/ath/ath11k/pci.c
582
ath11k_dbg(ab, ATH11K_DBG_BOOT, "pci_mem 0x%p\n", ab->mem);
drivers/net/wireless/ath/ath11k/pci.c
598
pci_iounmap(pci_dev, ab->mem);
drivers/net/wireless/ath/ath11k/pci.c
599
ab->mem = NULL;
drivers/net/wireless/ath/ath11k/pci.c
85
ab->mem + ATH11K_PCI_WINDOW_REG_ADDRESS);
drivers/net/wireless/ath/ath11k/pci.c
86
ioread32(ab->mem + ATH11K_PCI_WINDOW_REG_ADDRESS);
drivers/net/wireless/ath/ath11k/pcic.c
170
iowrite32(value, ab->mem + offset);
drivers/net/wireless/ath/ath11k/pcic.c
200
val = ioread32(ab->mem + offset);
drivers/net/wireless/ath/ath11k/qmi.c
2187
ab->mem = bar_addr_va;
drivers/net/wireless/ath/ath11k/qmi.c
2191
ab->mem_ce = ab->mem;
drivers/net/wireless/ath/ath12k/ahb.c
128
return ioread32(ab->mem + offset);
drivers/net/wireless/ath/ath12k/ahb.c
137
iowrite32(value, ab->mem + offset);
drivers/net/wireless/ath/ath12k/ahb.c
909
ab->mem = devm_platform_get_and_ioremap_resource(pdev, 0, &mem_res);
drivers/net/wireless/ath/ath12k/ahb.c
910
if (IS_ERR(ab->mem)) {
drivers/net/wireless/ath/ath12k/ahb.c
911
ret = dev_err_probe(&pdev->dev, PTR_ERR(ab->mem), "ioremap error\n");
drivers/net/wireless/ath/ath12k/ahb.c
959
devm_iounmap(ab->dev, ab->mem);
drivers/net/wireless/ath/ath12k/ahb.c
962
ab->mem = NULL;
drivers/net/wireless/ath/ath12k/ahb.c
970
if (ab->mem)
drivers/net/wireless/ath/ath12k/ahb.c
971
devm_iounmap(ab->dev, ab->mem);
drivers/net/wireless/ath/ath12k/ahb.c
976
ab->mem = NULL;
drivers/net/wireless/ath/ath12k/core.h
951
void __iomem *mem;
drivers/net/wireless/ath/ath12k/hal.c
556
(unsigned long)ab->mem,
drivers/net/wireless/ath/ath12k/hal.c
566
(unsigned long)ab->mem,
drivers/net/wireless/ath/ath12k/mhi.c
202
mhi_ctrl->regs = ab->mem;
drivers/net/wireless/ath/ath12k/pci.c
1164
val = ioread32(ab->mem + offset);
drivers/net/wireless/ath/ath12k/pci.c
1177
val = ioread32(ab->mem +
drivers/net/wireless/ath/ath12k/pci.c
1180
val = ioread32(ab->mem + window_start +
drivers/net/wireless/ath/ath12k/pci.c
1185
val = ioread32(ab->mem + window_start +
drivers/net/wireless/ath/ath12k/pci.c
1212
iowrite32(value, ab->mem + offset);
drivers/net/wireless/ath/ath12k/pci.c
1225
iowrite32(value, ab->mem +
drivers/net/wireless/ath/ath12k/pci.c
1228
iowrite32(value, ab->mem + window_start +
drivers/net/wireless/ath/ath12k/pci.c
123
ab->mem + ab_pci->window_reg_addr);
drivers/net/wireless/ath/ath12k/pci.c
1233
iowrite32(value, ab->mem + window_start +
drivers/net/wireless/ath/ath12k/pci.c
124
ioread32(ab->mem + ab_pci->window_reg_addr);
drivers/net/wireless/ath/ath12k/pci.c
143
iowrite32(WINDOW_ENABLE_BIT | window, ab_pci->ab->mem + ab_pci->window_reg_addr);
drivers/net/wireless/ath/ath12k/pci.c
176
ab->mem + ab_pci->window_reg_addr);
drivers/net/wireless/ath/ath12k/pci.c
177
ioread32(ab->mem + ab_pci->window_reg_addr);
drivers/net/wireless/ath/ath12k/pci.c
855
ab->mem = pci_iomap(pdev, ATH12K_PCI_BAR_NUM, 0);
drivers/net/wireless/ath/ath12k/pci.c
856
if (!ab->mem) {
drivers/net/wireless/ath/ath12k/pci.c
862
ath12k_dbg(ab, ATH12K_DBG_BOOT, "boot pci_mem 0x%p\n", ab->mem);
drivers/net/wireless/ath/ath12k/pci.c
878
pci_iounmap(pci_dev, ab->mem);
drivers/net/wireless/ath/ath12k/pci.c
879
ab->mem = NULL;
drivers/net/wireless/ath/ath12k/wifi7/hal.c
323
(u32 *)((unsigned long)ab->mem + reg_base);
drivers/net/wireless/ath/ath12k/wifi7/hal.c
329
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath12k/wifi7/hal.c
334
(u32 *)((unsigned long)ab->mem + reg_base +
drivers/net/wireless/ath/ath12k/wifi7/hal.c
341
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath12k/wifi7/hal.c
388
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath12k/wifi7/hal.c
391
(unsigned long)ab->mem);
drivers/net/wireless/ath/ath5k/ahb.c
109
mem = ioremap(res->start, resource_size(res));
drivers/net/wireless/ath/ath5k/ahb.c
110
if (mem == NULL) {
drivers/net/wireless/ath/ath5k/ahb.c
132
ah->iobase = mem;
drivers/net/wireless/ath/ath5k/ahb.c
183
iounmap(mem);
drivers/net/wireless/ath/ath5k/ahb.c
91
void __iomem *mem;
drivers/net/wireless/ath/ath5k/pci.c
163
void __iomem *mem;
drivers/net/wireless/ath/ath5k/pci.c
240
mem = pci_iomap(pdev, 0, 0);
drivers/net/wireless/ath/ath5k/pci.c
241
if (!mem) {
drivers/net/wireless/ath/ath5k/pci.c
266
ah->iobase = mem; /* So we can unmap it on detach */
drivers/net/wireless/ath/ath5k/pci.c
280
pci_iounmap(pdev, mem);
drivers/net/wireless/ath/ath9k/ahb.c
109
hw_name, mem, irq);
drivers/net/wireless/ath/ath9k/ahb.c
61
void __iomem *mem;
drivers/net/wireless/ath/ath9k/ahb.c
67
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/net/wireless/ath/ath9k/ahb.c
68
if (IS_ERR(mem)) {
drivers/net/wireless/ath/ath9k/ahb.c
70
return PTR_ERR(mem);
drivers/net/wireless/ath/ath9k/ahb.c
90
sc->mem = mem;
drivers/net/wireless/ath/ath9k/ath9k.h
994
void __iomem *mem;
drivers/net/wireless/ath/ath9k/ath9k_pci_owl_loader.c
105
pci_iounmap(pdev, mem);
drivers/net/wireless/ath/ath9k/ath9k_pci_owl_loader.c
38
void __iomem *mem;
drivers/net/wireless/ath/ath9k/ath9k_pci_owl_loader.c
67
mem = pci_iomap(pdev, 0, 0);
drivers/net/wireless/ath/ath9k/ath9k_pci_owl_loader.c
68
if (!mem) {
drivers/net/wireless/ath/ath9k/ath9k_pci_owl_loader.c
96
iowrite32(val, mem + reg);
drivers/net/wireless/ath/ath9k/init.c
181
iowrite32(val, sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/init.c
184
iowrite32(val, sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/init.c
197
val = ioread32(sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/init.c
200
val = ioread32(sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/init.c
219
val = ioread32(sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/init.c
222
iowrite32(val, sc->mem + reg_offset);
drivers/net/wireless/ath/ath9k/pci.c
955
sc->mem = pcim_iomap_table(pdev)[0];
drivers/net/wireless/ath/ath9k/pci.c
990
hw_name, sc->mem, pdev->irq);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
761
struct brcmf_core *mem;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
763
mem = brcmf_chip_get_core(&ci->pub, BCMA_CORE_ARM_CR4);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
764
if (mem) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
765
mem_core = container_of(mem, struct brcmf_core_priv, pub);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
773
mem = brcmf_chip_get_core(&ci->pub, BCMA_CORE_SYS_MEM);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
774
if (mem) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
775
mem_core = container_of(mem, struct brcmf_core_priv,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
784
mem = brcmf_chip_get_core(&ci->pub,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
786
if (!mem) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c
790
mem_core = container_of(mem, struct brcmf_core_priv,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
371
void (*frob_mem)(void *ctx, u32 mem_addr, void *mem, size_t buflen);
drivers/net/wireless/intel/iwlwifi/mei/main.c
249
struct iwl_mei_shared_mem_ptrs *mem = &mei->shared_mem;
drivers/net/wireless/intel/iwlwifi/mei/main.c
255
mem->ctrl = mei_cldev_dma_map(cldev, HBM_DMA_BUF_ID_WLAN, mem_size);
drivers/net/wireless/intel/iwlwifi/mei/main.c
256
if (IS_ERR(mem->ctrl)) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
257
int ret = PTR_ERR(mem->ctrl);
drivers/net/wireless/intel/iwlwifi/mei/main.c
259
mem->ctrl = NULL;
drivers/net/wireless/intel/iwlwifi/mei/main.c
264
memset(mem->ctrl, 0, mem_size);
drivers/net/wireless/intel/iwlwifi/mei/main.c
290
struct iwl_mei_shared_mem_ptrs *mem = &mei->shared_mem;
drivers/net/wireless/intel/iwlwifi/mei/main.c
296
mem->ctrl->sap_id = cpu_to_le32(SAP_CONTROL_BLOCK_ID);
drivers/net/wireless/intel/iwlwifi/mei/main.c
298
mem->ctrl->size = cpu_to_le32(sizeof(*mem->ctrl));
drivers/net/wireless/intel/iwlwifi/mei/main.c
300
h2m = &mem->ctrl->dir[SAP_DIRECTION_HOST_TO_ME];
drivers/net/wireless/intel/iwlwifi/mei/main.c
301
m2h = &mem->ctrl->dir[SAP_DIRECTION_ME_TO_HOST];
drivers/net/wireless/intel/iwlwifi/mei/main.c
315
q_head = (void *)(mem->ctrl + 1);
drivers/net/wireless/intel/iwlwifi/mei/main.c
320
mem->q_head[dir][queue] = q_head;
drivers/net/wireless/intel/iwlwifi/mei/main.c
322
le32_to_cpu(mem->ctrl->dir[dir].q_ctrl_blk[queue].size);
drivers/net/wireless/intel/iwlwifi/mei/main.c
323
mem->q_size[dir][queue] =
drivers/net/wireless/intel/iwlwifi/mei/main.c
324
le32_to_cpu(mem->ctrl->dir[dir].q_ctrl_blk[queue].size);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1038
static void iwl_mvm_frob_mem(void *ctx, u32 mem_addr, void *mem, size_t buflen)
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1079
memset((u8 *)mem + start - mem_addr, 0xAA, end - start);
drivers/net/wireless/marvell/mwifiex/fw.h
2421
struct host_cmd_ds_mem_access mem;
drivers/net/wireless/marvell/mwifiex/sta_cmd.c
1092
struct host_cmd_ds_mem_access *mem_access = (void *)&cmd->params.mem;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
770
struct host_cmd_ds_mem_access *mem = (void *)&resp->params.mem;
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
772
priv->mem_rw.addr = le32_to_cpu(mem->addr);
drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c
773
priv->mem_rw.value = le32_to_cpu(mem->value);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.h
471
} mem[4];
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
267
void __iomem *mem;
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
274
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
275
if (IS_ERR(mem))
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
276
return PTR_ERR(mem);
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
309
rt2x00dev->csr.base = mem;
drivers/net/wireless/ti/wl1251/io.c
173
partition->mem.start = mem_start;
drivers/net/wireless/ti/wl1251/io.c
174
partition->mem.size = mem_size;
drivers/net/wireless/ti/wl1251/wl1251.h
131
struct wl1251_partition mem;
drivers/net/wireless/ti/wl12xx/main.c
294
.mem = {
drivers/net/wireless/ti/wl12xx/main.c
491
.mem = {
drivers/net/wireless/ti/wl12xx/main.c
511
.mem = {
drivers/net/wireless/ti/wl12xx/main.c
530
.mem = {
drivers/net/wireless/ti/wl12xx/main.c
549
.mem = {
drivers/net/wireless/ti/wl12xx/main.c
641
memcpy(&wl->conf.mem, &wl12xx_default_priv_conf.mem_wl127x,
drivers/net/wireless/ti/wl12xx/main.c
642
sizeof(wl->conf.mem));
drivers/net/wireless/ti/wl12xx/main.c
665
memcpy(&wl->conf.mem, &wl12xx_default_priv_conf.mem_wl127x,
drivers/net/wireless/ti/wl12xx/main.c
666
sizeof(wl->conf.mem));
drivers/net/wireless/ti/wl18xx/main.c
432
.mem = {
drivers/net/wireless/ti/wl18xx/main.c
582
.mem = { .start = 0x00A00000, .size = 0x00012000 },
drivers/net/wireless/ti/wl18xx/main.c
588
.mem = { .start = 0x00000000, .size = 0x00014000 },
drivers/net/wireless/ti/wl18xx/main.c
594
.mem = { .start = 0x00700000, .size = 0x0000030c },
drivers/net/wireless/ti/wl18xx/main.c
600
.mem = { .start = 0x00800000, .size = 0x000050FC },
drivers/net/wireless/ti/wl18xx/main.c
606
.mem = { .start = WL18XX_PHY_INIT_MEM_ADDR,
drivers/net/wireless/ti/wlcore/acx.c
959
struct conf_memory_settings *mem;
drivers/net/wireless/ti/wlcore/acx.c
970
mem = &wl->conf.mem;
drivers/net/wireless/ti/wlcore/acx.c
973
mem_conf->num_stations = mem->num_stations;
drivers/net/wireless/ti/wlcore/acx.c
974
mem_conf->rx_mem_block_num = mem->rx_block_num;
drivers/net/wireless/ti/wlcore/acx.c
975
mem_conf->tx_min_mem_block_num = mem->tx_min_block_num;
drivers/net/wireless/ti/wlcore/acx.c
976
mem_conf->num_ssid_profiles = mem->ssid_profiles;
drivers/net/wireless/ti/wlcore/acx.c
978
mem_conf->dyn_mem_enable = mem->dynamic_memory;
drivers/net/wireless/ti/wlcore/acx.c
979
mem_conf->tx_free_req = mem->min_req_tx_blocks;
drivers/net/wireless/ti/wlcore/acx.c
980
mem_conf->rx_free_req = mem->min_req_rx_blocks;
drivers/net/wireless/ti/wlcore/acx.c
981
mem_conf->tx_min = mem->tx_min;
drivers/net/wireless/ti/wlcore/boot.c
187
partition.mem.start = dest;
drivers/net/wireless/ti/wlcore/boot.c
194
partition_limit = wl->ptable[PART_DOWN].mem.size;
drivers/net/wireless/ti/wlcore/boot.c
202
wl->ptable[PART_DOWN].mem.size;
drivers/net/wireless/ti/wlcore/boot.c
203
partition.mem.start = addr;
drivers/net/wireless/ti/wlcore/conf.h
1141
struct conf_memory_settings mem;
drivers/net/wireless/ti/wlcore/debugfs.c
1048
part.mem.start = *ppos;
drivers/net/wireless/ti/wlcore/debugfs.c
1049
part.mem.size = bytes;
drivers/net/wireless/ti/wlcore/debugfs.c
1129
part.mem.start = *ppos;
drivers/net/wireless/ti/wlcore/debugfs.c
1130
part.mem.size = bytes;
drivers/net/wireless/ti/wlcore/debugfs.c
1286
DEBUGFS_ADD_PREFIX(dev, mem, rootdir);
drivers/net/wireless/ti/wlcore/io.c
132
p->mem.start, p->mem.size);
drivers/net/wireless/ti/wlcore/io.c
140
ret = wlcore_raw_write32(wl, HW_PART0_START_ADDR, p->mem.start);
drivers/net/wireless/ti/wlcore/io.c
144
ret = wlcore_raw_write32(wl, HW_PART0_SIZE_ADDR, p->mem.size);
drivers/net/wireless/ti/wlcore/io.c
69
if ((addr >= part->mem.start) &&
drivers/net/wireless/ti/wlcore/io.c
70
(addr < part->mem.start + part->mem.size))
drivers/net/wireless/ti/wlcore/io.c
71
return addr - part->mem.start;
drivers/net/wireless/ti/wlcore/io.c
74
return addr - part->reg.start + part->mem.size;
drivers/net/wireless/ti/wlcore/io.c
77
return addr - part->mem2.start + part->mem.size +
drivers/net/wireless/ti/wlcore/io.c
81
return addr - part->mem3.start + part->mem.size +
drivers/net/wireless/ti/wlcore/wlcore.h
137
struct wlcore_partition mem;
drivers/ntb/hw/intel/ntb_hw_gen1.c
1648
int rc, mem;
drivers/ntb/hw/intel/ntb_hw_gen1.c
1740
mem = pci_select_bars(pdev, IORESOURCE_MEM);
drivers/ntb/hw/intel/ntb_hw_gen1.c
1741
ndev->bar4_split = hweight32(mem) ==
drivers/ntb/hw/intel/ntb_hw_gen1.c
1744
mem, ndev->bar4_split);
drivers/nvdimm/btt.c
1105
void *mem = kmap_local_page(page);
drivers/nvdimm/btt.c
1107
ret = arena_read_bytes(arena, nsoff, mem + off, len, NVDIMM_IO_ATOMIC);
drivers/nvdimm/btt.c
1108
kunmap_local(mem);
drivers/nvdimm/btt.c
1118
void *mem = kmap_local_page(page);
drivers/nvdimm/btt.c
1120
ret = arena_write_bytes(arena, nsoff, mem + off, len, NVDIMM_IO_ATOMIC);
drivers/nvdimm/btt.c
1121
kunmap_local(mem);
drivers/nvdimm/btt.c
1128
void *mem = kmap_local_page(page);
drivers/nvdimm/btt.c
1130
memset(mem + off, 0, len);
drivers/nvdimm/btt.c
1131
kunmap_local(mem);
drivers/nvdimm/btt.c
1150
void *mem;
drivers/nvdimm/btt.c
1160
mem = bvec_kmap_local(&bv);
drivers/nvdimm/btt.c
1162
ret = arena_write_bytes(arena, meta_nsoff, mem, cur_len,
drivers/nvdimm/btt.c
1165
ret = arena_read_bytes(arena, meta_nsoff, mem, cur_len,
drivers/nvdimm/btt.c
1168
kunmap_local(mem);
drivers/nvdimm/core.c
101
nvdimm_map->mem = memremap(offset, size, flags);
drivers/nvdimm/core.c
105
if (!nvdimm_map->mem)
drivers/nvdimm/core.c
132
memunmap(nvdimm_map->mem);
drivers/nvdimm/core.c
174
return nvdimm_map->mem;
drivers/nvdimm/core.c
59
void *mem;
drivers/nvdimm/pmem.c
128
void *mem;
drivers/nvdimm/pmem.c
131
mem = kmap_local_page(page);
drivers/nvdimm/pmem.c
133
memcpy_flushcache(pmem_addr, mem + off, chunk);
drivers/nvdimm/pmem.c
134
kunmap_local(mem);
drivers/nvdimm/pmem.c
147
void *mem;
drivers/nvdimm/pmem.c
150
mem = kmap_local_page(page);
drivers/nvdimm/pmem.c
152
rem = copy_mc_to_kernel(mem + off, pmem_addr, chunk);
drivers/nvdimm/pmem.c
153
kunmap_local(mem);
drivers/nvme/host/rdma.c
1376
sig_attrs->mem.sig_type = IB_SIG_TYPE_NONE;
drivers/nvme/host/rdma.c
1386
nvme_rdma_set_sig_domain(bi, cmd, &sig_attrs->mem, control,
drivers/nvme/target/rdma.c
602
nvmet_rdma_set_sig_domain(bi, cmd, &sig_attrs->mem, control,
drivers/nvme/target/rdma.c
613
nvmet_rdma_set_sig_domain(bi, cmd, &sig_attrs->mem, control,
drivers/nvmem/rmem.c
124
struct reserved_mem *mem;
drivers/nvmem/rmem.c
132
mem = of_reserved_mem_lookup(dev->of_node);
drivers/nvmem/rmem.c
133
if (!mem) {
drivers/nvmem/rmem.c
137
priv->mem = mem;
drivers/nvmem/rmem.c
143
config.size = mem->size;
drivers/nvmem/rmem.c
17
struct reserved_mem *mem;
drivers/nvmem/rmem.c
38
if ((phys_addr_t)offset + bytes > priv->mem->size)
drivers/nvmem/rmem.c
51
addr = memremap(priv->mem->base, priv->mem->size, MEMREMAP_WB);
drivers/nvmem/rmem.c
83
if (header.size > priv->mem->size)
drivers/of/address.c
1148
void __iomem *mem;
drivers/of/address.c
1159
mem = ioremap_np(res.start, resource_size(&res));
drivers/of/address.c
1161
mem = ioremap(res.start, resource_size(&res));
drivers/of/address.c
1163
if (!mem) {
drivers/of/address.c
1168
return mem;
drivers/of/fdt.c
127
pp = unflatten_dt_alloc(mem, sizeof(struct property),
drivers/of/fdt.c
176
pp = unflatten_dt_alloc(mem, sizeof(struct property) + len,
drivers/of/fdt.c
193
void **mem,
drivers/of/fdt.c
210
np = unflatten_dt_alloc(mem, sizeof(struct device_node) + len,
drivers/of/fdt.c
226
populate_properties(blob, offset, mem, np, pathp, dryrun);
drivers/of/fdt.c
271
void *mem,
drivers/of/fdt.c
279
void *base = mem;
drivers/of/fdt.c
309
ret = populate_node(blob, offset, &mem, nps[depth],
drivers/of/fdt.c
332
return mem - base;
drivers/of/fdt.c
358
void *mem;
drivers/of/fdt.c
390
mem = dt_alloc(size + 4, __alignof__(struct device_node));
drivers/of/fdt.c
391
if (!mem)
drivers/of/fdt.c
394
memset(mem, 0, size);
drivers/of/fdt.c
396
*(__be32 *)(mem + size) = cpu_to_be32(0xdeadbeef);
drivers/of/fdt.c
398
pr_debug(" unflattening %p...\n", mem);
drivers/of/fdt.c
401
ret = unflatten_dt_nodes(blob, mem, dad, mynodes);
drivers/of/fdt.c
403
if (be32_to_cpup(mem + size) != 0xdeadbeef)
drivers/of/fdt.c
405
be32_to_cpup(mem + size));
drivers/of/fdt.c
416
return mem;
drivers/of/fdt.c
444
void *mem;
drivers/of/fdt.c
447
mem = __unflatten_device_tree(blob, dad, mynodes, &kernel_tree_alloc,
drivers/of/fdt.c
451
return mem;
drivers/of/fdt.c
82
static void *unflatten_dt_alloc(void **mem, unsigned long size,
drivers/of/fdt.c
87
*mem = PTR_ALIGN(*mem, align);
drivers/of/fdt.c
88
res = *mem;
drivers/of/fdt.c
89
*mem += size;
drivers/of/fdt.c
96
void **mem,
drivers/of/kexec.c
279
scratch_mem = image->kho.scratch->mem;
drivers/pci/controller/dwc/pci-keystone.c
411
struct resource *mem;
drivers/pci/controller/dwc/pci-keystone.c
419
mem = entry->res;
drivers/pci/controller/dwc/pci-keystone.c
420
start = mem->start;
drivers/pci/controller/dwc/pci-keystone.c
421
end = mem->end;
drivers/pci/controller/dwc/pcie-designware-ep.c
1057
epc->mem->window.page_size);
drivers/pci/controller/dwc/pcie-designware-ep.c
1366
epc->mem->window.page_size);
drivers/pci/controller/dwc/pcie-designware-ep.c
629
*pci_size = ALIGN(ofst + *pci_size, epc->mem->window.page_size);
drivers/pci/controller/dwc/pcie-designware-host.c
1186
void __iomem *mem;
drivers/pci/controller/dwc/pcie-designware-host.c
1207
mem = ioremap(pci->pp.msg_res->start, pci->region_align);
drivers/pci/controller/dwc/pcie-designware-host.c
1208
if (!mem)
drivers/pci/controller/dwc/pcie-designware-host.c
1212
writel(0, mem);
drivers/pci/controller/dwc/pcie-designware-host.c
1214
iounmap(mem);
drivers/pci/controller/dwc/pcie-designware.c
1165
ll->vaddr.mem = dmam_alloc_coherent(pci->dev, ll->sz,
drivers/pci/controller/dwc/pcie-designware.c
1167
if (!ll->vaddr.mem)
drivers/pci/controller/dwc/pcie-designware.c
1176
ll->vaddr.mem = dmam_alloc_coherent(pci->dev, ll->sz,
drivers/pci/controller/dwc/pcie-designware.c
1178
if (!ll->vaddr.mem)
drivers/pci/controller/pci-mvebu.c
1413
mvebu_mbus_get_pcie_mem_aperture(&pcie->mem);
drivers/pci/controller/pci-mvebu.c
1414
if (resource_size(&pcie->mem) == 0) {
drivers/pci/controller/pci-mvebu.c
1419
pcie->mem.name = "PCI MEM";
drivers/pci/controller/pci-mvebu.c
1420
pci_add_resource(&bridge->windows, &pcie->mem);
drivers/pci/controller/pci-mvebu.c
1421
ret = devm_request_resource(dev, &iomem_resource, &pcie->mem);
drivers/pci/controller/pci-mvebu.c
89
struct resource mem;
drivers/pci/controller/pci-v3-semi.c
522
struct resource *mem;
drivers/pci/controller/pci-v3-semi.c
537
mem = win->res;
drivers/pci/controller/pci-v3-semi.c
538
if (mem->flags & IORESOURCE_PREFETCH) {
drivers/pci/controller/pci-v3-semi.c
539
mem->name = "V3 PCI PRE-MEM";
drivers/pci/controller/pci-v3-semi.c
540
v3->pre_mem = mem->start;
drivers/pci/controller/pci-v3-semi.c
541
v3->pre_bus_addr = mem->start - win->offset;
drivers/pci/controller/pci-v3-semi.c
543
mem, &v3->pre_bus_addr);
drivers/pci/controller/pci-v3-semi.c
544
if (resource_size(mem) != SZ_256M) {
drivers/pci/controller/pci-v3-semi.c
549
(mem->start != v3->non_pre_mem + SZ_256M)) {
drivers/pci/controller/pci-v3-semi.c
564
mem->name = "V3 PCI NON-PRE-MEM";
drivers/pci/controller/pci-v3-semi.c
565
v3->non_pre_mem = mem->start;
drivers/pci/controller/pci-v3-semi.c
566
v3->non_pre_bus_addr = mem->start - win->offset;
drivers/pci/controller/pci-v3-semi.c
568
mem, &v3->non_pre_bus_addr);
drivers/pci/controller/pci-v3-semi.c
569
if (resource_size(mem) != SZ_256M) {
drivers/pci/controller/pci-versatile.c
69
int i, myslot = -1, mem = 1;
drivers/pci/controller/pci-versatile.c
93
writel(entry->res->start >> 28, PCI_IMAP(mem));
drivers/pci/controller/pci-versatile.c
94
writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem));
drivers/pci/controller/pci-versatile.c
95
mem++;
drivers/pci/controller/pcie-iproc-bcma.c
58
pcie->mem.start = bdev->addr_s[0];
drivers/pci/controller/pcie-iproc-bcma.c
59
pcie->mem.end = bdev->addr_s[0] + SZ_128M - 1;
drivers/pci/controller/pcie-iproc-bcma.c
60
pcie->mem.name = "PCIe MEM space";
drivers/pci/controller/pcie-iproc-bcma.c
61
pcie->mem.flags = IORESOURCE_MEM;
drivers/pci/controller/pcie-iproc-bcma.c
62
pci_add_resource(&bridge->windows, &pcie->mem);
drivers/pci/controller/pcie-iproc.h
92
struct resource mem;
drivers/pci/controller/pcie-mediatek.c
670
struct resource *mem = NULL;
drivers/pci/controller/pcie-mediatek.c
678
mem = entry->res;
drivers/pci/controller/pcie-mediatek.c
679
if (!mem)
drivers/pci/controller/pcie-mediatek.c
742
val = lower_32_bits(mem->start) |
drivers/pci/controller/pcie-mediatek.c
743
AHB2PCIE_SIZE(fls(resource_size(mem)));
drivers/pci/controller/pcie-mediatek.c
746
val = upper_32_bits(mem->start);
drivers/pci/endpoint/pci-epc-mem.c
101
epc->mem = epc->windows[0];
drivers/pci/endpoint/pci-epc-mem.c
108
mem = epc->windows[i];
drivers/pci/endpoint/pci-epc-mem.c
109
kfree(mem->bitmap);
drivers/pci/endpoint/pci-epc-mem.c
110
kfree(mem);
drivers/pci/endpoint/pci-epc-mem.c
150
struct pci_epc_mem *mem;
drivers/pci/endpoint/pci-epc-mem.c
157
mem = epc->windows[i];
drivers/pci/endpoint/pci-epc-mem.c
158
kfree(mem->bitmap);
drivers/pci/endpoint/pci-epc-mem.c
159
kfree(mem);
drivers/pci/endpoint/pci-epc-mem.c
164
epc->mem = NULL;
drivers/pci/endpoint/pci-epc-mem.c
182
struct pci_epc_mem *mem;
drivers/pci/endpoint/pci-epc-mem.c
190
mem = epc->windows[i];
drivers/pci/endpoint/pci-epc-mem.c
191
if (size > mem->window.size)
drivers/pci/endpoint/pci-epc-mem.c
194
align_size = ALIGN(size, mem->window.page_size);
drivers/pci/endpoint/pci-epc-mem.c
195
order = pci_epc_mem_get_order(mem, align_size);
drivers/pci/endpoint/pci-epc-mem.c
197
mutex_lock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
198
pageno = bitmap_find_free_region(mem->bitmap, mem->pages,
drivers/pci/endpoint/pci-epc-mem.c
201
page_shift = ilog2(mem->window.page_size);
drivers/pci/endpoint/pci-epc-mem.c
202
*phys_addr = mem->window.phys_base +
drivers/pci/endpoint/pci-epc-mem.c
206
bitmap_release_region(mem->bitmap,
drivers/pci/endpoint/pci-epc-mem.c
208
mutex_unlock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
211
mutex_unlock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
214
mutex_unlock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
224
struct pci_epc_mem *mem;
drivers/pci/endpoint/pci-epc-mem.c
228
mem = epc->windows[i];
drivers/pci/endpoint/pci-epc-mem.c
23
static int pci_epc_mem_get_order(struct pci_epc_mem *mem, size_t size)
drivers/pci/endpoint/pci-epc-mem.c
230
if (phys_addr >= mem->window.phys_base &&
drivers/pci/endpoint/pci-epc-mem.c
231
phys_addr < (mem->window.phys_base + mem->window.size))
drivers/pci/endpoint/pci-epc-mem.c
232
return mem;
drivers/pci/endpoint/pci-epc-mem.c
250
struct pci_epc_mem *mem;
drivers/pci/endpoint/pci-epc-mem.c
256
mem = pci_epc_get_matching_window(epc, phys_addr);
drivers/pci/endpoint/pci-epc-mem.c
257
if (!mem) {
drivers/pci/endpoint/pci-epc-mem.c
26
unsigned int page_shift = ilog2(mem->window.page_size);
drivers/pci/endpoint/pci-epc-mem.c
262
page_size = mem->window.page_size;
drivers/pci/endpoint/pci-epc-mem.c
265
pageno = (phys_addr - mem->window.phys_base) >> page_shift;
drivers/pci/endpoint/pci-epc-mem.c
267
order = pci_epc_mem_get_order(mem, size);
drivers/pci/endpoint/pci-epc-mem.c
268
mutex_lock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
269
bitmap_release_region(mem->bitmap, pageno, order);
drivers/pci/endpoint/pci-epc-mem.c
270
mutex_unlock(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
51
struct pci_epc_mem *mem = NULL;
drivers/pci/endpoint/pci-epc-mem.c
77
mem = kzalloc_obj(*mem);
drivers/pci/endpoint/pci-epc-mem.c
78
if (!mem) {
drivers/pci/endpoint/pci-epc-mem.c
87
kfree(mem);
drivers/pci/endpoint/pci-epc-mem.c
92
mem->window.phys_base = windows[i].phys_base;
drivers/pci/endpoint/pci-epc-mem.c
93
mem->window.size = windows[i].size;
drivers/pci/endpoint/pci-epc-mem.c
94
mem->window.page_size = page_size;
drivers/pci/endpoint/pci-epc-mem.c
95
mem->bitmap = bitmap;
drivers/pci/endpoint/pci-epc-mem.c
96
mem->pages = pages;
drivers/pci/endpoint/pci-epc-mem.c
97
mutex_init(&mem->lock);
drivers/pci/endpoint/pci-epc-mem.c
98
epc->windows[i] = mem;
drivers/pci/hotplug/ibmphp.h
360
u32 mem;
drivers/pci/hotplug/ibmphp.h
681
struct resource_node *mem[6];
drivers/pci/hotplug/ibmphp_pci.c
1023
if (mem)
drivers/pci/hotplug/ibmphp_pci.c
1024
ibmphp_remove_resource(mem);
drivers/pci/hotplug/ibmphp_pci.c
1034
func->mem[i] = NULL;
drivers/pci/hotplug/ibmphp_pci.c
1150
amount->mem += len[count];
drivers/pci/hotplug/ibmphp_pci.c
1168
if ((amount->mem) && (amount->mem < MEMBRIDGE))
drivers/pci/hotplug/ibmphp_pci.c
1169
amount->mem = MEMBRIDGE;
drivers/pci/hotplug/ibmphp_pci.c
1196
struct resource_node *mem;
drivers/pci/hotplug/ibmphp_pci.c
1279
if (ibmphp_find_resource(bus, start_address, &mem, MEM) < 0) {
drivers/pci/hotplug/ibmphp_pci.c
1283
if (mem) {
drivers/pci/hotplug/ibmphp_pci.c
1284
debug("mem->start = %x\n", mem->start);
drivers/pci/hotplug/ibmphp_pci.c
1286
ibmphp_remove_resource(mem);
drivers/pci/hotplug/ibmphp_pci.c
1310
struct resource_node *mem = NULL;
drivers/pci/hotplug/ibmphp_pci.c
1395
if (ibmphp_find_resource(bus, start_address, &mem, MEM) < 0) {
drivers/pci/hotplug/ibmphp_pci.c
1399
if (mem) {
drivers/pci/hotplug/ibmphp_pci.c
1400
debug("mem->start = %x\n", mem->start);
drivers/pci/hotplug/ibmphp_pci.c
1402
ibmphp_remove_resource(mem);
drivers/pci/hotplug/ibmphp_pci.c
1569
if (cur_func->mem[i]) {
drivers/pci/hotplug/ibmphp_pci.c
1572
ibmphp_remove_resource(cur_func->mem[i]);
drivers/pci/hotplug/ibmphp_pci.c
1573
cur_func->mem[i] = NULL;
drivers/pci/hotplug/ibmphp_pci.c
1603
static int add_new_bus(struct bus_node *bus, struct resource_node *io, struct resource_node *mem, struct resource_node *pfmem, u8 parent_busno)
drivers/pci/hotplug/ibmphp_pci.c
1631
if (mem) {
drivers/pci/hotplug/ibmphp_pci.c
1636
mem_range->start = mem->start;
drivers/pci/hotplug/ibmphp_pci.c
1637
mem_range->end = mem->end;
drivers/pci/hotplug/ibmphp_pci.c
316
} else if (cur_func->mem[i]) {
drivers/pci/hotplug/ibmphp_pci.c
317
ibmphp_remove_resource(cur_func->mem[i]);
drivers/pci/hotplug/ibmphp_pci.c
318
cur_func->mem[i] = NULL;
drivers/pci/hotplug/ibmphp_pci.c
345
struct resource_node *mem[6];
drivers/pci/hotplug/ibmphp_pci.c
488
mem[count] = kzalloc_obj(struct resource_node);
drivers/pci/hotplug/ibmphp_pci.c
489
if (!mem[count])
drivers/pci/hotplug/ibmphp_pci.c
492
mem[count]->type = MEM;
drivers/pci/hotplug/ibmphp_pci.c
493
mem[count]->busno = func->busno;
drivers/pci/hotplug/ibmphp_pci.c
494
mem[count]->devfunc = PCI_DEVFN(func->device,
drivers/pci/hotplug/ibmphp_pci.c
496
mem[count]->len = len[count];
drivers/pci/hotplug/ibmphp_pci.c
497
if (ibmphp_check_resource(mem[count], 0) == 0) {
drivers/pci/hotplug/ibmphp_pci.c
498
ibmphp_add_resource(mem[count]);
drivers/pci/hotplug/ibmphp_pci.c
499
func->mem[count] = mem[count];
drivers/pci/hotplug/ibmphp_pci.c
503
kfree(mem[count]);
drivers/pci/hotplug/ibmphp_pci.c
506
pci_bus_write_config_dword(ibmphp_pci_bus, devfn, address[count], func->mem[count]->start);
drivers/pci/hotplug/ibmphp_pci.c
508
debug("b4 writing, start address is %x\n", func->mem[count]->start);
drivers/pci/hotplug/ibmphp_pci.c
561
struct resource_node *mem = NULL;
drivers/pci/hotplug/ibmphp_pci.c
750
func->mem[count] = bus_mem[count];
drivers/pci/hotplug/ibmphp_pci.c
758
pci_bus_write_config_dword(ibmphp_pci_bus, devfn, address[count], func->mem[count]->start);
drivers/pci/hotplug/ibmphp_pci.c
780
debug("amount_needed->mem = %x\n", amount_needed->mem);
drivers/pci/hotplug/ibmphp_pci.c
795
func->mem[count] = NULL;
drivers/pci/hotplug/ibmphp_pci.c
824
if (!amount_needed->mem) {
drivers/pci/hotplug/ibmphp_pci.c
828
debug("it wants %x memory behind the bridge\n", amount_needed->mem);
drivers/pci/hotplug/ibmphp_pci.c
829
mem = kzalloc_obj(*mem);
drivers/pci/hotplug/ibmphp_pci.c
830
if (!mem) {
drivers/pci/hotplug/ibmphp_pci.c
834
mem->type = MEM;
drivers/pci/hotplug/ibmphp_pci.c
835
mem->busno = func->busno;
drivers/pci/hotplug/ibmphp_pci.c
836
mem->devfunc = PCI_DEVFN(func->device, func->function);
drivers/pci/hotplug/ibmphp_pci.c
837
mem->len = amount_needed->mem;
drivers/pci/hotplug/ibmphp_pci.c
838
if (ibmphp_check_resource(mem, 1) == 0) {
drivers/pci/hotplug/ibmphp_pci.c
839
ibmphp_add_resource(mem);
drivers/pci/hotplug/ibmphp_pci.c
901
rc = add_new_bus(bus, io, mem, pfmem, func->busno);
drivers/pci/hotplug/ibmphp_pci.c
903
rc = add_new_bus(bus, io, mem, pfmem, 0xFF);
drivers/pci/hotplug/ibmphp_res.c
1667
struct resource_node *mem;
drivers/pci/hotplug/ibmphp_res.c
1690
mem = kzalloc_obj(struct resource_node);
drivers/pci/hotplug/ibmphp_res.c
1691
if (!mem)
drivers/pci/hotplug/ibmphp_res.c
1694
mem->type = MEM;
drivers/pci/hotplug/ibmphp_res.c
1695
mem->busno = pfmem_cur->busno;
drivers/pci/hotplug/ibmphp_res.c
1696
mem->devfunc = pfmem_cur->devfunc;
drivers/pci/hotplug/ibmphp_res.c
1697
mem->start = pfmem_cur->start;
drivers/pci/hotplug/ibmphp_res.c
1698
mem->end = pfmem_cur->end;
drivers/pci/hotplug/ibmphp_res.c
1699
mem->len = pfmem_cur->len;
drivers/pci/hotplug/ibmphp_res.c
1700
if (ibmphp_add_resource(mem) < 0)
drivers/pci/hotplug/ibmphp_res.c
1702
pfmem_cur->rangeno = mem->rangeno;
drivers/pci/hotplug/ibmphp_res.c
1914
struct resource_node *mem;
drivers/pci/hotplug/ibmphp_res.c
2043
if (ibmphp_find_resource(bus_cur, start_address, &mem, MEM)) {
drivers/pci/hotplug/ibmphp_res.c
2044
mem = kzalloc_obj(struct resource_node);
drivers/pci/hotplug/ibmphp_res.c
2045
if (!mem) {
drivers/pci/hotplug/ibmphp_res.c
2049
mem->type = MEM;
drivers/pci/hotplug/ibmphp_res.c
2050
mem->busno = bus_cur->busno;
drivers/pci/hotplug/ibmphp_res.c
2051
mem->devfunc = ((device << 3) | (function & 0x7));
drivers/pci/hotplug/ibmphp_res.c
2052
mem->start = start_address;
drivers/pci/hotplug/ibmphp_res.c
2053
mem->end = end_address + 0xfffff;
drivers/pci/hotplug/ibmphp_res.c
2054
mem->len = mem->end - mem->start + 1;
drivers/pci/hotplug/ibmphp_res.c
2055
ibmphp_add_resource(mem);
drivers/pci/ide.c
247
struct resource *mem, *pref;
drivers/pci/ide.c
298
mem = pci_resource_n(br, PCI_BRIDGE_MEM_WINDOW);
drivers/pci/ide.c
300
if (resource_assigned(mem))
drivers/pci/ide.c
301
pcibios_resource_to_bus(br->bus, &mem_assoc, mem);
drivers/pci/p2pdma.c
1118
if (state->mem == p2p_pgmap->mem)
drivers/pci/p2pdma.c
1121
state->mem = p2p_pgmap->mem;
drivers/pci/p2pdma.c
1122
state->map = pci_p2pdma_map_type(p2p_pgmap->mem, dev);
drivers/pci/p2pdma.c
217
to_pci_dev(pgmap->mem->owner)->p2pdma, 1);
drivers/pci/p2pdma.c
28
struct p2pdma_provider mem[PCI_STD_NUM_BARS];
drivers/pci/p2pdma.c
284
p2p->mem[i].owner = &pdev->dev;
drivers/pci/p2pdma.c
285
p2p->mem[i].bus_offset =
drivers/pci/p2pdma.c
329
return &p2p->mem[bar];
drivers/pci/p2pdma.c
33
struct p2pdma_provider *mem;
drivers/pci/p2pdma.c
368
sysfs_remove_file_from_group(&p2p_pgmap->mem->owner->kobj,
drivers/pci/p2pdma.c
387
struct p2pdma_provider *mem;
drivers/pci/p2pdma.c
413
mem = pcim_p2pdma_provider(pdev, bar);
drivers/pci/p2pdma.c
418
if (WARN_ON(!mem))
drivers/pci/p2pdma.c
431
p2p_pgmap->mem = mem;
drivers/pcmcia/cistpl.c
1022
static u_char *parse_mem(u_char *p, u_char *q, cistpl_mem_t *mem)
drivers/pcmcia/cistpl.c
103
s->cis_virt = ioremap(mem->res->start, s->map_size);
drivers/pcmcia/cistpl.c
1030
mem->nwin = (*p & 0x07) + 1;
drivers/pcmcia/cistpl.c
1037
for (i = 0; i < mem->nwin; i++) {
drivers/pcmcia/cistpl.c
105
mem->card_start = card_offset;
drivers/pcmcia/cistpl.c
1055
mem->win[i].len = len << 8;
drivers/pcmcia/cistpl.c
1056
mem->win[i].card_addr = ca << 8;
drivers/pcmcia/cistpl.c
1057
mem->win[i].host_addr = ha << 8;
drivers/pcmcia/cistpl.c
106
mem->flags = flags;
drivers/pcmcia/cistpl.c
108
ret = s->ops->set_mem_map(s, mem);
drivers/pcmcia/cistpl.c
1158
entry->mem.nwin = 0;
drivers/pcmcia/cistpl.c
1161
entry->mem.nwin = 1;
drivers/pcmcia/cistpl.c
1162
entry->mem.win[0].len = get_unaligned_le16(p) << 8;
drivers/pcmcia/cistpl.c
1163
entry->mem.win[0].card_addr = 0;
drivers/pcmcia/cistpl.c
1164
entry->mem.win[0].host_addr = 0;
drivers/pcmcia/cistpl.c
1170
entry->mem.nwin = 1;
drivers/pcmcia/cistpl.c
1171
entry->mem.win[0].len = get_unaligned_le16(p) << 8;
drivers/pcmcia/cistpl.c
1172
entry->mem.win[0].card_addr = get_unaligned_le16(p + 2) << 8;
drivers/pcmcia/cistpl.c
1173
entry->mem.win[0].host_addr = 0;
drivers/pcmcia/cistpl.c
1179
p = parse_mem(p, q, &entry->mem);
drivers/pcmcia/cistpl.c
118
s->cis_virt = ioremap(mem->static_start, s->map_size);
drivers/pcmcia/cistpl.c
89
pccard_mem_map *mem = &s->cis_mem;
drivers/pcmcia/cistpl.c
92
if (!(s->features & SS_CAP_STATIC_MAP) && (mem->res == NULL)) {
drivers/pcmcia/cistpl.c
93
mem->res = pcmcia_find_mem_region(0, s->map_size,
drivers/pcmcia/cistpl.c
95
if (mem->res == NULL) {
drivers/pcmcia/cs_internal.h
34
struct resource mem[MAX_WIN]; /* mem areas */
drivers/pcmcia/ds.c
557
c->mem[i].name = p_dev->devname;
drivers/pcmcia/ds.c
558
c->mem[i].flags = IORESOURCE_MEM;
drivers/pcmcia/ds.c
564
p_dev->resource[i] = &p_dev->function_config->mem[i-MAX_IO_WIN];
drivers/pcmcia/electra_cf.c
180
struct resource mem, io;
drivers/pcmcia/electra_cf.c
185
err = of_address_to_resource(np, 0, &mem);
drivers/pcmcia/electra_cf.c
201
cf->mem_phys = mem.start;
drivers/pcmcia/electra_cf.c
202
cf->mem_size = PAGE_ALIGN(resource_size(&mem));
drivers/pcmcia/electra_cf.c
218
cf->iomem.end = (unsigned long)cf->mem_base + (mem.end - mem.start);
drivers/pcmcia/i82092.c
390
pccard_mem_map mem = { .res = &res, };
drivers/pcmcia/i82092.c
397
mem.map = i;
drivers/pcmcia/i82092.c
398
i82092aa_set_mem_map(sock, &mem);
drivers/pcmcia/i82092.c
591
struct pccard_mem_map *mem)
drivers/pcmcia/i82092.c
600
pcibios_resource_to_bus(sock_info->dev->bus, &region, mem->res);
drivers/pcmcia/i82092.c
602
map = mem->map;
drivers/pcmcia/i82092.c
606
if ((mem->card_start > 0x3ffffff) || (region.start > region.end) ||
drivers/pcmcia/i82092.c
607
(mem->speed > 1000)) {
drivers/pcmcia/i82092.c
613
mem->card_start);
drivers/pcmcia/i82092.c
624
if (mem->flags & MAP_16BIT)
drivers/pcmcia/i82092.c
626
if (mem->flags & MAP_0WS)
drivers/pcmcia/i82092.c
633
switch (to_cycles(mem->speed)) {
drivers/pcmcia/i82092.c
651
i = ((mem->card_start - region.start) >> 12) & 0x3fff;
drivers/pcmcia/i82092.c
652
if (mem->flags & MAP_WRPROT)
drivers/pcmcia/i82092.c
654
if (mem->flags & MAP_ATTRIB)
drivers/pcmcia/i82092.c
659
if (mem->flags & MAP_ACTIVE)
drivers/pcmcia/i82092aa.h
20
static int i82092aa_set_mem_map(struct pcmcia_socket *socket, struct pccard_mem_map *mem);
drivers/pcmcia/i82365.c
1061
static int i365_set_mem_map(u_short sock, struct pccard_mem_map *mem)
drivers/pcmcia/i82365.c
1067
"%#x)\n", sock, mem->map, mem->flags, mem->speed,
drivers/pcmcia/i82365.c
1068
(unsigned long long)mem->res->start,
drivers/pcmcia/i82365.c
1069
(unsigned long long)mem->res->end, mem->card_start);
drivers/pcmcia/i82365.c
1071
map = mem->map;
drivers/pcmcia/i82365.c
1072
if ((map > 4) || (mem->card_start > 0x3ffffff) ||
drivers/pcmcia/i82365.c
1073
(mem->res->start > mem->res->end) || (mem->speed > 1000))
drivers/pcmcia/i82365.c
1075
if ((mem->res->start > 0xffffff) || (mem->res->end > 0xffffff))
drivers/pcmcia/i82365.c
1083
i = (mem->res->start >> 12) & 0x0fff;
drivers/pcmcia/i82365.c
1084
if (mem->flags & MAP_16BIT) i |= I365_MEM_16BIT;
drivers/pcmcia/i82365.c
1085
if (mem->flags & MAP_0WS) i |= I365_MEM_0WS;
drivers/pcmcia/i82365.c
1088
i = (mem->res->end >> 12) & 0x0fff;
drivers/pcmcia/i82365.c
1089
switch (to_cycles(mem->speed)) {
drivers/pcmcia/i82365.c
1097
i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff;
drivers/pcmcia/i82365.c
1098
if (mem->flags & MAP_WRPROT) i |= I365_MEM_WRPROT;
drivers/pcmcia/i82365.c
1099
if (mem->flags & MAP_ATTRIB) i |= I365_MEM_REG;
drivers/pcmcia/i82365.c
1103
if (mem->flags & MAP_ACTIVE)
drivers/pcmcia/i82365.c
1195
static int pcic_set_mem_map(struct pcmcia_socket *s, struct pccard_mem_map *mem)
drivers/pcmcia/i82365.c
1201
LOCKED(i365_set_mem_map(sock, mem));
drivers/pcmcia/i82365.c
1209
pccard_mem_map mem = { .res = &res, };
drivers/pcmcia/i82365.c
1216
mem.map = i;
drivers/pcmcia/i82365.c
1217
pcic_set_mem_map(s, &mem);
drivers/pcmcia/pcmcia_cis.c
226
cistpl_mem_t *mem = (cfg->mem.nwin) ? &cfg->mem : &dflt->mem;
drivers/pcmcia/pcmcia_cis.c
229
if (mem->nwin == 0)
drivers/pcmcia/pcmcia_cis.c
232
p_dev->resource[2]->start = mem->win[0].host_addr;
drivers/pcmcia/pcmcia_cis.c
233
p_dev->resource[2]->end = mem->win[0].len;
drivers/pcmcia/pcmcia_cis.c
236
p_dev->card_addr = mem->win[0].card_addr;
drivers/pcmcia/pd6729.c
468
struct pccard_mem_map *mem)
drivers/pcmcia/pd6729.c
475
map = mem->map;
drivers/pcmcia/pd6729.c
481
if ((mem->res->start > mem->res->end) || (mem->speed > 1000)) {
drivers/pcmcia/pd6729.c
492
i = (mem->res->start >> 12) & 0x0fff;
drivers/pcmcia/pd6729.c
493
if (mem->flags & MAP_16BIT)
drivers/pcmcia/pd6729.c
495
if (mem->flags & MAP_0WS)
drivers/pcmcia/pd6729.c
501
i = (mem->res->end >> 12) & 0x0fff;
drivers/pcmcia/pd6729.c
502
switch (to_cycles(mem->speed)) {
drivers/pcmcia/pd6729.c
520
indirect_write(socket, PD67_EXT_DATA, mem->res->start >> 24);
drivers/pcmcia/pd6729.c
524
i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff;
drivers/pcmcia/pd6729.c
525
if (mem->flags & MAP_WRPROT)
drivers/pcmcia/pd6729.c
527
if (mem->flags & MAP_ATTRIB) {
drivers/pcmcia/pd6729.c
538
if (mem->flags & MAP_ACTIVE)
drivers/pcmcia/pd6729.c
549
pccard_mem_map mem = { .res = &res, };
drivers/pcmcia/pd6729.c
557
mem.map = i;
drivers/pcmcia/pd6729.c
558
pd6729_set_mem_map(sock, &mem);
drivers/pcmcia/pxa2xx_base.c
168
pxa2xx_pcmcia_mcmem(sock, timing.mem, clk),
drivers/pcmcia/sa11xx_base.c
159
p+=sprintf(p, "common : %uns (%uns)\n", timing.mem,
drivers/pcmcia/sa11xx_base.c
89
bs_mem = skt->ops->get_timing(skt, cpu_clock, timing.mem);
drivers/pcmcia/soc_common.c
138
timing->mem =
drivers/pcmcia/soc_common.h
31
unsigned short mem;
drivers/pcmcia/tcic.c
731
static int tcic_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem)
drivers/pcmcia/tcic.c
738
"%#llx-%#llx, %#x)\n", psock, mem->map, mem->flags,
drivers/pcmcia/tcic.c
739
mem->speed, (unsigned long long)mem->res->start,
drivers/pcmcia/tcic.c
740
(unsigned long long)mem->res->end, mem->card_start);
drivers/pcmcia/tcic.c
741
if ((mem->map > 3) || (mem->card_start > 0x3ffffff) ||
drivers/pcmcia/tcic.c
742
(mem->res->start > 0xffffff) || (mem->res->end > 0xffffff) ||
drivers/pcmcia/tcic.c
743
(mem->res->start > mem->res->end) || (mem->speed > 1000))
drivers/pcmcia/tcic.c
746
addr = TCIC_MWIN(psock, mem->map);
drivers/pcmcia/tcic.c
748
base = mem->res->start; len = mem->res->end - mem->res->start;
drivers/pcmcia/tcic.c
757
mmap = mem->card_start - mem->res->start;
drivers/pcmcia/tcic.c
759
if (mem->flags & MAP_ATTRIB) mmap |= TCIC_MMAP_REG;
drivers/pcmcia/tcic.c
764
ctl |= to_cycles(mem->speed) & TCIC_MCTL_WSCNT_MASK;
drivers/pcmcia/tcic.c
765
ctl |= (mem->flags & MAP_16BIT) ? 0 : TCIC_MCTL_B8;
drivers/pcmcia/tcic.c
766
ctl |= (mem->flags & MAP_WRPROT) ? TCIC_MCTL_WP : 0;
drivers/pcmcia/tcic.c
767
ctl |= (mem->flags & MAP_ACTIVE) ? TCIC_MCTL_ENA : 0;
drivers/pcmcia/tcic.c
781
pccard_mem_map mem = { .res = &res, };
drivers/pcmcia/tcic.c
788
mem.map = i;
drivers/pcmcia/tcic.c
789
tcic_set_mem_map(s, &mem);
drivers/pcmcia/yenta_socket.c
442
static int yenta_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem)
drivers/pcmcia/yenta_socket.c
451
pcibios_resource_to_bus(socket->dev->bus, &region, mem->res);
drivers/pcmcia/yenta_socket.c
453
map = mem->map;
drivers/pcmcia/yenta_socket.c
456
card_start = mem->card_start;
drivers/pcmcia/yenta_socket.c
459
(card_start >> 26) || mem->speed > 1000)
drivers/pcmcia/yenta_socket.c
472
if (mem->flags & MAP_16BIT)
drivers/pcmcia/yenta_socket.c
474
if (mem->flags & MAP_0WS)
drivers/pcmcia/yenta_socket.c
479
switch (to_cycles(mem->speed)) {
drivers/pcmcia/yenta_socket.c
495
if (mem->flags & MAP_WRPROT)
drivers/pcmcia/yenta_socket.c
497
if (mem->flags & MAP_ATTRIB)
drivers/pcmcia/yenta_socket.c
501
if (mem->flags & MAP_ACTIVE)
drivers/pcmcia/yenta_socket.c
555
pccard_mem_map mem = { .res = &res, };
drivers/pcmcia/yenta_socket.c
563
mem.map = i;
drivers/pcmcia/yenta_socket.c
564
yenta_set_mem_map(&socket->socket, &mem);
drivers/phy/microchip/lan966x_serdes.c
50
static void lan_rmw_(u32 val, u32 mask, void __iomem *mem, u32 offset)
drivers/phy/microchip/lan966x_serdes.c
54
v = readl(mem + offset);
drivers/phy/microchip/lan966x_serdes.c
56
writel(v, mem + offset);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2253
void __iomem *mem;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2259
mem = devm_ioremap(dev, res.start, resource_size(&res));
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2260
if (!mem)
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2262
return mem;
drivers/pinctrl/pinctrl-equilibrium.c
308
void __iomem *mem;
drivers/pinctrl/pinctrl-equilibrium.c
315
mem = bank->membase;
drivers/pinctrl/pinctrl-equilibrium.c
326
writel(pmx, mem + (offset * 4));
drivers/pinctrl/pinctrl-equilibrium.c
373
static int get_drv_cur(void __iomem *mem, unsigned int offset)
drivers/pinctrl/pinctrl-equilibrium.c
378
return PARSE_DRV_CURRENT(readl(mem + REG_DRCC(idx)), pin_offset);
drivers/pinctrl/pinctrl-equilibrium.c
404
void __iomem *mem;
drivers/pinctrl/pinctrl-equilibrium.c
412
mem = bank->membase;
drivers/pinctrl/pinctrl-equilibrium.c
425
val = !!(readl(mem + REG_PUEN) & BIT(offset));
drivers/pinctrl/pinctrl-equilibrium.c
428
val = !!(readl(mem + REG_PDEN) & BIT(offset));
drivers/pinctrl/pinctrl-equilibrium.c
431
val = !!(readl(mem + REG_OD) & BIT(offset));
drivers/pinctrl/pinctrl-equilibrium.c
434
val = get_drv_cur(mem, offset);
drivers/pinctrl/pinctrl-equilibrium.c
437
val = !!(readl(mem + REG_SRC) & BIT(offset));
drivers/pinctrl/pinctrl-equilibrium.c
469
void __iomem *mem;
drivers/pinctrl/pinctrl-equilibrium.c
483
mem = bank->membase;
drivers/pinctrl/pinctrl-equilibrium.c
488
mem += REG_PUEN;
drivers/pinctrl/pinctrl-equilibrium.c
492
mem += REG_PDEN;
drivers/pinctrl/pinctrl-equilibrium.c
496
mem += REG_OD;
drivers/pinctrl/pinctrl-equilibrium.c
500
mem += REG_DRCC(offset / DRV_CUR_PINS);
drivers/pinctrl/pinctrl-equilibrium.c
505
mem += REG_SRC;
drivers/pinctrl/pinctrl-equilibrium.c
523
regval = readl(mem);
drivers/pinctrl/pinctrl-equilibrium.c
525
writel(regval, mem);
drivers/pinctrl/pinctrl-falcon.c
246
void __iomem *mem = info->membase[PORT(pin)];
drivers/pinctrl/pinctrl-falcon.c
251
!!pad_getbit(mem, LTQ_PADC_DCC, PORT_PIN(pin)));
drivers/pinctrl/pinctrl-falcon.c
256
!!pad_getbit(mem, LTQ_PADC_SRC, PORT_PIN(pin)));
drivers/pinctrl/pinctrl-falcon.c
260
if (pad_getbit(mem, LTQ_PADC_PDEN, PORT_PIN(pin)))
drivers/pinctrl/pinctrl-falcon.c
262
else if (pad_getbit(mem, LTQ_PADC_PUEN, PORT_PIN(pin)))
drivers/pinctrl/pinctrl-falcon.c
283
void __iomem *mem = info->membase[PORT(pin)];
drivers/pinctrl/pinctrl-falcon.c
313
pad_w32(mem, BIT(PORT_PIN(pin)), reg);
drivers/pinctrl/pinctrl-falcon.c
314
if (!(pad_r32(mem, reg) & BIT(PORT_PIN(pin))))
drivers/pinctrl/renesas/gpio.c
296
struct sh_pfc_window *mem)
drivers/pinctrl/renesas/gpio.c
305
chip->mem = mem;
drivers/pinctrl/renesas/gpio.c
32
struct sh_pfc_window *mem;
drivers/pinctrl/renesas/gpio.c
58
void __iomem *mem = address - chip->mem->phys + chip->mem->virt;
drivers/pinctrl/renesas/gpio.c
60
return sh_pfc_read_raw_reg(mem, dreg->reg_width);
drivers/pinctrl/renesas/gpio.c
67
void __iomem *mem = address - chip->mem->phys + chip->mem->virt;
drivers/pinctrl/renesas/gpio.c
69
sh_pfc_write_raw_reg(mem, dreg->reg_width, value);
drivers/pinctrl/renesas/pinctrl-rza1.c
48
#define RZA1_ADDR(mem, reg, port) ((mem) + (reg) + ((port) * 4))
drivers/pinctrl/renesas/pinctrl-rza1.c
576
void __iomem *mem = RZA1_ADDR(port->base, reg, port->id);
drivers/pinctrl/renesas/pinctrl-rza1.c
577
u16 val = ioread16(mem);
drivers/pinctrl/renesas/pinctrl-rza1.c
584
iowrite16(val, mem);
drivers/pinctrl/renesas/pinctrl-rza1.c
590
void __iomem *mem = RZA1_ADDR(port->base, reg, port->id);
drivers/pinctrl/renesas/pinctrl-rza1.c
592
return !!(ioread16(mem) & BIT(bit));
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3978
const u8 *mem = void_mem;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3988
s += scnprintf(s, 4, "%02x ", mem[offset]);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3995
u8 ch = mem[offset];
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
4007
mem += 16;
drivers/platform/x86/asus-armoury.c
683
u32 mem;
drivers/platform/x86/asus-armoury.c
685
err = armoury_get_devstate(attr, &mem, ASUS_WMI_DEVID_APU_MEM);
drivers/platform/x86/asus-armoury.c
690
if (mem == 0x100)
drivers/platform/x86/asus-armoury.c
694
if (apu_mem_map[i] == mem)
drivers/platform/x86/asus-armoury.c
698
pr_warn("Unrecognised value for APU mem 0x%08x\n", mem);
drivers/platform/x86/asus-armoury.c
706
u32 requested, mem;
drivers/platform/x86/asus-armoury.c
714
mem = apu_mem_map[requested];
drivers/platform/x86/asus-armoury.c
716
err = armoury_set_devstate(attr, mem, NULL, ASUS_WMI_DEVID_APU_MEM);
drivers/platform/x86/asus-armoury.c
718
pr_warn("Failed to set apu_mem 0x%x: %d\n", mem, err);
drivers/platform/x86/intel/ehl_pse_io.c
41
data->mem = DEFINE_RES_MEM(start + offset, EHL_PSE_IO_DEV_SIZE);
drivers/platform/x86/intel/telemetry/pltdrv.c
878
void __iomem *mem;
drivers/platform/x86/intel/telemetry/pltdrv.c
889
mem = devm_platform_ioremap_resource(pdev, 0);
drivers/platform/x86/intel/telemetry/pltdrv.c
890
if (IS_ERR(mem))
drivers/platform/x86/intel/telemetry/pltdrv.c
891
return PTR_ERR(mem);
drivers/platform/x86/intel/telemetry/pltdrv.c
893
telm_conf->pss_config.regmap = mem;
drivers/platform/x86/intel/telemetry/pltdrv.c
895
mem = devm_platform_ioremap_resource(pdev, 1);
drivers/platform/x86/intel/telemetry/pltdrv.c
896
if (IS_ERR(mem))
drivers/platform/x86/intel/telemetry/pltdrv.c
897
return PTR_ERR(mem);
drivers/platform/x86/intel/telemetry/pltdrv.c
899
telm_conf->ioss_config.regmap = mem;
drivers/platform/x86/intel/vsec_tpmi.c
411
void __iomem *mem;
drivers/platform/x86/intel/vsec_tpmi.c
431
mem = ioremap(off, size);
drivers/platform/x86/intel/vsec_tpmi.c
432
if (!mem) {
drivers/platform/x86/intel/vsec_tpmi.c
437
memcpy_fromio(buffer, mem, size);
drivers/platform/x86/intel/vsec_tpmi.c
442
iounmap(mem);
drivers/platform/x86/intel/vsec_tpmi.c
461
void __iomem *mem;
drivers/platform/x86/intel/vsec_tpmi.c
494
mem = ioremap(pfs->vsec_offset + punit * size, size);
drivers/platform/x86/intel/vsec_tpmi.c
495
if (!mem) {
drivers/platform/x86/intel/vsec_tpmi.c
500
writel(value, mem + addr);
drivers/platform/x86/intel/vsec_tpmi.c
502
iounmap(mem);
drivers/platform/x86/intel/vsec_tpmi.c
562
void __iomem *mem;
drivers/platform/x86/intel/vsec_tpmi.c
569
mem = devm_ioremap(&auxdev->dev, pfs->vsec_offset, size);
drivers/platform/x86/intel/vsec_tpmi.c
570
if (!mem)
drivers/platform/x86/intel/vsec_tpmi.c
574
tpmi_info->tpmi_control_mem = mem;
drivers/platform/x86/intel_scu_ipc.c
545
release_mem_region(data->mem.start, resource_size(&data->mem));
drivers/platform/x86/intel_scu_ipc.c
588
if (!request_mem_region(data->mem.start, resource_size(&data->mem), "intel_scu_ipc")) {
drivers/platform/x86/intel_scu_ipc.c
593
ipc_base = ioremap(data->mem.start, resource_size(&data->mem));
drivers/platform/x86/intel_scu_ipc.c
626
release_mem_region(data->mem.start, resource_size(&data->mem));
drivers/platform/x86/intel_scu_pcidrv.c
27
scu_data.mem = pdev->resource[0];
drivers/platform/x86/intel_scu_pltdrv.c
31
scu_data.mem = *res;
drivers/platform/x86/p2sb.c
180
struct resource *mem)
drivers/platform/x86/p2sb.c
190
memcpy(mem, &cache->res, sizeof(*mem));
drivers/platform/x86/p2sb.c
196
struct resource *mem)
drivers/platform/x86/p2sb.c
206
p2sb_read_bar0(pdev, mem);
drivers/platform/x86/p2sb.c
229
int p2sb_bar(struct pci_bus *bus, unsigned int devfn, struct resource *mem)
drivers/platform/x86/p2sb.c
239
return p2sb_read_from_cache(bus, devfn, mem);
drivers/platform/x86/p2sb.c
241
return p2sb_read_from_dev(bus, devfn, mem);
drivers/platform/x86/p2sb.c
66
static void p2sb_read_bar0(struct pci_dev *pdev, struct resource *mem)
drivers/platform/x86/p2sb.c
71
memset(mem, 0, sizeof(*mem));
drivers/platform/x86/p2sb.c
78
mem->start = bar0->start;
drivers/platform/x86/p2sb.c
79
mem->end = bar0->end;
drivers/platform/x86/p2sb.c
80
mem->flags = bar0->flags;
drivers/platform/x86/p2sb.c
81
mem->desc = bar0->desc;
drivers/pmdomain/rockchip/pm-domains.c
155
#define DOMAIN_M_G_SD(_name, pwr, status, req, idle, ack, g_mask, mem, wakeup, keepon) \
drivers/pmdomain/rockchip/pm-domains.c
241
#define DOMAIN_RK3562(name, pwr, req, g_mask, mem, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
242
DOMAIN_M_G_SD(name, pwr, pwr, req, req, req, g_mask, mem, wakeup, false)
drivers/pnp/base.h
74
struct pnp_mem mem;
drivers/pnp/interface.c
154
struct pnp_mem *mem)
drivers/pnp/interface.c
159
space, (unsigned long long) mem->min,
drivers/pnp/interface.c
160
(unsigned long long) mem->max,
drivers/pnp/interface.c
161
(unsigned long long) mem->align,
drivers/pnp/interface.c
162
(unsigned long long) mem->size);
drivers/pnp/interface.c
163
if (mem->flags & IORESOURCE_MEM_WRITEABLE)
drivers/pnp/interface.c
165
if (mem->flags & IORESOURCE_MEM_CACHEABLE)
drivers/pnp/interface.c
167
if (mem->flags & IORESOURCE_MEM_RANGELENGTH)
drivers/pnp/interface.c
169
if (mem->flags & IORESOURCE_MEM_SHADOWABLE)
drivers/pnp/interface.c
171
if (mem->flags & IORESOURCE_MEM_EXPANSIONROM)
drivers/pnp/interface.c
173
switch (mem->flags & IORESOURCE_MEM_TYPE_MASK) {
drivers/pnp/interface.c
197
pnp_print_mem(buffer, space, &option->u.mem);
drivers/pnp/manager.c
277
ret = pnp_assign_mem(dev, &option->u.mem, nmem++);
drivers/pnp/pnpacpi/rsparser.c
419
if (p->info.mem.write_protect == ACPI_READ_WRITE_MEMORY)
drivers/pnp/pnpacpi/rsparser.c
438
if (p->info.mem.write_protect == ACPI_READ_WRITE_MEMORY)
drivers/pnp/pnpacpi/rsparser.c
883
unsigned int port = 0, irq = 0, dma = 0, mem = 0;
drivers/pnp/pnpacpi/rsparser.c
911
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpacpi/rsparser.c
912
mem++;
drivers/pnp/pnpacpi/rsparser.c
916
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpacpi/rsparser.c
917
mem++;
drivers/pnp/pnpacpi/rsparser.c
921
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpacpi/rsparser.c
922
mem++;
drivers/pnp/pnpbios/rsparser.c
664
int port = 0, irq = 0, dma = 0, mem = 0;
drivers/pnp/pnpbios/rsparser.c
686
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpbios/rsparser.c
687
mem++;
drivers/pnp/pnpbios/rsparser.c
694
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpbios/rsparser.c
695
mem++;
drivers/pnp/pnpbios/rsparser.c
702
pnp_get_resource(dev, IORESOURCE_MEM, mem));
drivers/pnp/pnpbios/rsparser.c
703
mem++;
drivers/pnp/resource.c
127
struct pnp_mem *mem;
drivers/pnp/resource.c
133
mem = &option->u.mem;
drivers/pnp/resource.c
134
mem->min = min;
drivers/pnp/resource.c
135
mem->max = max;
drivers/pnp/resource.c
136
mem->align = align;
drivers/pnp/resource.c
137
mem->size = size;
drivers/pnp/resource.c
138
mem->flags = flags;
drivers/pnp/resource.c
650
struct pnp_mem *mem;
drivers/pnp/resource.c
665
mem = &option->u.mem;
drivers/pnp/resource.c
666
if (mem->min == start && mem->size == size)
drivers/pnp/support.c
110
struct pnp_mem *mem;
drivers/pnp/support.c
134
mem = &option->u.mem;
drivers/pnp/support.c
137
(unsigned long long) mem->min,
drivers/pnp/support.c
138
(unsigned long long) mem->max,
drivers/pnp/support.c
139
(unsigned long long) mem->align,
drivers/pnp/support.c
140
(unsigned long long) mem->size, mem->flags);
drivers/ptp/ptp_ocp.c
2033
struct signal_reg __iomem *reg = ext->mem;
drivers/ptp/ptp_ocp.c
2123
struct signal_reg __iomem *reg = ext->mem;
drivers/ptp/ptp_ocp.c
2164
struct ts_reg __iomem *reg = ext->mem;
drivers/ptp/ptp_ocp.c
2198
struct ts_reg __iomem *reg = ext->mem;
drivers/ptp/ptp_ocp.c
2248
ext->mem = ptp_ocp_get_mem(bp, r);
drivers/ptp/ptp_ocp.c
2249
if (IS_ERR(ext->mem)) {
drivers/ptp/ptp_ocp.c
2250
err = PTR_ERR(ext->mem);
drivers/ptp/ptp_ocp.c
2317
void __iomem *mem;
drivers/ptp/ptp_ocp.c
2319
mem = ptp_ocp_get_mem(bp, r);
drivers/ptp/ptp_ocp.c
2320
if (IS_ERR(mem))
drivers/ptp/ptp_ocp.c
2321
return PTR_ERR(mem);
drivers/ptp/ptp_ocp.c
2323
bp_assign_entry(bp, r, mem);
drivers/ptp/ptp_ocp.c
2359
bp->signal_out[i]->mem);
drivers/ptp/ptp_ocp.c
262
void __iomem *mem;
drivers/ptp/ptp_ocp.c
4022
struct signal_reg __iomem *reg = bp->signal_out[nr]->mem;
drivers/ptp/ptp_ocp.c
4138
ts_reg = bp->ts0->mem;
drivers/ptp/ptp_ocp.c
4146
ts_reg = bp->ts1->mem;
drivers/ptp/ptp_ocp.c
4154
ts_reg = bp->ts2->mem;
drivers/ptp/ptp_ocp.c
4162
ts_reg = bp->ts3->mem;
drivers/ptp/ptp_ocp.c
4170
ts_reg = bp->ts4->mem;
drivers/ptp/ptp_ocp.c
4178
ts_reg = bp->pps->mem;
drivers/remoteproc/da8xx_remoteproc.c
207
drproc->mem = devm_kcalloc(dev, num_mems, sizeof(*drproc->mem),
drivers/remoteproc/da8xx_remoteproc.c
209
if (!drproc->mem)
drivers/remoteproc/da8xx_remoteproc.c
215
drproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
drivers/remoteproc/da8xx_remoteproc.c
216
if (IS_ERR(drproc->mem[i].cpu_addr)) {
drivers/remoteproc/da8xx_remoteproc.c
219
return PTR_ERR(drproc->mem[i].cpu_addr);
drivers/remoteproc/da8xx_remoteproc.c
221
drproc->mem[i].bus_addr = res->start;
drivers/remoteproc/da8xx_remoteproc.c
222
drproc->mem[i].dev_addr =
drivers/remoteproc/da8xx_remoteproc.c
224
drproc->mem[i].size = resource_size(res);
drivers/remoteproc/da8xx_remoteproc.c
227
mem_names[i], &drproc->mem[i].bus_addr,
drivers/remoteproc/da8xx_remoteproc.c
228
drproc->mem[i].size, drproc->mem[i].cpu_addr,
drivers/remoteproc/da8xx_remoteproc.c
229
drproc->mem[i].dev_addr);
drivers/remoteproc/da8xx_remoteproc.c
68
struct da8xx_rproc_mem *mem;
drivers/remoteproc/imx_dsp_rproc.c
648
struct rproc_mem_entry *mem)
drivers/remoteproc/imx_dsp_rproc.c
653
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/imx_dsp_rproc.c
656
&mem->dma, mem->len);
drivers/remoteproc/imx_dsp_rproc.c
660
mem->va = va;
drivers/remoteproc/imx_dsp_rproc.c
666
struct rproc_mem_entry *mem)
drivers/remoteproc/imx_dsp_rproc.c
668
iounmap(mem->va);
drivers/remoteproc/imx_dsp_rproc.c
687
struct rproc_mem_entry *mem;
drivers/remoteproc/imx_dsp_rproc.c
702
mem = rproc_mem_entry_init(dev, NULL, (dma_addr_t)att->sa,
drivers/remoteproc/imx_dsp_rproc.c
706
if (mem)
drivers/remoteproc/imx_dsp_rproc.c
711
rproc_add_carveout(rproc, mem);
drivers/remoteproc/imx_dsp_rproc.c
733
mem = rproc_mem_entry_init(dev, NULL, (dma_addr_t)res.start,
drivers/remoteproc/imx_dsp_rproc.c
738
if (!mem)
drivers/remoteproc/imx_dsp_rproc.c
742
rproc_add_carveout(rproc, mem);
drivers/remoteproc/imx_rproc.c
113
struct imx_rproc_mem mem[IMX_RPROC_MEM_MAX];
drivers/remoteproc/imx_rproc.c
531
if (sys >= priv->mem[i].sys_addr && sys + len <
drivers/remoteproc/imx_rproc.c
532
priv->mem[i].sys_addr + priv->mem[i].size) {
drivers/remoteproc/imx_rproc.c
533
unsigned int offset = sys - priv->mem[i].sys_addr;
drivers/remoteproc/imx_rproc.c
535
va = (__force void *)(priv->mem[i].cpu_addr + offset);
drivers/remoteproc/imx_rproc.c
547
struct rproc_mem_entry *mem)
drivers/remoteproc/imx_rproc.c
552
dev_dbg(dev, "map memory: %p+%zx\n", &mem->dma, mem->len);
drivers/remoteproc/imx_rproc.c
553
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/imx_rproc.c
556
&mem->dma, mem->len);
drivers/remoteproc/imx_rproc.c
561
mem->va = va;
drivers/remoteproc/imx_rproc.c
567
struct rproc_mem_entry *mem)
drivers/remoteproc/imx_rproc.c
569
dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
drivers/remoteproc/imx_rproc.c
570
iounmap(mem->va);
drivers/remoteproc/imx_rproc.c
609
struct rproc_mem_entry *mem;
drivers/remoteproc/imx_rproc.c
636
mem = rproc_mem_entry_init(priv->dev, NULL, (dma_addr_t)res.start,
drivers/remoteproc/imx_rproc.c
641
if (!mem)
drivers/remoteproc/imx_rproc.c
645
rproc_add_carveout(rproc, mem);
drivers/remoteproc/imx_rproc.c
777
priv->mem[b].cpu_addr = devm_ioremap(&pdev->dev,
drivers/remoteproc/imx_rproc.c
780
priv->mem[b].cpu_addr = devm_ioremap_wc(&pdev->dev,
drivers/remoteproc/imx_rproc.c
782
if (!priv->mem[b].cpu_addr) {
drivers/remoteproc/imx_rproc.c
786
priv->mem[b].sys_addr = att->sa;
drivers/remoteproc/imx_rproc.c
787
priv->mem[b].size = att->size;
drivers/remoteproc/imx_rproc.c
814
priv->mem[b].cpu_addr = devm_ioremap_resource_wc(&pdev->dev, &res);
drivers/remoteproc/imx_rproc.c
815
if (!priv->mem[b].cpu_addr) {
drivers/remoteproc/imx_rproc.c
819
priv->mem[b].sys_addr = res.start;
drivers/remoteproc/imx_rproc.c
820
priv->mem[b].size = resource_size(&res);
drivers/remoteproc/imx_rproc.c
822
priv->rsc_table = priv->mem[b].cpu_addr;
drivers/remoteproc/ingenic_rproc.c
169
struct resource *mem;
drivers/remoteproc/ingenic_rproc.c
193
mem = platform_get_resource_byname(pdev, IORESOURCE_MEM,
drivers/remoteproc/ingenic_rproc.c
196
vpu->mem_info[i].base = devm_ioremap_resource(dev, mem);
drivers/remoteproc/ingenic_rproc.c
203
vpu->mem_info[i].len = resource_size(mem);
drivers/remoteproc/keystone_remoteproc.c
262
bus_addr = ksproc->mem[i].bus_addr;
drivers/remoteproc/keystone_remoteproc.c
263
dev_addr = ksproc->mem[i].dev_addr;
drivers/remoteproc/keystone_remoteproc.c
264
size = ksproc->mem[i].size;
drivers/remoteproc/keystone_remoteproc.c
271
va = ksproc->mem[i].cpu_addr + offset;
drivers/remoteproc/keystone_remoteproc.c
279
va = ksproc->mem[i].cpu_addr + offset;
drivers/remoteproc/keystone_remoteproc.c
305
ksproc->mem = devm_kcalloc(ksproc->dev, num_mems,
drivers/remoteproc/keystone_remoteproc.c
306
sizeof(*ksproc->mem), GFP_KERNEL);
drivers/remoteproc/keystone_remoteproc.c
307
if (!ksproc->mem)
drivers/remoteproc/keystone_remoteproc.c
313
ksproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
drivers/remoteproc/keystone_remoteproc.c
314
if (IS_ERR(ksproc->mem[i].cpu_addr)) {
drivers/remoteproc/keystone_remoteproc.c
317
return PTR_ERR(ksproc->mem[i].cpu_addr);
drivers/remoteproc/keystone_remoteproc.c
319
ksproc->mem[i].bus_addr = res->start;
drivers/remoteproc/keystone_remoteproc.c
320
ksproc->mem[i].dev_addr =
drivers/remoteproc/keystone_remoteproc.c
322
ksproc->mem[i].size = resource_size(res);
drivers/remoteproc/keystone_remoteproc.c
325
memset((__force void *)ksproc->mem[i].cpu_addr, 0,
drivers/remoteproc/keystone_remoteproc.c
326
ksproc->mem[i].size);
drivers/remoteproc/keystone_remoteproc.c
58
struct keystone_rproc_mem *mem;
drivers/remoteproc/omap_remoteproc.c
111
struct omap_rproc_mem *mem;
drivers/remoteproc/omap_remoteproc.c
1188
oproc->mem = devm_kcalloc(dev, num_mems, sizeof(*oproc->mem),
drivers/remoteproc/omap_remoteproc.c
1190
if (!oproc->mem)
drivers/remoteproc/omap_remoteproc.c
1201
oproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
drivers/remoteproc/omap_remoteproc.c
1202
if (IS_ERR(oproc->mem[i].cpu_addr)) {
drivers/remoteproc/omap_remoteproc.c
1205
return PTR_ERR(oproc->mem[i].cpu_addr);
drivers/remoteproc/omap_remoteproc.c
1207
oproc->mem[i].bus_addr = res->start;
drivers/remoteproc/omap_remoteproc.c
1208
oproc->mem[i].dev_addr = data->mems[i].dev_addr;
drivers/remoteproc/omap_remoteproc.c
1209
oproc->mem[i].size = resource_size(res);
drivers/remoteproc/omap_remoteproc.c
1212
data->mems[i].name, &oproc->mem[i].bus_addr,
drivers/remoteproc/omap_remoteproc.c
1213
oproc->mem[i].size, oproc->mem[i].cpu_addr,
drivers/remoteproc/omap_remoteproc.c
1214
oproc->mem[i].dev_addr);
drivers/remoteproc/omap_remoteproc.c
747
if (da >= oproc->mem[i].dev_addr && da + len <=
drivers/remoteproc/omap_remoteproc.c
748
oproc->mem[i].dev_addr + oproc->mem[i].size) {
drivers/remoteproc/omap_remoteproc.c
749
offset = da - oproc->mem[i].dev_addr;
drivers/remoteproc/omap_remoteproc.c
751
return (__force void *)(oproc->mem[i].cpu_addr +
drivers/remoteproc/rcar_rproc.c
23
struct rproc_mem_entry *mem)
drivers/remoteproc/rcar_rproc.c
28
dev_dbg(dev, "map memory: %pa+%zx\n", &mem->dma, mem->len);
drivers/remoteproc/rcar_rproc.c
29
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/rcar_rproc.c
32
&mem->dma, mem->len);
drivers/remoteproc/rcar_rproc.c
37
mem->va = va;
drivers/remoteproc/rcar_rproc.c
43
struct rproc_mem_entry *mem)
drivers/remoteproc/rcar_rproc.c
45
dev_dbg(&rproc->dev, "unmap memory: %pa\n", &mem->dma);
drivers/remoteproc/rcar_rproc.c
46
iounmap(mem->va);
drivers/remoteproc/rcar_rproc.c
55
struct rproc_mem_entry *mem;
drivers/remoteproc/rcar_rproc.c
73
mem = rproc_mem_entry_init(dev, NULL,
drivers/remoteproc/rcar_rproc.c
80
if (!mem)
drivers/remoteproc/rcar_rproc.c
83
rproc_add_carveout(rproc, mem);
drivers/remoteproc/remoteproc_core.c
252
struct rproc_mem_entry *carveout, *mem = NULL;
drivers/remoteproc/remoteproc_core.c
264
mem = carveout;
drivers/remoteproc/remoteproc_core.c
269
return mem;
drivers/remoteproc/remoteproc_core.c
287
struct rproc_mem_entry *mem, u32 da, u32 len)
drivers/remoteproc/remoteproc_core.c
293
if (len > mem->len) {
drivers/remoteproc/remoteproc_core.c
298
if (da != FW_RSC_ADDR_ANY && mem->da == FW_RSC_ADDR_ANY) {
drivers/remoteproc/remoteproc_core.c
301
} else if (da != FW_RSC_ADDR_ANY && mem->da != FW_RSC_ADDR_ANY) {
drivers/remoteproc/remoteproc_core.c
302
delta = da - mem->da;
drivers/remoteproc/remoteproc_core.c
311
if (delta + len > mem->len) {
drivers/remoteproc/remoteproc_core.c
328
struct rproc_mem_entry *mem;
drivers/remoteproc/remoteproc_core.c
337
mem = rproc_find_carveout_by_name(rproc, "vdev%dvring%d", rvdev->index,
drivers/remoteproc/remoteproc_core.c
339
if (mem) {
drivers/remoteproc/remoteproc_core.c
340
if (rproc_check_carveout_da(rproc, mem, rsc->vring[i].da, size))
drivers/remoteproc/remoteproc_core.c
344
mem = rproc_mem_entry_init(dev, NULL, 0,
drivers/remoteproc/remoteproc_core.c
350
if (!mem) {
drivers/remoteproc/remoteproc_core.c
355
rproc_add_carveout(rproc, mem);
drivers/remoteproc/remoteproc_core.c
51
struct rproc_mem_entry *mem);
drivers/remoteproc/remoteproc_core.c
53
struct rproc_mem_entry *mem);
drivers/remoteproc/remoteproc_core.c
681
struct rproc_mem_entry *mem)
drivers/remoteproc/remoteproc_core.c
689
va = dma_alloc_coherent(dev->parent, mem->len, &dma, GFP_KERNEL);
drivers/remoteproc/remoteproc_core.c
693
mem->len);
drivers/remoteproc/remoteproc_core.c
698
va, &dma, mem->len);
drivers/remoteproc/remoteproc_core.c
700
if (mem->da != FW_RSC_ADDR_ANY && !rproc->domain) {
drivers/remoteproc/remoteproc_core.c
707
if (mem->da != (u32)dma)
drivers/remoteproc/remoteproc_core.c
729
if (mem->da != FW_RSC_ADDR_ANY && rproc->domain) {
drivers/remoteproc/remoteproc_core.c
736
ret = iommu_map(rproc->domain, mem->da, dma, mem->len,
drivers/remoteproc/remoteproc_core.c
737
mem->flags, GFP_KERNEL);
drivers/remoteproc/remoteproc_core.c
750
mapping->da = mem->da;
drivers/remoteproc/remoteproc_core.c
751
mapping->len = mem->len;
drivers/remoteproc/remoteproc_core.c
755
mem->da, &dma);
drivers/remoteproc/remoteproc_core.c
758
if (mem->da == FW_RSC_ADDR_ANY) {
drivers/remoteproc/remoteproc_core.c
763
mem->da = (u32)dma;
drivers/remoteproc/remoteproc_core.c
766
mem->dma = dma;
drivers/remoteproc/remoteproc_core.c
767
mem->va = va;
drivers/remoteproc/remoteproc_core.c
774
dma_free_coherent(dev->parent, mem->len, va, dma);
drivers/remoteproc/remoteproc_core.c
789
struct rproc_mem_entry *mem)
drivers/remoteproc/remoteproc_core.c
794
dma_free_coherent(dev->parent, mem->len, mem->va, mem->dma);
drivers/remoteproc/remoteproc_core.c
887
void rproc_add_carveout(struct rproc *rproc, struct rproc_mem_entry *mem)
drivers/remoteproc/remoteproc_core.c
889
list_add_tail(&mem->node, &rproc->carveouts);
drivers/remoteproc/remoteproc_core.c
917
struct rproc_mem_entry *mem;
drivers/remoteproc/remoteproc_core.c
920
mem = kzalloc_obj(*mem);
drivers/remoteproc/remoteproc_core.c
921
if (!mem)
drivers/remoteproc/remoteproc_core.c
922
return mem;
drivers/remoteproc/remoteproc_core.c
924
mem->va = va;
drivers/remoteproc/remoteproc_core.c
925
mem->dma = dma;
drivers/remoteproc/remoteproc_core.c
926
mem->da = da;
drivers/remoteproc/remoteproc_core.c
927
mem->len = len;
drivers/remoteproc/remoteproc_core.c
928
mem->alloc = alloc;
drivers/remoteproc/remoteproc_core.c
929
mem->release = release;
drivers/remoteproc/remoteproc_core.c
930
mem->rsc_offset = FW_RSC_ADDR_ANY;
drivers/remoteproc/remoteproc_core.c
931
mem->of_resm_idx = -1;
drivers/remoteproc/remoteproc_core.c
934
vsnprintf(mem->name, sizeof(mem->name), name, args);
drivers/remoteproc/remoteproc_core.c
937
return mem;
drivers/remoteproc/remoteproc_core.c
960
struct rproc_mem_entry *mem;
drivers/remoteproc/remoteproc_core.c
963
mem = kzalloc_obj(*mem);
drivers/remoteproc/remoteproc_core.c
964
if (!mem)
drivers/remoteproc/remoteproc_core.c
965
return mem;
drivers/remoteproc/remoteproc_core.c
967
mem->da = da;
drivers/remoteproc/remoteproc_core.c
968
mem->len = len;
drivers/remoteproc/remoteproc_core.c
969
mem->rsc_offset = FW_RSC_ADDR_ANY;
drivers/remoteproc/remoteproc_core.c
970
mem->of_resm_idx = of_resm_idx;
drivers/remoteproc/remoteproc_core.c
973
vsnprintf(mem->name, sizeof(mem->name), name, args);
drivers/remoteproc/remoteproc_core.c
976
return mem;
drivers/remoteproc/remoteproc_virtio.c
111
struct rproc_mem_entry *mem;
drivers/remoteproc/remoteproc_virtio.c
126
mem = rproc_find_carveout_by_name(rproc, "vdev%dvring%d", rvdev->index,
drivers/remoteproc/remoteproc_virtio.c
128
if (!mem || !mem->va)
drivers/remoteproc/remoteproc_virtio.c
132
addr = mem->va;
drivers/remoteproc/remoteproc_virtio.c
161
rsc->vring[id].da = mem->da;
drivers/remoteproc/remoteproc_virtio.c
375
struct rproc_mem_entry *mem;
drivers/remoteproc/remoteproc_virtio.c
385
mem = rproc_find_carveout_by_name(rproc, "vdev%dbuffer", rvdev->index);
drivers/remoteproc/remoteproc_virtio.c
386
if (mem) {
drivers/remoteproc/remoteproc_virtio.c
389
if (mem->of_resm_idx != -1) {
drivers/remoteproc/remoteproc_virtio.c
394
mem->of_resm_idx);
drivers/remoteproc/remoteproc_virtio.c
400
if (mem->va) {
drivers/remoteproc/remoteproc_virtio.c
403
pa = rproc_va_to_pa(mem->va);
drivers/remoteproc/remoteproc_virtio.c
406
pa = (phys_addr_t)mem->dma;
drivers/remoteproc/remoteproc_virtio.c
411
mem->da,
drivers/remoteproc/remoteproc_virtio.c
412
mem->len);
drivers/remoteproc/st_remoteproc.c
100
&mem->dma, mem->len);
drivers/remoteproc/st_remoteproc.c
105
mem->va = va;
drivers/remoteproc/st_remoteproc.c
111
struct rproc_mem_entry *mem)
drivers/remoteproc/st_remoteproc.c
113
iounmap(mem->va);
drivers/remoteproc/st_remoteproc.c
122
struct rproc_mem_entry *mem;
drivers/remoteproc/st_remoteproc.c
138
mem = rproc_mem_entry_init(dev, NULL,
drivers/remoteproc/st_remoteproc.c
148
mem = rproc_of_resm_mem_entry_init(dev, index,
drivers/remoteproc/st_remoteproc.c
154
if (!mem)
drivers/remoteproc/st_remoteproc.c
157
rproc_add_carveout(rproc, mem);
drivers/remoteproc/st_remoteproc.c
92
struct rproc_mem_entry *mem)
drivers/remoteproc/st_remoteproc.c
97
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/st_slim_rproc.c
143
fw_rev = readl(slim_rproc->mem[ST_SLIM_DMEM].cpu_addr +
drivers/remoteproc/st_slim_rproc.c
183
if (da != slim_rproc->mem[i].bus_addr)
drivers/remoteproc/st_slim_rproc.c
186
if (len <= slim_rproc->mem[i].size) {
drivers/remoteproc/st_slim_rproc.c
188
va = (__force void *)slim_rproc->mem[i].cpu_addr;
drivers/remoteproc/st_slim_rproc.c
252
slim_rproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
drivers/remoteproc/st_slim_rproc.c
253
if (IS_ERR(slim_rproc->mem[i].cpu_addr)) {
drivers/remoteproc/st_slim_rproc.c
255
err = PTR_ERR(slim_rproc->mem[i].cpu_addr);
drivers/remoteproc/st_slim_rproc.c
258
slim_rproc->mem[i].bus_addr = res->start;
drivers/remoteproc/st_slim_rproc.c
259
slim_rproc->mem[i].size = resource_size(res);
drivers/remoteproc/stm32_rproc.c
117
struct rproc_mem_entry *mem)
drivers/remoteproc/stm32_rproc.c
122
dev_dbg(dev, "map memory: %pad+%zx\n", &mem->dma, mem->len);
drivers/remoteproc/stm32_rproc.c
123
va = (__force void *)ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/stm32_rproc.c
126
&mem->dma, mem->len);
drivers/remoteproc/stm32_rproc.c
131
mem->va = va;
drivers/remoteproc/stm32_rproc.c
137
struct rproc_mem_entry *mem)
drivers/remoteproc/stm32_rproc.c
139
dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma);
drivers/remoteproc/stm32_rproc.c
140
iounmap((__force __iomem void *)mem->va);
drivers/remoteproc/stm32_rproc.c
216
struct rproc_mem_entry *mem;
drivers/remoteproc/stm32_rproc.c
237
mem = rproc_mem_entry_init(dev, NULL,
drivers/remoteproc/stm32_rproc.c
244
if (mem)
drivers/remoteproc/stm32_rproc.c
249
mem = rproc_of_resm_mem_entry_init(dev, index,
drivers/remoteproc/stm32_rproc.c
255
if (!mem) {
drivers/remoteproc/stm32_rproc.c
259
rproc_add_carveout(rproc, mem);
drivers/remoteproc/ti_k3_common.c
373
bus_addr = kproc->mem[i].bus_addr;
drivers/remoteproc/ti_k3_common.c
374
dev_addr = kproc->mem[i].dev_addr;
drivers/remoteproc/ti_k3_common.c
375
size = kproc->mem[i].size;
drivers/remoteproc/ti_k3_common.c
380
va = kproc->mem[i].cpu_addr + offset;
drivers/remoteproc/ti_k3_common.c
387
va = kproc->mem[i].cpu_addr + offset;
drivers/remoteproc/ti_k3_common.c
418
kproc->mem = devm_kcalloc(kproc->dev, num_mems,
drivers/remoteproc/ti_k3_common.c
419
sizeof(*kproc->mem), GFP_KERNEL);
drivers/remoteproc/ti_k3_common.c
420
if (!kproc->mem)
drivers/remoteproc/ti_k3_common.c
439
kproc->mem[i].cpu_addr = devm_ioremap_wc(dev, res->start,
drivers/remoteproc/ti_k3_common.c
441
if (!kproc->mem[i].cpu_addr) {
drivers/remoteproc/ti_k3_common.c
446
kproc->mem[i].bus_addr = res->start;
drivers/remoteproc/ti_k3_common.c
447
kproc->mem[i].dev_addr = data->mems[i].dev_addr;
drivers/remoteproc/ti_k3_common.c
448
kproc->mem[i].size = resource_size(res);
drivers/remoteproc/ti_k3_common.c
451
data->mems[i].name, &kproc->mem[i].bus_addr,
drivers/remoteproc/ti_k3_common.c
452
kproc->mem[i].size, kproc->mem[i].cpu_addr,
drivers/remoteproc/ti_k3_common.c
453
kproc->mem[i].dev_addr);
drivers/remoteproc/ti_k3_common.h
84
struct k3_rproc_mem *mem;
drivers/remoteproc/ti_k3_r5_remoteproc.c
377
memset_io(kproc->mem[0].cpu_addr, 0x00, kproc->mem[0].size);
drivers/remoteproc/ti_k3_r5_remoteproc.c
380
memset_io(kproc->mem[1].cpu_addr, 0x00, kproc->mem[1].size);
drivers/remoteproc/ti_k3_r5_remoteproc.c
791
WARN_ON(kproc->mem[0].size != SZ_64K);
drivers/remoteproc/ti_k3_r5_remoteproc.c
792
WARN_ON(kproc->mem[1].size != SZ_64K);
drivers/remoteproc/ti_k3_r5_remoteproc.c
794
kproc->mem[0].size /= 2;
drivers/remoteproc/ti_k3_r5_remoteproc.c
795
kproc->mem[1].size /= 2;
drivers/remoteproc/ti_k3_r5_remoteproc.c
798
kproc->mem[0].size, kproc->mem[1].size);
drivers/remoteproc/ti_k3_r5_remoteproc.c
912
kproc->mem[0].dev_addr = loczrama ? 0 : K3_R5_TCM_DEV_ADDR;
drivers/remoteproc/ti_k3_r5_remoteproc.c
913
kproc->mem[1].dev_addr = loczrama ? K3_R5_TCM_DEV_ADDR : 0;
drivers/remoteproc/ti_k3_r5_remoteproc.c
929
kproc->mem = devm_kcalloc(kproc->dev, num_mems, sizeof(*kproc->mem),
drivers/remoteproc/ti_k3_r5_remoteproc.c
931
if (!kproc->mem)
drivers/remoteproc/ti_k3_r5_remoteproc.c
949
kproc->mem[i].dev_addr = core->loczrama ?
drivers/remoteproc/ti_k3_r5_remoteproc.c
952
kproc->mem[i].dev_addr = core->loczrama ?
drivers/remoteproc/ti_k3_r5_remoteproc.c
957
&kproc->mem[i].bus_addr, data->mems[i].name);
drivers/remoteproc/wkup_m3_rproc.c
103
if (da >= wkupm3->mem[i].dev_addr && da + len <=
drivers/remoteproc/wkup_m3_rproc.c
104
wkupm3->mem[i].dev_addr + wkupm3->mem[i].size) {
drivers/remoteproc/wkup_m3_rproc.c
105
offset = da - wkupm3->mem[i].dev_addr;
drivers/remoteproc/wkup_m3_rproc.c
107
va = (__force void *)(wkupm3->mem[i].cpu_addr + offset);
drivers/remoteproc/wkup_m3_rproc.c
193
wkupm3->mem[i].cpu_addr = devm_ioremap_resource(dev, res);
drivers/remoteproc/wkup_m3_rproc.c
194
if (IS_ERR(wkupm3->mem[i].cpu_addr))
drivers/remoteproc/wkup_m3_rproc.c
195
return dev_err_probe(dev, PTR_ERR(wkupm3->mem[i].cpu_addr),
drivers/remoteproc/wkup_m3_rproc.c
197
wkupm3->mem[i].bus_addr = res->start;
drivers/remoteproc/wkup_m3_rproc.c
198
wkupm3->mem[i].size = resource_size(res);
drivers/remoteproc/wkup_m3_rproc.c
208
wkupm3->mem[i].dev_addr = be32_to_cpu(*addrp) - l4_offset;
drivers/remoteproc/wkup_m3_rproc.c
52
struct wkup_m3_mem mem[WKUPM3_MEM_MAX];
drivers/remoteproc/xlnx_r5_remoteproc.c
454
struct rproc_mem_entry *mem)
drivers/remoteproc/xlnx_r5_remoteproc.c
458
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/xlnx_r5_remoteproc.c
462
mem->va = (void *)va;
drivers/remoteproc/xlnx_r5_remoteproc.c
477
struct rproc_mem_entry *mem)
drivers/remoteproc/xlnx_r5_remoteproc.c
479
iounmap((void __iomem *)mem->va);
drivers/remoteproc/xlnx_r5_remoteproc.c
585
static int tcm_mem_unmap(struct rproc *rproc, struct rproc_mem_entry *mem)
drivers/remoteproc/xlnx_r5_remoteproc.c
587
iounmap((void __iomem *)mem->va);
drivers/remoteproc/xlnx_r5_remoteproc.c
603
struct rproc_mem_entry *mem)
drivers/remoteproc/xlnx_r5_remoteproc.c
607
va = ioremap_wc(mem->dma, mem->len);
drivers/remoteproc/xlnx_r5_remoteproc.c
612
mem->va = (void *)va;
drivers/remoteproc/xlnx_r5_remoteproc.c
615
memset_io(va, 0, mem->len);
drivers/reset/reset-microchip-sparx5.c
133
void __iomem *mem;
drivers/reset/reset-microchip-sparx5.c
135
mem = devm_platform_get_and_ioremap_resource(pdev, index, &res);
drivers/reset/reset-microchip-sparx5.c
136
if (IS_ERR(mem)) {
drivers/reset/reset-microchip-sparx5.c
138
return PTR_ERR(mem);
drivers/reset/reset-microchip-sparx5.c
141
map = devm_regmap_init_mmio(&pdev->dev, mem, &sparx5_reset_regmap_config);
drivers/rtc/rtc-pcf8583.c
132
static int pcf8583_read_mem(struct i2c_client *client, struct rtc_mem *mem)
drivers/rtc/rtc-pcf8583.c
144
.len = mem->nr,
drivers/rtc/rtc-pcf8583.c
145
.buf = mem->data,
drivers/rtc/rtc-pcf8583.c
149
if (mem->loc < 8)
drivers/rtc/rtc-pcf8583.c
152
addr[0] = mem->loc;
drivers/rtc/rtc-pcf8583.c
157
static int pcf8583_write_mem(struct i2c_client *client, struct rtc_mem *mem)
drivers/rtc/rtc-pcf8583.c
162
if (mem->loc < 8 || mem->nr > 8)
drivers/rtc/rtc-pcf8583.c
165
buf[0] = mem->loc;
drivers/rtc/rtc-pcf8583.c
166
memcpy(buf + 1, mem->data, mem->nr);
drivers/rtc/rtc-pcf8583.c
168
ret = i2c_master_send(client, buf, mem->nr + 1);
drivers/rtc/rtc-pcf8583.c
169
return ret == mem->nr + 1 ? 0 : -EIO;
drivers/rtc/rtc-pcf8583.c
176
struct rtc_mem mem = {
drivers/rtc/rtc-pcf8583.c
199
pcf8583_read_mem(client, &mem))
drivers/s390/block/dasd_int.h
728
dasd_init_chunklist(struct list_head *chunk_list, void *mem,
drivers/s390/block/dasd_int.h
734
chunk = (struct dasd_mchunk *) mem;
drivers/s390/block/dasd_int.h
762
dasd_free_chunk(struct list_head *chunk_list, void *mem)
drivers/s390/block/dasd_int.h
768
((char *) mem - sizeof(struct dasd_mchunk));
drivers/s390/char/sclp_early_core.c
312
int __init sclp_early_get_memsize(unsigned long *mem)
drivers/s390/char/sclp_early_core.c
324
*mem = rnsize * rnmax;
drivers/s390/char/sclp_mem.c
194
struct memory_block *mem;
drivers/s390/char/sclp_mem.c
234
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(addr)));
drivers/s390/char/sclp_mem.c
235
put_device(&mem->dev);
drivers/s390/char/sclp_mem.c
240
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(addr)));
drivers/s390/char/sclp_mem.c
241
if (mem->state != MEM_OFFLINE) {
drivers/s390/char/sclp_mem.c
242
put_device(&mem->dev);
drivers/s390/char/sclp_mem.c
247
put_device(&mem->dev);
drivers/s390/char/sclp_mem.c
283
struct memory_block *mem;
drivers/s390/char/sclp_mem.c
297
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(sclp_mem->id * block_size)));
drivers/s390/char/sclp_mem.c
298
if (!mem) {
drivers/s390/char/sclp_mem.c
301
put_device(&mem->dev);
drivers/s390/char/sclp_mem.c
343
static int sclp_create_configured_mem(struct memory_block *mem, void *argument)
drivers/s390/char/sclp_mem.c
350
id = mem->dev.id;
drivers/s390/cio/cmf.c
1244
cmf_activate(cmb_area.mem, CMF_ON);
drivers/s390/cio/cmf.c
378
struct cmb *mem;
drivers/s390/cio/cmf.c
455
cmb = cmb_area.mem;
drivers/s390/cio/cmf.c
463
if (cmb - cmb_area.mem >= cmb_area.num_channels) {
drivers/s390/cio/cmf.c
481
struct cmb *mem;
drivers/s390/cio/cmf.c
498
if (!cmb_area.mem) {
drivers/s390/cio/cmf.c
504
mem = (void *)__get_free_pages(GFP_KERNEL, get_order(size));
drivers/s390/cio/cmf.c
507
if (cmb_area.mem) {
drivers/s390/cio/cmf.c
509
free_pages((unsigned long)mem, get_order(size));
drivers/s390/cio/cmf.c
510
} else if (!mem) {
drivers/s390/cio/cmf.c
516
memset(mem, 0, size);
drivers/s390/cio/cmf.c
517
cmb_area.mem = mem;
drivers/s390/cio/cmf.c
518
cmf_activate(cmb_area.mem, CMF_ON);
drivers/s390/cio/cmf.c
553
free_pages((unsigned long)cmb_area.mem, get_order(size));
drivers/s390/cio/cmf.c
554
cmb_area.mem = NULL;
drivers/s390/cio/cmf.c
572
offset = mme ? (struct cmb *)cmb_data->hw_block - cmb_area.mem : 0;
drivers/s390/crypto/zcrypt_ccamisc.c
1025
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
1114
free_cprbmem(mem, PARMBSIZE, false, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
1125
void *mem;
drivers/s390/crypto/zcrypt_ccamisc.c
1139
mem = (xflags & ZCRYPT_XFLAG_NOMEMALLOC) ?
drivers/s390/crypto/zcrypt_ccamisc.c
1142
if (!mem)
drivers/s390/crypto/zcrypt_ccamisc.c
1146
token = (u8 *)mem;
drivers/s390/crypto/zcrypt_ccamisc.c
1203
mempool_free(mem, cprb_mempool);
drivers/s390/crypto/zcrypt_ccamisc.c
1216
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
1265
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
1370
free_cprbmem(mem, PARMBSIZE, true, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
1382
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
1429
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
1518
free_cprbmem(mem, PARMBSIZE, true, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
1534
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
1554
rc = alloc_and_prep_cprbmem(parmbsize, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
1623
free_cprbmem(mem, parmbsize, false, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
1634
void *mem;
drivers/s390/crypto/zcrypt_ccamisc.c
1654
mem = (xflags & ZCRYPT_XFLAG_NOMEMALLOC) ?
drivers/s390/crypto/zcrypt_ccamisc.c
1657
if (!mem)
drivers/s390/crypto/zcrypt_ccamisc.c
1659
rarray = (u8 *)mem;
drivers/s390/crypto/zcrypt_ccamisc.c
1660
varray = (u8 *)mem + PAGE_SIZE / 2;
drivers/s390/crypto/zcrypt_ccamisc.c
1707
mempool_free(mem, cprb_mempool);
drivers/s390/crypto/zcrypt_ccamisc.c
288
static void free_cprbmem(void *mem, size_t paramblen, bool scrub, u32 xflags)
drivers/s390/crypto/zcrypt_ccamisc.c
290
if (mem && scrub)
drivers/s390/crypto/zcrypt_ccamisc.c
291
memzero_explicit(mem, 2 * (sizeof(struct CPRBX) + paramblen));
drivers/s390/crypto/zcrypt_ccamisc.c
294
mempool_free(mem, cprb_mempool);
drivers/s390/crypto/zcrypt_ccamisc.c
296
kfree(mem);
drivers/s390/crypto/zcrypt_ccamisc.c
325
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
363
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
456
free_cprbmem(mem, PARMBSIZE, false, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
468
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
505
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
596
free_cprbmem(mem, PARMBSIZE, true, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
609
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
652
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
746
free_cprbmem(mem, PARMBSIZE, true, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
774
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ccamisc.c
847
rc = alloc_and_prep_cprbmem(PARMBSIZE, &mem,
drivers/s390/crypto/zcrypt_ccamisc.c
958
free_cprbmem(mem, PARMBSIZE, false, xflags);
drivers/s390/crypto/zcrypt_ccamisc.c
977
u8 *mem, *ptr;
drivers/s390/crypto/zcrypt_ep11misc.c
1359
void *mem;
drivers/s390/crypto/zcrypt_ep11misc.c
1377
mem = (xflags & ZCRYPT_XFLAG_NOMEMALLOC) ?
drivers/s390/crypto/zcrypt_ep11misc.c
1380
if (!mem)
drivers/s390/crypto/zcrypt_ep11misc.c
1382
kek = (u8 *)mem;
drivers/s390/crypto/zcrypt_ep11misc.c
1417
mempool_free(mem, cprb_mempool);
drivers/s390/crypto/zcrypt_ep11misc.c
386
static void free_cprbmem(void *mem, size_t payload_len, bool scrub, u32 xflags)
drivers/s390/crypto/zcrypt_ep11misc.c
388
if (mem && scrub)
drivers/s390/crypto/zcrypt_ep11misc.c
389
memzero_explicit(mem, sizeof(struct ep11_cprb) + payload_len);
drivers/s390/crypto/zcrypt_ep11misc.c
392
mempool_free(mem, cprb_mempool);
drivers/s390/crypto/zcrypt_ep11misc.c
394
kfree(mem);
drivers/scsi/aacraid/comminit.c
259
static void aac_queue_init(struct aac_dev * dev, struct aac_queue * q, u32 *mem, int qsize)
drivers/scsi/aacraid/comminit.c
268
q->headers.producer = (__le32 *)mem;
drivers/scsi/aacraid/comminit.c
269
q->headers.consumer = (__le32 *)(mem+1);
drivers/scsi/be2iscsi/be_cmds.c
695
struct be_dma_mem *mem)
drivers/scsi/be2iscsi/be_cmds.c
698
u64 dma = (u64) mem->dma;
drivers/scsi/be2iscsi/be_cmds.c
700
buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
drivers/scsi/be2iscsi/be_main.c
2990
struct be_dma_mem *mem = &q->dma_mem;
drivers/scsi/be2iscsi/be_main.c
2995
mem->size = len * entry_size;
drivers/scsi/be2iscsi/be_main.c
2996
mem->va = vaddress;
drivers/scsi/be2iscsi/be_main.c
2997
if (!mem->va)
drivers/scsi/be2iscsi/be_main.c
2999
memset(mem->va, 0, mem->size);
drivers/scsi/be2iscsi/be_main.c
3009
struct be_dma_mem *mem;
drivers/scsi/be2iscsi/be_main.c
3022
mem = &eq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3032
mem->va = eq_vaddress;
drivers/scsi/be2iscsi/be_main.c
3041
mem->dma = paddr;
drivers/scsi/be2iscsi/be_main.c
3059
mem = &eq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3060
if (mem->va)
drivers/scsi/be2iscsi/be_main.c
3063
mem->va, mem->dma);
drivers/scsi/be2iscsi/be_main.c
3073
struct be_dma_mem *mem;
drivers/scsi/be2iscsi/be_main.c
3088
mem = &cq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3105
mem->dma = paddr;
drivers/scsi/be2iscsi/be_main.c
3122
mem = &cq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3123
if (mem->va)
drivers/scsi/be2iscsi/be_main.c
3126
mem->va, mem->dma);
drivers/scsi/be2iscsi/be_main.c
3140
struct be_dma_mem *mem;
drivers/scsi/be2iscsi/be_main.c
3147
mem = &dq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3162
mem->dma = (unsigned long)mem_descr->mem_array[idx].
drivers/scsi/be2iscsi/be_main.c
3192
struct be_dma_mem *mem;
drivers/scsi/be2iscsi/be_main.c
3199
mem = &dataq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3215
mem->dma = (unsigned long)mem_descr->mem_array[idx].
drivers/scsi/be2iscsi/be_main.c
3314
struct be_dma_mem *mem = &q->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3315
if (mem->va) {
drivers/scsi/be2iscsi/be_main.c
3316
dma_free_coherent(&phba->pcidev->dev, mem->size,
drivers/scsi/be2iscsi/be_main.c
3317
mem->va, mem->dma);
drivers/scsi/be2iscsi/be_main.c
3318
mem->va = NULL;
drivers/scsi/be2iscsi/be_main.c
3325
struct be_dma_mem *mem = &q->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3330
mem->size = len * entry_size;
drivers/scsi/be2iscsi/be_main.c
3331
mem->va = dma_alloc_coherent(&phba->pcidev->dev, mem->size, &mem->dma,
drivers/scsi/be2iscsi/be_main.c
3333
if (!mem->va)
drivers/scsi/csiostor/csio_init.c
67
unsigned int mem = (uintptr_t)file->private_data & 3;
drivers/scsi/csiostor/csio_init.c
68
struct csio_hw *hw = file->private_data - mem;
drivers/scsi/csiostor/csio_init.c
82
if (mem == MEM_MC)
drivers/scsi/csiostor/csio_init.c
86
ret = hw->chip_ops->chip_edc_read(hw, mem, pos,
drivers/scsi/elx/libefc_sli/sli4.c
72
config->payload.mem.addr.low = cpu_to_le32(lower_32_bits(dma->phys));
drivers/scsi/elx/libefc_sli/sli4.c
73
config->payload.mem.addr.high = cpu_to_le32(upper_32_bits(dma->phys));
drivers/scsi/elx/libefc_sli/sli4.c
74
config->payload.mem.length =
drivers/scsi/elx/libefc_sli/sli4.h
2607
struct sli4_bufptr mem;
drivers/scsi/mpt3sas/mpt3sas_config.c
200
struct config_request *mem)
drivers/scsi/mpt3sas/mpt3sas_config.c
204
if (mem->sz > ioc->config_page_sz) {
drivers/scsi/mpt3sas/mpt3sas_config.c
205
mem->page = dma_alloc_coherent(&ioc->pdev->dev, mem->sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
206
&mem->page_dma, GFP_KERNEL);
drivers/scsi/mpt3sas/mpt3sas_config.c
207
if (!mem->page) {
drivers/scsi/mpt3sas/mpt3sas_config.c
209
__func__, mem->sz);
drivers/scsi/mpt3sas/mpt3sas_config.c
213
mem->page = ioc->config_page;
drivers/scsi/mpt3sas/mpt3sas_config.c
214
mem->page_dma = ioc->config_page_dma;
drivers/scsi/mpt3sas/mpt3sas_config.c
216
ioc->config_vaddr = mem->page;
drivers/scsi/mpt3sas/mpt3sas_config.c
231
struct config_request *mem)
drivers/scsi/mpt3sas/mpt3sas_config.c
233
if (mem->sz > ioc->config_page_sz)
drivers/scsi/mpt3sas/mpt3sas_config.c
234
dma_free_coherent(&ioc->pdev->dev, mem->sz, mem->page,
drivers/scsi/mpt3sas/mpt3sas_config.c
235
mem->page_dma);
drivers/scsi/mpt3sas/mpt3sas_config.c
304
struct config_request mem;
drivers/scsi/mpt3sas/mpt3sas_config.c
315
memset(&mem, 0, sizeof(struct config_request));
drivers/scsi/mpt3sas/mpt3sas_config.c
328
mem.sz = mpi_request->Header.PageLength * 4;
drivers/scsi/mpt3sas/mpt3sas_config.c
330
mem.sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4;
drivers/scsi/mpt3sas/mpt3sas_config.c
331
r = _config_alloc_config_dma_memory(ioc, &mem);
drivers/scsi/mpt3sas/mpt3sas_config.c
339
MPT3_CONFIG_COMMON_WRITE_SGLFLAGS | mem.sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
340
mem.page_dma);
drivers/scsi/mpt3sas/mpt3sas_config.c
341
memcpy(mem.page, config_page, min_t(u16, mem.sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
346
MPT3_CONFIG_COMMON_SGLFLAGS | mem.sz, mem.page_dma);
drivers/scsi/mpt3sas/mpt3sas_config.c
347
memset(mem.page, 0, min_t(u16, mem.sz, config_page_sz));
drivers/scsi/mpt3sas/mpt3sas_config.c
449
u8 *p = (u8 *)mem.page;
drivers/scsi/mpt3sas/mpt3sas_config.c
460
_debug_dump_config(p, min_t(u16, mem.sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
476
_debug_dump_config(p, min_t(u16, mem.sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
483
memcpy(config_page, mem.page, min_t(u16, mem.sz,
drivers/scsi/mpt3sas/mpt3sas_config.c
489
_config_free_config_dma_memory(ioc, &mem);
drivers/scsi/qedi/qedi_main.c
318
uinfo->mem[0].addr = (unsigned long)udev->uctrl;
drivers/scsi/qedi/qedi_main.c
319
uinfo->mem[0].size = sizeof(struct qedi_uio_ctrl);
drivers/scsi/qedi/qedi_main.c
320
uinfo->mem[0].memtype = UIO_MEM_LOGICAL;
drivers/scsi/qedi/qedi_main.c
322
uinfo->mem[1].addr = (unsigned long)udev->ll2_ring;
drivers/scsi/qedi/qedi_main.c
323
uinfo->mem[1].size = udev->ll2_ring_size;
drivers/scsi/qedi/qedi_main.c
324
uinfo->mem[1].memtype = UIO_MEM_LOGICAL;
drivers/scsi/qedi/qedi_main.c
326
uinfo->mem[2].addr = (unsigned long)udev->ll2_buf;
drivers/scsi/qedi/qedi_main.c
327
uinfo->mem[2].size = udev->ll2_buf_size;
drivers/scsi/qedi/qedi_main.c
328
uinfo->mem[2].memtype = UIO_MEM_LOGICAL;
drivers/scsi/qla2xxx/qla_bsg.c
1195
ql84_mgmt->mgmt.mgmtp.u.mem.start_addr);
drivers/scsi/qla2xxx/qla_bsg.c
1246
cpu_to_le32(ql84_mgmt->mgmt.mgmtp.u.mem.start_addr);
drivers/scsi/qla2xxx/qla_bsg.h
98
} mem; /* for QLA84_MGMT_READ/WRITE_MEM */
drivers/scsi/qla2xxx/qla_tmpl.c
35
qla27xx_insertbuf(void *mem, ulong size, void *buf, ulong *len)
drivers/scsi/qla2xxx/qla_tmpl.c
37
if (buf && mem && size) {
drivers/scsi/qla2xxx/qla_tmpl.c
39
memcpy(buf, mem, size);
drivers/scsi/sun3_scsi.c
516
struct resource *irq, *mem;
drivers/scsi/sun3_scsi.c
538
mem = platform_get_resource(pdev, IORESOURCE_MEM, i);
drivers/scsi/sun3_scsi.c
539
if (!irq || !mem)
drivers/scsi/sun3_scsi.c
542
ioaddr = sun3_ioremap(mem->start, resource_size(mem),
drivers/scsi/sun3_scsi.c
565
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/scsi/sun3_scsi.c
566
if (!irq || !mem)
drivers/scsi/sun3_scsi.c
569
ioaddr = ioremap(mem->start, resource_size(mem));
drivers/scsi/sun3_scsi.c
590
hostdata->base = mem->start;
drivers/soundwire/intel.c
59
static ssize_t intel_sprintf(void __iomem *mem, bool l,
drivers/soundwire/intel.c
65
value = intel_readl(mem, reg);
drivers/soundwire/intel.c
67
value = intel_readw(mem, reg);
drivers/soundwire/intel_ace2x_debugfs.c
24
static ssize_t intel_sprintf(void __iomem *mem, bool l,
drivers/soundwire/intel_ace2x_debugfs.c
30
value = intel_readl(mem, reg);
drivers/soundwire/intel_ace2x_debugfs.c
32
value = intel_readw(mem, reg);
drivers/spi/atmel-quadspi.c
1395
aq->mem = devm_ioremap_resource(&pdev->dev, res);
drivers/spi/atmel-quadspi.c
1396
if (IS_ERR(aq->mem))
drivers/spi/atmel-quadspi.c
1397
return dev_err_probe(&pdev->dev, PTR_ERR(aq->mem),
drivers/spi/atmel-quadspi.c
275
void __iomem *mem;
drivers/spi/atmel-quadspi.c
298
int (*transfer)(struct spi_mem *mem, const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
480
static bool atmel_qspi_supports_op(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
483
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
484
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/atmel-quadspi.c
654
static int atmel_qspi_transfer(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
657
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
669
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
675
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
811
static int atmel_qspi_dma_rx_xfer(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
816
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
836
static int atmel_qspi_dma_tx_xfer(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
841
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
861
static int atmel_qspi_dma_transfer(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
867
ret = spi_controller_dma_map_mem_op_data(mem->spi->controller, op,
drivers/spi/atmel-quadspi.c
873
ret = atmel_qspi_dma_rx_xfer(mem, op, &sgt, loff);
drivers/spi/atmel-quadspi.c
875
ret = atmel_qspi_dma_tx_xfer(mem, op, &sgt, loff);
drivers/spi/atmel-quadspi.c
877
spi_controller_dma_unmap_mem_op_data(mem->spi->controller, op, &sgt);
drivers/spi/atmel-quadspi.c
882
static int atmel_qspi_sama7g5_transfer(struct spi_mem *mem,
drivers/spi/atmel-quadspi.c
886
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
904
ret = atmel_qspi_dma_transfer(mem, op, offset);
drivers/spi/atmel-quadspi.c
908
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
922
ret = atmel_qspi_dma_transfer(mem, op, offset);
drivers/spi/atmel-quadspi.c
926
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
944
static int atmel_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/atmel-quadspi.c
946
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
969
err = aq->ops->transfer(mem, op, offset);
drivers/spi/spi-airoha-snfi.c
518
static bool airoha_snand_supports_op(struct spi_mem *mem,
drivers/spi/spi-airoha-snfi.c
521
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-airoha-snfi.c
537
u8 *txrx_buf = spi_get_ctldata(desc->mem->spi);
drivers/spi/spi-airoha-snfi.c
549
if (!airoha_snand_supports_op(desc->mem, &desc->info.op_tmpl))
drivers/spi/spi-airoha-snfi.c
558
struct spi_device *spi = desc->mem->spi;
drivers/spi/spi-airoha-snfi.c
751
struct spi_device *spi = desc->mem->spi;
drivers/spi/spi-airoha-snfi.c
932
static int airoha_snand_exec_op(struct spi_mem *mem,
drivers/spi/spi-airoha-snfi.c
940
as_ctrl = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amd.c
440
static bool amd_spi_supports_op(struct spi_mem *mem,
drivers/spi/spi-amd.c
443
struct amd_spi *amd_spi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amd.c
470
if (op->max_freq < mem->spi->controller->min_speed_hz)
drivers/spi/spi-amd.c
473
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-amd.c
476
static int amd_spi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-amd.c
478
struct amd_spi *amd_spi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amd.c
733
static int amd_spi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-amd.c
738
amd_spi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amlogic-spifc-a1.c
254
static int amlogic_spifc_a1_exec_op(struct spi_mem *mem,
drivers/spi/spi-amlogic-spifc-a1.c
258
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amlogic-spifc-a1.c
294
static int amlogic_spifc_a1_adjust_op_size(struct spi_mem *mem,
drivers/spi/spi-amlogic-spifc-a4.c
792
static int aml_sfc_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
799
sfc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-amlogic-spifc-a4.c
801
spi = mem->spi;
drivers/spi/spi-amlogic-spifc-a4.c
831
static int aml_sfc_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
836
sfc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-aspeed-smc.c
287
static bool aspeed_spi_supports_mem_op(struct spi_mem *mem,
drivers/spi/spi-aspeed-smc.c
308
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-aspeed-smc.c
313
static int do_aspeed_spi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-aspeed-smc.c
316
struct aspeed_spi *aspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-aspeed-smc.c
317
struct aspeed_spi_chip *chip = &aspi->chips[spi_get_chipselect(mem->spi, 0)];
drivers/spi/spi-aspeed-smc.c
376
static int aspeed_spi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-aspeed-smc.c
381
ret = do_aspeed_spi_exec_mem_op(mem, op);
drivers/spi/spi-aspeed-smc.c
383
dev_err(&mem->spi->dev, "operation failed: %d\n", ret);
drivers/spi/spi-aspeed-smc.c
387
static const char *aspeed_spi_get_name(struct spi_mem *mem)
drivers/spi/spi-aspeed-smc.c
389
struct aspeed_spi *aspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-aspeed-smc.c
393
spi_get_chipselect(mem->spi, 0));
drivers/spi/spi-aspeed-smc.c
698
struct aspeed_spi *aspi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-aspeed-smc.c
699
struct aspeed_spi_chip *chip = &aspi->chips[spi_get_chipselect(desc->mem->spi, 0)];
drivers/spi/spi-aspeed-smc.c
712
chip->clk_freq = desc->mem->spi->max_speed_hz;
drivers/spi/spi-aspeed-smc.c
765
struct aspeed_spi *aspi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-aspeed-smc.c
766
struct aspeed_spi_chip *chip = &aspi->chips[spi_get_chipselect(desc->mem->spi, 0)];
drivers/spi/spi-atcspi200.c
256
static int atcspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
260
spi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-atcspi200.c
364
static int atcspi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
366
struct spi_device *spi_dev = mem->spi;
drivers/spi/spi-ath79.c
137
static int ath79_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-ath79.c
140
struct ath79_spi *sp = ath79_spidev_to_sp(mem->spi);
drivers/spi/spi-ath79.c
143
if (spi_get_chipselect(mem->spi, 0) || spi_get_csgpiod(mem->spi, 0))
drivers/spi/spi-axiado.c
554
static int ax_spi_mem_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-axiado.c
556
struct spi_device *spi = mem->spi;
drivers/spi/spi-axiado.c
699
static int ax_spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-axiado.c
701
struct spi_device *spi = mem->spi;
drivers/spi/spi-bcm-qspi.c
1190
static int bcm_qspi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-bcm-qspi.c
1193
struct spi_device *spi = mem->spi;
drivers/spi/spi-bcm63xx-hsspi.c
714
static bool bcm63xx_hsspi_mem_supports_op(struct spi_mem *mem,
drivers/spi/spi-bcm63xx-hsspi.c
717
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-cadence-quadspi.c
1448
static int cqspi_mem_process(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1450
struct cqspi_st *cqspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-quadspi.c
1453
f_pdata = &cqspi->f_pdata[spi_get_chipselect(mem->spi, 0)];
drivers/spi/spi-cadence-quadspi.c
1476
static int cqspi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1479
struct cqspi_st *cqspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-quadspi.c
1500
dev_err(&mem->spi->dev, "resume failed with %d\n", ret);
drivers/spi/spi-cadence-quadspi.c
1505
ret = cqspi_mem_process(mem, op);
drivers/spi/spi-cadence-quadspi.c
1511
dev_err(&mem->spi->dev, "operation failed with %d\n", ret);
drivers/spi/spi-cadence-quadspi.c
1520
static bool cqspi_supports_mem_op(struct spi_mem *mem,
drivers/spi/spi-cadence-quadspi.c
1523
struct cqspi_st *cqspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-quadspi.c
1558
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-cadence-quadspi.c
1729
static const char *cqspi_get_name(struct spi_mem *mem)
drivers/spi/spi-cadence-quadspi.c
1731
struct cqspi_st *cqspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-quadspi.c
1735
spi_get_chipselect(mem->spi, 0));
drivers/spi/spi-cadence-xspi.c
740
struct spi_mem *mem,
drivers/spi/spi-cadence-xspi.c
745
if (cdns_xspi->cur_cs != spi_get_chipselect(mem->spi, 0))
drivers/spi/spi-cadence-xspi.c
746
cdns_xspi->cur_cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-cadence-xspi.c
752
static int cdns_xspi_mem_op_execute(struct spi_mem *mem,
drivers/spi/spi-cadence-xspi.c
756
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-xspi.c
759
ret = cdns_xspi_mem_op(cdns_xspi, mem, op);
drivers/spi/spi-cadence-xspi.c
764
static int marvell_xspi_mem_op_execute(struct spi_mem *mem,
drivers/spi/spi-cadence-xspi.c
768
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-cadence-xspi.c
771
cdns_mrvl_xspi_setup_clock(cdns_xspi, mem->spi->max_speed_hz);
drivers/spi/spi-cadence-xspi.c
773
ret = cdns_xspi_mem_op(cdns_xspi, mem, op);
drivers/spi/spi-cadence-xspi.c
778
static bool cdns_xspi_supports_op(struct spi_mem *mem,
drivers/spi/spi-cadence-xspi.c
781
struct spi_device *spi = mem->spi;
drivers/spi/spi-cadence-xspi.c
823
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-cadence-xspi.c
829
static int cdns_xspi_adjust_mem_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
832
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-dw-core.c
491
static int dw_spi_adjust_mem_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
499
static bool dw_spi_supports_mem_op(struct spi_mem *mem,
drivers/spi/spi-dw-core.c
506
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-dw-core.c
675
static int dw_spi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
677
struct dw_spi *dws = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-dw-core.c
705
dw_spi_update_config(dws, mem->spi, &cfg);
drivers/spi/spi-dw-core.c
742
ret = dw_spi_write_then_read(dws, mem->spi);
drivers/spi/spi-dw-core.c
760
dw_spi_stop_mem_op(dws, mem->spi);
drivers/spi/spi-dw-mmio.c
318
struct resource *mem;
drivers/spi/spi-dw-mmio.c
330
dws->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/spi/spi-dw-mmio.c
334
dws->paddr = mem->start;
drivers/spi/spi-fsl-espi.c
663
static int fsl_espi_probe(struct device *dev, struct resource *mem,
drivers/spi/spi-fsl-espi.c
702
espi->reg_base = devm_ioremap_resource(dev, mem);
drivers/spi/spi-fsl-espi.c
759
struct resource mem;
drivers/spi/spi-fsl-espi.c
772
ret = of_address_to_resource(np, 0, &mem);
drivers/spi/spi-fsl-espi.c
780
return fsl_espi_probe(dev, &mem, irq, num_cs);
drivers/spi/spi-fsl-lib.c
81
void mpc8xxx_spi_probe(struct device *dev, struct resource *mem,
drivers/spi/spi-fsl-lib.h
107
extern void mpc8xxx_spi_probe(struct device *dev, struct resource *mem,
drivers/spi/spi-fsl-qspi.c
390
static bool fsl_qspi_supports_op(struct spi_mem *mem,
drivers/spi/spi-fsl-qspi.c
393
struct fsl_qspi *q = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-fsl-qspi.c
435
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-fsl-qspi.c
668
static int fsl_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
670
struct fsl_qspi *q = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-fsl-qspi.c
682
fsl_qspi_select_mem(q, mem->spi, op);
drivers/spi/spi-fsl-qspi.c
730
static int fsl_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
732
struct fsl_qspi *q = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-fsl-qspi.c
837
static const char *fsl_qspi_get_name(struct spi_mem *mem)
drivers/spi/spi-fsl-qspi.c
839
struct fsl_qspi *q = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-fsl-qspi.c
840
struct device *dev = &mem->spi->dev;
drivers/spi/spi-fsl-qspi.c
853
spi_get_chipselect(mem->spi, 0));
drivers/spi/spi-fsl-spi.c
529
struct resource *mem, unsigned int irq)
drivers/spi/spi-fsl-spi.c
546
mpc8xxx_spi_probe(dev, mem, irq);
drivers/spi/spi-fsl-spi.c
564
mpc8xxx_spi->reg_base = devm_ioremap_resource(dev, mem);
drivers/spi/spi-fsl-spi.c
621
dev_info(dev, "at MMIO %pa (irq = %d), %s mode\n", &mem->start,
drivers/spi/spi-fsl-spi.c
639
struct resource mem;
drivers/spi/spi-fsl-spi.c
681
ret = of_address_to_resource(np, 0, &mem);
drivers/spi/spi-fsl-spi.c
691
host = fsl_spi_probe(dev, &mem, irq);
drivers/spi/spi-fsl-spi.c
730
struct resource *mem;
drivers/spi/spi-fsl-spi.c
737
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/spi/spi-fsl-spi.c
738
if (!mem)
drivers/spi/spi-fsl-spi.c
745
host = fsl_spi_probe(&pdev->dev, mem, irq);
drivers/spi/spi-gxp.c
200
static int do_gxp_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
202
struct gxp_spi *spifi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-gxp.c
203
struct gxp_spi_chip *chip = &spifi->chips[spi_get_chipselect(mem->spi, 0)];
drivers/spi/spi-gxp.c
221
static int gxp_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
225
ret = do_gxp_exec_mem_op(mem, op);
drivers/spi/spi-gxp.c
227
dev_err(&mem->spi->dev, "operation failed: %d", ret);
drivers/spi/spi-hisi-sfc-v3xx.c
147
static int hisi_sfc_v3xx_adjust_op_size(struct spi_mem *mem,
drivers/spi/spi-hisi-sfc-v3xx.c
150
struct spi_device *spi = mem->spi;
drivers/spi/spi-hisi-sfc-v3xx.c
171
static bool hisi_sfc_v3xx_supports_op(struct spi_mem *mem,
drivers/spi/spi-hisi-sfc-v3xx.c
174
struct spi_device *spi = mem->spi;
drivers/spi/spi-hisi-sfc-v3xx.c
186
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-hisi-sfc-v3xx.c
359
static int hisi_sfc_v3xx_exec_op(struct spi_mem *mem,
drivers/spi/spi-hisi-sfc-v3xx.c
363
struct spi_device *spi = mem->spi;
drivers/spi/spi-intel.c
185
const struct spi_mem *mem,
drivers/spi/spi-intel.c
459
const struct spi_mem *mem)
drivers/spi/spi-intel.c
462
if (!mem)
drivers/spi/spi-intel.c
464
return (spi_get_chipselect(mem->spi, 0) == 1) ? ispi->chip0_size : 0;
drivers/spi/spi-intel.c
467
static int intel_spi_read_reg(struct intel_spi *ispi, const struct spi_mem *mem,
drivers/spi/spi-intel.c
471
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
490
static int intel_spi_write_reg(struct intel_spi *ispi, const struct spi_mem *mem,
drivers/spi/spi-intel.c
494
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
551
static int intel_spi_read(struct intel_spi *ispi, const struct spi_mem *mem,
drivers/spi/spi-intel.c
555
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
612
static int intel_spi_write(struct intel_spi *ispi, const struct spi_mem *mem,
drivers/spi/spi-intel.c
616
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
675
static int intel_spi_erase(struct intel_spi *ispi, const struct spi_mem *mem,
drivers/spi/spi-intel.c
679
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
713
static int intel_spi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-intel.c
757
static bool intel_spi_supports_mem_op(struct spi_mem *mem,
drivers/spi/spi-intel.c
760
struct intel_spi *ispi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-intel.c
789
static int intel_spi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-intel.c
791
struct intel_spi *ispi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-intel.c
798
return iop->exec_op(ispi, mem, iop, op);
drivers/spi/spi-intel.c
801
static const char *intel_spi_get_name(struct spi_mem *mem)
drivers/spi/spi-intel.c
803
const struct intel_spi *ispi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-intel.c
814
struct intel_spi *ispi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-intel.c
828
struct intel_spi *ispi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-intel.c
838
ret = iop->exec_op(ispi, desc->mem, iop, &op);
drivers/spi/spi-intel.c
845
struct intel_spi *ispi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-intel.c
854
ret = iop->exec_op(ispi, desc->mem, iop, &op);
drivers/spi/spi-mem.c
1010
polling_delay_us, timeout_ms * 1000, false, mem,
drivers/spi/spi-mem.c
1024
struct spi_mem *mem;
drivers/spi/spi-mem.c
1026
mem = devm_kzalloc(&spi->dev, sizeof(*mem), GFP_KERNEL);
drivers/spi/spi-mem.c
1027
if (!mem)
drivers/spi/spi-mem.c
1030
mem->spi = spi;
drivers/spi/spi-mem.c
1033
mem->name = ctlr->mem_ops->get_name(mem);
drivers/spi/spi-mem.c
1035
mem->name = dev_name(&spi->dev);
drivers/spi/spi-mem.c
1037
if (IS_ERR_OR_NULL(mem->name))
drivers/spi/spi-mem.c
1038
return PTR_ERR_OR_ZERO(mem->name);
drivers/spi/spi-mem.c
1040
spi_set_drvdata(spi, mem);
drivers/spi/spi-mem.c
1042
return memdrv->probe(mem);
drivers/spi/spi-mem.c
1048
struct spi_mem *mem = spi_get_drvdata(spi);
drivers/spi/spi-mem.c
1051
memdrv->remove(mem);
drivers/spi/spi-mem.c
1057
struct spi_mem *mem = spi_get_drvdata(spi);
drivers/spi/spi-mem.c
1060
memdrv->shutdown(mem);
drivers/spi/spi-mem.c
107
static int spi_check_buswidth_req(struct spi_mem *mem, u8 buswidth, bool tx)
drivers/spi/spi-mem.c
109
u32 mode = mem->spi->mode;
drivers/spi/spi-mem.c
145
static bool spi_mem_check_buswidth(struct spi_mem *mem,
drivers/spi/spi-mem.c
148
if (spi_check_buswidth_req(mem, op->cmd.buswidth, true))
drivers/spi/spi-mem.c
152
spi_check_buswidth_req(mem, op->addr.buswidth, true))
drivers/spi/spi-mem.c
156
spi_check_buswidth_req(mem, op->dummy.buswidth, true))
drivers/spi/spi-mem.c
160
spi_check_buswidth_req(mem, op->data.buswidth,
drivers/spi/spi-mem.c
167
bool spi_mem_default_supports_op(struct spi_mem *mem,
drivers/spi/spi-mem.c
170
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
204
if (op->max_freq && mem->spi->controller->min_speed_hz &&
drivers/spi/spi-mem.c
205
op->max_freq < mem->spi->controller->min_speed_hz)
drivers/spi/spi-mem.c
209
op->max_freq < mem->spi->max_speed_hz) {
drivers/spi/spi-mem.c
214
return spi_mem_check_buswidth(mem, op);
drivers/spi/spi-mem.c
254
static bool spi_mem_internal_supports_op(struct spi_mem *mem,
drivers/spi/spi-mem.c
257
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
260
return ctlr->mem_ops->supports_op(mem, op);
drivers/spi/spi-mem.c
262
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-mem.c
280
bool spi_mem_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mem.c
283
spi_mem_adjust_op_freq(mem, (struct spi_mem_op *)op);
drivers/spi/spi-mem.c
288
return spi_mem_internal_supports_op(mem, op);
drivers/spi/spi-mem.c
292
static int spi_mem_access_start(struct spi_mem *mem)
drivers/spi/spi-mem.c
294
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
319
static void spi_mem_access_end(struct spi_mem *mem)
drivers/spi/spi-mem.c
321
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
385
int spi_mem_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mem.c
388
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
395
spi_mem_adjust_op_freq(mem, (struct spi_mem_op *)op);
drivers/spi/spi-mem.c
397
dev_vdbg(&mem->spi->dev, "[cmd: 0x%02x][%dB addr: %#8llx][%2dB dummy][%4dB data %s] %d%c-%d%c-%d%c-%d%c @ %uHz\n",
drivers/spi/spi-mem.c
406
op->max_freq ? op->max_freq : mem->spi->max_speed_hz);
drivers/spi/spi-mem.c
412
if (!spi_mem_internal_supports_op(mem, op))
drivers/spi/spi-mem.c
415
if (ctlr->mem_ops && ctlr->mem_ops->exec_op && !spi_get_csgpiod(mem->spi, 0)) {
drivers/spi/spi-mem.c
416
ret = spi_mem_access_start(mem);
drivers/spi/spi-mem.c
420
trace_spi_mem_start_op(mem, op);
drivers/spi/spi-mem.c
421
ret = ctlr->mem_ops->exec_op(mem, op);
drivers/spi/spi-mem.c
422
trace_spi_mem_stop_op(mem, op);
drivers/spi/spi-mem.c
424
spi_mem_access_end(mem);
drivers/spi/spi-mem.c
433
spi_mem_add_op_stats(mem->spi->pcpu_statistics, op, ret);
drivers/spi/spi-mem.c
505
ret = spi_sync(mem->spi, &msg);
drivers/spi/spi-mem.c
531
const char *spi_mem_get_name(struct spi_mem *mem)
drivers/spi/spi-mem.c
533
return mem->name;
drivers/spi/spi-mem.c
552
int spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
554
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
558
return ctlr->mem_ops->adjust_op_size(mem, op);
drivers/spi/spi-mem.c
563
if (len > spi_max_transfer_size(mem->spi))
drivers/spi/spi-mem.c
567
spi_max_transfer_size(mem->spi),
drivers/spi/spi-mem.c
568
spi_max_message_size(mem->spi) -
drivers/spi/spi-mem.c
588
void spi_mem_adjust_op_freq(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
590
if (!op->max_freq || op->max_freq > mem->spi->max_speed_hz)
drivers/spi/spi-mem.c
591
op->max_freq = mem->spi->max_speed_hz;
drivers/spi/spi-mem.c
614
u64 spi_mem_calc_op_duration(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
619
spi_mem_adjust_op_freq(mem, op);
drivers/spi/spi-mem.c
656
ret = spi_mem_adjust_op_size(desc->mem, &op);
drivers/spi/spi-mem.c
660
ret = spi_mem_exec_op(desc->mem, &op);
drivers/spi/spi-mem.c
676
ret = spi_mem_adjust_op_size(desc->mem, &op);
drivers/spi/spi-mem.c
680
ret = spi_mem_exec_op(desc->mem, &op);
drivers/spi/spi-mem.c
701
spi_mem_dirmap_create(struct spi_mem *mem,
drivers/spi/spi-mem.c
704
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
720
desc->mem = mem;
drivers/spi/spi-mem.c
723
ret = spi_mem_access_start(mem);
drivers/spi/spi-mem.c
731
spi_mem_access_end(mem);
drivers/spi/spi-mem.c
736
if (!spi_mem_supports_op(desc->mem, &desc->info.op_tmpl))
drivers/spi/spi-mem.c
760
struct spi_controller *ctlr = desc->mem->spi->controller;
drivers/spi/spi-mem.c
789
devm_spi_mem_dirmap_create(struct device *dev, struct spi_mem *mem,
drivers/spi/spi-mem.c
799
desc = spi_mem_dirmap_create(mem, info);
drivers/spi/spi-mem.c
857
struct spi_controller *ctlr = desc->mem->spi->controller;
drivers/spi/spi-mem.c
869
ret = spi_mem_access_start(desc->mem);
drivers/spi/spi-mem.c
875
spi_mem_access_end(desc->mem);
drivers/spi/spi-mem.c
903
struct spi_controller *ctlr = desc->mem->spi->controller;
drivers/spi/spi-mem.c
915
ret = spi_mem_access_start(desc->mem);
drivers/spi/spi-mem.c
921
spi_mem_access_end(desc->mem);
drivers/spi/spi-mem.c
935
static int spi_mem_read_status(struct spi_mem *mem,
drivers/spi/spi-mem.c
942
ret = spi_mem_exec_op(mem, op);
drivers/spi/spi-mem.c
970
int spi_mem_poll_status(struct spi_mem *mem,
drivers/spi/spi-mem.c
977
struct spi_controller *ctlr = mem->spi->controller;
drivers/spi/spi-mem.c
986
if (ctlr->mem_ops && ctlr->mem_ops->poll_status && !spi_get_csgpiod(mem->spi, 0)) {
drivers/spi/spi-mem.c
987
ret = spi_mem_access_start(mem);
drivers/spi/spi-mem.c
991
ret = ctlr->mem_ops->poll_status(mem, op, mask, match,
drivers/spi/spi-mem.c
995
spi_mem_access_end(mem);
drivers/spi/spi-mem.c
999
if (!spi_mem_supports_op(mem, op))
drivers/spi/spi-microchip-core-qspi.c
460
static int mchp_coreqspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
463
(mem->spi->controller);
drivers/spi/spi-microchip-core-qspi.c
472
dev_err(&mem->spi->dev, "Timeout waiting on QSPI ready.\n");
drivers/spi/spi-microchip-core-qspi.c
476
err = mchp_coreqspi_setup_clock(qspi, mem->spi, op->max_freq);
drivers/spi/spi-microchip-core-qspi.c
532
static bool mchp_coreqspi_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
534
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-microchip-core-qspi.c
559
static int mchp_coreqspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mt65xx.c
1000
struct mtk_spi *mdata = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mt65xx.c
1009
mtk_spi_hw_init(mem->spi->controller, mem->spi);
drivers/spi/spi-mt65xx.c
1010
mtk_spi_prepare_transfer(mem->spi->controller, op->max_freq);
drivers/spi/spi-mt65xx.c
1030
mtk_spi_setup_packet(mem->spi->controller);
drivers/spi/spi-mt65xx.c
1128
mtk_spi_mem_setup_dma_xfer(mem->spi->controller, op);
drivers/spi/spi-mt65xx.c
1130
mtk_spi_enable_transfer(mem->spi->controller);
drivers/spi/spi-mt65xx.c
1133
ret = mtk_spi_transfer_wait(mem, op);
drivers/spi/spi-mt65xx.c
903
static int mtk_spi_mem_adjust_op_size(struct spi_mem *mem,
drivers/spi/spi-mt65xx.c
920
static bool mtk_spi_mem_supports_op(struct spi_mem *mem,
drivers/spi/spi-mt65xx.c
923
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mt65xx.c
967
static int mtk_spi_transfer_wait(struct spi_mem *mem,
drivers/spi/spi-mt65xx.c
970
struct mtk_spi *mdata = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mt65xx.c
982
ms = div_u64(ms, mem->spi->max_speed_hz);
drivers/spi/spi-mt65xx.c
997
static int mtk_spi_mem_exec_op(struct spi_mem *mem,
drivers/spi/spi-mtk-nor.c
276
static int mtk_nor_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
278
struct mtk_nor *sp = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mtk-nor.c
311
static bool mtk_nor_supports_op(struct spi_mem *mem,
drivers/spi/spi-mtk-nor.c
314
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mtk-nor.c
599
static int mtk_nor_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
601
struct mtk_nor *sp = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mtk-snfi.c
1241
static bool mtk_snand_supports_op(struct spi_mem *mem,
drivers/spi/spi-mtk-snfi.c
1244
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mtk-snfi.c
1255
static int mtk_snand_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1257
struct mtk_snand *ms = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mtk-snfi.c
1282
static int mtk_snand_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1284
struct mtk_snand *ms = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mxic.c
399
struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-mxic.c
406
writel(mxic_spi_prep_hc_cfg(desc->mem->spi, 0, desc->info.op_tmpl.data.swap16),
drivers/spi/spi-mxic.c
415
LMODE_SLV_ACT(spi_get_chipselect(desc->mem->spi, 0)) |
drivers/spi/spi-mxic.c
444
struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-mxic.c
451
writel(mxic_spi_prep_hc_cfg(desc->mem->spi, 0, desc->info.op_tmpl.data.swap16),
drivers/spi/spi-mxic.c
460
LMODE_SLV_ACT(spi_get_chipselect(desc->mem->spi, 0)) |
drivers/spi/spi-mxic.c
485
static bool mxic_spi_mem_supports_op(struct spi_mem *mem,
drivers/spi/spi-mxic.c
499
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-mxic.c
504
struct mxic_spi *mxic = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-mxic.c
512
if (!mxic_spi_mem_supports_op(desc->mem, &desc->info.op_tmpl))
drivers/spi/spi-mxic.c
518
static int mxic_spi_mem_exec_op(struct spi_mem *mem,
drivers/spi/spi-mxic.c
521
struct mxic_spi *mxic = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-mxic.c
529
writel(mxic_spi_prep_hc_cfg(mem->spi, HC_CFG_MAN_CS_EN, op->data.swap16),
drivers/spi/spi-mxic.c
535
mxic->regs + SS_CTRL(spi_get_chipselect(mem->spi, 0)));
drivers/spi/spi-npcm-fiu.c
291
spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
292
struct npcm_fiu_chip *chip = &fiu->chip[spi_get_chipselect(desc->mem->spi, 0)];
drivers/spi/spi-npcm-fiu.c
318
spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
319
struct npcm_fiu_chip *chip = &fiu->chip[spi_get_chipselect(desc->mem->spi, 0)];
drivers/spi/spi-npcm-fiu.c
334
static int npcm_fiu_uma_read(struct spi_mem *mem,
drivers/spi/spi-npcm-fiu.c
339
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
348
(spi_get_chipselect(mem->spi, 0) <<
drivers/spi/spi-npcm-fiu.c
390
static int npcm_fiu_uma_write(struct spi_mem *mem,
drivers/spi/spi-npcm-fiu.c
395
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
403
(spi_get_chipselect(mem->spi, 0) <<
drivers/spi/spi-npcm-fiu.c
440
static int npcm_fiu_manualwrite(struct spi_mem *mem,
drivers/spi/spi-npcm-fiu.c
444
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
456
(spi_get_chipselect(mem->spi, 0) <<
drivers/spi/spi-npcm-fiu.c
461
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, true, NULL, 0);
drivers/spi/spi-npcm-fiu.c
467
ret = npcm_fiu_uma_write(mem, op, data[0], false,
drivers/spi/spi-npcm-fiu.c
477
ret = npcm_fiu_uma_write(mem, op, data[0], false,
drivers/spi/spi-npcm-fiu.c
489
static int npcm_fiu_read(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
505
ret = npcm_fiu_uma_read(mem, op, addr, true, buf_ptr,
drivers/spi/spi-npcm-fiu.c
543
static int npcm_fiu_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
546
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
547
struct npcm_fiu_chip *chip = &fiu->chip[spi_get_chipselect(mem->spi, 0)];
drivers/spi/spi-npcm-fiu.c
566
ret = npcm_fiu_uma_read(mem, op, op->addr.val, false,
drivers/spi/spi-npcm-fiu.c
569
ret = npcm_fiu_read(mem, op);
drivers/spi/spi-npcm-fiu.c
573
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
584
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
588
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
592
ret = npcm_fiu_manualwrite(mem, op);
drivers/spi/spi-npcm-fiu.c
601
spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-npcm-fiu.c
602
struct npcm_fiu_chip *chip = &fiu->chip[spi_get_chipselect(desc->mem->spi, 0)];
drivers/spi/spi-npcm-fiu.c
621
spi_get_chipselect(desc->mem->spi, 0))),
drivers/spi/spi-nxp-fspi.c
1027
static int nxp_fspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
1029
struct nxp_fspi *f = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-fspi.c
1045
nxp_fspi_select_mem(f, mem->spi, op);
drivers/spi/spi-nxp-fspi.c
1073
static int nxp_fspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
1075
struct nxp_fspi *f = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-fspi.c
1228
static const char *nxp_fspi_get_name(struct spi_mem *mem)
drivers/spi/spi-nxp-fspi.c
1230
struct nxp_fspi *f = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-fspi.c
1231
struct device *dev = &mem->spi->dev;
drivers/spi/spi-nxp-fspi.c
1240
spi_get_chipselect(mem->spi, 0));
drivers/spi/spi-nxp-fspi.c
469
static bool nxp_fspi_supports_op(struct spi_mem *mem,
drivers/spi/spi-nxp-fspi.c
472
struct nxp_fspi *f = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-fspi.c
519
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-nxp-xspi.c
1010
nxp_xspi_select_mem(xspi, mem->spi, op);
drivers/spi/spi-nxp-xspi.c
1032
static int nxp_xspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
1034
struct nxp_xspi *xspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-xspi.c
1155
static const char *nxp_xspi_get_name(struct spi_mem *mem)
drivers/spi/spi-nxp-xspi.c
1157
struct nxp_xspi *xspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-xspi.c
1158
struct device *dev = &mem->spi->dev;
drivers/spi/spi-nxp-xspi.c
1167
spi_get_chipselect(mem->spi, 0));
drivers/spi/spi-nxp-xspi.c
373
static bool nxp_xspi_supports_op(struct spi_mem *mem,
drivers/spi/spi-nxp-xspi.c
376
struct nxp_xspi *xspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-nxp-xspi.c
412
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-nxp-xspi.c
988
static int nxp_xspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
990
struct nxp_xspi *xspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-pic32.c
710
struct resource *mem;
drivers/spi/spi-pic32.c
713
pic32s->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/spi/spi-pic32.c
717
pic32s->dma_base = mem->start;
drivers/spi/spi-qcom-qspi.c
669
static int qcom_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1455
static bool qcom_spi_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1457
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-qpic-snand.c
1471
static int qcom_spi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1473
struct qcom_nand_controller *snandc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-realtek-rtl-snand.c
326
static int rtl_snand_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
328
struct rtl_snand *snand = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-realtek-rtl-snand.c
329
int cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-realtek-rtl-snand.c
53
static bool rtl_snand_supports_op(struct spi_mem *mem,
drivers/spi/spi-realtek-rtl-snand.c
56
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-rockchip-sfc.c
318
struct spi_mem *mem,
drivers/spi/spi-rockchip-sfc.c
323
u8 cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-rockchip-sfc.c
524
static int rockchip_sfc_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-rockchip-sfc.c
526
struct rockchip_sfc *sfc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-rockchip-sfc.c
529
u8 cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-rockchip-sfc.c
548
rockchip_sfc_xfer_setup(sfc, mem, op, len);
drivers/spi/spi-rockchip-sfc.c
573
static int rockchip_sfc_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-rockchip-sfc.c
575
struct rockchip_sfc *sfc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-rockchip.c
762
struct resource *mem;
drivers/spi/spi-rockchip.c
782
rs->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/spi/spi-rockchip.c
892
rs->dma_addr_tx = mem->start + ROCKCHIP_SPI_TXDR;
drivers/spi/spi-rockchip.c
893
rs->dma_addr_rx = mem->start + ROCKCHIP_SPI_RXDR;
drivers/spi/spi-rpc-if.c
100
rpcif_spi_mem_prepare(desc->mem->spi, &desc->info.op_tmpl, &offs, &len);
drivers/spi/spi-rpc-if.c
108
spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-rpc-if.c
113
if (!rpcif_spi_mem_supports_op(desc->mem, &desc->info.op_tmpl))
drivers/spi/spi-rpc-if.c
125
static int rpcif_spi_mem_exec_op(struct spi_mem *mem,
drivers/spi/spi-rpc-if.c
129
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-rpc-if.c
131
rpcif_spi_mem_prepare(mem->spi, op, NULL, NULL);
drivers/spi/spi-rpc-if.c
64
static bool rpcif_spi_mem_supports_op(struct spi_mem *mem,
drivers/spi/spi-rpc-if.c
67
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-rpc-if.c
81
struct rpcif *rpc = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-rpc-if.c
86
rpcif_spi_mem_prepare(desc->mem->spi, &desc->info.op_tmpl, &offs, &len);
drivers/spi/spi-rpc-if.c
95
spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-sg2044-nor.c
388
static int sg2044_spifmc_exec_op(struct spi_mem *mem,
drivers/spi/spi-sg2044-nor.c
393
spifmc = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-sn-f-ospi.c
264
struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
267
struct spi_device *spi = mem->spi;
drivers/spi/spi-sn-f-ospi.c
338
static int f_ospi_indir_prepare_op(struct f_ospi *ospi, struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
350
f_ospi_config_indir_protocol(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
403
static int f_ospi_indir_read(struct f_ospi *ospi, struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
413
ret = f_ospi_indir_prepare_op(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
457
static int f_ospi_indir_write(struct f_ospi *ospi, struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
467
ret = f_ospi_indir_prepare_op(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
504
static int f_ospi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
506
struct f_ospi *ospi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-sn-f-ospi.c
511
err = f_ospi_indir_read(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
517
err = f_ospi_indir_write(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
528
static bool f_ospi_supports_op_width(struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
554
static bool f_ospi_supports_op(struct spi_mem *mem,
drivers/spi/spi-sn-f-ospi.c
563
if (!f_ospi_supports_op_width(mem, op))
drivers/spi/spi-sn-f-ospi.c
566
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-sn-f-ospi.c
569
static int f_ospi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
553
static int stm32_ospi_poll_status(struct spi_mem *mem,
drivers/spi/spi-stm32-ospi.c
560
struct stm32_ospi *ospi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-stm32-ospi.c
575
ret = stm32_ospi_send(mem->spi, op);
drivers/spi/spi-stm32-ospi.c
583
static int stm32_ospi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
585
struct stm32_ospi *ospi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-stm32-ospi.c
598
ret = stm32_ospi_send(mem->spi, op);
drivers/spi/spi-stm32-ospi.c
608
struct stm32_ospi *ospi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-stm32-ospi.c
626
struct stm32_ospi *ospi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-stm32-ospi.c
654
ret = stm32_ospi_send(desc->mem->spi, &op);
drivers/spi/spi-stm32-qspi.c
451
static int stm32_qspi_poll_status(struct spi_mem *mem, const struct spi_mem_op *op,
drivers/spi/spi-stm32-qspi.c
457
struct stm32_qspi *qspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-stm32-qspi.c
460
if (!spi_mem_supports_op(mem, op))
drivers/spi/spi-stm32-qspi.c
474
ret = stm32_qspi_send(mem->spi, op);
drivers/spi/spi-stm32-qspi.c
482
static int stm32_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
484
struct stm32_qspi *qspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-stm32-qspi.c
497
ret = stm32_qspi_send(mem->spi, op);
drivers/spi/spi-stm32-qspi.c
507
struct stm32_qspi *qspi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-stm32-qspi.c
525
struct stm32_qspi *qspi = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-stm32-qspi.c
552
ret = stm32_qspi_send(desc->mem->spi, &op);
drivers/spi/spi-sun6i.c
633
struct resource *mem;
drivers/spi/spi-sun6i.c
645
sspi->base_addr = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/spi/spi-sun6i.c
725
sspi->dma_addr_tx = mem->start + SUN6I_TXDATA_REG;
drivers/spi/spi-sun6i.c
726
sspi->dma_addr_rx = mem->start + SUN6I_RXDATA_REG;
drivers/spi/spi-ti-qspi.c
575
static int ti_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-ti-qspi.c
577
struct ti_qspi *qspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-ti-qspi.c
604
static int ti_qspi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-ti-qspi.c
607
struct ti_qspi *qspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-ti-qspi.c
623
if (!qspi->mmap_enabled || qspi->current_cs != spi_get_chipselect(mem->spi, 0)) {
drivers/spi/spi-ti-qspi.c
625
ti_qspi_enable_memory_map(mem->spi);
drivers/spi/spi-ti-qspi.c
627
ti_qspi_setup_mmap_read(mem->spi, op->cmd.opcode, op->data.buswidth,
drivers/spi/spi-ti-qspi.c
634
!spi_controller_dma_map_mem_op_data(mem->spi->controller, op,
drivers/spi/spi-ti-qspi.c
637
spi_controller_dma_unmap_mem_op_data(mem->spi->controller,
drivers/spi/spi-wpcm-fiu.c
137
int (*exec)(struct spi_mem *mem, const struct spi_mem_op *op);
drivers/spi/spi-wpcm-fiu.c
150
static int wpcm_fiu_normal_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
152
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
160
ret = wpcm_fiu_do_uma(fiu, spi_get_chipselect(mem->spi, 0), op->addr.nbytes == 3,
drivers/spi/spi-wpcm-fiu.c
177
static int wpcm_fiu_fast_read_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
195
static int wpcm_fiu_4ba_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
197
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
198
int cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-wpcm-fiu.c
240
static int wpcm_fiu_rdid_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
242
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
243
int cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-wpcm-fiu.c
277
static int wpcm_fiu_dummy_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
279
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
280
int cs = spi_get_chipselect(mem->spi, 0);
drivers/spi/spi-wpcm-fiu.c
322
static bool wpcm_fiu_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
324
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-wpcm-fiu.c
352
static int wpcm_fiu_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
354
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
360
return shape->exec(mem, op);
drivers/spi/spi-wpcm-fiu.c
367
static int wpcm_fiu_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
377
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
378
int cs = spi_get_chipselect(desc->mem->spi, 0);
drivers/spi/spi-wpcm-fiu.c
401
struct wpcm_fiu_spi *fiu = spi_controller_get_devdata(desc->mem->spi->controller);
drivers/spi/spi-wpcm-fiu.c
402
int cs = spi_get_chipselect(desc->mem->spi, 0);
drivers/spi/spi-zynq-qspi.c
225
static bool zynq_qspi_supports_op(struct spi_mem *mem,
drivers/spi/spi-zynq-qspi.c
228
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-zynq-qspi.c
536
static int zynq_qspi_exec_mem_op(struct spi_mem *mem,
drivers/spi/spi-zynq-qspi.c
539
struct zynq_qspi *xqspi = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-zynq-qspi.c
543
zynq_qspi_chipselect(mem->spi, true);
drivers/spi/spi-zynq-qspi.c
544
zynq_qspi_config_op(xqspi, mem->spi, op);
drivers/spi/spi-zynq-qspi.c
620
zynq_qspi_chipselect(mem->spi, false);
drivers/spi/spi-zynqmp-gqspi.c
1044
static int zynqmp_qspi_exec_op(struct spi_mem *mem,
drivers/spi/spi-zynqmp-gqspi.c
1048
spi_controller_get_devdata(mem->spi->controller);
drivers/spi/spi-zynqmp-gqspi.c
1057
zynqmp_qspi_chipselect(mem->spi, false);
drivers/spi/spi-zynqmp-gqspi.c
1182
zynqmp_qspi_chipselect(mem->spi, true);
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
18
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
25
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
32
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
40
enum ia_css_isp_memories mem);
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
46
enum ia_css_isp_memories mem);
drivers/staging/media/atomisp/pci/runtime/isp_param/interface/ia_css_isp_param.h
52
enum ia_css_isp_memories mem);
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
102
unsigned int mem, pclass;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
105
for (mem = 0; mem < IA_CSS_NUM_MEMORIES; mem++)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
111
size = mem_initializers->params[pclass][mem].size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
112
mem_params->params[pclass][mem].size = size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
113
mem_params->params[pclass][mem].address = NULL;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
114
css_params->params[pclass][mem].size = size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
115
css_params->params[pclass][mem].address = 0x0;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
117
mem_params->params[pclass][mem].address = kvcalloc(1,
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
120
if (!mem_params->params[pclass][mem].address) {
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
125
css_params->params[pclass][mem].address = hmm_alloc(size);
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
126
if (!css_params->params[pclass][mem].address) {
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
145
unsigned int mem, pclass;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
147
for (mem = 0; mem < IA_CSS_NUM_MEMORIES; mem++) {
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
149
kvfree(mem_params->params[pclass][mem].address);
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
150
if (css_params->params[pclass][mem].address)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
151
hmm_free(css_params->params[pclass][mem].address);
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
152
mem_params->params[pclass][mem].address = NULL;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
153
css_params->params[pclass][mem].address = 0x0;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
179
unsigned int mem;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
18
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
181
for (mem = 0; mem < N_IA_CSS_ISP_MEMORIES; mem++)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
183
size_t size = host->params[pclass][mem].size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
184
ia_css_ptr ddr_mem_ptr = ddr->params[pclass][mem].address;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
185
char *host_mem_ptr = host->params[pclass][mem].address;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
187
if (size != ddr->params[pclass][mem].size)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
21
mem_init->params[pclass][mem].address = address;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
22
mem_init->params[pclass][mem].size = (uint32_t)size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
29
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
32
mem_init->params[pclass][mem].address = address;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
33
mem_init->params[pclass][mem].size = (uint32_t)size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
40
enum ia_css_isp_memories mem,
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
43
mem_init->params[pclass][mem].address = address;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
44
mem_init->params[pclass][mem].size = (uint32_t)size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
52
enum ia_css_isp_memories mem)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
54
return &mem_init->params[pclass][mem];
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
61
enum ia_css_isp_memories mem)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
63
return &mem_init->params[pclass][mem];
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
70
enum ia_css_isp_memories mem)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
72
return &mem_init->params[pclass][mem];
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
81
unsigned int pclass, mem;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
85
for (mem = 0; mem < IA_CSS_NUM_MEMORIES; mem++) {
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
86
if (!mem_params->params[pclass][mem].address)
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
88
isp_mem_if->params[pclass][mem].size = mem_params->params[pclass][mem].size;
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
90
isp_mem_if->params[pclass][mem].address =
drivers/staging/media/atomisp/pci/runtime/isp_param/src/isp_param.c
91
css_params->params[pclass][mem].address;
drivers/staging/media/atomisp/pci/sh_css_params.c
2966
enum ia_css_isp_memories mem)
drivers/staging/media/atomisp/pci/sh_css_params.c
2973
IA_CSS_PARAM_CLASS_PARAM, mem);
drivers/staging/media/atomisp/pci/sh_css_params.c
3134
unsigned int mem;
drivers/staging/media/atomisp/pci/sh_css_params.c
3153
for (mem = 0; mem < IA_CSS_NUM_MEMORIES; mem++) {
drivers/staging/media/atomisp/pci/sh_css_params.c
3155
[pipeline->pipe_id][stage->stage_num][mem] = false;
drivers/staging/media/atomisp/pci/sh_css_params.c
3263
unsigned int mem;
drivers/staging/media/atomisp/pci/sh_css_params.c
3597
for (mem = 0; mem < N_IA_CSS_MEMORIES; mem++) {
drivers/staging/media/atomisp/pci/sh_css_params.c
3600
IA_CSS_PARAM_CLASS_PARAM, mem);
drivers/staging/media/atomisp/pci/sh_css_params.c
3604
buff_realloced = reallocate_buffer(&ddr_map->isp_mem_param[stage_num][mem],
drivers/staging/media/atomisp/pci/sh_css_params.c
3605
&ddr_map_size->isp_mem_param[stage_num][mem],
drivers/staging/media/atomisp/pci/sh_css_params.c
3607
params->isp_mem_params_changed[pipe_id][stage_num][mem],
drivers/staging/media/atomisp/pci/sh_css_params.c
3613
if (params->isp_mem_params_changed[pipe_id][stage_num][mem] || buff_realloced) {
drivers/staging/media/atomisp/pci/sh_css_params.c
3615
ddr_map->isp_mem_param[stage_num][mem],
drivers/staging/media/atomisp/pci/sh_css_params.c
3616
ddr_map_size->isp_mem_param[stage_num][mem], mem);
drivers/staging/media/atomisp/pci/sh_css_params.c
3893
unsigned int i, j, mem;
drivers/staging/media/atomisp/pci/sh_css_params.c
3902
for (mem = 0; mem < N_IA_CSS_MEMORIES; mem++) {
drivers/staging/media/atomisp/pci/sh_css_params.c
3903
params->isp_mem_params_changed[i][j][mem] = true;
drivers/staging/media/av7110/av7110.c
1224
u8 *mem = (u8 *)(budget->grabbing);
drivers/staging/media/av7110/av7110.c
1246
pr_info("%s(): %02x Event Counter 1 0x%04x\n", __func__, mem[olddma],
drivers/staging/media/av7110/av7110.c
1252
dvb_dmx_swfilter_packets(demux, mem + olddma, (newdma - olddma) / 188);
drivers/staging/media/av7110/av7110.c
1255
dvb_dmx_swfilter_packets(demux, mem + olddma, (TS_BUFLEN - olddma) / 188);
drivers/staging/media/av7110/av7110.c
1256
dvb_dmx_swfilter_packets(demux, mem, newdma / 188);
drivers/staging/media/ipu3/ipu3-css-fw.c
74
enum imgu_abi_memories mem,
drivers/staging/media/ipu3/ipu3-css-fw.c
82
bi->info.isp.sp.mem_initializers.params[cls][mem].size)
drivers/staging/media/ipu3/ipu3-css-fw.h
189
enum imgu_abi_memories mem,
drivers/staging/media/ipu3/ipu3-css.c
1172
&css_pipe->aux_frames[j].mem[i]);
drivers/staging/media/ipu3/ipu3-css.c
1177
&css_pipe->aux_frames[j].mem[i]);
drivers/staging/media/ipu3/ipu3-css.c
1197
&css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_REF].mem[i],
drivers/staging/media/ipu3/ipu3-css.c
1203
&css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_TNR].mem[i],
drivers/staging/media/ipu3/ipu3-css.c
1253
&css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_REF].mem[i],
drivers/staging/media/ipu3/ipu3-css.c
1274
&css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_TNR].mem[i],
drivers/staging/media/ipu3/ipu3-css.c
177
static inline void writes(const void *mem, ssize_t count, void __iomem *addr)
drivers/staging/media/ipu3/ipu3-css.c
180
const u32 *buf = mem;
drivers/staging/media/ipu3/ipu3-css.c
790
css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_REF].mem[i].daddr;
drivers/staging/media/ipu3/ipu3-css.c
792
css_pipe->aux_frames[IPU3_CSS_AUX_FRAME_REF].mem[i].daddr +
drivers/staging/media/ipu3/ipu3-css.c
840
.mem[i].daddr;
drivers/staging/media/ipu3/ipu3-css.h
128
struct imgu_css_map mem[IPU3_CSS_AUX_FRAMES];
drivers/staging/rtl8723bs/hal/sdio_ops.c
389
u8 *mem
drivers/staging/rtl8723bs/hal/sdio_ops.c
406
err = _sd_read(intfhdl, addr, cnt, mem);
drivers/staging/rtl8723bs/hal/sdio_ops.c
433
u8 *mem
drivers/staging/rtl8723bs/hal/sdio_ops.c
439
struct xmit_buf *xmitbuf = (struct xmit_buf *)mem;
drivers/staging/rtl8723bs/include/rtw_recv.h
314
uint mem[RECVFRAME_HDR_ALIGN>>2];
drivers/target/target_core_configfs.c
2575
struct t10_alua_lba_map_member *mem;
drivers/target/target_core_configfs.c
2588
list_for_each_entry(mem, &map->lba_map_mem_list,
drivers/target/target_core_configfs.c
2590
switch (mem->lba_map_mem_alua_state) {
drivers/target/target_core_configfs.c
2608
mem->lba_map_mem_alua_pg_id, state);
drivers/target/target_core_user.c
1832
if (info->mem[vma->vm_pgoff].size == 0)
drivers/target/target_core_user.c
1903
addr = (void *)(unsigned long)info->mem[mi].addr + offset;
drivers/target/target_core_user.c
2248
info->mem[0].name = "tcm-user command & data buffer";
drivers/target/target_core_user.c
2249
info->mem[0].addr = (phys_addr_t)(uintptr_t)udev->mb_addr;
drivers/target/target_core_user.c
2250
info->mem[0].size = data_size + udev->cmdr_size + CMDR_OFF;
drivers/target/target_core_user.c
2251
info->mem[0].memtype = UIO_MEM_NONE;
drivers/tty/serial/8250/8250_men_mcb.c
101
iounmap(mem);
drivers/tty/serial/8250/8250_men_mcb.c
183
struct resource *mem;
drivers/tty/serial/8250/8250_men_mcb.c
187
mem = mcb_get_resource(mdev, IORESOURCE_MEM);
drivers/tty/serial/8250/8250_men_mcb.c
188
if (mem == NULL)
drivers/tty/serial/8250/8250_men_mcb.c
197
res = init_serial_data(mdev, mem, data);
drivers/tty/serial/8250/8250_men_mcb.c
217
uart.port.mapbase = (unsigned long) mem->start
drivers/tty/serial/8250/8250_men_mcb.c
83
void __iomem *mem;
drivers/tty/serial/8250/8250_men_mcb.c
91
mem = ioremap(mem_res->start + MEN_Z025_REGISTER_OFFSET,
drivers/tty/serial/8250/8250_men_mcb.c
93
if (!mem) {
drivers/tty/serial/8250/8250_men_mcb.c
99
reg_value = MEN_READ_REGISTER(mem);
drivers/tty/serial/cpm_uart.c
1256
void __iomem *mem, *pram;
drivers/tty/serial/cpm_uart.c
1286
mem = of_iomap(np, 0);
drivers/tty/serial/cpm_uart.c
1287
if (!mem)
drivers/tty/serial/cpm_uart.c
1292
pinfo->sccp = mem;
drivers/tty/serial/cpm_uart.c
1297
pinfo->smcp = mem;
drivers/tty/serial/cpm_uart.c
1315
pinfo->port.mapbase = (unsigned long)mem;
drivers/tty/serial/cpm_uart.c
1361
iounmap(mem);
drivers/tty/serial/cpm_uart.h
86
u32 mem = (u32)pinfo->mem_addr;
drivers/tty/serial/cpm_uart.h
88
if (likely(val >= mem && val < mem + pinfo->mem_size)) {
drivers/tty/serial/cpm_uart.h
89
offset = val - mem;
drivers/tty/serial/men_z135_uart.c
122
struct resource *mem;
drivers/tty/serial/men_z135_uart.c
728
mcb_release_mem(uart->mem);
drivers/tty/serial/men_z135_uart.c
735
struct resource *mem;
drivers/tty/serial/men_z135_uart.c
737
mem = mcb_request_mem(uart->mdev, dev_name(&mdev->dev));
drivers/tty/serial/men_z135_uart.c
738
if (IS_ERR(mem))
drivers/tty/serial/men_z135_uart.c
739
return PTR_ERR(mem);
drivers/tty/serial/men_z135_uart.c
741
port->mapbase = mem->start;
drivers/tty/serial/men_z135_uart.c
742
uart->mem = mem;
drivers/tty/serial/men_z135_uart.c
744
port->membase = ioremap(mem->start, resource_size(mem));
drivers/tty/serial/men_z135_uart.c
746
mcb_release_mem(mem);
drivers/tty/serial/men_z135_uart.c
804
struct resource *mem;
drivers/tty/serial/men_z135_uart.c
818
mem = &mdev->mem;
drivers/tty/serial/men_z135_uart.c
831
uart->port.mapbase = mem->start;
drivers/tty/serial/omap-serial.c
1543
struct resource *mem;
drivers/tty/serial/omap-serial.c
1567
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/tty/serial/omap-serial.c
1606
up->port.mapbase = mem->start;
drivers/tty/serial/sifive.c
1019
ssp->port.mapbase = mem->start;
drivers/tty/serial/sifive.c
976
struct resource *mem;
drivers/tty/serial/sifive.c
985
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/uio/uio.c
106
struct uio_mem *mem = map->mem;
drivers/uio/uio.c
114
return entry->show(mem, buf);
drivers/uio/uio.c
291
struct uio_mem *mem;
drivers/uio/uio.c
297
mem = &idev->info->mem[mi];
drivers/uio/uio.c
298
if (mem->size == 0)
drivers/uio/uio.c
315
map->mem = mem;
drivers/uio/uio.c
316
mem->map = map;
drivers/uio/uio.c
370
mem = &idev->info->mem[mi];
drivers/uio/uio.c
371
map = mem->map;
drivers/uio/uio.c
382
struct uio_mem *mem;
drivers/uio/uio.c
386
mem = &idev->info->mem[i];
drivers/uio/uio.c
387
if (mem->size == 0)
drivers/uio/uio.c
389
kobject_put(&mem->map->kobj);
drivers/uio/uio.c
45
struct uio_mem *mem;
drivers/uio/uio.c
49
static ssize_t map_name_show(struct uio_mem *mem, char *buf)
drivers/uio/uio.c
51
if (unlikely(!mem->name))
drivers/uio/uio.c
52
mem->name = "";
drivers/uio/uio.c
54
return sprintf(buf, "%s\n", mem->name);
drivers/uio/uio.c
57
static ssize_t map_addr_show(struct uio_mem *mem, char *buf)
drivers/uio/uio.c
59
return sprintf(buf, "%pa\n", &mem->addr);
drivers/uio/uio.c
62
static ssize_t map_size_show(struct uio_mem *mem, char *buf)
drivers/uio/uio.c
64
return sprintf(buf, "%pa\n", &mem->size);
drivers/uio/uio.c
67
static ssize_t map_offset_show(struct uio_mem *mem, char *buf)
drivers/uio/uio.c
677
if (idev->info->mem[vma->vm_pgoff].size == 0)
drivers/uio/uio.c
69
return sprintf(buf, "0x%llx\n", (unsigned long long)mem->offs);
drivers/uio/uio.c
711
addr = (void *)(unsigned long)idev->info->mem[mi].addr + offset;
drivers/uio/uio.c
712
if (idev->info->mem[mi].memtype == UIO_MEM_LOGICAL)
drivers/uio/uio.c
746
struct uio_mem *mem;
drivers/uio/uio.c
750
mem = idev->info->mem + mi;
drivers/uio/uio.c
752
if (mem->addr & ~PAGE_MASK)
drivers/uio/uio.c
754
if (vma->vm_end - vma->vm_start > mem->size)
drivers/uio/uio.c
758
if (idev->info->mem[mi].memtype == UIO_MEM_PHYS)
drivers/uio/uio.c
772
mem->addr >> PAGE_SHIFT,
drivers/uio/uio.c
780
struct uio_mem *mem;
drivers/uio/uio.c
789
mem = idev->info->mem + mi;
drivers/uio/uio.c
791
if (mem->addr & ~PAGE_MASK)
drivers/uio/uio.c
793
if (mem->dma_addr & ~PAGE_MASK)
drivers/uio/uio.c
795
if (!mem->dma_device)
drivers/uio/uio.c
797
if (vma->vm_end - vma->vm_start > mem->size)
drivers/uio/uio.c
800
dev_warn(mem->dma_device,
drivers/uio/uio.c
809
addr = (void *)(uintptr_t)mem->addr;
drivers/uio/uio.c
810
ret = dma_mmap_coherent(mem->dma_device,
drivers/uio/uio.c
813
mem->dma_addr,
drivers/uio/uio.c
846
actual_pages = ((idev->info->mem[mi].addr & ~PAGE_MASK)
drivers/uio/uio.c
847
+ idev->info->mem[mi].size + PAGE_SIZE -1) >> PAGE_SHIFT;
drivers/uio/uio.c
858
switch (idev->info->mem[mi].memtype) {
drivers/uio/uio_cif.c
105
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_cif.c
29
void __iomem *plx_intscr = dev_info->mem[0].internal_addr
drivers/uio/uio_cif.c
56
info->mem[0].addr = pci_resource_start(dev, 0);
drivers/uio/uio_cif.c
57
if (!info->mem[0].addr)
drivers/uio/uio_cif.c
59
info->mem[0].internal_addr = pci_ioremap_bar(dev, 0);
drivers/uio/uio_cif.c
60
if (!info->mem[0].internal_addr)
drivers/uio/uio_cif.c
63
info->mem[0].size = pci_resource_len(dev, 0);
drivers/uio/uio_cif.c
64
info->mem[0].memtype = UIO_MEM_PHYS;
drivers/uio/uio_cif.c
65
info->mem[1].addr = pci_resource_start(dev, 2);
drivers/uio/uio_cif.c
66
info->mem[1].size = pci_resource_len(dev, 2);
drivers/uio/uio_cif.c
67
info->mem[1].memtype = UIO_MEM_PHYS;
drivers/uio/uio_cif.c
90
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_dfl.c
29
uiomem = &uioinfo->mem[0];
drivers/uio/uio_dmem_genirq.c
228
uiomem = &uioinfo->mem[0];
drivers/uio/uio_dmem_genirq.c
236
if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_dmem_genirq.c
249
priv->dmem_region_start = uiomem - &uioinfo->mem[0];
drivers/uio/uio_dmem_genirq.c
253
if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_dmem_genirq.c
266
while (uiomem < &uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_dmem_genirq.c
53
uiomem = &priv->uioinfo->mem[priv->dmem_region_start];
drivers/uio/uio_dmem_genirq.c
56
while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_dmem_genirq.c
82
uiomem = &priv->uioinfo->mem[priv->dmem_region_start];
drivers/uio/uio_dmem_genirq.c
87
while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_fsl_elbc_gpcm.c
158
void __iomem *reg_int_en = info->mem[0].internal_addr +
drivers/uio/uio_fsl_elbc_gpcm.c
161
void __iomem *reg_int_stat = info->mem[0].internal_addr +
drivers/uio/uio_fsl_elbc_gpcm.c
190
iowrite32(0, info->mem[0].internal_addr + win0_offset +
drivers/uio/uio_fsl_elbc_gpcm.c
200
iowrite32(0, info->mem[0].internal_addr + DPM_HOST_WIN0_OFFSET +
drivers/uio/uio_fsl_elbc_gpcm.c
379
info->mem[0].internal_addr = ioremap(res.start, resource_size(&res));
drivers/uio/uio_fsl_elbc_gpcm.c
380
if (!info->mem[0].internal_addr) {
drivers/uio/uio_fsl_elbc_gpcm.c
386
info->mem[0].name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%pOFn", node);
drivers/uio/uio_fsl_elbc_gpcm.c
387
if (!info->mem[0].name) {
drivers/uio/uio_fsl_elbc_gpcm.c
392
info->mem[0].addr = res.start;
drivers/uio/uio_fsl_elbc_gpcm.c
393
info->mem[0].size = resource_size(&res);
drivers/uio/uio_fsl_elbc_gpcm.c
394
info->mem[0].memtype = UIO_MEM_PHYS;
drivers/uio/uio_fsl_elbc_gpcm.c
433
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_fsl_elbc_gpcm.c
446
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_hv_generic.c
284
pdata->info.mem[TXRX_RING_MAP].name = "txrx_rings";
drivers/uio/uio_hv_generic.c
286
pdata->info.mem[TXRX_RING_MAP].addr
drivers/uio/uio_hv_generic.c
288
pdata->info.mem[TXRX_RING_MAP].size
drivers/uio/uio_hv_generic.c
290
pdata->info.mem[TXRX_RING_MAP].memtype = UIO_MEM_IOVA;
drivers/uio/uio_hv_generic.c
292
pdata->info.mem[INT_PAGE_MAP].name = "int_page";
drivers/uio/uio_hv_generic.c
293
pdata->info.mem[INT_PAGE_MAP].addr
drivers/uio/uio_hv_generic.c
295
pdata->info.mem[INT_PAGE_MAP].size = HV_HYP_PAGE_SIZE;
drivers/uio/uio_hv_generic.c
296
pdata->info.mem[INT_PAGE_MAP].memtype = UIO_MEM_LOGICAL;
drivers/uio/uio_hv_generic.c
298
pdata->info.mem[MON_PAGE_MAP].name = "monitor_page";
drivers/uio/uio_hv_generic.c
299
pdata->info.mem[MON_PAGE_MAP].addr
drivers/uio/uio_hv_generic.c
301
pdata->info.mem[MON_PAGE_MAP].size = HV_HYP_PAGE_SIZE;
drivers/uio/uio_hv_generic.c
302
pdata->info.mem[MON_PAGE_MAP].memtype = UIO_MEM_LOGICAL;
drivers/uio/uio_hv_generic.c
322
pdata->info.mem[RECV_BUF_MAP].name = pdata->recv_name;
drivers/uio/uio_hv_generic.c
323
pdata->info.mem[RECV_BUF_MAP].addr = (uintptr_t)pdata->recv_buf;
drivers/uio/uio_hv_generic.c
324
pdata->info.mem[RECV_BUF_MAP].size = RECV_BUFFER_SIZE;
drivers/uio/uio_hv_generic.c
325
pdata->info.mem[RECV_BUF_MAP].memtype = UIO_MEM_VIRTUAL;
drivers/uio/uio_hv_generic.c
343
pdata->info.mem[SEND_BUF_MAP].name = pdata->send_name;
drivers/uio/uio_hv_generic.c
344
pdata->info.mem[SEND_BUF_MAP].addr = (uintptr_t)pdata->send_buf;
drivers/uio/uio_hv_generic.c
345
pdata->info.mem[SEND_BUF_MAP].size = SEND_BUFFER_SIZE;
drivers/uio/uio_hv_generic.c
346
pdata->info.mem[SEND_BUF_MAP].memtype = UIO_MEM_VIRTUAL;
drivers/uio/uio_mf624.c
117
static int mf624_setup_mem(struct pci_dev *dev, int bar, struct uio_mem *mem, const char *name)
drivers/uio/uio_mf624.c
122
mem->name = name;
drivers/uio/uio_mf624.c
123
mem->addr = start & PAGE_MASK;
drivers/uio/uio_mf624.c
124
mem->offs = start & ~PAGE_MASK;
drivers/uio/uio_mf624.c
125
if (!mem->addr)
drivers/uio/uio_mf624.c
127
mem->size = ((start & ~PAGE_MASK) + len + PAGE_SIZE - 1) & PAGE_MASK;
drivers/uio/uio_mf624.c
128
mem->memtype = UIO_MEM_PHYS;
drivers/uio/uio_mf624.c
129
mem->internal_addr = pci_ioremap_bar(dev, bar);
drivers/uio/uio_mf624.c
130
if (!mem->internal_addr)
drivers/uio/uio_mf624.c
155
if (mf624_setup_mem(dev, 0, &info->mem[0], "PCI chipset, interrupts, status "
drivers/uio/uio_mf624.c
159
if (mf624_setup_mem(dev, 2, &info->mem[1], "ADC, DAC, DIO"))
drivers/uio/uio_mf624.c
163
if (mf624_setup_mem(dev, 4, &info->mem[2], "Counter/timer chip"))
drivers/uio/uio_mf624.c
180
iounmap(info->mem[2].internal_addr);
drivers/uio/uio_mf624.c
182
iounmap(info->mem[1].internal_addr);
drivers/uio/uio_mf624.c
184
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_mf624.c
205
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_mf624.c
206
iounmap(info->mem[1].internal_addr);
drivers/uio/uio_mf624.c
207
iounmap(info->mem[2].internal_addr);
drivers/uio/uio_mf624.c
35
void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR;
drivers/uio/uio_mf624.c
63
void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR;
drivers/uio/uio_mf624.c
90
void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR;
drivers/uio/uio_netx.c
110
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_netx.c
123
iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0);
drivers/uio/uio_netx.c
127
iounmap(info->mem[0].internal_addr);
drivers/uio/uio_netx.c
34
void __iomem *int_enable_reg = dev_info->mem[0].internal_addr
drivers/uio/uio_netx.c
36
void __iomem *int_status_reg = dev_info->mem[0].internal_addr
drivers/uio/uio_netx.c
81
info->mem[0].addr = pci_resource_start(dev, bar);
drivers/uio/uio_netx.c
82
if (!info->mem[0].addr)
drivers/uio/uio_netx.c
84
info->mem[0].internal_addr = ioremap(pci_resource_start(dev, bar),
drivers/uio/uio_netx.c
87
if (!info->mem[0].internal_addr)
drivers/uio/uio_netx.c
90
info->mem[0].size = pci_resource_len(dev, bar);
drivers/uio/uio_netx.c
91
info->mem[0].memtype = UIO_MEM_PHYS;
drivers/uio/uio_netx.c
98
iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0);
drivers/uio/uio_pci_generic.c
106
uiomem = &gdev->info.mem[0];
drivers/uio/uio_pci_generic.c
113
if (uiomem >= &gdev->info.mem[MAX_UIO_MAPS]) {
drivers/uio/uio_pci_generic.c
131
while (uiomem < &gdev->info.mem[MAX_UIO_MAPS]) {
drivers/uio/uio_pci_generic_sva.c
109
struct uio_mem *uiomem = &udev->info.mem[i];
drivers/uio/uio_pci_generic_sva.c
114
if (uiomem >= &udev->info.mem[MAX_UIO_MAPS]) {
drivers/uio/uio_pdrv_genirq.c
189
uiomem = &uioinfo->mem[0];
drivers/uio/uio_pdrv_genirq.c
197
if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_pdrv_genirq.c
213
while (uiomem < &uioinfo->mem[MAX_UIO_MAPS]) {
drivers/uio/uio_sercos3.c
108
info->mem[n].addr = pci_resource_start(dev, pci_bar);
drivers/uio/uio_sercos3.c
109
if (!info->mem[n].addr)
drivers/uio/uio_sercos3.c
111
info->mem[n].internal_addr = ioremap(pci_resource_start(dev, pci_bar),
drivers/uio/uio_sercos3.c
113
if (!info->mem[n].internal_addr)
drivers/uio/uio_sercos3.c
115
info->mem[n].size = pci_resource_len(dev, pci_bar);
drivers/uio/uio_sercos3.c
116
info->mem[n].memtype = UIO_MEM_PHYS;
drivers/uio/uio_sercos3.c
171
if (info->mem[i].internal_addr)
drivers/uio/uio_sercos3.c
172
iounmap(info->mem[i].internal_addr);
drivers/uio/uio_sercos3.c
189
if (info->mem[i].internal_addr)
drivers/uio/uio_sercos3.c
190
iounmap(info->mem[i].internal_addr);
drivers/uio/uio_sercos3.c
55
void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET;
drivers/uio/uio_sercos3.c
68
void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET;
drivers/uio/uio_sercos3.c
78
void __iomem *isr0 = info->mem[3].internal_addr + ISR0_OFFSET;
drivers/uio/uio_sercos3.c
79
void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET;
drivers/usb/core/devio.c
1840
as->urb->transfer_buffer = as->usbm->mem +
drivers/usb/core/devio.c
199
usbm->mem, usbm->dma_handle);
drivers/usb/core/devio.c
236
void *mem;
drivers/usb/core/devio.c
254
mem = hcd_buffer_alloc_pages(hcd,
drivers/usb/core/devio.c
256
if (!mem) {
drivers/usb/core/devio.c
261
memset(mem, 0, size);
drivers/usb/core/devio.c
263
usbm->mem = mem;
drivers/usb/core/devio.c
280
virt_to_phys(usbm->mem) >> PAGE_SHIFT,
drivers/usb/core/devio.c
286
if (dma_mmap_coherent(hcd->self.sysdev, vma, mem, dma_handle,
drivers/usb/core/devio.c
88
void *mem;
drivers/usb/gadget/composite.c
1413
unsigned mem;
drivers/usb/gadget/composite.c
1418
mem = sizeof(*uc);
drivers/usb/gadget/composite.c
1419
mem += sizeof(void *) * (n_gstrings + 1);
drivers/usb/gadget/composite.c
1420
mem += sizeof(struct usb_gadget_strings) * n_gstrings;
drivers/usb/gadget/composite.c
1421
mem += sizeof(struct usb_string) * (n_strings + 1) * (n_gstrings);
drivers/usb/gadget/composite.c
1422
uc = kmalloc(mem, GFP_KERNEL);
drivers/usb/gadget/config.c
74
void *mem;
drivers/usb/gadget/config.c
82
mem = kmalloc(bytes, GFP_KERNEL);
drivers/usb/gadget/config.c
83
if (!mem)
drivers/usb/gadget/config.c
90
tmp = mem;
drivers/usb/gadget/config.c
91
ret = mem;
drivers/usb/gadget/config.c
92
mem += (n_desc + 1) * sizeof(*tmp);
drivers/usb/gadget/config.c
94
memcpy(mem, *src, (*src)->bLength);
drivers/usb/gadget/config.c
95
*tmp = mem;
drivers/usb/gadget/config.c
97
mem += (*src)->bLength;
drivers/usb/gadget/function/f_uvc.c
482
#define UVC_COPY_DESCRIPTOR(mem, dst, desc) \
drivers/usb/gadget/function/f_uvc.c
484
memcpy(mem, desc, (desc)->bLength); \
drivers/usb/gadget/function/f_uvc.c
485
*(dst)++ = mem; \
drivers/usb/gadget/function/f_uvc.c
486
mem += (desc)->bLength; \
drivers/usb/gadget/function/f_uvc.c
489
#define UVC_COPY_DESCRIPTORS(mem, dst, src) \
drivers/usb/gadget/function/f_uvc.c
493
memcpy(mem, *__src, (*__src)->bLength); \
drivers/usb/gadget/function/f_uvc.c
494
*dst++ = mem; \
drivers/usb/gadget/function/f_uvc.c
495
mem += (*__src)->bLength; \
drivers/usb/gadget/function/f_uvc.c
499
#define UVC_COPY_XU_DESCRIPTOR(mem, dst, desc) \
drivers/usb/gadget/function/f_uvc.c
501
*(dst)++ = mem; \
drivers/usb/gadget/function/f_uvc.c
502
memcpy(mem, desc, 22); /* bLength to bNrInPins */ \
drivers/usb/gadget/function/f_uvc.c
503
mem += 22; \
drivers/usb/gadget/function/f_uvc.c
505
memcpy(mem, (desc)->baSourceID, (desc)->bNrInPins); \
drivers/usb/gadget/function/f_uvc.c
506
mem += (desc)->bNrInPins; \
drivers/usb/gadget/function/f_uvc.c
508
memcpy(mem, &(desc)->bControlSize, 1); \
drivers/usb/gadget/function/f_uvc.c
509
mem++; \
drivers/usb/gadget/function/f_uvc.c
511
memcpy(mem, (desc)->bmControls, (desc)->bControlSize); \
drivers/usb/gadget/function/f_uvc.c
512
mem += (desc)->bControlSize; \
drivers/usb/gadget/function/f_uvc.c
514
memcpy(mem, &(desc)->iExtension, 1); \
drivers/usb/gadget/function/f_uvc.c
515
mem++; \
drivers/usb/gadget/function/f_uvc.c
534
void *mem;
drivers/usb/gadget/function/f_uvc.c
617
mem = kmalloc((n_desc + 1) * sizeof(*src) + bytes, GFP_KERNEL);
drivers/usb/gadget/function/f_uvc.c
618
if (mem == NULL)
drivers/usb/gadget/function/f_uvc.c
621
hdr = mem;
drivers/usb/gadget/function/f_uvc.c
622
dst = mem;
drivers/usb/gadget/function/f_uvc.c
623
mem += (n_desc + 1) * sizeof(*src);
drivers/usb/gadget/function/f_uvc.c
626
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_iad);
drivers/usb/gadget/function/f_uvc.c
627
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_intf);
drivers/usb/gadget/function/f_uvc.c
629
uvc_control_header = mem;
drivers/usb/gadget/function/f_uvc.c
630
UVC_COPY_DESCRIPTORS(mem, dst,
drivers/usb/gadget/function/f_uvc.c
634
UVC_COPY_XU_DESCRIPTOR(mem, dst, &xu->desc);
drivers/usb/gadget/function/f_uvc.c
641
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_interrupt_ep);
drivers/usb/gadget/function/f_uvc.c
644
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_ss_interrupt_comp);
drivers/usb/gadget/function/f_uvc.c
646
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_interrupt_cs_ep);
drivers/usb/gadget/function/f_uvc.c
649
UVC_COPY_DESCRIPTOR(mem, dst, &uvc_streaming_intf_alt0);
drivers/usb/gadget/function/f_uvc.c
651
uvc_streaming_header = mem;
drivers/usb/gadget/function/f_uvc.c
652
UVC_COPY_DESCRIPTORS(mem, dst,
drivers/usb/gadget/function/f_uvc.c
657
UVC_COPY_DESCRIPTORS(mem, dst, uvc_streaming_std);
drivers/usb/gadget/function/uvc_queue.c
82
buf->mem = vb2_plane_vaddr(vb, 0);
drivers/usb/gadget/function/uvc_queue.h
36
void *mem;
drivers/usb/gadget/function/uvc_video.c
101
ret = uvc_video_encode_header(video, buf, mem, len);
drivers/usb/gadget/function/uvc_video.c
103
mem += ret;
drivers/usb/gadget/function/uvc_video.c
109
ret = uvc_video_encode_data(video, buf, mem, len);
drivers/usb/gadget/function/uvc_video.c
204
void *mem = req->buf;
drivers/usb/gadget/function/uvc_video.c
210
ret = uvc_video_encode_header(video, buf, mem, len);
drivers/usb/gadget/function/uvc_video.c
211
mem += ret;
drivers/usb/gadget/function/uvc_video.c
215
ret = uvc_video_encode_data(video, buf, mem, len);
drivers/usb/gadget/function/uvc_video.c
78
void *mem;
drivers/usb/gadget/function/uvc_video.c
81
mem = buf->mem + queue->buf_used;
drivers/usb/gadget/function/uvc_video.c
84
memcpy(data, mem, nbytes);
drivers/usb/gadget/function/uvc_video.c
94
void *mem = req->buf;
drivers/usb/gadget/udc/bdc/bdc.h
438
void *mem;
drivers/usb/host/ohci-da8xx.c
363
struct resource *mem;
drivers/usb/host/ohci-da8xx.c
422
hcd->regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
drivers/usb/host/ohci-da8xx.c
427
hcd->rsrc_start = mem->start;
drivers/usb/host/ohci-da8xx.c
428
hcd->rsrc_len = resource_size(mem);
drivers/usb/host/ohci-sm501.c
100
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
drivers/usb/host/ohci-sm501.c
101
if (mem == NULL) {
drivers/usb/host/ohci-sm501.c
107
if (!request_mem_region(mem->start, resource_size(mem), pdev->name)) {
drivers/usb/host/ohci-sm501.c
160
retval = usb_hcd_setup_local_mem(hcd, mem->start,
drivers/usb/host/ohci-sm501.c
161
mem->start - mem->parent->start,
drivers/usb/host/ohci-sm501.c
162
resource_size(mem));
drivers/usb/host/ohci-sm501.c
183
release_mem_region(mem->start, resource_size(mem));
drivers/usb/host/ohci-sm501.c
191
struct resource *mem;
drivers/usb/host/ohci-sm501.c
197
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
drivers/usb/host/ohci-sm501.c
198
release_mem_region(mem->start, resource_size(mem));
drivers/usb/host/ohci-sm501.c
92
struct resource *res, *mem;
drivers/usb/host/oxu210hp-hcd.c
1012
index = qh - &oxu->mem->qh_pool[0];
drivers/usb/host/oxu210hp-hcd.c
1045
qh = (struct ehci_qh *) &oxu->mem->qh_pool[i];
drivers/usb/host/oxu210hp-hcd.c
1146
oxu->mem->frame_list[i] = EHCI_LIST_END;
drivers/usb/host/oxu210hp-hcd.c
1163
oxu->periodic = (__le32 *) &oxu->mem->frame_list;
drivers/usb/host/oxu210hp-hcd.c
3075
oxu->mem = hcd->regs + OXU_SPH_MEM;
drivers/usb/host/oxu210hp-hcd.c
3081
oxu->mem = hcd->regs + OXU_OTG_MEM;
drivers/usb/host/oxu210hp-hcd.c
387
struct oxu_onchip_mem __iomem *mem;
drivers/usb/host/oxu210hp-hcd.c
917
qtd->buffer = (void *) &oxu->mem->db_pool[i];
drivers/usb/host/oxu210hp-hcd.c
941
index = (qtd->buffer - (void *) &oxu->mem->db_pool[0])
drivers/usb/host/oxu210hp-hcd.c
970
index = qtd - &oxu->mem->qtd_pool[0];
drivers/usb/host/oxu210hp-hcd.c
988
qtd = (struct ehci_qtd *) &oxu->mem->qtd_pool[i];
drivers/usb/isp1760/isp1760-core.c
481
int isp1760_register(struct resource *mem, int irq, unsigned long irqflags,
drivers/usb/isp1760/isp1760-core.c
540
hcd->base = devm_ioremap_resource(dev, mem);
drivers/usb/isp1760/isp1760-core.c
578
ret = isp1760_hcd_register(hcd, mem, irq,
drivers/usb/isp1760/isp1760-core.h
53
int isp1760_register(struct resource *mem, int irq, unsigned long irqflags,
drivers/usb/isp1760/isp1760-hcd.c
1006
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
1009
WARN_ON((slot < 0) || (slot > mem->slot_num - 1));
drivers/usb/isp1760/isp1760-hcd.c
1119
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
1120
int slot_num = mem->slot_num;
drivers/usb/isp1760/isp1760-hcd.c
1557
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
1564
for (slot = 0; slot < mem->slot_num; slot++)
drivers/usb/isp1760/isp1760-hcd.c
1774
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
1827
if (len > mem->blocks_size[ISP176x_BLOCK_NUM - 1])
drivers/usb/isp1760/isp1760-hcd.c
1828
this_qtd_len = mem->blocks_size[ISP176x_BLOCK_NUM - 1];
drivers/usb/isp1760/isp1760-hcd.c
2559
int isp1760_hcd_register(struct isp1760_hcd *priv, struct resource *mem,
drivers/usb/isp1760/isp1760-hcd.c
2592
hcd->rsrc_start = mem->start;
drivers/usb/isp1760/isp1760-hcd.c
2593
hcd->rsrc_len = resource_size(mem);
drivers/usb/isp1760/isp1760-hcd.c
584
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
590
for (i = 0, curr = 0; i < ARRAY_SIZE(mem->blocks); i++, curr += j) {
drivers/usb/isp1760/isp1760-hcd.c
591
for (j = 0; j < mem->blocks[i]; j++) {
drivers/usb/isp1760/isp1760-hcd.c
593
priv->memory_pool[curr + j].size = mem->blocks_size[i];
drivers/usb/isp1760/isp1760-hcd.c
600
mem->payload_area_size);
drivers/usb/isp1760/isp1760-hcd.c
606
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
614
for (i = 0; i < mem->payload_blocks; i++) {
drivers/usb/isp1760/isp1760-hcd.c
627
const struct isp1760_memory_layout *mem = priv->memory_layout;
drivers/usb/isp1760/isp1760-hcd.c
633
for (i = 0; i < mem->payload_blocks; i++) {
drivers/usb/isp1760/isp1760-hcd.h
75
int isp1760_hcd_register(struct isp1760_hcd *priv, struct resource *mem,
drivers/usb/isp1760/isp1760-hcd.h
83
struct resource *mem,
drivers/usb/mon/mon_text.c
735
static void mon_text_ctor(void *mem)
drivers/usb/mon/mon_text.c
741
memset(mem, 0xe5, sizeof(struct mon_event_text));
drivers/usb/musb/tusb6010.c
1109
struct resource *mem;
drivers/usb/musb/tusb6010.c
1120
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
drivers/usb/musb/tusb6010.c
1121
if (!mem) {
drivers/usb/musb/tusb6010.c
1126
musb->async = mem->start;
drivers/usb/musb/tusb6010.c
1129
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
drivers/usb/musb/tusb6010.c
1130
if (!mem) {
drivers/usb/musb/tusb6010.c
1135
musb->sync = mem->start;
drivers/usb/musb/tusb6010.c
1137
sync = ioremap(mem->start, resource_size(mem));
drivers/vdpa/mlx5/core/mr.c
101
struct mlx5_destroy_mkey_mem *mem)
drivers/vdpa/mlx5/core/mr.c
103
void *in = &mem->in;
drivers/vdpa/mlx5/core/mr.c
65
struct mlx5_create_mkey_mem *mem)
drivers/vdpa/mlx5/core/mr.c
67
void *in = &mem->in;
drivers/vdpa/mlx5/core/mr.c
92
struct mlx5_create_mkey_mem *mem)
drivers/vdpa/mlx5/core/mr.c
94
u32 mkey_index = MLX5_GET(create_mkey_out, mem->out, mkey_index);
drivers/vfio/vfio_iommu_spapr_tce.c
107
struct mm_iommu_table_group_mem_t *mem;
drivers/vfio/vfio_iommu_spapr_tce.c
115
mem = mm_iommu_get(container->mm, vaddr, size >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
116
if (!mem)
drivers/vfio/vfio_iommu_spapr_tce.c
120
if (tcemem->mem == mem) {
drivers/vfio/vfio_iommu_spapr_tce.c
131
mm_iommu_put(container->mm, mem);
drivers/vfio/vfio_iommu_spapr_tce.c
140
struct mm_iommu_table_group_mem_t *mem = NULL;
drivers/vfio/vfio_iommu_spapr_tce.c
148
mem = mm_iommu_get(container->mm, vaddr, entries);
drivers/vfio/vfio_iommu_spapr_tce.c
149
if (mem) {
drivers/vfio/vfio_iommu_spapr_tce.c
151
if (tcemem->mem == mem) {
drivers/vfio/vfio_iommu_spapr_tce.c
157
ret = mm_iommu_new(container->mm, vaddr, entries, &mem);
drivers/vfio/vfio_iommu_spapr_tce.c
168
tcemem->mem = mem;
drivers/vfio/vfio_iommu_spapr_tce.c
176
mm_iommu_put(container->mm, mem);
drivers/vfio/vfio_iommu_spapr_tce.c
394
struct mm_iommu_table_group_mem_t *mem;
drivers/vfio/vfio_iommu_spapr_tce.c
396
mem = mm_iommu_lookup(container->mm, tce, 1ULL << shift);
drivers/vfio/vfio_iommu_spapr_tce.c
397
if (!mem)
drivers/vfio/vfio_iommu_spapr_tce.c
400
ret = mm_iommu_ua_to_hpa(mem, tce, shift, phpa);
drivers/vfio/vfio_iommu_spapr_tce.c
404
*pmem = mem;
drivers/vfio/vfio_iommu_spapr_tce.c
412
struct mm_iommu_table_group_mem_t *mem = NULL;
drivers/vfio/vfio_iommu_spapr_tce.c
421
tbl->it_page_shift, &hpa, &mem);
drivers/vfio/vfio_iommu_spapr_tce.c
425
if (mem)
drivers/vfio/vfio_iommu_spapr_tce.c
426
mm_iommu_mapped_dec(mem);
drivers/vfio/vfio_iommu_spapr_tce.c
55
struct mm_iommu_table_group_mem_t *mem;
drivers/vfio/vfio_iommu_spapr_tce.c
557
struct mm_iommu_table_group_mem_t *mem = NULL;
drivers/vfio/vfio_iommu_spapr_tce.c
561
tce, tbl->it_page_shift, &hpa, &mem);
drivers/vfio/vfio_iommu_spapr_tce.c
576
if (mm_iommu_mapped_inc(mem))
drivers/vfio/vfio_iommu_spapr_tce.c
94
ret = mm_iommu_put(container->mm, tcemem->mem);
drivers/vhost/vhost.c
1969
struct vhost_memory mem, *newmem;
drivers/vhost/vhost.c
1975
if (copy_from_user(&mem, m, size))
drivers/vhost/vhost.c
1977
if (mem.padding)
drivers/vhost/vhost.c
1979
if (mem.nregions > max_mem_regions)
drivers/vhost/vhost.c
1981
newmem = kvzalloc_flex(*newmem, regions, mem.nregions);
drivers/vhost/vhost.c
1985
memcpy(newmem, &mem, size);
drivers/vhost/vhost.c
1987
flex_array_size(newmem, regions, mem.nregions))) {
drivers/vhost/vhost.c
1999
region < newmem->regions + mem.nregions;
drivers/video/fbdev/arkfb.c
561
int rv, mem, step;
drivers/video/fbdev/arkfb.c
587
mem = ((var->bits_per_pixel * var->xres_virtual) >> 3) * var->yres_virtual;
drivers/video/fbdev/arkfb.c
588
if (mem > info->screen_size)
drivers/video/fbdev/arkfb.c
591
mem >> 10, (unsigned int) (info->screen_size >> 10));
drivers/video/fbdev/aty/aty128fb.c
1000
par->mem = &ddr_sgram;
drivers/video/fbdev/aty/aty128fb.c
1003
par->mem = &sdr_sgram;
drivers/video/fbdev/aty/aty128fb.c
1427
const struct aty128_meminfo *m = par->mem;
drivers/video/fbdev/aty/aty128fb.c
1920
printk("%dM %s\n", par->vram_size / (1024*1024), par->mem->name);
drivers/video/fbdev/aty/aty128fb.c
1922
printk("%dk %s\n", par->vram_size / 1024, par->mem->name);
drivers/video/fbdev/aty/aty128fb.c
444
const struct aty128_meminfo *mem; /* onboard mem info */
drivers/video/fbdev/aty/aty128fb.c
994
par->mem = &sdr_128;
drivers/video/fbdev/aty/aty128fb.c
997
par->mem = &sdr_sgram;
drivers/video/fbdev/aty/atyfb_base.c
2956
u32 mem, chip_id;
drivers/video/fbdev/aty/atyfb_base.c
3047
mem = aty_ld_le32(MEM_CNTL, par);
drivers/video/fbdev/aty/atyfb_base.c
3050
switch (mem & 0x0f) {
drivers/video/fbdev/aty/atyfb_base.c
3052
mem = (mem & ~(0x0f)) | 2;
drivers/video/fbdev/aty/atyfb_base.c
3055
mem = (mem & ~(0x0f)) | 3;
drivers/video/fbdev/aty/atyfb_base.c
3058
mem = (mem & ~(0x0f)) | 4;
drivers/video/fbdev/aty/atyfb_base.c
3061
mem = (mem & ~(0x0f)) | 5;
drivers/video/fbdev/aty/atyfb_base.c
3067
mem &= ~(0x00700000);
drivers/video/fbdev/aty/atyfb_base.c
3069
mem &= ~(0xcf80e000); /* Turn off all undocumented bits. */
drivers/video/fbdev/aty/atyfb_base.c
3070
aty_st_le32(MEM_CNTL, mem, par);
drivers/video/fbdev/cirrusfb.c
1867
unsigned long mem;
drivers/video/fbdev/cirrusfb.c
1873
mem = ((SR14 & 7) + 1) << 20;
drivers/video/fbdev/cirrusfb.c
1878
mem = 512 * 1024;
drivers/video/fbdev/cirrusfb.c
1881
mem = 1024 * 1024;
drivers/video/fbdev/cirrusfb.c
1887
mem = 2048 * 1024;
drivers/video/fbdev/cirrusfb.c
1891
mem = 1024 * 1024;
drivers/video/fbdev/cirrusfb.c
1897
mem *= 2;
drivers/video/fbdev/cirrusfb.c
1901
return mem;
drivers/video/fbdev/core/sysmem.h
28
unsigned long *mem = dst->address;
drivers/video/fbdev/core/sysmem.h
30
mem[offset] = val;
drivers/video/fbdev/core/sysmem.h
36
unsigned long *mem = src->address;
drivers/video/fbdev/core/sysmem.h
38
return mem[offset];
drivers/video/fbdev/cyber2000fb.c
751
unsigned int mem;
drivers/video/fbdev/cyber2000fb.c
829
mem = var->xres_virtual * var->yres_virtual * (var->bits_per_pixel / 8);
drivers/video/fbdev/cyber2000fb.c
830
if (mem > cfb->fb.fix.smem_len)
drivers/video/fbdev/cyber2000fb.c
855
unsigned int mem;
drivers/video/fbdev/cyber2000fb.c
934
mem = cfb->fb.fix.line_length * var->yres_virtual;
drivers/video/fbdev/cyber2000fb.c
935
BUG_ON(mem > cfb->fb.fix.smem_len);
drivers/video/fbdev/g364fb.c
176
int mem, i;
drivers/video/fbdev/g364fb.c
212
mem = (r4030_read_reg32(JAZZ_R4030_CONFIG) >> 8) & 3;
drivers/video/fbdev/g364fb.c
213
fb_fix.smem_len = (1 << (mem * 2)) * 512 * 1024;
drivers/video/fbdev/i740fb.c
403
u32 bpp, base, dacspeed24, mem, freq;
drivers/video/fbdev/i740fb.c
488
mem = vxres * vyres * ((bpp + 1) / 8);
drivers/video/fbdev/i740fb.c
489
if (mem > info->screen_size) {
drivers/video/fbdev/i740fb.c
491
mem >> 10, info->screen_size >> 10);
drivers/video/fbdev/matrox/matroxfb_base.c
1244
static unsigned int mem; /* "matroxfb:mem:xxxxxM" */
drivers/video/fbdev/matrox/matroxfb_base.c
1723
if (mem < 1024) mem *= 1024;
drivers/video/fbdev/matrox/matroxfb_base.c
1724
if (mem < 0x00100000) mem *= 1024;
drivers/video/fbdev/matrox/matroxfb_base.c
1726
if (mem && (mem < memsize))
drivers/video/fbdev/matrox/matroxfb_base.c
1727
memsize = mem;
drivers/video/fbdev/matrox/matroxfb_base.c
2395
mem = simple_strtoul(this_opt+4, NULL, 0);
drivers/video/fbdev/matrox/matroxfb_base.c
2508
module_param(mem, int, 0);
drivers/video/fbdev/matrox/matroxfb_base.c
2509
MODULE_PARM_DESC(mem, "Size of available memory in MB, KB or B (2,4,8,12,16MB, default=autodetect)");
drivers/video/fbdev/matrox/matroxfb_crtc2.c
24
static int mem = 8192;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
26
module_param(mem, int, 0);
drivers/video/fbdev/matrox/matroxfb_crtc2.c
27
MODULE_PARM_DESC(mem, "Memory size reserved for dualhead (default=8MB)");
drivers/video/fbdev/matrox/matroxfb_crtc2.c
609
if (mem < 64)
drivers/video/fbdev/matrox/matroxfb_crtc2.c
610
mem *= 1024;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
611
if (mem < 64*1024)
drivers/video/fbdev/matrox/matroxfb_crtc2.c
612
mem *= 1024;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
613
mem &= ~0x00000FFF; /* PAGE_MASK? */
drivers/video/fbdev/matrox/matroxfb_crtc2.c
614
if (minfo->video.len_usable + mem <= minfo->video.len)
drivers/video/fbdev/matrox/matroxfb_crtc2.c
615
m2info->video.offbase = minfo->video.len - mem;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
616
else if (minfo->video.len < mem) {
drivers/video/fbdev/matrox/matroxfb_crtc2.c
619
m2info->video.borrowed = mem;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
620
minfo->video.len_usable -= mem;
drivers/video/fbdev/matrox/matroxfb_crtc2.c
624
m2info->video.len = m2info->video.len_usable = m2info->video.len_maximum = mem;
drivers/video/fbdev/metronomefb.c
148
static u8 calc_cksum(int start, int end, u8 *mem)
drivers/video/fbdev/metronomefb.c
154
tmp += mem[i];
drivers/video/fbdev/metronomefb.c
170
static int load_waveform(u8 *mem, size_t size, int m, int t,
drivers/video/fbdev/metronomefb.c
195
wfm_hdr = (struct waveform_hdr *) mem;
drivers/video/fbdev/metronomefb.c
205
cksum = calc_cksum(32, 47, mem);
drivers/video/fbdev/metronomefb.c
228
if (mem[i] > t) {
drivers/video/fbdev/metronomefb.c
238
cksum = calc_cksum(sizeof(*wfm_hdr), cksum_idx, mem);
drivers/video/fbdev/metronomefb.c
239
if (cksum != mem[cksum_idx]) {
drivers/video/fbdev/metronomefb.c
241
" %x != %x\n", cksum, mem[cksum_idx]);
drivers/video/fbdev/metronomefb.c
250
cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem);
drivers/video/fbdev/metronomefb.c
251
if (cksum != mem[cksum_idx]) {
drivers/video/fbdev/metronomefb.c
253
" %x != %x\n", cksum, mem[cksum_idx]);
drivers/video/fbdev/metronomefb.c
258
tta = get_unaligned_le32(mem + wmta + m * 4) & 0x00FFFFFF;
drivers/video/fbdev/metronomefb.c
262
cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem);
drivers/video/fbdev/metronomefb.c
263
if (cksum != mem[cksum_idx]) {
drivers/video/fbdev/metronomefb.c
265
" %x != %x\n", cksum, mem[cksum_idx]);
drivers/video/fbdev/metronomefb.c
271
wfm_idx = get_unaligned_le32(mem + tta + trn * 4) & 0x00FFFFFF;
drivers/video/fbdev/metronomefb.c
277
v = mem[wfm_idx++];
drivers/video/fbdev/metronomefb.c
279
while (((v = mem[wfm_idx++]) != wfm_hdr->swtb) &&
drivers/video/fbdev/metronomefb.c
289
rl = mem[wfm_idx++];
drivers/video/fbdev/metronomefb.c
297
cksum = calc_cksum(owfm_idx, cksum_idx, mem);
drivers/video/fbdev/metronomefb.c
298
if (cksum != mem[cksum_idx]) {
drivers/video/fbdev/metronomefb.c
300
" %x != %x\n", cksum, mem[cksum_idx]);
drivers/video/fbdev/omap2/omapfb/vrfb.c
339
struct resource *mem;
drivers/video/fbdev/omap2/omapfb/vrfb.c
357
mem = platform_get_resource(pdev, IORESOURCE_MEM, 1 + i);
drivers/video/fbdev/omap2/omapfb/vrfb.c
358
if (!mem) {
drivers/video/fbdev/omap2/omapfb/vrfb.c
364
ctxs[i].base = mem->start;
drivers/video/fbdev/s3fb.c
553
int rv, mem, step;
drivers/video/fbdev/s3fb.c
584
mem = ((var->bits_per_pixel * var->xres_virtual) >> 3) * var->yres_virtual;
drivers/video/fbdev/s3fb.c
585
if (mem > info->screen_size) {
drivers/video/fbdev/s3fb.c
587
mem >> 10, (unsigned int) (info->screen_size >> 10));
drivers/video/fbdev/simplefb.c
105
struct resource *mem = par->mem;
drivers/video/fbdev/simplefb.c
115
if (mem)
drivers/video/fbdev/simplefb.c
116
release_mem_region(mem->start, resource_size(mem));
drivers/video/fbdev/simplefb.c
538
struct resource *res, *mem;
drivers/video/fbdev/simplefb.c
562
mem = request_mem_region(res->start, resource_size(res), "simplefb");
drivers/video/fbdev/simplefb.c
563
if (!mem) {
drivers/video/fbdev/simplefb.c
570
mem = res;
drivers/video/fbdev/simplefb.c
583
info->fix.smem_start = mem->start;
drivers/video/fbdev/simplefb.c
584
info->fix.smem_len = resource_size(mem);
drivers/video/fbdev/simplefb.c
632
if (mem != res)
drivers/video/fbdev/simplefb.c
633
par->mem = mem; /* release in clean-up handler */
drivers/video/fbdev/simplefb.c
661
if (mem != res)
drivers/video/fbdev/simplefb.c
662
release_mem_region(mem->start, resource_size(mem));
drivers/video/fbdev/simplefb.c
76
struct resource *mem;
drivers/video/fbdev/sis/sis_main.c
6567
static unsigned int mem = 0;
drivers/video/fbdev/sis/sis_main.c
6634
if(mem)
drivers/video/fbdev/sis/sis_main.c
6635
sisfb_parm_mem = mem;
drivers/video/fbdev/sis/sis_main.c
6682
module_param(mem, int, 0);
drivers/video/fbdev/sis/sis_main.c
6707
MODULE_PARM_DESC(mem,
drivers/video/fbdev/sm501fb.c
1233
void __iomem *mem = info->regs;
drivers/video/fbdev/sm501fb.c
1239
smc501_readl(mem + reg));
drivers/video/fbdev/sm501fb.c
168
static int sm501_alloc_mem(struct sm501fb_info *inf, struct sm501_mem *mem,
drivers/video/fbdev/sm501fb.c
239
mem->size = size;
drivers/video/fbdev/sm501fb.c
240
mem->sm_addr = ptr;
drivers/video/fbdev/sm501fb.c
241
mem->k_addr = inf->fbmem + ptr;
drivers/video/fbdev/sm501fb.c
244
__func__, mem->sm_addr, mem->k_addr, why, size);
drivers/video/fbdev/sstfb.c
1303
mem = simple_strtoul (this_opt+4, NULL, 0);
drivers/video/fbdev/sstfb.c
1531
module_param(mem, int, 0);
drivers/video/fbdev/sstfb.c
1532
MODULE_PARM_DESC(mem, "Size of frame buffer memory in MB (1, 2, 4 MB, default=autodetect)");
drivers/video/fbdev/sstfb.c
834
if (mem >= 1 && mem <= 4) {
drivers/video/fbdev/sstfb.c
835
*memsize = (mem * 0x100000);
drivers/video/fbdev/sstfb.c
99
static int mem; /* mem size in MB, 0 = autodetect */
drivers/video/fbdev/udlfb.c
1165
static void dlfb_deferred_vfree(struct dlfb_data *dlfb, void *mem)
drivers/video/fbdev/udlfb.c
1170
d->mem = mem;
drivers/video/fbdev/udlfb.c
76
void *mem;
drivers/video/fbdev/udlfb.c
958
vfree(d->mem);
drivers/video/fbdev/vt8623fb.c
323
int rv, mem, step;
drivers/video/fbdev/vt8623fb.c
348
mem = ((var->bits_per_pixel * var->xres_virtual) >> 3) * var->yres_virtual;
drivers/video/fbdev/vt8623fb.c
349
if (mem > info->screen_size)
drivers/video/fbdev/vt8623fb.c
352
mem >> 10, (unsigned int) (info->screen_size >> 10));
drivers/video/fbdev/vt8623fb.c
357
if ((var->bits_per_pixel == 0) && (mem > (256*1024)))
drivers/video/fbdev/vt8623fb.c
360
mem >> 10);
drivers/watchdog/menz69_wdt.c
105
struct resource *mem;
drivers/watchdog/menz69_wdt.c
111
mem = mcb_request_mem(dev, "z069-wdt");
drivers/watchdog/menz69_wdt.c
112
if (IS_ERR(mem))
drivers/watchdog/menz69_wdt.c
113
return PTR_ERR(mem);
drivers/watchdog/menz69_wdt.c
115
drv->base = devm_ioremap(&dev->dev, mem->start, resource_size(mem));
drivers/watchdog/menz69_wdt.c
119
drv->mem = mem;
drivers/watchdog/menz69_wdt.c
135
mcb_release_mem(mem);
drivers/watchdog/menz69_wdt.c
144
mcb_release_mem(drv->mem);
drivers/watchdog/menz69_wdt.c
16
struct resource *mem;
drivers/xen/efi.c
313
.u.efi_info.mem.addr = phys_addr,
drivers/xen/efi.c
314
.u.efi_info.mem.size = U64_MAX - phys_addr,
drivers/xen/efi.c
324
out_md->phys_addr = info->mem.addr;
drivers/xen/efi.c
325
out_md->num_pages = info->mem.size >> EFI_PAGE_SHIFT;
drivers/xen/efi.c
326
out_md->type = info->mem.type;
drivers/xen/efi.c
327
out_md->attribute = info->mem.attr;
fs/btrfs/lzo.c
327
workspace->mem);
fs/btrfs/lzo.c
62
void *mem;
fs/btrfs/lzo.c
83
kvfree(workspace->mem);
fs/btrfs/lzo.c
95
workspace->mem = kvmalloc(LZO1X_MEM_COMPRESS, GFP_KERNEL | __GFP_NOWARN);
fs/btrfs/lzo.c
98
if (!workspace->mem || !workspace->buf || !workspace->cbuf)
fs/btrfs/zstd.c
366
kvfree(workspace->mem);
fs/btrfs/zstd.c
385
workspace->mem = kvmalloc(workspace->size, GFP_KERNEL | __GFP_NOWARN);
fs/btrfs/zstd.c
387
if (!workspace->mem || !workspace->buf)
fs/btrfs/zstd.c
423
stream = zstd_init_cstream(&workspace->params, len, workspace->mem, workspace->size);
fs/btrfs/zstd.c
46
void *mem;
fs/btrfs/zstd.c
606
ZSTD_BTRFS_MAX_INPUT, workspace->mem, workspace->size);
fs/btrfs/zstd.c
691
ZSTD_BTRFS_MAX_INPUT, workspace->mem, workspace->size);
fs/smb/client/smbdirect.c
1807
response = mempool_alloc(sc->recv_io.mem.pool, GFP_KERNEL);
fs/smb/client/smbdirect.c
1825
mempool_free(response, sc->recv_io.mem.pool);
fs/smb/client/smbdirect.c
1835
mempool_free(response, sc->recv_io.mem.pool);
fs/smb/client/smbdirect.c
1959
mempool_destroy(sc->send_io.mem.pool);
fs/smb/client/smbdirect.c
1960
kmem_cache_destroy(sc->send_io.mem.cache);
fs/smb/client/smbdirect.c
1962
mempool_destroy(sc->recv_io.mem.pool);
fs/smb/client/smbdirect.c
1963
kmem_cache_destroy(sc->recv_io.mem.cache);
fs/smb/client/smbdirect.c
2012
mempool_destroy(sc->recv_io.mem.pool);
fs/smb/client/smbdirect.c
2013
kmem_cache_destroy(sc->recv_io.mem.cache);
fs/smb/client/smbdirect.c
2014
mempool_destroy(sc->send_io.mem.pool);
fs/smb/client/smbdirect.c
2015
kmem_cache_destroy(sc->send_io.mem.cache);
fs/smb/client/smbdirect.c
2029
sc->send_io.mem.cache =
fs/smb/client/smbdirect.c
2035
if (!sc->send_io.mem.cache)
fs/smb/client/smbdirect.c
2038
sc->send_io.mem.pool =
fs/smb/client/smbdirect.c
2040
mempool_free_slab, sc->send_io.mem.cache);
fs/smb/client/smbdirect.c
2041
if (!sc->send_io.mem.pool)
fs/smb/client/smbdirect.c
2052
sc->recv_io.mem.cache =
fs/smb/client/smbdirect.c
2056
if (!sc->recv_io.mem.cache)
fs/smb/client/smbdirect.c
2059
sc->recv_io.mem.pool =
fs/smb/client/smbdirect.c
2061
mempool_free_slab, sc->recv_io.mem.cache);
fs/smb/client/smbdirect.c
2062
if (!sc->recv_io.mem.pool)
fs/smb/client/smbdirect.c
2074
mempool_destroy(sc->recv_io.mem.pool);
fs/smb/client/smbdirect.c
2076
kmem_cache_destroy(sc->recv_io.mem.cache);
fs/smb/client/smbdirect.c
2078
mempool_destroy(sc->send_io.mem.pool);
fs/smb/client/smbdirect.c
2080
kmem_cache_destroy(sc->send_io.mem.cache);
fs/smb/client/smbdirect.c
519
msg = mempool_alloc(sc->send_io.mem.pool, GFP_KERNEL);
fs/smb/client/smbdirect.c
557
mempool_free(msg, sc->send_io.mem.pool);
fs/smb/common/smbdirect/smbdirect_socket.h
163
} mem;
fs/smb/common/smbdirect/smbdirect_socket.h
228
} mem;
fs/smb/server/transport_rdma.c
2238
mempool_free(recvmsg, sc->recv_io.mem.pool);
fs/smb/server/transport_rdma.c
2240
mempool_destroy(sc->recv_io.mem.pool);
fs/smb/server/transport_rdma.c
2241
sc->recv_io.mem.pool = NULL;
fs/smb/server/transport_rdma.c
2243
kmem_cache_destroy(sc->recv_io.mem.cache);
fs/smb/server/transport_rdma.c
2244
sc->recv_io.mem.cache = NULL;
fs/smb/server/transport_rdma.c
2246
mempool_destroy(sc->send_io.mem.pool);
fs/smb/server/transport_rdma.c
2247
sc->send_io.mem.pool = NULL;
fs/smb/server/transport_rdma.c
2249
kmem_cache_destroy(sc->send_io.mem.cache);
fs/smb/server/transport_rdma.c
2250
sc->send_io.mem.cache = NULL;
fs/smb/server/transport_rdma.c
2261
sc->send_io.mem.cache = kmem_cache_create(name,
fs/smb/server/transport_rdma.c
2265
if (!sc->send_io.mem.cache)
fs/smb/server/transport_rdma.c
2268
sc->send_io.mem.pool = mempool_create(sp->send_credit_target,
fs/smb/server/transport_rdma.c
2270
sc->send_io.mem.cache);
fs/smb/server/transport_rdma.c
2271
if (!sc->send_io.mem.pool)
fs/smb/server/transport_rdma.c
2275
sc->recv_io.mem.cache = kmem_cache_create(name,
fs/smb/server/transport_rdma.c
2279
if (!sc->recv_io.mem.cache)
fs/smb/server/transport_rdma.c
2282
sc->recv_io.mem.pool =
fs/smb/server/transport_rdma.c
2284
mempool_free_slab, sc->recv_io.mem.cache);
fs/smb/server/transport_rdma.c
2285
if (!sc->recv_io.mem.pool)
fs/smb/server/transport_rdma.c
2289
recvmsg = mempool_alloc(sc->recv_io.mem.pool, KSMBD_DEFAULT_GFP);
fs/smb/server/transport_rdma.c
536
msg = mempool_alloc(sc->send_io.mem.pool, KSMBD_DEFAULT_GFP);
fs/smb/server/transport_rdma.c
565
mempool_free(msg, sc->send_io.mem.pool);
fs/squashfs/zstd_wrapper.c
24
void *mem;
fs/squashfs/zstd_wrapper.c
38
wksp->mem = vmalloc(wksp->mem_size);
fs/squashfs/zstd_wrapper.c
39
if (wksp->mem == NULL)
fs/squashfs/zstd_wrapper.c
56
vfree(wksp->mem);
fs/squashfs/zstd_wrapper.c
74
stream = zstd_init_dstream(wksp->window_size, wksp->mem, wksp->mem_size);
fs/userfaultfd.c
2120
static void init_once_userfaultfd_ctx(void *mem)
fs/userfaultfd.c
2122
struct userfaultfd_ctx *ctx = (struct userfaultfd_ctx *) mem;
include/acpi/acrestyp.h
256
struct acpi_memory_attribute mem;
include/drm/drm_format_helper.h
30
void *mem;
include/drm/drm_format_helper.h
38
.mem = (_mem), \
include/drm/ttm/ttm_bo.h
440
struct ttm_resource **mem,
include/drm/ttm/ttm_bo.h
448
struct ttm_resource *mem);
include/drm/ttm/ttm_bo.h
450
struct ttm_resource *mem);
include/drm/ttm/ttm_device.h
172
struct ttm_resource *mem);
include/drm/ttm/ttm_device.h
174
struct ttm_resource *mem);
include/drm/ttm/ttm_resource.h
515
struct ttm_resource *mem);
include/drm/ttm/ttm_resource.h
519
struct ttm_resource *mem);
include/linux/clk/ti.h
283
struct regmap *syscon, void __iomem *mem);
include/linux/clk/ti.h
284
void omap2_clk_legacy_provider_init(int index, void __iomem *mem);
include/linux/cper.h
602
int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg);
include/linux/cper.h
603
int cper_dimm_err_location(struct cper_mem_err_compact *mem, char *msg);
include/linux/crash_core.h
64
extern int crash_exclude_mem_range(struct crash_mem *mem,
include/linux/crash_core.h
67
extern int crash_prepare_elf64_headers(struct crash_mem *mem, int need_kernel_map,
include/linux/crypto.h
432
void crypto_destroy_tfm(void *mem, struct crypto_tfm *tfm);
include/linux/dm-io.h
63
struct dm_io_memory mem; /* Memory to use for io */
include/linux/dma/edma.h
23
void *mem;
include/linux/ehl_pse_io_aux.h
20
struct resource mem;
include/linux/firmware/cirrus/cs_dsp.h
175
const struct cs_dsp_region *mem;
include/linux/kbuild.h
10
#define OFFSET(sym, str, mem) \
include/linux/kbuild.h
11
DEFINE(sym, offsetof(struct str, mem))
include/linux/kexec.h
102
unsigned long mem;
include/linux/kexec.h
110
compat_ulong_t mem; /* User space sees this as a (void *) ... */
include/linux/kexec.h
189
unsigned long mem;
include/linux/kexec_handover.h
30
void kho_unpreserve_free(void *mem);
include/linux/kexec_handover.h
31
void kho_restore_free(void *mem);
include/linux/kexec_handover.h
81
static inline void kho_unpreserve_free(void *mem) { }
include/linux/kexec_handover.h
82
static inline void kho_restore_free(void *mem) { }
include/linux/kgdb.h
309
extern char *kgdb_mem2hex(char *mem, char *buf, int count);
include/linux/kgdb.h
310
extern int kgdb_hex2mem(char *buf, char *mem, int count);
include/linux/kgdb.h
92
extern char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs);
include/linux/kgdb.h
93
extern int dbg_set_reg(int regno, void *mem, struct pt_regs *regs);
include/linux/kvm_host.h
1234
const struct kvm_userspace_memory_region2 *mem);
include/linux/mcb.h
127
extern void mcb_release_mem(struct resource *mem);
include/linux/mcb.h
74
struct resource mem;
include/linux/memory.h
198
void memory_block_add_nid_early(struct memory_block *mem, int nid);
include/linux/mlx5/qp.h
460
} mem;
include/linux/module.h
455
struct module_memory mem[MOD_MEM_NUM_TYPES] __module_memory_align;
include/linux/module.h
633
base = (unsigned long)mod->mem[type].base;
include/linux/module.h
634
size = mod->mem[type].size;
include/linux/netfilter/x_tables.h
415
const char __percpu *mem;
include/linux/pci-epc.h
175
struct pci_epc_mem *mem;
include/linux/pci-p2pdma.h
172
struct p2pdma_provider *mem;
include/linux/platform_data/x86/intel_scu_ipc.h
20
struct resource mem;
include/linux/platform_data/x86/p2sb.h
17
int p2sb_bar(struct pci_bus *bus, unsigned int devfn, struct resource *mem);
include/linux/platform_data/x86/p2sb.h
21
static inline int p2sb_bar(struct pci_bus *bus, unsigned int devfn, struct resource *mem)
include/linux/platform_data/xtalk-bridge.h
12
struct resource mem;
include/linux/plist.h
172
#define plist_for_each_entry(pos, head, mem) \
include/linux/plist.h
173
list_for_each_entry(pos, &(head)->node_list, mem.node_list)
include/linux/remoteproc.h
344
int (*alloc)(struct rproc *rproc, struct rproc_mem_entry *mem);
include/linux/remoteproc.h
345
int (*release)(struct rproc *rproc, struct rproc_mem_entry *mem);
include/linux/remoteproc.h
674
void rproc_add_carveout(struct rproc *rproc, struct rproc_mem_entry *mem);
include/linux/remoteproc/st_slim_rproc.h
42
struct st_slim_mem mem[ST_SLIM_MEM_MAX];
include/linux/seq_buf.h
178
extern int seq_buf_putmem(struct seq_buf *s, const void *mem, unsigned int len);
include/linux/seq_buf.h
179
extern int seq_buf_putmem_hex(struct seq_buf *s, const void *mem,
include/linux/spi/spi-mem.h
251
struct spi_mem *mem;
include/linux/spi/spi-mem.h
281
static inline void spi_mem_set_drvdata(struct spi_mem *mem, void *data)
include/linux/spi/spi-mem.h
283
mem->drvpriv = data;
include/linux/spi/spi-mem.h
293
static inline void *spi_mem_get_drvdata(struct spi_mem *mem)
include/linux/spi/spi-mem.h
295
return mem->drvpriv;
include/linux/spi/spi-mem.h
345
int (*adjust_op_size)(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
346
bool (*supports_op)(struct spi_mem *mem,
include/linux/spi/spi-mem.h
348
int (*exec_op)(struct spi_mem *mem,
include/linux/spi/spi-mem.h
350
const char *(*get_name)(struct spi_mem *mem);
include/linux/spi/spi-mem.h
357
int (*poll_status)(struct spi_mem *mem,
include/linux/spi/spi-mem.h
401
int (*probe)(struct spi_mem *mem);
include/linux/spi/spi-mem.h
402
int (*remove)(struct spi_mem *mem);
include/linux/spi/spi-mem.h
403
void (*shutdown)(struct spi_mem *mem);
include/linux/spi/spi-mem.h
415
bool spi_mem_default_supports_op(struct spi_mem *mem,
include/linux/spi/spi-mem.h
434
bool spi_mem_default_supports_op(struct spi_mem *mem,
include/linux/spi/spi-mem.h
441
int spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
442
void spi_mem_adjust_op_freq(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
443
u64 spi_mem_calc_op_duration(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
445
bool spi_mem_supports_op(struct spi_mem *mem,
include/linux/spi/spi-mem.h
448
int spi_mem_exec_op(struct spi_mem *mem,
include/linux/spi/spi-mem.h
451
const char *spi_mem_get_name(struct spi_mem *mem);
include/linux/spi/spi-mem.h
454
spi_mem_dirmap_create(struct spi_mem *mem,
include/linux/spi/spi-mem.h
462
devm_spi_mem_dirmap_create(struct device *dev, struct spi_mem *mem,
include/linux/spi/spi-mem.h
467
int spi_mem_poll_status(struct spi_mem *mem,
include/linux/swap.h
388
extern unsigned long mem_cgroup_shrink_node(struct mem_cgroup *mem,
include/linux/swap.h
575
static inline int mem_cgroup_swappiness(struct mem_cgroup *mem)
include/linux/swiotlb.h
145
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
include/linux/swiotlb.h
147
if (!mem)
include/linux/swiotlb.h
165
if (paddr >= mem->defpool.start && paddr < mem->defpool.end)
include/linux/swiotlb.h
166
return &mem->defpool;
include/linux/swiotlb.h
174
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
include/linux/swiotlb.h
176
return mem && mem->force_bounce;
include/linux/trace_seq.h
111
extern void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len);
include/linux/trace_seq.h
112
extern void trace_seq_putmem_hex(struct trace_seq *s, const void *mem,
include/linux/trace_seq.h
166
trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len)
include/linux/trace_seq.h
169
static inline void trace_seq_putmem_hex(struct trace_seq *s, const void *mem,
include/linux/uio_driver.h
109
struct uio_mem mem[MAX_UIO_MAPS];
include/linux/vmalloc.h
146
extern void vm_unmap_ram(const void *mem, unsigned int count);
include/media/drv-intf/saa7146.h
128
unsigned char __iomem *mem; /* pointer to mapped IO memory */
include/media/drv-intf/saa7146.h
175
void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt);
include/media/drv-intf/saa7146.h
22
#define saa7146_write(sxy,adr,dat) writel((dat),(sxy->mem+(adr)))
include/media/drv-intf/saa7146.h
23
#define saa7146_read(sxy,adr) readl(sxy->mem+(adr))
include/media/vsp1.h
107
dma_addr_t mem[3];
include/media/vsp1.h
177
dma_addr_t mem;
include/media/vsp1.h
181
dma_addr_t mem;
include/media/vsp1.h
66
dma_addr_t mem[3];
include/net/inet_frag.h
146
return atomic_long_read(&fqdir->mem);
include/net/inet_frag.h
151
atomic_long_sub(val, &fqdir->mem);
include/net/inet_frag.h
156
atomic_long_add(val, &fqdir->mem);
include/net/inet_frag.h
26
atomic_long_t mem ____cacheline_aligned_in_smp;
include/net/page_pool/types.h
274
const struct xdp_mem_info *mem);
include/net/page_pool/types.h
283
const struct xdp_mem_info *mem)
include/net/sock.h
1900
void sock_kfree_s(struct sock *sk, void *mem, int size);
include/net/sock.h
1901
void sock_kzfree_s(struct sock *sk, void *mem, int size);
include/net/xdp.h
434
if (xdp->rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL)
include/net/xdp.h
443
xdp_frame->mem_type = xdp->rxq->mem.type;
include/net/xdp.h
497
int xdp_reg_mem_model(struct xdp_mem_info *mem,
include/net/xdp.h
499
void xdp_unreg_mem_model(struct xdp_mem_info *mem);
include/net/xdp.h
515
const struct xdp_mem_info *mem)
include/net/xdp.h
517
xdp_rxq->mem = *mem;
include/net/xdp.h
530
xdp_rxq->mem = (struct xdp_mem_info){ };
include/net/xdp.h
66
struct xdp_mem_info mem;
include/net/xdp_priv.h
10
struct xdp_mem_info mem;
include/pcmcia/cistpl.h
463
cistpl_mem_t mem;
include/pcmcia/cistpl.h
483
u_char mem;
include/pcmcia/ss.h
127
int (*set_mem_map)(struct pcmcia_socket *s, struct pccard_mem_map *mem);
include/ras/ras_event.h
27
TP_PROTO(struct cper_sec_mem_err *mem,
include/ras/ras_event.h
33
TP_ARGS(mem, err_seq, fru_id, fru_text, sev),
include/ras/ras_event.h
48
if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE)
include/ras/ras_event.h
49
__entry->etype = mem->error_type;
include/ras/ras_event.h
53
if (mem->validation_bits & CPER_MEM_VALID_PA)
include/ras/ras_event.h
54
__entry->pa = mem->physical_addr;
include/ras/ras_event.h
58
if (mem->validation_bits & CPER_MEM_VALID_PA_MASK)
include/ras/ras_event.h
59
__entry->pa_mask_lsb = (u8)__ffs64(mem->physical_addr_mask);
include/ras/ras_event.h
64
cper_mem_err_pack(mem, &__entry->data);
include/rdma/signature.h
89
struct ib_sig_domain mem;
include/sound/emu10k1.h
1549
struct snd_util_memblk mem;
include/sound/emu10k1.h
1557
#define snd_emu10k1_memblk_offset(blk) (((blk)->mapped_page << PAGE_SHIFT) | ((blk)->mem.offset & (PAGE_SIZE - 1)))
include/trace/events/spi-mem.h
20
TP_PROTO(struct spi_mem *mem, const struct spi_mem_op *op),
include/trace/events/spi-mem.h
21
TP_ARGS(mem, op),
include/trace/events/spi-mem.h
24
__string(name, mem->name)
include/trace/events/spi-mem.h
3
#define TRACE_SYSTEM spi-mem
include/trace/events/spi-mem.h
43
__entry->max_freq = op->max_freq ?: mem->spi->max_speed_hz;
include/trace/events/spi-mem.h
78
TP_PROTO(struct spi_mem *mem, const struct spi_mem_op *op),
include/trace/events/spi-mem.h
79
TP_ARGS(mem, op),
include/trace/events/spi-mem.h
82
__string(name, mem->name)
include/trace/events/xdp.h
319
__entry->mem_id = xa->mem.id;
include/trace/events/xdp.h
320
__entry->mem_type = xa->mem.type;
include/trace/events/xdp.h
349
__entry->mem_id = xa->mem.id;
include/trace/events/xdp.h
350
__entry->mem_type = xa->mem.type;
include/uapi/drm/amdxdna_accel.h
329
struct amdxdna_drm_query_aie_tile_metadata mem;
include/uapi/linux/kexec.h
67
const void *mem;
include/xen/interface/platform.h
258
} mem;
io_uring/zcrx.c
101
static void io_release_dmabuf(struct io_zcrx_mem *mem)
io_uring/zcrx.c
106
if (mem->sgt)
io_uring/zcrx.c
107
dma_buf_unmap_attachment_unlocked(mem->attach, mem->sgt,
io_uring/zcrx.c
109
if (mem->attach)
io_uring/zcrx.c
110
dma_buf_detach(mem->dmabuf, mem->attach);
io_uring/zcrx.c
111
if (mem->dmabuf)
io_uring/zcrx.c
112
dma_buf_put(mem->dmabuf);
io_uring/zcrx.c
114
mem->sgt = NULL;
io_uring/zcrx.c
115
mem->attach = NULL;
io_uring/zcrx.c
116
mem->dmabuf = NULL;
io_uring/zcrx.c
120
struct io_zcrx_mem *mem,
io_uring/zcrx.c
1224
if (area->mem.is_dmabuf)
io_uring/zcrx.c
137
mem->is_dmabuf = true;
io_uring/zcrx.c
138
mem->dmabuf = dma_buf_get(dmabuf_fd);
io_uring/zcrx.c
139
if (IS_ERR(mem->dmabuf)) {
io_uring/zcrx.c
140
ret = PTR_ERR(mem->dmabuf);
io_uring/zcrx.c
141
mem->dmabuf = NULL;
io_uring/zcrx.c
145
mem->attach = dma_buf_attach(mem->dmabuf, ifq->dev);
io_uring/zcrx.c
146
if (IS_ERR(mem->attach)) {
io_uring/zcrx.c
147
ret = PTR_ERR(mem->attach);
io_uring/zcrx.c
148
mem->attach = NULL;
io_uring/zcrx.c
152
mem->sgt = dma_buf_map_attachment_unlocked(mem->attach, DMA_FROM_DEVICE);
io_uring/zcrx.c
153
if (IS_ERR(mem->sgt)) {
io_uring/zcrx.c
154
ret = PTR_ERR(mem->sgt);
io_uring/zcrx.c
155
mem->sgt = NULL;
io_uring/zcrx.c
159
for_each_sgtable_dma_sg(mem->sgt, sg, i)
io_uring/zcrx.c
167
mem->size = len;
io_uring/zcrx.c
170
io_release_dmabuf(mem);
io_uring/zcrx.c
192
struct io_zcrx_mem *mem,
io_uring/zcrx.c
207
ret = sg_alloc_table_from_pages(&mem->page_sg_table, pages, nr_pages,
io_uring/zcrx.c
216
mem->account_pages = io_count_account_pages(pages, nr_pages);
io_uring/zcrx.c
217
ret = io_account_mem(ifq->user, ifq->mm_account, mem->account_pages);
io_uring/zcrx.c
219
mem->account_pages = 0;
io_uring/zcrx.c
221
mem->sgt = &mem->page_sg_table;
io_uring/zcrx.c
222
mem->pages = pages;
io_uring/zcrx.c
223
mem->nr_folios = nr_pages;
io_uring/zcrx.c
224
mem->size = area_reg->len;
io_uring/zcrx.c
228
static void io_release_area_mem(struct io_zcrx_mem *mem)
io_uring/zcrx.c
230
if (mem->is_dmabuf) {
io_uring/zcrx.c
231
io_release_dmabuf(mem);
io_uring/zcrx.c
234
if (mem->pages) {
io_uring/zcrx.c
235
unpin_user_pages(mem->pages, mem->nr_folios);
io_uring/zcrx.c
236
sg_free_table(mem->sgt);
io_uring/zcrx.c
237
mem->sgt = NULL;
io_uring/zcrx.c
238
kvfree(mem->pages);
io_uring/zcrx.c
243
struct io_zcrx_mem *mem,
io_uring/zcrx.c
262
return io_import_dmabuf(ifq, mem, area_reg);
io_uring/zcrx.c
263
return io_import_umem(ifq, mem, area_reg);
io_uring/zcrx.c
279
if (area->mem.is_dmabuf) {
io_uring/zcrx.c
280
io_release_dmabuf(&area->mem);
io_uring/zcrx.c
282
dma_unmap_sgtable(ifq->dev, &area->mem.page_sg_table,
io_uring/zcrx.c
295
if (!area->mem.is_dmabuf) {
io_uring/zcrx.c
296
ret = dma_map_sgtable(ifq->dev, &area->mem.page_sg_table,
io_uring/zcrx.c
303
if (ret && !area->mem.is_dmabuf)
io_uring/zcrx.c
304
dma_unmap_sgtable(ifq->dev, &area->mem.page_sg_table,
io_uring/zcrx.c
417
io_release_area_mem(&area->mem);
io_uring/zcrx.c
419
if (area->mem.account_pages)
io_uring/zcrx.c
421
area->mem.account_pages);
io_uring/zcrx.c
460
ret = io_import_area(ifq, &area->mem, area_reg);
io_uring/zcrx.c
464
if (buf_size_shift > io_area_max_shift(&area->mem)) {
io_uring/zcrx.c
470
nr_iovs = area->mem.size >> ifq->niov_shift;
io_uring/zcrx.c
52
lockdep_assert(!area->mem.is_dmabuf);
io_uring/zcrx.c
55
return area->mem.pages[net_iov_idx(niov) << niov_pages_shift];
io_uring/zcrx.c
58
static int io_area_max_shift(struct io_zcrx_mem *mem)
io_uring/zcrx.c
60
struct sg_table *sgt = mem->sgt;
io_uring/zcrx.c
74
struct sg_table *sgt = area->mem.sgt;
io_uring/zcrx.h
38
struct io_zcrx_mem mem;
kernel/bpf/cpumap.c
199
rxq.mem.type = xdpf->mem_type;
kernel/crash_core.c
176
int crash_prepare_elf64_headers(struct crash_mem *mem, int need_kernel_map,
kernel/crash_core.c
189
nr_phdr += mem->nr_ranges;
kernel/crash_core.c
251
for (i = 0; i < mem->nr_ranges; i++) {
kernel/crash_core.c
252
mstart = mem->ranges[i].start;
kernel/crash_core.c
253
mend = mem->ranges[i].end;
kernel/crash_core.c
291
int crash_exclude_mem_range(struct crash_mem *mem,
kernel/crash_core.c
297
for (i = 0; i < mem->nr_ranges; i++) {
kernel/crash_core.c
298
start = mem->ranges[i].start;
kernel/crash_core.c
299
end = mem->ranges[i].end;
kernel/crash_core.c
324
memmove(&mem->ranges[i], &mem->ranges[i + 1],
kernel/crash_core.c
325
(mem->nr_ranges - (i + 1)) * sizeof(mem->ranges[i]));
kernel/crash_core.c
327
mem->nr_ranges--;
kernel/crash_core.c
330
if (mem->nr_ranges >= mem->max_nr_ranges)
kernel/crash_core.c
333
memmove(&mem->ranges[i + 2], &mem->ranges[i + 1],
kernel/crash_core.c
334
(mem->nr_ranges - (i + 1)) * sizeof(mem->ranges[i]));
kernel/crash_core.c
336
mem->ranges[i].end = p_start - 1;
kernel/crash_core.c
337
mem->ranges[i + 1].start = p_end + 1;
kernel/crash_core.c
338
mem->ranges[i + 1].end = end;
kernel/crash_core.c
341
mem->nr_ranges++;
kernel/crash_core.c
343
mem->ranges[i].end = p_start - 1;
kernel/crash_core.c
345
mem->ranges[i].start = p_end + 1;
kernel/crash_core.c
609
unsigned long mem;
kernel/crash_core.c
614
mem = image->segment[n].mem;
kernel/crash_core.c
615
ptr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
kernel/crash_core_test.c
10
struct crash_mem *mem;
kernel/crash_core_test.c
21
mem = kunit_kzalloc(test, alloc_size, GFP_KERNEL);
kernel/crash_core_test.c
22
if (!mem) {
kernel/crash_core_test.c
27
mem->max_nr_ranges = max_ranges;
kernel/crash_core_test.c
28
mem->nr_ranges = nr_initial_ranges;
kernel/crash_core_test.c
30
memcpy(mem->ranges, initial_ranges,
kernel/crash_core_test.c
34
return mem;
kernel/crash_core_test.c
73
struct crash_mem *mem;
kernel/crash_core_test.c
78
mem = create_crash_mem(test, params->initial_max_ranges,
kernel/crash_core_test.c
80
if (!mem)
kernel/crash_core_test.c
83
ret = crash_exclude_mem_range(mem, params->exclude_start, params->exclude_end);
kernel/crash_core_test.c
89
assert_ranges_equal(test, mem->ranges, mem->nr_ranges,
kernel/crash_core_test.c
97
mem->nr_ranges,
kernel/crash_dump_dm_crypt.c
432
kbuf.mem = KEXEC_BUF_MEM_UNKNOWN;
kernel/crash_dump_dm_crypt.c
438
image->dm_crypt_keys_addr = kbuf.mem;
kernel/debug/gdbstub.c
238
char *kgdb_mem2hex(char *mem, char *buf, int count)
kernel/debug/gdbstub.c
249
err = copy_from_kernel_nofault(tmp, mem, count);
kernel/debug/gdbstub.c
267
int kgdb_hex2mem(char *buf, char *mem, int count)
kernel/debug/gdbstub.c
285
return copy_to_kernel_nofault(mem, tmp_raw, count);
kernel/debug/gdbstub.c
325
static int kgdb_ebin2mem(char *buf, char *mem, int count)
kernel/debug/gdbstub.c
337
return copy_to_kernel_nofault(mem, c, size);
kernel/dma/coherent.c
120
struct dma_coherent_mem *mem;
kernel/dma/coherent.c
123
mem = dma_init_coherent_memory(phys_addr, device_addr, size, false);
kernel/dma/coherent.c
124
if (IS_ERR(mem))
kernel/dma/coherent.c
125
return PTR_ERR(mem);
kernel/dma/coherent.c
127
ret = dma_assign_coherent_memory(dev, mem);
kernel/dma/coherent.c
129
_dma_release_coherent_memory(mem);
kernel/dma/coherent.c
142
struct dma_coherent_mem *mem,
kernel/dma/coherent.c
150
spin_lock_irqsave(&mem->spinlock, flags);
kernel/dma/coherent.c
152
if (unlikely(size > ((dma_addr_t)mem->size << PAGE_SHIFT)))
kernel/dma/coherent.c
155
pageno = bitmap_find_free_region(mem->bitmap, mem->size, order);
kernel/dma/coherent.c
162
*dma_handle = dma_get_device_base(dev, mem) +
kernel/dma/coherent.c
164
ret = mem->virt_base + ((dma_addr_t)pageno << PAGE_SHIFT);
kernel/dma/coherent.c
165
spin_unlock_irqrestore(&mem->spinlock, flags);
kernel/dma/coherent.c
169
spin_unlock_irqrestore(&mem->spinlock, flags);
kernel/dma/coherent.c
190
struct dma_coherent_mem *mem = dev_get_coherent_memory(dev);
kernel/dma/coherent.c
192
if (!mem)
kernel/dma/coherent.c
195
*ret = __dma_alloc_from_coherent(dev, mem, size, dma_handle);
kernel/dma/coherent.c
199
static int __dma_release_from_coherent(struct dma_coherent_mem *mem,
kernel/dma/coherent.c
202
if (mem && vaddr >= mem->virt_base && vaddr <
kernel/dma/coherent.c
203
(mem->virt_base + ((dma_addr_t)mem->size << PAGE_SHIFT))) {
kernel/dma/coherent.c
204
int page = (vaddr - mem->virt_base) >> PAGE_SHIFT;
kernel/dma/coherent.c
207
spin_lock_irqsave(&mem->spinlock, flags);
kernel/dma/coherent.c
208
bitmap_release_region(mem->bitmap, page, order);
kernel/dma/coherent.c
209
spin_unlock_irqrestore(&mem->spinlock, flags);
kernel/dma/coherent.c
229
struct dma_coherent_mem *mem = dev_get_coherent_memory(dev);
kernel/dma/coherent.c
231
return __dma_release_from_coherent(mem, order, vaddr);
kernel/dma/coherent.c
234
static int __dma_mmap_from_coherent(struct dma_coherent_mem *mem,
kernel/dma/coherent.c
237
if (mem && vaddr >= mem->virt_base && vaddr + size <=
kernel/dma/coherent.c
238
(mem->virt_base + ((dma_addr_t)mem->size << PAGE_SHIFT))) {
kernel/dma/coherent.c
240
int start = (vaddr - mem->virt_base) >> PAGE_SHIFT;
kernel/dma/coherent.c
246
unsigned long pfn = mem->pfn_base + start + off;
kernel/dma/coherent.c
274
struct dma_coherent_mem *mem = dev_get_coherent_memory(dev);
kernel/dma/coherent.c
276
return __dma_mmap_from_coherent(mem, vma, vaddr, size, ret);
kernel/dma/coherent.c
31
struct dma_coherent_mem * mem)
kernel/dma/coherent.c
313
struct dma_coherent_mem *mem;
kernel/dma/coherent.c
315
mem = dma_init_coherent_memory(phys_addr, phys_addr, size, true);
kernel/dma/coherent.c
316
if (IS_ERR(mem))
kernel/dma/coherent.c
317
return PTR_ERR(mem);
kernel/dma/coherent.c
318
dma_coherent_default_memory = mem;
kernel/dma/coherent.c
33
if (mem->use_dev_dma_pfn_offset)
kernel/dma/coherent.c
339
struct dma_coherent_mem *mem = rmem->priv;
kernel/dma/coherent.c
34
return phys_to_dma(dev, PFN_PHYS(mem->pfn_base));
kernel/dma/coherent.c
341
if (!mem) {
kernel/dma/coherent.c
342
mem = dma_init_coherent_memory(rmem->base, rmem->base,
kernel/dma/coherent.c
344
if (IS_ERR(mem))
kernel/dma/coherent.c
345
return PTR_ERR(mem);
kernel/dma/coherent.c
346
rmem->priv = mem;
kernel/dma/coherent.c
35
return mem->device_base;
kernel/dma/coherent.c
350
if (mem->device_base + rmem->size - 1 >
kernel/dma/coherent.c
354
dma_assign_coherent_memory(dev, mem);
kernel/dma/coherent.c
77
static void _dma_release_coherent_memory(struct dma_coherent_mem *mem)
kernel/dma/coherent.c
79
if (!mem)
kernel/dma/coherent.c
82
memunmap(mem->virt_base);
kernel/dma/coherent.c
83
bitmap_free(mem->bitmap);
kernel/dma/coherent.c
84
kfree(mem);
kernel/dma/coherent.c
88
struct dma_coherent_mem *mem)
kernel/dma/coherent.c
96
dev->dma_mem = mem;
kernel/dma/direct.c
480
p2pdma_state.mem, sg_phys(sg));
kernel/dma/swiotlb.c
1000
static void dec_transient_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
1002
atomic_long_sub(nslots, &mem->transient_nslabs);
kernel/dma/swiotlb.c
1006
static void inc_transient_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
1009
static void dec_transient_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
1160
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
1166
list_for_each_entry_rcu(pool, &mem->pools, node) {
kernel/dma/swiotlb.c
1200
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
1219
if (!mem->can_grow)
kernel/dma/swiotlb.c
1222
schedule_work(&mem->dyn_alloc);
kernel/dma/swiotlb.c
1242
inc_transient_used(mem, pool->nslabs);
kernel/dma/swiotlb.c
1306
static unsigned long mem_used(struct io_tlb_mem *mem)
kernel/dma/swiotlb.c
1308
return atomic_long_read(&mem->total_used);
kernel/dma/swiotlb.c
1340
static unsigned long mem_used(struct io_tlb_mem *mem)
kernel/dma/swiotlb.c
1347
list_for_each_entry_rcu(pool, &mem->pools, node)
kernel/dma/swiotlb.c
1353
return mem_pool_used(&mem->defpool);
kernel/dma/swiotlb.c
1388
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
1397
if (!mem || !mem->nslabs) {
kernel/dma/swiotlb.c
1423
size, mem->nslabs, mem_used(mem));
kernel/dma/swiotlb.c
1459
struct io_tlb_pool *mem)
kernel/dma/swiotlb.c
1467
index = (tlb_addr - offset - mem->start) >> IO_TLB_SHIFT;
kernel/dma/swiotlb.c
1468
index -= mem->slots[index].pad_slots;
kernel/dma/swiotlb.c
1469
nslots = nr_slots(mem->slots[index].alloc_size + offset);
kernel/dma/swiotlb.c
1470
aindex = index / mem->area_nslabs;
kernel/dma/swiotlb.c
1471
area = &mem->areas[aindex];
kernel/dma/swiotlb.c
1479
BUG_ON(aindex >= mem->nareas);
kernel/dma/swiotlb.c
1483
count = mem->slots[index + nslots].list;
kernel/dma/swiotlb.c
1492
mem->slots[i].list = ++count;
kernel/dma/swiotlb.c
1493
mem->slots[i].orig_addr = INVALID_PHYS_ADDR;
kernel/dma/swiotlb.c
1494
mem->slots[i].alloc_size = 0;
kernel/dma/swiotlb.c
1495
mem->slots[i].pad_slots = 0;
kernel/dma/swiotlb.c
1503
io_tlb_offset(i) != IO_TLB_SEGSIZE - 1 && mem->slots[i].list;
kernel/dma/swiotlb.c
1505
mem->slots[i].list = ++count;
kernel/dma/swiotlb.c
1646
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
1648
return mem && mem->nslabs;
kernel/dma/swiotlb.c
1680
static unsigned long mem_transient_used(struct io_tlb_mem *mem)
kernel/dma/swiotlb.c
1682
return atomic_long_read(&mem->transient_nslabs);
kernel/dma/swiotlb.c
1687
struct io_tlb_mem *mem = data;
kernel/dma/swiotlb.c
1689
*val = mem_transient_used(mem);
kernel/dma/swiotlb.c
1699
struct io_tlb_mem *mem = data;
kernel/dma/swiotlb.c
1701
*val = mem_used(mem);
kernel/dma/swiotlb.c
1707
struct io_tlb_mem *mem = data;
kernel/dma/swiotlb.c
1709
*val = atomic_long_read(&mem->used_hiwater);
kernel/dma/swiotlb.c
1715
struct io_tlb_mem *mem = data;
kernel/dma/swiotlb.c
1721
atomic_long_set(&mem->used_hiwater, val);
kernel/dma/swiotlb.c
1729
static void swiotlb_create_debugfs_files(struct io_tlb_mem *mem,
kernel/dma/swiotlb.c
1732
mem->debugfs = debugfs_create_dir(dirname, io_tlb_default_mem.debugfs);
kernel/dma/swiotlb.c
1733
if (!mem->nslabs)
kernel/dma/swiotlb.c
1736
debugfs_create_ulong("io_tlb_nslabs", 0400, mem->debugfs, &mem->nslabs);
kernel/dma/swiotlb.c
1737
debugfs_create_file("io_tlb_used", 0400, mem->debugfs, mem,
kernel/dma/swiotlb.c
1739
debugfs_create_file("io_tlb_used_hiwater", 0600, mem->debugfs, mem,
kernel/dma/swiotlb.c
1742
debugfs_create_file("io_tlb_transient_nslabs", 0400, mem->debugfs,
kernel/dma/swiotlb.c
1743
mem, &fops_io_tlb_transient_used);
kernel/dma/swiotlb.c
1757
static inline void swiotlb_create_debugfs_files(struct io_tlb_mem *mem,
kernel/dma/swiotlb.c
1768
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
1774
if (!mem)
kernel/dma/swiotlb.c
1810
struct io_tlb_mem *mem = rmem->priv;
kernel/dma/swiotlb.c
1826
if (!mem) {
kernel/dma/swiotlb.c
1829
mem = kzalloc_obj(*mem);
kernel/dma/swiotlb.c
1830
if (!mem)
kernel/dma/swiotlb.c
1832
pool = &mem->defpool;
kernel/dma/swiotlb.c
1836
kfree(mem);
kernel/dma/swiotlb.c
1843
kfree(mem);
kernel/dma/swiotlb.c
1851
mem->force_bounce = true;
kernel/dma/swiotlb.c
1852
mem->for_alloc = true;
kernel/dma/swiotlb.c
1854
spin_lock_init(&mem->lock);
kernel/dma/swiotlb.c
1855
INIT_LIST_HEAD_RCU(&mem->pools);
kernel/dma/swiotlb.c
1857
add_mem_pool(mem, pool);
kernel/dma/swiotlb.c
1859
rmem->priv = mem;
kernel/dma/swiotlb.c
1861
swiotlb_create_debugfs_files(mem, rmem->name);
kernel/dma/swiotlb.c
1864
dev->dma_io_tlb_mem = mem;
kernel/dma/swiotlb.c
230
struct io_tlb_pool *mem = &io_tlb_default_mem.defpool;
kernel/dma/swiotlb.c
232
if (!mem->nslabs) {
kernel/dma/swiotlb.c
237
pr_info("mapped [mem %pa-%pa] (%luMB)\n", &mem->start, &mem->end,
kernel/dma/swiotlb.c
238
(mem->nslabs << IO_TLB_SHIFT) >> 20);
kernel/dma/swiotlb.c
259
struct io_tlb_pool *mem = &io_tlb_default_mem.defpool;
kernel/dma/swiotlb.c
262
if (!mem->nslabs || mem->late_alloc)
kernel/dma/swiotlb.c
264
bytes = PAGE_ALIGN(mem->nslabs << IO_TLB_SHIFT);
kernel/dma/swiotlb.c
265
set_memory_decrypted((unsigned long)mem->vaddr, bytes >> PAGE_SHIFT);
kernel/dma/swiotlb.c
268
static void swiotlb_init_io_tlb_pool(struct io_tlb_pool *mem, phys_addr_t start,
kernel/dma/swiotlb.c
274
mem->nslabs = nslabs;
kernel/dma/swiotlb.c
275
mem->start = start;
kernel/dma/swiotlb.c
276
mem->end = mem->start + bytes;
kernel/dma/swiotlb.c
277
mem->late_alloc = late_alloc;
kernel/dma/swiotlb.c
278
mem->nareas = nareas;
kernel/dma/swiotlb.c
279
mem->area_nslabs = nslabs / mem->nareas;
kernel/dma/swiotlb.c
281
for (i = 0; i < mem->nareas; i++) {
kernel/dma/swiotlb.c
282
spin_lock_init(&mem->areas[i].lock);
kernel/dma/swiotlb.c
283
mem->areas[i].index = 0;
kernel/dma/swiotlb.c
284
mem->areas[i].used = 0;
kernel/dma/swiotlb.c
287
for (i = 0; i < mem->nslabs; i++) {
kernel/dma/swiotlb.c
288
mem->slots[i].list = min(IO_TLB_SEGSIZE - io_tlb_offset(i),
kernel/dma/swiotlb.c
289
mem->nslabs - i);
kernel/dma/swiotlb.c
290
mem->slots[i].orig_addr = INVALID_PHYS_ADDR;
kernel/dma/swiotlb.c
291
mem->slots[i].alloc_size = 0;
kernel/dma/swiotlb.c
292
mem->slots[i].pad_slots = 0;
kernel/dma/swiotlb.c
296
mem->vaddr = vaddr;
kernel/dma/swiotlb.c
305
static void add_mem_pool(struct io_tlb_mem *mem, struct io_tlb_pool *pool)
kernel/dma/swiotlb.c
308
spin_lock(&mem->lock);
kernel/dma/swiotlb.c
309
list_add_rcu(&pool->node, &mem->pools);
kernel/dma/swiotlb.c
310
mem->nslabs += pool->nslabs;
kernel/dma/swiotlb.c
311
spin_unlock(&mem->lock);
kernel/dma/swiotlb.c
313
mem->nslabs = pool->nslabs;
kernel/dma/swiotlb.c
356
struct io_tlb_pool *mem = &io_tlb_default_mem.defpool;
kernel/dma/swiotlb.c
397
alloc_size = PAGE_ALIGN(array_size(sizeof(*mem->slots), nslabs));
kernel/dma/swiotlb.c
398
mem->slots = memblock_alloc(alloc_size, PAGE_SIZE);
kernel/dma/swiotlb.c
399
if (!mem->slots) {
kernel/dma/swiotlb.c
405
mem->areas = memblock_alloc(array_size(sizeof(struct io_tlb_area),
kernel/dma/swiotlb.c
407
if (!mem->areas) {
kernel/dma/swiotlb.c
412
swiotlb_init_io_tlb_pool(mem, __pa(tlb), nslabs, false, nareas);
kernel/dma/swiotlb.c
413
add_mem_pool(&io_tlb_default_mem, mem);
kernel/dma/swiotlb.c
432
struct io_tlb_pool *mem = &io_tlb_default_mem.defpool;
kernel/dma/swiotlb.c
497
area_order = get_order(array_size(sizeof(*mem->areas), nareas));
kernel/dma/swiotlb.c
498
mem->areas = (struct io_tlb_area *)
kernel/dma/swiotlb.c
500
if (!mem->areas)
kernel/dma/swiotlb.c
503
mem->slots = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
kernel/dma/swiotlb.c
504
get_order(array_size(sizeof(*mem->slots), nslabs)));
kernel/dma/swiotlb.c
505
if (!mem->slots)
kernel/dma/swiotlb.c
510
swiotlb_init_io_tlb_pool(mem, virt_to_phys(vstart), nslabs, true,
kernel/dma/swiotlb.c
512
add_mem_pool(&io_tlb_default_mem, mem);
kernel/dma/swiotlb.c
518
free_pages((unsigned long)mem->areas, area_order);
kernel/dma/swiotlb.c
526
struct io_tlb_pool *mem = &io_tlb_default_mem.defpool;
kernel/dma/swiotlb.c
534
if (!mem->nslabs)
kernel/dma/swiotlb.c
538
tbl_vaddr = (unsigned long)phys_to_virt(mem->start);
kernel/dma/swiotlb.c
539
tbl_size = PAGE_ALIGN(mem->end - mem->start);
kernel/dma/swiotlb.c
540
slots_size = PAGE_ALIGN(array_size(sizeof(*mem->slots), mem->nslabs));
kernel/dma/swiotlb.c
543
if (mem->late_alloc) {
kernel/dma/swiotlb.c
544
area_order = get_order(array_size(sizeof(*mem->areas),
kernel/dma/swiotlb.c
545
mem->nareas));
kernel/dma/swiotlb.c
546
free_pages((unsigned long)mem->areas, area_order);
kernel/dma/swiotlb.c
548
free_pages((unsigned long)mem->slots, get_order(slots_size));
kernel/dma/swiotlb.c
550
memblock_free_late(__pa(mem->areas),
kernel/dma/swiotlb.c
551
array_size(sizeof(*mem->areas), mem->nareas));
kernel/dma/swiotlb.c
552
memblock_free_late(mem->start, tbl_size);
kernel/dma/swiotlb.c
553
memblock_free_late(__pa(mem->slots), slots_size);
kernel/dma/swiotlb.c
556
memset(mem, 0, sizeof(*mem));
kernel/dma/swiotlb.c
735
struct io_tlb_mem *mem =
kernel/dma/swiotlb.c
740
default_nareas, mem->phys_limit, GFP_KERNEL);
kernel/dma/swiotlb.c
746
add_mem_pool(mem, pool);
kernel/dma/swiotlb.c
778
struct io_tlb_mem *mem = dev->dma_io_tlb_mem;
kernel/dma/swiotlb.c
782
list_for_each_entry_rcu(pool, &mem->pools, node) {
kernel/dma/swiotlb.c
859
enum dma_data_direction dir, struct io_tlb_pool *mem)
kernel/dma/swiotlb.c
861
int index = (tlb_addr - mem->start) >> IO_TLB_SHIFT;
kernel/dma/swiotlb.c
862
phys_addr_t orig_addr = mem->slots[index].orig_addr;
kernel/dma/swiotlb.c
863
size_t alloc_size = mem->slots[index].alloc_size;
kernel/dma/swiotlb.c
865
unsigned char *vaddr = mem->vaddr + tlb_addr - mem->start;
kernel/dma/swiotlb.c
952
static unsigned int wrap_area_index(struct io_tlb_pool *mem, unsigned int index)
kernel/dma/swiotlb.c
954
if (index >= mem->area_nslabs)
kernel/dma/swiotlb.c
966
static void inc_used_and_hiwater(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
970
new_used = atomic_long_add_return(nslots, &mem->total_used);
kernel/dma/swiotlb.c
971
old_hiwater = atomic_long_read(&mem->used_hiwater);
kernel/dma/swiotlb.c
975
} while (!atomic_long_try_cmpxchg(&mem->used_hiwater,
kernel/dma/swiotlb.c
979
static void dec_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
981
atomic_long_sub(nslots, &mem->total_used);
kernel/dma/swiotlb.c
985
static void inc_used_and_hiwater(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
988
static void dec_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
995
static void inc_transient_used(struct io_tlb_mem *mem, unsigned int nslots)
kernel/dma/swiotlb.c
997
atomic_long_add(nslots, &mem->transient_nslabs);
kernel/kexec.c
298
ksegments[i].mem = in.mem;
kernel/kexec_core.c
134
mstart = image->segment[i].mem;
kernel/kexec_core.c
153
mstart = image->segment[i].mem;
kernel/kexec_core.c
158
pstart = image->segment[j].mem;
kernel/kexec_core.c
206
mstart = image->segment[i].mem;
kernel/kexec_core.c
224
accept_memory(image->segment[i].mem, image->segment[i].memsz);
kernel/kexec_core.c
270
mstart = image->segment[i].mem;
kernel/kexec_core.c
439
mstart = image->segment[i].mem;
kernel/kexec_core.c
812
maddr = segment->mem;
kernel/kexec_core.c
889
maddr = segment->mem;
kernel/kexec_core.c
971
addr = image->segment[idx].mem;
kernel/kexec_elf.c
416
kbuf->mem = KEXEC_BUF_MEM_UNKNOWN;
kernel/kexec_elf.c
420
load_addr = kbuf->mem;
kernel/kexec_file.c
1035
bss_addr = kbuf->mem + kbuf->bufsz;
kernel/kexec_file.c
1064
kbuf->image->start = kbuf->mem + offset + entry_off;
kernel/kexec_file.c
1084
kbuf->image->start += kbuf->mem + offset;
kernel/kexec_file.c
1092
sechdrs[i].sh_addr = kbuf->mem + offset;
kernel/kexec_file.c
438
i, ksegment->buf, ksegment->bufsz, ksegment->mem,
kernel/kexec_file.c
511
kbuf->mem = temp_start;
kernel/kexec_file.c
553
kbuf->mem = temp_start;
kernel/kexec_file.c
673
unsigned long mem;
kernel/kexec_file.c
690
mem = page_to_boot_pfn(p) << PAGE_SHIFT;
kernel/kexec_file.c
692
if (kimage_is_destination_range(kbuf->image, mem, mem + kbuf->memsz)) {
kernel/kexec_file.c
694
pr_debug("CMA overlaps existing mem: 0x%lx+0x%lx\n", mem, kbuf->memsz);
kernel/kexec_file.c
699
kbuf->mem = page_to_boot_pfn(p) << PAGE_SHIFT;
kernel/kexec_file.c
720
if (kbuf->mem != KEXEC_BUF_MEM_UNKNOWN)
kernel/kexec_file.c
794
ksegment->mem = kbuf->mem;
kernel/kexec_file.c
865
sha_regions[j].start = ksegment->mem;
kernel/liveupdate/kexec_handover.c
1204
void kho_unpreserve_free(void *mem)
kernel/liveupdate/kexec_handover.c
1208
if (!mem)
kernel/liveupdate/kexec_handover.c
1211
folio = virt_to_folio(mem);
kernel/liveupdate/kexec_handover.c
1229
void kho_restore_free(void *mem)
kernel/liveupdate/kexec_handover.c
1233
if (!mem)
kernel/liveupdate/kexec_handover.c
1236
folio = kho_restore_folio(__pa(mem));
kernel/liveupdate/kexec_handover.c
1564
.mem = KEXEC_BUF_MEM_UNKNOWN,
kernel/liveupdate/luo_file.c
176
void *mem;
kernel/liveupdate/luo_file.c
184
mem = kho_alloc_preserve(size);
kernel/liveupdate/luo_file.c
185
if (IS_ERR(mem))
kernel/liveupdate/luo_file.c
186
return PTR_ERR(mem);
kernel/liveupdate/luo_file.c
188
file_set->files = mem;
kernel/module/debug_kmemleak.c
18
!mod->mem[type].is_rox)
kernel/module/debug_kmemleak.c
19
kmemleak_no_scan(mod->mem[type].base);
kernel/module/kallsyms.c
119
struct module_memory *mod_mem_data = &mod->mem[MOD_DATA];
kernel/module/kallsyms.c
120
struct module_memory *mod_mem_init_data = &mod->mem[MOD_INIT_DATA];
kernel/module/kallsyms.c
178
void *data_base = mod->mem[MOD_DATA].base;
kernel/module/kallsyms.c
179
void *init_data_base = mod->mem[MOD_INIT_DATA].base;
kernel/module/kallsyms.c
265
mod_mem = &mod->mem[MOD_INIT_TEXT];
kernel/module/kallsyms.c
267
mod_mem = &mod->mem[MOD_TEXT];
kernel/module/kdb.c
29
kdb_printf("%-20s%8u", mod->name, mod->mem[MOD_TEXT].size);
kernel/module/kdb.c
30
kdb_printf("/%8u", mod->mem[MOD_RODATA].size);
kernel/module/kdb.c
31
kdb_printf("/%8u", mod->mem[MOD_RO_AFTER_INIT].size);
kernel/module/kdb.c
32
kdb_printf("/%8u", mod->mem[MOD_DATA].size);
kernel/module/kdb.c
44
kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base);
kernel/module/kdb.c
45
kdb_printf("/0x%px", mod->mem[MOD_RODATA].base);
kernel/module/kdb.c
46
kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base);
kernel/module/kdb.c
47
kdb_printf("/0x%px", mod->mem[MOD_DATA].base);
kernel/module/main.c
1004
unsigned int size = mk->mod->mem[MOD_TEXT].size;
kernel/module/main.c
1008
size += mk->mod->mem[type].size;
kernel/module/main.c
1023
size += mk->mod->mem[type].size;
kernel/module/main.c
1037
size += mk->mod->mem[type].size;
kernel/module/main.c
121
struct module_memory *mod_mem = &mod->mem[type];
kernel/module/main.c
1314
unsigned int size = PAGE_ALIGN(mod->mem[type].size);
kernel/module/main.c
1318
mod->mem[type].size = size;
kernel/module/main.c
1329
mod->mem[type].is_rox = execmem_is_rox(execmem_type);
kernel/module/main.c
1342
if (!mod->mem[type].is_rox)
kernel/module/main.c
1346
mod->mem[type].base = ptr;
kernel/module/main.c
1354
struct module_memory *mem = &mod->mem[type];
kernel/module/main.c
1356
if (mem->is_rox)
kernel/module/main.c
1357
execmem_restore_rox(mem->base, mem->size);
kernel/module/main.c
1363
struct module_memory *mem = &mod->mem[type];
kernel/module/main.c
1365
execmem_free(mem->base);
kernel/module/main.c
1371
struct module_memory *mod_mem = &mod->mem[type];
kernel/module/main.c
1383
lockdep_free_key_range(mod->mem[MOD_DATA].base, mod->mem[MOD_DATA].size);
kernel/module/main.c
1646
mod->mem[type].size += arch_mod_section_prepend(mod, section);
kernel/module/main.c
1647
offset = ALIGN(mod->mem[type].size, sechdr->sh_addralign ?: 1);
kernel/module/main.c
1648
mod->mem[type].size = offset + sechdr->sh_size;
kernel/module/main.c
2738
if (!mod->mem[type].size) {
kernel/module/main.c
2739
mod->mem[type].base = NULL;
kernel/module/main.c
2781
dest = mod->mem[type].base + offset;
kernel/module/main.c
2840
const struct module_memory *mod_mem = &mod->mem[type];
kernel/module/main.c
3025
const struct module_memory *mod_mem = &mod->mem[type];
kernel/module/main.c
3039
freeinit->init_text = mod->mem[MOD_INIT_TEXT].base;
kernel/module/main.c
3040
freeinit->init_data = mod->mem[MOD_INIT_DATA].base;
kernel/module/main.c
3041
freeinit->init_rodata = mod->mem[MOD_INIT_RODATA].base;
kernel/module/main.c
3077
ftrace_free_mem(mod, mod->mem[MOD_INIT_TEXT].base,
kernel/module/main.c
3078
mod->mem[MOD_INIT_TEXT].base + mod->mem[MOD_INIT_TEXT].size);
kernel/module/main.c
3096
mod->mem[type].base = NULL;
kernel/module/main.c
3097
mod->mem[type].size = 0;
kernel/module/procfs.c
70
size += mod->mem[type].size;
kernel/module/procfs.c
95
value = m->private ? NULL : mod->mem[MOD_TEXT].base;
kernel/module/strict_rwx.c
18
const struct module_memory *mod_mem = &mod->mem[type];
kernel/module/strict_rwx.c
36
const struct module_memory *mem = &mod->mem[type];
kernel/module/strict_rwx.c
39
if (mem->is_rox)
kernel/module/strict_rwx.c
40
ret = execmem_restore_rox(mem->base, mem->size);
kernel/module/tree_lookup.c
81
mod->mem[type].mtn.mod = mod;
kernel/module/tree_lookup.c
82
if (mod->mem[type].size)
kernel/module/tree_lookup.c
83
__mod_tree_insert(&mod->mem[type].mtn, &mod_tree);
kernel/module/tree_lookup.c
90
if (mod->mem[type].size)
kernel/module/tree_lookup.c
91
__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
kernel/module/tree_lookup.c
98
if (mod->mem[type].size)
kernel/module/tree_lookup.c
99
__mod_tree_remove(&mod->mem[type].mtn, &mod_tree);
kernel/relay.c
110
void *mem;
kernel/relay.c
126
mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL);
kernel/relay.c
127
if (!mem)
kernel/relay.c
131
return mem;
kernel/trace/ftrace.c
572
offset = rec->ip - (unsigned long)mod->mem[MOD_TEXT].base;
kernel/trace/trace.c
5439
entry->mod_addr = (unsigned long)mod->mem[MOD_TEXT].base;
kernel/trace/trace.c
9280
module_delta->delta[i] = (unsigned long)mod->mem[MOD_TEXT].base
kernel/trace/trace_seq.c
284
void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len)
kernel/trace/trace_seq.c
296
seq_buf_putmem(&s->seq, mem, len);
kernel/trace/trace_seq.c
310
void trace_seq_putmem_hex(struct trace_seq *s, const void *mem,
kernel/trace/trace_seq.c
327
seq_buf_putmem_hex(&s->seq, mem, len);
lib/crypto/aes.c
187
static __always_inline void aes_prefetch(const void *mem, size_t len)
lib/crypto/aes.c
190
*(volatile const u8 *)(mem + i);
lib/seq_buf.c
236
int seq_buf_putmem(struct seq_buf *s, const void *mem, unsigned int len)
lib/seq_buf.c
241
memcpy(s->buffer + s->len, mem, len);
lib/seq_buf.c
264
int seq_buf_putmem_hex(struct seq_buf *s, const void *mem,
lib/seq_buf.c
268
const unsigned char *data = mem;
lib/test_bpf.c
1917
u64 mem, upd, res;
lib/test_bpf.c
1926
mem = 0x0123456789abcdefULL;
lib/test_bpf.c
1929
mem = 0x01234567U;
lib/test_bpf.c
1939
res = mem;
lib/test_bpf.c
1942
__bpf_alu_result(&res, mem, upd, BPF_OP(op));
lib/test_bpf.c
1951
i += __bpf_ld_imm64(&insn[i], R0, mem);
lib/test_bpf.c
1955
i += __bpf_ld_imm64(&insn[i], R0, ~mem);
lib/test_bpf.c
1966
cmp = mem; /* Expect value from memory */
lib/test_bpf.c
1970
cmp = ~mem; /* Expect value to be preserved */
lib/test_bpf.c
1992
src = mem; /* Expect fetched value from mem */
lib/test_bpf.c
2021
i += __bpf_ld_imm64(&insn[i], R1, mem);
lib/test_bpf.c
2023
i += __bpf_ld_imm64(&insn[i], R1, mem);
lib/tests/fortify_kunit.c
941
char *mem = haystack + strlen("Where oh where is ");
lib/tests/fortify_kunit.c
948
mem);
lib/tests/fortify_kunit.c
962
char *mem = haystack + strlen("Where oh where is ");
lib/tests/fortify_kunit.c
969
mem);
lib/tests/fortify_kunit.c
983
char *mem = haystack + 1;
lib/tests/fortify_kunit.c
991
mem);
lib/zlib_deflate/deflate.c
196
deflate_workspace *mem;
lib/zlib_deflate/deflate.c
210
mem = (deflate_workspace *) strm->workspace;
lib/zlib_deflate/deflate.c
226
next = (char *) mem;
lib/zlib_deflate/deflate.c
227
next += sizeof(*mem);
lib/zlib_deflate/deflate.c
233
mem->window_memory = (Byte *) PTR_ALIGN(next, PAGE_SIZE);
lib/zlib_deflate/deflate.c
235
mem->window_memory = (Byte *) next;
lib/zlib_deflate/deflate.c
238
mem->prev_memory = (Pos *) next;
lib/zlib_deflate/deflate.c
240
mem->head_memory = (Pos *) next;
lib/zlib_deflate/deflate.c
242
mem->overlay_memory = next;
lib/zlib_deflate/deflate.c
244
s = (deflate_state *) &(mem->deflate_memory);
lib/zlib_deflate/deflate.c
258
s->window = (Byte *) mem->window_memory;
lib/zlib_deflate/deflate.c
259
s->prev = (Pos *) mem->prev_memory;
lib/zlib_deflate/deflate.c
260
s->head = (Pos *) mem->head_memory;
lib/zlib_deflate/deflate.c
264
overlay = (ush *) mem->overlay_memory;
mm/cma_debug.c
100
cma_release(cma, mem->p, mem->n);
mm/cma_debug.c
101
count -= mem->n;
mm/cma_debug.c
102
kfree(mem);
mm/cma_debug.c
104
cma_release(cma, mem->p, count);
mm/cma_debug.c
105
mem->p += count;
mm/cma_debug.c
106
mem->n -= count;
mm/cma_debug.c
108
cma_add_to_cma_mem_list(cma, mem);
mm/cma_debug.c
111
cma_add_to_cma_mem_list(cma, mem);
mm/cma_debug.c
131
struct cma_mem *mem;
mm/cma_debug.c
134
mem = kzalloc_obj(*mem);
mm/cma_debug.c
135
if (!mem)
mm/cma_debug.c
140
kfree(mem);
mm/cma_debug.c
144
mem->p = p;
mm/cma_debug.c
145
mem->n = count;
mm/cma_debug.c
147
cma_add_to_cma_mem_list(cma, mem);
mm/cma_debug.c
69
static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem)
mm/cma_debug.c
72
hlist_add_head(&mem->node, &cma->mem_head);
mm/cma_debug.c
78
struct cma_mem *mem = NULL;
mm/cma_debug.c
82
mem = hlist_entry(cma->mem_head.first, struct cma_mem, node);
mm/cma_debug.c
83
hlist_del_init(&mem->node);
mm/cma_debug.c
87
return mem;
mm/cma_debug.c
92
struct cma_mem *mem = NULL;
mm/cma_debug.c
95
mem = cma_get_entry_from_list(cma);
mm/cma_debug.c
96
if (mem == NULL)
mm/cma_debug.c
99
if (mem->n <= count) {
mm/hmm.c
819
return pci_p2pdma_bus_addr_map(p2pdma_state->mem, paddr);
mm/memory_hotplug.c
1345
static int online_memory_block(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
1347
mem->online_type = mhp_get_default_online_type();
mm/memory_hotplug.c
1348
return device_online(&mem->dev);
mm/memory_hotplug.c
1425
struct memory_block *mem;
mm/memory_hotplug.c
1427
mem = find_memory_block(pfn_to_section_nr(PFN_DOWN(cur_start)));
mm/memory_hotplug.c
1428
if (WARN_ON_ONCE(!mem))
mm/memory_hotplug.c
1431
altmap = mem->altmap;
mm/memory_hotplug.c
1432
mem->altmap = NULL;
mm/memory_hotplug.c
2119
static int check_memblock_offlined_cb(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
2123
*nid = mem->nid;
mm/memory_hotplug.c
2124
if (unlikely(mem->state != MEM_OFFLINE)) {
mm/memory_hotplug.c
2127
beginpa = PFN_PHYS(section_nr_to_pfn(mem->start_section_nr));
mm/memory_hotplug.c
2137
static int count_memory_range_altmaps_cb(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
2141
if (mem->altmap)
mm/memory_hotplug.c
2163
static int check_no_memblock_for_node_cb(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
2172
return mem->nid == nid ? -EEXIST : 0;
mm/memory_hotplug.c
2326
static int try_offline_memory_block(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
2338
page = pfn_to_online_page(section_nr_to_pfn(mem->start_section_nr));
mm/memory_hotplug.c
2342
rc = device_offline(&mem->dev);
mm/memory_hotplug.c
2355
static int try_reonline_memory_block(struct memory_block *mem, void *arg)
mm/memory_hotplug.c
2361
mem->online_type = **online_types;
mm/memory_hotplug.c
2362
rc = device_online(&mem->dev);
mm/mempool.c
720
struct kmem_cache *mem = pool_data;
mm/mempool.c
721
VM_BUG_ON(mem->ctor);
mm/mempool.c
722
return kmem_cache_alloc_noprof(mem, gfp_mask);
mm/mempool.c
728
struct kmem_cache *mem = pool_data;
mm/mempool.c
729
kmem_cache_free(mem, element);
mm/nommu.c
328
void vm_unmap_ram(const void *mem, unsigned int count)
mm/page_poison.c
48
static void check_poison_mem(struct page *page, unsigned char *mem, size_t bytes)
mm/page_poison.c
54
start = memchr_inv(mem, PAGE_POISON, bytes);
mm/page_poison.c
58
for (end = mem + bytes - 1; end > start; end--) {
mm/slab_common.c
1230
void *mem = (void *)p;
mm/slab_common.c
1232
ks = ksize(mem);
mm/slab_common.c
1234
kasan_unpoison_range(mem, ks);
mm/slab_common.c
1235
memzero_explicit(mem, ks);
mm/slab_common.c
1237
kfree(mem);
mm/vmalloc.c
3025
void vm_unmap_ram(const void *mem, unsigned int count)
mm/vmalloc.c
3028
unsigned long addr = (unsigned long)kasan_reset_tag(mem);
mm/vmalloc.c
3037
kasan_poison_vmalloc(mem, size);
mm/vmalloc.c
3040
debug_check_no_locks_freed(mem, size);
mm/vmalloc.c
3072
void *mem;
mm/vmalloc.c
3075
mem = vb_alloc(size, GFP_KERNEL);
mm/vmalloc.c
3076
if (IS_ERR(mem))
mm/vmalloc.c
3078
addr = (unsigned long)mem;
mm/vmalloc.c
3089
mem = (void *)addr;
mm/vmalloc.c
3094
vm_unmap_ram(mem, count);
mm/vmalloc.c
3103
mem = kasan_unpoison_vmalloc(mem, size, KASAN_VMALLOC_PROT_NORMAL);
mm/vmalloc.c
3105
return mem;
mm/vmstat.c
228
int mem; /* memory in 128 MB units */
mm/vmstat.c
260
mem = zone_managed_pages(zone) >> (27 - PAGE_SHIFT);
mm/vmstat.c
262
threshold = 2 * fls(num_online_cpus()) * (1 + fls(mem));
net/bpf/test_run.c
107
struct xdp_mem_info mem;
net/bpf/test_run.c
145
frm->mem_type = new_ctx->rxq->mem.type;
net/bpf/test_run.c
178
err = xdp_reg_mem_model(&xdp->mem, MEM_TYPE_PAGE_POOL, pp);
net/bpf/test_run.c
188
xdp->rxq.mem.type = MEM_TYPE_PAGE_POOL;
net/bpf/test_run.c
189
xdp->rxq.mem.id = pp->xdp_mem_id;
net/bpf/test_run.c
206
xdp_unreg_mem_model(&xdp->mem);
net/bpf/test_run.c
238
head->frame->mem_type = head->orig_ctx.rxq->mem.type;
net/core/filter.c
4174
if (rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL)
net/core/filter.c
4201
enum xdp_mem_type mem_type = xdp->rxq->mem.type;
net/core/page_pool.c
1235
const struct xdp_mem_info *mem)
net/core/page_pool.c
1239
pool->xdp_mem_id = mem->id;
net/core/sock.c
2875
void *mem;
net/core/sock.c
2880
mem = kmalloc(size, priority);
net/core/sock.c
2881
if (mem)
net/core/sock.c
2882
return mem;
net/core/sock.c
2896
void *mem;
net/core/sock.c
2898
mem = sock_kmalloc(sk, size, priority);
net/core/sock.c
2899
if (mem)
net/core/sock.c
2900
memcpy(mem, src, size);
net/core/sock.c
2901
return mem;
net/core/sock.c
2909
static inline void __sock_kfree_s(struct sock *sk, void *mem, int size,
net/core/sock.c
2912
if (WARN_ON_ONCE(!mem))
net/core/sock.c
2915
kfree_sensitive(mem);
net/core/sock.c
2917
kfree(mem);
net/core/sock.c
2921
void sock_kfree_s(struct sock *sk, void *mem, int size)
net/core/sock.c
2923
__sock_kfree_s(sk, mem, size, false);
net/core/sock.c
2927
void sock_kzfree_s(struct sock *sk, void *mem, int size)
net/core/sock.c
2929
__sock_kfree_s(sk, mem, size, true);
net/core/sock.c
4012
void sk_get_meminfo(const struct sock *sk, u32 *mem)
net/core/sock.c
4014
memset(mem, 0, sizeof(*mem) * SK_MEMINFO_VARS);
net/core/sock.c
4016
mem[SK_MEMINFO_RMEM_ALLOC] = sk_rmem_alloc_get(sk);
net/core/sock.c
4017
mem[SK_MEMINFO_RCVBUF] = READ_ONCE(sk->sk_rcvbuf);
net/core/sock.c
4018
mem[SK_MEMINFO_WMEM_ALLOC] = sk_wmem_alloc_get(sk);
net/core/sock.c
4019
mem[SK_MEMINFO_SNDBUF] = READ_ONCE(sk->sk_sndbuf);
net/core/sock.c
4020
mem[SK_MEMINFO_FWD_ALLOC] = READ_ONCE(sk->sk_forward_alloc);
net/core/sock.c
4021
mem[SK_MEMINFO_WMEM_QUEUED] = READ_ONCE(sk->sk_wmem_queued);
net/core/sock.c
4022
mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc);
net/core/sock.c
4023
mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
net/core/sock.c
4024
mem[SK_MEMINFO_DROPS] = sk_drops_read(sk);
net/core/sock_diag.c
68
u32 mem[SK_MEMINFO_VARS];
net/core/sock_diag.c
70
sk_get_meminfo(sk, mem);
net/core/sock_diag.c
72
return nla_put(skb, attrtype, sizeof(mem), &mem);
net/core/xdp.c
117
void xdp_unreg_mem_model(struct xdp_mem_info *mem)
net/core/xdp.c
120
int type = mem->type;
net/core/xdp.c
121
int id = mem->id;
net/core/xdp.c
124
mem->id = 0;
net/core/xdp.c
125
mem->type = 0;
net/core/xdp.c
144
xdp_unreg_mem_model(&xdp_rxq->mem);
net/core/xdp.c
271
static struct xdp_mem_allocator *__xdp_reg_mem_model(struct xdp_mem_info *mem,
net/core/xdp.c
283
mem->type = type;
net/core/xdp.c
310
mem->id = id;
net/core/xdp.c
311
xdp_alloc->mem = *mem;
net/core/xdp.c
317
ida_free(&mem_id_pool, mem->id);
net/core/xdp.c
318
mem->id = 0;
net/core/xdp.c
324
page_pool_use_xdp_mem(allocator, mem_allocator_disconnect, mem);
net/core/xdp.c
335
int xdp_reg_mem_model(struct xdp_mem_info *mem,
net/core/xdp.c
340
xdp_alloc = __xdp_reg_mem_model(mem, type, allocator);
net/core/xdp.c
357
xdp_alloc = __xdp_reg_mem_model(&xdp_rxq->mem, type, allocator);
net/core/xdp.c
383
struct xdp_mem_info mem;
net/core/xdp.c
385
return xdp_reg_mem_model(&mem, MEM_TYPE_PAGE_POOL, pool);
net/core/xdp.c
398
struct xdp_mem_info mem = {
net/core/xdp.c
403
xdp_unreg_mem_model(&mem);
net/core/xdp.c
418
struct xdp_mem_info mem = {
net/core/xdp.c
423
xdp_rxq_info_attach_mem_model(xdp_rxq, &mem);
net/core/xdp.c
45
BUILD_BUG_ON(sizeof_field(struct xdp_mem_allocator, mem.id)
net/core/xdp.c
545
__xdp_return(netmem, xdp->rxq->mem.type, true, NULL);
net/core/xdp.c
559
xdp->rxq->mem.type, true, xdp);
net/core/xdp.c
562
__xdp_return(virt_to_netmem(xdp->data), xdp->rxq->mem.type, true, xdp);
net/core/xdp.c
58
return xa->mem.id != mem_id;
net/core/xdp.c
64
.key_offset = offsetof(struct xdp_mem_allocator, mem.id),
net/core/xdp.c
65
.key_len = sizeof_field(struct xdp_mem_allocator, mem.id),
net/core/xdp.c
657
if (rxq->mem.type == MEM_TYPE_PAGE_POOL)
net/core/xdp.c
80
ida_free(&mem_id_pool, xa->mem.id);
net/netfilter/x_tables.c
1096
void *mem;
net/netfilter/x_tables.c
1134
mem = vmalloc(len);
net/netfilter/x_tables.c
1135
if (!mem)
net/netfilter/x_tables.c
1138
if (copy_from_sockptr_offset(mem, arg, offset, len) == 0)
net/netfilter/x_tables.c
1139
return mem;
net/netfilter/x_tables.c
1141
vfree(mem);
net/netfilter/x_tables.c
1395
struct xt_counters *mem;
net/netfilter/x_tables.c
1397
if (counters == 0 || counters > INT_MAX / sizeof(*mem))
net/netfilter/x_tables.c
1400
counters *= sizeof(*mem);
net/netfilter/x_tables.c
1956
if (!state->mem) {
net/netfilter/x_tables.c
1957
state->mem = __alloc_percpu(XT_PCPU_BLOCK_SIZE,
net/netfilter/x_tables.c
1959
if (!state->mem)
net/netfilter/x_tables.c
1962
counter->pcnt = (__force unsigned long)(state->mem + state->off);
net/netfilter/x_tables.c
1965
state->mem = NULL;
net/netfilter/x_tables.c
730
size_t mem;
net/netfilter/x_tables.c
740
mem = sizeof(struct compat_delta) * number;
net/netfilter/x_tables.c
741
if (mem > XT_MAX_TABLE_SIZE)
net/netfilter/x_tables.c
744
xt[af].compat_tab = vmalloc(mem);
net/sched/sch_fq_codel.c
145
unsigned int mem = 0;
net/sched/sch_fq_codel.c
170
mem += get_codel_cb(skb)->mem_usage;
net/sched/sch_fq_codel.c
178
q->memory_usage -= mem;
net/sctp/diag.c
164
u32 mem[SK_MEMINFO_VARS];
net/sctp/diag.c
171
mem[SK_MEMINFO_WMEM_ALLOC] = amt;
net/sctp/diag.c
176
mem[SK_MEMINFO_RMEM_ALLOC] = amt;
net/sctp/diag.c
177
mem[SK_MEMINFO_RCVBUF] = sk->sk_rcvbuf;
net/sctp/diag.c
178
mem[SK_MEMINFO_SNDBUF] = sk->sk_sndbuf;
net/sctp/diag.c
179
mem[SK_MEMINFO_FWD_ALLOC] = sk->sk_forward_alloc;
net/sctp/diag.c
180
mem[SK_MEMINFO_WMEM_QUEUED] = sk->sk_wmem_queued;
net/sctp/diag.c
181
mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc);
net/sctp/diag.c
182
mem[SK_MEMINFO_BACKLOG] = READ_ONCE(sk->sk_backlog.len);
net/sctp/diag.c
183
mem[SK_MEMINFO_DROPS] = sk_drops_read(sk);
net/sctp/diag.c
185
if (nla_put(skb, INET_DIAG_SKMEMINFO, sizeof(mem), &mem) < 0)
net/tipc/crypto.c
692
u8 *mem;
net/tipc/crypto.c
705
mem = kmalloc(len, GFP_ATOMIC);
net/tipc/crypto.c
706
if (!mem)
net/tipc/crypto.c
709
*iv = (u8 *)PTR_ALIGN(mem + crypto_ctx_size,
net/tipc/crypto.c
716
return (void *)mem;
net/tls/tls_sw.c
1527
u8 *mem;
net/tls/tls_sw.c
1561
mem = kmalloc(aead_size + struct_size(dctx, sg, size_add(n_sgin, n_sgout)),
net/tls/tls_sw.c
1563
if (!mem) {
net/tls/tls_sw.c
1569
aead_req = (struct aead_request *)mem;
net/tls/tls_sw.c
1570
dctx = (struct tls_decrypt_ctx *)(mem + aead_size);
net/tls/tls_sw.c
1676
kfree(mem);
net/xdp/xsk.c
381
if (xdp->rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL) {
samples/seccomp/user-trap.c
112
int ret = -1, mem;
samples/seccomp/user-trap.c
132
mem = open(path, O_RDONLY);
samples/seccomp/user-trap.c
133
if (mem < 0) {
samples/seccomp/user-trap.c
159
if (lseek(mem, req->data.args[0], SEEK_SET) < 0) {
samples/seccomp/user-trap.c
164
ret = read(mem, source, sizeof(source));
samples/seccomp/user-trap.c
170
if (lseek(mem, req->data.args[1], SEEK_SET) < 0) {
samples/seccomp/user-trap.c
175
ret = read(mem, target, sizeof(target));
samples/seccomp/user-trap.c
201
close(mem);
samples/vfio-mdev/mbochs.c
96
MODULE_PARM_DESC(mem, "megabytes available to " MBOCHS_NAME " devices");
scripts/dtc/data.c
44
struct data data_copy_mem(const char *mem, int len)
scripts/dtc/data.c
51
memcpy(d.val, mem, len);
scripts/dtc/dtc.h
170
struct data data_copy_mem(const char *mem, int len);
security/integrity/ima/ima_kexec.c
192
image->ima_buffer_addr = kbuf.mem;
security/integrity/ima/ima_kexec.c
199
kbuf.mem);
sound/isa/msnd/msnd_pinnacle.c
1070
mem[idx] = pnp_mem_start(pnp_dev, 0);
sound/isa/msnd/msnd_pinnacle.c
1082
chip->base = mem[idx];
sound/isa/msnd/msnd_pinnacle.c
648
static int snd_msnd_write_cfg_mem(struct snd_msnd *chip, int cfg, int num, int mem)
sound/isa/msnd/msnd_pinnacle.c
652
mem >>= 8;
sound/isa/msnd/msnd_pinnacle.c
653
wmem = (u16)(mem & 0xfff);
sound/isa/msnd/msnd_pinnacle.c
677
u16 io1, u16 irq, int mem)
sound/isa/msnd/msnd_pinnacle.c
687
if (snd_msnd_write_cfg_mem(chip, cfg, num, mem))
sound/isa/msnd/msnd_pinnacle.c
718
static long mem[SNDRV_CARDS] = SNDRV_DEFAULT_PORT;
sound/isa/msnd/msnd_pinnacle.c
758
module_param_hw_array(mem, long, iomem, NULL, 0444);
sound/isa/msnd/msnd_pinnacle.c
779
if (irq[i] == SNDRV_AUTO_PORT || mem[i] == SNDRV_AUTO_PORT) {
sound/isa/msnd/msnd_pinnacle.c
818
if (!(mem[i] == 0xb0000 ||
sound/isa/msnd/msnd_pinnacle.c
819
mem[i] == 0xc8000 ||
sound/isa/msnd/msnd_pinnacle.c
820
mem[i] == 0xd0000 ||
sound/isa/msnd/msnd_pinnacle.c
821
mem[i] == 0xd8000 ||
sound/isa/msnd/msnd_pinnacle.c
822
mem[i] == 0xe0000 ||
sound/isa/msnd/msnd_pinnacle.c
823
mem[i] == 0xe8000)) {
sound/isa/msnd/msnd_pinnacle.c
883
switch (mem[idx]) {
sound/isa/msnd/msnd_pinnacle.c
914
irq[idx], mem[idx]);
sound/isa/msnd/msnd_pinnacle.c
973
chip->base = mem[idx];
sound/pci/cs5530.c
84
void __iomem *mem;
sound/pci/cs5530.c
94
mem = pcim_iomap_region(pci, 0, "CS5530");
sound/pci/cs5530.c
95
if (IS_ERR(mem))
sound/pci/cs5530.c
96
return PTR_ERR(mem);
sound/pci/cs5530.c
98
map = readw(mem + 0x18);
sound/pci/emu10k1/memory.c
215
blk = get_emu10k1_memblk(p, mem.list);
sound/pci/emu10k1/memory.c
228
blk->mem.offset = aligned_page_offset(page); /* set aligned offset */
sound/pci/emu10k1/memory.c
446
p = blk->mem.list.prev;
sound/pci/emu10k1/memory.c
448
q = get_emu10k1_memblk(p, mem.list);
sound/pci/emu10k1/memory.c
453
p = blk->mem.list.next;
sound/pci/emu10k1/memory.c
455
q = get_emu10k1_memblk(p, mem.list);
sound/pci/emu10k1/memory.c
563
if (snd_BUG_ON(offset + size > p->mem.size))
sound/pci/emu10k1/memory.c
623
if (snd_BUG_ON(offset + size > p->mem.size))
sound/pci/emu10k1/memory.c
83
blk->first_page = get_aligned_page(blk->mem.offset);
sound/pci/emu10k1/memory.c
84
blk->last_page = get_aligned_page(blk->mem.offset + blk->mem.size - 1);
sound/pci/mixart/mixart.c
1084
iounmap(mgr->mem[i].virt);
sound/pci/mixart/mixart.c
1263
mgr->mem[i].phys = pci_resource_start(pci, i);
sound/pci/mixart/mixart.c
1264
mgr->mem[i].virt = pci_ioremap_bar(pci, i);
sound/pci/mixart/mixart.c
1265
if (!mgr->mem[i].virt) {
sound/pci/mixart/mixart.c
1267
mgr->mem[i].phys);
sound/pci/mixart/mixart.c
1320
mgr->mem[0].phys, mgr->mem[1].phys, mgr->irq, i);
sound/pci/mixart/mixart.h
62
struct mem_area mem[2];
sound/pci/mixart/mixart_hwdep.h
31
#define MIXART_MEM(mgr,x) ((mgr)->mem[0].virt + (x))
sound/pci/mixart/mixart_hwdep.h
32
#define MIXART_REG(mgr,x) ((mgr)->mem[1].virt + (x))
sound/soc/atmel/atmel-i2s.c
617
struct resource *mem;
sound/soc/atmel/atmel-i2s.c
636
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/atmel/atmel-i2s.c
706
dev->playback.addr = (dma_addr_t)mem->start + ATMEL_I2SC_THR;
sound/soc/atmel/atmel-i2s.c
708
dev->capture.addr = (dma_addr_t)mem->start + ATMEL_I2SC_RHR;
sound/soc/atmel/mchp-i2s-mcc.c
1029
struct resource *mem;
sound/soc/atmel/mchp-i2s-mcc.c
1040
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/atmel/mchp-i2s-mcc.c
1101
dev->playback.addr = (dma_addr_t)mem->start + MCHP_I2SMCC_THR;
sound/soc/atmel/mchp-i2s-mcc.c
1102
dev->capture.addr = (dma_addr_t)mem->start + MCHP_I2SMCC_RHR;
sound/soc/atmel/mchp-spdifrx.c
1078
struct resource *mem;
sound/soc/atmel/mchp-spdifrx.c
1091
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/atmel/mchp-spdifrx.c
1150
dev->capture.addr = (dma_addr_t)mem->start + SPDIFRX_RHR;
sound/soc/atmel/mchp-spdiftx.c
784
struct resource *mem;
sound/soc/atmel/mchp-spdiftx.c
797
base = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/atmel/mchp-spdiftx.c
852
dev->playback.addr = (dma_addr_t)mem->start + SPDIFTX_CDR;
sound/soc/codecs/cs35l41-lib.c
1426
dsp->mem = cs35l41_dsp1_regions;
sound/soc/codecs/cs35l45.c
1361
dsp->cs_dsp.mem = cs35l45_dsp1_regions;
sound/soc/codecs/cs35l56-shared.c
880
cs_dsp->mem = cs35l56_dsp1_regions;
sound/soc/codecs/cs47l15.c
1417
cs47l15->core.adsp[0].cs_dsp.mem = cs47l15_dsp1_regions;
sound/soc/codecs/cs47l24.c
1247
cs47l24->core.adsp[i].cs_dsp.mem = cs47l24_dsp_regions[i - 1];
sound/soc/codecs/cs47l35.c
1700
cs47l35->core.adsp[i].cs_dsp.mem = cs47l35_dsp_regions[i];
sound/soc/codecs/cs47l85.c
2646
cs47l85->core.adsp[i].cs_dsp.mem = cs47l85_dsp_regions[i];
sound/soc/codecs/cs47l90.c
2557
cs47l90->core.adsp[i].cs_dsp.mem = cs47l90_dsp_regions[i];
sound/soc/codecs/cs47l92.c
2020
cs47l92->core.adsp[0].cs_dsp.mem = cs47l92_dsp1_regions;
sound/soc/codecs/cs48l32.c
3688
dsp->cs_dsp.mem = cs48l32_dsp1_regions;
sound/soc/codecs/wm2200.c
2217
wm2200->dsp[0].cs_dsp.mem = wm2200_dsp1_regions;
sound/soc/codecs/wm2200.c
2221
wm2200->dsp[1].cs_dsp.mem = wm2200_dsp2_regions;
sound/soc/codecs/wm5102.c
2070
wm5102->core.adsp[0].cs_dsp.mem = wm5102_dsp1_regions;
sound/soc/codecs/wm5110.c
2426
wm5110->core.adsp[i].cs_dsp.mem = wm5110_dsp_regions[i];
sound/soc/jz4740/jz4740-i2s.c
542
struct resource *mem;
sound/soc/jz4740/jz4740-i2s.c
552
regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/jz4740/jz4740-i2s.c
557
i2s->playback_dma_data.addr = mem->start + JZ_REG_AIC_FIFO;
sound/soc/jz4740/jz4740-i2s.c
560
i2s->capture_dma_data.addr = mem->start + JZ_REG_AIC_FIFO;
sound/soc/renesas/rcar/dma.c
62
static struct rsnd_mod mem = {
sound/soc/renesas/rcar/dma.c
746
rsnd_mod_name(mod[i] ? mod[i] : &mem),
sound/soc/renesas/rcar/dma.c
813
rsnd_mod_name(mod_from ? mod_from : &mem),
sound/soc/renesas/rcar/dma.c
814
rsnd_mod_name(mod_to ? mod_to : &mem));
sound/soc/renesas/rcar/dma.c
882
return rsnd_mod_init(NULL, &mem, &mem_ops, NULL, 0, 0);
sound/soc/tegra/tegra20_ac97.c
304
struct resource *mem;
sound/soc/tegra/tegra20_ac97.c
330
regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/tegra/tegra20_ac97.c
365
ac97->capture_dma_data.addr = mem->start + TEGRA20_AC97_FIFO_RX1;
sound/soc/tegra/tegra20_ac97.c
369
ac97->playback_dma_data.addr = mem->start + TEGRA20_AC97_FIFO_TX1;
sound/soc/tegra/tegra20_i2s.c
402
struct resource *mem;
sound/soc/tegra/tegra20_i2s.c
429
regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/tegra/tegra20_i2s.c
443
i2s->capture_dma_data.addr = mem->start + TEGRA20_I2S_FIFO2;
sound/soc/tegra/tegra20_i2s.c
447
i2s->playback_dma_data.addr = mem->start + TEGRA20_I2S_FIFO1;
sound/soc/tegra/tegra20_spdif.c
346
struct resource *mem;
sound/soc/tegra/tegra20_spdif.c
369
regs = devm_platform_get_and_ioremap_resource(pdev, 0, &mem);
sound/soc/tegra/tegra20_spdif.c
380
spdif->playback_dma_data.addr = mem->start + TEGRA20_SPDIF_DATA_OUT;
sound/soc/tegra/tegra210_mbdrc.c
1000
mem.flags = IORESOURCE_MEM;
sound/soc/tegra/tegra210_mbdrc.c
1001
regs = devm_ioremap_resource(dev, &mem);
sound/soc/tegra/tegra210_mbdrc.c
985
struct resource mem;
sound/soc/tegra/tegra210_mbdrc.c
993
err = of_address_to_resource(child, 0, &mem);
sound/soc/tegra/tegra210_peq.c
405
struct resource mem;
sound/soc/tegra/tegra210_peq.c
413
err = of_address_to_resource(child, 0, &mem);
sound/soc/tegra/tegra210_peq.c
420
mem.flags = IORESOURCE_MEM;
sound/soc/tegra/tegra210_peq.c
421
regs = devm_ioremap_resource(dev, &mem);
sound/soc/ti/davinci-i2s.c
784
struct resource *mem, *res;
sound/soc/ti/davinci-i2s.c
789
mem = platform_get_resource_byname(pdev, IORESOURCE_MEM, "mpu");
sound/soc/ti/davinci-i2s.c
790
if (!mem) {
sound/soc/ti/davinci-i2s.c
793
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
sound/soc/ti/davinci-i2s.c
794
if (!mem) {
sound/soc/ti/davinci-i2s.c
800
io_base = devm_ioremap_resource(&pdev->dev, mem);
sound/soc/ti/davinci-i2s.c
816
dma_data->addr = (dma_addr_t)(mem->start + DAVINCI_MCBSP_DXR_REG);
sound/soc/ti/davinci-i2s.c
831
dma_data->addr = (dma_addr_t)(mem->start + DAVINCI_MCBSP_DRR_REG);
sound/soc/ti/davinci-mcasp.c
2544
struct resource *mem, *dat;
sound/soc/ti/davinci-mcasp.c
2560
mem = platform_get_resource_byname(pdev, IORESOURCE_MEM, "mpu");
sound/soc/ti/davinci-mcasp.c
2561
if (!mem) {
sound/soc/ti/davinci-mcasp.c
2564
mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
sound/soc/ti/davinci-mcasp.c
2565
if (!mem) {
sound/soc/ti/davinci-mcasp.c
2571
mcasp->base = devm_ioremap_resource(&pdev->dev, mem);
sound/soc/ti/davinci-mcasp.c
2666
dma_data->addr = mem->start + davinci_mcasp_txdma_offset(mcasp->pdata);
sound/soc/ti/davinci-mcasp.c
2678
mem->start + davinci_mcasp_rxdma_offset(mcasp->pdata);
tools/accounting/delaytop.c
170
SORT_FIELD(mem, m, MODE_DEFAULT | MODE_MEMVERBOSE),
tools/accounting/delaytop.c
977
TASK_AVG(tasks[i], mem),
tools/accounting/delaytop.c
988
TASK_AVG(tasks[i], mem));
tools/lib/perf/include/perf/event.h
11
#define event_contains(obj, mem) ((obj).header.size > offsetof(typeof(obj), mem))
tools/mm/thp_swap_allocator_test.c
50
void *mem = NULL;
tools/mm/thp_swap_allocator_test.c
52
if (posix_memalign(&mem, alignment, size) != 0) {
tools/mm/thp_swap_allocator_test.c
56
return mem;
tools/mm/thp_swap_allocator_test.c
65
static void random_madvise_dontneed(void *mem, size_t mem_size,
tools/mm/thp_swap_allocator_test.c
75
addr = (char *)mem + offset;
tools/mm/thp_swap_allocator_test.c
83
static void random_swapin(void *mem, size_t mem_size,
tools/mm/thp_swap_allocator_test.c
93
addr = (char *)mem + offset;
tools/perf/builtin-c2c.c
3077
.mem = true, /* Only enable memory event */
tools/perf/builtin-mem.c
105
if (mem->cpu_list)
tools/perf/builtin-mem.c
121
(mem->operation & MEM_OPERATION_LOAD) &&
tools/perf/builtin-mem.c
122
(mem->operation & MEM_OPERATION_STORE)) {
tools/perf/builtin-mem.c
126
if (mem->operation & MEM_OPERATION_LOAD)
tools/perf/builtin-mem.c
129
if (mem->operation & MEM_OPERATION_STORE)
tools/perf/builtin-mem.c
138
if (mem->phys_addr)
tools/perf/builtin-mem.c
141
if (mem->data_page_size)
tools/perf/builtin-mem.c
150
if (mem->all_user)
tools/perf/builtin-mem.c
153
if (mem->all_kernel)
tools/perf/builtin-mem.c
156
if (mem->cpu_list) {
tools/perf/builtin-mem.c
158
rec_argv[i++] = mem->cpu_list;
tools/perf/builtin-mem.c
186
struct perf_mem *mem = container_of(tool, struct perf_mem, tool);
tools/perf/builtin-mem.c
200
if (al.filtered || (mem->hide_unresolved && al.sym == NULL))
tools/perf/builtin-mem.c
226
if (mem->phys_addr) {
tools/perf/builtin-mem.c
232
if (mem->data_page_size) {
tools/perf/builtin-mem.c
264
static int report_raw_events(struct perf_mem *mem)
tools/perf/builtin-mem.c
268
.mem = true, /* Only enable memory event */
tools/perf/builtin-mem.c
275
.force = mem->force,
tools/perf/builtin-mem.c
280
perf_tool__init(&mem->tool, /*ordered_events=*/true);
tools/perf/builtin-mem.c
281
mem->tool.sample = process_sample_event;
tools/perf/builtin-mem.c
282
mem->tool.mmap = perf_event__process_mmap;
tools/perf/builtin-mem.c
283
mem->tool.mmap2 = perf_event__process_mmap2;
tools/perf/builtin-mem.c
284
mem->tool.comm = perf_event__process_comm;
tools/perf/builtin-mem.c
285
mem->tool.lost = perf_event__process_lost;
tools/perf/builtin-mem.c
286
mem->tool.fork = perf_event__process_fork;
tools/perf/builtin-mem.c
287
mem->tool.attr = perf_event__process_attr;
tools/perf/builtin-mem.c
288
mem->tool.build_id = perf_event__process_build_id;
tools/perf/builtin-mem.c
289
mem->tool.namespaces = perf_event__process_namespaces;
tools/perf/builtin-mem.c
290
mem->tool.auxtrace_info = perf_event__process_auxtrace_info;
tools/perf/builtin-mem.c
291
mem->tool.auxtrace = perf_event__process_auxtrace;
tools/perf/builtin-mem.c
292
mem->tool.auxtrace_error = perf_event__process_auxtrace_error;
tools/perf/builtin-mem.c
294
session = perf_session__new(&data, &mem->tool);
tools/perf/builtin-mem.c
301
if (mem->cpu_list) {
tools/perf/builtin-mem.c
302
ret = perf_session__cpu_bitmap(session, mem->cpu_list,
tools/perf/builtin-mem.c
303
mem->cpu_bitmap);
tools/perf/builtin-mem.c
314
if (mem->phys_addr)
tools/perf/builtin-mem.c
317
if (mem->data_page_size)
tools/perf/builtin-mem.c
329
static char *get_sort_order(struct perf_mem *mem)
tools/perf/builtin-mem.c
331
bool has_extra_options = (mem->phys_addr | mem->data_page_size) ? true : false;
tools/perf/builtin-mem.c
334
if (mem->sort_key)
tools/perf/builtin-mem.c
335
scnprintf(sort, sizeof(sort), "--sort=%s", mem->sort_key);
tools/perf/builtin-mem.c
336
else if (mem->data_type)
tools/perf/builtin-mem.c
342
else if (!(mem->operation & MEM_OPERATION_LOAD)) {
tools/perf/builtin-mem.c
351
if (mem->phys_addr)
tools/perf/builtin-mem.c
354
if (mem->data_page_size)
tools/perf/builtin-mem.c
358
if (mem->data_type && !strstr(sort, "type"))
tools/perf/builtin-mem.c
364
static int __cmd_report(int argc, const char **argv, struct perf_mem *mem,
tools/perf/builtin-mem.c
378
if (mem->dump_raw)
tools/perf/builtin-mem.c
379
return report_raw_events(mem);
tools/perf/builtin-mem.c
390
new_sort_order = get_sort_order(mem);
tools/perf/builtin-mem.c
475
struct perf_mem mem = {
tools/perf/builtin-mem.c
484
OPT_CALLBACK('t', "type", &mem.operation,
tools/perf/builtin-mem.c
487
OPT_STRING('C', "cpu", &mem.cpu_list, "cpu",
tools/perf/builtin-mem.c
489
OPT_BOOLEAN('f', "force", &mem.force, "don't complain, do it"),
tools/perf/builtin-mem.c
492
OPT_BOOLEAN('p', "phys-data", &mem.phys_addr, "Record/Report sample physical addresses"),
tools/perf/builtin-mem.c
493
OPT_BOOLEAN(0, "data-page-size", &mem.data_page_size, "Record/Report sample data address page size"),
tools/perf/builtin-mem.c
497
OPT_CALLBACK('e', "event", &mem, "event",
tools/perf/builtin-mem.c
501
OPT_BOOLEAN('U', "all-user", &mem.all_user, "collect only user level data"),
tools/perf/builtin-mem.c
502
OPT_BOOLEAN('K', "all-kernel", &mem.all_kernel, "collect only kernel level data"),
tools/perf/builtin-mem.c
506
OPT_BOOLEAN('D', "dump-raw-samples", &mem.dump_raw,
tools/perf/builtin-mem.c
508
OPT_BOOLEAN('U', "hide-unresolved", &mem.hide_unresolved,
tools/perf/builtin-mem.c
516
OPT_STRING('s', "sort", &mem.sort_key, "key[,key2...]",
tools/perf/builtin-mem.c
518
OPT_BOOLEAN('T', "type-profile", &mem.data_type,
tools/perf/builtin-mem.c
52
struct perf_mem *mem = (struct perf_mem *)opt->value;
tools/perf/builtin-mem.c
532
if (!argc || !(strncmp(argv[0], "rec", 3) || mem.operation))
tools/perf/builtin-mem.c
535
if (!mem.input_name || !strlen(mem.input_name)) {
tools/perf/builtin-mem.c
537
mem.input_name = "-";
tools/perf/builtin-mem.c
539
mem.input_name = "perf.data";
tools/perf/builtin-mem.c
543
ret = __cmd_record(argc, argv, &mem, record_options);
tools/perf/builtin-mem.c
545
ret = __cmd_report(argc, argv, &mem, report_options);
tools/perf/builtin-mem.c
68
mem->operation = 0;
tools/perf/builtin-mem.c
72
static int __cmd_record(int argc, const char **argv, struct perf_mem *mem,
tools/perf/tests/mem.c
63
DEFINE_SUITE("Test data source output", mem);
tools/perf/tests/tests.h
128
DECLARE_SUITE(mem);
tools/perf/util/arm-spe.c
1838
if (spe->synth_opts.mem) {
tools/perf/util/auxtrace.c
1472
synth_opts->mem = true;
tools/perf/util/auxtrace.c
1716
synth_opts->mem = true;
tools/perf/util/auxtrace.h
144
bool mem;
tools/perf/util/capstone.c
273
if (op->mem.base != X86_REG_RIP)
tools/perf/util/capstone.c
277
orig_addr = addr + insn->size + op->mem.disp;
tools/perf/util/cpumap.c
577
unsigned int cpu, mem;
tools/perf/util/cpumap.c
603
if (dent1->d_type != DT_DIR || sscanf(dent1->d_name, "node%u", &mem) < 1)
tools/perf/util/cpumap.c
618
cpunode_map[cpu] = mem;
tools/perf/util/cputopo.c
323
u64 mem;
tools/perf/util/cputopo.c
337
if (sscanf(buf, "%*s %*d %31s %"PRIu64, field, &mem) != 2)
tools/perf/util/cputopo.c
340
node->mem_total = mem;
tools/perf/util/cputopo.c
342
node->mem_free = mem;
tools/perf/util/debuginfo.c
157
GElf_Shdr mem, *shdr;
tools/perf/util/debuginfo.c
176
shdr = gelf_getshdr(scn, &mem);
tools/perf/util/header.c
1335
#define for_each_memory(mem, dir) \
tools/perf/util/header.c
1339
sscanf(ent->d_name, "memory%u", &mem) == 1)
tools/perf/util/header.c
1838
char *mem = argv_i;
tools/perf/util/header.c
1848
free(mem);
tools/perf/util/header.c
682
uint64_t mem;
tools/perf/util/header.c
694
n = sscanf(buf, "%*s %"PRIu64, &mem);
tools/perf/util/header.c
696
ret = do_write(ff, &mem, sizeof(mem));
tools/power/acpi/os_specific/service_layers/osunixxf.c
532
void *mem;
tools/power/acpi/os_specific/service_layers/osunixxf.c
534
mem = (void *)malloc((size_t) size);
tools/power/acpi/os_specific/service_layers/osunixxf.c
535
return (mem);
tools/power/acpi/os_specific/service_layers/osunixxf.c
553
void *mem;
tools/power/acpi/os_specific/service_layers/osunixxf.c
555
mem = (void *)calloc(1, (size_t) size);
tools/power/acpi/os_specific/service_layers/osunixxf.c
556
return (mem);
tools/power/acpi/os_specific/service_layers/osunixxf.c
572
void acpi_os_free(void *mem)
tools/power/acpi/os_specific/service_layers/osunixxf.c
575
free(mem);
tools/power/cpupower/debug/i386/dump_psb.c
152
char *mem = NULL;
tools/power/cpupower/debug/i386/dump_psb.c
182
mem = mmap(mem, 0x100000 - 0xc0000, PROT_READ, MAP_SHARED, fd, 0xc0000);
tools/power/cpupower/debug/i386/dump_psb.c
185
for (p = mem; p - mem < LEN; p+=16) {
tools/power/cpupower/debug/i386/dump_psb.c
192
munmap(mem, LEN);
tools/testing/memblock/tests/common.h
169
static inline void assert_mem_content(void *mem, int size, int flags)
tools/testing/memblock/tests/common.h
172
ASSERT_MEM_NE(mem, 0, size);
tools/testing/memblock/tests/common.h
174
ASSERT_MEM_EQ(mem, 0, size);
tools/testing/selftests/arm64/abi/hwcap.c
163
u64 __attribute__ ((aligned (16))) mem[2] = { 10, 20 };
tools/testing/selftests/arm64/abi/hwcap.c
164
register u64 *memp asm ("x0") = mem;
tools/testing/selftests/arm64/abi/hwcap.c
177
float __attribute__ ((aligned (16))) mem;
tools/testing/selftests/arm64/abi/hwcap.c
178
register float *memp asm ("x0") = &mem;
tools/testing/selftests/bpf/benchs/bench_htab_mem.c
300
double loop, mem;
tools/testing/selftests/bpf/benchs/bench_htab_mem.c
303
mem = res->gp_ct / 1048576.0;
tools/testing/selftests/bpf/benchs/bench_htab_mem.c
305
printf("per-prod-op %7.2lfk/s, memory usage %7.2lfMiB\n", loop, mem);
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
381
void *mem;
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
387
mem = mmap(NULL, page_size * 2, PROT_READ | PROT_WRITE,
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
389
if (!ASSERT_OK_PTR(mem, "mmap(page_size * 2)")) {
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
399
value.udata = mem + page_size - offsetof(struct user_data, b);
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
405
value.udata = mem + page_size - sizeof(struct user_data);
tools/testing/selftests/bpf/prog_tests/task_local_storage.c
412
munmap(mem, page_size * 2);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
136
struct my_struct *mem,
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
139
if (!mem)
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
143
mem->x +
tools/testing/selftests/bpf/progs/test_xdp_meta.c
13
#define ctx_ptr(ctx, mem) (void *)(unsigned long)ctx->mem
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
227
char mem[16];
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
239
subprog_untrusted((void *)mem);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
243
subprog_untrusted((void *)mem + offset);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
297
subprog_void_untrusted(mem);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
301
subprog_void_untrusted((void *)mem + offset);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
305
subprog_char_untrusted(mem + offset);
tools/testing/selftests/bpf/progs/verifier_global_ptr_args.c
306
subprog_enum_untrusted((void *)mem + offset);
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
79
__noinline int global_unsupp(const int *mem)
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
81
if (!mem)
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
83
return mem[100]; /* BOOM */
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
859
__bpf_kfunc void bpf_kfunc_call_test_mem_len_pass1(void *mem, int mem__sz)
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
863
__bpf_kfunc void bpf_kfunc_call_test_mem_len_fail1(void *mem, int len)
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
867
__bpf_kfunc void bpf_kfunc_call_test_mem_len_fail2(u64 *mem, int len)
tools/testing/selftests/bpf/test_kmods/bpf_testmod_kfunc.h
117
void bpf_kfunc_call_test_mem_len_fail2(__u64 *mem, int len) __ksym;
tools/testing/selftests/bpf/test_kmods/bpf_testmod_kfunc.h
129
void bpf_kfunc_call_test_mem_len_fail1(void *mem, int len);
tools/testing/selftests/bpf/test_kmods/bpf_testmod_kfunc.h
93
void bpf_kfunc_call_test_mem_len_pass1(void *mem, int len) __ksym;
tools/testing/selftests/bpf/trace_helpers.c
456
GElf_Phdr mem, *phdr;
tools/testing/selftests/bpf/trace_helpers.c
459
phdr = gelf_getphdr(elf, i, &mem);
tools/testing/selftests/cgroup/test_zswap.c
271
char *mem;
tools/testing/selftests/cgroup/test_zswap.c
273
mem = (char *)malloc(memsize);
tools/testing/selftests/cgroup/test_zswap.c
274
if (!mem)
tools/testing/selftests/cgroup/test_zswap.c
286
memcpy(&mem[i], buf, pagesize);
tools/testing/selftests/cgroup/test_zswap.c
298
if (memcmp(&mem[i], buf, pagesize)) {
tools/testing/selftests/cgroup/test_zswap.c
318
free(mem);
tools/testing/selftests/cgroup/test_zswap.c
66
char *mem = (char *)malloc(size);
tools/testing/selftests/cgroup/test_zswap.c
69
if (!mem)
tools/testing/selftests/cgroup/test_zswap.c
72
mem[i] = 'a';
tools/testing/selftests/cgroup/test_zswap.c
76
if (mem[i] != 'a')
tools/testing/selftests/cgroup/test_zswap.c
80
free(mem);
tools/testing/selftests/cgroup/test_zswap.c
87
char *mem = (char *)malloc(size);
tools/testing/selftests/cgroup/test_zswap.c
89
if (!mem)
tools/testing/selftests/cgroup/test_zswap.c
92
mem[i] = 'a';
tools/testing/selftests/cgroup/test_zswap.c
93
free(mem);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1009
provider->memcpy_from_device(tmp_mem, mem,
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1069
struct memory_buffer *mem;
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1073
mem = provider->alloc(getpagesize() * NUM_PAGES);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1074
if (!mem) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1102
if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1118
if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1134
if (bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1157
provider->free(mem);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1251
static int do_client(struct memory_buffer *mem)
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1298
if (bind_tx_queue(ifindex, mem->fd, &ys)) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1357
provider->memcpy_to_device(mem, 0, line, line_size);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1403
struct memory_buffer *mem;
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1519
mem = provider->alloc(getpagesize() * NUM_PAGES);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1520
if (!mem) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1525
ret = is_server ? do_server(mem) : do_client(mem);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
1532
provider->free(mem);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
829
static int do_server(struct memory_buffer *mem)
tools/testing/selftests/drivers/net/hw/ncdevmem.c
877
if (bind_rx_queue(ifindex, mem->fd, create_queues(), num_queues, &ys)) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
882
tmp_mem = malloc(mem->size);
tools/testing/selftests/kexec/test_kexec_jump.c
50
segment.mem = (void *)0x400000;
tools/testing/selftests/kvm/guest_memfd_test.c
101
kvm_mbind(mem + page_size * 2, page_size * 2, MPOL_BIND, &nodemask_0, maxnode, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
102
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem + page_size * 2, MPOL_F_ADDR);
tools/testing/selftests/kvm/guest_memfd_test.c
108
kvm_mbind(mem, total_size, MPOL_DEFAULT, NULL, 0, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
109
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR);
tools/testing/selftests/kvm/guest_memfd_test.c
115
ret = mbind(mem, page_size, 999, &nodemask_0, maxnode, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
119
kvm_munmap(mem, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
129
char *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
135
mem = kvm_mmap(total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
138
pages[i] = (char *)mem + page_size * i;
tools/testing/selftests/kvm/guest_memfd_test.c
141
memset(mem, 0xaa, page_size);
tools/testing/selftests/kvm/guest_memfd_test.c
148
memset(mem, 0xaa, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
163
memset(mem, 0xaa, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
171
kvm_munmap(mem, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
177
char *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
180
mem = kvm_mmap(map_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
182
TEST_EXPECT_SIGBUS(memset(mem, val, map_size));
tools/testing/selftests/kvm/guest_memfd_test.c
183
TEST_EXPECT_SIGBUS((void)READ_ONCE(mem[accessible_size]));
tools/testing/selftests/kvm/guest_memfd_test.c
186
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
188
kvm_munmap(mem, map_size);
tools/testing/selftests/kvm/guest_memfd_test.c
203
char *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
205
mem = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
206
TEST_ASSERT_EQ(mem, MAP_FAILED);
tools/testing/selftests/kvm/guest_memfd_test.c
208
mem = mmap(NULL, total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
209
TEST_ASSERT_EQ(mem, MAP_FAILED);
tools/testing/selftests/kvm/guest_memfd_test.c
409
static void guest_code(uint8_t *mem, uint64_t size)
tools/testing/selftests/kvm/guest_memfd_test.c
414
__GUEST_ASSERT(mem[i] == 0xaa,
tools/testing/selftests/kvm/guest_memfd_test.c
415
"Guest expected 0xaa at offset %lu, got 0x%x", i, mem[i]);
tools/testing/selftests/kvm/guest_memfd_test.c
417
memset(mem, 0xff, size);
tools/testing/selftests/kvm/guest_memfd_test.c
433
uint8_t *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
44
void *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
454
mem = kvm_mmap(size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
455
memset(mem, 0xaa, size);
tools/testing/selftests/kvm/guest_memfd_test.c
456
kvm_munmap(mem, size);
tools/testing/selftests/kvm/guest_memfd_test.c
46
mem = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
464
mem = kvm_mmap(size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
466
TEST_ASSERT_EQ(mem[i], 0xff);
tools/testing/selftests/kvm/guest_memfd_test.c
47
TEST_ASSERT(mem == MAP_FAILED, "Copy-on-write not allowed by guest_memfd.");
tools/testing/selftests/kvm/guest_memfd_test.c
53
char *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
57
mem = kvm_mmap(total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
59
memset(mem, val, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
61
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
68
TEST_ASSERT_EQ(READ_ONCE(mem[i]), 0x00);
tools/testing/selftests/kvm/guest_memfd_test.c
70
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
72
memset(mem, val, page_size);
tools/testing/selftests/kvm/guest_memfd_test.c
74
TEST_ASSERT_EQ(READ_ONCE(mem[i]), val);
tools/testing/selftests/kvm/guest_memfd_test.c
76
kvm_munmap(mem, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
85
char *mem;
tools/testing/selftests/kvm/guest_memfd_test.c
91
mem = kvm_mmap(total_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/guest_memfd_test.c
94
kvm_mbind(mem, page_size * 2, MPOL_INTERLEAVE, &nodemask_0, maxnode, 0);
tools/testing/selftests/kvm/guest_memfd_test.c
95
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR);
tools/testing/selftests/kvm/include/kvm_syscalls.h
55
void *mem;
tools/testing/selftests/kvm/include/kvm_syscalls.h
57
mem = mmap(NULL, size, prot, flags, fd, offset);
tools/testing/selftests/kvm/include/kvm_syscalls.h
58
TEST_ASSERT(mem != MAP_FAILED, __KVM_SYSCALL_ERROR("mmap()",
tools/testing/selftests/kvm/include/kvm_syscalls.h
60
return mem;
tools/testing/selftests/kvm/include/kvm_syscalls.h
76
__KVM_SYSCALL_DEFINE(munmap, 2, void *, mem, size_t, size);
tools/testing/selftests/kvm/include/kvm_util.h
739
#define vcpu_arch_put_guest(mem, val) do { (mem) = (val); } while (0)
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
54
#define vcpu_arch_put_guest(mem, __val) \
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
56
const typeof(mem) val = (__val); \
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
59
(mem) = val; \
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
62
: "+m" (mem) \
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
65
uint64_t __old = READ_ONCE(mem); \
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
68
: [ptr] "+m" (mem), [old] "+a" (__old) \
tools/testing/selftests/kvm/mmu_stress_test.c
291
void *mem;
tools/testing/selftests/kvm/mmu_stress_test.c
344
mem = kvm_mmap(slot_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd);
tools/testing/selftests/kvm/mmu_stress_test.c
346
TEST_ASSERT(!madvise(mem, slot_size, MADV_NOHUGEPAGE), "madvise() failed");
tools/testing/selftests/kvm/mmu_stress_test.c
350
((uint8_t *)mem)[i] = 0xaa;
tools/testing/selftests/kvm/mmu_stress_test.c
361
vm_set_user_memory_region(vm, slot, 0, gpa, slot_size, mem);
tools/testing/selftests/kvm/mmu_stress_test.c
385
mprotect(mem, slot_size, PROT_READ);
tools/testing/selftests/kvm/mmu_stress_test.c
390
mprotect(mem, slot_size, PROT_READ | PROT_WRITE);
tools/testing/selftests/kvm/mmu_stress_test.c
415
kvm_munmap(mem, slot_size / 2);
tools/testing/selftests/kvm/s390/irq_routing.c
30
vm_paddr_t mem;
tools/testing/selftests/kvm/s390/irq_routing.c
39
mem = vm_phy_pages_alloc(vm, 2, 4096 * 42, 0);
tools/testing/selftests/kvm/s390/irq_routing.c
44
routing->entries[0].u.adapter.summary_addr = (uintptr_t)mem;
tools/testing/selftests/kvm/s390/irq_routing.c
45
routing->entries[0].u.adapter.ind_addr = (uintptr_t)mem;
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
50
char *mem, *page0, *page1, *page2, tmp;
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
64
mem = mmap(0, 3 * pagesize, PROT_READ, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
65
TEST_ASSERT(mem != MAP_FAILED, "mmap() failed");
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
68
madvise(mem, 3 * pagesize, MADV_NOHUGEPAGE);
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
70
page0 = mem;
tools/testing/selftests/kvm/set_memory_region_test.c
415
void *mem, *mem_aligned, *mem_extra;
tools/testing/selftests/kvm/set_memory_region_test.c
437
mem = kvm_mmap((size_t)max_mem_slots * MEM_REGION_SIZE + alignment,
tools/testing/selftests/kvm/set_memory_region_test.c
440
mem_aligned = (void *)(((size_t) mem + alignment - 1) & ~(alignment - 1));
tools/testing/selftests/kvm/set_memory_region_test.c
458
kvm_munmap(mem, (size_t)max_mem_slots * MEM_REGION_SIZE + alignment);
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
32
uint8_t *mem = (uint8_t *)gpa; \
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
36
__GUEST_ASSERT(mem[i] == pattern, \
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
38
pattern, i, gpa + i, mem[i]); \
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
41
static void memcmp_h(uint8_t *mem, uint64_t gpa, uint8_t pattern, size_t size)
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
46
TEST_ASSERT(mem[i] == pattern,
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
48
pattern, gpa + i, mem[i]);
tools/testing/selftests/mm/cow.c
1055
char *mem, *dummy;
tools/testing/selftests/mm/cow.c
1062
mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, flags, -1, 0);
tools/testing/selftests/mm/cow.c
1063
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
1070
memset(mem, 1, hugetlbsize);
tools/testing/selftests/mm/cow.c
1084
fn(mem, hugetlbsize, true);
tools/testing/selftests/mm/cow.c
1086
munmap(mem, hugetlbsize);
tools/testing/selftests/mm/cow.c
119
static int child_memcmp_fn(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
126
memcpy(old, mem, size);
tools/testing/selftests/mm/cow.c
1276
static void do_test_anon_thp_collapse(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1293
ret = mprotect(mem + pagesize, pagesize, PROT_READ);
tools/testing/selftests/mm/cow.c
1299
ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
tools/testing/selftests/mm/cow.c
1309
ret = madvise(mem, size, MADV_COLLAPSE);
tools/testing/selftests/mm/cow.c
1321
ret = madvise(mem + size / 2, size / 2, MADV_DONTFORK);
tools/testing/selftests/mm/cow.c
1330
ret = madvise(mem, size / 2, MADV_DONTFORK);
tools/testing/selftests/mm/cow.c
134
return memcmp(old, mem, size);
tools/testing/selftests/mm/cow.c
1350
exit(child_memcmp_fn(mem, size, &comm_pipes));
tools/testing/selftests/mm/cow.c
1353
exit(child_memcmp_fn(mem, size / 2, &comm_pipes));
tools/testing/selftests/mm/cow.c
1356
exit(child_memcmp_fn(mem + size / 2, size / 2,
tools/testing/selftests/mm/cow.c
137
static int child_vmsplice_memcmp_fn(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1376
ret = madvise(mem, size, MADV_DOFORK);
tools/testing/selftests/mm/cow.c
1387
ret = madvise(mem, size, MADV_COLLAPSE);
tools/testing/selftests/mm/cow.c
1401
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
141
.iov_base = mem,
tools/testing/selftests/mm/cow.c
1420
static void test_anon_thp_collapse_unshared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1424
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UNSHARED);
tools/testing/selftests/mm/cow.c
1427
static void test_anon_thp_collapse_fully_shared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1431
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_FULLY_SHARED);
tools/testing/selftests/mm/cow.c
1434
static void test_anon_thp_collapse_lower_shared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1438
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_LOWER_SHARED);
tools/testing/selftests/mm/cow.c
1441
static void test_anon_thp_collapse_upper_shared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
1445
do_test_anon_thp_collapse(mem, size, ANON_THP_COLLAPSE_UPPER_SHARED);
tools/testing/selftests/mm/cow.c
1510
typedef void (*non_anon_test_fn)(char *mem, const char *smem, size_t size);
tools/testing/selftests/mm/cow.c
1512
static void test_cow(char *mem, const char *smem, size_t size)
tools/testing/selftests/mm/cow.c
1520
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
153
memcpy(old, mem, size);
tools/testing/selftests/mm/cow.c
1532
static void test_ro_pin(char *mem, const char *smem, size_t size)
tools/testing/selftests/mm/cow.c
1534
do_test_ro_pin(mem, size, RO_PIN_TEST, false);
tools/testing/selftests/mm/cow.c
1537
static void test_ro_fast_pin(char *mem, const char *smem, size_t size)
tools/testing/selftests/mm/cow.c
1539
do_test_ro_pin(mem, size, RO_PIN_TEST, true);
tools/testing/selftests/mm/cow.c
1544
char *mem, *smem;
tools/testing/selftests/mm/cow.c
1548
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/cow.c
1550
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
1564
if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
tools/testing/selftests/mm/cow.c
1569
fn(mem, smem, pagesize);
tools/testing/selftests/mm/cow.c
1571
munmap(mem, pagesize);
tools/testing/selftests/mm/cow.c
1578
char *mem, *smem, *mmap_mem, *mmap_smem;
tools/testing/selftests/mm/cow.c
1608
mem = (char *)(((uintptr_t)mmap_mem + pmdsize) & ~(pmdsize - 1));
tools/testing/selftests/mm/cow.c
1611
ret = madvise(mem, pmdsize, MADV_HUGEPAGE);
tools/testing/selftests/mm/cow.c
1629
if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
tools/testing/selftests/mm/cow.c
1634
if (!pagemap_is_populated(pagemap_fd, mem + pagesize) ||
tools/testing/selftests/mm/cow.c
1640
fn(mem, smem, pmdsize);
tools/testing/selftests/mm/cow.c
1649
char *mem, *smem;
tools/testing/selftests/mm/cow.c
166
if (munmap(mem, size) < 0)
tools/testing/selftests/mm/cow.c
1669
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
tools/testing/selftests/mm/cow.c
1670
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
1683
if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
tools/testing/selftests/mm/cow.c
1688
fn(mem, smem, pagesize);
tools/testing/selftests/mm/cow.c
1690
munmap(mem, pagesize);
tools/testing/selftests/mm/cow.c
1699
char *mem, *smem;
tools/testing/selftests/mm/cow.c
1727
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd, 0);
tools/testing/selftests/mm/cow.c
1728
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
1741
if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
tools/testing/selftests/mm/cow.c
1746
fn(mem, smem, pagesize);
tools/testing/selftests/mm/cow.c
1748
munmap(mem, pagesize);
tools/testing/selftests/mm/cow.c
1759
char *mem, *smem;
tools/testing/selftests/mm/cow.c
1782
mem = mmap(NULL, hugetlbsize, PROT_READ | PROT_WRITE, MAP_PRIVATE, fd,
tools/testing/selftests/mm/cow.c
1784
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
1797
if (!populate_page_checked(mem) || !populate_page_checked(smem)) {
tools/testing/selftests/mm/cow.c
1802
fn(mem, smem, hugetlbsize);
tools/testing/selftests/mm/cow.c
1804
munmap(mem, hugetlbsize);
tools/testing/selftests/mm/cow.c
184
typedef int (*child_fn)(char *mem, size_t size, struct comm_pipes *comm_pipes);
tools/testing/selftests/mm/cow.c
186
static void do_test_cow_in_parent(char *mem, size_t size, bool do_mprotect,
tools/testing/selftests/mm/cow.c
205
exit(fn(mem, size, &comm_pipes));
tools/testing/selftests/mm/cow.c
216
ret = mprotect(mem, size, PROT_READ);
tools/testing/selftests/mm/cow.c
225
ret = mprotect(mem, size, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/cow.c
236
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
263
static void test_cow_in_parent(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
265
do_test_cow_in_parent(mem, size, false, child_memcmp_fn, false);
tools/testing/selftests/mm/cow.c
268
static void test_cow_in_parent_mprotect(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
270
do_test_cow_in_parent(mem, size, true, child_memcmp_fn, false);
tools/testing/selftests/mm/cow.c
273
static void test_vmsplice_in_child(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
275
do_test_cow_in_parent(mem, size, false, child_vmsplice_memcmp_fn,
tools/testing/selftests/mm/cow.c
279
static void test_vmsplice_in_child_mprotect(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
282
do_test_cow_in_parent(mem, size, true, child_vmsplice_memcmp_fn,
tools/testing/selftests/mm/cow.c
286
static void do_test_vmsplice_in_parent(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
290
.iov_base = mem,
tools/testing/selftests/mm/cow.c
302
memcpy(old, mem, size);
tools/testing/selftests/mm/cow.c
335
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
351
if (munmap(mem, size) < 0) {
tools/testing/selftests/mm/cow.c
400
static void test_vmsplice_before_fork(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
402
do_test_vmsplice_in_parent(mem, size, true, is_hugetlb);
tools/testing/selftests/mm/cow.c
405
static void test_vmsplice_after_fork(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
407
do_test_vmsplice_in_parent(mem, size, false, is_hugetlb);
tools/testing/selftests/mm/cow.c
411
static void do_test_iouring(char *mem, size_t size, bool use_fork)
tools/testing/selftests/mm/cow.c
460
iov.iov_base = mem;
tools/testing/selftests/mm/cow.c
496
ret = mprotect(mem, size, PROT_READ);
tools/testing/selftests/mm/cow.c
504
ret = mprotect(mem, size, PROT_READ | PROT_WRITE);
tools/testing/selftests/mm/cow.c
516
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
523
io_uring_prep_write_fixed(sqe, fd, mem, size, 0, 0);
tools/testing/selftests/mm/cow.c
559
if (!memcmp(mem, tmp, size)) {
tools/testing/selftests/mm/cow.c
583
static void test_iouring_ro(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
585
do_test_iouring(mem, size, false);
tools/testing/selftests/mm/cow.c
588
static void test_iouring_fork(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
590
do_test_iouring(mem, size, true);
tools/testing/selftests/mm/cow.c
602
static void do_test_ro_pin(char *mem, size_t size, enum ro_pin_test test,
tools/testing/selftests/mm/cow.c
675
ret = mprotect(mem, size, PROT_READ);
tools/testing/selftests/mm/cow.c
677
ret |= mprotect(mem, size, PROT_READ | PROT_WRITE);
tools/testing/selftests/mm/cow.c
689
args.addr = (__u64)(uintptr_t)mem;
tools/testing/selftests/mm/cow.c
704
memset(mem, 0xff, size);
tools/testing/selftests/mm/cow.c
716
if (!memcmp(mem, tmp, size)) {
tools/testing/selftests/mm/cow.c
744
static void test_ro_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
746
do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, false);
tools/testing/selftests/mm/cow.c
749
static void test_ro_fast_pin_on_shared(char *mem, size_t size, bool is_hugetlb)
tools/testing/selftests/mm/cow.c
751
do_test_ro_pin(mem, size, RO_PIN_TEST_SHARED, true);
tools/testing/selftests/mm/cow.c
754
static void test_ro_pin_on_ro_previously_shared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
757
do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, false);
tools/testing/selftests/mm/cow.c
760
static void test_ro_fast_pin_on_ro_previously_shared(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
763
do_test_ro_pin(mem, size, RO_PIN_TEST_PREVIOUSLY_SHARED, true);
tools/testing/selftests/mm/cow.c
766
static void test_ro_pin_on_ro_exclusive(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
769
do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, false);
tools/testing/selftests/mm/cow.c
772
static void test_ro_fast_pin_on_ro_exclusive(char *mem, size_t size,
tools/testing/selftests/mm/cow.c
775
do_test_ro_pin(mem, size, RO_PIN_TEST_RO_EXCLUSIVE, true);
tools/testing/selftests/mm/cow.c
778
typedef void (*test_fn)(char *mem, size_t size, bool hugetlb);
tools/testing/selftests/mm/cow.c
782
char *mem;
tools/testing/selftests/mm/cow.c
785
mem = mmap(NULL, pagesize, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/cow.c
787
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/cow.c
793
ret = madvise(mem, pagesize, MADV_NOHUGEPAGE);
tools/testing/selftests/mm/cow.c
802
memset(mem, 1, pagesize);
tools/testing/selftests/mm/cow.c
805
madvise(mem, pagesize, MADV_PAGEOUT);
tools/testing/selftests/mm/cow.c
806
if (!pagemap_is_swapped(pagemap_fd, mem)) {
tools/testing/selftests/mm/cow.c
813
fn(mem, pagesize, false);
tools/testing/selftests/mm/cow.c
815
munmap(mem, pagesize);
tools/testing/selftests/mm/cow.c
843
char *mem, *mmap_mem, *tmp, *mremap_mem = MAP_FAILED;
tools/testing/selftests/mm/cow.c
858
mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1));
tools/testing/selftests/mm/cow.c
860
ret = madvise(mem, thpsize, MADV_HUGEPAGE);
tools/testing/selftests/mm/cow.c
871
mem[0] = 1;
tools/testing/selftests/mm/cow.c
872
if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
tools/testing/selftests/mm/cow.c
877
memset(mem, 1, thpsize);
tools/testing/selftests/mm/cow.c
892
ret = mprotect(mem + pagesize, pagesize, PROT_READ);
tools/testing/selftests/mm/cow.c
898
ret = mprotect(mem + pagesize, pagesize, PROT_READ | PROT_WRITE);
tools/testing/selftests/mm/cow.c
911
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTNEED);
tools/testing/selftests/mm/cow.c
932
tmp = mremap(mem + mremap_size, mremap_size, mremap_size,
tools/testing/selftests/mm/cow.c
947
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DONTFORK);
tools/testing/selftests/mm/cow.c
963
ret = madvise(mem + pagesize, thpsize - pagesize, MADV_DOFORK);
tools/testing/selftests/mm/cow.c
978
madvise(mem, size, MADV_PAGEOUT);
tools/testing/selftests/mm/cow.c
979
if (!range_is_swapped(mem, size)) {
tools/testing/selftests/mm/cow.c
989
fn(mem, size, false);
tools/testing/selftests/mm/gup_longterm.c
135
mem = mmap(NULL, size, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/gup_longterm.c
137
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/gup_longterm.c
149
memset(mem, 0, size);
tools/testing/selftests/mm/gup_longterm.c
158
ret = mprotect(mem, size, PROT_READ);
tools/testing/selftests/mm/gup_longterm.c
192
args.addr = (__u64)(uintptr_t)mem;
tools/testing/selftests/mm/gup_longterm.c
255
iov.iov_base = mem;
tools/testing/selftests/mm/gup_longterm.c
295
munmap(mem, size);
tools/testing/selftests/mm/gup_longterm.c
95
char *mem;
tools/testing/selftests/mm/memfd_secret.c
100
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/memfd_secret.c
109
memset(mem, PATTERN, page_size);
tools/testing/selftests/mm/memfd_secret.c
111
iov.iov_base = mem;
tools/testing/selftests/mm/memfd_secret.c
120
munmap(mem, page_size);
tools/testing/selftests/mm/memfd_secret.c
130
char *mem;
tools/testing/selftests/mm/memfd_secret.c
132
if (read(pipefd[0], &mem, sizeof(mem)) < 0) {
tools/testing/selftests/mm/memfd_secret.c
139
riov.iov_base = mem;
tools/testing/selftests/mm/memfd_secret.c
154
char *mem;
tools/testing/selftests/mm/memfd_secret.c
157
if (read(pipefd[0], &mem, sizeof(mem)) < 0) {
tools/testing/selftests/mm/memfd_secret.c
175
if (ptrace(PTRACE_PEEKDATA, ppid, mem, 0))
tools/testing/selftests/mm/memfd_secret.c
206
char *mem;
tools/testing/selftests/mm/memfd_secret.c
224
mem = mmap(NULL, page_size, prot, mode, fd, 0);
tools/testing/selftests/mm/memfd_secret.c
225
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/memfd_secret.c
230
memset(mem, PATTERN, page_size);
tools/testing/selftests/mm/memfd_secret.c
232
if (write(pipefd[1], &mem, sizeof(mem)) < 0) {
tools/testing/selftests/mm/memfd_secret.c
63
char *mem;
tools/testing/selftests/mm/memfd_secret.c
69
mem = mmap(NULL, len, prot, mode, fd, 0);
tools/testing/selftests/mm/memfd_secret.c
70
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/memfd_secret.c
74
munmap(mem, len);
tools/testing/selftests/mm/memfd_secret.c
77
mem = mmap(NULL, len, prot, mode, fd, 0);
tools/testing/selftests/mm/memfd_secret.c
78
if (mem != MAP_FAILED) {
tools/testing/selftests/mm/memfd_secret.c
80
munmap(mem, len);
tools/testing/selftests/mm/memfd_secret.c
92
char *mem;
tools/testing/selftests/mm/memfd_secret.c
99
mem = mmap(NULL, page_size, prot, mode, fd, 0);
tools/testing/selftests/mm/mkdirty.c
101
if (*mem != 0) {
tools/testing/selftests/mm/mkdirty.c
110
lseek(mem_fd, (uintptr_t) mem, SEEK_SET);
tools/testing/selftests/mm/mkdirty.c
112
if (ret != 1 || *mem != data) {
tools/testing/selftests/mm/mkdirty.c
117
do_test_write_sigsegv(mem);
tools/testing/selftests/mm/mkdirty.c
119
munmap(mem, pagesize);
tools/testing/selftests/mm/mkdirty.c
124
char *mem, *mmap_mem;
tools/testing/selftests/mm/mkdirty.c
131
mem = mmap_thp_range(PROT_READ, &mmap_mem, &mmap_size);
tools/testing/selftests/mm/mkdirty.c
132
if (mem == MAP_FAILED)
tools/testing/selftests/mm/mkdirty.c
140
lseek(mem_fd, (uintptr_t) mem, SEEK_SET);
tools/testing/selftests/mm/mkdirty.c
142
if (ret != 1 || *mem != data) {
tools/testing/selftests/mm/mkdirty.c
148
if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
tools/testing/selftests/mm/mkdirty.c
153
do_test_write_sigsegv(mem);
tools/testing/selftests/mm/mkdirty.c
160
char *mem;
tools/testing/selftests/mm/mkdirty.c
164
mem = mmap(NULL, pagesize, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANON,
tools/testing/selftests/mm/mkdirty.c
166
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/mkdirty.c
172
memset(mem, 1, pagesize);
tools/testing/selftests/mm/mkdirty.c
173
if (mprotect(mem, pagesize, PROT_READ)) {
tools/testing/selftests/mm/mkdirty.c
179
if (syscall(__NR_mbind, mem, pagesize, MPOL_LOCAL, NULL, 0x7fful,
tools/testing/selftests/mm/mkdirty.c
185
do_test_write_sigsegv(mem);
tools/testing/selftests/mm/mkdirty.c
187
munmap(mem, pagesize);
tools/testing/selftests/mm/mkdirty.c
192
char *mem, *mmap_mem;
tools/testing/selftests/mm/mkdirty.c
197
mem = mmap_thp_range(PROT_READ|PROT_WRITE, &mmap_mem, &mmap_size);
tools/testing/selftests/mm/mkdirty.c
198
if (mem == MAP_FAILED)
tools/testing/selftests/mm/mkdirty.c
205
memset(mem, 1, pagesize);
tools/testing/selftests/mm/mkdirty.c
206
if (mprotect(mem, thpsize, PROT_READ)) {
tools/testing/selftests/mm/mkdirty.c
212
if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
tools/testing/selftests/mm/mkdirty.c
218
if (syscall(__NR_mbind, mem, thpsize, MPOL_LOCAL, NULL, 0x7fful,
tools/testing/selftests/mm/mkdirty.c
224
do_test_write_sigsegv(mem);
tools/testing/selftests/mm/mkdirty.c
231
char *mem, *mmap_mem;
tools/testing/selftests/mm/mkdirty.c
236
mem = mmap_thp_range(PROT_READ|PROT_WRITE, &mmap_mem, &mmap_size);
tools/testing/selftests/mm/mkdirty.c
237
if (mem == MAP_FAILED)
tools/testing/selftests/mm/mkdirty.c
244
memset(mem, 1, pagesize);
tools/testing/selftests/mm/mkdirty.c
245
if (mprotect(mem, thpsize, PROT_READ)) {
tools/testing/selftests/mm/mkdirty.c
251
if (!pagemap_is_populated(pagemap_fd, mem + thpsize - pagesize)) {
tools/testing/selftests/mm/mkdirty.c
257
if (mprotect(mem + thpsize - pagesize, pagesize,
tools/testing/selftests/mm/mkdirty.c
263
do_test_write_sigsegv(mem);
tools/testing/selftests/mm/mkdirty.c
41
static void do_test_write_sigsegv(char *mem)
tools/testing/selftests/mm/mkdirty.c
43
char orig = *mem;
tools/testing/selftests/mm/mkdirty.c
53
*mem = orig + 1;
tools/testing/selftests/mm/mkdirty.c
58
ksft_test_result(ret == 1 && *mem == orig,
tools/testing/selftests/mm/mkdirty.c
65
char *mem, *mmap_mem;
tools/testing/selftests/mm/mkdirty.c
73
mem = (char *)(((uintptr_t)mmap_mem + thpsize) & ~(thpsize - 1));
tools/testing/selftests/mm/mkdirty.c
75
if (madvise(mem, thpsize, MADV_HUGEPAGE)) {
tools/testing/selftests/mm/mkdirty.c
83
return mem;
tools/testing/selftests/mm/mkdirty.c
89
char *mem;
tools/testing/selftests/mm/mkdirty.c
94
mem = mmap(NULL, pagesize, PROT_READ, MAP_PRIVATE|MAP_ANON, -1, 0);
tools/testing/selftests/mm/mkdirty.c
95
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/pagemap_ioctl.c
1057
char *mem, *fmem;
tools/testing/selftests/mm/pagemap_ioctl.c
1065
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1066
if (mem == MAP_FAILED || vec == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1069
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1070
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1072
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
1076
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1079
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1082
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1085
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1089
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1090
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1093
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1094
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1096
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1097
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1099
memset(mem, 0, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1101
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1103
ksft_test_result(ret >= 0 && vec[0].start == (uintptr_t)mem && LEN(vec[0]) == vec_size &&
tools/testing/selftests/mm/pagemap_ioctl.c
1109
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1111
ksft_test_result(ret >= 0 && vec[0].start == (uintptr_t)mem && LEN(vec[0]) == vec_size &&
tools/testing/selftests/mm/pagemap_ioctl.c
1117
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1119
ksft_test_result(ret >= 0 && vec[0].start == (uintptr_t)mem && LEN(vec[0]) == vec_size &&
tools/testing/selftests/mm/pagemap_ioctl.c
1126
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1131
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1135
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1136
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1139
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1140
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1142
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1143
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1145
memset(mem, 0, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1147
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
1150
ksft_test_result(ret >= 0 && vec[0].start == (uintptr_t)mem && LEN(vec[0]) == vec_size &&
tools/testing/selftests/mm/pagemap_ioctl.c
1153
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1154
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1227
char *mem, *mem2;
tools/testing/selftests/mm/pagemap_ioctl.c
1237
mem = mmap(0, 2 * page_size, PROT_READ|PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1238
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1240
wp_init(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1241
wp_addr_range(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1244
memset(mem, 1, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1246
ret = pagemap_ioctl(mem, 2 * page_size, &vec, 1, 0, 0, PAGE_IS_WRITTEN,
tools/testing/selftests/mm/pagemap_ioctl.c
1254
wp_addr_range(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1256
ksft_test_result(pagemap_ioctl(mem, 2 * page_size, &vec, 1, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1261
mem2 = mmap(mem + page_size, page_size, PROT_READ|PROT_WRITE,
tools/testing/selftests/mm/pagemap_ioctl.c
1269
mprotect(mem, page_size, PROT_NONE);
tools/testing/selftests/mm/pagemap_ioctl.c
1270
mprotect(mem, 2 * page_size, PROT_READ);
tools/testing/selftests/mm/pagemap_ioctl.c
1271
mprotect(mem, 2 * page_size, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
1274
memset(mem, 2, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1277
mprotect(mem, page_size, PROT_NONE);
tools/testing/selftests/mm/pagemap_ioctl.c
1278
mprotect(mem, page_size, PROT_READ);
tools/testing/selftests/mm/pagemap_ioctl.c
1279
mprotect(mem, page_size, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
1281
ret = pagemap_ioctl(mem, 2 * page_size, &vec, 1, 0, 0, PAGE_IS_WRITTEN,
tools/testing/selftests/mm/pagemap_ioctl.c
1290
wp_addr_range(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1292
memset(mem, 'A', 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1294
ret = pagemap_ioctl(mem, 2 * page_size, &vec, 1, 0, 0, PAGE_IS_WRITTEN,
tools/testing/selftests/mm/pagemap_ioctl.c
1302
wp_free(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1303
munmap(mem, 2 * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1314
static ssize_t get_dirty_pages_reset(char *mem, unsigned int count,
tools/testing/selftests/mm/pagemap_ioctl.c
1324
arg.start = (uintptr_t)mem;
tools/testing/selftests/mm/pagemap_ioctl.c
1326
arg.end = (uintptr_t)(mem + count * page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1350
void *thread_proc(void *mem)
tools/testing/selftests/mm/pagemap_ioctl.c
1352
int *m = mem;
tools/testing/selftests/mm/pagemap_ioctl.c
1395
char *mem;
tools/testing/selftests/mm/pagemap_ioctl.c
1404
mem = mmap(NULL, 0x1000 * nthreads * pages_per_thread, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/pagemap_ioctl.c
1406
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1409
wp_init(mem, 0x1000 * nthreads * pages_per_thread);
tools/testing/selftests/mm/pagemap_ioctl.c
1410
wp_addr_range(mem, 0x1000 * nthreads * pages_per_thread);
tools/testing/selftests/mm/pagemap_ioctl.c
1412
memset(mem, 0, 0x1000 * nthreads * pages_per_thread);
tools/testing/selftests/mm/pagemap_ioctl.c
1414
count = get_dirty_pages_reset(mem, nthreads * pages_per_thread, 1, page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1416
count = get_dirty_pages_reset(mem, nthreads * pages_per_thread, 1, page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1421
pthread_create(&th, NULL, thread_proc, mem + 0x1000 * i * pages_per_thread);
tools/testing/selftests/mm/pagemap_ioctl.c
1431
count = get_dirty_pages_reset(mem, nthreads * pages_per_thread, 1,
tools/testing/selftests/mm/pagemap_ioctl.c
1442
c = get_dirty_pages_reset(mem, nthreads * pages_per_thread, 1, page_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1488
char *mmap_mem, *mem;
tools/testing/selftests/mm/pagemap_ioctl.c
1492
mem = mmap(NULL, mem_size, PROT_READ, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1493
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1498
(void)((volatile char *)mem)[i];
tools/testing/selftests/mm/pagemap_ioctl.c
1500
ret = pagemap_ioctl(mem, mem_size, &vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1508
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1523
mem = (char *)(((uintptr_t)mmap_mem + hpage_size) & ~(hpage_size - 1));
tools/testing/selftests/mm/pagemap_ioctl.c
1525
ret = madvise(mem, hpage_size, MADV_HUGEPAGE);
tools/testing/selftests/mm/pagemap_ioctl.c
1527
FORCE_READ(*mem);
tools/testing/selftests/mm/pagemap_ioctl.c
1529
ret = pagemap_ioctl(mem, hpage_size, &vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
1547
char *mem, *map, *fmem;
tools/testing/selftests/mm/pagemap_ioctl.c
1571
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1572
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1574
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1575
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1577
base_tests("Page testing:", mem, mem_size, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1579
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1580
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1584
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1585
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
1587
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1588
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1590
base_tests("Large Page testing:", mem, mem_size, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1592
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1593
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1609
mem = gethugetlb_mem(mem_size, &shmid);
tools/testing/selftests/mm/pagemap_ioctl.c
1610
if (mem) {
tools/testing/selftests/mm/pagemap_ioctl.c
1611
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1612
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1614
base_tests("Hugetlb shmem testing:", mem, mem_size, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1616
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1623
mem = gethugetlb_mem(mem_size, NULL);
tools/testing/selftests/mm/pagemap_ioctl.c
1624
if (mem) {
tools/testing/selftests/mm/pagemap_ioctl.c
1625
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1626
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1628
base_tests("Hugetlb mem testing:", mem, mem_size, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1630
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
164
char *mem;
tools/testing/selftests/mm/pagemap_ioctl.c
1640
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1641
if (mem != MAP_FAILED) {
tools/testing/selftests/mm/pagemap_ioctl.c
1642
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1643
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
1645
base_tests("Hugetlb shmem testing:", mem, mem_size, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
1647
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
171
mem = shmat(*shmid, 0, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
172
if (mem == (char *)-1) {
tools/testing/selftests/mm/pagemap_ioctl.c
177
mem = mmap(NULL, size, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/pagemap_ioctl.c
179
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
183
return mem;
tools/testing/selftests/mm/pagemap_ioctl.c
189
char *mem, *vec;
tools/testing/selftests/mm/pagemap_ioctl.c
192
mem = mmap(NULL, mem_size, PROT_NONE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
193
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
196
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
199
mprotect(mem, mem_size/8, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
200
mprotect(mem + 1 * mem_size/8, mem_size/8, PROT_READ);
tools/testing/selftests/mm/pagemap_ioctl.c
201
mprotect(mem + 2 * mem_size/8, mem_size/8, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
202
mprotect(mem + 3 * mem_size/8, mem_size/8, PROT_READ);
tools/testing/selftests/mm/pagemap_ioctl.c
203
mprotect(mem + 4 * mem_size/8, mem_size/8, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
204
mprotect(mem + 5 * mem_size/8, mem_size/8, PROT_NONE);
tools/testing/selftests/mm/pagemap_ioctl.c
205
mprotect(mem + 6 * mem_size/8, mem_size/8, PROT_READ|PROT_WRITE);
tools/testing/selftests/mm/pagemap_ioctl.c
206
mprotect(mem + 7 * mem_size/8, mem_size/8, PROT_READ);
tools/testing/selftests/mm/pagemap_ioctl.c
208
wp_addr_range(mem + (mem_size/16), mem_size - 2 * (mem_size/8));
tools/testing/selftests/mm/pagemap_ioctl.c
209
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
214
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
221
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
222
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
244
char *mem, *m[2];
tools/testing/selftests/mm/pagemap_ioctl.c
258
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
259
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
262
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
263
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
266
ksft_test_result(pagemap_ioctl(mem, 0, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
270
ksft_test_result(pagemap_ioctl(mem, mem_size, NULL, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
274
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
278
ksft_test_result(pagemap_ioctl(mem, mem_size, 0, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
282
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, -1,
tools/testing/selftests/mm/pagemap_ioctl.c
286
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
291
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
295
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
299
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
303
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
308
ksft_test_result(pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
315
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
322
mem[i]++;
tools/testing/selftests/mm/pagemap_ioctl.c
324
ret = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0, PAGE_IS_WRITTEN, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
333
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
339
ret2 = pagemap_ioctl(mem, mem_size, vec, 2, 0, 0, PAGE_IS_WRITTEN, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
344
ret3 = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
356
mem[i]++;
tools/testing/selftests/mm/pagemap_ioctl.c
357
mem[(mem_size/page_size - 1) * page_size]++;
tools/testing/selftests/mm/pagemap_ioctl.c
359
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
365
ret2 = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
377
memset(mem, -1, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
380
ret = pagemap_ioctl(mem + page_size, 2 * page_size, vec, 1,
tools/testing/selftests/mm/pagemap_ioctl.c
386
ret2 = pagemap_ioctl(mem, mem_size, vec2, vec_size, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
392
vec[0].start == (uintptr_t)(mem + page_size) &&
tools/testing/selftests/mm/pagemap_ioctl.c
393
ret2 == 2 && LEN(vec2[0]) == 1 && vec2[0].start == (uintptr_t)mem &&
tools/testing/selftests/mm/pagemap_ioctl.c
395
vec2[1].start == (uintptr_t)(mem + 3 * page_size),
tools/testing/selftests/mm/pagemap_ioctl.c
398
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
399
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
443
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
444
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
447
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
448
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
450
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
457
mem[i * page_size]++;
tools/testing/selftests/mm/pagemap_ioctl.c
459
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
467
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
475
ret = pagemap_ioctl(mem, mem_size, vec, vec_size,
tools/testing/selftests/mm/pagemap_ioctl.c
486
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
487
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
498
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
499
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
502
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
503
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
507
walk_end = (long)mem;
tools/testing/selftests/mm/pagemap_ioctl.c
511
mem[i]++;
tools/testing/selftests/mm/pagemap_ioctl.c
517
ret = pagemap_ioc((void *)walk_end, mem_size-(walk_end - (long)mem), vec,
tools/testing/selftests/mm/pagemap_ioctl.c
537
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
538
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
548
mem = mmap(NULL, mem_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON, -1, 0);
tools/testing/selftests/mm/pagemap_ioctl.c
549
if (mem == MAP_FAILED)
tools/testing/selftests/mm/pagemap_ioctl.c
552
wp_init(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
553
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
555
memset(mem, 0, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
557
ret = pagemap_ioc(mem, 0, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
561
ksft_test_result(ret == 0 && walk_end == (long)mem,
tools/testing/selftests/mm/pagemap_ioctl.c
564
ret = pagemap_ioc(mem, 0, vec, vec_size, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
568
ksft_test_result(ret == 0 && walk_end == (long)mem,
tools/testing/selftests/mm/pagemap_ioctl.c
571
ret = pagemap_ioc(mem, 0, vec, 0, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
575
ksft_test_result(ret == 0 && walk_end == (long)mem,
tools/testing/selftests/mm/pagemap_ioctl.c
578
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
582
ksft_test_result(ret == 1 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
585
ret = pagemap_ioc(mem, mem_size, vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
589
ksft_test_result(ret == 1 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
592
ret = pagemap_ioc(mem, mem_size, vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
596
ksft_test_result(ret == 1 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
599
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
603
ksft_test_result(ret == 1 && walk_end == (long)(mem + mem_size/2),
tools/testing/selftests/mm/pagemap_ioctl.c
606
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
610
ksft_test_result(ret == 1 && walk_end == (long)(mem + page_size),
tools/testing/selftests/mm/pagemap_ioctl.c
613
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
617
ksft_test_result(ret == 1 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
620
wp_addr_range(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
622
mem[i]++;
tools/testing/selftests/mm/pagemap_ioctl.c
624
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
628
ksft_test_result((unsigned long)ret == vec_size/2 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
631
ret = pagemap_ioc(mem, mem_size, vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
635
ksft_test_result(ret == 1 && walk_end == (long)(mem + page_size * 2),
tools/testing/selftests/mm/pagemap_ioctl.c
638
ret = pagemap_ioc(mem, mem_size, vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
642
ksft_test_result(ret == 1 && walk_end == (long)(mem + page_size * 2),
tools/testing/selftests/mm/pagemap_ioctl.c
645
ret = pagemap_ioc(mem, mem_size, vec, vec_size/2, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
649
ksft_test_result((unsigned long)ret == vec_size/2 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
652
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
656
ksft_test_result((unsigned long)ret == vec_size/2 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
659
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
663
ksft_test_result((unsigned long)ret == vec_size/2 && walk_end == (long)(mem + mem_size),
tools/testing/selftests/mm/pagemap_ioctl.c
666
ret = pagemap_ioc(mem, mem_size, vec, vec_size, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
670
ksft_test_result(ret == 1 && walk_end == (long)(mem + page_size * 2),
tools/testing/selftests/mm/pagemap_ioctl.c
674
wp_free(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
675
munmap(mem, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
680
int base_tests(char *prefix, char *mem, unsigned long long mem_size, int skip)
tools/testing/selftests/mm/pagemap_ioctl.c
702
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
710
memset(mem, -1, mem_size);
tools/testing/selftests/mm/pagemap_ioctl.c
712
written = pagemap_ioctl(mem, mem_size, vec, 1, 0, 0, PAGE_IS_WRITTEN, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
721
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
726
memset(mem + page_size, 0, mem_size - (2 * page_size));
tools/testing/selftests/mm/pagemap_ioctl.c
728
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
736
written = pagemap_ioctl(mem, mem_size, vec, 1, 0, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
745
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
750
mem[vec_size/2 * page_size]++;
tools/testing/selftests/mm/pagemap_ioctl.c
752
written = pagemap_ioctl(mem, mem_size, vec, vec_size, 0, 0, PAGE_IS_WRITTEN,
tools/testing/selftests/mm/pagemap_ioctl.c
761
written = pagemap_ioctl(mem, mem_size, vec, 1, PM_SCAN_WP_MATCHING | PM_SCAN_CHECK_WPASYNC,
tools/testing/selftests/mm/pagemap_ioctl.c
766
mem[vec_size/2 * page_size]++;
tools/testing/selftests/mm/pagemap_ioctl.c
767
mem[(vec_size/2 + 1) * page_size]++;
tools/testing/selftests/mm/pagemap_ioctl.c
769
written = pagemap_ioctl(&mem[vec_size/2 * page_size], 2 * page_size, vec, 1, 0,
tools/testing/selftests/mm/pagemap_ioctl.c
774
ksft_test_result(written == 1 && vec[0].start == (uintptr_t)(&mem[vec_size/2 * page_size])
tools/testing/selftests/mm/prctl_thp_disable.c
42
char *mem, *mmap_mem;
tools/testing/selftests/mm/prctl_thp_disable.c
54
mem = (char *)(((uintptr_t)mmap_mem + pmdsize) & ~(pmdsize - 1));
tools/testing/selftests/mm/prctl_thp_disable.c
57
madvise(mem, pmdsize, MADV_HUGEPAGE);
tools/testing/selftests/mm/prctl_thp_disable.c
59
madvise(mem, pmdsize, MADV_NOHUGEPAGE);
tools/testing/selftests/mm/prctl_thp_disable.c
62
memset(mem, 1, pmdsize);
tools/testing/selftests/mm/prctl_thp_disable.c
65
madvise(mem, pmdsize, MADV_COLLAPSE);
tools/testing/selftests/mm/prctl_thp_disable.c
68
mprotect(mem, pmdsize, PROT_READ);
tools/testing/selftests/mm/prctl_thp_disable.c
70
ret = check_huge_anon(mem, 1, pmdsize);
tools/testing/selftests/mm/uffd-wp-mremap.c
101
if (mem == MAP_FAILED) {
tools/testing/selftests/mm/uffd-wp-mremap.c
102
mem = NULL;
tools/testing/selftests/mm/uffd-wp-mremap.c
106
assert(((uintptr_t)mem & (size - 1)) == 0);
tools/testing/selftests/mm/uffd-wp-mremap.c
113
mem[0] = 1;
tools/testing/selftests/mm/uffd-wp-mremap.c
114
for (addr = mem; addr < mem + size; addr += pagesize) {
tools/testing/selftests/mm/uffd-wp-mremap.c
116
munmap(mem, size);
tools/testing/selftests/mm/uffd-wp-mremap.c
117
mem = NULL;
tools/testing/selftests/mm/uffd-wp-mremap.c
121
memset(mem, 1, size);
tools/testing/selftests/mm/uffd-wp-mremap.c
126
return mem;
tools/testing/selftests/mm/uffd-wp-mremap.c
129
static bool check_uffd_wp_state(void *mem, size_t size, bool expect)
tools/testing/selftests/mm/uffd-wp-mremap.c
134
for (addr = mem; addr < mem + size; addr += pagesize) {
tools/testing/selftests/mm/uffd-wp-mremap.c
139
(addr - mem) / pagesize);
tools/testing/selftests/mm/uffd-wp-mremap.c
161
void *mem = NULL;
tools/testing/selftests/mm/uffd-wp-mremap.c
173
mem = alloc_one_folio(size, private, hugetlb);
tools/testing/selftests/mm/uffd-wp-mremap.c
174
if (!mem) {
tools/testing/selftests/mm/uffd-wp-mremap.c
187
if (uffd_register(gopts->uffd, mem, size, false, true, false)) {
tools/testing/selftests/mm/uffd-wp-mremap.c
192
wp_prms.range.start = (uintptr_t)mem;
tools/testing/selftests/mm/uffd-wp-mremap.c
200
madvise(mem, size, MADV_PAGEOUT);
tools/testing/selftests/mm/uffd-wp-mremap.c
201
if (!range_is_swapped(mem, size)) {
tools/testing/selftests/mm/uffd-wp-mremap.c
208
if (!check_uffd_wp_state(mem, size, true))
tools/testing/selftests/mm/uffd-wp-mremap.c
221
if (mremap(mem, size, size, MREMAP_FIXED | MREMAP_MAYMOVE, addr) == MAP_FAILED) {
tools/testing/selftests/mm/uffd-wp-mremap.c
226
mem = addr;
tools/testing/selftests/mm/uffd-wp-mremap.c
229
if (!check_uffd_wp_state(mem, size, false))
tools/testing/selftests/mm/uffd-wp-mremap.c
239
if (mem)
tools/testing/selftests/mm/uffd-wp-mremap.c
240
munmap(mem, size);
tools/testing/selftests/mm/uffd-wp-mremap.c
50
char *mmap_mem, *mem;
tools/testing/selftests/mm/uffd-wp-mremap.c
56
mem = (char *)(((uintptr_t)mmap_mem + size - 1) & ~(size - 1));
tools/testing/selftests/mm/uffd-wp-mremap.c
57
munmap(mmap_mem, mem - mmap_mem);
tools/testing/selftests/mm/uffd-wp-mremap.c
58
munmap(mem + size, mmap_mem + mmap_size - mem - size);
tools/testing/selftests/mm/uffd-wp-mremap.c
60
return mem;
tools/testing/selftests/mm/uffd-wp-mremap.c
68
char *mem, *addr;
tools/testing/selftests/mm/uffd-wp-mremap.c
91
mem = mmap_aligned(size, prot, flags);
tools/testing/selftests/mm/uffd-wp-mremap.c
98
mem = mmap(NULL, size, prot, flags, -1, 0);
tools/testing/selftests/net/tls.c
675
void *mem = malloc(16384);
tools/testing/selftests/net/tls.c
684
struct iovec vec = { (void *)mem, send_len };
tools/testing/selftests/net/tls.c
692
EXPECT_NE(recv(self->cfd, mem, send_len, 0), -1);
tools/testing/selftests/net/tls.c
695
free(mem);
tools/testing/selftests/nolibc/nolibc-test.c
1132
void *mem;
tools/testing/selftests/nolibc/nolibc-test.c
1175
mem = mmap(NULL, mem_length, PROT_READ, MAP_SHARED, fd, pa_offset);
tools/testing/selftests/nolibc/nolibc-test.c
1176
if (mem == MAP_FAILED) {
tools/testing/selftests/nolibc/nolibc-test.c
1181
mem = mremap(mem, mem_length, mem_length * 2, MREMAP_MAYMOVE, 0);
tools/testing/selftests/nolibc/nolibc-test.c
1182
if (mem == MAP_FAILED) {
tools/testing/selftests/nolibc/nolibc-test.c
1183
munmap(mem, mem_length);
tools/testing/selftests/nolibc/nolibc-test.c
1188
ret = munmap(mem, mem_length * 2);
tools/testing/selftests/riscv/hwprobe/cbo.c
135
prefetch_i(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
136
prefetch_r(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
137
prefetch_w(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
168
cbo_clean(&mem[block_size]);
tools/testing/selftests/riscv/hwprobe/cbo.c
172
cbo_flush(&mem[block_size]);
tools/testing/selftests/riscv/hwprobe/cbo.c
193
cbo_zero(&mem[block_size]);
tools/testing/selftests/riscv/hwprobe/cbo.c
205
cbo_zero(&mem[i * block_size]);
tools/testing/selftests/riscv/hwprobe/cbo.c
212
if (mem[i * block_size + j] != expected) {
tools/testing/selftests/riscv/hwprobe/cbo.c
27
static char mem[4096] __aligned(4096) = { [0 ... 4095] = 0xa5 };
tools/testing/selftests/riscv/hwprobe/cbo.c
76
cbo_inval(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
85
cbo_clean(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
89
cbo_flush(&mem[0]);
tools/testing/selftests/riscv/hwprobe/cbo.c
98
cbo_zero(&mem[0]);
tools/testing/vsock/util.c
814
static void mem_invert(unsigned char *mem, size_t size)
tools/testing/vsock/util.c
819
mem[i] = ~mem[i];
tools/usb/usbip/libsrc/names.c
144
void *mem;
tools/usb/usbip/libsrc/names.c
157
p->mem = calloc(1, size);
tools/usb/usbip/libsrc/names.c
158
if (!p->mem) {
tools/usb/usbip/libsrc/names.c
166
return p->mem;
tools/usb/usbip/libsrc/names.c
179
if (pool->mem)
tools/usb/usbip/libsrc/names.c
180
free(pool->mem);
tools/virtio/vhost_net_test.c
238
dev->mem = malloc(offsetof(struct vhost_memory, regions) +
tools/virtio/vhost_net_test.c
239
sizeof(dev->mem->regions[0]));
tools/virtio/vhost_net_test.c
240
assert(dev->mem);
tools/virtio/vhost_net_test.c
241
memset(dev->mem, 0, offsetof(struct vhost_memory, regions) +
tools/virtio/vhost_net_test.c
242
sizeof(dev->mem->regions[0]));
tools/virtio/vhost_net_test.c
243
dev->mem->nregions = 1;
tools/virtio/vhost_net_test.c
244
dev->mem->regions[0].guest_phys_addr = (long)dev->buf;
tools/virtio/vhost_net_test.c
245
dev->mem->regions[0].userspace_addr = (long)dev->buf;
tools/virtio/vhost_net_test.c
246
dev->mem->regions[0].memory_size = dev->buf_size;
tools/virtio/vhost_net_test.c
248
r = ioctl(dev->control, VHOST_SET_MEM_TABLE, dev->mem);
tools/virtio/vhost_net_test.c
54
struct vhost_memory *mem;
tools/virtio/virtio_test.c
141
dev->mem = malloc(offsetof(struct vhost_memory, regions) +
tools/virtio/virtio_test.c
142
sizeof dev->mem->regions[0]);
tools/virtio/virtio_test.c
143
assert(dev->mem);
tools/virtio/virtio_test.c
144
memset(dev->mem, 0, offsetof(struct vhost_memory, regions) +
tools/virtio/virtio_test.c
145
sizeof dev->mem->regions[0]);
tools/virtio/virtio_test.c
146
dev->mem->nregions = 1;
tools/virtio/virtio_test.c
147
dev->mem->regions[0].guest_phys_addr = (long)dev->buf;
tools/virtio/virtio_test.c
148
dev->mem->regions[0].userspace_addr = (long)dev->buf;
tools/virtio/virtio_test.c
149
dev->mem->regions[0].memory_size = dev->buf_size;
tools/virtio/virtio_test.c
150
r = ioctl(dev->control, VHOST_SET_MEM_TABLE, dev->mem);
tools/virtio/virtio_test.c
46
struct vhost_memory *mem;
virt/kvm/kvm_main.c
1581
const struct kvm_userspace_memory_region2 *mem)
virt/kvm/kvm_main.c
1589
if (mem->flags & KVM_MEM_GUEST_MEMFD)
virt/kvm/kvm_main.c
1598
!(mem->flags & KVM_MEM_GUEST_MEMFD))
virt/kvm/kvm_main.c
1601
if (mem->flags & ~valid_flags)
virt/kvm/kvm_main.c
2002
const struct kvm_userspace_memory_region2 *mem)
virt/kvm/kvm_main.c
2014
r = check_memory_region_flags(kvm, mem);
virt/kvm/kvm_main.c
2018
as_id = mem->slot >> 16;
virt/kvm/kvm_main.c
2019
id = (u16)mem->slot;
virt/kvm/kvm_main.c
2022
if ((mem->memory_size & (PAGE_SIZE - 1)) ||
virt/kvm/kvm_main.c
2023
(mem->memory_size != (unsigned long)mem->memory_size))
virt/kvm/kvm_main.c
2025
if (mem->guest_phys_addr & (PAGE_SIZE - 1))
virt/kvm/kvm_main.c
2028
if ((mem->userspace_addr & (PAGE_SIZE - 1)) ||
virt/kvm/kvm_main.c
2029
(mem->userspace_addr != untagged_addr(mem->userspace_addr)) ||
virt/kvm/kvm_main.c
2030
!access_ok((void __user *)(unsigned long)mem->userspace_addr,
virt/kvm/kvm_main.c
2031
mem->memory_size))
virt/kvm/kvm_main.c
2033
if (mem->flags & KVM_MEM_GUEST_MEMFD &&
virt/kvm/kvm_main.c
2034
(mem->guest_memfd_offset & (PAGE_SIZE - 1) ||
virt/kvm/kvm_main.c
2035
mem->guest_memfd_offset + mem->memory_size < mem->guest_memfd_offset))
virt/kvm/kvm_main.c
2039
if (mem->guest_phys_addr + mem->memory_size < mem->guest_phys_addr)
virt/kvm/kvm_main.c
2049
(mem->memory_size >> PAGE_SHIFT) > KVM_MEM_MAX_NR_PAGES)
virt/kvm/kvm_main.c
2060
if (!mem->memory_size) {
virt/kvm/kvm_main.c
2070
base_gfn = (mem->guest_phys_addr >> PAGE_SHIFT);
virt/kvm/kvm_main.c
2071
npages = (mem->memory_size >> PAGE_SHIFT);
virt/kvm/kvm_main.c
2084
if (mem->flags & KVM_MEM_GUEST_MEMFD)
virt/kvm/kvm_main.c
2086
if ((mem->userspace_addr != old->userspace_addr) ||
virt/kvm/kvm_main.c
2088
((mem->flags ^ old->flags) & (KVM_MEM_READONLY | KVM_MEM_GUEST_MEMFD)))
virt/kvm/kvm_main.c
2093
else if (mem->flags != old->flags)
virt/kvm/kvm_main.c
2112
new->flags = mem->flags;
virt/kvm/kvm_main.c
2113
new->userspace_addr = mem->userspace_addr;
virt/kvm/kvm_main.c
2114
if (mem->flags & KVM_MEM_GUEST_MEMFD) {
virt/kvm/kvm_main.c
2115
r = kvm_gmem_bind(kvm, new, mem->guest_memfd, mem->guest_memfd_offset);
virt/kvm/kvm_main.c
2127
if (mem->flags & KVM_MEM_GUEST_MEMFD)
virt/kvm/kvm_main.c
2135
const struct kvm_userspace_memory_region2 *mem)
virt/kvm/kvm_main.c
2137
if (WARN_ON_ONCE(mem->slot < KVM_USER_MEM_SLOTS))
virt/kvm/kvm_main.c
2140
if (WARN_ON_ONCE(mem->flags))
virt/kvm/kvm_main.c
2143
return kvm_set_memory_region(kvm, mem);
virt/kvm/kvm_main.c
2148
struct kvm_userspace_memory_region2 *mem)
virt/kvm/kvm_main.c
2150
if ((u16)mem->slot >= KVM_USER_MEM_SLOTS)
virt/kvm/kvm_main.c
2154
return kvm_set_memory_region(kvm, mem);
virt/kvm/kvm_main.c
5178
struct kvm_userspace_memory_region2 mem;
virt/kvm/kvm_main.c
5186
memset(&mem, 0, sizeof(mem));
virt/kvm/kvm_main.c
5200
if (copy_from_user(&mem, argp, size))
virt/kvm/kvm_main.c
5205
(mem.flags & ~KVM_SET_USER_MEMORY_REGION_V1_FLAGS))
virt/kvm/kvm_main.c
5208
r = kvm_vm_ioctl_set_memory_region(kvm, &mem);