Symbol: dma_addr
arch/alpha/kernel/pci_iommu.c
343
static void alpha_pci_unmap_phys(struct device *dev, dma_addr_t dma_addr,
arch/alpha/kernel/pci_iommu.c
353
if (dma_addr >= __direct_map_base
arch/alpha/kernel/pci_iommu.c
354
&& dma_addr < __direct_map_base + __direct_map_size) {
arch/alpha/kernel/pci_iommu.c
358
dma_addr, size, __builtin_return_address(0));
arch/alpha/kernel/pci_iommu.c
363
if (dma_addr > 0xffffffff) {
arch/alpha/kernel/pci_iommu.c
365
dma_addr, size, __builtin_return_address(0));
arch/alpha/kernel/pci_iommu.c
370
if (!arena || dma_addr < arena->dma_base)
arch/alpha/kernel/pci_iommu.c
373
dma_ofs = (dma_addr - arena->dma_base) >> PAGE_SHIFT;
arch/alpha/kernel/pci_iommu.c
377
dma_addr, arena->dma_base, arena->size);
arch/alpha/kernel/pci_iommu.c
382
npages = iommu_num_pages(dma_addr, size, PAGE_SIZE);
arch/alpha/kernel/pci_iommu.c
392
alpha_mv.mv_pci_tbi(hose, dma_addr, dma_addr + size - 1);
arch/alpha/kernel/pci_iommu.c
397
dma_addr, size, npages, __builtin_return_address(0));
arch/alpha/kernel/pci_iommu.c
451
void *cpu_addr, dma_addr_t dma_addr,
arch/alpha/kernel/pci_iommu.c
455
dma_unmap_single(&pdev->dev, dma_addr, size, DMA_BIDIRECTIONAL);
arch/alpha/kernel/pci_iommu.c
459
dma_addr, size, __builtin_return_address(0));
arch/arm/mach-rpc/dma.c
178
idma->dma_addr = idma->dma.sg->dma_address;
arch/arm/mach-rpc/dma.c
32
dma_addr_t dma_addr;
arch/arm/mach-rpc/dma.c
59
idma->cur_addr = idma->dma_addr;
arch/arm/mach-rpc/dma.c
73
idma->dma_addr += end - offset;
arch/arm/mach-rpc/dma.c
78
idma->dma_addr = idma->dma.sg->dma_address;
arch/arm/mm/dma-mapping.c
1114
void *cpu_addr, dma_addr_t dma_addr, size_t size,
arch/arm/mm/dma-mapping.c
1167
void *cpu_addr, dma_addr_t dma_addr,
arch/arm/mm/dma-mapping.c
1371
dma_addr_t dma_addr;
arch/arm/mm/dma-mapping.c
1378
dma_addr = __alloc_iova(mapping, len);
arch/arm/mm/dma-mapping.c
1379
if (dma_addr == DMA_MAPPING_ERROR)
arch/arm/mm/dma-mapping.c
1380
return dma_addr;
arch/arm/mm/dma-mapping.c
1384
ret = iommu_map(mapping->domain, dma_addr, addr, len, prot, GFP_KERNEL);
arch/arm/mm/dma-mapping.c
1388
return dma_addr + offset_in_page(phys);
arch/arm/mm/dma-mapping.c
1390
__free_iova(mapping, dma_addr, len);
arch/arm/mm/dma-mapping.c
965
dma_addr_t dma_addr, iova;
arch/arm/mm/dma-mapping.c
968
dma_addr = __alloc_iova(mapping, size);
arch/arm/mm/dma-mapping.c
969
if (dma_addr == DMA_MAPPING_ERROR)
arch/arm/mm/dma-mapping.c
970
return dma_addr;
arch/arm/mm/dma-mapping.c
972
iova = dma_addr;
arch/arm/mm/dma-mapping.c
993
return dma_addr;
arch/arm/mm/dma-mapping.c
995
iommu_unmap(mapping->domain, dma_addr, iova-dma_addr);
arch/arm/mm/dma-mapping.c
996
__free_iova(mapping, dma_addr, size);
arch/mips/ath25/ar5312.c
63
u32 dma_addr = ar5312_rst_reg_read(AR5312_DMAADDR); /* clears error */
arch/mips/ath25/ar5312.c
66
proc_addr, proc1, dma_addr, dma1);
arch/mips/include/asm/mach-rc32434/dma_v.h
42
static inline void rc32434_start_dma(struct dma_reg *ch, u32 dma_addr)
arch/mips/include/asm/mach-rc32434/dma_v.h
45
__raw_writel(dma_addr, &ch->dmadptr);
arch/mips/include/asm/mach-rc32434/dma_v.h
48
static inline void rc32434_chain_dma(struct dma_reg *ch, u32 dma_addr)
arch/mips/include/asm/mach-rc32434/dma_v.h
50
__raw_writel(dma_addr, &ch->dmandptr);
arch/mips/jazz/jazzdma.c
541
static void jazz_dma_unmap_phys(struct device *dev, dma_addr_t dma_addr,
arch/mips/jazz/jazzdma.c
545
arch_sync_dma_for_cpu(vdma_log2phys(dma_addr), size, dir);
arch/mips/jazz/jazzdma.c
546
vdma_free(dma_addr);
arch/mips/loongson2ef/fuloong-2e/dma.c
11
return dma_addr & 0x7fffffff;
arch/mips/loongson2ef/fuloong-2e/dma.c
9
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
arch/mips/loongson2ef/lemote-2f/dma.c
11
if (dma_addr > 0x8fffffff)
arch/mips/loongson2ef/lemote-2f/dma.c
12
return dma_addr;
arch/mips/loongson2ef/lemote-2f/dma.c
13
return dma_addr & 0x0fffffff;
arch/mips/loongson2ef/lemote-2f/dma.c
9
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
arch/mips/pci/pci-ar2315.c
179
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
arch/mips/pci/pci-ar2315.c
181
return dma_addr - ar2315_dev_offset(dev);
arch/mips/pci/pci-xtalk-bridge.c
37
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
arch/mips/pci/pci-xtalk-bridge.c
39
return dma_addr & ~(0xffUL << 56);
arch/mips/sgi-ip32/ip32-dma.c
23
dma_addr_t dma_addr = paddr & RAM_OFFSET_MASK;
arch/mips/sgi-ip32/ip32-dma.c
26
dma_addr += CRIME_HI_MEM_BASE;
arch/mips/sgi-ip32/ip32-dma.c
27
return dma_addr;
arch/mips/sgi-ip32/ip32-dma.c
30
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
arch/mips/sgi-ip32/ip32-dma.c
32
phys_addr_t paddr = dma_addr & RAM_OFFSET_MASK;
arch/mips/sgi-ip32/ip32-dma.c
34
if (dma_addr >= 256*1024*1024)
arch/powerpc/include/asm/opal.h
236
uint64_t dma_addr, uint32_t npages);
arch/powerpc/kernel/iommu.c
390
static bool iommu_free_check(struct iommu_table *tbl, dma_addr_t dma_addr,
arch/powerpc/kernel/iommu.c
395
entry = dma_addr >> tbl->it_page_shift;
arch/powerpc/kernel/iommu.c
403
printk(KERN_INFO "\tdma_addr = 0x%llx\n", (u64)dma_addr);
arch/powerpc/kernel/iommu.c
437
static void __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr,
arch/powerpc/kernel/iommu.c
444
entry = dma_addr >> tbl->it_page_shift;
arch/powerpc/kernel/iommu.c
449
if (!iommu_free_check(tbl, dma_addr, npages))
arch/powerpc/kernel/iommu.c
459
static void iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr,
arch/powerpc/kernel/iommu.c
462
__iommu_free(tbl, dma_addr, npages);
arch/powerpc/kernel/iommu.c
477
dma_addr_t dma_next = 0, dma_addr;
arch/powerpc/kernel/iommu.c
533
dma_addr = entry << tbl->it_page_shift;
arch/powerpc/kernel/iommu.c
534
dma_addr |= (vaddr & ~IOMMU_PAGE_MASK(tbl));
arch/powerpc/kernel/iommu.c
537
npages, entry, dma_addr);
arch/powerpc/kernel/iommu.c
552
if (novmerge || (dma_addr != dma_next) ||
arch/powerpc/kernel/iommu.c
568
outs->dma_address = dma_addr;
arch/powerpc/kernel/iommu.c
573
dma_next = dma_addr + slen;
arch/powerpc/platforms/ps3/system-bus.c
621
static void ps3_unmap_phys(struct device *_dev, dma_addr_t dma_addr,
arch/powerpc/platforms/ps3/system-bus.c
627
result = ps3_dma_unmap(dev->d_region, dma_addr, size);
arch/powerpc/platforms/pseries/ibmebus.c
101
dma_addr_t dma_addr,
arch/powerpc/platforms/pseries/iommu.c
1063
static bool find_existing_ddw(struct device_node *pdn, u64 *dma_addr, int *window_shift,
arch/powerpc/platforms/pseries/iommu.c
1075
*dma_addr = be64_to_cpu(dma64->dma_base);
arch/powerpc/platforms/pseries/iommu.c
1406
static struct property *ddw_property_create(const char *propname, u32 liobn, u64 dma_addr,
arch/powerpc/platforms/pseries/iommu.c
1428
ddwprop->dma_base = cpu_to_be64(dma_addr);
arch/sparc/include/asm/iommu-common.h
50
u64 dma_addr, unsigned long npages,
arch/sparc/kernel/iommu-common.c
246
void iommu_tbl_range_free(struct iommu_map_table *iommu, u64 dma_addr,
arch/sparc/kernel/iommu-common.c
254
entry = (dma_addr - iommu->table_map_base) >> shift;
arch/sparc/kernel/iommu.c
448
dma_addr_t dma_next = 0, dma_addr;
arch/sparc/kernel/iommu.c
514
dma_addr = iommu->tbl.table_map_base +
arch/sparc/kernel/iommu.c
516
dma_addr |= (s->offset & ~IO_PAGE_MASK);
arch/sparc/kernel/iommu.c
531
if ((dma_addr != dma_next) ||
arch/sparc/kernel/iommu.c
546
outs->dma_address = dma_addr;
arch/sparc/kernel/iommu.c
552
dma_next = dma_addr + slen;
arch/sparc/kernel/pci_sun4v.c
484
dma_addr_t dma_next = 0, dma_addr;
arch/sparc/kernel/pci_sun4v.c
557
dma_addr = tbl->table_map_base + (entry << IO_PAGE_SHIFT);
arch/sparc/kernel/pci_sun4v.c
558
dma_addr |= (s->offset & ~IO_PAGE_MASK);
arch/sparc/kernel/pci_sun4v.c
574
if ((dma_addr != dma_next) ||
arch/sparc/kernel/pci_sun4v.c
589
outs->dma_address = dma_addr;
arch/sparc/kernel/pci_sun4v.c
595
dma_next = dma_addr + slen;
arch/sparc/mm/io-unit.c
273
dma_addr_t dma_addr, unsigned long attrs)
arch/sparc/mm/iommu.c
282
static void sbus_iommu_unmap_phys(struct device *dev, dma_addr_t dma_addr,
arch/sparc/mm/iommu.c
286
unsigned int busa = dma_addr & PAGE_MASK;
arch/sparc/mm/iommu.c
287
unsigned long off = dma_addr & ~PAGE_MASK;
arch/x86/kernel/amd_gart_64.c
246
static void gart_unmap_phys(struct device *dev, dma_addr_t dma_addr,
arch/x86/kernel/amd_gart_64.c
254
if (WARN_ON_ONCE(dma_addr == DMA_MAPPING_ERROR))
arch/x86/kernel/amd_gart_64.c
262
if (dma_addr < iommu_bus_base ||
arch/x86/kernel/amd_gart_64.c
263
dma_addr >= iommu_bus_base + iommu_size)
arch/x86/kernel/amd_gart_64.c
266
iommu_page = (dma_addr - iommu_bus_base)>>PAGE_SHIFT;
arch/x86/kernel/amd_gart_64.c
267
npages = iommu_num_pages(dma_addr, size, PAGE_SIZE);
arch/x86/kernel/amd_gart_64.c
465
gart_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_addr,
arch/x86/kernel/amd_gart_64.c
470
vaddr = dma_direct_alloc(dev, size, dma_addr, flag, attrs);
arch/x86/kernel/amd_gart_64.c
475
*dma_addr = dma_map_area(dev, virt_to_phys(vaddr), size,
arch/x86/kernel/amd_gart_64.c
478
if (unlikely(*dma_addr == DMA_MAPPING_ERROR))
arch/x86/kernel/amd_gart_64.c
482
dma_direct_free(dev, size, vaddr, *dma_addr, attrs);
arch/x86/kernel/amd_gart_64.c
489
dma_addr_t dma_addr, unsigned long attrs)
arch/x86/kernel/amd_gart_64.c
491
gart_unmap_phys(dev, dma_addr, size, DMA_BIDIRECTIONAL, 0);
arch/x86/kernel/amd_gart_64.c
492
dma_direct_free(dev, size, vaddr, dma_addr, attrs);
drivers/accel/amdxdna/aie2_message.c
378
dma_addr_t dma_addr;
drivers/accel/amdxdna/aie2_message.c
382
buff_addr = aie2_alloc_msg_buffer(ndev, &buf_sz, &dma_addr);
drivers/accel/amdxdna/aie2_message.c
391
req.dump_buff_addr = dma_addr;
drivers/accel/amdxdna/aie2_message.c
420
aie2_free_msg_buffer(ndev, buf_sz, buff_addr, dma_addr);
drivers/accel/amdxdna/aie2_message.c
430
dma_addr_t dma_addr;
drivers/accel/amdxdna/aie2_message.c
438
addr = aie2_alloc_msg_buffer(ndev, &buf_sz, &dma_addr);
drivers/accel/amdxdna/aie2_message.c
442
req.buf_addr = dma_addr;
drivers/accel/amdxdna/aie2_message.c
469
aie2_free_msg_buffer(ndev, buf_sz, addr, dma_addr);
drivers/accel/amdxdna/aie2_message.c
56
dma_addr_t *dma_addr)
drivers/accel/amdxdna/aie2_message.c
67
return dma_alloc_noncoherent(xdna->ddev.dev, *size, dma_addr,
drivers/accel/amdxdna/aie2_pci.h
340
dma_addr_t *dma_addr);
drivers/accel/amdxdna/aie2_pci.h
341
#define aie2_free_msg_buffer(ndev, size, buff_addr, dma_addr) \
drivers/accel/amdxdna/aie2_pci.h
343
dma_addr, DMA_FROM_DEVICE)
drivers/accel/ethosu/ethosu_job.c
55
writel_relaxed(lower_32_bits(bo->dma_addr), dev->regs + NPU_REG_BASEP(region));
drivers/accel/ethosu/ethosu_job.c
56
writel_relaxed(upper_32_bits(bo->dma_addr), dev->regs + NPU_REG_BASEP_HI(region));
drivers/accel/ethosu/ethosu_job.c
57
dev_dbg(dev->base.dev, "Region %d base addr = %pad\n", region, &bo->dma_addr);
drivers/accel/ethosu/ethosu_job.c
69
writel_relaxed(lower_32_bits(cmd_bo->dma_addr), dev->regs + NPU_REG_QBASE);
drivers/accel/ethosu/ethosu_job.c
70
writel_relaxed(upper_32_bits(cmd_bo->dma_addr), dev->regs + NPU_REG_QBASE_HI);
drivers/accel/ethosu/ethosu_job.c
76
"Submitted cmd at %pad to core\n", &cmd_bo->dma_addr);
drivers/accel/habanalabs/common/debugfs.c
358
dma_addr_t dma_addr;
drivers/accel/habanalabs/common/debugfs.c
368
npages = hl_get_sg_info(sg, &dma_addr);
drivers/accel/habanalabs/common/debugfs.c
376
dma_addr += (dev_entry->userptr_lookup -
drivers/accel/habanalabs/common/debugfs.c
386
(u64)dma_addr, userptr->pid,
drivers/accel/habanalabs/common/device.c
178
void hl_asic_dma_pool_free_caller(struct hl_device *hdev, void *vaddr, dma_addr_t dma_addr,
drivers/accel/habanalabs/common/device.c
181
hl_asic_dma_free_common(hdev, 0, vaddr, dma_addr, DMA_ALLOC_POOL, caller);
drivers/accel/habanalabs/common/habanalabs.h
159
#define hl_asic_dma_pool_free(hdev, vaddr, dma_addr) \
drivers/accel/habanalabs/common/habanalabs.h
160
hl_asic_dma_pool_free_caller(hdev, vaddr, dma_addr, __func__)
drivers/accel/habanalabs/common/habanalabs.h
1658
void *cpu_addr, dma_addr_t dma_addr, size_t size);
drivers/accel/habanalabs/common/habanalabs.h
1674
dma_addr_t dma_addr);
drivers/accel/habanalabs/common/habanalabs.h
3701
static inline u32 hl_get_sg_info(struct scatterlist *sg, dma_addr_t *dma_addr)
drivers/accel/habanalabs/common/habanalabs.h
3703
*dma_addr = sg_dma_address(sg);
drivers/accel/habanalabs/common/habanalabs.h
3705
return ((((*dma_addr) & (PAGE_SIZE - 1)) + sg_dma_len(sg)) +
drivers/accel/habanalabs/common/habanalabs.h
3763
void hl_asic_dma_pool_free_caller(struct hl_device *hdev, void *vaddr, dma_addr_t dma_addr,
drivers/accel/habanalabs/common/memory.c
850
dma_addr_t dma_addr;
drivers/accel/habanalabs/common/memory.c
872
npages = hl_get_sg_info(sg, &dma_addr);
drivers/accel/habanalabs/common/memory.c
877
(dma_addr & (huge_page_size - 1)))
drivers/accel/habanalabs/common/memory.c
901
npages = hl_get_sg_info(sg, &dma_addr);
drivers/accel/habanalabs/common/memory.c
906
phys_pg_pack->offset = dma_addr & (page_size - 1);
drivers/accel/habanalabs/common/memory.c
907
dma_addr &= page_mask;
drivers/accel/habanalabs/common/memory.c
911
phys_pg_pack->pages[j++] = dma_addr;
drivers/accel/habanalabs/common/memory.c
912
dma_addr += page_size;
drivers/accel/habanalabs/common/mmu/mmu.c
830
dma_addr_t dma_addr;
drivers/accel/habanalabs/common/mmu/mmu.c
855
&dma_addr,
drivers/accel/habanalabs/common/mmu/mmu.c
864
rc = gen_pool_add_virt(hr_priv->mmu_pgt_pool, virt_addr, (phys_addr_t) dma_addr,
drivers/accel/habanalabs/gaudi/gaudi.c
4163
void *cpu_addr, dma_addr_t dma_addr, size_t size)
drivers/accel/habanalabs/gaudi/gaudi.c
4180
(dma_addr - HOST_PHYS_BASE), size);
drivers/accel/habanalabs/gaudi/gaudi.c
4836
dma_addr_t dma_addr)
drivers/accel/habanalabs/gaudi/gaudi.c
4839
dma_addr_t fixed_dma_addr = dma_addr - HOST_PHYS_BASE;
drivers/accel/habanalabs/gaudi/gaudi.c
5148
dma_addr_t dma_addr, dma_addr_next;
drivers/accel/habanalabs/gaudi/gaudi.c
5198
dma_addr = sg_dma_address(sg);
drivers/accel/habanalabs/gaudi/gaudi.c
5211
if ((dma_addr + len == dma_addr_next) &&
drivers/accel/habanalabs/gaudi/gaudi.c
5229
new_dma_pkt->src_addr = cpu_to_le64(dma_addr);
drivers/accel/habanalabs/gaudi/gaudi.c
5233
new_dma_pkt->dst_addr = cpu_to_le64(dma_addr);
drivers/accel/habanalabs/gaudi/gaudi.c
5889
u32 size_to_dma, dma_addr_t dma_addr)
drivers/accel/habanalabs/gaudi/gaudi.c
5899
WREG32(mmDMA0_CORE_DST_BASE_LO + dma_offset, lower_32_bits(dma_addr));
drivers/accel/habanalabs/gaudi/gaudi.c
5900
WREG32(mmDMA0_CORE_DST_BASE_HI + dma_offset, upper_32_bits(dma_addr));
drivers/accel/habanalabs/gaudi/gaudi.c
5941
dma_addr_t dma_addr;
drivers/accel/habanalabs/gaudi/gaudi.c
5946
kernel_addr = hl_asic_dma_alloc_coherent(hdev, SZ_2M, &dma_addr, GFP_KERNEL | __GFP_ZERO);
drivers/accel/habanalabs/gaudi/gaudi.c
6009
dma_addr);
drivers/accel/habanalabs/gaudi/gaudi.c
6035
hl_asic_dma_free_coherent(hdev, SZ_2M, kernel_addr, dma_addr);
drivers/accel/habanalabs/gaudi2/gaudi2.c
4019
hl_asic_dma_pool_free(hdev, msg_info[i].kern_addr, msg_info[i].dma_addr);
drivers/accel/habanalabs/gaudi2/gaudi2.c
4034
GFP_KERNEL, &msg_info[i].dma_addr);
drivers/accel/habanalabs/gaudi2/gaudi2.c
6833
void *cpu_addr, dma_addr_t dma_addr, size_t size)
drivers/accel/habanalabs/gaudi2/gaudi2.c
6849
rc = dma_mmap_coherent(hdev->dev, vma, cpu_addr, dma_addr, size);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7116
static void gaudi2_dma_pool_free(struct hl_device *hdev, void *vaddr, dma_addr_t dma_addr)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7118
dma_pool_free(hdev->dma_pool, vaddr, dma_addr);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7388
rc = hl_hw_queue_send_cb_no_cmpl(hdev, hw_queue_id, pkt_size, msg_info->dma_addr);
drivers/accel/habanalabs/gaudi2/gaudi2P.h
470
dma_addr_t dma_addr;
drivers/accel/habanalabs/goya/goya.c
2872
void *cpu_addr, dma_addr_t dma_addr, size_t size)
drivers/accel/habanalabs/goya/goya.c
2880
(dma_addr - HOST_PHYS_BASE), size);
drivers/accel/habanalabs/goya/goya.c
3260
dma_addr_t dma_addr)
drivers/accel/habanalabs/goya/goya.c
3263
dma_addr_t fixed_dma_addr = dma_addr - HOST_PHYS_BASE;
drivers/accel/habanalabs/goya/goya.c
3755
dma_addr_t dma_addr, dma_addr_next;
drivers/accel/habanalabs/goya/goya.c
3815
dma_addr = sg_dma_address(sg);
drivers/accel/habanalabs/goya/goya.c
3828
if ((dma_addr + len == dma_addr_next) &&
drivers/accel/habanalabs/goya/goya.c
3847
new_dma_pkt->src_addr = cpu_to_le64(dma_addr);
drivers/accel/habanalabs/goya/goya.c
3851
new_dma_pkt->dst_addr = cpu_to_le64(dma_addr);
drivers/accel/ivpu/ivpu_mmu_context.c
240
u64 vpu_addr, dma_addr_t dma_addr, u64 prot)
drivers/accel/ivpu/ivpu_mmu_context.c
268
pte[pte_idx] = dma_addr | prot;
drivers/accel/ivpu/ivpu_mmu_context.c
275
dma_addr_t dma_addr, u64 prot)
drivers/accel/ivpu/ivpu_mmu_context.c
280
drm_WARN_ON(&vdev->drm, !IS_ALIGNED(dma_addr, size));
drivers/accel/ivpu/ivpu_mmu_context.c
285
int ret = ivpu_mmu_context_map_page(vdev, ctx, vpu_addr, dma_addr, prot);
drivers/accel/ivpu/ivpu_mmu_context.c
292
dma_addr += IVPU_MMU_PAGE_SIZE;
drivers/accel/ivpu/ivpu_mmu_context.c
311
u64 vpu_addr, dma_addr_t dma_addr, size_t size, u64 prot)
drivers/accel/ivpu/ivpu_mmu_context.c
318
IS_ALIGNED(vpu_addr | dma_addr, IVPU_MMU_CONT_PAGES_SIZE)) {
drivers/accel/ivpu/ivpu_mmu_context.c
319
ret = ivpu_mmu_context_map_cont_64k(vdev, ctx, vpu_addr, dma_addr, prot);
drivers/accel/ivpu/ivpu_mmu_context.c
322
ret = ivpu_mmu_context_map_page(vdev, ctx, vpu_addr, dma_addr, prot);
drivers/accel/ivpu/ivpu_mmu_context.c
330
dma_addr += map_size;
drivers/accel/ivpu/ivpu_mmu_context.c
460
dma_addr_t dma_addr = sg_dma_address(sg) - sg->offset;
drivers/accel/ivpu/ivpu_mmu_context.c
464
ctx->id, dma_addr, vpu_addr, size);
drivers/accel/ivpu/ivpu_mmu_context.c
472
ret = ivpu_mmu_context_map_pages(vdev, ctx, vpu_addr, dma_addr, size, prot);
drivers/accel/ivpu/ivpu_mmu_context.c
49
dma_addr_t dma_addr;
drivers/accel/ivpu/ivpu_mmu_context.c
529
dma_addr_t dma_addr = sg_dma_address(sg) - sg->offset;
drivers/accel/ivpu/ivpu_mmu_context.c
533
ctx->id, dma_addr, vpu_addr, size);
drivers/accel/ivpu/ivpu_mmu_context.c
59
dma_addr = dma_map_page(vdev->drm.dev, page, 0, PAGE_SIZE, DMA_BIDIRECTIONAL);
drivers/accel/ivpu/ivpu_mmu_context.c
60
if (dma_mapping_error(vdev->drm.dev, dma_addr))
drivers/accel/ivpu/ivpu_mmu_context.c
68
*dma = dma_addr;
drivers/accel/ivpu/ivpu_mmu_context.c
72
dma_unmap_page(vdev->drm.dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL);
drivers/accel/ivpu/ivpu_mmu_context.c
79
static void ivpu_pgtable_free_page(struct ivpu_device *vdev, u64 *cpu_addr, dma_addr_t dma_addr)
drivers/accel/ivpu/ivpu_mmu_context.c
86
dma_unmap_page(vdev->drm.dev, dma_addr & ~IVPU_MMU_ENTRY_FLAGS_MASK, PAGE_SIZE,
drivers/accel/qaic/qaic.h
97
dma_addr_t dma_addr;
drivers/accel/qaic/qaic_control.c
212
dma_addr_t dma_addr;
drivers/accel/qaic/qaic_control.c
307
qdev->dbc[dbc_id].dma_addr = resources->dma_addr;
drivers/accel/qaic/qaic_control.c
321
resources->dma_addr);
drivers/accel/qaic/qaic_control.c
640
dma_addr_t dma_addr;
drivers/accel/qaic/qaic_control.c
671
buf = dma_alloc_coherent(&qdev->pdev->dev, size, &dma_addr, GFP_KERNEL);
drivers/accel/qaic/qaic_control.c
687
out_trans->req_q_addr = cpu_to_le64(dma_addr);
drivers/accel/qaic/qaic_control.c
689
out_trans->rsp_q_addr = cpu_to_le64(dma_addr + size - nelem * get_dbc_rsp_elem_size());
drivers/accel/qaic/qaic_control.c
698
resources->dma_addr = dma_addr;
drivers/accel/qaic/qaic_control.c
705
dma_free_coherent(&qdev->pdev->dev, size, buf, dma_addr);
drivers/accel/qaic/qaic_data.c
2019
dma_free_coherent(&qdev->pdev->dev, dbc->total_size, dbc->req_q_base, dbc->dma_addr);
drivers/accel/qaic/qaic_data.c
2022
dbc->dma_addr = 0;
drivers/atm/fore200e.c
1015
dma_sync_single_for_cpu(fore200e->dev, buffer->data.dma_addr,
drivers/atm/fore200e.c
1021
dma_sync_single_for_device(fore200e->dev, buffer->data.dma_addr,
drivers/atm/fore200e.c
178
chunk->dma_addr = dma_map_single(fore200e->dev, chunk->align_addr,
drivers/atm/fore200e.c
180
if (dma_mapping_error(fore200e->dev, chunk->dma_addr)) {
drivers/atm/fore200e.c
193
dma_unmap_single(fore200e->dev, chunk->dma_addr, chunk->dma_size,
drivers/atm/fore200e.c
2095
FORE200E_DMA_INDEX(bsq->rbd_block.dma_addr, struct rbd_block, i);
drivers/atm/fore200e.c
210
&chunk->dma_addr, GFP_KERNEL);
drivers/atm/fore200e.c
2100
fore200e->bus->write(FORE200E_DMA_INDEX(bsq->status.dma_addr, enum status, i),
drivers/atm/fore200e.c
2150
FORE200E_DMA_INDEX(rxq->rpd.dma_addr, struct rpd, i);
drivers/atm/fore200e.c
2155
fore200e->bus->write(FORE200E_DMA_INDEX(rxq->status.dma_addr, enum status, i),
drivers/atm/fore200e.c
2158
fore200e->bus->write(FORE200E_DMA_INDEX(rxq->rpd.dma_addr, struct rpd, i),
drivers/atm/fore200e.c
2209
FORE200E_DMA_INDEX(txq->tpd.dma_addr, struct tpd, i);
drivers/atm/fore200e.c
2214
fore200e->bus->write(FORE200E_DMA_INDEX(txq->status.dma_addr, enum status, i),
drivers/atm/fore200e.c
224
chunk->dma_addr);
drivers/atm/fore200e.c
2261
fore200e->bus->write(FORE200E_DMA_INDEX(cmdq->status.dma_addr, enum status, i),
drivers/atm/fore200e.c
78
#define FORE200E_DMA_INDEX(dma_addr, type, index) ((dma_addr) + (index) * sizeof(type))
drivers/atm/fore200e.c
943
entry->rbd_block->rbd[ i ].buffer_haddr = buffer->data.dma_addr;
drivers/atm/fore200e.h
567
dma_addr_t dma_addr; /* DMA address of aligned chunk */
drivers/atm/iphase.c
1106
u_int buf_addr, dma_addr;
drivers/atm/iphase.c
1171
dma_addr = (buf_desc_ptr->dma_start_hi << 16) | buf_desc_ptr->dma_start_lo;
drivers/atm/iphase.c
1172
len = dma_addr - buf_addr;
drivers/atm/iphase.c
1996
iadev->tx_buf[i].dma_addr = dma_map_single(&iadev->pci->dev,
drivers/atm/iphase.c
2205
dma_unmap_single(&iadev->pci->dev, desc->dma_addr,
drivers/atm/iphase.c
2485
dma_unmap_single(&iadev->pci->dev, desc->dma_addr,
drivers/atm/iphase.c
3020
wr_ptr->sys_pkt_addr = iadev->tx_buf[desc-1].dma_addr;
drivers/atm/iphase.h
203
dma_addr_t dma_addr;
drivers/atm/solos-pci.c
1074
dma_unmap_single(&card->dev->dev, SKB_CB(oldskb)->dma_addr,
drivers/atm/solos-pci.c
1094
SKB_CB(skb)->dma_addr = dma_map_single(&card->dev->dev, data,
drivers/atm/solos-pci.c
1097
iowrite32(SKB_CB(skb)->dma_addr,
drivers/atm/solos-pci.c
1415
dma_unmap_single(&card->dev->dev, SKB_CB(skb)->dma_addr,
drivers/atm/solos-pci.c
1421
dma_unmap_single(&card->dev->dev, SKB_CB(skb)->dma_addr,
drivers/atm/solos-pci.c
782
dma_unmap_single(&card->dev->dev, SKB_CB(skb)->dma_addr,
drivers/atm/solos-pci.c
876
SKB_CB(skb)->dma_addr =
drivers/atm/solos-pci.c
879
iowrite32(SKB_CB(skb)->dma_addr,
drivers/atm/solos-pci.c
93
uint32_t dma_addr;
drivers/block/mtip32xx/mtip32xx.c
1551
dma_addr_t dma_addr = 0;
drivers/block/mtip32xx/mtip32xx.c
1561
&dma_addr,
drivers/block/mtip32xx/mtip32xx.c
1604
(xfer_sz ? dma_addr : 0),
drivers/block/mtip32xx/mtip32xx.c
1637
ATA_SECT_SIZE * xfer_sz, buf, dma_addr);
drivers/block/n64cart.c
64
dma_addr_t dma_addr;
drivers/block/n64cart.c
71
dma_addr = dma_map_bvec(dev, bv, DMA_FROM_DEVICE, 0);
drivers/block/n64cart.c
72
if (dma_mapping_error(dev, dma_addr))
drivers/block/n64cart.c
77
n64cart_write_reg(PI_DRAM_REG, dma_addr);
drivers/block/n64cart.c
83
dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE);
drivers/bus/mhi/host/boot.c
229
upper_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
232
lower_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
277
mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_HIGH, upper_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
278
mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_LOW, lower_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
310
dma_free_coherent(mhi_cntrl->cntrl_dev, mhi_buf->len, mhi_buf->buf, mhi_buf->dma_addr);
drivers/bus/mhi/host/boot.c
323
mhi_buf->buf, mhi_buf->dma_addr);
drivers/bus/mhi/host/boot.c
34
bhi_vec->dma_addr = cpu_to_le64(mhi_buf->dma_addr);
drivers/bus/mhi/host/boot.c
350
&mhi_buf->dma_addr, GFP_KERNEL);
drivers/bus/mhi/host/boot.c
398
vec_size, &mhi_buf->dma_addr,
drivers/bus/mhi/host/boot.c
41
upper_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
413
mhi_buf->buf, mhi_buf->dma_addr);
drivers/bus/mhi/host/boot.c
433
bhi_vec->dma_addr = cpu_to_le64(mhi_buf->dma_addr);
drivers/bus/mhi/host/boot.c
44
lower_32_bits(mhi_buf->dma_addr));
drivers/bus/mhi/host/boot.c
57
&mhi_buf->dma_addr, mhi_buf->len, sequence_id);
drivers/bus/mhi/host/internal.h
28
__le64 dma_addr;
drivers/char/agp/alpha-agp.c
17
dma_addr_t dma_addr;
drivers/char/agp/alpha-agp.c
21
dma_addr = vmf->address - vmf->vma->vm_start + agp->aperture.bus_base;
drivers/char/agp/alpha-agp.c
22
pa = agp->ops->translate(agp, dma_addr);
drivers/char/agp/intel-gtt.c
299
dma_addr_t dma_addr;
drivers/char/agp/intel-gtt.c
307
dma_addr = dma_map_page(&intel_private.pcidev->dev, page, 0,
drivers/char/agp/intel-gtt.c
309
if (dma_mapping_error(&intel_private.pcidev->dev, dma_addr)) {
drivers/char/agp/intel-gtt.c
314
intel_private.scratch_page_dma = dma_addr;
drivers/char/xillybus/xillybus.h
113
dma_addr_t dma_addr;
drivers/char/xillybus/xillybus.h
27
dma_addr_t dma_addr;
drivers/char/xillybus/xillybus_core.c
1092
channel->rd_buffers[bufidx]->dma_addr,
drivers/char/xillybus/xillybus_core.c
1310
channel->rd_buffers[bufidx]->dma_addr,
drivers/char/xillybus/xillybus_core.c
1331
channel->rd_buffers[bufidx]->dma_addr,
drivers/char/xillybus/xillybus_core.c
306
dma_unmap_single(data->device, data->dma_addr,
drivers/char/xillybus/xillybus_core.c
334
this->dma_addr = addr;
drivers/char/xillybus/xillybus_core.c
349
dma_addr_t dma_addr;
drivers/char/xillybus/xillybus_core.c
396
&dma_addr);
drivers/char/xillybus/xillybus_core.c
400
iowrite32((u32) (dma_addr & 0xffffffff),
drivers/char/xillybus/xillybus_core.c
402
iowrite32(((u32) ((((u64) dma_addr) >> 32) & 0xffffffff)),
drivers/char/xillybus/xillybus_core.c
407
this_buffer->dma_addr = dma_addr;
drivers/char/xillybus/xillybus_core.c
414
ep->msgbuf_dma_addr = dma_addr;
drivers/char/xillybus/xillybus_core.c
660
channel->wr_buffers[0]->dma_addr,
drivers/char/xillybus/xillybus_core.c
774
channel->wr_buffers[bufidx]->dma_addr,
drivers/char/xillybus/xillybus_core.c
789
channel->wr_buffers[bufidx]->dma_addr,
drivers/comedi/comedi_buf.c
34
buf->dma_addr);
drivers/comedi/comedi_buf.c
93
&buf->dma_addr, GFP_KERNEL);
drivers/comedi/comedi_fops.c
2562
buf->dma_addr, PAGE_SIZE);
drivers/comedi/drivers/mite.c
555
writel(mite_chan->ring->dma_addr,
drivers/comedi/drivers/mite.c
672
desc->addr = cpu_to_le32(async->buf_map->page_list[i].dma_addr);
drivers/comedi/drivers/mite.c
673
desc->next = cpu_to_le32(ring->dma_addr +
drivers/comedi/drivers/mite.c
682
desc->addr = cpu_to_le32(async->buf_map->page_list[i].dma_addr);
drivers/comedi/drivers/mite.c
686
desc->next = cpu_to_le32(ring->dma_addr);
drivers/comedi/drivers/mite.c
704
descs, ring->dma_addr);
drivers/comedi/drivers/mite.c
706
ring->dma_addr = 0;
drivers/comedi/drivers/mite.c
731
&ring->dma_addr, GFP_KERNEL);
drivers/comedi/drivers/mite.c
762
ring->dma_addr = 0;
drivers/comedi/drivers/mite.h
33
dma_addr_t dma_addr;
drivers/crypto/atmel-sha.c
720
ctx->dma_addr = dma_map_single(dd->dev, ctx->buffer,
drivers/crypto/atmel-sha.c
722
if (dma_mapping_error(dd->dev, ctx->dma_addr)) {
drivers/crypto/atmel-sha.c
731
return atmel_sha_xmit_start(dd, ctx->dma_addr, length, 0, 0, final);
drivers/crypto/atmel-sha.c
811
ctx->dma_addr = dma_map_single(dd->dev, ctx->buffer,
drivers/crypto/atmel-sha.c
813
if (dma_mapping_error(dd->dev, ctx->dma_addr)) {
drivers/crypto/atmel-sha.c
823
return atmel_sha_xmit_start(dd, ctx->dma_addr, count, 0,
drivers/crypto/atmel-sha.c
838
length, ctx->dma_addr, count, final);
drivers/crypto/atmel-sha.c
866
dma_unmap_single(dd->dev, ctx->dma_addr,
drivers/crypto/atmel-sha.c
870
dma_unmap_single(dd->dev, ctx->dma_addr, ctx->buflen +
drivers/crypto/atmel-sha.c
98
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
234
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
610
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
617
dma_addr = dma_map_page(dev, page, offset, size, dir);
drivers/crypto/axis/artpec6_crypto.c
618
if (dma_mapping_error(dev, dma_addr))
drivers/crypto/axis/artpec6_crypto.c
623
map->dma_addr = dma_addr;
drivers/crypto/axis/artpec6_crypto.c
626
*dma_addr_out = dma_addr;
drivers/crypto/axis/artpec6_crypto.c
686
dma_unmap_page(dev, map->dma_addr, map->size, map->dir);
drivers/crypto/axis/artpec6_crypto.c
714
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
718
&dma_addr);
drivers/crypto/axis/artpec6_crypto.c
722
return artpec6_crypto_setup_out_descr_phys(common, dma_addr,
drivers/crypto/axis/artpec6_crypto.c
770
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
774
DMA_FROM_DEVICE, &dma_addr);
drivers/crypto/axis/artpec6_crypto.c
778
return artpec6_crypto_setup_in_descr_phys(common, dma_addr, len, last);
drivers/crypto/axis/artpec6_crypto.c
851
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
863
&dma_addr);
drivers/crypto/axis/artpec6_crypto.c
868
dma_addr,
drivers/crypto/axis/artpec6_crypto.c
912
dma_addr_t dma_addr;
drivers/crypto/axis/artpec6_crypto.c
920
&dma_addr);
drivers/crypto/axis/artpec6_crypto.c
925
dma_addr,
drivers/crypto/caam/caamalg.c
3684
dma_addr_t dma_addr;
drivers/crypto/caam/caamalg.c
3701
dma_addr = dma_map_single_attrs(ctx->jrdev, ctx->sh_desc_enc,
drivers/crypto/caam/caamalg.c
3706
if (dma_mapping_error(ctx->jrdev, dma_addr)) {
drivers/crypto/caam/caamalg.c
3712
ctx->sh_desc_enc_dma = dma_addr;
drivers/crypto/caam/caamalg.c
3713
ctx->sh_desc_dec_dma = dma_addr + offsetof(struct caam_ctx,
drivers/crypto/caam/caamalg.c
3716
ctx->key_dma = dma_addr + offsetof(struct caam_ctx, key) -
drivers/crypto/caam/caamalg_qi2.c
1580
dma_addr_t dma_addr;
drivers/crypto/caam/caamalg_qi2.c
1590
dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc,
drivers/crypto/caam/caamalg_qi2.c
1593
if (dma_mapping_error(ctx->dev, dma_addr)) {
drivers/crypto/caam/caamalg_qi2.c
1599
ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
drivers/crypto/caam/caamalg_qi2.c
1600
ctx->key_dma = dma_addr + NUM_OP * sizeof(ctx->flc[0]);
drivers/crypto/caam/caamalg_qi2.c
4572
dma_addr_t dma_addr;
drivers/crypto/caam/caamalg_qi2.c
4588
dma_addr = dma_map_single_attrs(ctx->dev, ctx->flc, sizeof(ctx->flc),
drivers/crypto/caam/caamalg_qi2.c
4591
if (dma_mapping_error(ctx->dev, dma_addr)) {
drivers/crypto/caam/caamalg_qi2.c
4602
ctx->flc_dma[i] = dma_addr + i * sizeof(ctx->flc[i]);
drivers/crypto/caam/caamhash.c
1782
dma_addr_t dma_addr;
drivers/crypto/caam/caamhash.c
1833
dma_addr = dma_map_single_attrs(ctx->jrdev, ctx->sh_desc_update,
drivers/crypto/caam/caamhash.c
1837
if (dma_mapping_error(ctx->jrdev, dma_addr)) {
drivers/crypto/caam/caamhash.c
1850
ctx->sh_desc_update_dma = dma_addr;
drivers/crypto/caam/caamhash.c
1851
ctx->sh_desc_update_first_dma = dma_addr +
drivers/crypto/caam/caamhash.c
1855
ctx->sh_desc_fin_dma = dma_addr + offsetof(struct caam_hash_ctx,
drivers/crypto/caam/caamhash.c
1858
ctx->sh_desc_digest_dma = dma_addr + offsetof(struct caam_hash_ctx,
drivers/crypto/caam/key_gen.c
100
append_fifo_store(desc, dma_addr, adata->keylen,
drivers/crypto/caam/key_gen.c
121
dma_unmap_single(jrdev, dma_addr, local_max, DMA_BIDIRECTIONAL);
drivers/crypto/caam/key_gen.c
50
dma_addr_t dma_addr;
drivers/crypto/caam/key_gen.c
75
dma_addr = dma_map_single(jrdev, key_out, local_max, DMA_BIDIRECTIONAL);
drivers/crypto/caam/key_gen.c
76
if (dma_mapping_error(jrdev, dma_addr)) {
drivers/crypto/caam/key_gen.c
82
append_key(desc, dma_addr, keylen, CLASS_2 | KEY_DEST_CLASS_REG);
drivers/crypto/cavium/cpt/cptvf.h
46
dma_addr_t dma_addr;
drivers/crypto/cavium/cpt/cptvf_main.c
186
chunk->dma_addr);
drivers/crypto/cavium/cpt/cptvf_main.c
188
chunk->dma_addr = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
236
&curr->dma_addr,
drivers/crypto/cavium/cpt/cptvf_main.c
258
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/cavium/cpt/cptvf_main.c
266
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/cavium/cpt/cptvf_main.c
651
base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
100
sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
110
if (list[j].dma_addr) {
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
111
dma_unmap_single(&pdev->dev, list[i].dma_addr,
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
115
list[j].dma_addr = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
288
if (req->out[i].dma_addr)
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
290
req->out[i].dma_addr,
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
296
if (req->in[i].dma_addr)
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
298
req->in[i].dma_addr,
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
59
list[i].dma_addr = dma_map_single(&pdev->dev,
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
64
list[i].dma_addr))) {
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
80
sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
81
sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
82
sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
83
sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
92
sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
96
sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr);
drivers/crypto/cavium/cpt/request_manager.h
54
dma_addr_t dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
1000
areq_ctx->ccm_iv0_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
1009
dma_addr = dma_map_single(dev, areq_ctx->hkey, AES_BLOCK_SIZE,
drivers/crypto/ccree/cc_buffer_mgr.c
1011
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/ccree/cc_buffer_mgr.c
1017
areq_ctx->hkey_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
1019
dma_addr = dma_map_single(dev, &areq_ctx->gcm_len_block,
drivers/crypto/ccree/cc_buffer_mgr.c
1021
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/ccree/cc_buffer_mgr.c
1027
areq_ctx->gcm_block_len_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
1029
dma_addr = dma_map_single(dev, areq_ctx->gcm_iv_inc1,
drivers/crypto/ccree/cc_buffer_mgr.c
1032
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/ccree/cc_buffer_mgr.c
1039
areq_ctx->gcm_iv_inc1_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
1041
dma_addr = dma_map_single(dev, areq_ctx->gcm_iv_inc2,
drivers/crypto/ccree/cc_buffer_mgr.c
1044
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/ccree/cc_buffer_mgr.c
1051
areq_ctx->gcm_iv_inc2_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
954
dma_addr_t dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
977
dma_addr = dma_map_single(dev, areq_ctx->mac_buf, MAX_MAC_SIZE,
drivers/crypto/ccree/cc_buffer_mgr.c
979
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/ccree/cc_buffer_mgr.c
985
areq_ctx->mac_buf_dma_addr = dma_addr;
drivers/crypto/ccree/cc_buffer_mgr.c
990
dma_addr = dma_map_single(dev, addr, AES_BLOCK_SIZE,
drivers/crypto/ccree/cc_buffer_mgr.c
993
if (dma_mapping_error(dev, dma_addr)) {
drivers/crypto/chelsio/chcr_algo.c
1595
req_ctx->hctx_wr.dma_addr =
drivers/crypto/chelsio/chcr_algo.c
1599
req_ctx->hctx_wr. dma_addr)) {
drivers/crypto/chelsio/chcr_algo.c
1605
req_ctx->hctx_wr.dma_addr = 0;
drivers/crypto/chelsio/chcr_algo.c
2075
if (hctx_wr->dma_addr) {
drivers/crypto/chelsio/chcr_algo.c
2076
dma_unmap_single(&u_ctx->lldi.pdev->dev, hctx_wr->dma_addr,
drivers/crypto/chelsio/chcr_algo.c
2078
hctx_wr->dma_addr = 0;
drivers/crypto/chelsio/chcr_algo.c
2706
reqctx->hctx_wr.dma_addr);
drivers/crypto/chelsio/chcr_crypto.h
269
dma_addr_t dma_addr;
drivers/crypto/hisilicon/qm.c
693
int hisi_qm_mb(struct hisi_qm *qm, u8 cmd, dma_addr_t dma_addr, u16 queue,
drivers/crypto/hisilicon/qm.c
711
qm_mb_pre_init(&mailbox, cmd, dma_addr, queue, op);
drivers/crypto/img-hash.c
90
dma_addr_t dma_addr;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
246
rctx->src_dll.dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
247
rctx->dst_dll.dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
248
rctx->aad_src_dll.dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
249
rctx->aad_dst_dll.dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
368
dma_free_coherent(dev, dll->size, dll->vaddr, dll->dma_addr);
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
371
dll->dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
442
rctx->src_dll.dma_addr = rctx->dst_dll.dma_addr;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
499
rc = ocs_aes_bypass_op(tctx->aes_dev, rctx->dst_dll.dma_addr,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
500
rctx->src_dll.dma_addr, req->cryptlen);
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
550
rctx->dst_dll.dma_addr, rctx->src_dll.dma_addr,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
823
rctx->src_dll.dma_addr = rctx->dst_dll.dma_addr;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
824
rctx->aad_src_dll.dma_addr = rctx->aad_dst_dll.dma_addr;
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
878
rc = ocs_aes_bypass_op(tctx->aes_dev, rctx->aad_dst_dll.dma_addr,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
879
rctx->aad_src_dll.dma_addr, req->cryptlen);
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
902
rctx->instruction, rctx->dst_dll.dma_addr,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
903
rctx->src_dll.dma_addr, in_size,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
905
rctx->aad_src_dll.dma_addr, req->assoclen,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
912
rctx->dst_dll.dma_addr,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
913
rctx->src_dll.dma_addr, in_size,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
915
rctx->aad_src_dll.dma_addr, req->assoclen,
drivers/crypto/intel/keembay/ocs-aes.c
1435
dll_desc->dma_addr = DMA_MAPPING_ERROR;
drivers/crypto/intel/keembay/ocs-aes.c
1469
&dll_desc->dma_addr, GFP_KERNEL);
drivers/crypto/intel/keembay/ocs-aes.c
1481
ll[i].next = dll_desc->dma_addr + (sizeof(*ll) * (i + 1));
drivers/crypto/intel/keembay/ocs-aes.h
63
dma_addr_t dma_addr;
drivers/crypto/intel/keembay/ocs-hcu.c
125
dma_addr_t dma_addr;
drivers/crypto/intel/keembay/ocs-hcu.c
471
writel(dma_list->dma_addr, hcu_dev->io_base + OCS_HCU_DMA_NEXT_SRC_DESCR);
drivers/crypto/intel/keembay/ocs-hcu.c
501
&dma_list->dma_addr, GFP_KERNEL);
drivers/crypto/intel/keembay/ocs-hcu.c
520
dma_list->head, dma_list->dma_addr);
drivers/crypto/intel/keembay/ocs-hcu.c
563
old_tail->nxt_desc = dma_list->dma_addr +
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
12
prep_admin_req_msg(struct rl_sla *sla, dma_addr_t dma_addr,
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
17
req->init_cfg_ptr = dma_addr;
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
76
dma_addr_t dma_addr;
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
80
&dma_addr, GFP_KERNEL);
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
85
prep_admin_req_msg(sla, dma_addr, fw_params, &req, is_update);
drivers/crypto/intel/qat/qat_common/adf_rl_admin.c
89
dma_addr);
drivers/crypto/intel/qat/qat_common/adf_transport.c
172
ring_size_bytes, &ring->dma_addr,
drivers/crypto/intel/qat/qat_common/adf_transport.c
179
if (adf_check_ring_alignment(ring->dma_addr, ring_size_bytes)) {
drivers/crypto/intel/qat/qat_common/adf_transport.c
182
ring->base_addr, ring->dma_addr);
drivers/crypto/intel/qat/qat_common/adf_transport.c
193
ring_base = csr_ops->build_csr_ring_base_addr(ring->dma_addr,
drivers/crypto/intel/qat/qat_common/adf_transport.c
213
ring->dma_addr);
drivers/crypto/intel/qat/qat_common/adf_transport_internal.h
20
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx/otx_cptpf_ucode.c
179
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx/otx_cptpf_ucode.c
188
dma_addr =
drivers/crypto/marvell/octeontx/otx_cptpf_ucode.c
191
dma_addr = eng_grp->ucode[0].align_dma;
drivers/crypto/marvell/octeontx/otx_cptpf_ucode.c
199
writeq((u64) dma_addr, cpt->reg_base +
drivers/crypto/marvell/octeontx/otx_cptvf.h
30
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
174
chunk->dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
176
chunk->dma_addr = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
223
&curr->dma_addr, GFP_KERNEL);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
243
(u64)curr->dma_addr;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
252
*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
634
base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
111
list[i].dma_addr = dma_map_single(&pdev->dev,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
116
list[i].dma_addr))) {
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
131
sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
132
sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
133
sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
134
sg_ptr->ptr3 = cpu_to_be64(list[i * 4 + 3].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
142
sg_ptr->ptr2 = cpu_to_be64(list[i * 4 + 2].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
146
sg_ptr->ptr1 = cpu_to_be64(list[i * 4 + 1].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
150
sg_ptr->ptr0 = cpu_to_be64(list[i * 4 + 0].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
159
if (list[j].dma_addr) {
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
160
dma_unmap_single(&pdev->dev, list[i].dma_addr,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
164
list[j].dma_addr = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
39
(void *) req->in[i].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
50
(void *) req->out[i].dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
203
if (req->out[i].dma_addr)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
205
req->out[i].dma_addr,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
211
if (req->in[i].dma_addr)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
213
req->in[i].dma_addr,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
67
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
107
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
197
if (req->out[i].dma_addr)
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
199
req->out[i].dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
205
if (req->in[i].dma_addr)
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
207
req->in[i].dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
231
list[i].dma_addr = dma_map_single(&pdev->dev, list[i].vptr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
234
if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) {
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
246
sg_ptr->ptr0 = cpu_to_be64(list[i * SG_COMPS_MAX + 0].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
247
sg_ptr->ptr1 = cpu_to_be64(list[i * SG_COMPS_MAX + 1].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
248
sg_ptr->ptr2 = cpu_to_be64(list[i * SG_COMPS_MAX + 2].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
249
sg_ptr->ptr3 = cpu_to_be64(list[i * SG_COMPS_MAX + 3].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
257
sg_ptr->ptr2 = cpu_to_be64(list[i * SG_COMPS_MAX + 2].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
261
sg_ptr->ptr1 = cpu_to_be64(list[i * SG_COMPS_MAX + 1].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
265
sg_ptr->ptr0 = cpu_to_be64(list[i * SG_COMPS_MAX + 0].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
274
if (list[j].dma_addr) {
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
275
dma_unmap_single(&pdev->dev, list[j].dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
279
list[j].dma_addr = 0;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
300
list[i].dma_addr = dma_map_single(&pdev->dev, list[i].vptr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
303
if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) {
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
314
sg_ptr->ptr0 = list[i * SGV2_COMPS_MAX + 0].dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
315
sg_ptr->ptr1 = list[i * SGV2_COMPS_MAX + 1].dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
316
sg_ptr->ptr2 = list[i * SGV2_COMPS_MAX + 2].dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
326
sg_ptr->ptr1 = list[i * SGV2_COMPS_MAX + 1].dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
330
sg_ptr->ptr0 = list[i * SGV2_COMPS_MAX + 0].dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
339
if (list[j].dma_addr) {
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
340
dma_unmap_single(&pdev->dev, list[j].dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
344
list[j].dma_addr = 0;
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
177
iq->dma_addr = iq->real_dma_addr + OTX2_CPT_INST_GRP_QLEN_BYTES;
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
181
iq->dma_addr = PTR_ALIGN(iq->dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
198
lf_q_base.u = lfs->lf[slot].iqueue.dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
71
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
163
dma_addr_t dma_addr, int blkaddr)
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
167
(u64)dma_addr, blkaddr);
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
174
dma_addr_t dma_addr;
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
191
dma_addr = engs->ucode->dma;
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
199
ret = __write_ucode_base(cptpf, bit, dma_addr,
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
28
req->in[i].dma_addr);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
38
req->out[i].dma_addr);
drivers/crypto/sa2ul.c
706
static void sa_dump_sc(u8 *buf, dma_addr_t dma_addr)
drivers/crypto/sa2ul.c
709
dev_info(sa_k3_dev, "Security context dump:: 0x%pad\n", &dma_addr);
drivers/crypto/stm32/stm32-hash.c
168
dma_addr_t dma_addr;
drivers/crypto/talitos.c
101
dma_addr_t dma_addr = dma_map_single_attrs(dev, data, len, dir, attrs);
drivers/crypto/talitos.c
105
to_talitos_ptr(ptr, dma_addr, len, is_sec1);
drivers/crypto/talitos.c
46
static void to_talitos_ptr(struct talitos_ptr *ptr, dma_addr_t dma_addr,
drivers/crypto/talitos.c
49
ptr->ptr = cpu_to_be32(lower_32_bits(dma_addr));
drivers/crypto/talitos.c
54
ptr->eptr = upper_32_bits(dma_addr);
drivers/dibs/dibs_loopback.c
105
dmb->dma_addr = dmb_node->dma_addr;
drivers/dibs/dibs_loopback.c
203
dmb->dma_addr = dmb_node->dma_addr;
drivers/dibs/dibs_loopback.c
85
dmb_node->dma_addr = DIBS_DMA_ADDR_INVALID;
drivers/dibs/dibs_loopback.h
31
dma_addr_t dma_addr;
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
735
dma_addr_t dma_addr, size_t buf_len,
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
746
mem_width = __ffs(data_width | dma_addr | buf_len);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
764
dw_axi_dma_chan_prep_cyclic(struct dma_chan *dchan, dma_addr_t dma_addr,
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
772
dma_addr_t src_addr = dma_addr;
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
784
axi_block_len = calculate_block_len(chan, dma_addr, buf_len, direction);
drivers/dma/ep93xx_dma.c
1167
ep93xx_dma_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr,
drivers/dma/ep93xx_dma.c
1205
desc->src_addr = dma_addr + offset;
drivers/dma/ep93xx_dma.c
1209
desc->dst_addr = dma_addr + offset;
drivers/dma/fsl-edma-common.c
349
dma_addr_t cur_addr, dma_addr, old_addr;
drivers/dma/fsl-edma-common.c
385
dma_addr = fsl_edma_get_tcd_to_cpu(fsl_chan, edesc->tcd[i].vtcd, saddr);
drivers/dma/fsl-edma-common.c
387
dma_addr = fsl_edma_get_tcd_to_cpu(fsl_chan, edesc->tcd[i].vtcd, daddr);
drivers/dma/fsl-edma-common.c
390
if (cur_addr >= dma_addr && cur_addr < dma_addr + size) {
drivers/dma/fsl-edma-common.c
391
len += dma_addr + size - cur_addr;
drivers/dma/fsl-edma-common.c
591
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/fsl-edma-common.c
617
dma_buf_next = dma_addr;
drivers/dma/fsl-edma-common.c
639
if (dma_buf_next >= dma_addr + buf_len)
drivers/dma/fsl-edma-common.c
640
dma_buf_next = dma_addr;
drivers/dma/fsl-edma-common.h
496
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/idxd/device.c
767
dma_addr_t dma_addr;
drivers/dma/idxd/device.c
788
addr = dma_alloc_coherent(dev, size, &dma_addr, GFP_KERNEL);
drivers/dma/idxd/device.c
796
evl->dma = dma_addr;
drivers/dma/idxd/device.c
801
evlcfg.bits[0] = dma_addr & GENMASK(63, 12);
drivers/dma/imx-dma.c
847
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/imx-dma.c
878
imxdmac->sg_list[i].dma_address = dma_addr;
drivers/dma/imx-dma.c
880
dma_addr += period_len;
drivers/dma/imx-sdma.c
1726
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/imx-sdma.c
1765
bd->buffer_addr = dma_addr;
drivers/dma/imx-sdma.c
1781
i, period_len, (u64)dma_addr,
drivers/dma/imx-sdma.c
1787
dma_addr += period_len;
drivers/dma/mmp_tdma.c
423
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/mmp_tdma.c
468
desc->src_addr = dma_addr;
drivers/dma/mmp_tdma.c
472
desc->dst_addr = dma_addr;
drivers/dma/mmp_tdma.c
475
dma_addr += period_len;
drivers/dma/mxs-dma.c
574
struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len,
drivers/dma/mxs-dma.c
612
ccw->bufaddr = dma_addr;
drivers/dma/mxs-dma.c
624
dma_addr += period_len;
drivers/dma/pl330.c
2678
struct dma_chan *chan, dma_addr_t dma_addr, size_t len,
drivers/dma/pl330.c
2733
src = dma_addr;
drivers/dma/pl330.c
2740
dst = dma_addr;
drivers/dma/pl330.c
2757
dma_addr += period_len;
drivers/dma/ppc4xx/adma.c
3703
dma_addr_t dma_addr, addrs[2];
drivers/dma/ppc4xx/adma.c
3731
dma_addr = dma_map_page(chan->device->dev, pg, 0,
drivers/dma/ppc4xx/adma.c
3735
ppc440spe_adma_pq_set_src(sw_desc, dma_addr, 0);
drivers/dma/ppc4xx/adma.c
3737
addrs[0] = dma_addr;
drivers/dma/qcom/qcom_adm.c
125
dma_addr_t dma_addr;
drivers/dma/qcom/qcom_adm.c
440
async_desc->dma_addr = dma_map_single(adev->dev, async_desc->cpl,
drivers/dma/qcom/qcom_adm.c
443
if (dma_mapping_error(adev->dev, async_desc->dma_addr)) {
drivers/dma/qcom/qcom_adm.c
448
cple_addr = async_desc->dma_addr + ((void *)cple - async_desc->cpl);
drivers/dma/qcom/qcom_adm.c
454
*cple |= (async_desc->dma_addr + ADM_DESC_ALIGN) >> 3;
drivers/dma/qcom/qcom_adm.c
557
writel(ALIGN(async_desc->dma_addr, ADM_DESC_ALIGN) >> 3,
drivers/dma/qcom/qcom_adm.c
690
dma_unmap_single(async_desc->adev->dev, async_desc->dma_addr,
drivers/dma/ste_dma40.c
2180
dma_sync_single_for_device(chan->base->dev, desc->lli_pool.dma_addr,
drivers/dma/ste_dma40.c
2523
dma40_prep_dma_cyclic(struct dma_chan *chan, dma_addr_t dma_addr,
drivers/dma/ste_dma40.c
2537
sg_dma_address(&sg[i]) = dma_addr;
drivers/dma/ste_dma40.c
2539
dma_addr += period_len;
drivers/dma/ste_dma40.c
3422
pool->dma_addr = dma_map_single(base->dev, pool->base,
drivers/dma/ste_dma40.c
3425
if (dma_mapping_error(base->dev, pool->dma_addr)) {
drivers/dma/ste_dma40.c
3426
pool->dma_addr = 0;
drivers/dma/ste_dma40.c
3640
if (base->lcla_pool.dma_addr)
drivers/dma/ste_dma40.c
3641
dma_unmap_single(base->dev, base->lcla_pool.dma_addr,
drivers/dma/ste_dma40.c
365
dma_addr_t dma_addr;
drivers/dma/ste_dma40.c
421
dma_addr_t dma_addr;
drivers/dma/ste_dma40.c
687
d40d->lli_pool.dma_addr = 0;
drivers/dma/ste_dma40.c
692
d40d->lli_pool.dma_addr = dma_map_single(d40c->base->dev,
drivers/dma/ste_dma40.c
698
d40d->lli_pool.dma_addr)) {
drivers/dma/ste_dma40.c
701
d40d->lli_pool.dma_addr = 0;
drivers/dma/ste_dma40.c
711
if (d40d->lli_pool.dma_addr)
drivers/dma/ste_dma40.c
712
dma_unmap_single(d40c->base->dev, d40d->lli_pool.dma_addr,
drivers/dma/ste_dma40.c
948
pool->dma_addr, lcla_offset,
drivers/dma/xilinx/xdma.c
172
desc->next_desc = cpu_to_le64(block[1].dma_addr);
drivers/dma/xilinx/xdma.c
202
desc->next_desc = cpu_to_le64(block->dma_addr + ((i + 1) * XDMA_DESC_SIZE));
drivers/dma/xilinx/xdma.c
205
desc->next_desc = cpu_to_le64(block->dma_addr);
drivers/dma/xilinx/xdma.c
255
sw_desc->desc_blocks[i].dma_addr);
drivers/dma/xilinx/xdma.c
272
dma_addr_t dma_addr;
drivers/dma/xilinx/xdma.c
299
addr = dma_pool_alloc(chan->desc_pool, GFP_NOWAIT, &dma_addr);
drivers/dma/xilinx/xdma.c
304
sw_desc->desc_blocks[i].dma_addr = dma_addr;
drivers/dma/xilinx/xdma.c
356
val = lower_32_bits(block->dma_addr);
drivers/dma/xilinx/xdma.c
361
val = upper_32_bits(block->dma_addr);
drivers/dma/xilinx/xdma.c
51
dma_addr_t dma_addr;
drivers/dma/xilinx/xilinx_dpdma.c
188
dma_addr_t dma_addr;
drivers/dma/xilinx/xilinx_dpdma.c
508
dma_addr_t dma_addr[],
drivers/dma/xilinx/xilinx_dpdma.c
514
hw_desc->src_addr = lower_32_bits(dma_addr[0]);
drivers/dma/xilinx/xilinx_dpdma.c
518
upper_32_bits(dma_addr[0]));
drivers/dma/xilinx/xilinx_dpdma.c
523
addr[i - 1] = lower_32_bits(dma_addr[i]);
drivers/dma/xilinx/xilinx_dpdma.c
529
addr_msb = upper_32_bits(dma_addr[i]) & GENMASK(15, 0);
drivers/dma/xilinx/xilinx_dpdma.c
538
prev->hw.next_desc = lower_32_bits(sw_desc->dma_addr);
drivers/dma/xilinx/xilinx_dpdma.c
542
upper_32_bits(sw_desc->dma_addr));
drivers/dma/xilinx/xilinx_dpdma.c
557
dma_addr_t dma_addr;
drivers/dma/xilinx/xilinx_dpdma.c
559
sw_desc = dma_pool_zalloc(chan->desc_pool, GFP_ATOMIC, &dma_addr);
drivers/dma/xilinx/xilinx_dpdma.c
563
sw_desc->dma_addr = dma_addr;
drivers/dma/xilinx/xilinx_dpdma.c
579
dma_pool_free(chan->desc_pool, sw_desc, sw_desc->dma_addr);
drivers/dma/xilinx/xilinx_dpdma.c
603
dev_dbg(dev, "descriptor DMA addr: %pad\n", &sw_desc->dma_addr);
drivers/dma/xilinx/xilinx_dpdma.c
735
last->hw.next_desc = lower_32_bits(sw_desc->dma_addr);
drivers/dma/xilinx/xilinx_dpdma.c
739
upper_32_bits(sw_desc->dma_addr));
drivers/dma/xilinx/xilinx_dpdma.c
944
lower_32_bits(sw_desc->dma_addr));
drivers/dma/xilinx/xilinx_dpdma.c
948
upper_32_bits(sw_desc->dma_addr)));
drivers/firewire/core-iso.c
111
dma_addr_t dma_addr = buffer->dma_addrs[i];
drivers/firewire/core-iso.c
112
dma_unmap_page(card->device, dma_addr, PAGE_SIZE, buffer->direction);
drivers/firewire/core-iso.c
132
dma_addr_t dma_addr = buffer->dma_addrs[i];
drivers/firewire/core-iso.c
133
ssize_t offset = (ssize_t)completed - (ssize_t)dma_addr;
drivers/firewire/core-iso.c
70
dma_addr_t dma_addr = dma_map_page(card->device, buffer->pages[i], 0, PAGE_SIZE,
drivers/firewire/core-iso.c
72
if (dma_mapping_error(card->device, dma_addr))
drivers/firewire/core-iso.c
75
dma_addrs[i] = dma_addr;
drivers/firewire/ohci.c
3287
dma_addr_t dma_addr = buffer->dma_addrs[page];
drivers/firewire/ohci.c
3288
pd[i].data_address = cpu_to_le32(dma_addr + offset);
drivers/firewire/ohci.c
3291
dma_addr, offset, length,
drivers/firewire/ohci.c
3370
dma_addr_t dma_addr = buffer->dma_addrs[page];
drivers/firewire/ohci.c
3371
pd->data_address = cpu_to_le32(dma_addr + offset);
drivers/firewire/ohci.c
3373
dma_sync_single_range_for_device(device, dma_addr,
drivers/firewire/ohci.c
3433
dma_addr_t dma_addr = buffer->dma_addrs[page];
drivers/firewire/ohci.c
3434
d->data_address = cpu_to_le32(dma_addr + offset);
drivers/firewire/ohci.c
3437
dma_addr, offset, length,
drivers/firewire/ohci.c
549
dma_addr_t dma_addr = ctx->dma_addrs[i];
drivers/firewire/ohci.c
550
if (dma_addr)
drivers/firewire/ohci.c
551
dma_unmap_page(dev, dma_addr, PAGE_SIZE, DMA_FROM_DEVICE);
drivers/firewire/ohci.c
885
dma_addr_t dma_addr = dma_map_page(dev, pages[i], 0, PAGE_SIZE, DMA_FROM_DEVICE);
drivers/firewire/ohci.c
886
if (dma_mapping_error(dev, dma_addr))
drivers/firewire/ohci.c
888
dma_addrs[i] = dma_addr;
drivers/firewire/ohci.c
889
dma_sync_single_for_device(dev, dma_addr, PAGE_SIZE, DMA_FROM_DEVICE);
drivers/fpga/versal-fpga.c
25
dma_addr_t dma_addr = 0;
drivers/fpga/versal-fpga.c
29
kbuf = dma_alloc_coherent(mgr->dev.parent, size, &dma_addr, GFP_KERNEL);
drivers/fpga/versal-fpga.c
34
ret = zynqmp_pm_load_pdi(PDI_SRC_DDR, dma_addr);
drivers/fpga/versal-fpga.c
35
dma_free_coherent(mgr->dev.parent, size, kbuf, dma_addr);
drivers/fpga/zynqmp-fpga.c
44
dma_addr_t dma_addr;
drivers/fpga/zynqmp-fpga.c
51
kbuf = dma_alloc_coherent(priv->dev, size, &dma_addr, GFP_KERNEL);
drivers/fpga/zynqmp-fpga.c
62
ret = zynqmp_pm_fpga_load(dma_addr, size, eemi_flags);
drivers/fpga/zynqmp-fpga.c
64
dma_free_coherent(priv->dev, size, kbuf, dma_addr);
drivers/gpu/drm/adp/adp_drv.c
184
writel(obj->dma_addr + fb->offsets[0], adp->be + ADBE_FB_BASE);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
655
dma_addr_t dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
668
dma_addr = mem->bo->tbo.sg->sgl->dma_address;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
670
pr_debug("%d BO address before DMA mapping: %llx\n", mmio, dma_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
671
dma_addr = dma_map_resource(adev->dev, dma_addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
673
ret = dma_mapping_error(adev->dev, dma_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
676
pr_debug("%d BO address after DMA mapping: %llx\n", mmio, dma_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
678
ttm->sg = create_sg_table(dma_addr, mem->bo->tbo.sg->sgl->length);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
696
dma_unmap_resource(adev->dev, dma_addr, mem->bo->tbo.sg->sgl->length,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
125
dma_addr_t dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
145
dma_addr = dma_map_page(&adev->pdev->dev, p, 0, adev->gart.table_size,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
147
if (dma_mapping_error(&adev->pdev->dev, dma_addr)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
154
dev_info(adev->dev, "%s dma_addr:%pad\n", __func__, &dma_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
165
sg_dma_address(sg->sgl) = dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
348
int pages, dma_addr_t *dma_addr, uint64_t flags,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
361
page_base = dma_addr[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
421
int pages, dma_addr_t *dma_addr, uint64_t flags)
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
438
page_base = dma_addr[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
464
int pages, dma_addr_t *dma_addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.c
470
amdgpu_gart_map(adev, offset, pages, dma_addr, flags, adev->gart.ptr);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.h
63
int pages, dma_addr_t *dma_addr, uint64_t flags,
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.h
66
int pages, dma_addr_t *dma_addr, uint64_t flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_gart.h
68
int pages, dma_addr_t *dma_addr, uint64_t flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
57
dma_addr_t dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
66
&dma_addr, GFP_KERNEL);
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
70
ih->gpu_addr = dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
71
ih->wptr_addr = dma_addr + ih->ring_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
73
ih->rptr_addr = dma_addr + ih->ring_size + 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
262
dma_addr_t *dma_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
264
dma_addr = &bo->ttm->dma_address[mm_cur->start >> PAGE_SHIFT];
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
265
amdgpu_gart_map(adev, 0, num_pages, dma_addr, flags, cpu_addr);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1427
dma_addr_t *dma_addr, struct amdgpu_device *bo_adev,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1447
last_domain = dma_addr[i] & SVM_RANGE_VRAM_DOMAIN;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1448
dma_addr[i] &= ~SVM_RANGE_VRAM_DOMAIN;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1454
last_domain == (dma_addr[i + 1] & SVM_RANGE_VRAM_DOMAIN))
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1483
NULL, dma_addr, &vm->last_update);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1486
dma_addr[j] |= last_domain;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
153
svm_is_valid_dma_mapping_addr(struct device *dev, dma_addr_t dma_addr)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1544
prange->dma_addr[gpuidx],
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
155
return dma_addr && !dma_mapping_error(dev, dma_addr) &&
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
156
!(dma_addr & SVM_RANGE_VRAM_DOMAIN);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
165
dma_addr_t *addr = prange->dma_addr[gpuidx];
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
174
prange->dma_addr[gpuidx] = addr;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
236
void svm_range_dma_unmap_dev(struct device *dev, dma_addr_t *dma_addr,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
242
if (!dma_addr)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
246
if (!svm_is_valid_dma_mapping_addr(dev, dma_addr[i]))
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
248
pr_debug_ratelimited("unmap 0x%llx\n", dma_addr[i] >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
249
dma_unmap_page(dev, dma_addr[i], PAGE_SIZE, dir);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
250
dma_addr[i] = 0;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
257
dma_addr_t *dma_addr;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
265
dma_addr = prange->dma_addr[gpuidx];
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
266
if (!dma_addr)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
276
svm_range_dma_unmap_dev(dev, dma_addr, 0, prange->npages);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
301
if (prange->dma_addr[gpuidx]) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
302
kvfree(prange->dma_addr[gpuidx]);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
303
prange->dma_addr[gpuidx] = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
942
if (!src->dma_addr[i])
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
944
dst->dma_addr[i] = svm_range_copy_array(src->dma_addr[i],
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
945
sizeof(*src->dma_addr[i]), src->npages, 0, NULL);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
946
if (!dst->dma_addr[i])
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
993
r = svm_range_split_array(&new->dma_addr[i], &old->dma_addr[i],
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
994
sizeof(*old->dma_addr[i]), old->start,
drivers/gpu/drm/amd/amdkfd/kfd_svm.h
118
dma_addr_t *dma_addr[MAX_GPU_INSTANCE];
drivers/gpu/drm/amd/amdkfd/kfd_svm.h
183
void svm_range_dma_unmap_dev(struct device *dev, dma_addr_t *dma_addr,
drivers/gpu/drm/arm/display/komeda/komeda_framebuffer.c
141
- to_drm_gem_dma_obj(obj)->dma_addr;
drivers/gpu/drm/arm/display/komeda/komeda_framebuffer.c
265
return obj->dma_addr + offset;
drivers/gpu/drm/arm/malidp_mw.c
176
mw_state->addrs[i] = obj->dma_addr + fb->offsets[i];
drivers/gpu/drm/arm/malidp_planes.c
708
dma_addr_t dma_addr;
drivers/gpu/drm/arm/malidp_planes.c
723
dma_addr = drm_fb_dma_get_gem_addr(fb, plane->state,
drivers/gpu/drm/arm/malidp_planes.c
732
dma_addr = obj->dma_addr;
drivers/gpu/drm/arm/malidp_planes.c
735
malidp_hw_write(mp->hwdev, lower_32_bits(dma_addr), ptr);
drivers/gpu/drm/arm/malidp_planes.c
736
malidp_hw_write(mp->hwdev, upper_32_bits(dma_addr), ptr + 4);
drivers/gpu/drm/aspeed/aspeed_gfx_crtc.c
190
writel(gem->dma_addr, priv->base + CRT_ADDR);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
591
state->dscrs[i]->addr = gem->dma_addr + state->offsets[i];
drivers/gpu/drm/drm_fb_dma_helper.c
104
dma_addr += fb->pitches[plane] * block_start_y;
drivers/gpu/drm/drm_fb_dma_helper.c
105
dma_addr += block_size * num_hblocks;
drivers/gpu/drm/drm_fb_dma_helper.c
107
return dma_addr;
drivers/gpu/drm/drm_fb_dma_helper.c
78
dma_addr_t dma_addr;
drivers/gpu/drm/drm_fb_dma_helper.c
92
dma_addr = obj->dma_addr + fb->offsets[plane];
drivers/gpu/drm/drm_gem_dma_helper.c
150
&dma_obj->dma_addr,
drivers/gpu/drm/drm_gem_dma_helper.c
155
&dma_obj->dma_addr,
drivers/gpu/drm/drm_gem_dma_helper.c
240
dma_obj->vaddr, dma_obj->dma_addr,
drivers/gpu/drm/drm_gem_dma_helper.c
244
dma_obj->vaddr, dma_obj->dma_addr);
drivers/gpu/drm/drm_gem_dma_helper.c
409
drm_printf_indent(p, indent, "dma_addr=%pad\n", &dma_obj->dma_addr);
drivers/gpu/drm/drm_gem_dma_helper.c
436
dma_obj->dma_addr, obj->size);
drivers/gpu/drm/drm_gem_dma_helper.c
481
dma_obj->dma_addr = sg_dma_address(sgt->sgl);
drivers/gpu/drm/drm_gem_dma_helper.c
484
drm_dbg_prime(dev, "dma_addr = %pad, size = %zu\n", &dma_obj->dma_addr,
drivers/gpu/drm/drm_gem_dma_helper.c
547
dma_obj->dma_addr,
drivers/gpu/drm/drm_gpusvm.c
1144
struct drm_pagemap_addr *addr = &svm_pages->dma_addr[j];
drivers/gpu/drm/drm_gpusvm.c
1179
if (svm_pages->dma_addr) {
drivers/gpu/drm/drm_gpusvm.c
1180
kvfree(svm_pages->dma_addr);
drivers/gpu/drm/drm_gpusvm.c
1181
svm_pages->dma_addr = NULL;
drivers/gpu/drm/drm_gpusvm.c
1355
if (!svm_pages->dma_addr)
drivers/gpu/drm/drm_gpusvm.c
1470
if (!svm_pages->dma_addr) {
drivers/gpu/drm/drm_gpusvm.c
1473
svm_pages->dma_addr =
drivers/gpu/drm/drm_gpusvm.c
1474
kvmalloc_objs(*svm_pages->dma_addr, npages);
drivers/gpu/drm/drm_gpusvm.c
1475
if (!svm_pages->dma_addr) {
drivers/gpu/drm/drm_gpusvm.c
1515
svm_pages->dma_addr[j] =
drivers/gpu/drm/drm_gpusvm.c
1521
svm_pages->dma_addr[j].addr)) {
drivers/gpu/drm/drm_gpusvm.c
1548
svm_pages->dma_addr[j] = drm_pagemap_addr_encode
drivers/gpu/drm/drm_pagemap.c
236
dma_addr_t dma_addr;
drivers/gpu/drm/drm_pagemap.c
266
dma_addr = dma_map_page(dev, page, 0, page_size(page), dir);
drivers/gpu/drm/drm_pagemap.c
267
if (dma_mapping_error(dev, dma_addr))
drivers/gpu/drm/drm_pagemap.c
271
drm_pagemap_addr_encode(dma_addr,
drivers/gpu/drm/etnaviv/etnaviv_sched.c
39
u32 dma_addr, primid = 0;
drivers/gpu/drm/etnaviv/etnaviv_sched.c
54
dma_addr = gpu_read(gpu, VIVS_FE_DMA_ADDRESS);
drivers/gpu/drm/etnaviv/etnaviv_sched.c
55
change = dma_addr - gpu->hangcheck_dma_addr;
drivers/gpu/drm/etnaviv/etnaviv_sched.c
70
gpu->hangcheck_dma_addr = dma_addr;
drivers/gpu/drm/exynos/exynos5433_drm_decon.c
411
dma_addr_t dma_addr = exynos_drm_fb_dma_addr(fb, 0);
drivers/gpu/drm/exynos/exynos5433_drm_decon.c
439
writel(dma_addr, ctx->addr + DECON_VIDW0xADD0B0(win));
drivers/gpu/drm/exynos/exynos5433_drm_decon.c
441
val = dma_addr + pitch * state->src.h;
drivers/gpu/drm/exynos/exynos_drm_fb.c
152
return exynos_gem->dma_addr + fb->offsets[index];
drivers/gpu/drm/exynos/exynos_drm_fimc.c
554
fimc_write(ctx, buf->dma_addr[0], EXYNOS_CIIYSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimc.c
555
fimc_write(ctx, buf->dma_addr[1], EXYNOS_CIICBSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimc.c
556
fimc_write(ctx, buf->dma_addr[2], EXYNOS_CIICRSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimc.c
937
fimc_write(ctx, buf->dma_addr[0], EXYNOS_CIOYSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimc.c
938
fimc_write(ctx, buf->dma_addr[1], EXYNOS_CIOCBSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimc.c
939
fimc_write(ctx, buf->dma_addr[2], EXYNOS_CIOCRSA(0));
drivers/gpu/drm/exynos/exynos_drm_fimd.c
838
dma_addr_t dma_addr;
drivers/gpu/drm/exynos/exynos_drm_fimd.c
852
dma_addr = exynos_drm_fb_dma_addr(fb, 0) + offset;
drivers/gpu/drm/exynos/exynos_drm_fimd.c
853
val = (unsigned long)dma_addr;
drivers/gpu/drm/exynos/exynos_drm_fimd.c
858
val = (unsigned long)(dma_addr + size);
drivers/gpu/drm/exynos/exynos_drm_fimd.c
863
(unsigned long)dma_addr, val, size);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
207
dma_addr_t dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
220
dma_addr_t dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
301
node[nr].dma_addr =
drivers/gpu/drm/exynos/exynos_drm_g2d.c
370
lnode->cmdlist->data[lnode->cmdlist->last] = node->dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
444
return &g2d_userptr->dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
516
g2d_userptr->dma_addr = sgt->sgl[0].dma_address;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
528
return &g2d_userptr->dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
728
addr = &exynos_gem->dma_addr;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
807
writel_relaxed(node->dma_addr, g2d->regs + G2D_DMA_SFR_BASE_ADDR);
drivers/gpu/drm/exynos/exynos_drm_gem.c
288
exynos_gem->dma_addr, exynos_gem->size,
drivers/gpu/drm/exynos/exynos_drm_gem.c
31
if (exynos_gem->dma_addr) {
drivers/gpu/drm/exynos/exynos_drm_gem.c
419
exynos_gem->dma_addr, exynos_gem->size,
drivers/gpu/drm/exynos/exynos_drm_gem.c
457
exynos_gem->dma_addr = sg_dma_address(sgt->sgl);
drivers/gpu/drm/exynos/exynos_drm_gem.c
58
&exynos_gem->dma_addr, GFP_KERNEL,
drivers/gpu/drm/exynos/exynos_drm_gem.c
69
(unsigned long)exynos_gem->dma_addr, exynos_gem->size);
drivers/gpu/drm/exynos/exynos_drm_gem.c
77
if (!exynos_gem->dma_addr) {
drivers/gpu/drm/exynos/exynos_drm_gem.c
83
(unsigned long)exynos_gem->dma_addr, exynos_gem->size);
drivers/gpu/drm/exynos/exynos_drm_gem.c
86
(dma_addr_t)exynos_gem->dma_addr,
drivers/gpu/drm/exynos/exynos_drm_gem.h
44
dma_addr_t dma_addr;
drivers/gpu/drm/exynos/exynos_drm_gsc.c
629
gsc_write(buf->dma_addr[0], GSC_IN_BASE_ADDR_Y(buf_id));
drivers/gpu/drm/exynos/exynos_drm_gsc.c
630
gsc_write(buf->dma_addr[1], GSC_IN_BASE_ADDR_CB(buf_id));
drivers/gpu/drm/exynos/exynos_drm_gsc.c
631
gsc_write(buf->dma_addr[2], GSC_IN_BASE_ADDR_CR(buf_id));
drivers/gpu/drm/exynos/exynos_drm_gsc.c
954
gsc_write(buf->dma_addr[0], GSC_OUT_BASE_ADDR_Y(buf_id));
drivers/gpu/drm/exynos/exynos_drm_gsc.c
955
gsc_write(buf->dma_addr[1], GSC_OUT_BASE_ADDR_CB(buf_id));
drivers/gpu/drm/exynos/exynos_drm_gsc.c
956
gsc_write(buf->dma_addr[2], GSC_OUT_BASE_ADDR_CR(buf_id));
drivers/gpu/drm/exynos/exynos_drm_ipp.c
372
buf->dma_addr[i] = buf->exynos_gem[i]->dma_addr +
drivers/gpu/drm/exynos/exynos_drm_ipp.h
75
dma_addr_t dma_addr[MAX_FB_BUFFER];
drivers/gpu/drm/exynos/exynos_drm_rotator.c
156
rot_write(buf->dma_addr[0], ROT_SRC_BUF_ADDR(0));
drivers/gpu/drm/exynos/exynos_drm_rotator.c
157
rot_write(buf->dma_addr[1], ROT_SRC_BUF_ADDR(1));
drivers/gpu/drm/exynos/exynos_drm_rotator.c
201
rot_write(buf->dma_addr[0], ROT_DST_BUF_ADDR(0));
drivers/gpu/drm/exynos/exynos_drm_rotator.c
202
rot_write(buf->dma_addr[1], ROT_DST_BUF_ADDR(1));
drivers/gpu/drm/exynos/exynos_drm_scaler.c
163
scaler_write(src_buf->dma_addr[i], bases[i]);
drivers/gpu/drm/exynos/exynos_drm_scaler.c
226
scaler_write(dst_buf->dma_addr[i], bases[i]);
drivers/gpu/drm/exynos/exynos_mixer.c
613
dma_addr_t dma_addr;
drivers/gpu/drm/exynos/exynos_mixer.c
652
dma_addr = exynos_drm_fb_dma_addr(fb, 0)
drivers/gpu/drm/exynos/exynos_mixer.c
678
mixer_reg_write(ctx, MXR_GRAPHIC_BASE(win), dma_addr);
drivers/gpu/drm/fsl-dcu/fsl_dcu_drm_plane.c
140
DCU_CTRLDESCLN(index, 3), gem->dma_addr);
drivers/gpu/drm/hisilicon/kirin/kirin_drm_ade.c
556
u32 addr = (u32) obj->dma_addr + y * stride;
drivers/gpu/drm/hisilicon/kirin/kirin_drm_ade.c
559
ch + 1, y, in_h, stride, (u32) obj->dma_addr);
drivers/gpu/drm/i915/gvt/dmabuf.c
101
dma_addr = sg_dma_address(sg);
drivers/gpu/drm/i915/gvt/dmabuf.c
102
if (dma_addr)
drivers/gpu/drm/i915/gvt/dmabuf.c
103
intel_gvt_dma_unmap_guest_page(vgpu, dma_addr);
drivers/gpu/drm/i915/gvt/dmabuf.c
82
dma_addr_t dma_addr =
drivers/gpu/drm/i915/gvt/dmabuf.c
84
if (intel_gvt_dma_pin_guest_page(vgpu, dma_addr)) {
drivers/gpu/drm/i915/gvt/dmabuf.c
92
sg_dma_address(sg) = dma_addr;
drivers/gpu/drm/i915/gvt/dmabuf.c
98
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/gtt.c
1096
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/gtt.c
1110
PAGE_SIZE, &dma_addr);
drivers/gpu/drm/i915/gvt/gtt.c
1119
ops->set_pfn(&sub_se, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1151
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/gtt.c
1165
PAGE_SIZE, &dma_addr);
drivers/gpu/drm/i915/gvt/gtt.c
1169
ops->set_pfn(&entry, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
1182
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/gtt.c
1193
ret = intel_gvt_dma_map_guest_page(vgpu, gfn, PAGE_SIZE, &dma_addr);
drivers/gpu/drm/i915/gvt/gtt.c
1209
I915_GTT_PAGE_SIZE_2M, &dma_addr))
drivers/gpu/drm/i915/gvt/gtt.c
1221
pte_ops->set_pfn(&se, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
2156
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/gtt.c
2226
&dma_addr);
drivers/gpu/drm/i915/gvt/gtt.c
2235
ops->set_pfn(&m, dma_addr >> PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gvt.h
752
int intel_gvt_dma_pin_guest_page(struct intel_vgpu *vgpu, dma_addr_t dma_addr);
drivers/gpu/drm/i915/gvt/gvt.h
754
unsigned long size, dma_addr_t *dma_addr);
drivers/gpu/drm/i915/gvt/gvt.h
756
dma_addr_t dma_addr);
drivers/gpu/drm/i915/gvt/kvmgt.c
102
dma_addr_t dma_addr;
drivers/gpu/drm/i915/gvt/kvmgt.c
1610
unsigned long size, dma_addr_t *dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
1622
ret = gvt_dma_map_page(vgpu, gfn, dma_addr, size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1626
ret = __gvt_cache_add(vgpu, gfn, *dma_addr, size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1631
gvt_dma_unmap_page(vgpu, gfn, entry->dma_addr, entry->size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1634
ret = gvt_dma_map_page(vgpu, gfn, dma_addr, size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1638
ret = __gvt_cache_add(vgpu, gfn, *dma_addr, size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1643
*dma_addr = entry->dma_addr;
drivers/gpu/drm/i915/gvt/kvmgt.c
1650
gvt_dma_unmap_page(vgpu, gfn, *dma_addr, size);
drivers/gpu/drm/i915/gvt/kvmgt.c
1656
int intel_gvt_dma_pin_guest_page(struct intel_vgpu *vgpu, dma_addr_t dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
1665
entry = __gvt_cache_find_dma_addr(vgpu, dma_addr);
drivers/gpu/drm/i915/gvt/kvmgt.c
1679
gvt_dma_unmap_page(entry->vgpu, entry->gfn, entry->dma_addr,
drivers/gpu/drm/i915/gvt/kvmgt.c
1685
dma_addr_t dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
1693
entry = __gvt_cache_find_dma_addr(vgpu, dma_addr);
drivers/gpu/drm/i915/gvt/kvmgt.c
179
dma_addr_t *dma_addr, unsigned long size)
drivers/gpu/drm/i915/gvt/kvmgt.c
190
*dma_addr = dma_map_page(dev, page, 0, size, DMA_BIDIRECTIONAL);
drivers/gpu/drm/i915/gvt/kvmgt.c
191
if (dma_mapping_error(dev, *dma_addr)) {
drivers/gpu/drm/i915/gvt/kvmgt.c
202
dma_addr_t dma_addr, unsigned long size)
drivers/gpu/drm/i915/gvt/kvmgt.c
206
dma_unmap_page(dev, dma_addr, size, DMA_BIDIRECTIONAL);
drivers/gpu/drm/i915/gvt/kvmgt.c
211
dma_addr_t dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
219
if (dma_addr < itr->dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
221
else if (dma_addr > itr->dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
248
dma_addr_t dma_addr, unsigned long size)
drivers/gpu/drm/i915/gvt/kvmgt.c
259
new->dma_addr = dma_addr;
drivers/gpu/drm/i915/gvt/kvmgt.c
284
if (dma_addr < itr->dma_addr)
drivers/gpu/drm/i915/gvt/kvmgt.c
318
gvt_dma_unmap_page(vgpu, dma->gfn, dma->dma_addr, dma->size);
drivers/gpu/drm/i915/gvt/kvmgt.c
635
gvt_dma_unmap_page(vgpu, entry->gfn, entry->dma_addr,
drivers/gpu/drm/i915/i915_initial_plane.c
42
dma_addr_t dma_addr;
drivers/gpu/drm/i915/i915_initial_plane.c
56
dma_addr = intel_ggtt_read_entry(&ggtt->vm, base, &is_present, &is_local);
drivers/gpu/drm/i915/i915_initial_plane.c
69
if (dma_addr < mem->region.start || dma_addr > mem->region.end) {
drivers/gpu/drm/i915/i915_initial_plane.c
72
&dma_addr, mem->region.name, &mem->region.start, &mem->region.end);
drivers/gpu/drm/i915/i915_initial_plane.c
77
&dma_addr);
drivers/gpu/drm/i915/i915_initial_plane.c
79
plane_config->phys_base = dma_addr - mem->region.start;
drivers/gpu/drm/imagination/pvr_free_list.c
237
dma_addr_t dma_addr = sg_page_iter_dma_address(&dma_iter);
drivers/gpu/drm/imagination/pvr_free_list.c
238
u64 dma_pfn = dma_addr >>
drivers/gpu/drm/imagination/pvr_fw_mips.c
44
dma_addr_t dma_addr;
drivers/gpu/drm/imagination/pvr_fw_mips.c
74
&dma_addr));
drivers/gpu/drm/imagination/pvr_fw_mips.c
75
boot_data->stack_phys_addr = dma_addr;
drivers/gpu/drm/imagination/pvr_mmu.c
1489
child_table->backing_page.dma_addr);
drivers/gpu/drm/imagination/pvr_mmu.c
1550
child_table->backing_page.dma_addr);
drivers/gpu/drm/imagination/pvr_mmu.c
1607
dma_addr_t dma_addr, struct pvr_page_flags_raw flags)
drivers/gpu/drm/imagination/pvr_mmu.c
1613
pvr_page_table_l0_entry_raw_set(entry_raw, dma_addr, flags);
drivers/gpu/drm/imagination/pvr_mmu.c
1863
return ctx->page_table_l2.backing_page.dma_addr;
drivers/gpu/drm/imagination/pvr_mmu.c
2248
pvr_page_create(struct pvr_mmu_op_context *op_ctx, dma_addr_t dma_addr,
drivers/gpu/drm/imagination/pvr_mmu.c
2257
pvr_page_table_l0_insert(op_ctx, dma_addr, flags);
drivers/gpu/drm/imagination/pvr_mmu.c
230
dma_addr_t dma_addr;
drivers/gpu/drm/imagination/pvr_mmu.c
2517
dma_addr_t dma_addr = sg_dma_address(sgl) + offset;
drivers/gpu/drm/imagination/pvr_mmu.c
2536
err = pvr_page_create(op_ctx, dma_addr, page_flags);
drivers/gpu/drm/imagination/pvr_mmu.c
2545
dma_addr += PVR_DEVICE_PAGE_SIZE;
drivers/gpu/drm/imagination/pvr_mmu.c
2547
err = pvr_page_create(op_ctx, dma_addr, page_flags);
drivers/gpu/drm/imagination/pvr_mmu.c
266
dma_addr_t dma_addr;
drivers/gpu/drm/imagination/pvr_mmu.c
283
dma_addr = dma_map_page(dev, raw_page, 0, PVR_MMU_BACKING_PAGE_SIZE,
drivers/gpu/drm/imagination/pvr_mmu.c
285
if (dma_mapping_error(dev, dma_addr)) {
drivers/gpu/drm/imagination/pvr_mmu.c
290
page->dma_addr = dma_addr;
drivers/gpu/drm/imagination/pvr_mmu.c
333
dma_unmap_page(dev, page->dma_addr, PVR_MMU_BACKING_PAGE_SIZE,
drivers/gpu/drm/imagination/pvr_mmu.c
371
dma_sync_single_for_device(dev, page->dma_addr,
drivers/gpu/drm/imagination/pvr_mmu.c
736
dma_addr_t dma_addr,
drivers/gpu/drm/imagination/pvr_mmu.c
741
(dma_addr & ~ROGUE_MMUCTRL_PAGE_X_RANGE_CLRMSK) |
drivers/gpu/drm/imagination/pvr_vm_mips.c
182
dma_addr_t dma_addr;
drivers/gpu/drm/imagination/pvr_vm_mips.c
188
&dma_addr);
drivers/gpu/drm/imagination/pvr_vm_mips.c
192
pte = ((dma_addr >> ROGUE_MIPSFW_LOG2_PAGE_SIZE_4K)
drivers/gpu/drm/imx/dcss/dcss-plane.c
228
p1_ba = dma_obj->dma_addr + fb->offsets[0] +
drivers/gpu/drm/imx/dcss/dcss-plane.c
235
p1_ba = dma_obj->dma_addr + fb->offsets[0] +
drivers/gpu/drm/imx/dcss/dcss-plane.c
241
p2_ba = dma_obj->dma_addr + fb->offsets[1] +
drivers/gpu/drm/imx/ipuv3/ipuv3-plane.c
136
return dma_obj->dma_addr + fb->offsets[plane] + fb->pitches[plane] * y +
drivers/gpu/drm/imx/ipuv3/ipuv3-plane.c
155
return dma_obj->dma_addr + fb->offsets[1] + fb->pitches[1] * y +
drivers/gpu/drm/imx/ipuv3/ipuv3-plane.c
174
return dma_obj->dma_addr + fb->offsets[2] + fb->pitches[2] * y +
drivers/gpu/drm/mediatek/mtk_gem.c
123
&dma_obj->dma_addr,
drivers/gpu/drm/mediatek/mtk_gem.c
132
dma_obj->vaddr, &dma_obj->dma_addr,
drivers/gpu/drm/mediatek/mtk_gem.c
205
dma_obj->dma_addr, obj->size);
drivers/gpu/drm/mediatek/mtk_gem.c
227
dma_obj->dma_addr = sg_dma_address(sgt->sgl);
drivers/gpu/drm/mediatek/mtk_gem.c
32
dma_obj->vaddr, dma_obj->dma_addr);
drivers/gpu/drm/mediatek/mtk_gem.c
58
dma_obj->dma_addr, obj->size);
drivers/gpu/drm/mediatek/mtk_plane.c
128
addr = dma_obj->dma_addr;
drivers/gpu/drm/meson/meson_overlay.c
655
priv->viu.vd1_addr2 = gem->dma_addr + fb->offsets[2];
drivers/gpu/drm/meson/meson_overlay.c
667
priv->viu.vd1_addr1 = gem->dma_addr + fb->offsets[1];
drivers/gpu/drm/meson/meson_overlay.c
679
priv->viu.vd1_addr0 = gem->dma_addr + fb->offsets[0];
drivers/gpu/drm/meson/meson_plane.c
371
priv->viu.osd1_addr = gem->dma_addr;
drivers/gpu/drm/mxsfb/mxsfb_kms.c
356
dma_addr_t dma_addr;
drivers/gpu/drm/mxsfb/mxsfb_kms.c
392
dma_addr = drm_fb_dma_get_gem_addr(new_pstate->fb, new_pstate, 0);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
393
if (dma_addr) {
drivers/gpu/drm/mxsfb/mxsfb_kms.c
394
writel(dma_addr, mxsfb->base + mxsfb->devdata->cur_buf);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
395
writel(dma_addr, mxsfb->base + mxsfb->devdata->next_buf);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
545
dma_addr_t dma_addr;
drivers/gpu/drm/mxsfb/mxsfb_kms.c
547
dma_addr = drm_fb_dma_get_gem_addr(new_pstate->fb, new_pstate, 0);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
548
if (dma_addr)
drivers/gpu/drm/mxsfb/mxsfb_kms.c
549
writel(dma_addr, mxsfb->base + mxsfb->devdata->next_buf);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
560
dma_addr_t dma_addr;
drivers/gpu/drm/mxsfb/mxsfb_kms.c
563
dma_addr = drm_fb_dma_get_gem_addr(new_pstate->fb, new_pstate, 0);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
564
if (!dma_addr) {
drivers/gpu/drm/mxsfb/mxsfb_kms.c
575
dma_addr += 64;
drivers/gpu/drm/mxsfb/mxsfb_kms.c
577
writel(dma_addr, mxsfb->base + LCDC_AS_NEXT_BUF);
drivers/gpu/drm/mxsfb/mxsfb_kms.c
584
writel(dma_addr, mxsfb->base + LCDC_AS_BUF);
drivers/gpu/drm/nouveau/include/nvkm/core/falcon.h
35
int (*init)(struct nvkm_falcon *, u64 dma_addr, int xfer_len,
drivers/gpu/drm/nouveau/include/nvkm/core/falcon.h
49
int nvkm_falcon_dma_wr(struct nvkm_falcon *, const u8 *img, u64 dma_addr, u32 dma_base,
drivers/gpu/drm/nouveau/nouveau_dmem.c
165
dma_info->dma_addr = dma_map_page(dev, dpage, 0, page_size(dpage),
drivers/gpu/drm/nouveau/nouveau_dmem.c
168
if (dma_mapping_error(dev, dma_info->dma_addr))
drivers/gpu/drm/nouveau/nouveau_dmem.c
172
NOUVEAU_APER_HOST, dma_info->dma_addr,
drivers/gpu/drm/nouveau/nouveau_dmem.c
175
dma_unmap_page(dev, dma_info->dma_addr, page_size(dpage),
drivers/gpu/drm/nouveau/nouveau_dmem.c
270
dma_unmap_page(drm->dev->dev, dma_info.dma_addr, PAGE_SIZE,
drivers/gpu/drm/nouveau/nouveau_dmem.c
524
dma_unmap_page(chunk->drm->dev->dev, dma_info[i].dma_addr,
drivers/gpu/drm/nouveau/nouveau_dmem.c
748
dma_info->dma_addr = dma_map_page(dev, spage, 0, page_size(spage),
drivers/gpu/drm/nouveau/nouveau_dmem.c
751
if (dma_mapping_error(dev, dma_info->dma_addr))
drivers/gpu/drm/nouveau/nouveau_dmem.c
755
dma_info->dma_addr))
drivers/gpu/drm/nouveau/nouveau_dmem.c
758
dma_info->dma_addr = DMA_MAPPING_ERROR;
drivers/gpu/drm/nouveau/nouveau_dmem.c
775
dma_unmap_page(dev, dma_info->dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL);
drivers/gpu/drm/nouveau/nouveau_dmem.c
801
if (!dma_mapping_error(drm->dev->dev, dma_info[nr_dma].dma_addr))
drivers/gpu/drm/nouveau/nouveau_dmem.c
815
dma_unmap_page(drm->dev->dev, dma_info[nr_dma].dma_addr,
drivers/gpu/drm/nouveau/nouveau_dmem.c
92
dma_addr_t dma_addr;
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
56
nvkm_falcon_dma_wr(struct nvkm_falcon *falcon, const u8 *img, u64 dma_addr, u32 dma_base,
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
71
dma_addr += dma_base;
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
75
type, mem_base, len, dma_base, dma_addr - dma_base, dma_start);
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
79
ret = dma->init(falcon, dma_addr, dmalen, mem_type, sec, &cmd);
drivers/gpu/drm/nouveau/nvkm/falcon/ga102.c
48
ga102_flcn_dma_init(struct nvkm_falcon *falcon, u64 dma_addr, int xfer_len,
drivers/gpu/drm/nouveau/nvkm/falcon/ga102.c
57
nvkm_falcon_wr32(falcon, 0x110, dma_addr >> 8);
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
504
dma_addr_t dma_addr = node->dma_addrs[i];
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
505
if (dma_addr)
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
506
dma_unmap_page(dev, dma_addr, PAGE_SIZE,
drivers/gpu/drm/omapdrm/omap_fb.c
274
ret = omap_gem_pin(fb->obj[i], &plane->dma_addr);
drivers/gpu/drm/omapdrm/omap_fb.c
290
plane->dma_addr = 0;
drivers/gpu/drm/omapdrm/omap_fb.c
316
plane->dma_addr = 0;
drivers/gpu/drm/omapdrm/omap_fb.c
438
plane->dma_addr = 0;
drivers/gpu/drm/omapdrm/omap_fb.c
48
dma_addr_t dma_addr;
drivers/gpu/drm/omapdrm/omap_fb.c
97
return plane->dma_addr + offset;
drivers/gpu/drm/omapdrm/omap_fbdev.c
162
dma_addr_t dma_addr;
drivers/gpu/drm/omapdrm/omap_fbdev.c
221
ret = omap_gem_pin(bo, &dma_addr);
drivers/gpu/drm/omapdrm/omap_fbdev.c
240
fbi->fix.smem_start = dma_addr;
drivers/gpu/drm/omapdrm/omap_gem.c
1144
off, &omap_obj->dma_addr,
drivers/gpu/drm/omapdrm/omap_gem.c
116
dma_addr_t dma_addr;
drivers/gpu/drm/omapdrm/omap_gem.c
1219
omap_obj->dma_addr);
drivers/gpu/drm/omapdrm/omap_gem.c
1364
&omap_obj->dma_addr,
drivers/gpu/drm/omapdrm/omap_gem.c
1405
omap_obj->dma_addr = sg_dma_address(sgt->sgl);
drivers/gpu/drm/omapdrm/omap_gem.c
1502
entry->dma_addr = tiler_ssptr(block);
drivers/gpu/drm/omapdrm/omap_gem.c
1506
&entry->dma_addr,
drivers/gpu/drm/omapdrm/omap_gem.c
369
pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff;
drivers/gpu/drm/omapdrm/omap_gem.c
464
pfn = entry->dma_addr >> PAGE_SHIFT;
drivers/gpu/drm/omapdrm/omap_gem.c
67
dma_addr_t dma_addr;
drivers/gpu/drm/omapdrm/omap_gem.c
774
omap_obj->dma_addr = tiler_ssptr(block);
drivers/gpu/drm/omapdrm/omap_gem.c
777
DBG("got dma address: %pad", &omap_obj->dma_addr);
drivers/gpu/drm/omapdrm/omap_gem.c
797
int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr)
drivers/gpu/drm/omapdrm/omap_gem.c
826
if (dma_addr)
drivers/gpu/drm/omapdrm/omap_gem.c
827
*dma_addr = omap_obj->dma_addr;
drivers/gpu/drm/omapdrm/omap_gem.c
869
omap_obj->dma_addr = 0;
drivers/gpu/drm/omapdrm/omap_gem.c
897
int x, int y, dma_addr_t *dma_addr)
drivers/gpu/drm/omapdrm/omap_gem.c
906
*dma_addr = tiler_tsptr(omap_obj->block, orient, x, y);
drivers/gpu/drm/omapdrm/omap_gem.h
72
int omap_gem_pin(struct drm_gem_object *obj, dma_addr_t *dma_addr);
drivers/gpu/drm/omapdrm/omap_gem.h
80
int x, int y, dma_addr_t *dma_addr);
drivers/gpu/drm/radeon/radeon.h
628
dma_addr_t *dma_addr, uint32_t flags);
drivers/gpu/drm/radeon/radeon_gart.c
286
int pages, struct page **pagelist, dma_addr_t *dma_addr,
drivers/gpu/drm/radeon/radeon_gart.c
303
page_base = dma_addr[i];
drivers/gpu/drm/renesas/rcar-du/rcar_du_kms.c
401
dma_obj->dma_addr = 0;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
354
dma[i] = gem->dma_addr + fb->offsets[i];
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
295
gem->dma_addr, gem->base.size);
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
147
cfg.mem[i] = gem->dma_addr + fb->offsets[i];
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
55
sstate->dma[0] = gem->dma_addr + fb->offsets[0]
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
61
sstate->dma[1] = gem->dma_addr + fb->offsets[1]
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
170
&rk_obj->dma_addr, GFP_KERNEL,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
205
dma_free_attrs(drm->dev, obj->size, rk_obj->kvaddr, rk_obj->dma_addr,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
236
return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
41
rk_obj->dma_addr = rk_obj->mm.start;
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
43
ret = iommu_map_sgtable(private->domain, rk_obj->dma_addr, rk_obj->sgt,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
442
rk_obj->dma_addr, obj->size,
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
479
rk_obj->dma_addr = sg_dma_address(sg->sgl);
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
69
iommu_unmap(private->domain, rk_obj->dma_addr, rk_obj->size);
drivers/gpu/drm/rockchip/rockchip_drm_gem.h
17
dma_addr_t dma_addr;
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1016
dma_addr = rk_uv_obj->dma_addr + offset + fb->offsets[1];
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1018
VOP_WIN_SET(vop, win, uv_mst, dma_addr);
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
928
dma_addr_t dma_addr;
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
971
dma_addr = rk_obj->dma_addr + offset + fb->offsets[0];
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
978
dma_addr += (actual_h - 1) * fb->pitches[0];
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
990
VOP_AFBC_SET(vop, hdr_ptr, dma_addr);
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
997
VOP_WIN_SET(vop, win, yrgb_mst, dma_addr);
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
1214
yrgb_mst = rk_obj->dma_addr + offset + fb->offsets[0];
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
1226
uv_mst = rk_obj->dma_addr + offset + fb->offsets[1];
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
1953
fb_addr = rk_obj->dma_addr + fb->offsets[i];
drivers/gpu/drm/sprd/sprd_dpu.c
341
addr = dma_obj->dma_addr + fb->offsets[i];
drivers/gpu/drm/sti/sti_gdp.c
521
dma_addr_t dma_addr;
drivers/gpu/drm/sti/sti_gdp.c
528
base = dma_alloc_wc(gdp->dev, size, &dma_addr, GFP_KERNEL);
drivers/gpu/drm/sti/sti_gdp.c
537
if (dma_addr & 0xF) {
drivers/gpu/drm/sti/sti_gdp.c
542
gdp->node_list[i].top_field_paddr = dma_addr;
drivers/gpu/drm/sti/sti_gdp.c
546
dma_addr += sizeof(struct sti_gdp_node);
drivers/gpu/drm/sti/sti_gdp.c
548
if (dma_addr & 0xF) {
drivers/gpu/drm/sti/sti_gdp.c
553
gdp->node_list[i].btm_field_paddr = dma_addr;
drivers/gpu/drm/sti/sti_gdp.c
556
dma_addr += sizeof(struct sti_gdp_node);
drivers/gpu/drm/sti/sti_gdp.c
789
(unsigned long) dma_obj->dma_addr);
drivers/gpu/drm/sti/sti_gdp.c
793
top_field->gam_gdp_pml = (u32) dma_obj->dma_addr + fb->offsets[0];
drivers/gpu/drm/sti/sti_gdp.c
838
(unsigned long) dma_obj->dma_addr,
drivers/gpu/drm/sti/sti_hqvdp.c
1189
(unsigned long) dma_obj->dma_addr);
drivers/gpu/drm/sti/sti_hqvdp.c
1192
cmd->top.current_luma = (u32) dma_obj->dma_addr + fb->offsets[0];
drivers/gpu/drm/sti/sti_hqvdp.c
1193
cmd->top.current_chroma = (u32) dma_obj->dma_addr + fb->offsets[1];
drivers/gpu/drm/sti/sti_hqvdp.c
855
dma_addr_t dma_addr;
drivers/gpu/drm/sti/sti_hqvdp.c
862
&dma_addr,
drivers/gpu/drm/sti/sti_hqvdp.c
869
hqvdp->hqvdp_cmd_paddr = (u32)dma_addr;
drivers/gpu/drm/sun4i/sun4i_backend.c
335
dma_addr_t dma_addr;
drivers/gpu/drm/sun4i/sun4i_backend.c
344
dma_addr = drm_fb_dma_get_gem_addr(fb, state, 0);
drivers/gpu/drm/sun4i/sun4i_backend.c
345
DRM_DEBUG_DRIVER("Setting buffer address to %pad\n", &dma_addr);
drivers/gpu/drm/sun4i/sun4i_backend.c
348
return sun4i_backend_update_yuv_buffer(backend, fb, dma_addr);
drivers/gpu/drm/sun4i/sun4i_backend.c
351
lo_paddr = dma_addr << 3;
drivers/gpu/drm/sun4i/sun4i_backend.c
358
hi_paddr = dma_addr >> 29;
drivers/gpu/drm/sun4i/sun4i_frontend.c
164
dma_addr_t dma_addr;
drivers/gpu/drm/sun4i/sun4i_frontend.c
225
dma_addr = drm_fb_dma_get_gem_addr(fb, state, 0);
drivers/gpu/drm/sun4i/sun4i_frontend.c
226
DRM_DEBUG_DRIVER("Setting buffer #0 address to %pad\n", &dma_addr);
drivers/gpu/drm/sun4i/sun4i_frontend.c
227
regmap_write(frontend->regs, SUN4I_FRONTEND_BUF_ADDR0_REG, dma_addr);
drivers/gpu/drm/sun4i/sun4i_frontend.c
230
dma_addr = drm_fb_dma_get_gem_addr(fb, state, swap ? 2 : 1);
drivers/gpu/drm/sun4i/sun4i_frontend.c
232
&dma_addr);
drivers/gpu/drm/sun4i/sun4i_frontend.c
234
dma_addr);
drivers/gpu/drm/sun4i/sun4i_frontend.c
238
dma_addr = drm_fb_dma_get_gem_addr(fb, state, swap ? 1 : 2);
drivers/gpu/drm/sun4i/sun4i_frontend.c
240
&dma_addr);
drivers/gpu/drm/sun4i/sun4i_frontend.c
242
dma_addr);
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
128
dma_addr_t dma_addr;
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
137
DRM_DEBUG_DRIVER("Using GEM @ %pad\n", &gem->dma_addr);
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
141
dma_addr = gem->dma_addr + fb->offsets[0];
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
144
dma_addr += (state->src.x1 >> 16) * bpp;
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
145
dma_addr += (state->src.y1 >> 16) * fb->pitches[0];
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
153
DRM_DEBUG_DRIVER("Setting buffer address to %pad\n", &dma_addr);
drivers/gpu/drm/sun4i/sun8i_ui_layer.c
157
lower_32_bits(dma_addr));
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
202
dma_addr_t dma_addr;
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
216
DRM_DEBUG_DRIVER("Using GEM @ %pad\n", &gem->dma_addr);
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
219
dma_addr = gem->dma_addr + fb->offsets[i];
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
230
dma_addr += dx * format->cpp[i];
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
231
dma_addr += dy * fb->pitches[i];
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
242
i + 1, &dma_addr);
drivers/gpu/drm/sun4i/sun8i_vi_layer.c
247
lower_32_bits(dma_addr));
drivers/gpu/drm/tidss/tidss_dispc.c
2170
return gem->dma_addr + fb->offsets[0] + x * fb->format->cpp[0] +
drivers/gpu/drm/tidss/tidss_dispc.c
2187
return gem->dma_addr + fb->offsets[1] +
drivers/gpu/drm/tidss/tidss_dispc.c
2200
dma_addr_t dma_addr = dispc_plane_state_dma_addr(state);
drivers/gpu/drm/tidss/tidss_dispc.c
2207
dispc_vid_write(dispc, hw_plane, DISPC_VID_BA_0, dma_addr & 0xffffffff);
drivers/gpu/drm/tidss/tidss_dispc.c
2208
dispc_vid_write(dispc, hw_plane, DISPC_VID_BA_EXT_0, (u64)dma_addr >> 32);
drivers/gpu/drm/tidss/tidss_dispc.c
2209
dispc_vid_write(dispc, hw_plane, DISPC_VID_BA_1, dma_addr & 0xffffffff);
drivers/gpu/drm/tidss/tidss_dispc.c
2210
dispc_vid_write(dispc, hw_plane, DISPC_VID_BA_EXT_1, (u64)dma_addr >> 32);
drivers/gpu/drm/tilcdc/tilcdc_crtc.c
73
start = gem->dma_addr + fb->offsets[0] +
drivers/gpu/drm/tiny/arcpgu.c
232
arc_pgu_write(arcpgu, ARCPGU_REG_BUF0_ADDR, gem->dma_addr);
drivers/gpu/drm/ttm/ttm_pool.c
251
struct page *p, dma_addr_t *dma_addr)
drivers/gpu/drm/ttm/ttm_pool.c
267
*dma_addr = addr;
drivers/gpu/drm/ttm/ttm_pool.c
273
static void ttm_pool_unmap(struct ttm_pool *pool, dma_addr_t dma_addr,
drivers/gpu/drm/ttm/ttm_pool.c
280
dma_unmap_page(pool->dev, dma_addr, (long)num_pages << PAGE_SHIFT,
drivers/gpu/drm/ttm/ttm_pool.c
466
const dma_addr_t *dma_addr, enum ttm_caching caching)
drivers/gpu/drm/ttm/ttm_pool.c
475
if (dma_addr)
drivers/gpu/drm/ttm/ttm_pool.c
476
ttm_pool_unmap(pool, *dma_addr, nr);
drivers/gpu/drm/ttm/ttm_pool.c
505
if (!alloc->dma_addr)
drivers/gpu/drm/ttm/ttm_pool.c
509
*alloc->dma_addr++ = first_dma;
drivers/gpu/drm/ttm/ttm_pool.c
570
dma_addr_t *dma_addr = alloc->dma_addr ? &restore->first_dma : NULL;
drivers/gpu/drm/ttm/ttm_pool.c
573
dma_addr, restore->page_caching);
drivers/gpu/drm/ttm/ttm_pool.c
631
if (alloc->dma_addr) {
drivers/gpu/drm/ttm/ttm_pool.c
681
dma_addr_t *dma_addr = tt->dma_address ?
drivers/gpu/drm/ttm/ttm_pool.c
684
nr = ttm_pool_unmap_and_free(pool, p, dma_addr, caching);
drivers/gpu/drm/ttm/ttm_pool.c
694
alloc->dma_addr = tt->dma_address;
drivers/gpu/drm/ttm/ttm_pool.c
723
WARN_ON(alloc->dma_addr && !pool->dev);
drivers/gpu/drm/ttm/ttm_pool.c
83
dma_addr_t *dma_addr;
drivers/gpu/drm/ttm/ttm_pool.c
925
dma_addr_t *dma_addr = tt->dma_address ? &restore->first_dma : NULL;
drivers/gpu/drm/ttm/ttm_pool.c
928
dma_addr, restore->page_caching);
drivers/gpu/drm/v3d/v3d_mmu.c
94
dma_addr_t dma_addr = sg_dma_address(sgl);
drivers/gpu/drm/v3d/v3d_mmu.c
95
u32 pfn = dma_addr >> V3D_MMU_PAGE_SHIFT;
drivers/gpu/drm/vc4/vc4_bo.c
307
dma_free_wc(dev->dev, obj->size, bo->base.vaddr, bo->base.dma_addr);
drivers/gpu/drm/vc4/vc4_gem.c
131
bo_state[i].paddr = vc4_bo->base.dma_addr;
drivers/gpu/drm/vc4/vc4_gem.c
812
exec->ct0ca = exec->exec_bo->dma_addr + bin_offset;
drivers/gpu/drm/vc4/vc4_gem.c
817
exec->shader_rec_p = exec->exec_bo->dma_addr + shader_rec_offset;
drivers/gpu/drm/vc4/vc4_gem.c
821
exec->uniforms_p = exec->exec_bo->dma_addr + uniforms_offset;
drivers/gpu/drm/vc4/vc4_irq.c
107
V3D_WRITE(V3D_BPOA, bo->base.dma_addr + bin_bo_slot * vc4->bin_alloc_size);
drivers/gpu/drm/vc4/vc4_plane.c
1588
vc4_dlist_write(vc4_state, bo->dma_addr + fb->offsets[i] + offsets[i]);
drivers/gpu/drm/vc4/vc4_plane.c
1997
dma_addr_t paddr = bo->dma_addr + fb->offsets[i] + offsets[i];
drivers/gpu/drm/vc4/vc4_plane.c
2219
dma_addr_t dma_addr = bo->dma_addr + fb->offsets[0];
drivers/gpu/drm/vc4/vc4_plane.c
2235
value |= VC4_SET_FIELD(upper_32_bits(dma_addr) & 0xff,
drivers/gpu/drm/vc4/vc4_plane.c
2241
value = lower_32_bits(dma_addr);
drivers/gpu/drm/vc4/vc4_plane.c
2247
addr = (u32)dma_addr;
drivers/gpu/drm/vc4/vc4_render_cl.c
106
return bo->dma_addr + surf->offset + VC4_TILE_BUFFER_SIZE *
drivers/gpu/drm/vc4/vc4_render_cl.c
147
rcl_u32(setup, setup->color_read->dma_addr +
drivers/gpu/drm/vc4/vc4_render_cl.c
169
rcl_u32(setup, setup->zs_read->dma_addr +
drivers/gpu/drm/vc4/vc4_render_cl.c
237
(setup->zs_write->dma_addr + args->zs_write.offset) |
drivers/gpu/drm/vc4/vc4_render_cl.c
360
(setup->color_write ? (setup->color_write->dma_addr +
drivers/gpu/drm/vc4/vc4_render_cl.c
379
exec->ct1ca = setup->rcl->dma_addr;
drivers/gpu/drm/vc4/vc4_render_cl.c
380
exec->ct1ea = setup->rcl->dma_addr + setup->next_offset;
drivers/gpu/drm/vc4/vc4_txp.c
338
addr = gem->dma_addr + fb->offsets[0];
drivers/gpu/drm/vc4/vc4_v3d.c
273
if ((bo->base.dma_addr & 0xf0000000) ==
drivers/gpu/drm/vc4/vc4_v3d.c
274
((bo->base.dma_addr + bo->base.base.size - 1) & 0xf0000000)) {
drivers/gpu/drm/vc4/vc4_validate.c
296
*(uint32_t *)(validated + 5) = ib->dma_addr + offset;
drivers/gpu/drm/vc4/vc4_validate.c
402
bin_addr = vc4->bin_bo->base.dma_addr + bin_slot * vc4->bin_alloc_size;
drivers/gpu/drm/vc4/vc4_validate.c
610
*validated_p0 = tex->dma_addr + p0;
drivers/gpu/drm/vc4/vc4_validate.c
738
*validated_p0 = tex->dma_addr + p0;
drivers/gpu/drm/vc4/vc4_validate.c
837
*(uint32_t *)(pkt_v + o) = bo[i]->dma_addr + src_offset;
drivers/gpu/drm/vc4/vc4_validate.c
922
*(uint32_t *)(pkt_v + o) = vbo->dma_addr + offset;
drivers/gpu/drm/xe/xe_pt.c
734
xe_res_first_dma(range->base.pages.dma_addr, 0,
drivers/gpu/drm/xe/xe_pt.c
765
xe_res_first_dma(to_userptr_vma(vma)->userptr.pages.dma_addr, 0,
drivers/gpu/drm/xe/xe_res_cursor.h
173
const struct drm_pagemap_addr *addr = cur->dma_addr;
drivers/gpu/drm/xe/xe_res_cursor.h
192
cur->dma_addr = addr;
drivers/gpu/drm/xe/xe_res_cursor.h
216
cur->dma_addr = NULL;
drivers/gpu/drm/xe/xe_res_cursor.h
232
static inline void xe_res_first_dma(const struct drm_pagemap_addr *dma_addr,
drivers/gpu/drm/xe/xe_res_cursor.h
236
XE_WARN_ON(!dma_addr);
drivers/gpu/drm/xe/xe_res_cursor.h
243
cur->dma_seg_size = PAGE_SIZE << dma_addr->order;
drivers/gpu/drm/xe/xe_res_cursor.h
246
cur->dma_addr = dma_addr;
drivers/gpu/drm/xe/xe_res_cursor.h
278
if (cur->dma_addr) {
drivers/gpu/drm/xe/xe_res_cursor.h
325
if (cur->dma_addr)
drivers/gpu/drm/xe/xe_res_cursor.h
342
if (cur->dma_addr)
drivers/gpu/drm/xe/xe_res_cursor.h
343
return cur->dma_addr->proto == XE_INTERCONNECT_VRAM;
drivers/gpu/drm/xe/xe_res_cursor.h
59
const struct drm_pagemap_addr *dma_addr;
drivers/gpu/drm/xe/xe_res_cursor.h
95
cur->dma_addr = NULL;
drivers/gpu/drm/xe/xe_svm.h
214
const struct drm_pagemap_addr *dma_addr;
drivers/gpu/drm/xlnx/zynqmp_disp.c
1123
dma_addr_t dma_addr;
drivers/gpu/drm/xlnx/zynqmp_disp.c
1125
dma_addr = drm_fb_dma_get_gem_addr(state->fb, state, i);
drivers/gpu/drm/xlnx/zynqmp_disp.c
1130
dma->xt.src_start = dma_addr;
drivers/hid/intel-ish-hid/ishtp/loader.c
147
dma_addr_t dma_addr;
drivers/hid/intel-ish-hid/ishtp/loader.c
152
dma_addr = le64_to_cpu(fragment->fragment_tbl[i].ddr_adrs);
drivers/hid/intel-ish-hid/ishtp/loader.c
153
dma_free_coherent(dev->devc, fragment_size, dma_bufs[i], dma_addr);
drivers/hid/intel-ish-hid/ishtp/loader.c
175
dma_addr_t dma_addr;
drivers/hid/intel-ish-hid/ishtp/loader.c
181
dma_bufs[i] = dma_alloc_coherent(dev->devc, fragment_size, &dma_addr, GFP_KERNEL);
drivers/hid/intel-ish-hid/ishtp/loader.c
185
fragment->fragment_tbl[i].ddr_adrs = cpu_to_le64(dma_addr);
drivers/i2c/busses/i2c-amd-mp2-pci.c
102
writeq((u64)i2c_common->dma_addr,
drivers/i2c/busses/i2c-amd-mp2-plat.c
50
i2c_common->dma_addr = dma_map_single(dev_pci, i2c_common->dma_buf,
drivers/i2c/busses/i2c-amd-mp2-plat.c
54
if (unlikely(dma_mapping_error(dev_pci, i2c_common->dma_addr))) {
drivers/i2c/busses/i2c-amd-mp2-plat.c
71
dma_unmap_single(dev_pci, i2c_common->dma_addr,
drivers/i2c/busses/i2c-amd-mp2.h
162
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-at91-master.c
196
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-at91-master.c
208
dma_addr = dma_map_single(dev->dev, dev->buf, dev->buf_len,
drivers/i2c/busses/i2c-at91-master.c
210
if (dma_mapping_error(dev->dev, dma_addr)) {
drivers/i2c/busses/i2c-at91-master.c
228
sg_dma_address(sg) = dma_addr;
drivers/i2c/busses/i2c-at91-master.c
235
sg_dma_address(sg) = dma_addr + part1_len;
drivers/i2c/busses/i2c-at91-master.c
248
sg_dma_address(&dma->sg[0]) = dma_addr;
drivers/i2c/busses/i2c-at91-master.c
335
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-at91-master.c
346
dma_addr = dma_map_single(dev->dev, dev->buf, buf_len, DMA_FROM_DEVICE);
drivers/i2c/busses/i2c-at91-master.c
347
if (dma_mapping_error(dev->dev, dma_addr)) {
drivers/i2c/busses/i2c-at91-master.c
368
sg_dma_address(&dma->sg[0]) = dma_addr;
drivers/i2c/busses/i2c-imx-lpi2c.c
170
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-imx-lpi2c.c
778
dma_unmap_single(chan->device->dev, dma->dma_addr,
drivers/i2c/busses/i2c-imx-lpi2c.c
870
dma->dma_addr = dma_map_single(chan->device->dev,
drivers/i2c/busses/i2c-imx-lpi2c.c
872
if (dma_mapping_error(chan->device->dev, dma->dma_addr)) {
drivers/i2c/busses/i2c-imx-lpi2c.c
877
desc = dmaengine_prep_slave_single(chan, dma->dma_addr,
drivers/i2c/busses/i2c-ismt.c
409
dma_addr_t dma_addr = 0; /* address of the data buffer */
drivers/i2c/busses/i2c-ismt.c
605
dma_addr = dma_map_single(dev,
drivers/i2c/busses/i2c-ismt.c
610
if (dma_mapping_error(dev, dma_addr)) {
drivers/i2c/busses/i2c-ismt.c
616
dev_dbg(dev, " dma_addr = %pad\n", &dma_addr);
drivers/i2c/busses/i2c-ismt.c
618
desc->dptr_low = lower_32_bits(dma_addr);
drivers/i2c/busses/i2c-ismt.c
619
desc->dptr_high = upper_32_bits(dma_addr);
drivers/i2c/busses/i2c-ismt.c
632
dma_unmap_single(dev, dma_addr, dma_size, dma_direction);
drivers/i2c/busses/i2c-qcom-geni.c
116
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-qcom-geni.c
424
geni_se_rx_dma_unprep(&gi2c->se, gi2c->dma_addr, gi2c->xfer_len);
drivers/i2c/busses/i2c-qcom-geni.c
436
geni_se_tx_dma_unprep(&gi2c->se, gi2c->dma_addr, gi2c->xfer_len);
drivers/i2c/busses/i2c-qcom-geni.c
466
gi2c->dma_addr = rx_dma;
drivers/i2c/busses/i2c-qcom-geni.c
505
gi2c->dma_addr = tx_dma;
drivers/i2c/busses/i2c-qcom-geni.c
576
tx_multi_xfer->dma_addr[wr_idx],
drivers/i2c/busses/i2c-qcom-geni.c
581
kfree(tx_multi_xfer->dma_addr);
drivers/i2c/busses/i2c-qcom-geni.c
720
gi2c_gpi_xfer->dma_addr[msg_idx] = addr;
drivers/i2c/busses/i2c-qcom-geni.c
802
tx_multi_xfer->dma_addr = kzalloc_objs(dma_addr_t, num);
drivers/i2c/busses/i2c-qcom-geni.c
803
if (!tx_multi_xfer->dma_buf || !tx_multi_xfer->dma_addr) {
drivers/i2c/busses/i2c-qcom-geni.c
96
dma_addr_t *dma_addr;
drivers/i2c/busses/i2c-rcar.c
487
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-rcar.c
512
dma_addr = dma_map_single(chan->device->dev, buf, len, dir);
drivers/i2c/busses/i2c-rcar.c
513
if (dma_mapping_error(chan->device->dev, dma_addr)) {
drivers/i2c/busses/i2c-rcar.c
519
sg_dma_address(&priv->sg) = dma_addr;
drivers/i2c/busses/i2c-sh_mobile.c
514
dma_addr_t dma_addr;
drivers/i2c/busses/i2c-sh_mobile.c
529
dma_addr = dma_map_single(chan->device->dev, pd->dma_buf, pd->msg->len, dir);
drivers/i2c/busses/i2c-sh_mobile.c
530
if (dma_mapping_error(chan->device->dev, dma_addr)) {
drivers/i2c/busses/i2c-sh_mobile.c
536
sg_dma_address(&pd->sg) = dma_addr;
drivers/infiniband/core/rw.c
377
u64 dma_addr;
drivers/infiniband/core/rw.c
381
dma_addr = ib_dma_map_bvec(dev, &bv, dir);
drivers/infiniband/core/rw.c
382
if (ib_dma_mapping_error(dev, dma_addr))
drivers/infiniband/core/rw.c
386
ctx->single.sge.addr = dma_addr;
drivers/infiniband/core/rw.c
445
u64 dma_addr;
drivers/infiniband/core/rw.c
447
dma_addr = ib_dma_map_bvec(dev, &bv, dir);
drivers/infiniband/core/rw.c
448
if (ib_dma_mapping_error(dev, dma_addr))
drivers/infiniband/core/rw.c
452
sge->addr = dma_addr;
drivers/infiniband/core/verbs.c
2823
u64 dma_addr = sg_dma_address(sg) + sg_offset;
drivers/infiniband/core/verbs.c
2824
u64 prev_addr = dma_addr;
drivers/infiniband/core/verbs.c
2826
u64 end_dma_addr = dma_addr + dma_len;
drivers/infiniband/core/verbs.c
2827
u64 page_addr = dma_addr & page_mask;
drivers/infiniband/core/verbs.c
2834
if (i && (last_page_off != 0 || page_addr != dma_addr)) {
drivers/infiniband/core/verbs.c
2836
if (last_end_dma_addr != dma_addr)
drivers/infiniband/core/verbs.c
2851
mr->length += prev_addr - dma_addr;
drivers/infiniband/hw/bng_re/bng_fw.c
350
cpu_to_le64(sbuf->dma_addr));
drivers/infiniband/hw/bng_re/bng_fw.h
112
dma_addr_t dma_addr;
drivers/infiniband/hw/bng_re/bng_sp.c
129
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bng_re/bng_sp.c
59
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
563
if (fence->dma_addr) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
564
dma_unmap_single(dev, fence->dma_addr, BNXT_RE_FENCE_BYTES,
drivers/infiniband/hw/bnxt_re/ib_verbs.c
566
fence->dma_addr = 0;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
577
dma_addr_t dma_addr = 0;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
584
dma_addr = dma_map_single(dev, fence->va, BNXT_RE_FENCE_BYTES,
drivers/infiniband/hw/bnxt_re/ib_verbs.c
586
rc = dma_mapping_error(dev, dma_addr);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
590
fence->dma_addr = 0;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
593
fence->dma_addr = dma_addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.h
51
dma_addr_t dma_addr;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1458
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1531
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
735
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
747
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
334
cpu_to_le64(sbuf->dma_addr));
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
167
dma_addr_t dma_addr;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1071
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1111
dma_free_coherent(&rcfw->pdev->dev, sbuf.size, sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
112
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
203
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
727
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
790
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
806
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
816
req.resp_addr = cpu_to_le64(sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
855
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
978
&sbuf.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
984
req.resp_addr = cpu_to_le64(sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
998
dma_free_coherent(&rcfw->pdev->dev, sbuf.size, sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/cxgb4/cq.c
104
dma_unmap_addr_set(cq, mapping, cq->dma_addr);
drivers/infiniband/hw/cxgb4/cq.c
1131
mm->dma_addr = chp->cq.dma_addr;
drivers/infiniband/hw/cxgb4/cq.c
1140
mm2->dma_addr = 0;
drivers/infiniband/hw/cxgb4/cq.c
1147
&chp->cq.dma_addr);
drivers/infiniband/hw/cxgb4/cq.c
152
res->u.cq.iqaddr = cpu_to_be64(cq->dma_addr);
drivers/infiniband/hw/cxgb4/cq.c
99
&cq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
547
dma_addr_t dma_addr;
drivers/infiniband/hw/cxgb4/provider.c
117
mm->dma_addr = 0;
drivers/infiniband/hw/cxgb4/provider.c
141
dma_addr_t dma_addr;
drivers/infiniband/hw/cxgb4/provider.c
157
dma_addr = mm->dma_addr;
drivers/infiniband/hw/cxgb4/provider.c
182
vaddr, dma_addr, size);
drivers/infiniband/hw/cxgb4/qp.c
118
sq->dma_addr = c4iw_ocqp_pool_alloc(rdev, sq->memsize);
drivers/infiniband/hw/cxgb4/qp.c
119
if (!sq->dma_addr)
drivers/infiniband/hw/cxgb4/qp.c
121
sq->phys_addr = rdev->oc_mw_pa + sq->dma_addr -
drivers/infiniband/hw/cxgb4/qp.c
123
sq->queue = (__force union t4_wr *)(rdev->oc_mw_kva + sq->dma_addr -
drivers/infiniband/hw/cxgb4/qp.c
132
&(sq->dma_addr), GFP_KERNEL);
drivers/infiniband/hw/cxgb4/qp.c
136
dma_unmap_addr_set(sq, mapping, sq->dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
2289
sq_key_mm->dma_addr = qhp->wq.sq.dma_addr;
drivers/infiniband/hw/cxgb4/qp.c
2297
rq_key_mm->dma_addr = qhp->wq.rq.dma_addr;
drivers/infiniband/hw/cxgb4/qp.c
2306
sq_db_key_mm->dma_addr = 0;
drivers/infiniband/hw/cxgb4/qp.c
2317
rq_db_key_mm->dma_addr = 0;
drivers/infiniband/hw/cxgb4/qp.c
2329
ma_sync_key_mm->dma_addr = 0;
drivers/infiniband/hw/cxgb4/qp.c
2567
&wq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/cxgb4/qp.c
2571
dma_unmap_addr_set(wq, mapping, wq->dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
258
dma_unmap_addr_set(&wq->sq, mapping, wq->sq.dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
2627
res->u.srq.eqaddr = cpu_to_be64(wq->dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
263
&wq->rq.dma_addr,
drivers/infiniband/hw/cxgb4/qp.c
274
dma_unmap_addr_set(&wq->rq, mapping, wq->rq.dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
2787
srq_key_mm->dma_addr = srq->wq.dma_addr;
drivers/infiniband/hw/cxgb4/qp.c
2794
srq_db_key_mm->dma_addr = 0;
drivers/infiniband/hw/cxgb4/qp.c
346
res->u.sqrq.eqaddr = cpu_to_be64(wq->sq.dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
375
res->u.sqrq.eqaddr = cpu_to_be64(wq->rq.dma_addr);
drivers/infiniband/hw/cxgb4/qp.c
97
c4iw_ocqp_pool_free(rdev, sq->dma_addr, sq->memsize);
drivers/infiniband/hw/cxgb4/t4.h
334
dma_addr_t dma_addr;
drivers/infiniband/hw/cxgb4/t4.h
363
dma_addr_t dma_addr;
drivers/infiniband/hw/cxgb4/t4.h
400
dma_addr_t dma_addr;
drivers/infiniband/hw/cxgb4/t4.h
678
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa.h
102
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com.c
1120
efa_com_set_dma_addr(params->dma_addr, &cmd.ba.mem_addr_high,
drivers/infiniband/hw/efa/efa_com.c
1218
eeq->eqes, eeq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
1236
&params.dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
1246
eeq->dma_addr = params.dma_addr;
drivers/infiniband/hw/efa/efa_com.c
1256
eeq->eqes, params.dma_addr);
drivers/infiniband/hw/efa/efa_com.c
138
dma_alloc_coherent(aq->dmadev, size, &sq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
150
addr_high = upper_32_bits(sq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
151
addr_low = lower_32_bits(sq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
175
dma_alloc_coherent(aq->dmadev, size, &cq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
184
addr_high = upper_32_bits(cq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
185
addr_low = lower_32_bits(cq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
215
aenq->entries = dma_alloc_coherent(edev->dmadev, size, &aenq->dma_addr,
drivers/infiniband/hw/efa/efa_com.c
225
addr_low = lower_32_bits(aenq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
226
addr_high = upper_32_bits(aenq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
684
dma_free_coherent(edev->dmadev, size, sq->entries, sq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
687
dma_free_coherent(edev->dmadev, size, cq->entries, cq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
690
dma_free_coherent(edev->dmadev, size, aenq->entries, aenq->dma_addr);
drivers/infiniband/hw/efa/efa_com.c
793
aq->cq.entries, aq->cq.dma_addr);
drivers/infiniband/hw/efa/efa_com.c
796
aq->sq.entries, aq->sq.dma_addr);
drivers/infiniband/hw/efa/efa_com.h
121
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com.h
130
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com.h
26
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com.h
35
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com.h
90
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_com_cmd.c
179
efa_com_set_dma_addr(params->dma_addr,
drivers/infiniband/hw/efa/efa_com_cmd.h
73
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_io_defs.h
220
u64 dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1083
cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr);
drivers/infiniband/hw/efa/efa_verbs.c
1095
efa_free_mapped(dev, cq->cpu_addr, cq->dma_addr, cq->size, DMA_FROM_DEVICE);
drivers/infiniband/hw/efa/efa_verbs.c
110
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1229
cq->dma_addr = ib_umem_start_dma_addr(umem);
drivers/infiniband/hw/efa/efa_verbs.c
123
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1232
cq->cpu_addr = efa_zalloc_mapped(dev, &cq->dma_addr, cq->size,
drivers/infiniband/hw/efa/efa_verbs.c
1242
params.dma_addr = cq->dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1291
cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr);
drivers/infiniband/hw/efa/efa_verbs.c
1304
efa_free_mapped(dev, cq->cpu_addr, cq->dma_addr, cq->size,
drivers/infiniband/hw/efa/efa_verbs.c
1370
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1415
dma_addr = dma_map_single(&dev->pdev->dev,
drivers/infiniband/hw/efa/efa_verbs.c
1419
if (dma_mapping_error(&dev->pdev->dev, dma_addr)) {
drivers/infiniband/hw/efa/efa_verbs.c
1425
chunk_list->chunks[i].dma_addr = dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1427
"chunk[%u] mapped at [%pad]\n", i, &dma_addr);
drivers/infiniband/hw/efa/efa_verbs.c
1438
efa_com_set_dma_addr(dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
1447
dma_unmap_single(&dev->pdev->dev, chunk_list->chunks[i].dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
1464
dma_unmap_single(&dev->pdev->dev, chunk_list->chunks[i].dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
1476
dma_addr_t dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1478
dma_addr = dma_map_single(&dev->pdev->dev, pbl->pbl_buf,
drivers/infiniband/hw/efa/efa_verbs.c
1480
if (dma_mapping_error(&dev->pdev->dev, dma_addr)) {
drivers/infiniband/hw/efa/efa_verbs.c
1485
pbl->phys.continuous.dma_addr = dma_addr;
drivers/infiniband/hw/efa/efa_verbs.c
1488
&dma_addr, pbl->pbl_buf_size_in_bytes);
drivers/infiniband/hw/efa/efa_verbs.c
1597
dma_unmap_single(&dev->pdev->dev, pbl->phys.continuous.dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
1641
efa_com_set_dma_addr(pbl->phys.continuous.dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
1648
efa_com_set_dma_addr(pbl->phys.indirect.chunk_list.chunks[0].dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
185
static void *efa_zalloc_mapped(struct efa_dev *dev, dma_addr_t *dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
194
*dma_addr = dma_map_single(&dev->pdev->dev, addr, size, dir);
drivers/infiniband/hw/efa/efa_verbs.c
195
if (dma_mapping_error(&dev->pdev->dev, *dma_addr)) {
drivers/infiniband/hw/efa/efa_verbs.c
205
dma_addr_t dma_addr,
drivers/infiniband/hw/efa/efa_verbs.c
208
dma_unmap_single(&dev->pdev->dev, dma_addr, size, dir);
drivers/infiniband/hw/erdma/erdma_verbs.c
2097
dma_addr_t dma_addr;
drivers/infiniband/hw/erdma/erdma_verbs.c
2103
resp = dma_pool_zalloc(dev->resp_pool, GFP_KERNEL, &dma_addr);
drivers/infiniband/hw/erdma/erdma_verbs.c
2107
req.target_addr = dma_addr;
drivers/infiniband/hw/erdma/erdma_verbs.c
2124
dma_pool_free(dev->resp_pool, resp, dma_addr);
drivers/infiniband/hw/erdma/erdma_verbs.c
879
dma_addr_t *dma_addr)
drivers/infiniband/hw/erdma/erdma_verbs.c
910
*dma_addr = sg_dma_address(page->umem->sgt_append.sgt.sgl) +
drivers/infiniband/hw/hfi1/user_exp_rcv.c
141
dma_unmap_single(&dd->pcidev->dev, node->dma_addr,
drivers/infiniband/hw/hfi1/user_exp_rcv.c
754
node->dma_addr = phys;
drivers/infiniband/hw/hfi1/user_exp_rcv.c
828
node->dma_addr);
drivers/infiniband/hw/hfi1/user_exp_rcv.c
907
node->rcventry, node->npages, node->dma_addr);
drivers/infiniband/hw/hfi1/user_exp_rcv.h
36
dma_addr_t dma_addr;
drivers/infiniband/hw/hns/hns_roce_hem.c
1163
hem_list_link_bt(pre->addr + step, cur->dma_addr);
drivers/infiniband/hw/hns/hns_roce_hem.c
1254
hem_list_link_bt(cpu_base + offset, hem->dma_addr);
drivers/infiniband/hw/hns/hns_roce_hem.c
1286
phy_base = root_hem->dma_addr + total * BA_BYTE_LEN;
drivers/infiniband/hw/hns/hns_roce_hem.c
1339
hem_list->root_ba = root_hem->dma_addr;
drivers/infiniband/hw/hns/hns_roce_hem.c
919
dma_addr_t dma_addr;
drivers/infiniband/hw/hns/hns_roce_hem.c
945
&hem->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/hns/hns_roce_hem.c
967
hem->addr, hem->dma_addr);
drivers/infiniband/hw/hns/hns_roce_hem.c
992
hem->dma_addr = (dma_addr_t)phy_addr;
drivers/infiniband/hw/ionic/ionic_admin.c
485
.dma_addr = cpu_to_le64(q->dma),
drivers/infiniband/hw/ionic/ionic_controlpath.c
1178
.dma_addr = ionic_pgtbl_dma(buf, 0),
drivers/infiniband/hw/ionic/ionic_controlpath.c
1502
wr.wqe.cmd.mod_qp.dma_addr = cpu_to_le64(hdr_dma);
drivers/infiniband/hw/ionic/ionic_controlpath.c
690
wr.wqe.cmd.create_ah.dma_addr = cpu_to_le64(hdr_dma);
drivers/infiniband/hw/ionic/ionic_controlpath.c
812
.dma_addr = ionic_pgtbl_dma(&mr->buf, addr),
drivers/infiniband/hw/ionic/ionic_datapath.c
1084
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_datapath.c
1107
dma_addr = ionic_pgtbl_dma(&mr->buf, mr->ibmr.iova);
drivers/infiniband/hw/ionic/ionic_datapath.c
1108
wqe->reg_mr.dma_addr = cpu_to_be64(le64_to_cpu(dma_addr));
drivers/infiniband/hw/ionic/ionic_fw.h
475
__be64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
663
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
674
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
696
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
710
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
739
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_fw.h
820
__le64 dma_addr;
drivers/infiniband/hw/ionic/ionic_hw_stats.c
93
.dma_addr = cpu_to_le64(dma),
drivers/infiniband/hw/mlx5/odp.c
172
dma_addr_t dma_addr;
drivers/infiniband/hw/mlx5/odp.c
179
dma_addr = hmm_dma_map_pfn(dev->dma_device, &odp->map,
drivers/infiniband/hw/mlx5/odp.c
181
if (ib_dma_mapping_error(dev, dma_addr))
drivers/infiniband/hw/mlx5/odp.c
184
dma_addr |= MLX5_IB_MTT_READ;
drivers/infiniband/hw/mlx5/odp.c
186
dma_addr |= MLX5_IB_MTT_WRITE;
drivers/infiniband/hw/mlx5/odp.c
188
pas[i] = cpu_to_be64(dma_addr);
drivers/infiniband/hw/mthca/mthca_cmd.c
1538
int mthca_MAP_ICM_page(struct mthca_dev *dev, u64 dma_addr, u64 virt)
drivers/infiniband/hw/mthca/mthca_cmd.c
1550
inbox[1] = cpu_to_be64(dma_addr);
drivers/infiniband/hw/mthca/mthca_cmd.c
1559
(unsigned long long) dma_addr, (unsigned long long) virt);
drivers/infiniband/hw/mthca/mthca_cmd.h
278
int mthca_MAP_ICM_page(struct mthca_dev *dev, u64 dma_addr, u64 virt);
drivers/infiniband/sw/rxe/rxe_mr.c
259
static int rxe_set_page(struct ib_mr *ibmr, u64 dma_addr)
drivers/infiniband/sw/rxe/rxe_mr.c
268
u64 addr = dma_addr + i * PAGE_SIZE;
drivers/infiniband/sw/rxe/rxe_mr.c
357
static void rxe_mr_copy_dma(struct rxe_mr *mr, u64 dma_addr, void *addr,
drivers/infiniband/sw/rxe/rxe_mr.c
360
unsigned int page_offset = dma_addr & (PAGE_SIZE - 1);
drivers/infiniband/sw/rxe/rxe_mr.c
366
page = ib_virt_dma_to_page(dma_addr);
drivers/infiniband/sw/rxe/rxe_mr.c
378
dma_addr += bytes;
drivers/infiniband/ulp/iser/iscsi_iser.c
201
u64 dma_addr;
drivers/infiniband/ulp/iser/iscsi_iser.c
206
dma_addr = ib_dma_map_single(device->ib_device, (void *)tx_desc,
drivers/infiniband/ulp/iser/iscsi_iser.c
208
if (ib_dma_mapping_error(device->ib_device, dma_addr))
drivers/infiniband/ulp/iser/iscsi_iser.c
214
tx_desc->dma_addr = dma_addr;
drivers/infiniband/ulp/iser/iscsi_iser.c
215
tx_desc->tx_sg[0].addr = tx_desc->dma_addr;
drivers/infiniband/ulp/iser/iscsi_iser.c
372
ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr,
drivers/infiniband/ulp/iser/iscsi_iser.h
241
u64 dma_addr;
drivers/infiniband/ulp/iser/iscsi_iser.h
269
u64 dma_addr;
drivers/infiniband/ulp/iser/iser_initiator.c
153
tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE);
drivers/infiniband/ulp/iser/iser_initiator.c
227
u64 dma_addr;
drivers/infiniband/ulp/iser/iser_initiator.c
251
dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc,
drivers/infiniband/ulp/iser/iser_initiator.c
253
if (ib_dma_mapping_error(device->ib_device, dma_addr))
drivers/infiniband/ulp/iser/iser_initiator.c
256
rx_desc->dma_addr = dma_addr;
drivers/infiniband/ulp/iser/iser_initiator.c
259
rx_sg->addr = rx_desc->dma_addr;
drivers/infiniband/ulp/iser/iser_initiator.c
269
ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
drivers/infiniband/ulp/iser/iser_initiator.c
293
ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
drivers/infiniband/ulp/iser/iser_initiator.c
647
desc->dma_addr, ISER_RX_PAYLOAD_SIZE,
drivers/infiniband/ulp/iser/iser_initiator.c
665
desc->dma_addr, ISER_RX_PAYLOAD_SIZE,
drivers/infiniband/ulp/iser/iser_initiator.c
704
ib_dma_unmap_single(device->ib_device, desc->dma_addr,
drivers/infiniband/ulp/iser/iser_verbs.c
864
tx_desc->dma_addr, ISER_HEADERS_LEN,
drivers/infiniband/ulp/isert/ib_isert.c
1333
ib_dma_sync_single_for_cpu(ib_dev, rx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
1337
rx_desc->dma_addr, hdr->opcode, hdr->itt, hdr->flags,
drivers/infiniband/ulp/isert/ib_isert.c
1368
ib_dma_sync_single_for_device(ib_dev, rx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
1383
ib_dma_sync_single_for_cpu(ib_dev, isert_conn->login_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
1399
ib_dma_sync_single_for_device(ib_dev, isert_conn->login_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
1503
if (tx_desc->dma_addr != 0) {
drivers/infiniband/ulp/isert/ib_isert.c
1505
ib_dma_unmap_single(ib_dev, tx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
1507
tx_desc->dma_addr = 0;
drivers/infiniband/ulp/isert/ib_isert.c
152
u64 dma_addr;
drivers/infiniband/ulp/isert/ib_isert.c
163
dma_addr = ib_dma_map_single(ib_dev, rx_desc->buf,
drivers/infiniband/ulp/isert/ib_isert.c
165
if (ib_dma_mapping_error(ib_dev, dma_addr))
drivers/infiniband/ulp/isert/ib_isert.c
168
rx_desc->dma_addr = dma_addr;
drivers/infiniband/ulp/isert/ib_isert.c
171
rx_sg->addr = rx_desc->dma_addr + isert_get_hdr_offset(rx_desc);
drivers/infiniband/ulp/isert/ib_isert.c
182
ib_dma_unmap_single(ib_dev, rx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
203
ib_dma_unmap_single(ib_dev, rx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
324
ib_dma_unmap_single(ib_dev, isert_conn->login_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
339
isert_conn->login_desc->dma_addr = ib_dma_map_single(ib_dev,
drivers/infiniband/ulp/isert/ib_isert.c
342
ret = ib_dma_mapping_error(ib_dev, isert_conn->login_desc->dma_addr);
drivers/infiniband/ulp/isert/ib_isert.c
345
isert_conn->login_desc->dma_addr = 0;
drivers/infiniband/ulp/isert/ib_isert.c
370
ib_dma_unmap_single(ib_dev, isert_conn->login_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
784
ib_dma_sync_single_for_device(ib_dev, tx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
827
ib_dma_sync_single_for_cpu(ib_dev, tx_desc->dma_addr,
drivers/infiniband/ulp/isert/ib_isert.c
839
u64 dma_addr;
drivers/infiniband/ulp/isert/ib_isert.c
841
dma_addr = ib_dma_map_single(ib_dev, (void *)tx_desc,
drivers/infiniband/ulp/isert/ib_isert.c
843
if (ib_dma_mapping_error(ib_dev, dma_addr)) {
drivers/infiniband/ulp/isert/ib_isert.c
848
tx_desc->dma_addr = dma_addr;
drivers/infiniband/ulp/isert/ib_isert.c
849
tx_desc->tx_sg[0].addr = tx_desc->dma_addr;
drivers/infiniband/ulp/isert/ib_isert.c
889
sge.addr = isert_conn->login_desc->dma_addr +
drivers/infiniband/ulp/isert/ib_isert.h
128
u64 dma_addr;
drivers/infiniband/ulp/isert/ib_isert.h
90
u64 dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1047
sge[i].addr = req->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1059
req->iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2525
ib_dma_sync_single_for_cpu(clt_path->s.dev->ib_dev, iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2589
tx_iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
472
sge.addr = req->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
484
req->iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
541
ib_dma_sync_single_for_cpu(clt_path->s.dev->ib_dev, iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
567
ib_dma_sync_single_for_device(clt_path->s.dev->ib_dev, iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-pri.h
129
dma_addr_t dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1156
srv_path->dma_addr[id],
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1635
kfree(srv_path->dma_addr);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1822
srv_path->dma_addr = kzalloc_objs(*srv_path->dma_addr, srv->queue_depth);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1823
if (!srv_path->dma_addr)
drivers/infiniband/ulp/rtrs/rtrs-srv.c
1875
kfree(srv_path->dma_addr);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
200
dma_addr_t dma_addr = srv_path->dma_addr[id->msg_id];
drivers/infiniband/ulp/rtrs/rtrs-srv.c
222
plist->addr = dma_addr + offset;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
293
list.addr = srv_mr->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
300
srv_mr->iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
312
ib_dma_sync_single_for_device(srv_path->s.dev->ib_dev, dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
414
list.addr = srv_mr->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
421
srv_mr->iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
659
u64 dma_addr = sg_dma_address(s);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
660
u64 dma_addr_end = dma_addr + dma_len;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
663
srv_path->dma_addr[ind++] = dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
664
dma_addr += max_chunk_size;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
665
} while (dma_addr < dma_addr_end);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
903
tx_iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
949
ib_dma_sync_single_for_cpu(srv_path->s.dev->ib_dev, iu->dma_addr,
drivers/infiniband/ulp/rtrs/rtrs-srv.h
87
dma_addr_t *dma_addr;
drivers/infiniband/ulp/rtrs/rtrs.c
138
list.addr = iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs.c
39
iu->dma_addr = ib_dma_map_single(dma_dev, iu->buf, size, dir);
drivers/infiniband/ulp/rtrs/rtrs.c
40
if (ib_dma_mapping_error(dma_dev, iu->dma_addr)) {
drivers/infiniband/ulp/rtrs/rtrs.c
65
ib_dma_unmap_single(ibdev, iu->dma_addr, iu->size, iu->direction);
drivers/infiniband/ulp/rtrs/rtrs.c
78
list.addr = iu->dma_addr;
drivers/infiniband/ulp/srp/ib_srp.c
1001
dma_addr = ib_dma_map_single(ibdev, req->indirect_desc,
drivers/infiniband/ulp/srp/ib_srp.c
1004
if (ib_dma_mapping_error(ibdev, dma_addr)) {
drivers/infiniband/ulp/srp/ib_srp.c
1009
req->indirect_dma_addr = dma_addr;
drivers/infiniband/ulp/srp/ib_srp.c
1394
static void srp_map_desc(struct srp_map_state *state, dma_addr_t dma_addr,
drivers/infiniband/ulp/srp/ib_srp.c
1401
desc->va = cpu_to_be64(dma_addr);
drivers/infiniband/ulp/srp/ib_srp.c
988
dma_addr_t dma_addr;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-test.c
201
const dma_addr_t dma_addr,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-test.c
209
.cd_table.cdtab_dma = dma_addr,
drivers/iommu/dma-iommu.c
1262
dma_addr_t dma_addr)
drivers/iommu/dma-iommu.c
1302
if (cur_len && !s_iova_off && (dma_addr & seg_mask) &&
drivers/iommu/dma-iommu.c
1313
sg_dma_address(cur) = dma_addr + s_iova_off;
drivers/iommu/dma-iommu.c
1317
dma_addr += s_iova_len;
drivers/iommu/dma-iommu.c
1670
void *cpu_addr, dma_addr_t dma_addr, size_t size,
drivers/iommu/dma-iommu.c
1701
void *cpu_addr, dma_addr_t dma_addr, size_t size,
drivers/iommu/dma-iommu.c
812
static void __iommu_dma_unmap(struct device *dev, dma_addr_t dma_addr,
drivers/iommu/dma-iommu.c
818
size_t iova_off = iova_offset(iovad, dma_addr);
drivers/iommu/dma-iommu.c
822
dma_addr -= iova_off;
drivers/iommu/dma-iommu.c
827
unmapped = iommu_unmap_fast(domain, dma_addr, size, &iotlb_gather);
drivers/iommu/dma-iommu.c
832
iommu_dma_free_iova(domain, dma_addr, size, &iotlb_gather);
drivers/iommu/s390-iommu.c
320
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
322
unsigned int rsx = calc_rsx(dma_addr);
drivers/iommu/s390-iommu.c
350
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
352
unsigned int rfx = calc_rfx(dma_addr);
drivers/iommu/s390-iommu.c
380
return dma_walk_rs_table(rso, dma_addr, gfp);
drivers/iommu/s390-iommu.c
437
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
441
return dma_walk_rf_table(domain->dma_table, dma_addr, gfp);
drivers/iommu/s390-iommu.c
443
return dma_walk_rs_table(domain->dma_table, dma_addr, gfp);
drivers/iommu/s390-iommu.c
452
dma_addr_t dma_addr, gfp_t gfp)
drivers/iommu/s390-iommu.c
457
rto = dma_walk_region_tables(domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
461
rtx = calc_rtx(dma_addr);
drivers/iommu/s390-iommu.c
466
sx = calc_sx(dma_addr);
drivers/iommu/s390-iommu.c
471
px = calc_px(dma_addr);
drivers/iommu/s390-iommu.c
858
phys_addr_t pa, dma_addr_t dma_addr,
drivers/iommu/s390-iommu.c
868
entry = dma_walk_cpu_trans(s390_domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
875
dma_addr += PAGE_SIZE;
drivers/iommu/s390-iommu.c
882
dma_addr -= PAGE_SIZE;
drivers/iommu/s390-iommu.c
883
entry = dma_walk_cpu_trans(s390_domain, dma_addr, gfp);
drivers/iommu/s390-iommu.c
893
dma_addr_t dma_addr, unsigned long nr_pages)
drivers/iommu/s390-iommu.c
900
entry = dma_walk_cpu_trans(s390_domain, dma_addr, GFP_ATOMIC);
drivers/iommu/s390-iommu.c
906
dma_addr += PAGE_SIZE;
drivers/mailbox/bcm-pdc-mailbox.c
516
pdc_build_rxd(struct pdc_state *pdcs, dma_addr_t dma_addr,
drivers/mailbox/bcm-pdc-mailbox.c
526
rxd->addrlow = cpu_to_le32(lower_32_bits(dma_addr));
drivers/mailbox/bcm-pdc-mailbox.c
527
rxd->addrhigh = cpu_to_le32(upper_32_bits(dma_addr));
drivers/mailbox/bcm-pdc-mailbox.c
544
pdc_build_txd(struct pdc_state *pdcs, dma_addr_t dma_addr, u32 buf_len,
drivers/mailbox/bcm-pdc-mailbox.c
554
txd->addrlow = cpu_to_le32(lower_32_bits(dma_addr));
drivers/mailbox/bcm-pdc-mailbox.c
555
txd->addrhigh = cpu_to_le32(upper_32_bits(dma_addr));
drivers/media/common/saa7146/saa7146_core.c
224
dma_addr_t dma_addr = 0;
drivers/media/common/saa7146/saa7146_core.c
226
cpu = dma_alloc_coherent(&pci->dev, PAGE_SIZE, &dma_addr, GFP_KERNEL);
drivers/media/common/saa7146/saa7146_core.c
232
pt->dma = dma_addr;
drivers/media/common/saa7146/saa7146_vbi.c
11
dma_addr_t dma_addr;
drivers/media/common/saa7146/saa7146_vbi.c
126
dma_free_coherent(&dev->pci->dev, 4096, cpu, dma_addr);
drivers/media/common/saa7146/saa7146_vbi.c
131
dma_free_coherent(&dev->pci->dev, 4096, cpu, dma_addr);
drivers/media/common/saa7146/saa7146_vbi.c
25
cpu = dma_alloc_coherent(&dev->pci->dev, 4096, &dma_addr, GFP_KERNEL);
drivers/media/common/saa7146/saa7146_vbi.c
30
saa7146_write(dev, BASE_EVEN3, dma_addr);
drivers/media/common/saa7146/saa7146_vbi.c
31
saa7146_write(dev, BASE_ODD3, dma_addr+vbi_pixel_to_capture);
drivers/media/common/saa7146/saa7146_vbi.c
32
saa7146_write(dev, PROT_ADDR3, dma_addr+4096);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
185
buf->dma_addr, buf->attrs);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
197
&buf->dma_addr,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
222
buf->dma_addr = sg_dma_address(buf->dma_sgt->sgl);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
289
ret = dma_mmap_attrs(buf->dev, vma, buf->cookie, buf->dma_addr,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
303
__func__, (unsigned long)buf->dma_addr, vma->vm_start,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
31
dma_addr_t dma_addr;
drivers/media/common/videobuf2/videobuf2-dma-contig.c
488
ret = dma_get_sgtable_attrs(buf->dev, sgt, buf->cookie, buf->dma_addr,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
557
dma_unmap_resource(buf->dev, buf->dma_addr, buf->size,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
618
buf->dma_addr = dma_map_resource(buf->dev,
drivers/media/common/videobuf2/videobuf2-dma-contig.c
620
if (dma_mapping_error(buf->dev, buf->dma_addr)) {
drivers/media/common/videobuf2/videobuf2-dma-contig.c
660
buf->dma_addr = sg_dma_address(sgt->sgl);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
724
buf->dma_addr = sg_dma_address(sgt->sgl);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
753
buf->dma_addr = 0;
drivers/media/common/videobuf2/videobuf2-dma-contig.c
762
if (WARN_ON(buf->dma_addr))
drivers/media/common/videobuf2/videobuf2-dma-contig.c
77
return &buf->dma_addr;
drivers/media/dvb-frontends/rtl2832_sdr.c
132
dma_addr_t dma_addr[MAX_BULK_BUFS];
drivers/media/dvb-frontends/rtl2832_sdr.c
312
dev->dma_addr[dev->buf_num]);
drivers/media/dvb-frontends/rtl2832_sdr.c
333
&dev->dma_addr[dev->buf_num]);
drivers/media/dvb-frontends/rtl2832_sdr.c
343
(long long)dev->dma_addr[dev->buf_num]);
drivers/media/dvb-frontends/rtl2832_sdr.c
394
dev->urb_list[i]->transfer_dma = dev->dma_addr[i];
drivers/media/pci/cx23885/cx23885-alsa.c
407
substream->runtime->dma_addr = 0;
drivers/media/pci/cx25821/cx25821-alsa.c
549
substream->runtime->dma_addr = 0;
drivers/media/pci/cx88/cx88-alsa.c
497
substream->runtime->dma_addr = 0;
drivers/media/pci/dm1105/dm1105.c
348
dma_addr_t dma_addr;
drivers/media/pci/dm1105/dm1105.c
602
dm_writel(DM1105_STADR, (__force u32)cpu_to_le32(dev->dma_addr));
drivers/media/pci/dm1105/dm1105.c
608
6 * DM1105_DMA_BYTES, &dev->dma_addr,
drivers/media/pci/dm1105/dm1105.c
617
dev->dma_addr);
drivers/media/pci/dt3155/dt3155.c
150
dma_addr_t dma_addr;
drivers/media/pci/dt3155/dt3155.c
153
dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/pci/dt3155/dt3155.c
154
iowrite32(dma_addr, pd->regs + EVEN_DMA_START);
drivers/media/pci/dt3155/dt3155.c
155
iowrite32(dma_addr + pd->width, pd->regs + ODD_DMA_START);
drivers/media/pci/dt3155/dt3155.c
235
dma_addr_t dma_addr;
drivers/media/pci/dt3155/dt3155.c
264
dma_addr = vb2_dma_contig_plane_dma_addr(ivb, 0);
drivers/media/pci/dt3155/dt3155.c
265
iowrite32(dma_addr, ipd->regs + EVEN_DMA_START);
drivers/media/pci/dt3155/dt3155.c
266
iowrite32(dma_addr + ipd->width, ipd->regs + ODD_DMA_START);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
208
ctx->dma_buffer = ipu6_dma_alloc(adev, sizeall, &ctx->dma_addr,
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
219
ctx->config_vied_addr = ctx->dma_addr;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
223
config_host_addr->input_queue = ctx->dma_addr + offset;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
228
config_host_addr->output_queue = ctx->dma_addr + offset;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
234
config_host_addr->specific_addr = ctx->dma_addr + offset;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
246
res.vied_address = ctx->dma_addr + offset;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
255
res.vied_address = ctx->dma_addr + offset;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
319
ctx->dma_buffer, ctx->dma_addr, 0);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
90
dma_addr_t dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
245
set->output_pins[aq->fw_output].addr = ivb->dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
327
msg->dma_addr, sizeof(*buf),
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
362
dma = ivb->dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
39
ivb->dma_addr = sg_dma_address(sg->sgl);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
412
buf, msg->dma_addr, sizeof(*buf),
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
52
ivb->dma_addr = 0;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
772
addr = ivb->dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.h
37
dma_addr_t dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
559
stream_cfg, msg->dma_addr,
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
606
buf, msg->dma_addr,
drivers/media/pci/intel/ipu6/ipu6-isys.c
934
fwmsg->dma_addr, 0);
drivers/media/pci/intel/ipu6/ipu6-isys.c
938
fwmsg->dma_addr, 0);
drivers/media/pci/intel/ipu6/ipu6-isys.c
944
dma_addr_t dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys.c
950
&dma_addr, GFP_KERNEL, 0);
drivers/media/pci/intel/ipu6/ipu6-isys.c
953
addr->dma_addr = dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys.c
970
addr->dma_addr, 0);
drivers/media/pci/intel/ipu6/ipu6-isys.h
175
dma_addr_t dma_addr;
drivers/media/pci/intel/ipu6/ipu6.c
240
dma_addr_t dma_addr;
drivers/media/pci/intel/ipu6/ipu6.c
245
dma_addr = sg_dma_address(isp->isys->fw_sgt.sgl);
drivers/media/pci/intel/ipu6/ipu6.c
247
dma_addr = sg_dma_address(isp->psys->fw_sgt.sgl);
drivers/media/pci/intel/ipu6/ipu6.c
249
pg_offset = server_fw_addr - dma_addr;
drivers/media/pci/pluto2/pluto2.c
110
dma_addr_t dma_addr;
drivers/media/pci/pluto2/pluto2.c
226
pluto_writereg(pluto, REG_PCAR, pluto->dma_addr);
drivers/media/pci/pluto2/pluto2.c
231
pluto->dma_addr = dma_map_single(&pluto->pdev->dev, pluto->dma_buf,
drivers/media/pci/pluto2/pluto2.c
234
return dma_mapping_error(&pluto->pdev->dev, pluto->dma_addr);
drivers/media/pci/pluto2/pluto2.c
239
dma_unmap_single(&pluto->pdev->dev, pluto->dma_addr, TS_DMA_BYTES,
drivers/media/pci/pluto2/pluto2.c
279
dma_sync_single_for_cpu(&pluto->pdev->dev, pluto->dma_addr,
drivers/media/pci/pluto2/pluto2.c
313
dma_sync_single_for_device(&pluto->pdev->dev, pluto->dma_addr,
drivers/media/pci/saa7134/saa7134-alsa.c
730
substream->runtime->dma_addr = 0;
drivers/media/pci/saa7134/saa7134-core.c
220
dma_addr_t dma_addr = 0;
drivers/media/pci/saa7134/saa7134-core.c
222
cpu = dma_alloc_coherent(&pci->dev, SAA7134_PGTABLE_SIZE, &dma_addr,
drivers/media/pci/saa7134/saa7134-core.c
228
pt->dma = dma_addr;
drivers/media/pci/smipcie/smipcie-main.c
356
port->dma_addr[0]);
drivers/media/pci/smipcie/smipcie-main.c
362
port->dma_addr[1]);
drivers/media/pci/smipcie/smipcie-main.c
405
&port->dma_addr[0],
drivers/media/pci/smipcie/smipcie-main.c
418
&port->dma_addr[1],
drivers/media/pci/smipcie/smipcie-main.c
784
mem = port->dma_addr[0];
drivers/media/pci/smipcie/smipcie-main.c
799
mem = port->dma_addr[1];
drivers/media/pci/smipcie/smipcie.h
256
dma_addr_t dma_addr[2];
drivers/media/pci/solo6x10/solo6x10-p2m.c
125
dma_addr_t dma_addr, u32 ext_addr, u32 size,
drivers/media/pci/solo6x10/solo6x10-p2m.c
128
WARN_ON_ONCE(dma_addr & 0x03);
drivers/media/pci/solo6x10/solo6x10-p2m.c
141
desc->dma_addr = dma_addr;
drivers/media/pci/solo6x10/solo6x10-p2m.c
146
dma_addr_t dma_addr, u32 ext_addr, u32 size,
drivers/media/pci/solo6x10/solo6x10-p2m.c
151
solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat,
drivers/media/pci/solo6x10/solo6x10-p2m.c
173
solo_reg_write(solo_dev, SOLO_P2M_TAR_ADR(id), desc->dma_addr);
drivers/media/pci/solo6x10/solo6x10-p2m.c
32
dma_addr_t dma_addr;
drivers/media/pci/solo6x10/solo6x10-p2m.c
40
dma_addr = dma_map_single(&solo_dev->pdev->dev, sys_addr, size,
drivers/media/pci/solo6x10/solo6x10-p2m.c
42
if (dma_mapping_error(&solo_dev->pdev->dev, dma_addr))
drivers/media/pci/solo6x10/solo6x10-p2m.c
45
ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size,
drivers/media/pci/solo6x10/solo6x10-p2m.c
48
dma_unmap_single(&solo_dev->pdev->dev, dma_addr, size,
drivers/media/pci/solo6x10/solo6x10-p2m.c
93
desc[1].dma_addr);
drivers/media/pci/solo6x10/solo6x10.h
114
u32 dma_addr;
drivers/media/pci/solo6x10/solo6x10.h
342
dma_addr_t dma_addr, u32 ext_addr, u32 size,
drivers/media/pci/solo6x10/solo6x10.h
348
dma_addr_t dma_addr, u32 ext_addr, u32 size,
drivers/media/pci/tw5864/tw5864-core.c
169
tw_writel(TW5864_VLC_STREAM_BASE_ADDR, cur_frame->vlc.dma_addr);
drivers/media/pci/tw5864/tw5864-core.c
170
tw_writel(TW5864_MV_STREAM_BASE_ADDR, cur_frame->mv.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1020
dev->h264_buf[dev->h264_buf_w_index].vlc.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1022
dev->h264_buf[dev->h264_buf_w_index].mv.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1087
dev->h264_buf[i].vlc.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1090
dev->h264_buf[i].mv.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1208
dev->h264_buf[i].vlc.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1211
dev->h264_buf[i].mv.dma_addr);
drivers/media/pci/tw5864/tw5864-video.c
1329
dma_sync_single_for_cpu(&dev->pci->dev, frame->vlc.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
1331
dma_sync_single_for_cpu(&dev->pci->dev, frame->mv.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
1334
dma_sync_single_for_device(&dev->pci->dev, frame->vlc.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
1336
dma_sync_single_for_device(&dev->pci->dev, frame->mv.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
959
&frame->vlc.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
968
&frame->mv.dma_addr,
drivers/media/pci/tw5864/tw5864-video.c
974
frame->vlc.addr, frame->vlc.dma_addr);
drivers/media/pci/tw5864/tw5864.h
70
dma_addr_t dma_addr;
drivers/media/pci/tw686x/tw686x-audio.c
181
ac->buf[i].dma = rt->dma_addr + period_size * i;
drivers/media/platform/allegro-dvt/allegro-core.c
1423
msg.dma_addr = to_codec_addr(dev, paddr);
drivers/media/platform/allegro-dvt/allegro-core.c
1518
buffer->dma_addr = to_codec_addr(dev, al_buffer->paddr);
drivers/media/platform/allegro-dvt/allegro-mail.c
292
dst[i++] = buffer->dma_addr;
drivers/media/platform/allegro-dvt/allegro-mail.c
307
dst[i++] = msg->dma_addr;
drivers/media/platform/allegro-dvt/allegro-mail.h
178
u32 dma_addr;
drivers/media/platform/allegro-dvt/allegro-mail.h
193
u32 dma_addr;
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
126
y_dma_addr = cap->buff->dma_addr[C3_ISP_PLANE_Y];
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
127
uv_dma_addr = cap->buff->dma_addr[C3_ISP_PLANE_UV];
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
129
y_dma_addr = cap->dummy_buff.dma_addr;
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
130
uv_dma_addr = cap->dummy_buff.dma_addr;
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
230
&dummy_buff->dma_addr, GFP_KERNEL,
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
241
cap->dummy_buff.vaddr, cap->dummy_buff.dma_addr,
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
613
buf->dma_addr[i] = vb2_dma_contig_plane_dma_addr(vb, i);
drivers/media/platform/amlogic/c3/isp/c3-isp-common.h
100
dma_addr_t dma_addr[C3_ISP_NUM_PLANES];
drivers/media/platform/amlogic/c3/isp/c3-isp-common.h
114
dma_addr_t dma_addr;
drivers/media/platform/amlogic/c3/isp/c3-isp-common.h
142
dma_addr_t dma_addr;
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
202
buf->dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
23
u32 awb_dma_addr = stats->buff->dma_addr;
drivers/media/platform/broadcom/bcm2835-unicam.c
144
dma_addr_t dma_addr;
drivers/media/platform/broadcom/bcm2835-unicam.c
1602
buf->dma_addr = vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0);
drivers/media/platform/broadcom/bcm2835-unicam.c
2302
&node->dummy_buf.dma_addr,
drivers/media/platform/broadcom/bcm2835-unicam.c
2343
node->dummy_buf.dma_addr);
drivers/media/platform/broadcom/bcm2835-unicam.c
2366
node->dummy_buf.dma_addr);
drivers/media/platform/broadcom/bcm2835-unicam.c
649
dma_addr_t endaddr = buf->dma_addr +
drivers/media/platform/broadcom/bcm2835-unicam.c
653
unicam_reg_write(node->dev, UNICAM_IBSA0, buf->dma_addr);
drivers/media/platform/broadcom/bcm2835-unicam.c
656
unicam_reg_write(node->dev, UNICAM_DBSA0, buf->dma_addr);
drivers/media/platform/broadcom/bcm2835-unicam.c
672
return (unsigned int)(cur_addr - frm->dma_addr) / stride;
drivers/media/platform/marvell/mcam-core.c
1246
desc->dma_addr = sg_dma_address(sg);
drivers/media/platform/marvell/mcam-core.c
186
u32 dma_addr;
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
101
addr_ext = FIELD_PREP(MTK_JPEG_ADDR_MASK, upper_32_bits(dma_addr));
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
114
dma_addr_t dma_addr;
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
121
dma_addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0);
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
123
dma_addr_offsetmask = dma_addr & JPEG_ENC_DST_ADDR_OFFSET_MASK;
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
128
writel(dma_addr & ~0xf, base + JPEG_ENC_DST_ADDR0);
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
129
writel((dma_addr + size) & ~0xf, base + JPEG_ENC_STALL_ADDR0);
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
132
addr_ext = FIELD_PREP(MTK_JPEG_ADDR_MASK, upper_32_bits(dma_addr));
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
88
dma_addr_t dma_addr;
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
93
dma_addr = vb2_dma_contig_plane_dma_addr(src_buf, i) +
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
96
writel(lower_32_bits(dma_addr), base + JPEG_ENC_SRC_LUMA_ADDR);
drivers/media/platform/mediatek/jpeg/mtk_jpeg_enc_hw.c
98
writel(lower_32_bits(dma_addr), base + JPEG_ENC_SRC_CHROMA_ADDR);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
107
(unsigned long)mem->dma_addr, mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
109
dma_free_coherent(&plat_dev->dev, mem->size, mem->va, mem->dma_addr);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
111
mem->dma_addr = 0;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
66
mem->va = dma_alloc_attrs(&plat_dev->dev, mem->size, &mem->dma_addr,
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.c
75
(unsigned long)mem->dma_addr, mem->size);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.h
20
dma_addr_t dma_addr;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_util.h
25
dma_addr_t dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
306
pfb->base_y.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
311
pfb->base_c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
320
&pfb->base_y.dma_addr, &pfb->base_c.dma_addr, pfb->base_y.size);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
348
buf.dma_addr = vb2_dma_contig_plane_dma_addr(&src_buf->vb2_buf, 0);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
357
ctx->id, buf.va, &buf.dma_addr, buf.size, src_buf);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
469
src_mem.dma_addr = vb2_dma_contig_plane_dma_addr(&src_buf->vb2_buf, 0);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateful.c
472
src_buf->vb2_buf.index, src_mem.va, &src_mem.dma_addr, src_mem.size);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
287
pfb->base_y.dma_addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
292
pfb->base_c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
298
dst_buf->index, pfb, pfb->base_y.va, &pfb->base_y.dma_addr,
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
299
&pfb->base_c.dma_addr, pfb->base_y.size, ctx->decoded_frame_cnt);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
343
bs_src->dma_addr = vb2_dma_contig_plane_dma_addr(vb2_src, 0);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_stateless.c
353
ctx->id, bs_src->va, &bs_src->dma_addr, bs_src->size, vb2_src);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1000
vsi->cdf_tmp.buf = instance->cdf_temp.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
121
dma_addr_t dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1600
vsi->bs.dma_addr = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1603
vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1605
vsi->trans.dma_addr = lat_buf->ctx->msg_queue.wdma_wptr_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1608
vsi->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1610
vsi->rd_mv.dma_addr = lat_buf->rd_mv_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1619
work_buffer[i].mv_addr.buf = instance->mv[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1621
work_buffer[i].segid_addr.buf = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1623
work_buffer[i].cdf_addr.buf = instance->cdf[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1627
vsi->cdf_tmp.buf = instance->cdf_temp.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1630
vsi->tile.buf = instance->tile.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1634
vsi->cdf_table.buf = instance->cdf_table.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1636
vsi->iq_table.buf = instance->iq_table.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1669
u64 pa = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1762
if (vsi->trans.dma_addr_end - vsi->trans.dma_addr == vsi->ube.size)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1805
vsi->fb.y.dma_addr = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1807
vsi->fb.c.dma_addr = fb->base_y.dma_addr + size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1809
vsi->fb.c.dma_addr = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1829
vref->y.dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1831
vref->c.dma_addr = vref->y.dma_addr + size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1833
vref->c.dma_addr = vb2_dma_contig_plane_dma_addr(vb, 1);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1835
vsi->tile.dma_addr = lat_buf->tile_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
2115
vsi->trans.dma_addr_end += ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
2117
&pfc->vsi.trans.dma_addr, &pfc->vsi.trans.dma_addr_end);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
150
inst->vsi->pred_buf_dma = inst->pred_buf.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
181
inst->vsi->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
233
fb->base_y.va, (u64)fb->base_y.dma_addr);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
342
uint64_t y_fb_dma = fb ? (u64)fb->base_y.dma_addr : 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
343
uint64_t c_fb_dma = fb ? (u64)fb->base_c.dma_addr : 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_if.c
375
inst->vsi->dec.bs_dma = (uint64_t)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
168
inst->vsi_ctx.pred_buf_dma = inst->pred_buf.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
200
inst->vsi_ctx.mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
358
y_fb_dma = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
359
c_fb_dma = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_if.c
364
inst->vsi_ctx.dec.bs_dma = (uint64_t)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1032
y_fb_dma = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1033
c_fb_dma = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1037
inst->vsi_ctx_ext.bs.dma_addr = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1039
inst->vsi_ctx_ext.fb.y.dma_addr = y_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1040
inst->vsi_ctx_ext.fb.c.dma_addr = c_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1065
inst->vsi_ctx_ext.mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1128
y_fb_dma = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1129
c_fb_dma = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1133
inst->vsi_ctx.dec.bs_buf_addr = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
1169
inst->vsi_ctx.mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
138
u64 dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
494
inst->vsi_ext->bs.dma_addr = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
499
inst->vsi_ext->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
502
inst->vsi_ext->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
505
inst->vsi_ext->row_info.dma_addr = 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
508
inst->vsi_ext->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
511
inst->vsi_ext->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
515
inst->vsi_ext->trans.dma_addr = inst->ctx->msg_queue.wdma_wptr_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
535
y_fb_dma = (u64)fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
539
c_fb_dma = (u64)fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
543
inst->vsi_core_ext->fb.y.dma_addr = y_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
545
inst->vsi_core_ext->fb.c.dma_addr = c_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
551
inst->vsi_core_ext->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
554
inst->vsi_core_ext->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
557
inst->vsi_core_ext->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
560
inst->vsi_core_ext->row_info.dma_addr = 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
563
inst->vsi_core_ext->trans.dma_addr = share_info->trans_start;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
568
inst->vsi_core_ext->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
648
y_fb_dma = (u64)fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
652
c_fb_dma = (u64)fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
661
lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
663
lat_buf->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
665
inst->vsi_core->wdma_err_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
666
inst->vsi_core->slice_bc_start_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
667
inst->vsi_core->slice_bc_end_addr = lat_buf->slice_bc_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
673
inst->vsi_core->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
796
inst->vsi_ext->ube.dma_addr, (unsigned long)inst->vsi_ext->ube.size,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
797
inst->vsi_ext->err_map.dma_addr);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
800
inst->vsi_ext->slice_bc.dma_addr,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
802
inst->vsi_ext->trans.dma_addr, inst->vsi_ext->trans.dma_addr_end);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
810
share_info->trans_end = inst->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
838
share_info->trans_end = inst->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
915
inst->vsi->dec.bs_buf_addr = (uint64_t)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
931
inst->vsi->mv_buf_dma[i] = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
933
inst->vsi->wdma_start_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
934
inst->vsi->wdma_end_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
936
inst->vsi->wdma_err_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
937
inst->vsi->slice_bc_start_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
938
inst->vsi->slice_bc_end_addr = lat_buf->slice_bc_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
959
share_info->trans_end = inst->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_h264_req_multi_if.c
986
share_info->trans_end = inst->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
1035
share_info->trans.dma_addr_end = inst->ctx->msg_queue.wdma_addr.dma_addr +
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
274
dma_addr_t dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
740
inst->vsi->bs.dma_addr = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
762
inst->vsi->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
766
inst->vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
769
inst->vsi->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
772
inst->vsi->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
776
inst->vsi->trans.dma_addr = inst->ctx->msg_queue.wdma_wptr_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
779
share_info->trans.dma_addr = inst->vsi->trans.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
813
y_fb_dma = (u64)fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
818
c_fb_dma = (u64)fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
822
inst->vsi_core->fb.y.dma_addr = y_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
824
inst->vsi_core->fb.c.dma_addr = c_fb_dma;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
829
inst->vsi_core->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
832
inst->vsi_core->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
835
inst->vsi_core->slice_bc.dma_addr = lat_buf->slice_bc_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
838
inst->vsi_core->trans.dma_addr = share_info->trans.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
841
inst->vsi_core->wrap.dma_addr = inst->wrap_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_hevc_req_multi_if.c
846
inst->vsi_core->mv_buf_dma[i].dma_addr = mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
302
if (prev_y_dma == (uint64_t)fb->base_y.dma_addr) {
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
374
inst->vsi->dec.working_buf_dma = (uint64_t)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
445
y_fb_dma = fb ? (u64)fb->base_y.dma_addr : 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
446
c_fb_dma = fb ? (u64)fb->base_c.dma_addr : 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_if.c
452
dec->bs_dma = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
157
inst->vsi->dec.seg_id_buf_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
166
inst->vsi->dec.wrap_y_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
175
inst->vsi->dec.wrap_c_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
184
inst->vsi->dec.vld_wrapper_dma = (u64)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
343
y_fb_dma = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
348
c_fb_dma = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
350
inst->vsi->dec.bs_dma = (u64)bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp8_req_if.c
357
bs->size, (u64)bs->dma_addr,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
404
vsi->mv_buf.pa = (unsigned long)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
421
vsi->seg_id_buf.pa = (unsigned long)mem->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
667
inst->vsi->mv_buf.pa = (unsigned long)inst->mv_buf.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_if.c
672
inst->vsi->seg_id_buf.pa = (unsigned long)inst->seg_id_buf.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1001
vsi->counts.dma_addr = instance->counts.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1004
vsi->ube.dma_addr = lat_buf->ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1006
vsi->trans.dma_addr = lat_buf->ctx->msg_queue.wdma_wptr_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1009
vsi->err_map.dma_addr = lat_buf->wdma_err_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1112
pa = bs->dma_addr + offset;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1624
(unsigned long)vsi->trans.dma_addr,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1630
if (vsi->trans.dma_addr_end - vsi->trans.dma_addr ==
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1680
vsi->fb.y.dma_addr = fb->base_y.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1682
vsi->fb.c.dma_addr = fb->base_y.dma_addr + size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1684
vsi->fb.c.dma_addr = fb->base_c.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1715
vsi->ref[i].y.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1718
vsi->ref[i].c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1719
vsi->ref[i].y.dma_addr + size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1721
vsi->ref[i].c.dma_addr =
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1737
vsi->bs.buf.dma_addr = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1739
vsi->bs.frame.dma_addr = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1743
vsi->mv[i].dma_addr = instance->mv[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1747
vsi->seg[i].dma_addr = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1750
vsi->tile.dma_addr = instance->tile.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1752
vsi->prob.dma_addr = instance->prob.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1754
vsi->counts.dma_addr = instance->counts.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
2100
(unsigned long)pfc->vsi.trans.dma_addr,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
2105
ctx->msg_queue.wdma_addr.dma_addr);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
2185
pfc->vsi.trans.dma_addr_end += ctx->msg_queue.wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
296
dma_addr_t dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
984
vsi->bs.buf.dma_addr = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
986
vsi->bs.frame.dma_addr = bs->dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
990
vsi->mv[i].dma_addr = instance->mv[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
994
vsi->seg[i].dma_addr = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
997
vsi->tile.dma_addr = instance->tile.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
999
vsi->prob.dma_addr = instance->prob.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_drv_if.c
77
if ((bs->dma_addr & 63) != 0) {
drivers/media/platform/mediatek/vcodec/decoder/vdec_drv_if.c
84
if (((fb->base_y.dma_addr & 511) != 0) ||
drivers/media/platform/mediatek/vcodec/decoder/vdec_drv_if.c
85
((fb->base_c.dma_addr & 511) != 0)) {
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
311
msg_queue->wdma_rptr_addr = msg_queue->wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec_msg_queue.c
312
msg_queue->wdma_wptr_addr = msg_queue->wdma_addr.dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1016
bs_buf.dma_addr = vb2_dma_contig_plane_dma_addr(&dst_buf->vb2_buf, 0);
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1022
(u64)bs_buf.dma_addr, bs_buf.size);
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1156
frm_buf.fb_addr[i].dma_addr =
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1162
bs_buf.dma_addr = vb2_dma_contig_plane_dma_addr(&dst_buf->vb2_buf, 0);
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1167
(u64)frm_buf.fb_addr[0].dma_addr, frm_buf.fb_addr[0].size,
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1168
(u64)frm_buf.fb_addr[1].dma_addr, frm_buf.fb_addr[1].size,
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc.c
1169
(u64)frm_buf.fb_addr[2].dma_addr, frm_buf.fb_addr[2].size);
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
358
inst->work_bufs[i].dma_addr = 0;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
383
wb_34[i].iova = inst->work_bufs[i].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
385
wb[i].iova = inst->work_bufs[i].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
389
&inst->work_bufs[i].dma_addr,
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
681
tmp_bs_buf.dma_addr = bs_buf->dma_addr + hdr_sz + filler_sz;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
192
wb[i].iova = inst->work_bufs[i].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
196
&inst->work_bufs[i].dma_addr,
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
268
if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) &&
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
269
(frm_buf->fb_addr[1].dma_addr % 16 == 0) &&
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
270
(frm_buf->fb_addr[2].dma_addr % 16 == 0)) {
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
271
out.base.input_addr[0] = frm_buf->fb_addr[0].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
272
out.base.input_addr[1] = frm_buf->fb_addr[1].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
273
out.base.input_addr[2] = frm_buf->fb_addr[2].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
280
out.base.bs_addr = bs_buf->dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
313
if ((frm_buf->fb_addr[0].dma_addr % 16 == 0) &&
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
314
(frm_buf->fb_addr[1].dma_addr % 16 == 0) &&
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
315
(frm_buf->fb_addr[2].dma_addr % 16 == 0)) {
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
316
out.input_addr[0] = frm_buf->fb_addr[0].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
317
out.input_addr[1] = frm_buf->fb_addr[1].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
318
out.input_addr[2] = frm_buf->fb_addr[2].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc_vpu_if.c
325
out.bs_addr = bs_buf->dma_addr;
drivers/media/platform/nuvoton/npcm-video.c
1018
unsigned int dma_addr, void *vaddr)
drivers/media/platform/nuvoton/npcm-video.c
1029
npcm_video_ece_set_enc_dba(video, dma_addr);
drivers/media/platform/nuvoton/npcm-video.c
1056
dma_addr_t dma_addr;
drivers/media/platform/nuvoton/npcm-video.c
1092
dma_addr = vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0);
drivers/media/platform/nuvoton/npcm-video.c
1093
size = npcm_video_hextile(video, index, dma_addr, addr);
drivers/media/platform/nvidia/tegra-vde/h264.c
398
tegra_vde_writel(vde, vde->secure_bo->dma_addr, vde->sxe, 0x7c);
drivers/media/platform/nvidia/tegra-vde/h264.c
740
vde->frames[id].y_addr = tb->dma_addr[0];
drivers/media/platform/nvidia/tegra-vde/h264.c
741
vde->frames[id].cb_addr = tb->dma_addr[1];
drivers/media/platform/nvidia/tegra-vde/h264.c
742
vde->frames[id].cr_addr = tb->dma_addr[2];
drivers/media/platform/nvidia/tegra-vde/h264.c
743
vde->frames[id].aux_addr = tb->aux->dma_addr;
drivers/media/platform/nvidia/tegra-vde/h264.c
932
bitstream->dma_addr[0],
drivers/media/platform/nvidia/tegra-vde/v4l2.c
269
tb->dma_addr[i] = tb->dma_base[i] + offset;
drivers/media/platform/nvidia/tegra-vde/v4l2.c
274
swap(tb->dma_addr[1], tb->dma_addr[2]);
drivers/media/platform/nvidia/tegra-vde/vde.c
106
bo->dma_addr = iova_dma_addr(&vde->iova, bo->iova);
drivers/media/platform/nvidia/tegra-vde/vde.c
108
bo->dma_addr = sg_dma_address(bo->sgt.sgl);
drivers/media/platform/nvidia/tegra-vde/vde.h
160
dma_addr_t dma_addr[VB2_MAX_PLANES];
drivers/media/platform/nvidia/tegra-vde/vde.h
89
dma_addr_t dma_addr;
drivers/media/platform/nxp/imx-jpeg/mxc-jpeg.c
627
u32 dma_addr;
drivers/media/platform/nxp/imx-jpeg/mxc-jpeg.c
638
dma_addr = mxc_jpeg_get_plane_dma_addr(buf, plane_no);
drivers/media/platform/nxp/imx-jpeg/mxc-jpeg.c
642
plane_no, vaddr, dma_addr, payload);
drivers/media/platform/nxp/imx7-media-csi.c
1320
dma_addr_t dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
1327
dma_addr = vb2_dma_contig_plane_dma_addr(&buf->vbuf.vb2_buf, 0);
drivers/media/platform/nxp/imx7-media-csi.c
1366
imx7_csi_update_buf(csi, dma_addr, buf_num);
drivers/media/platform/nxp/imx7-media-csi.c
219
dma_addr_t dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
399
static void imx7_csi_update_buf(struct imx7_csi *csi, dma_addr_t dma_addr,
drivers/media/platform/nxp/imx7-media-csi.c
403
imx7_csi_reg_write(csi, dma_addr, CSI_CSIDMASA_FB2);
drivers/media/platform/nxp/imx7-media-csi.c
405
imx7_csi_reg_write(csi, dma_addr, CSI_CSIDMASA_FB1);
drivers/media/platform/nxp/imx7-media-csi.c
417
dma_addr_t dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
423
dma_addr = vb2_dma_contig_plane_dma_addr(vb2_buf, 0);
drivers/media/platform/nxp/imx7-media-csi.c
426
dma_addr = csi->underrun_buf.dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
429
imx7_csi_update_buf(csi, dma_addr, i);
drivers/media/platform/nxp/imx7-media-csi.c
456
dma_free_coherent(csi->dev, buf->len, buf->virt, buf->dma_addr);
drivers/media/platform/nxp/imx7-media-csi.c
459
buf->dma_addr = 0;
drivers/media/platform/nxp/imx7-media-csi.c
468
buf->virt = dma_alloc_coherent(csi->dev, buf->len, &buf->dma_addr,
drivers/media/platform/nxp/imx7-media-csi.c
738
dma_addr_t dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
753
dma_addr = vb2_dma_contig_plane_dma_addr(&next->vbuf.vb2_buf, 0);
drivers/media/platform/nxp/imx7-media-csi.c
756
dma_addr = csi->underrun_buf.dma_addr;
drivers/media/platform/nxp/imx7-media-csi.c
760
imx7_csi_update_buf(csi, dma_addr, csi->buf_num);
drivers/media/platform/nxp/imx8-isi/imx8-isi-core.h
396
void mxc_isi_channel_set_inbuf(struct mxc_isi_pipe *pipe, dma_addr_t dma_addr);
drivers/media/platform/nxp/imx8-isi/imx8-isi-hw.c
30
void mxc_isi_channel_set_inbuf(struct mxc_isi_pipe *pipe, dma_addr_t dma_addr)
drivers/media/platform/nxp/imx8-isi/imx8-isi-hw.c
32
mxc_isi_write(pipe, CHNL_IN_BUF_ADDR, lower_32_bits(dma_addr));
drivers/media/platform/nxp/imx8-isi/imx8-isi-hw.c
35
upper_32_bits(dma_addr));
drivers/media/platform/qcom/venus/core.h
335
dma_addr_t dma_addr;
drivers/media/platform/qcom/venus/helpers.c
1451
buf->dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/qcom/venus/helpers.c
430
bd->device_addr = buf->dma_addr;
drivers/media/platform/qcom/venus/helpers.c
508
fdata.device_addr = buf->dma_addr;
drivers/media/platform/renesas/vsp1/vsp1_vspx.c
258
&buffer_desc->dma_addr,
drivers/media/platform/renesas/vsp1/vsp1_vspx.c
287
buffer_desc->dma_addr);
drivers/media/platform/rockchip/rga/rga-hw.c
114
static void rga_cmd_set_src_addr(struct rga_ctx *ctx, dma_addr_t dma_addr)
drivers/media/platform/rockchip/rga/rga-hw.c
121
dest[reg >> 2] = dma_addr >> 4;
drivers/media/platform/rockchip/rga/rga-hw.c
127
static void rga_cmd_set_src1_addr(struct rga_ctx *ctx, dma_addr_t dma_addr)
drivers/media/platform/rockchip/rga/rga-hw.c
134
dest[reg >> 2] = dma_addr >> 4;
drivers/media/platform/rockchip/rga/rga-hw.c
140
static void rga_cmd_set_dst_addr(struct rga_ctx *ctx, dma_addr_t dma_addr)
drivers/media/platform/rockchip/rga/rga-hw.c
147
dest[reg >> 2] = dma_addr >> 4;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
717
&dummy_buf->dma_addr,
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
730
cap->buf.dummy.dma_addr, DMA_ATTR_NO_KERNEL_MAPPING);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
758
cap->buf.dummy.dma_addr >> shift);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
761
cap->buf.dummy.dma_addr >> shift);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
776
cap->buf.dummy.dma_addr >> shift);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
778
cap->buf.dummy.dma_addr >> shift);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
780
cap->buf.dummy.dma_addr >> shift);
drivers/media/platform/rockchip/rkisp1/rkisp1-common.h
292
dma_addr_t dma_addr;
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
174
dma_addr_t dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
178
if (video->buffers[i]->dma_addr[0] == dma_addr)
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
209
ivb->dma_addr[i] = vb2_dma_contig_plane_dma_addr(vb, i);
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
211
ivb->dma_addr[i];
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
216
&ivb->dma_addr[i]);
drivers/media/platform/samsung/exynos4-is/fimc-isp.h
103
dma_addr_t dma_addr[FIMC_ISP_MAX_PLANES];
drivers/media/platform/st/sti/delta/delta-mem.c
14
dma_addr_t dma_addr;
drivers/media/platform/st/sti/delta/delta-mem.c
18
addr = dma_alloc_attrs(delta->dev, size, &dma_addr,
drivers/media/platform/st/sti/delta/delta-mem.c
29
buf->paddr = dma_addr;
drivers/media/platform/ti/omap3isp/isph3a_aewb.c
37
isp_reg_writel(aewb->isp, aewb->active_buf->dma_addr,
drivers/media/platform/ti/omap3isp/isph3a_af.c
41
isp_reg_writel(af->isp, af->active_buf->dma_addr, OMAP3_ISP_IOMEM_H3A,
drivers/media/platform/ti/omap3isp/isphist.c
175
dma_addr_t dma_addr = hist->active_buf->dma_addr;
drivers/media/platform/ti/omap3isp/isphist.c
181
if (unlikely(!dma_addr)) {
drivers/media/platform/ti/omap3isp/isphist.c
203
tx = dmaengine_prep_slave_single(hist->dma_ch, dma_addr,
drivers/media/platform/ti/omap3isp/ispstat.c
348
buf->dma_addr);
drivers/media/platform/ti/omap3isp/ispstat.c
350
buf->dma_addr = 0;
drivers/media/platform/ti/omap3isp/ispstat.c
368
buf->virt_addr = dma_alloc_coherent(dev, size, &buf->dma_addr,
drivers/media/platform/ti/omap3isp/ispstat.c
373
ret = dma_get_sgtable(dev, &buf->sgt, buf->virt_addr, buf->dma_addr,
drivers/media/platform/ti/omap3isp/ispstat.c
376
dma_free_coherent(dev, size, buf->virt_addr, buf->dma_addr);
drivers/media/platform/ti/omap3isp/ispstat.c
378
buf->dma_addr = 0;
drivers/media/platform/ti/omap3isp/ispstat.c
449
stat->subdev.name, i, &buf->dma_addr, buf->virt_addr);
drivers/media/platform/ti/omap3isp/ispstat.c
68
dma_sync(stat->isp->dev, buf->dma_addr, 0, MAGIC_SIZE, dir);
drivers/media/platform/ti/omap3isp/ispstat.c
69
dma_sync(stat->isp->dev, buf->dma_addr + (buf_size & PAGE_MASK),
drivers/media/platform/ti/omap3isp/ispstat.h
38
dma_addr_t dma_addr;
drivers/media/platform/ti/vpe/vip.c
2298
if (sc->loaded_coeff_h != port->sc_coeff_h.dma_addr ||
drivers/media/platform/ti/vpe/vip.c
2305
sc->loaded_coeff_h = port->sc_coeff_h.dma_addr;
drivers/media/platform/ti/vpe/vip.c
2310
if (sc->loaded_coeff_v != port->sc_coeff_v.dma_addr ||
drivers/media/platform/ti/vpe/vip.c
2317
sc->loaded_coeff_v = port->sc_coeff_v.dma_addr;
drivers/media/platform/ti/vpe/vip.c
644
dma_addr_t dma_addr = 0;
drivers/media/platform/ti/vpe/vip.c
737
fmt->vpdma_fmt[plane], dma_addr,
drivers/media/platform/ti/vpe/vip.c
837
dma_addr_t dma_addr;
drivers/media/platform/ti/vpe/vip.c
846
dma_addr = vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0);
drivers/media/platform/ti/vpe/vip.c
849
__func__, buf->vb.vb2_buf.index, &dma_addr);
drivers/media/platform/ti/vpe/vip.c
851
dma_addr = 0;
drivers/media/platform/ti/vpe/vip.c
857
dma_addr, stream->write_desc, drop_data, 0);
drivers/media/platform/ti/vpe/vip.c
860
dma_addr += stream->bytesperline * stream->height;
drivers/media/platform/ti/vpe/vip.c
862
dma_addr, stream->write_desc + 1,
drivers/media/platform/ti/vpe/vpdma.c
1113
write_reg(vpdma, VPDMA_LIST_ADDR, (u32) fw_dma_buf.dma_addr);
drivers/media/platform/ti/vpe/vpdma.c
394
buf->dma_addr = dma_map_single(dev, buf->addr, buf->size,
drivers/media/platform/ti/vpe/vpdma.c
396
if (dma_mapping_error(dev, buf->dma_addr)) {
drivers/media/platform/ti/vpe/vpdma.c
416
dma_unmap_single(dev, buf->dma_addr, buf->size,
drivers/media/platform/ti/vpe/vpdma.c
541
write_reg(vpdma, VPDMA_LIST_ADDR, (u32) list->buf.dma_addr);
drivers/media/platform/ti/vpe/vpdma.c
572
dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.c
582
dtd->start_addr = dma_addr;
drivers/media/platform/ti/vpe/vpdma.c
588
write_desc_addr = list->buf.dma_addr + offset;
drivers/media/platform/ti/vpe/vpdma.c
655
WARN_ON(blk->dma_addr & VPDMA_DESC_ALIGN);
drivers/media/platform/ti/vpe/vpdma.c
662
cfd->payload_addr = (u32) blk->dma_addr;
drivers/media/platform/ti/vpe/vpdma.c
684
WARN_ON(adb->dma_addr & VPDMA_DESC_ALIGN);
drivers/media/platform/ti/vpe/vpdma.c
691
cfd->payload_addr = (u32) adb->dma_addr;
drivers/media/platform/ti/vpe/vpdma.c
825
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.c
828
vpdma_rawchan_add_out_dtd(list, width, stride, c_rect, fmt, dma_addr,
drivers/media/platform/ti/vpe/vpdma.c
835
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.c
856
dma_addr += rect.top * stride + (rect.left * depth >> 3);
drivers/media/platform/ti/vpe/vpdma.c
869
dtd->start_addr = (u32) dma_addr;
drivers/media/platform/ti/vpe/vpdma.c
904
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.c
925
dma_addr += rect.top * stride + (rect.left * depth >> 3);
drivers/media/platform/ti/vpe/vpdma.c
940
dtd->start_addr = (u32) dma_addr;
drivers/media/platform/ti/vpe/vpdma.h
20
dma_addr_t dma_addr;
drivers/media/platform/ti/vpe/vpdma.h
226
struct vpdma_desc_list *list, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.h
245
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.h
249
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpdma.h
254
const struct vpdma_data_format *fmt, dma_addr_t dma_addr,
drivers/media/platform/ti/vpe/vpe.c
1029
dma_addr_t dma_addr;
drivers/media/platform/ti/vpe/vpe.c
1036
dma_addr = ctx->mv_buf_dma[mv_buf_selector];
drivers/media/platform/ti/vpe/vpe.c
1052
dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/ti/vpe/vpe.c
1056
dma_addr = vb2_dma_contig_plane_dma_addr(vb, plane);
drivers/media/platform/ti/vpe/vpe.c
1060
if (!dma_addr) {
drivers/media/platform/ti/vpe/vpe.c
1067
dma_addr += offset;
drivers/media/platform/ti/vpe/vpe.c
1081
vpdma_fmt, dma_addr, MAX_OUT_WIDTH_REG1,
drivers/media/platform/ti/vpe/vpe.c
1097
dma_addr_t dma_addr;
drivers/media/platform/ti/vpe/vpe.c
1105
dma_addr = ctx->mv_buf_dma[mv_buf_selector];
drivers/media/platform/ti/vpe/vpe.c
1118
dma_addr = vb2_dma_contig_plane_dma_addr(vb, 0);
drivers/media/platform/ti/vpe/vpe.c
1122
dma_addr = vb2_dma_contig_plane_dma_addr(vb, plane);
drivers/media/platform/ti/vpe/vpe.c
1126
if (!dma_addr) {
drivers/media/platform/ti/vpe/vpe.c
1133
dma_addr += offset;
drivers/media/platform/ti/vpe/vpe.c
1166
dma_addr += pix->width * height * bpp;
drivers/media/platform/ti/vpe/vpe.c
1184
&q_data->c_rect, vpdma_fmt, dma_addr,
drivers/media/platform/ti/vpe/vpe.c
1261
if (ctx->dev->loaded_mmrs != ctx->mmr_adb.dma_addr || ctx->load_mmrs) {
drivers/media/platform/ti/vpe/vpe.c
1267
ctx->dev->loaded_mmrs = ctx->mmr_adb.dma_addr;
drivers/media/platform/ti/vpe/vpe.c
1271
if (sc->loaded_coeff_h != ctx->sc_coeff_h.dma_addr ||
drivers/media/platform/ti/vpe/vpe.c
1277
sc->loaded_coeff_h = ctx->sc_coeff_h.dma_addr;
drivers/media/platform/ti/vpe/vpe.c
1281
if (sc->loaded_coeff_v != ctx->sc_coeff_v.dma_addr ||
drivers/media/platform/ti/vpe/vpe.c
1287
sc->loaded_coeff_v = ctx->sc_coeff_v.dma_addr;
drivers/media/platform/verisilicon/hantro_g1_h264_dec.c
199
dma_addr_t dma_addr = hantro_h264_get_ref_buf(ctx, i);
drivers/media/platform/verisilicon/hantro_g1_h264_dec.c
201
vdpu_write_relaxed(vpu, dma_addr, G1_REG_ADDR_REF(i));
drivers/media/platform/verisilicon/hantro_h264.c
360
dma_addr_t dma_addr = 0;
drivers/media/platform/verisilicon/hantro_h264.c
365
dma_addr = hantro_get_ref(ctx, dpb[dpb_idx].reference_ts);
drivers/media/platform/verisilicon/hantro_h264.c
367
if (!dma_addr) {
drivers/media/platform/verisilicon/hantro_h264.c
377
dma_addr = hantro_get_dec_buf_addr(ctx, buf);
drivers/media/platform/verisilicon/hantro_h264.c
385
return dma_addr | flags;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_h264_dec.c
416
dma_addr_t dma_addr = hantro_h264_get_ref_buf(ctx, i);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_h264_dec.c
418
vdpu_write_relaxed(vpu, dma_addr, VDPU_REG_REFER_BASE(i));
drivers/media/usb/airspy/airspy.c
120
dma_addr_t dma_addr[MAX_BULK_BUFS];
drivers/media/usb/airspy/airspy.c
356
s->dma_addr[s->buf_num]);
drivers/media/usb/airspy/airspy.c
375
&s->dma_addr[s->buf_num]);
drivers/media/usb/airspy/airspy.c
384
(long long)s->dma_addr[s->buf_num]);
drivers/media/usb/airspy/airspy.c
433
s->urb_list[i]->transfer_dma = s->dma_addr[i];
drivers/media/usb/as102/as102_drv.h
65
dma_addr_t dma_addr;
drivers/media/usb/as102/as102_usb_drv.c
239
dev->dma_addr);
drivers/media/usb/as102/as102_usb_drv.c
249
&dev->dma_addr);
drivers/media/usb/as102/as102_usb_drv.c
269
urb->transfer_dma = dev->dma_addr + (i * AS102_USB_BUF_SIZE);
drivers/media/usb/b2c2/flexcop-usb.c
420
fc_usb->iso_buffer, fc_usb->dma_addr);
drivers/media/usb/b2c2/flexcop-usb.c
439
bufsize, GFP_KERNEL, &fc_usb->dma_addr);
drivers/media/usb/b2c2/flexcop-usb.h
26
dma_addr_t dma_addr;
drivers/media/usb/dvb-usb-v2/dvb_usb.h
301
dma_addr_t dma_addr[MAX_NO_URBS_FOR_DATA_STREAM];
drivers/media/usb/dvb-usb-v2/usb_urb.c
243
(long long)stream->dma_addr[stream->buf_num]);
drivers/media/usb/dvb-usb/dvb-usb.h
352
dma_addr_t dma_addr[MAX_NO_URBS_FOR_DATA_STREAM];
drivers/media/usb/dvb-usb/usb-urb.c
102
stream->dma_addr[stream->buf_num]);
drivers/media/usb/dvb-usb/usb-urb.c
122
&stream->dma_addr[stream->buf_num]) ) == NULL) {
drivers/media/usb/dvb-usb/usb-urb.c
129
stream->buf_list[stream->buf_num], (long long)stream->dma_addr[stream->buf_num]);
drivers/media/usb/dvb-usb/usb-urb.c
162
stream->urb_list[i]->transfer_dma = stream->dma_addr[i];
drivers/media/usb/dvb-usb/usb-urb.c
200
urb->transfer_dma = stream->dma_addr[i];
drivers/media/usb/hackrf/hackrf.c
135
dma_addr_t dma_addr[MAX_BULK_BUFS];
drivers/media/usb/hackrf/hackrf.c
612
dev->dma_addr[dev->buf_num]);
drivers/media/usb/hackrf/hackrf.c
631
&dev->dma_addr[dev->buf_num]);
drivers/media/usb/hackrf/hackrf.c
641
(long long)dev->dma_addr[dev->buf_num]);
drivers/media/usb/hackrf/hackrf.c
697
dev->urb_list[i]->transfer_dma = dev->dma_addr[i];
drivers/message/fusion/mptbase.c
1050
mpt_add_sge(void *pAddr, u32 flagslength, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1054
pSge->Address = cpu_to_le32(dma_addr);
drivers/message/fusion/mptbase.c
1067
mpt_add_sge_64bit(void *pAddr, u32 flagslength, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1071
(lower_32_bits(dma_addr));
drivers/message/fusion/mptbase.c
1073
(upper_32_bits(dma_addr));
drivers/message/fusion/mptbase.c
1088
mpt_add_sge_64bit_1078(void *pAddr, u32 flagslength, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1094
(lower_32_bits(dma_addr));
drivers/message/fusion/mptbase.c
1095
tmp = (u32)(upper_32_bits(dma_addr));
drivers/message/fusion/mptbase.c
1100
if ((((u64)dma_addr + MPI_SGE_LENGTH(flagslength)) >> 32) == 9) {
drivers/message/fusion/mptbase.c
1107
(unsigned long long)dma_addr,
drivers/message/fusion/mptbase.c
1126
mpt_add_chain(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1133
pChain->Address = cpu_to_le32(dma_addr);
drivers/message/fusion/mptbase.c
1146
mpt_add_chain_64bit(void *pAddr, u8 next, u16 length, dma_addr_t dma_addr)
drivers/message/fusion/mptbase.c
1149
u32 tmp = dma_addr & 0xFFFFFFFF;
drivers/message/fusion/mptbase.c
1158
tmp = (u32)(upper_32_bits(dma_addr));
drivers/message/fusion/mptbase.h
594
typedef void (*MPT_ADD_SGE)(void *pAddr, u32 flagslength, dma_addr_t dma_addr);
drivers/message/fusion/mptbase.h
596
dma_addr_t dma_addr);
drivers/message/fusion/mptctl.c
1078
dma_addr_t dma_addr;
drivers/message/fusion/mptctl.c
1082
dma_addr = dma_map_single(&ioc->pcidev->dev,
drivers/message/fusion/mptctl.c
1085
sgl->Address = dma_addr;
drivers/message/fusion/mptctl.c
1132
dma_addr_t dma_addr;
drivers/message/fusion/mptctl.c
1139
dma_addr = sglbuf[i].Address;
drivers/message/fusion/mptctl.c
1144
dma_addr);
drivers/message/fusion/mptctl.c
1177
dma_addr_t dma_addr;
drivers/message/fusion/mptctl.c
1181
dma_addr = sg->Address;
drivers/message/fusion/mptctl.c
1184
dma_unmap_single(&ioc->pcidev->dev, dma_addr, len,
drivers/message/fusion/mptctl.c
1187
dma_addr);
drivers/message/fusion/mptctl.c
1197
dma_addr_t dma_addr;
drivers/message/fusion/mptctl.c
1201
dma_addr = sg->Address;
drivers/message/fusion/mptctl.c
1204
dma_unmap_single(&ioc->pcidev->dev, dma_addr, len, dir);
drivers/message/fusion/mptctl.c
1205
dma_free_coherent(&ioc->pcidev->dev, len, kptr, dma_addr);
drivers/misc/fastrpc.c
1041
pages[i].addr = ctx->maps[i]->dma_addr;
drivers/misc/fastrpc.c
1068
pages[i].addr = ctx->buf->dma_addr -
drivers/misc/fastrpc.c
1100
pages[i].addr = ctx->maps[i]->dma_addr;
drivers/misc/fastrpc.c
1182
msg->addr = ctx->buf ? ctx->buf->dma_addr : 0;
drivers/misc/fastrpc.c
1338
err = qcom_scm_assign_mem(fl->cctx->remote_heap->dma_addr,
drivers/misc/fastrpc.c
1345
&fl->cctx->remote_heap->dma_addr,
drivers/misc/fastrpc.c
1366
pages[0].addr = fl->cctx->remote_heap->dma_addr;
drivers/misc/fastrpc.c
1395
err = qcom_scm_assign_mem(fl->cctx->remote_heap->dma_addr,
drivers/misc/fastrpc.c
1400
&fl->cctx->remote_heap->dma_addr, fl->cctx->remote_heap->size, err);
drivers/misc/fastrpc.c
1490
pages[0].addr = imem->dma_addr;
drivers/misc/fastrpc.c
1948
pages.addr = buf->dma_addr;
drivers/misc/fastrpc.c
197
dma_addr_t dma_addr;
drivers/misc/fastrpc.c
1976
err = qcom_scm_assign_mem(buf->dma_addr, (u64)buf->size,
drivers/misc/fastrpc.c
1981
&buf->dma_addr, buf->size, err);
drivers/misc/fastrpc.c
2095
pages.addr = map->dma_addr;
drivers/misc/fastrpc.c
220
dma_addr_t dma_addr;
drivers/misc/fastrpc.c
348
err = qcom_scm_assign_mem(map->dma_addr, map->len,
drivers/misc/fastrpc.c
353
&map->dma_addr, map->len, err);
drivers/misc/fastrpc.c
418
fastrpc_ipa_to_dma_addr(buf->fl->cctx, buf->dma_addr));
drivers/misc/fastrpc.c
437
buf->dma_addr = 0;
drivers/misc/fastrpc.c
442
buf->virt = dma_alloc_coherent(dev, buf->size, &buf->dma_addr,
drivers/misc/fastrpc.c
468
buf->dma_addr += fastrpc_sid_offset(fl->cctx, fl->sctx);
drivers/misc/fastrpc.c
711
fastrpc_ipa_to_dma_addr(buffer->fl->cctx, buffer->dma_addr),
drivers/misc/fastrpc.c
761
fastrpc_ipa_to_dma_addr(buf->fl->cctx, buf->dma_addr), size);
drivers/misc/fastrpc.c
818
map->dma_addr = sg_phys(map->table->sgl);
drivers/misc/fastrpc.c
820
map->dma_addr = fastrpc_compute_dma_addr(fl, sg_dma_address(map->table->sgl));
drivers/misc/fastrpc.c
846
err = qcom_scm_assign_mem(map->dma_addr, (u64)map->len, &src_perms, dst_perms, 2);
drivers/misc/fastrpc.c
850
&map->dma_addr, map->len, err);
drivers/misc/genwqe/card_base.h
168
dma_addr_t dma_addr; /* physical DMA address */
drivers/misc/genwqe/card_dev.c
153
dma_addr_t *dma_addr,
drivers/misc/genwqe/card_dev.c
168
if (dma_addr)
drivers/misc/genwqe/card_dev.c
169
*dma_addr = m->dma_addr +
drivers/misc/genwqe/card_dev.c
213
(unsigned long)dma_map->dma_addr);
drivers/misc/genwqe/card_dev.c
219
dma_map->dma_addr);
drivers/misc/genwqe/card_dev.c
412
dma_map->dma_addr);
drivers/misc/genwqe/card_dev.c
458
&dma_map->dma_addr);
drivers/misc/genwqe/card_dev.c
465
*(dma_addr_t *)dma_map->k_vaddr = dma_map->dma_addr;
drivers/misc/genwqe/card_dev.c
487
dma_map->dma_addr);
drivers/misc/genwqe/card_dev.c
507
dma_addr_t dma_addr;
drivers/misc/genwqe/card_dev.c
540
xbuf = __genwqe_alloc_consistent(cd, FLASH_BLOCK, &dma_addr);
drivers/misc/genwqe/card_dev.c
563
__func__, (unsigned long)dma_addr, crc, tocopy,
drivers/misc/genwqe/card_dev.c
578
*(__be64 *)&req->__asiv[0] = cpu_to_be64(dma_addr);
drivers/misc/genwqe/card_dev.c
590
*(__be64 *)&req->asiv[0] = cpu_to_be64(dma_addr);
drivers/misc/genwqe/card_dev.c
635
__genwqe_free_consistent(cd, FLASH_BLOCK, xbuf, dma_addr);
drivers/misc/genwqe/card_dev.c
643
dma_addr_t dma_addr;
drivers/misc/genwqe/card_dev.c
676
xbuf = __genwqe_alloc_consistent(cd, FLASH_BLOCK, &dma_addr);
drivers/misc/genwqe/card_dev.c
690
__func__, (unsigned long)dma_addr, tocopy,
drivers/misc/genwqe/card_dev.c
704
*(__be64 *)&cmd->__asiv[0] = cpu_to_be64(dma_addr);
drivers/misc/genwqe/card_dev.c
712
*(__be64 *)&cmd->asiv[0] = cpu_to_be64(dma_addr);
drivers/misc/genwqe/card_dev.c
765
__genwqe_free_consistent(cd, FLASH_BLOCK, xbuf, dma_addr);
drivers/mmc/host/mmci_stm32_sdmmc.c
233
u32 dma_addr;
drivers/mmc/host/mmci_stm32_sdmmc.c
236
dma_addr = idma->bounce_dma_addr;
drivers/mmc/host/mmci_stm32_sdmmc.c
238
dma_addr = sg_dma_address(data->sg);
drivers/mmc/host/mmci_stm32_sdmmc.c
240
writel_relaxed(dma_addr,
drivers/mmc/host/mtk-sd.c
2091
dma_addr_t dma_addr;
drivers/mmc/host/mtk-sd.c
2096
dma_addr = dma->gpd_addr + sizeof(struct mt_gpdma_desc);
drivers/mmc/host/mtk-sd.c
2101
gpd->next = lower_32_bits(dma_addr);
drivers/mmc/host/mtk-sd.c
2103
gpd->gpd_info |= (upper_32_bits(dma_addr) & 0xf) << 24;
drivers/mmc/host/mtk-sd.c
2105
dma_addr = dma->bd_addr;
drivers/mmc/host/mtk-sd.c
2108
gpd->gpd_info |= (upper_32_bits(dma_addr) & 0xf) << 28;
drivers/mmc/host/mtk-sd.c
2112
dma_addr = dma->bd_addr + sizeof(*bd) * (i + 1);
drivers/mmc/host/mtk-sd.c
2113
bd[i].next = lower_32_bits(dma_addr);
drivers/mmc/host/mtk-sd.c
2115
bd[i].bd_info |= (upper_32_bits(dma_addr) & 0xf) << 24;
drivers/mmc/host/sunplus-mmc.c
433
dma_addr_t dma_addr;
drivers/mmc/host/sunplus-mmc.c
445
dma_addr = sg_dma_address(sg);
drivers/mmc/host/sunplus-mmc.c
448
writel(dma_addr, host->base + SPMMC_DMA_BASE_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
451
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_1_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
454
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_2_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
457
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_3_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
460
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_4_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
463
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_5_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
466
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_6_ADDR_REG);
drivers/mmc/host/sunplus-mmc.c
469
writel(dma_addr, host->base + SPMMC_SDRAM_SECTOR_7_ADDR_REG);
drivers/mmc/host/uniphier-sd.c
257
dma_addr_t dma_addr;
drivers/mmc/host/uniphier-sd.c
290
dma_addr = sg_dma_address(data->sg);
drivers/mmc/host/uniphier-sd.c
291
writel(lower_32_bits(dma_addr), host->ctl + UNIPHIER_SD_DMA_ADDR_L);
drivers/mmc/host/uniphier-sd.c
292
writel(upper_32_bits(dma_addr), host->ctl + UNIPHIER_SD_DMA_ADDR_H);
drivers/mmc/host/wbsd.c
1375
host->dma_addr = dma_map_single(mmc_dev(host->mmc), host->dma_buffer,
drivers/mmc/host/wbsd.c
1377
if (dma_mapping_error(mmc_dev(host->mmc), host->dma_addr))
drivers/mmc/host/wbsd.c
1383
if ((host->dma_addr & 0xffff) != 0)
drivers/mmc/host/wbsd.c
1388
else if (host->dma_addr >= 0x1000000)
drivers/mmc/host/wbsd.c
1401
dma_unmap_single(mmc_dev(host->mmc), host->dma_addr,
drivers/mmc/host/wbsd.c
1403
host->dma_addr = 0;
drivers/mmc/host/wbsd.c
1423
dma_unmap_single(mmc_dev(host->mmc), host->dma_addr,
drivers/mmc/host/wbsd.c
1432
host->dma_addr = 0;
drivers/mmc/host/wbsd.c
621
set_dma_addr(host->dma, host->dma_addr);
drivers/mmc/host/wbsd.h
158
dma_addr_t dma_addr; /* Physical address for same */
drivers/mtd/nand/onenand/onenand_samsung.c
133
void __iomem *dma_addr;
drivers/mtd/nand/onenand/onenand_samsung.c
520
void __iomem *base = onenand->dma_addr;
drivers/mtd/nand/onenand/onenand_samsung.c
564
void __iomem *base = onenand->dma_addr;
drivers/mtd/nand/onenand/onenand_samsung.c
586
void __iomem *base = onenand->dma_addr;
drivers/mtd/nand/onenand/onenand_samsung.c
635
!onenand->dma_addr || count != mtd->writesize)
drivers/mtd/nand/onenand/onenand_samsung.c
896
onenand->dma_addr = devm_platform_ioremap_resource(pdev, 1);
drivers/mtd/nand/onenand/onenand_samsung.c
897
if (IS_ERR(onenand->dma_addr))
drivers/mtd/nand/onenand/onenand_samsung.c
898
return PTR_ERR(onenand->dma_addr);
drivers/mtd/nand/raw/arasan-nand-controller.c
400
dma_addr_t dma_addr;
drivers/mtd/nand/raw/arasan-nand-controller.c
423
dma_addr = dma_map_single(nfc->dev, (void *)buf, len, DMA_FROM_DEVICE);
drivers/mtd/nand/raw/arasan-nand-controller.c
424
if (dma_mapping_error(nfc->dev, dma_addr)) {
drivers/mtd/nand/raw/arasan-nand-controller.c
429
writel_relaxed(lower_32_bits(dma_addr), nfc->base + DMA_ADDR0_REG);
drivers/mtd/nand/raw/arasan-nand-controller.c
430
writel_relaxed(upper_32_bits(dma_addr), nfc->base + DMA_ADDR1_REG);
drivers/mtd/nand/raw/arasan-nand-controller.c
435
dma_unmap_single(nfc->dev, dma_addr, len, DMA_FROM_DEVICE);
drivers/mtd/nand/raw/arasan-nand-controller.c
517
dma_addr_t dma_addr;
drivers/mtd/nand/raw/arasan-nand-controller.c
547
dma_addr = dma_map_single(nfc->dev, (void *)buf, len, DMA_TO_DEVICE);
drivers/mtd/nand/raw/arasan-nand-controller.c
548
if (dma_mapping_error(nfc->dev, dma_addr)) {
drivers/mtd/nand/raw/arasan-nand-controller.c
553
writel_relaxed(lower_32_bits(dma_addr), nfc->base + DMA_ADDR0_REG);
drivers/mtd/nand/raw/arasan-nand-controller.c
554
writel_relaxed(upper_32_bits(dma_addr), nfc->base + DMA_ADDR1_REG);
drivers/mtd/nand/raw/arasan-nand-controller.c
558
dma_unmap_single(nfc->dev, dma_addr, len, DMA_TO_DEVICE);
drivers/mtd/nand/raw/denali.c
548
dma_addr_t dma_addr, int page, bool write)
drivers/mtd/nand/raw/denali.c
566
denali->host_write(denali, mode, lower_32_bits(dma_addr));
drivers/mtd/nand/raw/denali.c
569
denali->host_write(denali, mode, upper_32_bits(dma_addr));
drivers/mtd/nand/raw/denali.c
573
dma_addr_t dma_addr, int page, bool write)
drivers/mtd/nand/raw/denali.c
587
denali->host_write(denali, mode | ((dma_addr >> 16) << 8), 0x2200);
drivers/mtd/nand/raw/denali.c
590
denali->host_write(denali, mode | ((dma_addr & 0xffff) << 8), 0x2300);
drivers/mtd/nand/raw/denali.c
656
dma_addr_t dma_addr;
drivers/mtd/nand/raw/denali.c
661
dma_addr = dma_map_single(denali->dev, buf, size, dir);
drivers/mtd/nand/raw/denali.c
662
if (dma_mapping_error(denali->dev, dma_addr)) {
drivers/mtd/nand/raw/denali.c
692
denali->setup_dma(denali, dma_addr, page, write);
drivers/mtd/nand/raw/denali.c
702
dma_unmap_single(denali->dev, dma_addr, size, dir);
drivers/mtd/nand/raw/denali.h
385
void (*setup_dma)(struct denali_controller *denali, dma_addr_t dma_addr,
drivers/mtd/nand/raw/fsmc_nand.c
491
dma_addr_t dma_dst, dma_src, dma_addr;
drivers/mtd/nand/raw/fsmc_nand.c
505
dma_addr = dma_map_single(dma_dev->dev, buffer, len, direction);
drivers/mtd/nand/raw/fsmc_nand.c
506
if (dma_mapping_error(dma_dev->dev, dma_addr))
drivers/mtd/nand/raw/fsmc_nand.c
510
dma_src = dma_addr;
drivers/mtd/nand/raw/fsmc_nand.c
514
dma_dst = dma_addr;
drivers/mtd/nand/raw/fsmc_nand.c
550
dma_unmap_single(dma_dev->dev, dma_addr, len, direction);
drivers/mtd/nand/raw/loongson-nand-controller.c
416
dma_addr_t dma_addr;
drivers/mtd/nand/raw/loongson-nand-controller.c
421
dma_addr = dma_map_single(dev, buf, op->orig_len, data_dir);
drivers/mtd/nand/raw/loongson-nand-controller.c
422
if (dma_mapping_error(dev, dma_addr)) {
drivers/mtd/nand/raw/loongson-nand-controller.c
427
dma_buf = dma_alloc_coherent(dev, op->len, &dma_addr, GFP_KERNEL);
drivers/mtd/nand/raw/loongson-nand-controller.c
435
desc = dmaengine_prep_slave_single(chan, dma_addr, op->len, xfer_dir, DMA_PREP_INTERRUPT);
drivers/mtd/nand/raw/loongson-nand-controller.c
465
dma_free_coherent(dev, op->len, dma_buf, dma_addr);
drivers/mtd/nand/raw/loongson-nand-controller.c
467
dma_unmap_single(dev, dma_addr, op->orig_len, data_dir);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
436
dma_addr_t dma_addr;
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
453
dma_addr = dma_map_single(nand->dev, (void *)addr, len, DMA_TO_DEVICE);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
454
ret = dma_mapping_error(nand->dev, dma_addr);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
459
dma_sync_single_for_device(nand->dev, dma_addr, len, DMA_TO_DEVICE);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
462
writel(dma_addr, nand->regs + MA35_NFI_REG_DMASA);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
471
dma_unmap_single(nand->dev, dma_addr, len, DMA_TO_DEVICE);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
481
dma_addr_t dma_addr;
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
493
dma_addr = dma_map_single(nand->dev, (void *)addr, len, DMA_FROM_DEVICE);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
494
ret = dma_mapping_error(nand->dev, dma_addr);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
501
writel(dma_addr, nand->regs + MA35_NFI_REG_DMASA);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
510
dma_unmap_single(nand->dev, dma_addr, len, DMA_FROM_DEVICE);
drivers/mtd/nand/raw/renesas-nand-controller.c
415
dma_addr_t dma_addr;
drivers/mtd/nand/raw/renesas-nand-controller.c
427
dma_addr = dma_map_single(rnandc->dev, rnandc->buf, mtd->writesize,
drivers/mtd/nand/raw/renesas-nand-controller.c
429
if (dma_mapping_error(rnandc->dev, dma_addr))
drivers/mtd/nand/raw/renesas-nand-controller.c
432
writel(dma_addr, rnandc->regs + DMA_ADDR_LOW_REG);
drivers/mtd/nand/raw/renesas-nand-controller.c
440
dma_unmap_single(rnandc->dev, dma_addr, mtd->writesize, DMA_FROM_DEVICE);
drivers/mtd/nand/raw/renesas-nand-controller.c
597
dma_addr_t dma_addr;
drivers/mtd/nand/raw/renesas-nand-controller.c
610
dma_addr = dma_map_single(rnandc->dev, (void *)rnandc->buf, mtd->writesize,
drivers/mtd/nand/raw/renesas-nand-controller.c
612
if (dma_mapping_error(rnandc->dev, dma_addr))
drivers/mtd/nand/raw/renesas-nand-controller.c
615
writel(dma_addr, rnandc->regs + DMA_ADDR_LOW_REG);
drivers/mtd/nand/raw/renesas-nand-controller.c
623
dma_unmap_single(rnandc->dev, dma_addr, mtd->writesize, DMA_TO_DEVICE);
drivers/mtd/nand/raw/sh_flctl.c
383
dma_addr_t dma_addr;
drivers/mtd/nand/raw/sh_flctl.c
397
dma_addr = dma_map_single(chan->device->dev, buf, len, dir);
drivers/mtd/nand/raw/sh_flctl.c
399
if (!dma_mapping_error(chan->device->dev, dma_addr))
drivers/mtd/nand/raw/sh_flctl.c
400
desc = dmaengine_prep_slave_single(chan, dma_addr, len,
drivers/mtd/nand/raw/sh_flctl.c
443
dma_unmap_single(chan->device->dev, dma_addr, len, dir);
drivers/mtd/nand/raw/tegra_nand.c
502
dma_addr_t dma_addr = 0, dma_addr_oob = 0;
drivers/mtd/nand/raw/tegra_nand.c
532
dma_addr = dma_map_single(ctrl->dev, buf, mtd->writesize, dir);
drivers/mtd/nand/raw/tegra_nand.c
533
ret = dma_mapping_error(ctrl->dev, dma_addr);
drivers/mtd/nand/raw/tegra_nand.c
540
writel_relaxed(dma_addr, ctrl->regs + DATA_PTR);
drivers/mtd/nand/raw/tegra_nand.c
614
dma_unmap_single(ctrl->dev, dma_addr, mtd->writesize, dir);
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
1100
dma_addr_t dma_addr[KVASER_PCIEFD_DMA_COUNT];
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
1110
&dma_addr[i],
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
1113
if (!pcie->dma_data[i] || !dma_addr[i]) {
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
1118
pcie->driver_data->ops->kvaser_pciefd_write_dma_map(pcie, dma_addr[i], i);
drivers/net/dsa/lantiq/lantiq_gswip.c
123
dma_addr_t dma_addr;
drivers/net/dsa/lantiq/lantiq_gswip.c
150
fw_addr = dmam_alloc_coherent(dev, size, &dma_addr, GFP_KERNEL);
drivers/net/dsa/lantiq/lantiq_gswip.c
153
dev_addr = ALIGN(dma_addr, XRX200_GPHY_FW_ALIGN);
drivers/net/ethernet/3com/3c59x.c
2120
dma_addr_t dma_addr;
drivers/net/ethernet/3com/3c59x.c
2155
dma_addr = dma_map_single(vp->gendev, skb->data, skb->len,
drivers/net/ethernet/3com/3c59x.c
2157
if (dma_mapping_error(vp->gendev, dma_addr))
drivers/net/ethernet/3com/3c59x.c
2160
vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/3c59x.c
2165
dma_addr = dma_map_single(vp->gendev, skb->data,
drivers/net/ethernet/3com/3c59x.c
2167
if (dma_mapping_error(vp->gendev, dma_addr))
drivers/net/ethernet/3com/3c59x.c
2170
vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/3c59x.c
2176
dma_addr = skb_frag_dma_map(vp->gendev, frag,
drivers/net/ethernet/3com/3c59x.c
2180
if (dma_mapping_error(vp->gendev, dma_addr)) {
drivers/net/ethernet/3com/3c59x.c
2196
cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/3c59x.c
2205
dma_addr = dma_map_single(vp->gendev, skb->data, skb->len, DMA_TO_DEVICE);
drivers/net/ethernet/3com/3c59x.c
2206
if (dma_mapping_error(vp->gendev, dma_addr))
drivers/net/ethernet/3com/3c59x.c
2208
vp->tx_ring[entry].addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/typhoon.c
1574
r->physAddr = cpu_to_le32(rxb->dma_addr);
drivers/net/ethernet/3com/typhoon.c
1589
dma_addr_t dma_addr;
drivers/net/ethernet/3com/typhoon.c
1608
dma_addr = dma_map_single(&tp->pdev->dev, skb->data, PKT_BUF_SZ,
drivers/net/ethernet/3com/typhoon.c
1617
r->physAddr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/typhoon.c
1619
rxb->dma_addr = dma_addr;
drivers/net/ethernet/3com/typhoon.c
1634
dma_addr_t dma_addr;
drivers/net/ethernet/3com/typhoon.c
1650
dma_addr = rxb->dma_addr;
drivers/net/ethernet/3com/typhoon.c
1664
dma_sync_single_for_cpu(&tp->pdev->dev, dma_addr,
drivers/net/ethernet/3com/typhoon.c
1667
dma_sync_single_for_device(&tp->pdev->dev, dma_addr,
drivers/net/ethernet/3com/typhoon.c
1675
dma_unmap_single(&tp->pdev->dev, dma_addr, PKT_BUF_SZ,
drivers/net/ethernet/3com/typhoon.c
1790
dma_unmap_single(&tp->pdev->dev, rxb->dma_addr,
drivers/net/ethernet/3com/typhoon.c
251
dma_addr_t dma_addr;
drivers/net/ethernet/actions/owl-emac.c
117
struct sk_buff *skb, dma_addr_t dma_addr)
drivers/net/ethernet/actions/owl-emac.c
121
dma_unmap_single(dev, dma_addr, skb_tailroom(skb), DMA_FROM_DEVICE);
drivers/net/ethernet/actions/owl-emac.c
133
struct sk_buff *skb, dma_addr_t dma_addr)
drivers/net/ethernet/actions/owl-emac.c
137
dma_unmap_single(dev, dma_addr, skb_headlen(skb), DMA_TO_DEVICE);
drivers/net/ethernet/actions/owl-emac.c
186
dma_addr_t dma_addr;
drivers/net/ethernet/actions/owl-emac.c
194
dma_addr = owl_emac_dma_map_rx(priv, skb);
drivers/net/ethernet/actions/owl-emac.c
195
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/actions/owl-emac.c
203
desc->buf_addr = dma_addr;
drivers/net/ethernet/actions/owl-emac.c
207
ring->skbs_dma[i] = dma_addr;
drivers/net/ethernet/actions/owl-emac.c
498
dma_addr_t dma_addr;
drivers/net/ethernet/actions/owl-emac.c
507
dma_addr = owl_emac_dma_map_tx(priv, skb);
drivers/net/ethernet/actions/owl-emac.c
508
if (dma_mapping_error(owl_emac_get_dev(priv), dma_addr)) {
drivers/net/ethernet/actions/owl-emac.c
525
owl_emac_dma_unmap_tx(priv, skb, dma_addr);
drivers/net/ethernet/actions/owl-emac.c
531
ring->skbs_dma[tx_head] = dma_addr;
drivers/net/ethernet/actions/owl-emac.c
538
WRITE_ONCE(desc->buf_addr, dma_addr);
drivers/net/ethernet/actions/owl-emac.c
574
dma_addr_t dma_addr;
drivers/net/ethernet/actions/owl-emac.c
576
dma_addr = owl_emac_dma_map_tx(priv, skb);
drivers/net/ethernet/actions/owl-emac.c
577
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/actions/owl-emac.c
600
owl_emac_dma_unmap_tx(priv, skb, dma_addr);
drivers/net/ethernet/actions/owl-emac.c
606
ring->skbs_dma[tx_head] = dma_addr;
drivers/net/ethernet/actions/owl-emac.c
613
WRITE_ONCE(desc->buf_addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
236
u32 dma_addr;
drivers/net/ethernet/aeroflex/greth.c
253
dma_addr = dma_map_single(greth->dev,
drivers/net/ethernet/aeroflex/greth.c
258
if (dma_mapping_error(greth->dev, dma_addr)) {
drivers/net/ethernet/aeroflex/greth.c
265
greth_write_bd(&rx_bd[i].addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
282
dma_addr = dma_map_single(greth->dev,
drivers/net/ethernet/aeroflex/greth.c
287
if (dma_mapping_error(greth->dev, dma_addr)) {
drivers/net/ethernet/aeroflex/greth.c
292
greth_write_bd(&rx_bd[i].addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
305
dma_addr = dma_map_single(greth->dev,
drivers/net/ethernet/aeroflex/greth.c
310
if (dma_mapping_error(greth->dev, dma_addr)) {
drivers/net/ethernet/aeroflex/greth.c
315
greth_write_bd(&tx_bd[i].addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
399
u32 status, dma_addr, ctrl;
drivers/net/ethernet/aeroflex/greth.c
426
dma_addr = greth_read_bd(&bdp->addr);
drivers/net/ethernet/aeroflex/greth.c
428
memcpy((unsigned char *) phys_to_virt(dma_addr), skb->data, skb->len);
drivers/net/ethernet/aeroflex/greth.c
430
dma_sync_single_for_device(greth->dev, dma_addr, skb->len, DMA_TO_DEVICE);
drivers/net/ethernet/aeroflex/greth.c
467
u32 status, dma_addr;
drivers/net/ethernet/aeroflex/greth.c
508
dma_addr = dma_map_single(greth->dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE);
drivers/net/ethernet/aeroflex/greth.c
510
if (unlikely(dma_mapping_error(greth->dev, dma_addr)))
drivers/net/ethernet/aeroflex/greth.c
513
greth_write_bd(&bdp->addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
540
dma_addr = skb_frag_dma_map(greth->dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/aeroflex/greth.c
543
if (unlikely(dma_mapping_error(greth->dev, dma_addr)))
drivers/net/ethernet/aeroflex/greth.c
546
greth_write_bd(&bdp->addr, dma_addr);
drivers/net/ethernet/aeroflex/greth.c
746
u32 status, dma_addr;
drivers/net/ethernet/aeroflex/greth.c
762
dma_addr = greth_read_bd(&bdp->addr);
drivers/net/ethernet/aeroflex/greth.c
800
dma_addr,
drivers/net/ethernet/aeroflex/greth.c
805
greth_print_rx_packet(phys_to_virt(dma_addr), pkt_len);
drivers/net/ethernet/aeroflex/greth.c
807
skb_put_data(skb, phys_to_virt(dma_addr),
drivers/net/ethernet/aeroflex/greth.c
825
dma_sync_single_for_device(greth->dev, dma_addr, MAX_FRAME_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/aeroflex/greth.c
862
u32 status, dma_addr;
drivers/net/ethernet/aeroflex/greth.c
900
dma_addr = dma_map_single(greth->dev,
drivers/net/ethernet/aeroflex/greth.c
905
if (!dma_mapping_error(greth->dev, dma_addr)) {
drivers/net/ethernet/aeroflex/greth.c
930
greth_write_bd(&bdp->addr, dma_addr);
drivers/net/ethernet/agere/et131x.c
2429
dma_addr_t dma_addr;
drivers/net/ethernet/agere/et131x.c
2458
dma_addr = dma_map_single(&adapter->pdev->dev,
drivers/net/ethernet/agere/et131x.c
2463
dma_addr))
drivers/net/ethernet/agere/et131x.c
2466
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2467
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2471
dma_addr = dma_map_single(&adapter->pdev->dev,
drivers/net/ethernet/agere/et131x.c
2476
dma_addr))
drivers/net/ethernet/agere/et131x.c
2479
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2480
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2484
dma_addr = dma_map_single(&adapter->pdev->dev,
drivers/net/ethernet/agere/et131x.c
2490
dma_addr))
drivers/net/ethernet/agere/et131x.c
2493
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2494
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2499
dma_addr = skb_frag_dma_map(&adapter->pdev->dev,
drivers/net/ethernet/agere/et131x.c
2504
if (dma_mapping_error(&adapter->pdev->dev, dma_addr))
drivers/net/ethernet/agere/et131x.c
2507
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2508
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2601
dma_addr = desc[frag].addr_lo;
drivers/net/ethernet/agere/et131x.c
2602
dma_addr |= (u64)desc[frag].addr_hi << 32;
drivers/net/ethernet/agere/et131x.c
2603
dma_unmap_page(&adapter->pdev->dev, dma_addr,
drivers/net/ethernet/agere/et131x.c
2610
dma_addr = desc[frag].addr_lo;
drivers/net/ethernet/agere/et131x.c
2611
dma_addr |= (u64)desc[frag].addr_hi << 32;
drivers/net/ethernet/agere/et131x.c
2612
dma_unmap_single(&adapter->pdev->dev, dma_addr,
drivers/net/ethernet/agere/et131x.c
2676
u64 dma_addr;
drivers/net/ethernet/agere/et131x.c
2689
dma_addr = desc->addr_lo;
drivers/net/ethernet/agere/et131x.c
2690
dma_addr |= (u64)desc->addr_hi << 32;
drivers/net/ethernet/agere/et131x.c
2693
dma_addr,
drivers/net/ethernet/airoha/airoha_eth.c
1004
&dma_addr, GFP_KERNEL);
drivers/net/ethernet/airoha/airoha_eth.c
1012
airoha_qdma_wr(qdma, REG_TX_IRQ_BASE(id), dma_addr);
drivers/net/ethernet/airoha/airoha_eth.c
1051
if (!e->dma_addr)
drivers/net/ethernet/airoha/airoha_eth.c
1054
dma_unmap_single(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
1057
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
1071
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_eth.c
1095
dma_addr = rmem->base;
drivers/net/ethernet/airoha/airoha_eth.c
1102
if (!dmam_alloc_coherent(eth->dev, size, &dma_addr,
drivers/net/ethernet/airoha/airoha_eth.c
1107
airoha_qdma_wr(qdma, REG_FWD_BUF_BASE, dma_addr);
drivers/net/ethernet/airoha/airoha_eth.c
1110
if (!dmam_alloc_coherent(eth->dev, size, &dma_addr, GFP_KERNEL))
drivers/net/ethernet/airoha/airoha_eth.c
1113
airoha_qdma_wr(qdma, REG_FWD_DSCP_BASE, dma_addr);
drivers/net/ethernet/airoha/airoha_eth.c
2001
e->dma_addr = addr;
drivers/net/ethernet/airoha/airoha_eth.c
2043
dma_unmap_single(dev->dev.parent, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
2045
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
563
e->dma_addr = page_pool_get_dma_addr(page) + offset;
drivers/net/ethernet/airoha/airoha_eth.c
568
WRITE_ONCE(desc->addr, cpu_to_le32(e->dma_addr));
drivers/net/ethernet/airoha/airoha_eth.c
627
dma_sync_single_for_cpu(eth->dev, e->dma_addr,
drivers/net/ethernet/airoha/airoha_eth.c
752
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_eth.c
772
&dma_addr, GFP_KERNEL);
drivers/net/ethernet/airoha/airoha_eth.c
778
airoha_qdma_wr(qdma, REG_RX_RING_BASE(qid), dma_addr);
drivers/net/ethernet/airoha/airoha_eth.c
806
dma_sync_single_for_cpu(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
909
dma_unmap_single(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
911
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
956
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_eth.c
970
&dma_addr, GFP_KERNEL);
drivers/net/ethernet/airoha/airoha_eth.c
985
airoha_qdma_wr(qdma, REG_TX_RING_BASE(qid), dma_addr);
drivers/net/ethernet/airoha/airoha_eth.c
999
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_eth.h
178
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_npu.c
168
dma_addr_t dma_addr;
drivers/net/ethernet/airoha/airoha_npu.c
171
dma_addr = dma_map_single(npu->dev, p, size, DMA_TO_DEVICE);
drivers/net/ethernet/airoha/airoha_npu.c
172
ret = dma_mapping_error(npu->dev, dma_addr);
drivers/net/ethernet/airoha/airoha_npu.c
178
regmap_write(npu->regmap, REG_CR_MBQ0_CTRL(0) + offset, dma_addr);
drivers/net/ethernet/airoha/airoha_npu.c
194
dma_unmap_single(npu->dev, dma_addr, size, DMA_TO_DEVICE);
drivers/net/ethernet/altera/altera_msgdma.c
111
csrwr32(lower_32_bits(buffer->dma_addr), priv->tx_dma_desc,
drivers/net/ethernet/altera/altera_msgdma.c
113
csrwr32(upper_32_bits(buffer->dma_addr), priv->tx_dma_desc,
drivers/net/ethernet/altera/altera_msgdma.c
156
dma_addr_t dma_addr = rxbuffer->dma_addr;
drivers/net/ethernet/altera/altera_msgdma.c
166
csrwr32(lower_32_bits(dma_addr), priv->rx_dma_desc,
drivers/net/ethernet/altera/altera_msgdma.c
168
csrwr32(upper_32_bits(dma_addr), priv->rx_dma_desc,
drivers/net/ethernet/altera/altera_sgdma.c
182
buffer->dma_addr, /* address of packet to xmit */
drivers/net/ethernet/altera/altera_sgdma.c
356
rxbuffer->dma_addr, /* write addr for rx dma */
drivers/net/ethernet/altera/altera_tse.h
364
dma_addr_t dma_addr;
drivers/net/ethernet/altera/altera_tse_main.c
202
rxbuffer->dma_addr = dma_map_single(priv->device, rxbuffer->skb->data,
drivers/net/ethernet/altera/altera_tse_main.c
206
if (dma_mapping_error(priv->device, rxbuffer->dma_addr)) {
drivers/net/ethernet/altera/altera_tse_main.c
211
rxbuffer->dma_addr &= (dma_addr_t)~3;
drivers/net/ethernet/altera/altera_tse_main.c
219
dma_addr_t dma_addr = rxbuffer->dma_addr;
drivers/net/ethernet/altera/altera_tse_main.c
223
if (dma_addr)
drivers/net/ethernet/altera/altera_tse_main.c
224
dma_unmap_single(priv->device, dma_addr,
drivers/net/ethernet/altera/altera_tse_main.c
229
rxbuffer->dma_addr = 0;
drivers/net/ethernet/altera/altera_tse_main.c
238
if (buffer->dma_addr) {
drivers/net/ethernet/altera/altera_tse_main.c
240
dma_unmap_page(priv->device, buffer->dma_addr,
drivers/net/ethernet/altera/altera_tse_main.c
243
dma_unmap_single(priv->device, buffer->dma_addr,
drivers/net/ethernet/altera/altera_tse_main.c
245
buffer->dma_addr = 0;
drivers/net/ethernet/altera/altera_tse_main.c
397
dma_unmap_single(priv->device, priv->rx_ring[entry].dma_addr,
drivers/net/ethernet/altera/altera_tse_main.c
548
dma_addr_t dma_addr;
drivers/net/ethernet/altera/altera_tse_main.c
569
dma_addr = dma_map_single(priv->device, skb->data, nopaged_len,
drivers/net/ethernet/altera/altera_tse_main.c
571
if (dma_mapping_error(priv->device, dma_addr)) {
drivers/net/ethernet/altera/altera_tse_main.c
578
buffer->dma_addr = dma_addr;
drivers/net/ethernet/amazon/ena/ena_com.c
121
cq->entries = dma_alloc_coherent(admin_queue->q_dmadev, size, &cq->dma_addr, GFP_KERNEL);
drivers/net/ethernet/amazon/ena/ena_com.c
143
aenq->entries = dma_alloc_coherent(ena_dev->dmadev, size, &aenq->dma_addr, GFP_KERNEL);
drivers/net/ethernet/amazon/ena/ena_com.c
153
addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(aenq->dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
154
addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(aenq->dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
1625
dma_free_coherent(ena_dev->dmadev, size, sq->entries, sq->dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
1630
dma_free_coherent(ena_dev->dmadev, size, cq->entries, cq->dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
1635
dma_free_coherent(ena_dev->dmadev, size, aenq->entries, aenq->dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
2010
addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(admin_queue->sq.dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
2011
addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(admin_queue->sq.dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
2016
addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(admin_queue->cq.dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
2017
addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(admin_queue->cq.dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
99
sq->entries = dma_alloc_coherent(admin_queue->q_dmadev, size, &sq->dma_addr, GFP_KERNEL);
drivers/net/ethernet/amazon/ena/ena_com.h
187
dma_addr_t dma_addr;
drivers/net/ethernet/amazon/ena/ena_com.h
195
dma_addr_t dma_addr;
drivers/net/ethernet/amazon/ena/ena_com.h
254
dma_addr_t dma_addr;
drivers/net/ethernet/amazon/ena/ena_netdev.c
534
rx_info->dma_addr = dma;
drivers/net/ethernet/amazon/ena/ena_netdev.c
547
dma_unmap_page_attrs(rx_ring->dev, rx_info->dma_addr, ENA_PAGE_SIZE, DMA_BIDIRECTIONAL,
drivers/net/ethernet/amazon/ena/ena_netdev.h
174
dma_addr_t dma_addr;
drivers/net/ethernet/amd/au1000_eth.c
1136
&aup->dma_addr, 0);
drivers/net/ethernet/amd/au1000_eth.c
1238
pDB->dma_addr = aup->dma_addr + MAX_BUF_SIZE * i;
drivers/net/ethernet/amd/au1000_eth.c
1249
aup->rx_dma_ring[i]->buff_stat = lower_32_bits(pDB->dma_addr);
drivers/net/ethernet/amd/au1000_eth.c
1258
aup->tx_dma_ring[i]->buff_stat = lower_32_bits(pDB->dma_addr);
drivers/net/ethernet/amd/au1000_eth.c
1313
aup->vaddr, aup->dma_addr);
drivers/net/ethernet/amd/au1000_eth.c
1346
aup->vaddr, aup->dma_addr);
drivers/net/ethernet/amd/au1000_eth.c
823
prxd->buff_stat = lower_32_bits(pDB->dma_addr) | RX_DMA_ENABLE;
drivers/net/ethernet/amd/au1000_eth.c
999
ptxd->buff_stat = lower_32_bits(pDB->dma_addr) | TX_DMA_ENABLE;
drivers/net/ethernet/amd/au1000_eth.h
110
dma_addr_t dma_addr; /* dma address of rx/tx buffers */
drivers/net/ethernet/amd/au1000_eth.h
32
dma_addr_t dma_addr;
drivers/net/ethernet/amd/pds_core/core.h
86
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
102
raw_desc->m0 = cpu_to_le64(SET_BITS(PKT_ADDRL, dma_addr) |
drivers/net/ethernet/apm/xgene-v2/main.c
176
static dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
193
pkt_buf = dma_alloc_coherent(dev, XGENE_ENET_STD_MTU, &dma_addr,
drivers/net/ethernet/apm/xgene-v2/main.c
206
upper_32_bits(dma_addr)));
drivers/net/ethernet/apm/xgene-v2/main.c
209
tx_ring->pkt_info[tail].dma_addr = dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
214
raw_desc->m0 = cpu_to_le64(SET_BITS(PKT_ADDRL, dma_addr) |
drivers/net/ethernet/apm/xgene-v2/main.c
240
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
262
dma_addr = tx_ring->pkt_info[head].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
266
dma_free_coherent(dev, XGENE_ENET_STD_MTU, pkt_buf, dma_addr);
drivers/net/ethernet/apm/xgene-v2/main.c
290
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
314
dma_addr = rx_ring->pkt_info[head].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
316
dma_unmap_single(dev, dma_addr, XGENE_ENET_STD_MTU,
drivers/net/ethernet/apm/xgene-v2/main.c
361
dma_free_coherent(dev, size, ring->desc_addr, ring->dma_addr);
drivers/net/ethernet/apm/xgene-v2/main.c
373
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
378
dma_addr = ring->pkt_info[i].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
383
dma_unmap_single(dev, dma_addr, XGENE_ENET_STD_MTU,
drivers/net/ethernet/apm/xgene-v2/main.c
415
ring->desc_addr = dma_alloc_coherent(dev, size, &ring->dma_addr,
drivers/net/ethernet/apm/xgene-v2/main.c
554
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
568
dma_addr = tx_ring->pkt_info[i].dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
570
dma_free_coherent(dev, XGENE_ENET_STD_MTU, pkt_buf, dma_addr);
drivers/net/ethernet/apm/xgene-v2/main.c
72
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
84
dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE);
drivers/net/ethernet/apm/xgene-v2/main.c
85
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/apm/xgene-v2/main.c
92
ring->pkt_info[tail].dma_addr = dma_addr;
drivers/net/ethernet/apm/xgene-v2/main.c
99
upper_32_bits(dma_addr)));
drivers/net/ethernet/apm/xgene-v2/ring.c
24
next_dma = ring->dma_addr + (offset * XGENE_ENET_DESC_SIZE);
drivers/net/ethernet/apm/xgene-v2/ring.c
37
dma_addr_t dma_addr = ring->dma_addr;
drivers/net/ethernet/apm/xgene-v2/ring.c
39
xge_wr_csr(pdata, DMATXDESCL, dma_addr);
drivers/net/ethernet/apm/xgene-v2/ring.c
40
xge_wr_csr(pdata, DMATXDESCH, upper_32_bits(dma_addr));
drivers/net/ethernet/apm/xgene-v2/ring.c
49
dma_addr_t dma_addr = ring->dma_addr;
drivers/net/ethernet/apm/xgene-v2/ring.c
51
xge_wr_csr(pdata, DMARXDESCL, dma_addr);
drivers/net/ethernet/apm/xgene-v2/ring.c
52
xge_wr_csr(pdata, DMARXDESCH, upper_32_bits(dma_addr));
drivers/net/ethernet/apm/xgene-v2/ring.h
68
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene-v2/ring.h
75
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
104
raw_desc->m1 = cpu_to_le64(SET_VAL(DATAADDR, dma_addr) |
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
126
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
146
dma_addr = dma_map_single(dev, skb->data, len, DMA_FROM_DEVICE);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
147
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
155
raw_desc->m1 = cpu_to_le64(SET_VAL(DATAADDR, dma_addr) |
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
178
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
187
dma_addr = GET_VAL(DATAADDR, le64_to_cpu(raw_desc->m1));
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
188
dma_unmap_single(dev, dma_addr, XGENE_ENET_MAX_MTU,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
197
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
205
dma_addr = buf_pool->frag_dma_addr[i];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
206
dma_unmap_page(dev, dma_addr, PAGE_SIZE,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
413
dma_addr_t dma_addr, pbuf_addr, *frag_dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
438
dma_addr = dma_map_single(dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
439
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
445
raw_desc->m1 = cpu_to_le64(SET_VAL(DATAADDR, dma_addr) |
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
490
dma_addr = pbuf_addr + offset;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
497
xgene_set_addr_len(exp_desc, i, dma_addr, hw_len);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
502
xgene_set_addr_len(exp_bufs, idx, dma_addr,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
507
xgene_set_addr_len(exp_desc, i, dma_addr,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
512
xgene_set_addr_len(exp_bufs, idx, dma_addr, hw_len);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
525
dma_addr = dma_map_single(dev, exp_bufs,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
528
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
533
exp_desc[2] = cpu_to_le64(SET_VAL(DATAADDR, dma_addr) |
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
613
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
633
dma_addr = GET_VAL(DATAADDR, le64_to_cpu(desc[i ^ 1]));
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
634
dma_unmap_page(dev, dma_addr, PAGE_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
676
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
73
dma_addr_t dma_addr;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
738
dma_addr = GET_VAL(DATAADDR, le64_to_cpu(desc[i ^ 1]));
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
739
dma_unmap_page(dev, dma_addr, PAGE_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
96
dma_addr = dma_map_page(dev, page, 0,
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
98
if (unlikely(dma_mapping_error(dev, dma_addr))) {
drivers/net/ethernet/atheros/ag71xx.c
1187
ring->buf[i].rx.dma_addr,
drivers/net/ethernet/atheros/ag71xx.c
1214
buf->rx.dma_addr = dma_map_single(&ag->pdev->dev, data, ag->rx_buf_size,
drivers/net/ethernet/atheros/ag71xx.c
1216
if (dma_mapping_error(&ag->pdev->dev, buf->rx.dma_addr)) {
drivers/net/ethernet/atheros/ag71xx.c
1221
desc->data = (u32)buf->rx.dma_addr + offset;
drivers/net/ethernet/atheros/ag71xx.c
1506
dma_addr_t dma_addr;
drivers/net/ethernet/atheros/ag71xx.c
1517
dma_addr = dma_map_single(&ag->pdev->dev, skb->data, skb->len,
drivers/net/ethernet/atheros/ag71xx.c
1519
if (dma_mapping_error(&ag->pdev->dev, dma_addr)) {
drivers/net/ethernet/atheros/ag71xx.c
1528
n = ag71xx_fill_dma_desc(ring, (u32)dma_addr,
drivers/net/ethernet/atheros/ag71xx.c
1564
dma_unmap_single(&ag->pdev->dev, dma_addr, skb->len, DMA_TO_DEVICE);
drivers/net/ethernet/atheros/ag71xx.c
1640
dma_unmap_single(&ag->pdev->dev, ring->buf[i].rx.dma_addr,
drivers/net/ethernet/atheros/ag71xx.c
311
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/asp2/bcmasp.h
166
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
140
dma_unmap_addr_set(txcb, dma_addr, 0);
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
301
dma_unmap_addr_set(txcb, dma_addr, mapping);
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
422
mapping = dma_unmap_addr(txcb, dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
175
ring->cpu_addr = dma_alloc_coherent(dev, size, &ring->dma_addr, GFP_KERNEL);
drivers/net/ethernet/broadcom/bcm4908_enet.c
191
dma_free_coherent(dev, size, ring->cpu_addr, ring->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
205
dma_free_coherent(dev, size, rx_ring->cpu_addr, rx_ring->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
210
dma_free_coherent(dev, size, tx_ring->cpu_addr, tx_ring->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
278
slot->dma_addr = dma_map_single(dev, slot->buf + ENET_RX_BUF_DMA_OFFSET,
drivers/net/ethernet/broadcom/bcm4908_enet.c
280
err = dma_mapping_error(dev, slot->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
293
buf_desc->addr = cpu_to_le32(slot->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
313
(uint32_t)ring->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
330
dma_unmap_single(dev, slot->dma_addr, slot->len, DMA_FROM_DEVICE);
drivers/net/ethernet/broadcom/bcm4908_enet.c
556
slot->dma_addr = dma_map_single(dev, skb->data, skb->len, DMA_TO_DEVICE);
drivers/net/ethernet/broadcom/bcm4908_enet.c
557
if (unlikely(dma_mapping_error(dev, slot->dma_addr)))
drivers/net/ethernet/broadcom/bcm4908_enet.c
570
buf_desc->addr = cpu_to_le32((uint32_t)slot->dma_addr);
drivers/net/ethernet/broadcom/bcm4908_enet.c
58
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bcm4908_enet.c
620
dma_unmap_single(dev, slot.dma_addr, ENET_RX_BUF_DMA_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/broadcom/bcm4908_enet.c
667
dma_unmap_single(dev, slot->dma_addr, slot->len, DMA_TO_DEVICE);
drivers/net/ethernet/broadcom/bcm4908_enet.c
74
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bcmsysport.c
1335
dma_unmap_addr_set(cb, dma_addr, mapping);
drivers/net/ethernet/broadcom/bcmsysport.c
1722
if (dma_unmap_addr(cb, dma_addr))
drivers/net/ethernet/broadcom/bcmsysport.c
1724
dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/bcmsysport.c
658
dma_unmap_addr_set(cb, dma_addr, 0);
drivers/net/ethernet/broadcom/bcmsysport.c
690
dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/bcmsysport.c
695
dma_unmap_addr_set(cb, dma_addr, mapping);
drivers/net/ethernet/broadcom/bcmsysport.c
861
dma_unmap_single(kdev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/bcmsysport.c
867
} else if (dma_unmap_addr(cb, dma_addr)) {
drivers/net/ethernet/broadcom/bcmsysport.c
869
dma_unmap_page(kdev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/bcmsysport.c
871
dma_unmap_addr_set(cb, dma_addr, 0);
drivers/net/ethernet/broadcom/bcmsysport.h
666
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/broadcom/bgmac.c
125
dma_desc->addr_low = cpu_to_le32(lower_32_bits(slot->dma_addr));
drivers/net/ethernet/broadcom/bgmac.c
126
dma_desc->addr_high = cpu_to_le32(upper_32_bits(slot->dma_addr));
drivers/net/ethernet/broadcom/bgmac.c
162
slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb),
drivers/net/ethernet/broadcom/bgmac.c
164
if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
drivers/net/ethernet/broadcom/bgmac.c
180
slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0,
drivers/net/ethernet/broadcom/bgmac.c
182
if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr)))
drivers/net/ethernet/broadcom/bgmac.c
211
dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb),
drivers/net/ethernet/broadcom/bgmac.c
220
dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE);
drivers/net/ethernet/broadcom/bgmac.c
262
dma_unmap_single(dma_dev, slot->dma_addr, len,
drivers/net/ethernet/broadcom/bgmac.c
265
dma_unmap_page(dma_dev, slot->dma_addr, len,
drivers/net/ethernet/broadcom/bgmac.c
279
slot->dma_addr = 0;
drivers/net/ethernet/broadcom/bgmac.c
337
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bgmac.c
352
dma_addr = dma_map_single(dma_dev, buf + BGMAC_RX_BUF_OFFSET,
drivers/net/ethernet/broadcom/bgmac.c
354
if (dma_mapping_error(dma_dev, dma_addr)) {
drivers/net/ethernet/broadcom/bgmac.c
362
slot->dma_addr = dma_addr;
drivers/net/ethernet/broadcom/bgmac.c
391
dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[desc_idx].dma_addr));
drivers/net/ethernet/broadcom/bgmac.c
392
dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[desc_idx].dma_addr));
drivers/net/ethernet/broadcom/bgmac.c
404
dma_sync_single_for_cpu(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE,
drivers/net/ethernet/broadcom/bgmac.c
408
dma_sync_single_for_device(dma_dev, slot->dma_addr, BGMAC_RX_BUF_SIZE,
drivers/net/ethernet/broadcom/bgmac.c
430
dma_addr_t dma_addr = slot->dma_addr;
drivers/net/ethernet/broadcom/bgmac.c
441
dma_unmap_single(dma_dev, dma_addr,
drivers/net/ethernet/broadcom/bgmac.c
540
if (!slot->dma_addr)
drivers/net/ethernet/broadcom/bgmac.c
544
dma_unmap_single(dma_dev, slot->dma_addr,
drivers/net/ethernet/broadcom/bgmac.c
547
dma_unmap_page(dma_dev, slot->dma_addr,
drivers/net/ethernet/broadcom/bgmac.c
561
if (!slot->dma_addr)
drivers/net/ethernet/broadcom/bgmac.c
564
dma_unmap_single(dma_dev, slot->dma_addr,
drivers/net/ethernet/broadcom/bgmac.c
568
slot->dma_addr = 0;
drivers/net/ethernet/broadcom/bgmac.h
396
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
772
dma_addr_t dma_addr,
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
786
dma_sync_single_for_cpu(bd->dev, dma_addr, len, bn->rx_dir);
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
818
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
920
dma_addr = rx_buf->mapping;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
923
skb = bnge_copy_skb(bnapi, data_ptr, len, dma_addr);
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
926
skb = bnge_rx_skb(bn, rxr, cons, data, data_ptr, dma_addr, len);
drivers/net/ethernet/broadcom/bnx2.c
3025
unsigned int len, unsigned int hdr_len, dma_addr_t dma_addr,
drivers/net/ethernet/broadcom/bnx2.c
3045
dma_unmap_single(&bp->pdev->dev, dma_addr, bp->rx_buf_use_size,
drivers/net/ethernet/broadcom/bnx2.c
3164
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bnx2.c
3178
dma_addr = dma_unmap_addr(rx_buf, mapping);
drivers/net/ethernet/broadcom/bnx2.c
3180
dma_sync_single_for_cpu(&bp->pdev->dev, dma_addr,
drivers/net/ethernet/broadcom/bnx2.c
3239
skb = bnx2_rx_skb(bp, rxr, data, len, hdr_len, dma_addr,
drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
2046
void bnx2x_write_dmae(struct bnx2x *bp, dma_addr_t dma_addr, u32 dst_addr,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
14658
dma_addr_t addr = ctl->data.io.dma_addr;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
585
void bnx2x_write_dmae(struct bnx2x *bp, dma_addr_t dma_addr, u32 dst_addr,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
605
dmae.src_addr_lo = U64_LO(dma_addr);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
606
dmae.src_addr_hi = U64_HI(dma_addr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1141
dma_addr_t dma_addr,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1155
dma_addr -= bp->rx_dma_offset;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1156
dma_sync_single_for_cpu(&bp->pdev->dev, dma_addr, rxr->rx_page_size,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1173
dma_addr_t dma_addr,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1189
dma_addr -= bp->rx_dma_offset;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1190
dma_sync_single_for_cpu(&bp->pdev->dev, dma_addr, rxr->rx_page_size,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1220
dma_addr_t dma_addr,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1234
dma_sync_single_for_cpu(&bp->pdev->dev, dma_addr, bp->rx_buf_use_size,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2119
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2231
dma_addr = rx_buf->mapping;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2262
skb = bnxt_copy_skb(bnapi, data_ptr, len, dma_addr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2264
skb = bnxt_copy_xdp(bnapi, &xdp, len, dma_addr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2283
skb = bp->rx_skb_func(bp, rxr, cons, data, data_ptr, dma_addr,
drivers/net/ethernet/broadcom/cnic.c
1111
uinfo->mem[1].dma_addr = cp->status_blk_map;
drivers/net/ethernet/broadcom/cnic.c
1123
uinfo->mem[1].dma_addr = cp->status_blk_map;
drivers/net/ethernet/broadcom/cnic.c
1133
uinfo->mem[2].dma_addr = udev->l2_ring_map;
drivers/net/ethernet/broadcom/cnic.c
1139
uinfo->mem[3].dma_addr = udev->l2_buf_map;
drivers/net/ethernet/broadcom/cnic.c
214
io->dma_addr = addr;
drivers/net/ethernet/broadcom/cnic_if.h
139
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1874
dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1878
dma_unmap_page(dev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1881
dma_unmap_addr_set(cb, dma_addr, 0);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1886
} else if (dma_unmap_addr(cb, dma_addr)) {
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1888
dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1891
dma_unmap_addr_set(cb, dma_addr, 0);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1906
if (dma_unmap_addr(cb, dma_addr)) {
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1907
dma_unmap_single(dev, dma_unmap_addr(cb, dma_addr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1909
dma_unmap_addr_set(cb, dma_addr, 0);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2189
dma_unmap_addr_set(tx_cb_ptr, dma_addr, mapping);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2285
dma_unmap_addr_set(cb, dma_addr, mapping);
drivers/net/ethernet/broadcom/genet/bcmgenet.h
473
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/broadcom/tg3.c
6854
dma_addr_t dma_addr;
drivers/net/ethernet/broadcom/tg3.c
6863
dma_addr = dma_unmap_addr(ri, mapping);
drivers/net/ethernet/broadcom/tg3.c
6869
dma_addr = dma_unmap_addr(ri, mapping);
drivers/net/ethernet/broadcom/tg3.c
6908
dma_unmap_single(&tp->pdev->dev, dma_addr, skb_size,
drivers/net/ethernet/broadcom/tg3.c
6937
dma_sync_single_for_cpu(&tp->pdev->dev, dma_addr, len,
drivers/net/ethernet/broadcom/tg3.c
6942
dma_sync_single_for_device(&tp->pdev->dev, dma_addr,
drivers/net/ethernet/brocade/bna/bfa_cee.c
176
bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa);
drivers/net/ethernet/brocade/bna/bfa_ioc.h
50
#define bfa_dma_be_addr_set(dma_addr, pa) \
drivers/net/ethernet/brocade/bna/bfa_ioc.h
51
__bfa_dma_be_addr_set(&dma_addr, (u64)pa)
drivers/net/ethernet/brocade/bna/bfa_ioc.h
53
__bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa)
drivers/net/ethernet/brocade/bna/bfa_ioc.h
55
dma_addr->a32.addr_lo = (u32) htonl(pa);
drivers/net/ethernet/brocade/bna/bfa_ioc.h
56
dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa));
drivers/net/ethernet/brocade/bna/bfi_cna.h
106
union bfi_addr_u dma_addr;
drivers/net/ethernet/brocade/bna/bfi_cna.h
125
union bfi_addr_u dma_addr;
drivers/net/ethernet/brocade/bna/bfi_cna.h
49
union bfi_addr_u dma_addr;
drivers/net/ethernet/brocade/bna/bnad.c
105
dma_unmap_addr(&unmap->vectors[0], dma_addr),
drivers/net/ethernet/brocade/bna/bnad.c
107
dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0);
drivers/net/ethernet/brocade/bna/bnad.c
120
dma_unmap_addr(&unmap->vectors[vector], dma_addr),
drivers/net/ethernet/brocade/bna/bnad.c
123
dma_unmap_addr_set(&unmap->vectors[vector], dma_addr, 0);
drivers/net/ethernet/brocade/bna/bnad.c
2920
dma_addr_t dma_addr;
drivers/net/ethernet/brocade/bna/bnad.c
299
dma_unmap_addr(&unmap->vector, dma_addr),
drivers/net/ethernet/brocade/bna/bnad.c
3014
dma_addr = dma_map_single(&bnad->pcidev->dev, skb->data,
drivers/net/ethernet/brocade/bna/bnad.c
3016
if (dma_mapping_error(&bnad->pcidev->dev, dma_addr)) {
drivers/net/ethernet/brocade/bna/bnad.c
3021
BNA_SET_DMA_ADDR(dma_addr, &txqent->vector[0].host_addr);
drivers/net/ethernet/brocade/bna/bnad.c
3023
dma_unmap_addr_set(&unmap->vectors[0], dma_addr, dma_addr);
drivers/net/ethernet/brocade/bna/bnad.c
303
dma_unmap_addr_set(&unmap->vector, dma_addr, 0);
drivers/net/ethernet/brocade/bna/bnad.c
3050
dma_addr = skb_frag_dma_map(&bnad->pcidev->dev, frag,
drivers/net/ethernet/brocade/bna/bnad.c
3052
if (dma_mapping_error(&bnad->pcidev->dev, dma_addr)) {
drivers/net/ethernet/brocade/bna/bnad.c
3062
BNA_SET_DMA_ADDR(dma_addr, &txqent->vector[vect_id].host_addr);
drivers/net/ethernet/brocade/bna/bnad.c
3064
dma_unmap_addr_set(&unmap->vectors[vect_id], dma_addr,
drivers/net/ethernet/brocade/bna/bnad.c
3065
dma_addr);
drivers/net/ethernet/brocade/bna/bnad.c
314
dma_unmap_addr(&unmap->vector, dma_addr),
drivers/net/ethernet/brocade/bna/bnad.c
318
dma_unmap_addr_set(&unmap->vector, dma_addr, 0);
drivers/net/ethernet/brocade/bna/bnad.c
348
dma_addr_t dma_addr;
drivers/net/ethernet/brocade/bna/bnad.c
376
dma_addr = dma_map_page(&bnad->pcidev->dev, page, page_offset,
drivers/net/ethernet/brocade/bna/bnad.c
378
if (dma_mapping_error(&bnad->pcidev->dev, dma_addr)) {
drivers/net/ethernet/brocade/bna/bnad.c
387
dma_unmap_addr_set(&unmap->vector, dma_addr, dma_addr);
drivers/net/ethernet/brocade/bna/bnad.c
397
BNA_SET_DMA_ADDR(dma_addr, &rxent->host_addr);
drivers/net/ethernet/brocade/bna/bnad.c
421
dma_addr_t dma_addr;
drivers/net/ethernet/brocade/bna/bnad.c
439
dma_addr = dma_map_single(&bnad->pcidev->dev, skb->data,
drivers/net/ethernet/brocade/bna/bnad.c
441
if (dma_mapping_error(&bnad->pcidev->dev, dma_addr)) {
drivers/net/ethernet/brocade/bna/bnad.c
449
dma_unmap_addr_set(&unmap->vector, dma_addr, dma_addr);
drivers/net/ethernet/brocade/bna/bnad.c
453
BNA_SET_DMA_ADDR(dma_addr, &rxent->host_addr);
drivers/net/ethernet/brocade/bna/bnad.c
549
dma_unmap_addr(&unmap->vector, dma_addr),
drivers/net/ethernet/brocade/bna/bnad.c
577
dma_unmap_addr(&unmap->vector, dma_addr),
drivers/net/ethernet/brocade/bna/bnad.h
216
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/brocade/bna/bnad.h
227
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/cavium/liquidio/octeon_iq.h
271
u64 dma_addr;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
432
#define lio_dma_alloc(oct, size, dma_addr) \
drivers/net/ethernet/cavium/liquidio/octeon_network.h
433
dma_alloc_coherent(&(oct)->pci_dev->dev, size, dma_addr, GFP_KERNEL)
drivers/net/ethernet/cavium/liquidio/octeon_network.h
434
#define lio_dma_free(oct, size, virt_addr, dma_addr) \
drivers/net/ethernet/cavium/liquidio/octeon_network.h
435
dma_free_coherent(&(oct)->pci_dev->dev, size, virt_addr, dma_addr)
drivers/net/ethernet/cavium/liquidio/octeon_network.h
452
dma_addr_t dma_addr;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
464
dma_addr = pg_info->dma;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
470
dma_addr += pg_info->page_offset;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
472
return (u64)dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
741
u64 dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
752
(dma_addr_t *)&dma_addr);
drivers/net/ethernet/cavium/liquidio/request_manager.c
758
sc->dma_addr = dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
847
lio_dma_free(oct, sc->size, sc, sc->dma_addr);
drivers/net/ethernet/cavium/liquidio/request_manager.c
863
u64 dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
893
dma_addr = sc->dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
898
sc->dma_addr = dma_addr;
drivers/net/ethernet/cavium/liquidio/request_manager.c
911
sc->dmadptr = dma_addr + offset;
drivers/net/ethernet/cavium/liquidio/request_manager.c
921
sc->dmarptr = dma_addr + offset;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
537
u64 dma_addr, cpu_addr;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
542
dma_addr = *((u64 *)((void *)cqe_rx + (7 * sizeof(u64))));
drivers/net/ethernet/cavium/thunder/nicvf_main.c
544
cpu_addr = nicvf_iova_to_phys(nic, dma_addr);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
563
dma_addr -= offset;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
574
dma_addr &= PAGE_MASK;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
575
dma_unmap_page_attrs(&nic->pdev->dev, dma_addr,
drivers/net/ethernet/cavium/thunder/nicvf_main.c
590
nicvf_xdp_sq_append_pkt(nic, sq, (u64)xdp.data, dma_addr, len);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
605
dma_addr &= PAGE_MASK;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
606
dma_unmap_page_attrs(&nic->pdev->dev, dma_addr,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1236
u64 bufaddr, u64 dma_addr, u16 len)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1249
nicvf_sq_add_gather_subdesc(sq, qentry, len, dma_addr);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
137
pgcache->dma_addr = 0;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1555
u64 dma_addr;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1576
dma_addr = dma_map_page_attrs(&nic->pdev->dev, virt_to_page(skb->data),
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1579
if (dma_mapping_error(&nic->pdev->dev, dma_addr)) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1584
nicvf_sq_add_gather_subdesc(sq, qentry, size, dma_addr);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1595
dma_addr = dma_map_page_attrs(&nic->pdev->dev,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1600
if (dma_mapping_error(&nic->pdev->dev, dma_addr)) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1608
nicvf_sq_add_gather_subdesc(sq, qentry, size, dma_addr);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1637
static void nicvf_unmap_rcv_buffer(struct nicvf *nic, u64 dma_addr,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1655
dma_addr &= PAGE_MASK;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1657
dma_unmap_page_attrs(&nic->pdev->dev, dma_addr, len,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
210
if (rbdr->is_xdp && pgcache && pgcache->dma_addr) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
211
*rbuf = pgcache->dma_addr;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
225
pgcache->dma_addr = *rbuf + XDP_PACKET_HEADROOM;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
218
u64 dma_addr;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
320
static inline u64 nicvf_iova_to_phys(struct nicvf *nic, dma_addr_t dma_addr)
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
324
return iommu_iova_to_phys(nic->iommu_domain, dma_addr);
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
325
return dma_addr;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
346
u64 bufaddr, u64 dma_addr, u16 len);
drivers/net/ethernet/chelsio/cxgb/sge.c
1054
dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
1059
dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
1072
dma_unmap_single(&pdev->dev, dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
1096
dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
1253
dma_unmap_addr_set(ce, dma_addr, mapping);
drivers/net/ethernet/chelsio/cxgb/sge.c
1279
dma_unmap_addr_set(ce, dma_addr, mapping);
drivers/net/ethernet/chelsio/cxgb/sge.c
155
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
161
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
181
dma_addr_t dma_addr; /* DMA addr HW command descriptor Q */
drivers/net/ethernet/chelsio/cxgb/sge.c
196
dma_addr_t dma_addr; /* DMA addr HW freelist descriptor Q */
drivers/net/ethernet/chelsio/cxgb/sge.c
205
dma_addr_t dma_addr; /* DMA addr HW response descriptor Q */
drivers/net/ethernet/chelsio/cxgb/sge.c
506
dma_unmap_single(&pdev->dev, dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
526
sge->respQ.dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
539
q->dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
561
&q->dma_addr, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb/sge.c
598
dma_alloc_coherent(&pdev->dev, size, &sge->respQ.dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
623
dma_unmap_addr(ce, dma_addr),
drivers/net/ethernet/chelsio/cxgb/sge.c
663
q->dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
689
&q->dma_addr, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb/sge.c
750
setup_ring_params(ap, sge->cmdQ[0].dma_addr, sge->cmdQ[0].size,
drivers/net/ethernet/chelsio/cxgb/sge.c
752
setup_ring_params(ap, sge->cmdQ[1].dma_addr, sge->cmdQ[1].size,
drivers/net/ethernet/chelsio/cxgb/sge.c
754
setup_ring_params(ap, sge->freelQ[0].dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
757
setup_ring_params(ap, sge->freelQ[1].dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
764
setup_ring_params(ap, sge->respQ.dma_addr, sge->respQ.size,
drivers/net/ethernet/chelsio/cxgb/sge.c
842
dma_unmap_addr_set(ce, dma_addr, mapping);
drivers/net/ethernet/chelsio/cxgb3/sge.c
122
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2153
dma_unmap_addr(sd, dma_addr),
drivers/net/ethernet/chelsio/cxgb3/sge.c
361
dma_unmap_single(&pdev->dev, dma_unmap_addr(d, dma_addr),
drivers/net/ethernet/chelsio/cxgb3/sge.c
417
dma_unmap_addr_set(sd, dma_addr, mapping);
drivers/net/ethernet/chelsio/cxgb3/sge.c
518
dma_unmap_addr_set(sd, dma_addr, mapping);
drivers/net/ethernet/chelsio/cxgb3/sge.c
793
dma_unmap_addr(sd, dma_addr),
drivers/net/ethernet/chelsio/cxgb3/sge.c
797
dma_unmap_addr(sd, dma_addr),
drivers/net/ethernet/chelsio/cxgb3/sge.c
812
dma_unmap_single(&adap->pdev->dev, dma_unmap_addr(sd, dma_addr),
drivers/net/ethernet/chelsio/cxgb3/sge.c
846
dma_addr_t dma_addr = dma_unmap_addr(sd, dma_addr);
drivers/net/ethernet/chelsio/cxgb3/sge.c
853
dma_sync_single_for_cpu(&adap->pdev->dev, dma_addr,
drivers/net/ethernet/chelsio/cxgb3/sge.c
856
dma_sync_single_for_device(&adap->pdev->dev, dma_addr,
drivers/net/ethernet/chelsio/cxgb3/sge.c
881
dma_sync_single_for_cpu(&adap->pdev->dev, dma_addr, len,
drivers/net/ethernet/chelsio/cxgb4/sge.c
136
dma_addr_t dma_addr;
drivers/net/ethernet/chelsio/cxgb4/sge.c
193
return d->dma_addr & ~(dma_addr_t)RX_BUF_FLAGS;
drivers/net/ethernet/chelsio/cxgb4/sge.c
198
return !(d->dma_addr & RX_UNMAPPED_BUF);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3828
d->dma_addr |= RX_UNMAPPED_BUF;
drivers/net/ethernet/chelsio/cxgb4/sge.c
402
unsigned int rx_buf_size_idx = d->dma_addr & RX_BUF_SIZE;
drivers/net/ethernet/chelsio/cxgb4/sge.c
518
sd->dma_addr = mapping; /* includes size low bits */
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
170
dma_addr_t dma_addr; /* PCI DMA address (if mapped) */
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1720
sdesc->dma_addr |= RX_UNMAPPED_BUF;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
197
return sdesc->dma_addr & ~(dma_addr_t)(RX_LARGE_BUF | RX_UNMAPPED_BUF);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
209
return !(sdesc->dma_addr & RX_UNMAPPED_BUF);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
459
return (s->fl_pg_order > 0 && (sdesc->dma_addr & RX_LARGE_BUF)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
572
dma_addr_t dma_addr)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
575
sdesc->dma_addr = dma_addr;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
609
dma_addr_t dma_addr;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
645
dma_addr = dma_map_page(adapter->pdev_dev, page, 0,
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
648
if (unlikely(dma_mapping_error(adapter->pdev_dev, dma_addr))) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
660
dma_addr |= RX_LARGE_BUF;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
661
*d++ = cpu_to_be64(dma_addr);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
663
set_rx_sw_desc(sdesc, page, dma_addr);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
684
dma_addr = dma_map_page(adapter->pdev_dev, page, 0, PAGE_SIZE,
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
686
if (unlikely(dma_mapping_error(adapter->pdev_dev, dma_addr))) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
690
*d++ = cpu_to_be64(dma_addr);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
692
set_rx_sw_desc(sdesc, page, dma_addr);
drivers/net/ethernet/cisco/enic/enic.h
346
static inline int enic_dma_map_check(struct enic *enic, dma_addr_t dma_addr)
drivers/net/ethernet/cisco/enic/enic.h
348
if (unlikely(dma_mapping_error(&enic->pdev->dev, dma_addr))) {
drivers/net/ethernet/cisco/enic/enic_main.c
536
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_main.c
541
dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag, 0,
drivers/net/ethernet/cisco/enic/enic_main.c
544
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
546
enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag),
drivers/net/ethernet/cisco/enic/enic_main.c
561
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_main.c
564
dma_addr = dma_map_single(&enic->pdev->dev, skb->data, head_len,
drivers/net/ethernet/cisco/enic/enic_main.c
566
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
574
enic_queue_wq_desc(wq, skb, dma_addr, head_len, vlan_tag_insert,
drivers/net/ethernet/cisco/enic/enic_main.c
597
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_main.c
600
dma_addr = dma_map_single(&enic->pdev->dev, skb->data, head_len,
drivers/net/ethernet/cisco/enic/enic_main.c
602
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
610
enic_queue_wq_desc_csum_l4(wq, skb, dma_addr, head_len, csum_offset,
drivers/net/ethernet/cisco/enic/enic_main.c
674
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_main.c
694
dma_addr = dma_map_single(&enic->pdev->dev,
drivers/net/ethernet/cisco/enic/enic_main.c
697
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
699
enic_queue_wq_desc_tso(wq, skb, dma_addr, len, mss, hdr_len,
drivers/net/ethernet/cisco/enic/enic_main.c
720
dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag,
drivers/net/ethernet/cisco/enic/enic_main.c
723
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
725
enic_queue_wq_desc_cont(wq, skb, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_main.c
760
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_main.c
763
dma_addr = dma_map_single(&enic->pdev->dev, skb->data, head_len,
drivers/net/ethernet/cisco/enic/enic_main.c
765
if (unlikely(enic_dma_map_check(enic, dma_addr)))
drivers/net/ethernet/cisco/enic/enic_main.c
768
enic_queue_wq_desc_ex(wq, skb, dma_addr, head_len, mss_or_csum, 0,
drivers/net/ethernet/cisco/enic/enic_res.h
102
void *os_buf, dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/enic_res.h
106
enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_res.h
114
dma_addr_t dma_addr, unsigned int len)
drivers/net/ethernet/cisco/enic/enic_res.h
122
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/net/ethernet/cisco/enic/enic_res.h
125
vnic_rq_post(rq, os_buf, os_buf_index, dma_addr, len, wrid);
drivers/net/ethernet/cisco/enic/enic_res.h
34
void *os_buf, dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/enic_res.h
45
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/net/ethernet/cisco/enic/enic_res.h
55
vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop, desc_skip_cnt,
drivers/net/ethernet/cisco/enic/enic_res.h
60
void *os_buf, dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/enic_res.h
63
enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_res.h
69
dma_addr_t dma_addr, unsigned int len, int vlan_tag_insert,
drivers/net/ethernet/cisco/enic/enic_res.h
72
enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_res.h
79
void *os_buf, dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/enic_res.h
83
enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_res.h
91
void *os_buf, dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/enic_res.h
95
enic_queue_wq_desc_ex(wq, os_buf, dma_addr, len,
drivers/net/ethernet/cisco/enic/enic_rq.c
279
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/enic_rq.c
285
enic_queue_rq_desc(rq, buf->os_buf, os_buf_index, buf->dma_addr,
drivers/net/ethernet/cisco/enic/enic_rq.c
298
dma_addr = page_pool_get_dma_addr(page) + offset;
drivers/net/ethernet/cisco/enic/enic_rq.c
299
enic_queue_rq_desc(rq, (void *)page, os_buf_index, dma_addr, len);
drivers/net/ethernet/cisco/enic/enic_rq.c
362
dma_sync_single_for_cpu(&enic->pdev->dev, buf->dma_addr,
drivers/net/ethernet/cisco/enic/enic_rq.c
378
buf->dma_addr = 0;
drivers/net/ethernet/cisco/enic/enic_wq.c
46
dma_unmap_single(&enic->pdev->dev, buf->dma_addr, buf->len,
drivers/net/ethernet/cisco/enic/enic_wq.c
49
dma_unmap_page(&enic->pdev->dev, buf->dma_addr, buf->len,
drivers/net/ethernet/cisco/enic/vnic_rq.h
110
dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/cisco/enic/vnic_rq.h
117
buf->dma_addr = dma_addr;
drivers/net/ethernet/cisco/enic/vnic_rq.h
57
dma_addr_t dma_addr;
drivers/net/ethernet/cisco/enic/vnic_wq.h
119
void *os_buf, dma_addr_t dma_addr,
drivers/net/ethernet/cisco/enic/vnic_wq.h
131
buf->dma_addr = dma_addr;
drivers/net/ethernet/cisco/enic/vnic_wq.h
42
dma_addr_t dma_addr;
drivers/net/ethernet/engleder/tsnep_main.c
44
#define DMA_ADDR_HIGH(dma_addr) ((u32)(((dma_addr) >> 32) & 0xFFFFFFFF))
drivers/net/ethernet/engleder/tsnep_main.c
46
#define DMA_ADDR_HIGH(dma_addr) ((u32)(0))
drivers/net/ethernet/engleder/tsnep_main.c
48
#define DMA_ADDR_LOW(dma_addr) ((u32)((dma_addr) & 0xFFFFFFFF))
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
116
dma_addr_t dma_addr,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
122
TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid),
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
132
__field(dma_addr_t, dma_addr)
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
144
__entry->dma_addr = dma_addr;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
157
&__entry->dma_addr,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
167
dma_addr_t dma_addr,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
171
TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid)
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
179
dma_addr_t dma_addr,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-trace.h
183
TP_ARGS(netdev, vaddr, size, dma_addr, map_size, bpid)
drivers/net/ethernet/freescale/fec_main.c
4681
dma_addr_t dma_addr;
drivers/net/ethernet/freescale/fec_main.c
4701
dma_addr = dma_map_single(&fep->pdev->dev, xdpf->data,
drivers/net/ethernet/freescale/fec_main.c
4703
if (dma_mapping_error(&fep->pdev->dev, dma_addr))
drivers/net/ethernet/freescale/fec_main.c
4714
dma_addr = page_pool_get_dma_addr(page) +
drivers/net/ethernet/freescale/fec_main.c
4716
dma_sync_single_for_device(&fep->pdev->dev, dma_addr,
drivers/net/ethernet/freescale/fec_main.c
4727
bdp->cbd_bufaddr = cpu_to_fec32(dma_addr);
drivers/net/ethernet/fungible/funcore/fun_queue.c
101
sq_depth - 1, dma_addr, 0,
drivers/net/ethernet/fungible/funcore/fun_queue.c
126
u8 cqe_size_log2, u32 cq_depth, dma_addr_t dma_addr,
drivers/net/ethernet/fungible/funcore/fun_queue.c
146
cq_depth - 1, dma_addr, tailroom,
drivers/net/ethernet/fungible/funcore/fun_queue.c
21
int numa_node, dma_addr_t *dma_addr, void **sw_va,
drivers/net/ethernet/fungible/funcore/fun_queue.c
37
va = dma_alloc_coherent(dma_dev, dma_sz, dma_addr, GFP_KERNEL);
drivers/net/ethernet/fungible/funcore/fun_queue.c
46
dma_free_coherent(dma_dev, dma_sz, va, *dma_addr);
drivers/net/ethernet/fungible/funcore/fun_queue.c
58
bool wb, void *hw_va, dma_addr_t dma_addr, void *sw_va)
drivers/net/ethernet/fungible/funcore/fun_queue.c
65
dma_free_coherent(dma_dev, sz, hw_va, dma_addr);
drivers/net/ethernet/fungible/funcore/fun_queue.c
76
u8 sqe_size_log2, u32 sq_depth, dma_addr_t dma_addr,
drivers/net/ethernet/fungible/funcore/fun_queue.c
94
wb_addr = dma_addr + (sq_depth << sqe_size_log2);
drivers/net/ethernet/fungible/funcore/fun_queue.h
133
u8 sqe_size_log2, u32 sq_depth, dma_addr_t dma_addr,
drivers/net/ethernet/fungible/funcore/fun_queue.h
138
u8 cqe_size_log2, u32 cq_depth, dma_addr_t dma_addr,
drivers/net/ethernet/fungible/funcore/fun_queue.h
144
int numa_node, dma_addr_t *dma_addr, void **sw_va,
drivers/net/ethernet/fungible/funcore/fun_queue.h
147
bool wb, void *hw_va, dma_addr_t dma_addr, void *sw_va);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
110
rb->dma_addr = dma_map_page(q->dma_dev, p, 0, PAGE_SIZE,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
112
if (unlikely(dma_mapping_error(q->dma_dev, rb->dma_addr))) {
drivers/net/ethernet/fungible/funeth/funeth_rx.c
130
dma_unmap_page(q->dma_dev, rb->dma_addr, PAGE_SIZE,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
269
dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
277
FUN_EPRQ_RQBUF_INIT(buf->dma_addr);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
320
buf->dma_addr + q->buf_offset,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
559
q->rqes[i] = FUN_EPRQ_RQBUF_INIT(b->dma_addr);
drivers/net/ethernet/fungible/funeth/funeth_rx.c
58
dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
77
dma_sync_single_for_device(q->dma_dev, buf->dma_addr,
drivers/net/ethernet/fungible/funeth/funeth_rx.c
88
dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE,
drivers/net/ethernet/fungible/funeth/funeth_tx.c
644
&q->dma_addr, (void **)&q->info,
drivers/net/ethernet/fungible/funeth/funeth_tx.c
670
q->desc, q->dma_addr, q->info);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
696
q->dma_addr, fp->tx_coal_count, fp->tx_coal_usec,
drivers/net/ethernet/fungible/funeth/funeth_txrx.h
115
dma_addr_t dma_addr; /* DMA address of descriptor ring */
drivers/net/ethernet/fungible/funeth/funeth_txrx.h
150
dma_addr_t dma_addr; /* DMA address of page start */
drivers/net/ethernet/google/gve/gve_tx_dqo.c
756
void **va, dma_addr_t *dma_addr)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
762
*dma_addr = tx->dqo.qpl->page_buses[page_id] + offset;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
774
dma_addr_t dma_addr;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
786
gve_tx_buf_get_addr(tx, index, &va, &dma_addr);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
792
dma_sync_single_for_device(tx->dev, dma_addr,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
796
dma_addr,
drivers/net/ethernet/hisilicon/hisi_femac.c
145
dma_addr_t dma_addr;
drivers/net/ethernet/hisilicon/hisi_femac.c
147
dma_addr = priv->txq.dma_phys[pos];
drivers/net/ethernet/hisilicon/hisi_femac.c
148
dma_unmap_single(priv->dev, dma_addr, skb->len, DMA_TO_DEVICE);
drivers/net/ethernet/hisilicon/hisi_femac.c
390
dma_addr_t dma_addr;
drivers/net/ethernet/hisilicon/hisi_femac.c
402
dma_addr = rxq->dma_phys[pos];
drivers/net/ethernet/hisilicon/hisi_femac.c
403
dma_unmap_single(priv->dev, dma_addr, MAX_FRAME_SIZE,
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
133
&cmdq_buf->dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
150
dma_pool_free(cmdqs->cmdq_buf_pool, cmdq_buf->buf, cmdq_buf->dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
174
hinic_set_sge(&sge_resp->sge, buf_out->dma_addr, buf_out->size);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
231
hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.h
112
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
636
eq->dma_addr = devm_kcalloc(&pdev->dev, eq->num_pages,
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
637
sizeof(*eq->dma_addr), GFP_KERNEL);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
638
if (!eq->dma_addr)
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
651
&eq->dma_addr[pg],
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
659
val = upper_32_bits(eq->dma_addr[pg]);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
664
val = lower_32_bits(eq->dma_addr[pg]);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
682
eq->dma_addr[pg]);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
687
devm_kfree(&pdev->dev, eq->dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
704
eq->dma_addr[pg]);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
707
devm_kfree(&pdev->dev, eq->dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.h
191
dma_addr_t *dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
107
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
109
*vaddr = dma_alloc_coherent(&pdev->dev, page_sz, &dma_addr,
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
116
*paddr = (u64)dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
126
dma_free_coherent(&pdev->dev, page_sz, *vaddr, dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
426
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
428
dma_addr = (dma_addr_t)be64_to_cpu(*paddr);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
430
dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
472
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
475
&dma_addr, GFP_KERNEL);
drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c
482
*paddr = cpu_to_be64(dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
128
dma_addr_t *dma_addr)
drivers/net/ethernet/huawei/hinic/hinic_rx.c
150
*dma_addr = addr;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
163
static void rx_unmap_skb(struct hinic_rxq *rxq, dma_addr_t dma_addr)
drivers/net/ethernet/huawei/hinic/hinic_rx.c
170
dma_unmap_single(&pdev->dev, dma_addr, rxq->rq->buf_sz,
drivers/net/ethernet/huawei/hinic/hinic_rx.c
181
dma_addr_t dma_addr)
drivers/net/ethernet/huawei/hinic/hinic_rx.c
183
rx_unmap_skb(rxq, dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
199
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
211
skb = rx_alloc_skb(rxq, &dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
215
hinic_set_sge(&sge, dma_addr, skb->len);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
220
rx_free_skb(rxq, skb, dma_addr);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
139
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
142
dma_addr = dma_map_single(&pdev->dev, skb->data, skb_headlen(skb),
drivers/net/ethernet/huawei/hinic/hinic_tx.c
144
if (dma_mapping_error(&pdev->dev, dma_addr)) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
149
hinic_set_sge(&sges[0], dma_addr, skb_headlen(skb));
drivers/net/ethernet/huawei/hinic/hinic_tx.c
154
dma_addr = skb_frag_dma_map(&pdev->dev, frag, 0,
drivers/net/ethernet/huawei/hinic/hinic_tx.c
157
if (dma_mapping_error(&pdev->dev, dma_addr)) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
162
hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag));
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
131
&cmd_buf->dma_addr);
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
158
dma_pool_free(cmdqs->cmd_buf_pool, cmd_buf->buf, cmd_buf->dma_addr);
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
307
hinic3_set_sge(sge, buf_out->dma_addr, cpu_to_le32(CMDQ_BUF_SIZE));
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
321
hinic3_set_sge(&wqe->buf_desc.sge, buf_in->dma_addr,
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.h
96
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
495
u64 dma_addr, offset;
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
508
dma_addr = mq->dma_buf_paddr + offset;
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
509
dma_msg->dma_addr_high = cpu_to_le32(upper_32_bits(dma_addr));
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
510
dma_msg->dma_addr_low = cpu_to_le32(lower_32_bits(dma_addr));
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
137
dma_addr_t dma_addr, u16 len)
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
142
rq_wqe->buf_hi_addr = cpu_to_le32(upper_32_bits(dma_addr));
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
143
rq_wqe->buf_lo_addr = cpu_to_le32(lower_32_bits(dma_addr));
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
150
dma_addr_t dma_addr;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
161
dma_addr = page_pool_get_dma_addr(rx_info->page) +
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
163
rq_wqe_buf_set(rxq->rq, rxq->next_to_update, dma_addr,
drivers/net/ethernet/i825xx/lasi_82596.c
182
sizeof(struct i596_dma), &lp->dma_addr,
drivers/net/ethernet/i825xx/lasi_82596.c
194
lp->dma, lp->dma_addr, DMA_BIDIRECTIONAL);
drivers/net/ethernet/i825xx/lasi_82596.c
207
lp->dma_addr, DMA_BIDIRECTIONAL);
drivers/net/ethernet/i825xx/lib82596.c
1039
tx_cmd->dma_addr = dma_map_single(dev->dev.parent, skb->data,
drivers/net/ethernet/i825xx/lib82596.c
1041
tbd->data = SWAP32(tx_cmd->dma_addr);
drivers/net/ethernet/i825xx/lib82596.c
1208
tx_cmd->dma_addr,
drivers/net/ethernet/i825xx/lib82596.c
205
dma_addr_t dma_addr;
drivers/net/ethernet/i825xx/lib82596.c
326
dma_addr_t dma_addr;
drivers/net/ethernet/i825xx/lib82596.c
369
return lp->dma_addr + ((unsigned long)v - (unsigned long)lp->dma);
drivers/net/ethernet/i825xx/lib82596.c
498
dma_addr_t dma_addr;
drivers/net/ethernet/i825xx/lib82596.c
504
dma_addr = dma_map_single(dev->dev.parent, skb->data,
drivers/net/ethernet/i825xx/lib82596.c
511
rbd->b_data = SWAP32(dma_addr);
drivers/net/ethernet/i825xx/lib82596.c
720
dma_addr_t dma_addr;
drivers/net/ethernet/i825xx/lib82596.c
737
dma_addr = dma_map_single(dev->dev.parent,
drivers/net/ethernet/i825xx/lib82596.c
742
rbd->b_data = SWAP32(dma_addr);
drivers/net/ethernet/i825xx/lib82596.c
840
tx_cmd->dma_addr,
drivers/net/ethernet/i825xx/sni_82596.c
135
&lp->dma_addr, GFP_KERNEL);
drivers/net/ethernet/i825xx/sni_82596.c
146
lp->dma_addr);
drivers/net/ethernet/i825xx/sni_82596.c
163
lp->dma_addr);
drivers/net/ethernet/ibm/ibmveth.c
172
pool->dma_addr = kzalloc_objs(dma_addr_t, pool->size);
drivers/net/ethernet/ibm/ibmveth.c
173
if (!pool->dma_addr) {
drivers/net/ethernet/ibm/ibmveth.c
182
kfree(pool->dma_addr);
drivers/net/ethernet/ibm/ibmveth.c
183
pool->dma_addr = NULL;
drivers/net/ethernet/ibm/ibmveth.c
221
dma_addr_t dma_addr;
drivers/net/ethernet/ibm/ibmveth.c
258
dma_addr = dma_map_single(dev, skb->data,
drivers/net/ethernet/ibm/ibmveth.c
261
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/ibm/ibmveth.c
267
pool->dma_addr[index] = dma_addr;
drivers/net/ethernet/ibm/ibmveth.c
271
dma_addr = pool->dma_addr[index];
drivers/net/ethernet/ibm/ibmveth.c
285
descs[filled].fields.address = dma_addr;
drivers/net/ethernet/ibm/ibmveth.c
341
dma_addr = pool->dma_addr[index];
drivers/net/ethernet/ibm/ibmveth.c
344
if (dma_addr &&
drivers/net/ethernet/ibm/ibmveth.c
345
!dma_mapping_error(dev, dma_addr))
drivers/net/ethernet/ibm/ibmveth.c
346
dma_unmap_single(dev, dma_addr,
drivers/net/ethernet/ibm/ibmveth.c
421
if (pool->skbuff && pool->dma_addr) {
drivers/net/ethernet/ibm/ibmveth.c
426
pool->dma_addr[i],
drivers/net/ethernet/ibm/ibmveth.c
435
if (pool->dma_addr) {
drivers/net/ethernet/ibm/ibmveth.c
436
kfree(pool->dma_addr);
drivers/net/ethernet/ibm/ibmveth.c
437
pool->dma_addr = NULL;
drivers/net/ethernet/ibm/ibmveth.c
488
adapter->rx_buff_pool[pool].dma_addr[index],
drivers/net/ethernet/ibm/ibmveth.h
141
dma_addr_t *dma_addr;
drivers/net/ethernet/ibm/ibmvnic.c
1803
adapter->vpd->dma_addr =
drivers/net/ethernet/ibm/ibmvnic.c
1806
if (dma_mapping_error(dev, adapter->vpd->dma_addr)) {
drivers/net/ethernet/ibm/ibmvnic.c
1819
crq.get_vpd.ioba = cpu_to_be32(adapter->vpd->dma_addr);
drivers/net/ethernet/ibm/ibmvnic.c
2373
u64 dma_addr;
drivers/net/ethernet/ibm/ibmvnic.c
2379
dma_addr = (u64)ind_bufp->indir_dma;
drivers/net/ethernet/ibm/ibmvnic.c
2387
rc = send_subcrq_indirect(adapter, handle, dma_addr, entries);
drivers/net/ethernet/ibm/ibmvnic.c
2395
rc, entries, &dma_addr, handle);
drivers/net/ethernet/ibm/ibmvnic.c
5324
dma_unmap_single(dev, adapter->vpd->dma_addr, adapter->vpd->len,
drivers/net/ethernet/ibm/ibmvnic.c
784
dma_addr_t dma_addr;
drivers/net/ethernet/ibm/ibmvnic.c
828
dma_addr = ltb->addr + offset;
drivers/net/ethernet/ibm/ibmvnic.c
832
pool->rx_buff[bufidx].dma = dma_addr;
drivers/net/ethernet/ibm/ibmvnic.c
843
sub_crq->rx_add.ioba = cpu_to_be32(dma_addr);
drivers/net/ethernet/ibm/ibmvnic.h
889
dma_addr_t dma_addr;
drivers/net/ethernet/intel/e100.c
1387
cb->u.dump_buffer_addr = cpu_to_le32(nic->dma_addr +
drivers/net/ethernet/intel/e100.c
1548
nic->dma_addr + offsetof(struct mem, stats))))
drivers/net/ethernet/intel/e100.c
1737
dma_addr_t dma_addr;
drivers/net/ethernet/intel/e100.c
1740
dma_addr = dma_map_single(&nic->pdev->dev, skb->data, skb->len,
drivers/net/ethernet/intel/e100.c
1743
if (dma_mapping_error(&nic->pdev->dev, dma_addr))
drivers/net/ethernet/intel/e100.c
1758
cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd);
drivers/net/ethernet/intel/e100.c
1762
cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/intel/e100.c
1889
cb->dma_addr = nic->cbs_dma_addr + i * sizeof(struct cb);
drivers/net/ethernet/intel/e100.c
1910
e100_exec_cmd(nic, ruc_start, rx->dma_addr);
drivers/net/ethernet/intel/e100.c
1923
rx->dma_addr = dma_map_single(&nic->pdev->dev, rx->skb->data,
drivers/net/ethernet/intel/e100.c
1926
if (dma_mapping_error(&nic->pdev->dev, rx->dma_addr)) {
drivers/net/ethernet/intel/e100.c
1929
rx->dma_addr = 0;
drivers/net/ethernet/intel/e100.c
1938
put_unaligned_le32(rx->dma_addr, &prev_rfd->link);
drivers/net/ethernet/intel/e100.c
1940
rx->prev->dma_addr,
drivers/net/ethernet/intel/e100.c
1961
dma_sync_single_for_cpu(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
1981
dma_sync_single_for_device(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
1995
dma_unmap_single(&nic->pdev->dev, rx->dma_addr, RFD_BUF_LEN,
drivers/net/ethernet/intel/e100.c
2098
new_before_last_rx->dma_addr,
drivers/net/ethernet/intel/e100.c
2107
old_before_last_rx->dma_addr,
drivers/net/ethernet/intel/e100.c
2113
old_before_last_rx->dma_addr,
drivers/net/ethernet/intel/e100.c
2138
rx->dma_addr, RFD_BUF_LEN,
drivers/net/ethernet/intel/e100.c
2181
dma_sync_single_for_device(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
2381
dma_sync_single_for_cpu(&nic->pdev->dev, nic->rx_to_clean->dma_addr,
drivers/net/ethernet/intel/e100.c
2768
&nic->dma_addr, GFP_KERNEL);
drivers/net/ethernet/intel/e100.c
2776
nic->mem, nic->dma_addr);
drivers/net/ethernet/intel/e100.c
418
dma_addr_t dma_addr;
drivers/net/ethernet/intel/e100.c
496
dma_addr_t dma_addr;
drivers/net/ethernet/intel/e100.c
579
dma_addr_t dma_addr;
drivers/net/ethernet/intel/e100.c
649
u32 dma_addr = nic->dma_addr + offsetof(struct mem, selftest);
drivers/net/ethernet/intel/e100.c
657
iowrite32(selftest | dma_addr, &nic->csr->port);
drivers/net/ethernet/intel/e100.c
812
static int e100_exec_cmd(struct nic *nic, u8 cmd, dma_addr_t dma_addr)
drivers/net/ethernet/intel/e100.c
834
iowrite32(dma_addr, &nic->csr->scb.gen_ptr);
drivers/net/ethernet/intel/e100.c
878
nic->cb_to_send->dma_addr))) {
drivers/net/ethernet/marvell/mvneta.c
2110
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvneta.c
2124
dma_addr = dma_map_single(dev, data, len,
drivers/net/ethernet/marvell/mvneta.c
2126
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/marvell/mvneta.c
2135
dma_addr = page_pool_get_dma_addr(page);
drivers/net/ethernet/marvell/mvneta.c
2137
dma_addr += skb_frag_off(frag);
drivers/net/ethernet/marvell/mvneta.c
2139
dma_addr += sizeof(*xdpf) + xdpf->headroom;
drivers/net/ethernet/marvell/mvneta.c
2140
dma_sync_single_for_device(dev, dma_addr, len,
drivers/net/ethernet/marvell/mvneta.c
2147
tx_desc->buf_phys_addr = dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1521
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
1047
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
1073
buf = mvpp2_buf_alloc(port, bm_pool, pp, &dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
1078
mvpp2_bm_pool_put(port, bm_pool->id, dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
194
dma_addr_t dma_addr)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
198
addr = dma_addr & ~MVPP2_TX_DESC_ALIGN;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
199
offset = dma_addr & MVPP2_TX_DESC_ALIGN;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3591
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3596
&dma_addr, &phys_addr, GFP_ATOMIC);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3600
mvpp2_bm_pool_put(port, pool, dma_addr, phys_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3679
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3699
dma_addr = dma_map_single(port->dev->dev.parent, xdpf->data,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3702
if (unlikely(dma_mapping_error(port->dev->dev.parent, dma_addr))) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3713
dma_addr = page_pool_get_dma_addr(page) +
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3715
dma_sync_single_for_device(port->dev->dev.parent, dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3721
mvpp2_txdesc_dma_addr_set(port, tx_desc, dma_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3869
dma_addr_t dma_addr, dma_addr_next;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3873
dma_addr = mvpp2_rxdesc_cookie_get(port, rx_desc);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3886
mvpp2_bm_pool_put(port, pool, dma_addr, phys_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3889
dma_addr = dma_addr_next;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3921
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3936
dma_addr = mvpp2_rxdesc_dma_addr_get(port, rx_desc);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3949
dma_sync_single_for_cpu(dev->dev.parent, dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4032
dma_unmap_single_attrs(dev->dev.parent, dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4056
mvpp2_bm_pool_put(port, pool, dma_addr, phys_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
406
&bm_pool->dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
414
bm_pool->virt_addr, bm_pool->dma_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
421
lower_32_bits(bm_pool->dma_addr));
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
463
dma_addr_t *dma_addr,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
468
*dma_addr = mvpp2_thread_read(priv, thread,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
482
*dma_addr |= (u64)dma_addr_highbits << 32;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
576
bm_pool->dma_addr);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
768
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
778
dma_addr = page_pool_get_dma_addr(page);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
781
dma_addr = dma_map_single(port->dev->dev.parent, data,
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
784
if (unlikely(dma_mapping_error(port->dev->dev.parent, dma_addr))) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
789
*buf_dma_addr = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
247
iq->dma_addr = iq->real_dma_addr + CN10K_CPT_INST_GRP_QLEN_BYTES;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
251
iq->dma_addr = PTR_ALIGN(iq->dma_addr, OTX2_ALIGN);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
281
otx2_write64(pf, CN10K_CPT_LF_Q_BASE, pf->ipsec.iq.dma_addr);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
871
u64 dma_addr, *iova = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
898
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
899
if (dma_mapping_error(pfvf->dev, dma_addr))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
904
*iova++ = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
905
*cpt_iova++ = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
908
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.h
73
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2154
otx2_dma_unmap_page(pfvf, sg->dma_addr[seg],
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1142
u64 dma_addr, int len, int *offset, u16 flags);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2765
u64 dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2768
dma_addr = otx2_dma_map_page(pf, virt_to_page(xdpf->data),
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2771
if (dma_mapping_error(pf->dev, dma_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2774
err = otx2_xdp_sq_append_pkt(pf, xdpf, dma_addr, xdpf->len,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2777
otx2_dma_unmap_page(pf, dma_addr, xdpf->len, DMA_TO_DEVICE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
120
otx2_dma_unmap_page(pfvf, sg->dma_addr[0], sg->size[0], DMA_TO_DEVICE);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1418
u64 dma_addr, int len, int *offset, u16 flags)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1429
*iova = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1432
sq->sg[sq->head].dma_addr[0] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
666
u64 dma_addr, *iova = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
688
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
689
if (dma_mapping_error(pfvf->dev, dma_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
694
*iova++ = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
697
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
852
u64 dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
864
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
865
if (dma_mapping_error(pfvf->dev, dma_addr))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
869
sg->dma_addr[sg->num_segs] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
888
return sg->dma_addr[0] + (seg_addr - (u64)skb->data);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
894
return sg->dma_addr[seg] + offset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
923
*iova++ = list->dma_addr[seg];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
966
list.dma_addr[0] =
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
979
list.dma_addr[list.num_segs] =
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
173
static inline u64 otx2_iova_to_phys(void *iommu_domain, dma_addr_t dma_addr)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
177
return iommu_iova_to_phys(iommu_domain, dma_addr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
178
return dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
83
u64 dma_addr[OTX2_MAX_FRAGS_IN_SQE];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
240
dma_addr_t dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
242
dma_addr = xsk_buff_raw_get_dma(pool, xdp_desc[i].addr);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
243
otx2_xsk_sq_append_pkt(pfvf, dma_addr, xdp_desc[i].len, queue);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1305
dma_addr_t dma_addr;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1323
dma_addr = dma_map_single(eth->dma_dev,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1327
if (unlikely(dma_mapping_error(eth->dma_dev, dma_addr)))
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1334
txd->txd1 = dma_addr + i * MTK_QDMA_PAGE_SIZE;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1341
txd->txd3 |= TX_DMA_PREP_ADDR64(dma_addr + i * MTK_QDMA_PAGE_SIZE);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1911
static void *mtk_page_pool_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1920
*dma_addr = page_pool_get_dma_addr(page) + MTK_PP_HEADROOM;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2190
dma_addr_t dma_addr = DMA_MAPPING_ERROR;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2249
&dma_addr,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2297
dma_addr = dma_map_single(eth->dma_dev,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2301
dma_addr))) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2369
rxd->rxd1 = (unsigned int)dma_addr;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2372
if (unlikely(dma_addr == DMA_MAPPING_ERROR))
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2376
addr64 = RX_DMA_PREP_ADDR64(dma_addr);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2819
dma_addr_t dma_addr;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2825
&dma_addr, GFP_KERNEL);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2837
dma_addr = dma_map_single(eth->dma_dev,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2841
dma_addr))) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2846
rxd->rxd1 = (unsigned int)dma_addr;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2855
rxd->rxd2 |= RX_DMA_PREP_ADDR64(dma_addr);
drivers/net/ethernet/mediatek/mtk_star_emac.c
1115
desc_data.dma_addr = mtk_star_dma_map_tx(priv, skb);
drivers/net/ethernet/mediatek/mtk_star_emac.c
1116
if (dma_mapping_error(dev, desc_data.dma_addr))
drivers/net/ethernet/mediatek/mtk_star_emac.c
1326
desc_data.dma_addr = new_dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
1631
&priv->dma_addr,
drivers/net/ethernet/mediatek/mtk_star_emac.c
227
dma_addr_t dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
262
dma_addr_t dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
324
desc_data->dma_addr = ring->dma_addrs[ring->tail];
drivers/net/ethernet/mediatek/mtk_star_emac.c
350
ring->dma_addrs[ring->head] = desc_data->dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
356
WRITE_ONCE(desc->data_ptr, desc_data->dma_addr);
drivers/net/ethernet/mediatek/mtk_star_emac.c
411
dma_unmap_single(dev, desc_data->dma_addr,
drivers/net/ethernet/mediatek/mtk_star_emac.c
428
return dma_unmap_single(dev, desc_data->dma_addr,
drivers/net/ethernet/mediatek/mtk_star_emac.c
516
val = (unsigned int)priv->dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
700
dma_addr_t dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
708
dma_addr = mtk_star_dma_map_rx(priv, skb);
drivers/net/ethernet/mediatek/mtk_star_emac.c
709
if (dma_mapping_error(dev, dma_addr)) {
drivers/net/ethernet/mediatek/mtk_star_emac.c
715
desc->data_ptr = dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
719
ring->dma_addrs[i] = dma_addr;
drivers/net/ethernet/mediatek/mtk_star_emac.c
737
desc_data.dma_addr = ring->dma_addrs[i];
drivers/net/ethernet/mellanox/mlx4/icm.c
118
&buf->dma_addr, gfp_mask);
drivers/net/ethernet/mellanox/mlx4/icm.c
124
buf->dma_addr);
drivers/net/ethernet/mellanox/mlx4/icm.c
336
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx4/icm.c
341
dma_addr = chunk->buf[i].dma_addr;
drivers/net/ethernet/mellanox/mlx4/icm.c
347
dma_addr = sg_dma_address(&chunk->sg[i]);
drivers/net/ethernet/mellanox/mlx4/icm.c
360
*dma_handle = dma_addr + dma_offset;
drivers/net/ethernet/mellanox/mlx4/icm.c
76
chunk->buf[i].dma_addr);
drivers/net/ethernet/mellanox/mlx4/icm.h
128
return iter->chunk->buf[iter->page_idx].dma_addr;
drivers/net/ethernet/mellanox/mlx4/icm.h
53
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
101
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
103
xdptxd->dma_addr = dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
115
(union mlx5e_xdp_info) { .frame.dma_addr = dma_addr });
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
125
dma_addr = page_pool_get_dma_addr(page) + (xdpf->data - (void *)xdpf);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
126
dma_sync_single_for_device(sq->pdev, dma_addr, xdptxd->len, DMA_BIDIRECTIONAL);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
145
xdptxd->dma_addr = dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
477
tmp.dma_addr = xdptxdf->dma_arr ? xdptxdf->dma_arr[0] :
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
541
dma_addr_t dma_addr = xdptxd->dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
603
dma_addr += inline_hdr_sz;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
609
dseg->addr = cpu_to_be64(dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
670
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
675
dma_addr = xdpi.frame.dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
677
dma_unmap_single(sq->pdev, dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
68
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
688
dma_addr = xdpi.frame.dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
690
dma_unmap_single(sq->pdev, dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
876
xdptxd->dma_addr = dma_map_single(sq->pdev, xdptxd->data,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
879
if (unlikely(dma_mapping_error(sq->pdev, xdptxd->dma_addr)))
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
909
dma_unmap_single(sq->pdev, xdptxd->dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
926
(union mlx5e_xdp_info) { .frame.dma_addr = xdptxd->dma_addr });
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
933
{ .frame.dma_addr = dma_arr[j] });
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
96
dma_addr = dma_map_single(sq->pdev, xdptxd->data, xdptxd->len,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
98
if (dma_mapping_error(sq->pdev, dma_addr)) {
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
220
dseg->addr = cpu_to_be64(xdptxd->dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
88
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
104
xsk_buff_raw_dma_sync_for_device(pool, xdptxd.dma_addr, xdptxd.len);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
99
xdptxd.dma_addr = xsk_buff_raw_get_dma(pool, desc.addr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.h
176
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
524
aso->dma_addr = dma_map_single(pdev, aso->ctx, sizeof(aso->ctx),
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
526
err = dma_mapping_error(pdev, aso->dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
544
dma_unmap_single(pdev, aso->dma_addr, sizeof(aso->ctx),
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
562
dma_unmap_single(pdev, aso->dma_addr, sizeof(aso->ctx),
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
611
cpu_to_be32(lower_32_bits(aso->dma_addr) | ASO_CTRL_READ_EN);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_offload.c
612
ctrl->va_h = cpu_to_be32(upper_32_bits(aso->dma_addr));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
27
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
271
buf->dma_addr = dma_map_single(pdev, &buf->progress,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
273
if (unlikely(dma_mapping_error(pdev, buf->dma_addr))) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
304
psv->va = cpu_to_be64(buf->dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
319
dma_unmap_single(pdev, buf->dma_addr, PROGRESS_PARAMS_PADDED_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
448
dma_sync_single_for_cpu(dev, buf->dma_addr, PROGRESS_PARAMS_PADDED_SIZE,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
467
dma_unmap_single(dev, buf->dma_addr, PROGRESS_PARAMS_PADDED_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
720
dma_addr_t dma_addr = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
739
dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
741
if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
744
dseg->addr = cpu_to_be64(dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
747
mlx5e_dma_push_netmem(sq, skb_frag_netmem(frag), dma_addr, fsz);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
1290
aso_ctrl->va_l = cpu_to_be32(umr->dma_addr | ASO_CTRL_READ_EN);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
1291
aso_ctrl->va_h = cpu_to_be32((u64)umr->dma_addr >> 32);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
165
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
175
dma_addr = dma_map_single(dma_device, umr->ctx, sizeof(umr->ctx), DMA_BIDIRECTIONAL);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
176
err = dma_mapping_error(dma_device, dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
188
umr->dma_addr = dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
195
dma_unmap_single(dma_device, dma_addr, sizeof(umr->ctx), DMA_BIDIRECTIONAL);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
206
dma_unmap_single(&mdev->pdev->dev, umr->dma_addr, sizeof(umr->ctx), DMA_BIDIRECTIONAL);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
92
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
185
dma_addr_t dma_addr = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
190
dma_addr = dma_map_single(sq->pdev, skb_data, headlen,
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
192
if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
195
dseg->addr = cpu_to_be64(dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
199
mlx5e_dma_push_single(sq, dma_addr, headlen);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
208
dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
210
if (unlikely(dma_mapping_error(sq->pdev, dma_addr)))
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
213
dseg->addr = cpu_to_be64(dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
217
mlx5e_dma_push_netmem(sq, skb_frag_netmem(frag), dma_addr, fsz);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
532
dseg->addr = cpu_to_be64(txd->dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
581
txd.dma_addr = dma_map_single(sq->pdev, txd.data, txd.len, DMA_TO_DEVICE);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
582
if (unlikely(dma_mapping_error(sq->pdev, txd.dma_addr)))
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
594
mlx5e_dma_push_single(sq, txd.dma_addr, txd.len);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
119
data->addr = cpu_to_be64(buf->sg[0].dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
159
data->addr = cpu_to_be64(buf->sg[sgi].dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
58
buf->sg[0].dma_addr = dma_map_single(dma_device, buf->sg[0].data,
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
60
err = dma_mapping_error(dma_device, buf->sg[0].dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
70
buf->sg[1].dma_addr = dma_map_single(dma_device, buf->sg[1].data,
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
72
err = dma_mapping_error(dma_device, buf->sg[1].dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
75
dma_unmap_single(dma_device, buf->sg[0].dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
91
dma_unmap_single(dma_device, buf->sg[1].dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
95
dma_unmap_single(dma_device, buf->sg[0].dma_addr,
drivers/net/ethernet/mellanox/mlx5/core/fpga/sdk.h
71
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1163
dma_addr_t dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1170
dma_addr = dma_map_single(dma_device, buf, size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1172
err = dma_mapping_error(dma_device, dma_addr);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1182
dma_unmap_single(dma_device, dma_addr, size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1188
mr->dma_addr = dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1198
dma_unmap_single(mlx5_core_dma_dev(mdev), mr->dma_addr, mr->size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
592
send_info->write.addr = (uintptr_t)send_ring->mr->dma_addr + buff_offset;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
607
send_info->read.addr = (uintptr_t)send_ring->sync_mr->dma_addr;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1448
dma_addr_t dma_addr;
drivers/net/ethernet/micrel/ksz884x.c
1306
dma_addr_t dma_addr;
drivers/net/ethernet/micrel/ksz884x.c
4024
&adapter->desc_pool.dma_addr, GFP_KERNEL);
drivers/net/ethernet/micrel/ksz884x.c
4035
adapter->desc_pool.phys = adapter->desc_pool.dma_addr + offset;
drivers/net/ethernet/micrel/ksz884x.c
4180
adapter->desc_pool.dma_addr);
drivers/net/ethernet/microchip/lan743x_main.h
1106
#define DMA_ADDR_HIGH32(dma_addr) ((u32)(((dma_addr) >> 32) & 0xFFFFFFFF))
drivers/net/ethernet/microchip/lan743x_main.h
1108
#define DMA_ADDR_HIGH32(dma_addr) ((u32)(0))
drivers/net/ethernet/microchip/lan743x_main.h
1110
#define DMA_ADDR_LOW32(dma_addr) ((u32)((dma_addr) & 0xFFFFFFFF))
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
31
*dataptr = lan966x->tx.dcbs_buf[dcb].dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
359
dcb_buf->dma_addr,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
368
dcb_buf->dma_addr,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
41
*dataptr = lan966x->tx.dcbs_buf[dcb].dma_addr + XDP_PACKET_HEADROOM;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
605
dma_addr_t dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
638
dma_addr = dma_map_single(lan966x->dev,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
642
if (dma_mapping_error(lan966x->dev, dma_addr)) {
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
657
dma_addr = page_pool_get_dma_addr(page);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
659
dma_addr + XDP_PACKET_HEADROOM,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
670
next_dcb_buf->dma_addr = dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
703
dma_addr_t dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
737
dma_addr = dma_map_single(lan966x->dev, skb->data, skb->len,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
739
if (dma_mapping_error(lan966x->dev, dma_addr)) {
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
751
next_dcb_buf->dma_addr = dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_main.h
220
dma_addr_t dma_addr;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
20
*dataptr = FDMA_PRIV(fdma)->tx.dbs[dcb].dma_addr;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
325
db_buf->dma_addr = dma_map_single(sparx5->dev,
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
329
if (dma_mapping_error(sparx5->dev, db_buf->dma_addr))
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
76
db->dma_addr,
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
178
dma_addr_t dma_addr;
drivers/net/ethernet/mscc/ocelot_fdma.c
111
dma_addr_t dma_addr,
drivers/net/ethernet/mscc/ocelot_fdma.c
114
u32 offset = dma_addr & 0x3;
drivers/net/ethernet/mscc/ocelot_fdma.c
117
dcb->datap = ALIGN_DOWN(dma_addr, 4);
drivers/net/ethernet/mscc/ocelot_fdma.c
141
rxb->dma_addr = mapping;
drivers/net/ethernet/mscc/ocelot_fdma.c
152
dma_addr_t dma_addr;
drivers/net/ethernet/mscc/ocelot_fdma.c
172
dma_addr = rxb->dma_addr + rxb->page_offset;
drivers/net/ethernet/mscc/ocelot_fdma.c
173
ocelot_fdma_dcb_set_data(dcb, dma_addr, OCELOT_FDMA_RXB_SIZE);
drivers/net/ethernet/mscc/ocelot_fdma.c
198
dma_unmap_addr_set(tx_buf, dma_addr, mapping);
drivers/net/ethernet/mscc/ocelot_fdma.c
309
dma_sync_single_range_for_device(ocelot->dev, old_rxb->dma_addr,
drivers/net/ethernet/mscc/ocelot_fdma.c
334
dma_sync_single_range_for_cpu(ocelot->dev, rxb->dma_addr,
drivers/net/ethernet/mscc/ocelot_fdma.c
343
dma_unmap_page(ocelot->dev, rxb->dma_addr, PAGE_SIZE,
drivers/net/ethernet/mscc/ocelot_fdma.c
508
dma_unmap_single(ocelot->dev, dma_unmap_addr(buf, dma_addr),
drivers/net/ethernet/mscc/ocelot_fdma.c
711
dma_unmap_page(ocelot->dev, rxb->dma_addr, PAGE_SIZE,
drivers/net/ethernet/mscc/ocelot_fdma.c
735
dma_unmap_single(ocelot->dev, dma_unmap_addr(txb, dma_addr),
drivers/net/ethernet/mscc/ocelot_fdma.h
110
dma_addr_t dma_addr;
drivers/net/ethernet/mscc/ocelot_fdma.h
78
DEFINE_DMA_UNMAP_ADDR(dma_addr);
drivers/net/ethernet/netronome/nfp/bpf/offload.c
500
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
525
dma_addr = dma_map_single(nn->dp.dev, img,
drivers/net/ethernet/netronome/nfp/bpf/offload.c
528
if (dma_mapping_error(nn->dp.dev, dma_addr)) {
drivers/net/ethernet/netronome/nfp/bpf/offload.c
534
nn_writeq(nn, NFP_NET_CFG_BPF_ADDR, dma_addr);
drivers/net/ethernet/netronome/nfp/bpf/offload.c
542
dma_unmap_single(nn->dp.dev, dma_addr, nfp_prog->prog_len * sizeof(u64),
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1045
rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1058
rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1087
nfp_net_dma_unmap_rx(dp, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1207
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1240
dma_addr = dma_map_single(dp->dev, skb->data, skb_headlen(skb),
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1242
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1250
txbuf->dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1259
nfp_desc_set_dma_addr_40b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1352
nfp_net_dma_sync_cpu_rx(dp, rxbuf->dma_addr + meta_off, data_len);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1372
nfp_net_dma_unmap_rx(dp, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
266
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
302
dma_addr = dma_map_single(dp->dev, skb->data, skb_headlen(skb),
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
304
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
312
txbuf->dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
321
nfp_desc_set_dma_addr_40b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
350
dma_addr = skb_frag_dma_map(dp->dev, frag, 0,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
352
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
357
tx_ring->txbufs[wr_idx].dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
362
nfp_desc_set_dma_addr_40b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
390
dma_unmap_page(dp->dev, tx_ring->txbufs[wr_idx].dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
393
tx_ring->txbufs[wr_idx].dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
399
dma_unmap_single(dp->dev, tx_ring->txbufs[wr_idx].dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
402
tx_ring->txbufs[wr_idx].dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
460
dma_unmap_single(dp->dev, tx_buf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
468
dma_unmap_page(dp->dev, tx_buf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
476
tx_buf->dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
552
nfp_nfd3_napi_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
569
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
570
if (dma_mapping_error(dp->dev, *dma_addr)) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
589
void *frag, dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
595
nfp_net_dma_sync_dev_rx(dp, dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
599
rx_ring->rxbufs[wr_idx].dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
609
dma_addr + dp->rx_dma_off);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
636
rx_ring->rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
830
nfp_nfd3_rx_give_one(dp, rx_ring, rxbuf->frag, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
868
nfp_nfd3_rx_give_one(dp, rx_ring, txbuf->frag, txbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
871
txbuf->dma_addr = rxbuf->dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
876
dma_sync_single_for_device(dp->dev, rxbuf->dma_addr + dma_off,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
883
nfp_desc_set_dma_addr_40b(txd, rxbuf->dma_addr + dma_off);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
989
nfp_net_dma_sync_cpu_rx(dp, rxbuf->dma_addr + meta_off,
drivers/net/ethernet/netronome/nfp/nfd3/nfd3.h
74
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
171
nfp_net_dma_unmap_rx(dp, tx_ring->txbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
187
txbufs[i].frag = nfp_net_rx_alloc_one(dp, &txbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
227
if (tx_ring->txbufs[i].dma_addr)
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
229
&tx_ring->txbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
60
dma_unmap_single(dp->dev, tx_buf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
65
dma_unmap_page(dp->dev, tx_buf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
73
tx_buf->dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
271
txbuf->dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
29
xsk_buff_raw_dma_sync_for_device(pool, xrxbuf->dma_addr + pkt_off,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
43
nfp_desc_set_dma_addr_40b(txd, xrxbuf->dma_addr + pkt_off);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1111
nfp_net_dma_sync_cpu_rx(dp, rxbuf->dma_addr + meta_off,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1164
rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1177
rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1206
nfp_net_dma_unmap_rx(dp, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1318
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1369
dma_addr = dma_map_single(dp->dev, skb->data, dma_len, DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1370
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1376
txbuf->dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1386
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1390
dma_addr += tmp_dlen + 1;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1397
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1401
dma_addr += dlen_type + 1;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1431
dma_unmap_single(dp->dev, txbuf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1514
nfp_net_dma_sync_cpu_rx(dp, rxbuf->dma_addr + meta_off, data_len);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1534
nfp_net_dma_unmap_rx(dp, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
267
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
313
dma_addr = dma_map_single(dp->dev, skb->data, dma_len, DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
314
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
320
txbuf->dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
338
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
348
dma_addr += tmp_dlen + 1;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
363
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
366
dma_addr += dlen_type + 1;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
374
dma_addr = skb_frag_dma_map(dp->dev, frag, 0, dma_len,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
376
if (dma_mapping_error(dp->dev, dma_addr))
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
379
txbuf->dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
442
dma_unmap_single(dp->dev, txbuf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
449
dma_unmap_page(dp->dev, txbuf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
514
dma_unmap_single(dev, txbuf->dma_addr, size, DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
524
dma_unmap_page(dev, txbuf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
572
nfp_nfdk_napi_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
589
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
590
if (dma_mapping_error(dp->dev, *dma_addr)) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
609
void *frag, dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
615
nfp_net_dma_sync_dev_rx(dp, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
619
rx_ring->rxbufs[wr_idx].dma_addr = dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
625
dma_addr + dp->rx_dma_off);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
649
rx_ring->rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
830
nfp_nfdk_rx_give_one(dp, rx_ring, rxbuf->frag, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
874
txbuf[1].dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
909
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
956
txbuf[1].dma_addr = rxbuf->dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
959
dma_sync_single_for_device(dp->dev, rxbuf->dma_addr + dma_off,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
965
dma_addr = rxbuf->dma_addr + dma_off;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
980
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
984
dma_addr += tmp_dlen + 1;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
991
nfp_desc_set_dma_addr_48b(txd, dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
995
dma_addr += dlen_type + 1;
drivers/net/ethernet/netronome/nfp/nfdk/nfdk.h
99
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
37
dma_unmap_single(dev, txbuf->dma_addr, size, DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
45
dma_unmap_page(dev, txbuf->dma_addr,
drivers/net/ethernet/netronome/nfp/nfp_net.h
121
#define nfp_desc_set_dma_addr_40b(desc, dma_addr) \
drivers/net/ethernet/netronome/nfp/nfp_net.h
124
dma_addr_t __addr = (dma_addr); \
drivers/net/ethernet/netronome/nfp/nfp_net.h
130
#define nfp_desc_set_dma_addr_48b(desc, dma_addr) \
drivers/net/ethernet/netronome/nfp/nfp_net.h
133
dma_addr_t __addr = (dma_addr); \
drivers/net/ethernet/netronome/nfp/nfp_net.h
287
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfp_net.h
296
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
51
if (rx_ring->rxbufs[i].dma_addr)
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
53
&rx_ring->rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
55
if (rx_ring->xsk_rxbufs[i].dma_addr)
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
57
&rx_ring->xsk_rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
148
nfp_net_dma_unmap_rx(dp, rx_ring->rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
150
rx_ring->rxbufs[i].dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
17
void *nfp_net_rx_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
173
rxbufs[i].frag = nfp_net_rx_alloc_one(dp, &rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
34
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
35
if (dma_mapping_error(dp->dev, *dma_addr)) {
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
103
void *nfp_net_rx_alloc_one(struct nfp_net_dp *dp, dma_addr_t *dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
17
nfp_net_dma_sync_dev_rx(const struct nfp_net_dp *dp, dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
19
dma_sync_single_for_device(dp->dev, dma_addr,
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
25
dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
27
dma_unmap_single_attrs(dp->dev, dma_addr,
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
33
dma_addr_t dma_addr,
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
36
dma_sync_single_for_cpu(dp->dev, dma_addr - NFP_NET_RX_BUF_HEADROOM,
drivers/net/ethernet/netronome/nfp/nfp_net_xsk.c
28
rx_ring->xsk_rxbufs[idx].dma_addr =
drivers/net/ethernet/netronome/nfp/nfp_net_xsk.c
34
rxbuf->dma_addr = 0;
drivers/net/ethernet/netronome/nfp/nfp_net_xsk.c
78
rx_ring->xsk_rxbufs[wr_idx].dma_addr);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
502
dma_addr_t dma_addr;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
550
chunks[i].dma_addr = addr;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
576
desc->descs[i].addr = cpu_to_le64(chunks[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
609
dma_unmap_single(dev, chunks[i].dma_addr, chunks[i].len,
drivers/net/ethernet/pensando/ionic/ionic_dev.h
204
dma_addr_t dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_if.h
2626
__le64 dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1083
dma_addr_t dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1085
dma_addr = dma_map_single(dev, data, len, DMA_TO_DEVICE);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1086
if (unlikely(dma_mapping_error(dev, dma_addr))) {
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1092
return dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1100
dma_addr_t dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1102
dma_addr = skb_frag_dma_map(dev, frag, offset, len, DMA_TO_DEVICE);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1103
if (unlikely(dma_mapping_error(dev, dma_addr))) {
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1109
return dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1117
dma_addr_t dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1122
dma_addr = ionic_tx_map_single(q, skb->data, skb_headlen(skb));
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1123
if (dma_addr == DMA_MAPPING_ERROR)
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1125
buf_info->dma_addr = dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1132
dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag));
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1133
if (dma_addr == DMA_MAPPING_ERROR)
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1135
buf_info->dma_addr = dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1149
dma_unmap_page(dev, buf_info->dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1152
dma_unmap_single(dev, desc_info->bufs[0].dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1167
dma_unmap_single(dev, buf_info->dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1171
dma_unmap_page(dev, buf_info->dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1505
frag_addr = buf_info->dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1567
buf_info->dma_addr);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1606
buf_info->dma_addr);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1633
elem->addr = cpu_to_le64(buf_info->dma_addr);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
313
dma_addr_t dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
322
dma_addr = page_pool_get_dma_addr(page) +
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
324
dma_sync_single_for_device(q->dev, dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
327
dma_addr = ionic_tx_map_single(q, frame->data, len);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
328
if (dma_addr == DMA_MAPPING_ERROR)
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
332
buf_info->dma_addr = dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
356
dma_addr = page_pool_get_dma_addr(pg) +
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
358
dma_sync_single_for_device(q->dev, dma_addr,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
362
dma_addr = ionic_tx_map_frag(q, frag, 0,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
364
if (dma_addr == DMA_MAPPING_ERROR) {
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
369
bi->dma_addr = dma_addr;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
373
elem->addr = cpu_to_le64(bi->dma_addr);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
382
0, (desc_info->nbufs - 1), buf_info->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1116
rfbuf->dma_addr, rfbuf->length,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1118
rfbuf->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1189
if (tpbuf->dma_addr) {
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1191
tpbuf->dma_addr, tpbuf->length,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1193
tpbuf->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1349
tpbuf->dma_addr = dma_map_page(adpt->netdev->dev.parent,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1355
tpbuf->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1359
TPD_BUFFER_ADDR_L_SET(tpd, lower_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1360
TPD_BUFFER_ADDR_H_SET(tpd, upper_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1369
tpbuf->dma_addr = dma_map_page(adpt->netdev->dev.parent,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1376
tpbuf->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1380
TPD_BUFFER_ADDR_L_SET(tpd, lower_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1381
TPD_BUFFER_ADDR_H_SET(tpd, upper_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1392
tpbuf->dma_addr = skb_frag_dma_map(adpt->netdev->dev.parent,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1396
tpbuf->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1400
TPD_BUFFER_ADDR_L_SET(tpd, lower_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1401
TPD_BUFFER_ADDR_H_SET(tpd, upper_32_bits(tpbuf->dma_addr));
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1424
dma_unmap_page(adpt->netdev->dev.parent, tpbuf->dma_addr,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1426
tpbuf->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
305
writel(upper_32_bits(adpt->tx_q.tpd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
308
writel(lower_32_bits(adpt->tx_q.tpd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
315
writel(upper_32_bits(adpt->rx_q.rfd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
318
writel(lower_32_bits(adpt->rx_q.rfd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
320
writel(lower_32_bits(adpt->rx_q.rrd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
598
if (tpbuf->dma_addr) {
drivers/net/ethernet/qualcomm/emac/emac-mac.c
600
tpbuf->dma_addr, tpbuf->length,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
602
tpbuf->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
635
if (rfbuf->dma_addr) {
drivers/net/ethernet/qualcomm/emac/emac-mac.c
636
dma_unmap_single(dev, rfbuf->dma_addr, rfbuf->length,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
638
rfbuf->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
669
tx_q->tpd.dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
687
tx_q->tpd.dma_addr = ring_header->dma_addr + ring_header->used;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
707
rx_q->rfd.dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
711
rx_q->rrd.dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
731
rx_q->rrd.dma_addr = ring_header->dma_addr + ring_header->used;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
735
rx_q->rfd.dma_addr = ring_header->dma_addr + ring_header->used;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
772
&ring_header->dma_addr,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
777
ring_header->used = ALIGN(ring_header->dma_addr, 8) -
drivers/net/ethernet/qualcomm/emac/emac-mac.c
778
ring_header->dma_addr;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
798
ring_header->v_addr, ring_header->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
801
ring_header->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
818
ring_header->v_addr, ring_header->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
821
ring_header->dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
834
adpt->tx_q.tpd.tpbuff[i].dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
841
adpt->rx_q.rfd.rfbuff[i].dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
875
while (!next_rxbuf->dma_addr) {
drivers/net/ethernet/qualcomm/emac/emac-mac.c
883
curr_rxbuf->dma_addr =
drivers/net/ethernet/qualcomm/emac/emac-mac.c
888
curr_rxbuf->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
896
emac_mac_rx_rfd_create(adpt, rx_q, curr_rxbuf->dma_addr);
drivers/net/ethernet/qualcomm/emac/emac-mac.h
135
dma_addr_t dma_addr; /* dma address */
drivers/net/ethernet/qualcomm/emac/emac-mac.h
146
dma_addr_t dma_addr; /* dma address */
drivers/net/ethernet/qualcomm/emac/emac-mac.h
153
dma_addr_t dma_addr; /* dma address */
drivers/net/ethernet/qualcomm/emac/emac-mac.h
164
dma_addr_t dma_addr; /* physical address */
drivers/net/ethernet/qualcomm/emac/emac-mac.h
197
dma_addr_t dma_addr; /* dma address */
drivers/net/ethernet/renesas/ravb_main.c
2125
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/ravb_main.c
2170
dma_addr = dma_map_single(ndev->dev.parent, buffer, len,
drivers/net/ethernet/renesas/ravb_main.c
2172
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/ravb_main.c
2177
desc->dptr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/ravb_main.c
2181
dma_addr = dma_map_single(ndev->dev.parent, buffer, len,
drivers/net/ethernet/renesas/ravb_main.c
2183
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/ravb_main.c
2190
dma_addr = dma_map_single(ndev->dev.parent, skb->data, skb->len,
drivers/net/ethernet/renesas/ravb_main.c
2192
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/ravb_main.c
2196
desc->dptr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/ravb_main.c
2205
dma_unmap_single(ndev->dev.parent, dma_addr,
drivers/net/ethernet/renesas/ravb_main.c
301
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/ravb_main.c
316
dma_addr = page_pool_get_dma_addr(rx_buff->page) + rx_buff->offset;
drivers/net/ethernet/renesas/ravb_main.c
317
dma_sync_single_for_device(ndev->dev.parent, dma_addr,
drivers/net/ethernet/renesas/ravb_main.c
319
rx_desc->dptr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/rswitch_main.c
1699
dma_addr_t dma_addr, u16 len, u8 die_dt)
drivers/net/ethernet/renesas/rswitch_main.c
1701
rswitch_desc_set_dptr(&desc->desc, dma_addr);
drivers/net/ethernet/renesas/rswitch_main.c
1742
dma_addr_t dma_addr, dma_addr_orig;
drivers/net/ethernet/renesas/rswitch_main.c
1772
dma_addr = dma_addr_orig + i * RSWITCH_DESC_BUF_SIZE;
drivers/net/ethernet/renesas/rswitch_main.c
1774
if (!rswitch_ext_desc_set(rdev, skb, desc, dma_addr, len, die_dt))
drivers/net/ethernet/renesas/rswitch_main.c
406
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rswitch_main.c
412
dma_addr = dma_map_single(ndev->dev.parent,
drivers/net/ethernet/renesas/rswitch_main.c
416
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/rswitch_main.c
420
rswitch_desc_set_dptr(&desc->desc, dma_addr);
drivers/net/ethernet/renesas/rswitch_main.c
441
dma_addr = rswitch_desc_get_dptr(&desc->desc);
drivers/net/ethernet/renesas/rswitch_main.c
442
dma_unmap_single(ndev->dev.parent, dma_addr,
drivers/net/ethernet/renesas/rswitch_main.c
473
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rswitch_main.c
479
dma_addr = dma_map_single(ndev->dev.parent,
drivers/net/ethernet/renesas/rswitch_main.c
483
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/rswitch_main.c
487
rswitch_desc_set_dptr(&desc->desc, dma_addr);
drivers/net/ethernet/renesas/rswitch_main.c
503
dma_addr = rswitch_desc_get_dptr(&desc->desc);
drivers/net/ethernet/renesas/rswitch_main.c
504
dma_unmap_single(ndev->dev.parent, dma_addr,
drivers/net/ethernet/renesas/rswitch_main.c
755
dma_addr_t dma_addr = rswitch_desc_get_dptr(&desc->desc);
drivers/net/ethernet/renesas/rswitch_main.c
760
dma_unmap_single(ndev->dev.parent, dma_addr, RSWITCH_MAP_BUF_SIZE,
drivers/net/ethernet/renesas/rtsn.c
1046
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rtsn.c
1068
dma_addr = dma_map_single(ndev->dev.parent, skb->data, skb->len,
drivers/net/ethernet/renesas/rtsn.c
1070
if (dma_mapping_error(ndev->dev.parent, dma_addr)) {
drivers/net/ethernet/renesas/rtsn.c
1078
desc->dptr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/rtsn.c
169
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rtsn.c
185
dma_addr = le32_to_cpu(desc->dptr);
drivers/net/ethernet/renesas/rtsn.c
186
dma_unmap_single(ndev->dev.parent, dma_addr, PKT_BUF_SZ,
drivers/net/ethernet/renesas/rtsn.c
221
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rtsn.c
231
dma_addr = dma_map_single(ndev->dev.parent, skb->data,
drivers/net/ethernet/renesas/rtsn.c
234
if (dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/rtsn.c
236
desc->dptr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/rtsn.c
390
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/rtsn.c
413
dma_addr = dma_map_single(ndev->dev.parent,
drivers/net/ethernet/renesas/rtsn.c
416
if (!dma_mapping_error(ndev->dev.parent, dma_addr))
drivers/net/ethernet/renesas/rtsn.c
418
rx_desc->dptr = cpu_to_le32((u32)dma_addr);
drivers/net/ethernet/renesas/sh_eth.c
1328
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/sh_eth.c
1349
dma_addr = dma_map_single(&mdp->pdev->dev, skb->data, buf_len,
drivers/net/ethernet/renesas/sh_eth.c
1351
if (dma_mapping_error(&mdp->pdev->dev, dma_addr)) {
drivers/net/ethernet/renesas/sh_eth.c
1360
rxdesc->addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/sh_eth.c
1593
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/sh_eth.c
1642
dma_addr = le32_to_cpu(rxdesc->addr);
drivers/net/ethernet/renesas/sh_eth.c
1645
phys_to_virt(ALIGN(dma_addr, 4)),
drivers/net/ethernet/renesas/sh_eth.c
1650
dma_unmap_single(&mdp->pdev->dev, dma_addr,
drivers/net/ethernet/renesas/sh_eth.c
1680
dma_addr = dma_map_single(&mdp->pdev->dev, skb->data,
drivers/net/ethernet/renesas/sh_eth.c
1682
if (dma_mapping_error(&mdp->pdev->dev, dma_addr)) {
drivers/net/ethernet/renesas/sh_eth.c
1689
rxdesc->addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/renesas/sh_eth.c
2497
dma_addr_t dma_addr;
drivers/net/ethernet/renesas/sh_eth.c
2521
dma_addr = dma_map_single(&mdp->pdev->dev, skb->data, skb->len,
drivers/net/ethernet/renesas/sh_eth.c
2523
if (dma_mapping_error(&mdp->pdev->dev, dma_addr)) {
drivers/net/ethernet/renesas/sh_eth.c
2527
txdesc->addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
48
dma_addr_t dma_addr;
drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
80
dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
81
writel(lower_32_bits(dma_addr),
drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
84
dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES);
drivers/net/ethernet/samsung/sxgbe/sxgbe_dma.c
85
writel(lower_32_bits(dma_addr),
drivers/net/ethernet/seeq/sgiseeq.c
203
dma_addr_t dma_addr;
drivers/net/ethernet/seeq/sgiseeq.c
209
dma_addr = dma_map_single(dev->dev.parent,
drivers/net/ethernet/seeq/sgiseeq.c
213
sp->rx_desc[i].rdma.pbuf = dma_addr;
drivers/net/ethernet/sfc/ef10.c
1919
MCDI_SET_QWORD(inbuf, MAC_STATS_IN_DMA_ADDR, stats_buf.dma_addr);
drivers/net/ethernet/sfc/ef10.c
2066
_efx_writed(efx, cpu_to_le32((u64)nic_data->mcdi_buf.dma_addr >> 32),
drivers/net/ethernet/sfc/ef10.c
2068
_efx_writed(efx, cpu_to_le32((u32)nic_data->mcdi_buf.dma_addr),
drivers/net/ethernet/sfc/ef10.c
2433
dma_addr_t dma_addr, unsigned int len)
drivers/net/ethernet/sfc/ef10.c
2440
dma_addr_t end = dma_addr + EFX_EF10_MAX_TX_DESCRIPTOR_LEN;
drivers/net/ethernet/sfc/ef10.c
2443
len = (end & (~(EFX_PAGE_SIZE - 1))) - dma_addr;
drivers/net/ethernet/sfc/ef10.c
2480
ESF_DZ_TX_KER_BUF_ADDR, buffer->dma_addr);
drivers/net/ethernet/sfc/ef10.c
2597
ESF_DZ_RX_KER_BUF_ADDR, rx_buf->dma_addr);
drivers/net/ethernet/sfc/ef100_nic.c
39
static u8 *ef100_mcdi_buf(struct efx_nic *efx, u8 bufid, dma_addr_t *dma_addr)
drivers/net/ethernet/sfc/ef100_nic.c
43
if (dma_addr)
drivers/net/ethernet/sfc/ef100_nic.c
44
*dma_addr = nic_data->mcdi_buf.dma_addr +
drivers/net/ethernet/sfc/ef100_nic.c
69
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/ef100_nic.c
70
u8 *pdu = ef100_mcdi_buf(efx, 0, &dma_addr);
drivers/net/ethernet/sfc/ef100_nic.c
82
_efx_writed(efx, cpu_to_le32((u64)dma_addr >> 32), efx_reg(efx, ER_GZ_MC_DB_LWRD));
drivers/net/ethernet/sfc/ef100_nic.c
83
_efx_writed(efx, cpu_to_le32((u32)dma_addr), efx_reg(efx, ER_GZ_MC_DB_HWRD));
drivers/net/ethernet/sfc/ef100_rx.c
205
EFX_POPULATE_QWORD_1(*rxd, ESF_GZ_RX_BUF_ADDR, rx_buf->dma_addr);
drivers/net/ethernet/sfc/ef100_tx.c
176
ESF_GZ_TX_SEND_ADDR, buffer->dma_addr);
drivers/net/ethernet/sfc/ef100_tx.c
328
ESF_GZ_TX_SEG_ADDR, buffer->dma_addr);
drivers/net/ethernet/sfc/falcon/falcon.c
1429
efx->stats_buffer.dma_addr);
drivers/net/ethernet/sfc/falcon/falcon.c
1726
(u64)efx->stats_buffer.dma_addr,
drivers/net/ethernet/sfc/falcon/falcon.c
2358
BUG_ON(efx->irq_status.dma_addr & 0x0f);
drivers/net/ethernet/sfc/falcon/falcon.c
2362
(u64)efx->irq_status.dma_addr,
drivers/net/ethernet/sfc/falcon/farch.c
1669
FRF_AZ_INT_ADR_KER, efx->irq_status.dma_addr);
drivers/net/ethernet/sfc/falcon/farch.c
172
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/farch.c
180
dma_addr = buffer->buf.dma_addr + (i * EF4_BUF_SIZE);
drivers/net/ethernet/sfc/falcon/farch.c
183
index, (unsigned long long)dma_addr);
drivers/net/ethernet/sfc/falcon/farch.c
186
FRF_AZ_BUF_ADR_FBUF, dma_addr >> 12,
drivers/net/ethernet/sfc/falcon/farch.c
232
BUG_ON(buffer->buf.dma_addr & (EF4_BUF_SIZE - 1));
drivers/net/ethernet/sfc/falcon/farch.c
242
(u64)buffer->buf.dma_addr, len,
drivers/net/ethernet/sfc/falcon/farch.c
258
(u64)buffer->buf.dma_addr, buffer->buf.len,
drivers/net/ethernet/sfc/falcon/farch.c
332
FSF_AZ_TX_KER_BUF_ADDR, buffer->dma_addr);
drivers/net/ethernet/sfc/falcon/farch.c
348
dma_addr_t dma_addr, unsigned int len)
drivers/net/ethernet/sfc/falcon/farch.c
351
unsigned int limit = (~dma_addr & (EF4_PAGE_SIZE - 1)) + 1;
drivers/net/ethernet/sfc/falcon/farch.c
355
if (EF4_WORKAROUND_5391(tx_queue->efx) && (dma_addr & 0xf))
drivers/net/ethernet/sfc/falcon/farch.c
356
len = min_t(unsigned int, len, 512 - (dma_addr & 0xf));
drivers/net/ethernet/sfc/falcon/farch.c
483
FSF_AZ_RX_KER_BUF_ADDR, rx_buf->dma_addr);
drivers/net/ethernet/sfc/falcon/net_driver.h
107
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/net_driver.h
1087
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/falcon/net_driver.h
148
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/net_driver.h
267
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/net_driver.h
289
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/nic.c
34
&buffer->dma_addr, gfp_flags);
drivers/net/ethernet/sfc/falcon/nic.c
45
buffer->addr, buffer->dma_addr);
drivers/net/ethernet/sfc/falcon/nic.h
389
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/falcon/rx.c
131
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/falcon/rx.c
159
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/falcon/rx.c
171
dma_addr =
drivers/net/ethernet/sfc/falcon/rx.c
176
dma_addr))) {
drivers/net/ethernet/sfc/falcon/rx.c
181
state->dma_addr = dma_addr;
drivers/net/ethernet/sfc/falcon/rx.c
184
dma_addr = state->dma_addr;
drivers/net/ethernet/sfc/falcon/rx.c
187
dma_addr += sizeof(struct ef4_rx_page_state);
drivers/net/ethernet/sfc/falcon/rx.c
193
rx_buf->dma_addr = dma_addr + efx->rx_ip_align;
drivers/net/ethernet/sfc/falcon/rx.c
200
dma_addr += efx->rx_page_buf_step;
drivers/net/ethernet/sfc/falcon/rx.c
221
state->dma_addr,
drivers/net/ethernet/sfc/falcon/rx.c
814
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/falcon/rx.c
88
dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len,
drivers/net/ethernet/sfc/falcon/tx.c
174
dma_addr_t dma_addr,
drivers/net/ethernet/sfc/falcon/tx.c
184
dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/falcon/tx.c
187
buffer->dma_addr = dma_addr;
drivers/net/ethernet/sfc/falcon/tx.c
190
dma_addr += dma_len;
drivers/net/ethernet/sfc/falcon/tx.c
204
dma_addr_t dma_addr, unmap_addr;
drivers/net/ethernet/sfc/falcon/tx.c
213
dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/sfc/falcon/tx.c
216
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/falcon/tx.c
218
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/falcon/tx.c
226
buffer = ef4_tx_map_chunk(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/falcon/tx.c
233
buffer->dma_offset = buffer->dma_addr - unmap_addr;
drivers/net/ethernet/sfc/falcon/tx.c
247
dma_addr = skb_frag_dma_map(dma_dev, fragment,
drivers/net/ethernet/sfc/falcon/tx.c
251
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/falcon/tx.c
253
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/falcon/tx.c
38
buffer->dma_addr = page_buf->dma_addr + offset;
drivers/net/ethernet/sfc/falcon/tx.c
50
dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset;
drivers/net/ethernet/sfc/falcon/tx.h
16
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/mcdi_functions.c
115
dma_addr = channel->eventq.dma_addr;
drivers/net/ethernet/sfc/mcdi_functions.c
117
MCDI_SET_ARRAY_QWORD(inbuf, INIT_EVQ_IN_DMA_ADDR, i, dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
118
dma_addr += EFX_BUF_SIZE;
drivers/net/ethernet/sfc/mcdi_functions.c
172
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/mcdi_functions.c
185
dma_addr = tx_queue->txd.dma_addr;
drivers/net/ethernet/sfc/mcdi_functions.c
188
tx_queue->queue, entries, (u64)dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
191
MCDI_SET_ARRAY_QWORD(inbuf, INIT_TXQ_IN_DMA_ADDR, i, dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
192
dma_addr += EFX_BUF_SIZE;
drivers/net/ethernet/sfc/mcdi_functions.c
285
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/mcdi_functions.c
309
dma_addr = rx_queue->rxd.dma_addr;
drivers/net/ethernet/sfc/mcdi_functions.c
312
efx_rx_queue_index(rx_queue), entries, (u64)dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
315
MCDI_SET_ARRAY_QWORD(inbuf, INIT_RXQ_IN_DMA_ADDR, i, dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
316
dma_addr += EFX_BUF_SIZE;
drivers/net/ethernet/sfc/mcdi_functions.c
80
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/mcdi_mon.c
151
hwmon->dma_buf.dma_addr);
drivers/net/ethernet/sfc/mcdi_port_common.c
1153
dma_addr_t dma_addr = efx->stats_buffer.dma_addr;
drivers/net/ethernet/sfc/mcdi_port_common.c
1159
MCDI_SET_QWORD(inbuf, MAC_STATS_IN_DMA_ADDR, dma_addr);
drivers/net/ethernet/sfc/mcdi_port_common.c
1231
(u64) efx->stats_buffer.dma_addr,
drivers/net/ethernet/sfc/net_driver.h
121
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/net_driver.h
1453
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/net_driver.h
148
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/net_driver.h
336
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/net_driver.h
359
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/nic.c
35
&buffer->dma_addr, gfp_flags);
drivers/net/ethernet/sfc/nic.c
46
buffer->addr, buffer->dma_addr);
drivers/net/ethernet/sfc/ptp.c
1028
ptp->start.dma_addr);
drivers/net/ethernet/sfc/rx.c
279
dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr,
drivers/net/ethernet/sfc/rx_common.c
165
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/rx_common.c
328
state->dma_addr,
drivers/net/ethernet/sfc/rx_common.c
377
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/rx_common.c
389
dma_addr =
drivers/net/ethernet/sfc/rx_common.c
394
dma_addr))) {
drivers/net/ethernet/sfc/rx_common.c
399
state->dma_addr = dma_addr;
drivers/net/ethernet/sfc/rx_common.c
402
dma_addr = state->dma_addr;
drivers/net/ethernet/sfc/rx_common.c
405
dma_addr += sizeof(struct efx_rx_page_state);
drivers/net/ethernet/sfc/rx_common.c
411
rx_buf->dma_addr = dma_addr + efx->rx_ip_align +
drivers/net/ethernet/sfc/rx_common.c
420
dma_addr += efx->rx_page_buf_step;
drivers/net/ethernet/sfc/rx_common.c
60
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/rx_common.h
71
dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len,
drivers/net/ethernet/sfc/siena/farch.c
1759
FRF_AZ_INT_ADR_KER, efx->irq_status.dma_addr);
drivers/net/ethernet/sfc/siena/farch.c
176
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/farch.c
184
dma_addr = buffer->buf.dma_addr + (i * EFX_BUF_SIZE);
drivers/net/ethernet/sfc/siena/farch.c
187
index, (unsigned long long)dma_addr);
drivers/net/ethernet/sfc/siena/farch.c
190
FRF_AZ_BUF_ADR_FBUF, dma_addr >> 12,
drivers/net/ethernet/sfc/siena/farch.c
239
BUG_ON(buffer->buf.dma_addr & (EFX_BUF_SIZE - 1));
drivers/net/ethernet/sfc/siena/farch.c
253
(u64)buffer->buf.dma_addr, len,
drivers/net/ethernet/sfc/siena/farch.c
269
(u64)buffer->buf.dma_addr, buffer->buf.len,
drivers/net/ethernet/sfc/siena/farch.c
343
FSF_AZ_TX_KER_BUF_ADDR, buffer->dma_addr);
drivers/net/ethernet/sfc/siena/farch.c
359
dma_addr_t dma_addr, unsigned int len)
drivers/net/ethernet/sfc/siena/farch.c
362
unsigned int limit = (~dma_addr & (EFX_PAGE_SIZE - 1)) + 1;
drivers/net/ethernet/sfc/siena/farch.c
477
FSF_AZ_RX_KER_BUF_ADDR, rx_buf->dma_addr);
drivers/net/ethernet/sfc/siena/mcdi_mon.c
151
hwmon->dma_buf.dma_addr);
drivers/net/ethernet/sfc/siena/mcdi_port_common.c
1160
dma_addr_t dma_addr = efx->stats_buffer.dma_addr;
drivers/net/ethernet/sfc/siena/mcdi_port_common.c
1166
MCDI_SET_QWORD(inbuf, MAC_STATS_IN_DMA_ADDR, dma_addr);
drivers/net/ethernet/sfc/siena/mcdi_port_common.c
1239
(u64) efx->stats_buffer.dma_addr,
drivers/net/ethernet/sfc/siena/net_driver.h
121
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/net_driver.h
1352
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/siena/net_driver.h
168
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/net_driver.h
325
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/net_driver.h
348
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/nic.c
35
&buffer->dma_addr, gfp_flags);
drivers/net/ethernet/sfc/siena/nic.c
46
buffer->addr, buffer->dma_addr);
drivers/net/ethernet/sfc/siena/nic.h
130
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/siena/ptp.c
1041
ptp->start.dma_addr);
drivers/net/ethernet/sfc/siena/rx.c
281
dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr,
drivers/net/ethernet/sfc/siena/rx_common.c
168
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/siena/rx_common.c
325
state->dma_addr,
drivers/net/ethernet/sfc/siena/rx_common.c
374
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/rx_common.c
386
dma_addr =
drivers/net/ethernet/sfc/siena/rx_common.c
391
dma_addr))) {
drivers/net/ethernet/sfc/siena/rx_common.c
396
state->dma_addr = dma_addr;
drivers/net/ethernet/sfc/siena/rx_common.c
399
dma_addr = state->dma_addr;
drivers/net/ethernet/sfc/siena/rx_common.c
402
dma_addr += sizeof(struct efx_rx_page_state);
drivers/net/ethernet/sfc/siena/rx_common.c
408
rx_buf->dma_addr = dma_addr + efx->rx_ip_align +
drivers/net/ethernet/sfc/siena/rx_common.c
417
dma_addr += efx->rx_page_buf_step;
drivers/net/ethernet/sfc/siena/rx_common.c
63
dma_unmap_page(&efx->pci_dev->dev, state->dma_addr,
drivers/net/ethernet/sfc/siena/rx_common.h
64
dma_sync_single_for_cpu(&efx->pci_dev->dev, rx_buf->dma_addr, len,
drivers/net/ethernet/sfc/siena/siena.c
308
BUG_ON(efx->irq_status.dma_addr & 0x0f);
drivers/net/ethernet/sfc/siena/siena.c
312
(unsigned long long)efx->irq_status.dma_addr,
drivers/net/ethernet/sfc/siena/siena_sriov.c
430
copy[1].from_addr = nic_data->vfdi_status.dma_addr + data_offset;
drivers/net/ethernet/sfc/siena/siena_sriov.c
816
if (!req->u.set_status_page.dma_addr || page_count > max_page_count) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
826
vf->status_addr = req->u.set_status_page.dma_addr;
drivers/net/ethernet/sfc/siena/siena_sriov.c
885
copy[0].to_addr = vf->buf.dma_addr;
drivers/net/ethernet/sfc/siena/siena_sriov.c
973
copy_req[k].from_addr = buffer->dma_addr;
drivers/net/ethernet/sfc/siena/tx.c
214
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/siena/tx.c
266
dma_addr = dma_map_single(&efx->pci_dev->dev,
drivers/net/ethernet/sfc/siena/tx.c
269
if (dma_mapping_error(&efx->pci_dev->dev, dma_addr))
drivers/net/ethernet/sfc/siena/tx.c
273
tx_buffer = efx_siena_tx_map_chunk(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/siena/tx.c
39
buffer->dma_addr = page_buf->dma_addr + offset;
drivers/net/ethernet/sfc/siena/tx_common.c
128
dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset;
drivers/net/ethernet/sfc/siena/tx_common.c
288
dma_addr_t dma_addr, size_t len)
drivers/net/ethernet/sfc/siena/tx_common.c
299
dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/siena/tx_common.c
304
buffer->dma_addr = dma_addr;
drivers/net/ethernet/sfc/siena/tx_common.c
307
dma_addr += dma_len;
drivers/net/ethernet/sfc/siena/tx_common.c
334
dma_addr_t dma_addr, unmap_addr;
drivers/net/ethernet/sfc/siena/tx_common.c
343
dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/sfc/siena/tx_common.c
346
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/siena/tx_common.c
348
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/siena/tx_common.c
359
efx_siena_tx_map_chunk(tx_queue, dma_addr, header_len);
drivers/net/ethernet/sfc/siena/tx_common.c
361
dma_addr += header_len;
drivers/net/ethernet/sfc/siena/tx_common.c
370
buffer = efx_siena_tx_map_chunk(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/siena/tx_common.c
377
buffer->dma_offset = buffer->dma_addr - unmap_addr;
drivers/net/ethernet/sfc/siena/tx_common.c
391
dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len,
drivers/net/ethernet/sfc/siena/tx_common.c
395
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/siena/tx_common.c
397
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/siena/tx_common.h
31
dma_addr_t dma_addr, size_t len);
drivers/net/ethernet/sfc/siena/vfdi.h
201
u64 dma_addr;
drivers/net/ethernet/sfc/tx.c
419
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/tx.c
47
buffer->dma_addr = page_buf->dma_addr + offset;
drivers/net/ethernet/sfc/tx.c
471
dma_addr = dma_map_single(&efx->pci_dev->dev,
drivers/net/ethernet/sfc/tx.c
474
if (dma_mapping_error(&efx->pci_dev->dev, dma_addr))
drivers/net/ethernet/sfc/tx.c
478
tx_buffer = efx_tx_map_chunk(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/tx.h
16
dma_addr_t dma_addr, unsigned int len);
drivers/net/ethernet/sfc/tx_common.c
165
dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset;
drivers/net/ethernet/sfc/tx_common.c
328
dma_addr_t dma_addr, size_t len)
drivers/net/ethernet/sfc/tx_common.c
339
dma_len = nic_type->tx_limit_len(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/tx_common.c
344
buffer->dma_addr = dma_addr;
drivers/net/ethernet/sfc/tx_common.c
347
dma_addr += dma_len;
drivers/net/ethernet/sfc/tx_common.c
374
dma_addr_t dma_addr, unmap_addr;
drivers/net/ethernet/sfc/tx_common.c
383
dma_addr = dma_map_single(dma_dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/sfc/tx_common.c
386
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/tx_common.c
388
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/tx_common.c
399
efx_tx_map_chunk(tx_queue, dma_addr, header_len);
drivers/net/ethernet/sfc/tx_common.c
401
dma_addr += header_len;
drivers/net/ethernet/sfc/tx_common.c
410
buffer = efx_tx_map_chunk(tx_queue, dma_addr, len);
drivers/net/ethernet/sfc/tx_common.c
417
buffer->dma_offset = buffer->dma_addr - unmap_addr;
drivers/net/ethernet/sfc/tx_common.c
431
dma_addr = skb_frag_dma_map(dma_dev, fragment, 0, len,
drivers/net/ethernet/sfc/tx_common.c
435
unmap_addr = dma_addr;
drivers/net/ethernet/sfc/tx_common.c
437
if (unlikely(dma_mapping_error(dma_dev, dma_addr)))
drivers/net/ethernet/sfc/tx_common.h
39
dma_addr_t dma_addr, size_t len);
drivers/net/ethernet/sfc/tx_tso.c
103
dma_addr_t dma_addr, unsigned int len,
drivers/net/ethernet/sfc/tx_tso.c
119
buffer->dma_addr = dma_addr;
drivers/net/ethernet/sfc/tx_tso.c
122
dma_addr, len);
drivers/net/ethernet/sfc/tx_tso.c
130
dma_addr += dma_len;
drivers/net/ethernet/sfc/tx_tso.c
175
dma_addr_t dma_addr;
drivers/net/ethernet/sfc/tx_tso.c
198
dma_addr = dma_map_single(dma_dev, skb->data,
drivers/net/ethernet/sfc/tx_tso.c
200
st->header_dma_addr = dma_addr;
drivers/net/ethernet/sfc/tx_tso.c
202
st->dma_addr = dma_addr + header_len;
drivers/net/ethernet/sfc/tx_tso.c
205
return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0;
drivers/net/ethernet/sfc/tx_tso.c
216
st->dma_addr = st->unmap_addr;
drivers/net/ethernet/sfc/tx_tso.c
253
efx_tx_queue_insert(tx_queue, st->dma_addr, n, &buffer);
drivers/net/ethernet/sfc/tx_tso.c
270
st->dma_addr += n;
drivers/net/ethernet/sfc/tx_tso.c
325
buffer->dma_addr = st->header_dma_addr;
drivers/net/ethernet/sfc/tx_tso.c
65
dma_addr_t dma_addr;
drivers/net/ethernet/socionext/netsec.c
1017
dma_sync_single_for_cpu(priv->dev, desc->dma_addr, pkt_len,
drivers/net/ethernet/socionext/netsec.c
1070
desc->dma_addr = dma_handle;
drivers/net/ethernet/socionext/netsec.c
1178
tx_desc.dma_addr = dma_map_single(priv->dev, skb->data,
drivers/net/ethernet/socionext/netsec.c
1180
if (dma_mapping_error(priv->dev, tx_desc.dma_addr)) {
drivers/net/ethernet/socionext/netsec.c
1220
dma_unmap_single(priv->dev, desc->dma_addr, desc->len,
drivers/net/ethernet/socionext/netsec.c
1341
desc->dma_addr = dma_handle;
drivers/net/ethernet/socionext/netsec.c
274
dma_addr_t dma_addr;
drivers/net/ethernet/socionext/netsec.c
624
de->data_buf_addr_up = upper_32_bits(desc->dma_addr);
drivers/net/ethernet/socionext/netsec.c
625
de->data_buf_addr_lw = lower_32_bits(desc->dma_addr);
drivers/net/ethernet/socionext/netsec.c
630
dring->desc[idx].dma_addr = desc->dma_addr;
drivers/net/ethernet/socionext/netsec.c
665
dma_unmap_single(priv->dev, desc->dma_addr, desc->len,
drivers/net/ethernet/socionext/netsec.c
811
de->data_buf_addr_up = upper_32_bits(desc->dma_addr);
drivers/net/ethernet/socionext/netsec.c
812
de->data_buf_addr_lw = lower_32_bits(desc->dma_addr);
drivers/net/ethernet/socionext/netsec.c
872
tx_desc.dma_addr = dma_handle;
drivers/net/ethernet/spacemit/k1_emac.c
330
if (!buf->dma_addr)
drivers/net/ethernet/spacemit/k1_emac.c
334
dma_unmap_page(&priv->pdev->dev, buf->dma_addr,
drivers/net/ethernet/spacemit/k1_emac.c
338
buf->dma_addr, buf->dma_len,
drivers/net/ethernet/spacemit/k1_emac.c
341
buf->dma_addr = 0;
drivers/net/ethernet/spacemit/k1_emac.c
378
dma_unmap_single(&priv->pdev->dev, rx_buf->dma_addr,
drivers/net/ethernet/spacemit/k1_emac.c
563
rx_buf->dma_addr = dma_map_single(&priv->pdev->dev, skb->data,
drivers/net/ethernet/spacemit/k1_emac.c
566
if (dma_mapping_error(&priv->pdev->dev, rx_buf->dma_addr)) {
drivers/net/ethernet/spacemit/k1_emac.c
577
rx_desc.buffer_addr_1 = rx_buf->dma_addr;
drivers/net/ethernet/spacemit/k1_emac.c
59
u64 dma_addr;
drivers/net/ethernet/spacemit/k1_emac.c
627
dma_unmap_single(&priv->pdev->dev, rx_buf->dma_addr,
drivers/net/ethernet/spacemit/k1_emac.c
715
tx_buf->buf[buf_idx].dma_addr = addr;
drivers/net/ethernet/spacemit/k1_emac.c
72
u64 dma_addr;
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
416
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
425
dma_addr = dma_rx_phy + i * sizeof(*p);
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
427
i, &dma_addr,
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
436
dma_addr = dma_rx_phy + i * sizeof(*extp);
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
438
i, &dma_addr,
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
449
dma_addr = dma_rx_phy + i * sizeof(*ep);
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
451
i, &dma_addr,
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
416
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
423
dma_addr = dma_rx_phy + i * sizeof(*ep);
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
427
i, &dma_addr,
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
262
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
269
dma_addr = dma_rx_phy + i * sizeof(*p);
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
273
i, &dma_addr,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1791
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1805
dma_addr = xsk_buff_xdp_get_dma(buf->xdp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1806
stmmac_set_desc_addr(priv, p, dma_addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2667
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2696
dma_addr = xsk_buff_raw_get_dma(pool, xdp_desc.addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2698
xsk_buff_raw_dma_sync_for_device(pool, dma_addr, xdp_desc.len);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2714
stmmac_set_desc_addr(priv, tx_desc, dma_addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5074
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5095
dma_addr = dma_map_single(priv->device, xdpf->data,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5097
if (dma_mapping_error(priv->device, dma_addr))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5104
dma_addr = page_pool_get_dma_addr(page) + sizeof(*xdpf) +
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5106
dma_sync_single_for_device(priv->device, dma_addr,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5112
tx_q->tx_skbuff_dma[entry].buf = dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5120
stmmac_set_desc_addr(priv, tx_desc, dma_addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5343
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5359
dma_addr = xsk_buff_xdp_get_dma(buf->xdp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5360
stmmac_set_desc_addr(priv, rx_desc, dma_addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6436
dma_addr_t dma_addr;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6441
dma_addr = dma_phy_addr + i * desc_size;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6443
i, &dma_addr,
drivers/net/ethernet/sun/cassini.c
1388
rxd[i].buffer = cpu_to_le64(page->dma_addr);
drivers/net/ethernet/sun/cassini.c
1948
dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off,
drivers/net/ethernet/sun/cassini.c
1952
page->dma_addr + off, i,
drivers/net/ethernet/sun/cassini.c
1978
dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off,
drivers/net/ethernet/sun/cassini.c
1987
page->dma_addr + off, i,
drivers/net/ethernet/sun/cassini.c
2013
page->dma_addr,
drivers/net/ethernet/sun/cassini.c
2017
page->dma_addr,
drivers/net/ethernet/sun/cassini.c
2052
dma_sync_single_for_cpu(&cp->pdev->dev, page->dma_addr + off,
drivers/net/ethernet/sun/cassini.c
2056
page->dma_addr + off, i,
drivers/net/ethernet/sun/cassini.c
2069
page->dma_addr,
drivers/net/ethernet/sun/cassini.c
2074
page->dma_addr,
drivers/net/ethernet/sun/cassini.c
2145
cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr);
drivers/net/ethernet/sun/cassini.c
2200
cpu_to_le64(new->dma_addr);
drivers/net/ethernet/sun/cassini.c
446
dma_unmap_page(&cp->pdev->dev, page->dma_addr, cp->page_size,
drivers/net/ethernet/sun/cassini.c
477
page->dma_addr = dma_map_page(&cp->pdev->dev, page->buffer, 0,
drivers/net/ethernet/sun/cassini.h
2700
dma_addr_t dma_addr;
drivers/net/ethernet/sun/niu.c
10000
u64 *dma_addr, gfp_t flag)
drivers/net/ethernet/sun/niu.c
10008
*dma_addr = __pa(page);
drivers/net/ethernet/sun/sungem.c
1560
dma_addr_t dma_addr;
drivers/net/ethernet/sun/sungem.c
1568
dma_addr = le64_to_cpu(rxd->buffer);
drivers/net/ethernet/sun/sungem.c
1569
dma_unmap_page(&gp->pdev->dev, dma_addr,
drivers/net/ethernet/sun/sungem.c
1592
dma_addr = le64_to_cpu(txd->buffer);
drivers/net/ethernet/sun/sungem.c
1593
dma_unmap_page(&gp->pdev->dev, dma_addr,
drivers/net/ethernet/sun/sungem.c
1610
dma_addr_t dma_addr;
drivers/net/ethernet/sun/sungem.c
1632
dma_addr = dma_map_page(&gp->pdev->dev,
drivers/net/ethernet/sun/sungem.c
1637
rxd->buffer = cpu_to_le64(dma_addr);
drivers/net/ethernet/sun/sungem.c
640
dma_addr_t dma_addr;
drivers/net/ethernet/sun/sungem.c
670
dma_addr = le64_to_cpu(txd->buffer);
drivers/net/ethernet/sun/sungem.c
673
dma_unmap_page(&gp->pdev->dev, dma_addr, dma_len,
drivers/net/ethernet/sun/sungem.c
763
dma_addr_t dma_addr;
drivers/net/ethernet/sun/sungem.c
804
dma_addr = le64_to_cpu(rxd->buffer);
drivers/net/ethernet/sun/sungem.c
813
dma_unmap_page(&gp->pdev->dev, dma_addr,
drivers/net/ethernet/sun/sungem.c
836
dma_sync_single_for_cpu(&gp->pdev->dev, dma_addr, len,
drivers/net/ethernet/sun/sungem.c
839
dma_sync_single_for_device(&gp->pdev->dev, dma_addr,
drivers/net/ethernet/sun/sunhme.c
1165
u32 dma_addr;
drivers/net/ethernet/sun/sunhme.c
1168
dma_addr = hme_read_desc32(hp, &rxd->rx_addr);
drivers/net/ethernet/sun/sunhme.c
1169
dma_unmap_single(hp->dma_dev, dma_addr,
drivers/net/ethernet/sun/sunhme.c
1180
u32 dma_addr;
drivers/net/ethernet/sun/sunhme.c
1187
dma_addr = hme_read_desc32(hp, &txd->tx_addr);
drivers/net/ethernet/sun/sunhme.c
1189
dma_unmap_single(hp->dma_dev, dma_addr,
drivers/net/ethernet/sun/sunhme.c
1194
dma_unmap_page(hp->dma_dev, dma_addr,
drivers/net/ethernet/sun/sunhme.c
1706
u32 flags, dma_addr, dma_len;
drivers/net/ethernet/sun/sunhme.c
1728
dma_addr = hme_read_desc32(hp, &this->tx_addr);
drivers/net/ethernet/sun/sunhme.c
1733
dma_unmap_single(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE);
drivers/net/ethernet/sun/sunhme.c
1735
dma_unmap_page(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE);
drivers/net/ethernet/sun/sunhme.c
1772
u32 dma_addr = hme_read_desc32(hp, &this->rx_addr);
drivers/net/ethernet/sun/sunhme.c
1790
dma_addr);
drivers/net/ethernet/sun/sunhme.c
1814
dma_unmap_single(hp->dma_dev, dma_addr, RX_BUF_ALLOC_SIZE, DMA_FROM_DEVICE);
drivers/net/ethernet/sun/sunhme.c
1833
dma_sync_single_for_cpu(hp->dma_dev, dma_addr, len + 2, DMA_FROM_DEVICE);
drivers/net/ethernet/sun/sunhme.c
1835
dma_sync_single_for_device(hp->dma_dev, dma_addr, len + 2, DMA_FROM_DEVICE);
drivers/net/ethernet/sun/sunhme.c
1839
dma_addr);
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
102
pool->dma_addr);
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
116
return addr ? pool->dma_addr + (addr - pool->cpumem) : 0;
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
122
return dma ? pool->cpumem + (dma - pool->dma_addr) : NULL;
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
19
dma_addr_t dma_addr;
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
39
pool->dma_addr);
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
85
&pool->dma_addr, GFP_KERNEL);
drivers/net/ethernet/ti/k3-cppi-desc-pool.c
91
(phys_addr_t)pool->dma_addr, pool->mem_size,
drivers/net/ethernet/ti/netcp_core.c
1086
dma_addr_t dma_addr;
drivers/net/ethernet/ti/netcp_core.c
1091
dma_addr = dma_map_single(dev, skb->data, pkt_len, DMA_TO_DEVICE);
drivers/net/ethernet/ti/netcp_core.c
1092
if (unlikely(dma_mapping_error(dev, dma_addr))) {
drivers/net/ethernet/ti/netcp_core.c
1100
dma_unmap_single(dev, dma_addr, pkt_len, DMA_TO_DEVICE);
drivers/net/ethernet/ti/netcp_core.c
1104
set_pkt_info(dma_addr, pkt_len, 0, desc);
drivers/net/ethernet/ti/netcp_core.c
1123
dma_addr = dma_map_page(dev, page, page_offset, buf_len,
drivers/net/ethernet/ti/netcp_core.c
1125
if (unlikely(!dma_addr)) {
drivers/net/ethernet/ti/netcp_core.c
1133
dma_unmap_page(dev, dma_addr, buf_len, DMA_TO_DEVICE);
drivers/net/ethernet/ti/netcp_core.c
1138
set_pkt_info(dma_addr, buf_len, 0, ndesc);
drivers/net/ethernet/ti/netcp_core.c
1149
&dma_addr, &dma_sz);
drivers/net/fddi/defxx.c
3006
dma_addr_t dma_addr;
drivers/net/fddi/defxx.c
3018
dma_addr = dma_map_single(bp->bus_dev,
drivers/net/fddi/defxx.c
3022
if (dma_mapping_error(bp->bus_dev, dma_addr)) {
drivers/net/fddi/defxx.c
3032
(u32)dma_addr;
drivers/net/fddi/defxx.c
3108
dma_addr_t dma_addr;
drivers/net/fddi/defxx.c
3117
dma_addr = bp->descr_block_virt->rcv_data[entry].long_1;
drivers/net/fddi/defxx.c
3119
dma_addr + RCV_BUFF_K_DESCR,
drivers/net/fddi/defxx.c
3169
dma_addr,
drivers/net/fddi/defxx.c
3194
dma_addr +
drivers/net/fddi/defxx.c
3300
dma_addr_t dma_addr;
drivers/net/fddi/defxx.c
3355
dma_addr = dma_map_single(bp->bus_dev, skb->data, skb->len,
drivers/net/fddi/defxx.c
3357
if (dma_mapping_error(bp->bus_dev, dma_addr)) {
drivers/net/fddi/defxx.c
3410
p_xmt_descr->long_1 = (u32)dma_addr;
drivers/net/fddi/skfp/h/osdef1st.h
106
dma_addr_t dma_addr;
drivers/net/fddi/skfp/h/osdef1st.h
111
dma_addr_t dma_addr;
drivers/net/fddi/skfp/skfddi.c
1183
txd->txd_os.dma_addr = dma_address; // save dma mapping
drivers/net/fddi/skfp/skfddi.c
1469
if (r->rxd_os.skb && r->rxd_os.dma_addr) {
drivers/net/fddi/skfp/skfddi.c
1473
r->rxd_os.dma_addr, MaxFrameSize,
drivers/net/fddi/skfp/skfddi.c
1475
r->rxd_os.dma_addr = 0;
drivers/net/fddi/skfp/skfddi.c
1509
dma_unmap_single(&(&smc->os.pdev)->dev, txd->txd_os.dma_addr,
drivers/net/fddi/skfp/skfddi.c
1511
txd->txd_os.dma_addr = 0;
drivers/net/fddi/skfp/skfddi.c
1716
rxd->rxd_os.dma_addr = b_addr;
drivers/net/fddi/skfp/skfddi.c
1730
rxd->rxd_os.dma_addr = b_addr;
drivers/net/fddi/skfp/skfddi.c
1783
rxd->rxd_os.dma_addr = b_addr;
drivers/net/fddi/skfp/skfddi.c
1840
rxd->rxd_os.dma_addr, MaxFrameSize,
drivers/net/thunderbolt/main.c
509
dma_addr_t dma_addr;
drivers/net/thunderbolt/main.c
524
dma_addr = dma_map_page(dma_dev, tf->page, 0,
drivers/net/thunderbolt/main.c
526
if (dma_mapping_error(dma_dev, dma_addr)) {
drivers/net/thunderbolt/main.c
531
tf->frame.buffer_phy = dma_addr;
drivers/net/thunderbolt/main.c
534
trace_tbnet_alloc_rx_frame(index, tf->page, dma_addr,
drivers/net/thunderbolt/main.c
591
dma_addr_t dma_addr;
drivers/net/thunderbolt/main.c
599
dma_addr = dma_map_page(dma_dev, tf->page, 0, TBNET_FRAME_SIZE,
drivers/net/thunderbolt/main.c
601
if (dma_mapping_error(dma_dev, dma_addr)) {
drivers/net/thunderbolt/main.c
609
tf->frame.buffer_phy = dma_addr;
drivers/net/thunderbolt/main.c
614
trace_tbnet_alloc_tx_frame(i, tf->page, dma_addr, DMA_TO_DEVICE);
drivers/net/vmxnet3/vmxnet3_drv.c
1420
vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/vmxnet3/vmxnet3_drv.c
1429
*dma_addr = page_pool_get_dma_addr(page) + pp->p.offset;
drivers/net/vmxnet3/vmxnet3_drv.c
1651
BUG_ON(rxd->addr != rbi->dma_addr ||
drivers/net/vmxnet3/vmxnet3_drv.c
1783
rbi->dma_addr,
drivers/net/vmxnet3/vmxnet3_drv.c
1789
rbi->dma_addr = new_dma_addr;
drivers/net/vmxnet3/vmxnet3_drv.c
1790
rxd->addr = cpu_to_le64(rbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_drv.c
1857
rbi->dma_addr, rbi->len,
drivers/net/vmxnet3/vmxnet3_drv.c
1864
rbi->dma_addr = new_dma_addr;
drivers/net/vmxnet3/vmxnet3_drv.c
1865
rxd->addr = cpu_to_le64(rbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_drv.c
383
dma_unmap_single(&pdev->dev, tbi->dma_addr, tbi->len,
drivers/net/vmxnet3/vmxnet3_drv.c
386
dma_unmap_page(&pdev->dev, tbi->dma_addr, tbi->len,
drivers/net/vmxnet3/vmxnet3_drv.c
691
&rbi->dma_addr,
drivers/net/vmxnet3/vmxnet3_drv.c
709
rbi->dma_addr = dma_map_single(
drivers/net/vmxnet3/vmxnet3_drv.c
714
rbi->dma_addr)) {
drivers/net/vmxnet3/vmxnet3_drv.c
734
rbi->dma_addr = dma_map_page(
drivers/net/vmxnet3/vmxnet3_drv.c
739
rbi->dma_addr)) {
drivers/net/vmxnet3/vmxnet3_drv.c
751
gd->rxd.addr = cpu_to_le64(rbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_drv.c
850
tbi->dma_addr = dma_map_single(&adapter->pdev->dev,
drivers/net/vmxnet3/vmxnet3_drv.c
853
if (dma_mapping_error(&adapter->pdev->dev, tbi->dma_addr))
drivers/net/vmxnet3/vmxnet3_drv.c
861
gdesc->txd.addr = cpu_to_le64(tbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_drv.c
892
tbi->dma_addr = skb_frag_dma_map(&adapter->pdev->dev, frag,
drivers/net/vmxnet3/vmxnet3_drv.c
895
if (dma_mapping_error(&adapter->pdev->dev, tbi->dma_addr))
drivers/net/vmxnet3/vmxnet3_drv.c
903
gdesc->txd.addr = cpu_to_le64(tbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_int.h
210
dma_addr_t dma_addr;
drivers/net/vmxnet3/vmxnet3_int.h
294
dma_addr_t dma_addr;
drivers/net/vmxnet3/vmxnet3_xdp.c
144
tbi->dma_addr = dma_map_single(&adapter->pdev->dev,
drivers/net/vmxnet3/vmxnet3_xdp.c
147
if (dma_mapping_error(&adapter->pdev->dev, tbi->dma_addr)) {
drivers/net/vmxnet3/vmxnet3_xdp.c
154
tbi->dma_addr = page_pool_get_dma_addr(page) +
drivers/net/vmxnet3/vmxnet3_xdp.c
157
tbi->dma_addr, buf_size,
drivers/net/vmxnet3/vmxnet3_xdp.c
166
gdesc->txd.addr = cpu_to_le64(tbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_xdp.c
424
rbi->dma_addr = new_dma_addr;
drivers/net/vmxnet3/vmxnet3_xdp.c
425
rxd->addr = cpu_to_le64(rbi->dma_addr);
drivers/net/vmxnet3/vmxnet3_xdp.h
39
void *vmxnet3_pp_get_buff(struct page_pool *pp, dma_addr_t *dma_addr,
drivers/net/wan/fsl_qmc_hdlc.c
259
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE);
drivers/net/wan/fsl_qmc_hdlc.c
309
desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data,
drivers/net/wan/fsl_qmc_hdlc.c
311
ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr);
drivers/net/wan/fsl_qmc_hdlc.c
315
ret = qmc_chan_read_submit(qmc_hdlc->qmc_chan, desc->dma_addr, desc->dma_size,
drivers/net/wan/fsl_qmc_hdlc.c
32
dma_addr_t dma_addr;
drivers/net/wan/fsl_qmc_hdlc.c
323
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_FROM_DEVICE);
drivers/net/wan/fsl_qmc_hdlc.c
341
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_TO_DEVICE);
drivers/net/wan/fsl_qmc_hdlc.c
358
desc->dma_addr = dma_map_single(qmc_hdlc->dev, desc->skb->data,
drivers/net/wan/fsl_qmc_hdlc.c
360
ret = dma_mapping_error(qmc_hdlc->dev, desc->dma_addr);
drivers/net/wan/fsl_qmc_hdlc.c
366
ret = qmc_chan_write_submit(qmc_hdlc->qmc_chan, desc->dma_addr, desc->dma_size,
drivers/net/wan/fsl_qmc_hdlc.c
369
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size, DMA_TO_DEVICE);
drivers/net/wan/fsl_qmc_hdlc.c
629
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size,
drivers/net/wan/fsl_qmc_hdlc.c
656
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size,
drivers/net/wan/fsl_qmc_hdlc.c
666
dma_unmap_single(qmc_hdlc->dev, desc->dma_addr, desc->dma_size,
drivers/net/wireless/ath/wcn36xx/dxe.c
178
wcn_ch->cpu_addr = dma_alloc_coherent(dev, size, &wcn_ch->dma_addr,
drivers/net/wireless/ath/wcn36xx/dxe.c
188
cur_ctl->desc_phy_addr = wcn_ch->dma_addr +
drivers/net/wireless/ath/wcn36xx/dxe.c
233
dma_free_coherent(dev, size,wcn_ch->cpu_addr, wcn_ch->dma_addr);
drivers/net/wireless/ath/wcn36xx/dxe.c
589
dma_addr_t dma_addr;
drivers/net/wireless/ath/wcn36xx/dxe.c
630
dma_addr = dxe->dst_addr_l;
drivers/net/wireless/ath/wcn36xx/dxe.c
636
dma_unmap_single(wcn->dev, dma_addr, WCN36XX_PKT_SIZE,
drivers/net/wireless/ath/wcn36xx/dxe.h
433
dma_addr_t dma_addr;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2280
packet->dma_addr = dma_map_single(&priv->pci_dev->dev,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2284
if (dma_mapping_error(&priv->pci_dev->dev, packet->dma_addr)) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2465
dma_unmap_single(&priv->pci_dev->dev, packet->dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2501
priv->rx_queue.drv[i].host_addr = packet->dma_addr;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2548
dma_unmap_single(&priv->pci_dev->dev, packet->dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2583
priv->rx_queue.drv[i].host_addr = packet->dma_addr;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2674
dma_sync_single_for_cpu(&priv->pci_dev->dev, packet->dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4576
priv->rx_queue.drv[i].host_addr = packet->dma_addr;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4586
priv->rx_buffers[j].dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
4641
priv->rx_buffers[i].dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2100.h
333
dma_addr_t dma_addr;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
3398
rxq->pool[i].dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2200.c
3711
ipw_write32(priv, base, q->dma_addr);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
3731
&q->q.dma_addr, GFP_KERNEL);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
3809
q->dma_addr);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5107
rxb->dma_addr);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5153
rxb->dma_addr =
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5189
rxq->pool[i].dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8230
dma_sync_single_for_cpu(&priv->pci_dev->dev, rxb->dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8383
dma_unmap_single(&priv->pci_dev->dev, rxb->dma_addr,
drivers/net/wireless/intel/ipw2x00/ipw2200.h
370
dma_addr_t dma_addr; /**< physical addr for BD's */
drivers/net/wireless/intel/ipw2x00/ipw2200.h
699
dma_addr_t dma_addr;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
910
il3945_dma_addr2rbd_ptr(struct il_priv *il, dma_addr_t dma_addr)
drivers/net/wireless/intel/iwlegacy/3945-mac.c
912
return cpu_to_le32((u32) dma_addr);
drivers/net/wireless/intel/iwlegacy/3945.c
2203
shared_data->tx_base_ptr[txq_id] = cpu_to_le32((u32) txq->q.dma_addr);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
234
il4965_dma_addr2rbd_ptr(struct il_priv *il, dma_addr_t dma_addr)
drivers/net/wireless/intel/iwlegacy/4965-mac.c
236
return cpu_to_le32((u32) (dma_addr >> 8));
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4002
il_wr(il, FH49_MEM_CBBC_QUEUE(txq_id), txq->q.dma_addr >> 8);
drivers/net/wireless/intel/iwlegacy/common.c
2790
txq->tfds, txq->q.dma_addr);
drivers/net/wireless/intel/iwlegacy/common.c
2870
txq->tfds, txq->q.dma_addr);
drivers/net/wireless/intel/iwlegacy/common.c
2982
dma_alloc_coherent(dev, tfd_sz, &txq->q.dma_addr, GFP_KERNEL);
drivers/net/wireless/intel/iwlegacy/common.h
123
dma_addr_t dma_addr; /* physical addr for BD's */
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
794
dma_addr_t dma_addr;
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
253
cpu_to_le64(trans_pcie->txqs.txq[trans->conf.cmd_queue]->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c
222
cpu_to_le64(trans_pcie->txqs.txq[trans->conf.cmd_queue]->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
607
dma_addr_t dma_addr;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
651
res = IWL_TSO_PAGE_INFO(addr)->dma_addr;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
137
static inline __le32 iwl_pcie_dma_addr2rbd_ptr(dma_addr_t dma_addr)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
139
return cpu_to_le32((u32)(dma_addr >> 8));
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1061
cmd.old.tfdq_addr = cpu_to_le64(txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1079
cmd.new.u.add.tfdq_dram_addr = cpu_to_le64(txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
43
info->dma_addr = phys;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
862
txq->tfds, txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1800
info->dma_addr = phys;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
220
dma_unmap_page(trans->dev, info->dma_addr, PAGE_SIZE,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
462
txq->tfds, txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
463
txq->dma_addr = 0;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
554
txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
558
txq->dma_addr >> 8);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
761
&txq->dma_addr, GFP_KERNEL);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
786
dma_free_coherent(trans->dev, tfd_sz, txq->tfds, txq->dma_addr);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
988
trans_pcie->txqs.txq[txq_id]->dma_addr >> 8);
drivers/net/wireless/intersil/p54/p54pci.c
199
dma_addr_t dma_addr;
drivers/net/wireless/intersil/p54/p54pci.c
217
dma_addr = le32_to_cpu(desc->host_addr);
drivers/net/wireless/intersil/p54/p54pci.c
218
dma_sync_single_for_cpu(&priv->pdev->dev, dma_addr,
drivers/net/wireless/intersil/p54/p54pci.c
224
dma_unmap_single(&priv->pdev->dev, dma_addr,
drivers/net/wireless/intersil/p54/p54pci.c
231
dma_sync_single_for_device(&priv->pdev->dev, dma_addr,
drivers/net/wireless/intersil/p54/p54pci.h
58
__le32 dma_addr;
drivers/net/wireless/intersil/p54/p54usb.c
770
P54U_WRITE(NET2280_DEV_U32, &devreg->dma_addr,
drivers/net/wireless/marvell/mwl8k.c
2218
dma_addr_t dma_addr;
drivers/net/wireless/marvell/mwl8k.c
2254
dma_addr = dma_map_single(&priv->pdev->dev, cmd, dma_size,
drivers/net/wireless/marvell/mwl8k.c
2256
if (dma_mapping_error(&priv->pdev->dev, dma_addr)) {
drivers/net/wireless/marvell/mwl8k.c
2262
iowrite32(dma_addr, regs + MWL8K_HIU_GEN_PTR);
drivers/net/wireless/marvell/mwl8k.c
2274
dma_unmap_single(&priv->pdev->dev, dma_addr, dma_size,
drivers/net/wireless/marvell/mwl8k.c
610
dma_addr_t dma_addr;
drivers/net/wireless/marvell/mwl8k.c
613
dma_addr = dma_map_single(&priv->pdev->dev, data, length,
drivers/net/wireless/marvell/mwl8k.c
615
if (dma_mapping_error(&priv->pdev->dev, dma_addr))
drivers/net/wireless/marvell/mwl8k.c
618
iowrite32(dma_addr, regs + MWL8K_HIU_GEN_PTR);
drivers/net/wireless/marvell/mwl8k.c
644
dma_unmap_single(&priv->pdev->dev, dma_addr, length, DMA_TO_DEVICE);
drivers/net/wireless/mediatek/mt76/dma.c
135
dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/dma.c
294
entry->dma_addr[0] = buf->addr;
drivers/net/wireless/mediatek/mt76/dma.c
31
t->dma_addr = addr;
drivers/net/wireless/mediatek/mt76/dma.c
334
entry->dma_addr[0] = buf[0].addr;
drivers/net/wireless/mediatek/mt76/dma.c
342
entry->dma_addr[1] = buf[1].addr;
drivers/net/wireless/mediatek/mt76/dma.c
382
dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0],
drivers/net/wireless/mediatek/mt76/dma.c
386
dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1],
drivers/net/wireless/mediatek/mt76/dma.c
466
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/dma.c
483
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/dma.c
525
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/dma.c
530
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/dma.c
537
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/dma.c
677
tx_info.buf[n].addr = t->dma_addr;
drivers/net/wireless/mediatek/mt76/dma.c
701
dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/dma.c
704
dma_sync_single_for_device(dev->dma_dev, t->dma_addr, dev->drv->txwi_size,
drivers/net/wireless/mediatek/mt76/mt76.h
218
dma_addr_t dma_addr[2];
drivers/net/wireless/mediatek/mt76/mt76.h
443
dma_addr_t dma_addr;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1768
static u32 mt7996_msdu_page_hash_from_addr(dma_addr_t dma_addr)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1773
while (dma_addr) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1774
val += (u32)((dma_addr & 0xff) + i) % MT7996_RRO_MSDU_PG_HASH_SIZE;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1775
dma_addr >>= 8;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1783
mt7996_rro_msdu_page_get(struct mt7996_dev *dev, dma_addr_t dma_addr)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1785
u32 hash = mt7996_msdu_page_hash_from_addr(dma_addr);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1792
if (p->dma_addr == dma_addr) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1816
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1847
dma_addr_t dma_addr, void *data)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1860
p->dma_addr = dma_addr;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1863
hash = mt7996_msdu_page_hash_from_addr(dma_addr);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1914
dma_addr_t dma_addr = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1936
dma_addr = FIELD_GET(WED_RRO_ADDR_HEAD_HIGH_MASK, data);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1937
dma_addr <<= 32;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1939
dma_addr |= le32_to_cpu(e->head_low);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1944
p = mt7996_rro_msdu_page_get(dev, dma_addr);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1948
dma_sync_single_for_cpu(mdev->dma_dev, p->dma_addr,
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1967
dma_sync_single_for_cpu(mdev->dma_dev, t->dma_addr,
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1971
t->dma_addr = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2030
dma_addr =
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2033
dma_addr <<= 32;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2034
dma_addr |= le32_to_cpu(pinfo->pg_low);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2036
dma_addr = le32_to_cpu(pinfo->pg_low);
drivers/net/wireless/mediatek/mt76/mt7996/mt7996.h
313
dma_addr_t dma_addr;
drivers/net/wireless/mediatek/mt76/mt7996/mt7996.h
840
dma_addr_t dma_addr, void *data);
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
15
u32 val, hif1_ofs = 0, dma_addr;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
154
dma_addr_t dma_addr;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
169
&dma_addr, GFP_KERNEL))
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
174
dma_addr, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
183
&dma_addr, GFP_KERNEL))
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
188
dma_addr, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
196
&dma_addr, GFP_KERNEL))
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
201
dma_addr, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
40
dma_addr = phy_addr;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
42
dma_addr += MT_RXQ_RING_BASE(MT_RXQ_RRO_BAND1) + 0x90 +
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
45
dma_addr += MT_RXQ_RING_BASE(MT_RXQ_RRO_BAND0) + 0x80;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
48
dma_addr, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
63
dma_addr = phy_addr;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
65
dma_addr += MT_TXQ_RING_BASE(0) + 0x150 + hif1_ofs;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
67
dma_addr += MT_TXQ_RING_BASE(0) + 0x120;
drivers/net/wireless/mediatek/mt76/mt7996/npu.c
71
dma_addr, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/npu.c
33
e->dma_addr[0] = page_pool_get_dma_addr(page) + offset;
drivers/net/wireless/mediatek/mt76/npu.c
36
desc[q->head].addr = e->dma_addr[0];
drivers/net/wireless/mediatek/mt76/npu.c
52
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/npu.c
83
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/tx.c
881
t->dma_addr = phys;
drivers/ntb/ntb_transport.c
1319
mw->dma_addr = 0;
drivers/ntb/ntb_transport.c
225
dma_addr_t dma_addr;
drivers/ntb/ntb_transport.c
769
mw->alloc_addr, mw->dma_addr);
drivers/ntb/ntb_transport.c
780
dma_addr_t dma_addr;
drivers/ntb/ntb_transport.c
797
&dma_addr, GFP_KERNEL,
drivers/ntb/ntb_transport.c
812
if (!IS_ALIGNED(dma_addr, align)) {
drivers/ntb/ntb_transport.c
815
dma_addr = ALIGN(dma_addr, align);
drivers/ntb/ntb_transport.c
824
mw->dma_addr = dma_addr;
drivers/ntb/ntb_transport.c
829
dma_free_coherent(ntb_dev, mw->alloc_size, alloc_addr, dma_addr);
drivers/ntb/ntb_transport.c
882
rc = ntb_mw_set_trans(nt->ndev, PIDX, num_mw, mw->dma_addr,
drivers/nvme/host/apple.c
377
dma_addr_t dma_addr = iod->first_dma;
drivers/nvme/host/apple.c
384
dma_pool_free(anv->prp_page_pool, prp_list, dma_addr);
drivers/nvme/host/apple.c
385
dma_addr = next_dma_addr;
drivers/nvme/host/apple.c
433
u64 dma_addr = sg_dma_address(sg);
drivers/nvme/host/apple.c
434
int offset = dma_addr & (NVME_CTRL_PAGE_SIZE - 1);
drivers/nvme/host/apple.c
448
dma_addr += (NVME_CTRL_PAGE_SIZE - offset);
drivers/nvme/host/apple.c
451
dma_addr = sg_dma_address(sg);
drivers/nvme/host/apple.c
456
iod->first_dma = dma_addr;
drivers/nvme/host/apple.c
471
iod->first_dma = dma_addr;
drivers/nvme/host/apple.c
490
prp_list[i++] = cpu_to_le64(dma_addr);
drivers/nvme/host/apple.c
492
dma_addr += NVME_CTRL_PAGE_SIZE;
drivers/nvme/host/apple.c
501
dma_addr = sg_dma_address(sg);
drivers/nvme/host/fc.c
921
fc_dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
drivers/nvme/host/fc.c
923
return dev ? dma_mapping_error(dev, dma_addr) : 0;
drivers/nvme/host/pci.c
1127
dma_addr_t dma_addr, int entries)
drivers/nvme/host/pci.c
1129
sge->addr = cpu_to_le64(dma_addr);
drivers/nvme/host/pci.c
1185
dma_addr_t dma_addr;
drivers/nvme/host/pci.c
1192
dma_addr = dma_map_bvec(nvmeq->dev->dev, &bv, rq_dma_dir(req), 0);
drivers/nvme/host/pci.c
1193
if (dma_mapping_error(nvmeq->dev->dev, dma_addr))
drivers/nvme/host/pci.c
1200
iod->cmd.common.dptr.sgl.addr = cpu_to_le64(dma_addr);
drivers/nvme/host/pci.c
1206
iod->cmd.common.dptr.prp1 = cpu_to_le64(dma_addr);
drivers/nvme/host/pci.c
1210
cpu_to_le64(dma_addr + first_prp_len);
drivers/nvme/host/pci.c
2493
u64 dma_addr = dev->host_mem_descs_dma;
drivers/nvme/host/pci.c
2501
c.features.dword13 = cpu_to_le32(lower_32_bits(dma_addr));
drivers/nvme/host/pci.c
2502
c.features.dword14 = cpu_to_le32(upper_32_bits(dma_addr));
drivers/nvme/host/pci.c
2602
dma_addr_t dma_addr;
drivers/nvme/host/pci.c
2605
bufs[i] = dma_alloc_attrs(dev->dev, len, &dma_addr, GFP_KERNEL,
drivers/nvme/host/pci.c
2610
descs[i].addr = cpu_to_le64(dma_addr);
drivers/nvme/host/pci.c
838
dma_addr_t dma_addr = nvme_pci_first_desc_dma_addr(&iod->cmd);
drivers/nvme/host/pci.c
843
dma_addr);
drivers/nvme/host/pci.c
852
dma_addr);
drivers/nvme/host/pci.c
853
dma_addr = next_dma_addr;
drivers/nvme/target/fc.c
280
fc_dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
drivers/nvme/target/fc.c
282
return dev ? dma_mapping_error(dev, dma_addr) : 0;
drivers/nvme/target/pci-epf.c
370
dma_addr_t dma_addr;
drivers/nvme/target/pci-epf.c
394
dma_addr = dma_map_single(dma_dev, seg->buf, seg->length, dir);
drivers/nvme/target/pci-epf.c
395
ret = dma_mapping_error(dma_dev, dma_addr);
drivers/nvme/target/pci-epf.c
405
desc = dmaengine_prep_slave_single(chan, dma_addr, seg->length,
drivers/nvme/target/pci-epf.c
428
dma_unmap_single(dma_dev, dma_addr, seg->length, dir);
drivers/nvmem/zynqmp_nvmem.c
122
zynqmp_pm_efuse_access(dma_addr, (u32 *)&ret);
drivers/nvmem/zynqmp_nvmem.c
141
efuse, dma_addr);
drivers/nvmem/zynqmp_nvmem.c
65
dma_addr_t dma_addr;
drivers/nvmem/zynqmp_nvmem.c
99
&dma_addr, GFP_KERNEL);
drivers/of/unittest.c
1136
dma_addr_t dma_addr;
drivers/of/unittest.c
1148
dma_addr = phys_to_dma(dev_bogus, expect_paddr);
drivers/of/unittest.c
1153
unittest(dma_addr == expect_dma_addr,
drivers/of/unittest.c
1155
&dma_addr, expect_dma_addr, np);
drivers/parport/parport_pc.c
589
dma_addr_t dma_addr, dma_handle;
drivers/parport/parport_pc.c
600
dma_addr = dma_handle = dma_map_single(dev, (void *)buf, length,
drivers/parport/parport_pc.c
606
dma_addr = priv->dma_handle;
drivers/parport/parport_pc.c
635
set_dma_addr(port->dma, dma_addr);
drivers/parport/parport_pc.c
651
dma_addr += count;
drivers/parport/parport_pc.c
689
dma_addr -= count;
drivers/pci/controller/pcie-rzg3s-host.c
190
dma_addr_t dma_addr;
drivers/pci/controller/pcie-rzg3s-host.c
698
msi->dma_addr = dma_map_single(dev, (void *)msi->pages, size * 2,
drivers/pci/controller/pcie-rzg3s-host.c
700
if (dma_mapping_error(dev, msi->dma_addr)) {
drivers/pci/controller/pcie-rzg3s-host.c
727
if (msi->dma_addr < base || msi->dma_addr > base + mask)
drivers/pci/controller/pcie-rzg3s-host.c
739
msi->window_base = ALIGN(msi->dma_addr, size);
drivers/pci/controller/pcie-rzg3s-host.c
740
if (msi->window_base < msi->dma_addr) {
drivers/pci/controller/pcie-rzg3s-host.c
750
dma_unmap_single(dev, msi->dma_addr, size * 2, DMA_BIDIRECTIONAL);
drivers/pci/controller/pcie-rzg3s-host.c
787
dma_unmap_single(host->dev, msi->dma_addr, size * 2, DMA_BIDIRECTIONAL);
drivers/pci/switch/switchtec.c
1303
writeq(stdev->dma_mrpc_dma_addr, &stdev->mmio_mrpc->dma_addr);
drivers/pci/switch/switchtec.c
1647
writeq(0, &stdev->mmio_mrpc->dma_addr);
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
609
if ((bulk->dma_addr != (dma_addr_t)(uintptr_t)bulk_params->dma_addr) ||
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1382
if (bulk->dma_addr && service->instance) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1470
pagelistinfo->pagelist, pagelistinfo->dma_addr);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1508
dma_addr_t dma_addr;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1539
pagelist = dma_alloc_coherent(instance->state->dev, pagelist_size, &dma_addr,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1560
pagelistinfo->dma_addr = dma_addr;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1764
bulk->dma_addr = pagelistinfo->dma_addr;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1820
bulk->dma_addr = 0;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2125
localport, bulk->actual, &bulk->dma_addr);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3096
dir_char, bulk->size, &bulk->dma_addr, bulk->cb_data);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3110
payload[0] = lower_32_bits(bulk->dma_addr);
drivers/platform/x86/amd/pmc/mp2_stb.c
105
mp2->vslbase = dmam_alloc_coherent(dev, len, &mp2->dma_addr, GFP_KERNEL);
drivers/platform/x86/amd/pmc/mp2_stb.c
95
writeq(mp2->dma_addr, mp2->mmio + AMD_C2P_MSG1);
drivers/platform/x86/amd/pmc/pmc.h
81
dma_addr_t dma_addr;
drivers/remoteproc/mtk_common.h
152
dma_addr_t dma_addr;
drivers/remoteproc/mtk_scp.c
1046
&scp->dma_addr, GFP_KERNEL);
drivers/remoteproc/mtk_scp.c
1062
scp->dma_addr);
drivers/remoteproc/mtk_scp.c
759
offset = da - scp->dma_addr;
drivers/remoteproc/mtk_scp.c
789
da >= scp->dma_addr &&
drivers/remoteproc/mtk_scp.c
790
(da + len) <= scp->dma_addr + scp_sizes->max_dram_size) {
drivers/remoteproc/mtk_scp.c
791
offset = da - scp->dma_addr;
drivers/remoteproc/xlnx_r5_remoteproc.c
542
dma_addr_t dma_addr;
drivers/remoteproc/xlnx_r5_remoteproc.c
549
dma_addr = (dma_addr_t)sram->sram_res.start;
drivers/remoteproc/xlnx_r5_remoteproc.c
555
dma_addr,
drivers/remoteproc/xlnx_r5_remoteproc.c
570
sram->sram_res.name, dma_addr, da, len);
drivers/s390/cio/css.c
1070
dma_addr_t dma_addr;
drivers/s390/cio/css.c
1077
cpu_addr = dma_alloc_coherent(dma_dev, PAGE_SIZE, &dma_addr,
drivers/s390/cio/css.c
1082
dma_addr, PAGE_SIZE, -1);
drivers/s390/cio/css.c
1118
dma_addr_t dma_addr;
drivers/s390/cio/css.c
1124
addr = gen_pool_dma_alloc(gp_dma, size, &dma_addr);
drivers/s390/cio/css.c
1127
addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP);
drivers/s390/cio/css.c
1130
gen_pool_add_virt(gp_dma, (unsigned long)addr, dma_addr, chunk_size, -1);
drivers/s390/cio/css.c
1131
addr = gen_pool_dma_alloc(gp_dma, size, dma_handle ? &dma_addr : NULL);
drivers/s390/cio/css.c
1134
*dma_handle = (__force dma32_t)dma_addr;
drivers/s390/net/ism_drv.c
225
dma_unmap_page(&ism->pdev->dev, dmb->dma_addr, dmb->dmb_len,
drivers/s390/net/ism_drv.c
260
dmb->dma_addr = dma_map_page(&ism->pdev->dev,
drivers/s390/net/ism_drv.c
263
if (dma_mapping_error(&ism->pdev->dev, dmb->dma_addr)) {
drivers/s390/net/ism_drv.c
293
cmd.request.dmb = dmb->dma_addr;
drivers/scsi/aic94xx/aic94xx_hwi.c
487
unsigned long dma_addr = OCM_BASE_ADDR;
drivers/scsi/aic94xx/aic94xx_hwi.c
490
dma_addr -= asd_ha->hw_prof.max_ddbs * ASD_DDB_SIZE;
drivers/scsi/aic94xx/aic94xx_hwi.c
491
asd_write_reg_addr(asd_ha, DEVCTXBASE, (dma_addr_t) dma_addr);
drivers/scsi/aic94xx/aic94xx_hwi.c
501
unsigned long dma_addr;
drivers/scsi/aic94xx/aic94xx_hwi.c
523
dma_addr = ALIGN((unsigned long) dma_handle, ASD_DDB_SIZE);
drivers/scsi/aic94xx/aic94xx_hwi.c
524
dma_addr -= asd_ha->hw_prof.max_ddbs * ASD_DDB_SIZE;
drivers/scsi/aic94xx/aic94xx_hwi.c
525
dma_handle = (dma_addr_t) dma_addr;
drivers/scsi/aic94xx/aic94xx_hwi.c
539
unsigned long dma_addr;
drivers/scsi/aic94xx/aic94xx_hwi.c
559
dma_addr = ALIGN((unsigned long) dma_handle, ASD_SCB_SIZE);
drivers/scsi/aic94xx/aic94xx_hwi.c
560
dma_addr -= asd_ha->hw_prof.max_scbs * ASD_SCB_SIZE;
drivers/scsi/aic94xx/aic94xx_hwi.c
561
dma_handle = (dma_addr_t) dma_addr;
drivers/scsi/am53c974.c
243
static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len)
drivers/scsi/am53c974.c
261
base = dma_addr & ((1U << 24) - 1U);
drivers/scsi/atari_scsi.c
162
unsigned long addr = SCSI_DMA_READ_P(dma_addr), end_addr;
drivers/scsi/atari_scsi.c
198
SCSI_DMA_READ_P(dma_addr));
drivers/scsi/atari_scsi.c
214
(SCSI_DMA_READ_P(dma_addr) - atari_dma_startaddr);
drivers/scsi/atari_scsi.c
337
phys_dst = SCSI_DMA_READ_P(dma_addr);
drivers/scsi/atari_scsi.c
458
SCSI_DMA_WRITE_P(dma_addr, addr);
drivers/scsi/bfa/bfa_ioc.c
3269
memcpy(ablk->cfg, ablk->dma_addr.kva,
drivers/scsi/bfa/bfa_ioc.c
3345
ablk->dma_addr.kva = dma_kva;
drivers/scsi/bfa/bfa_ioc.c
3346
ablk->dma_addr.pa = dma_pa;
drivers/scsi/bfa/bfa_ioc.c
3386
bfa_dma_be_addr_set(m->addr, ablk->dma_addr.pa);
drivers/scsi/bfa/bfa_ioc.h
175
#define bfa_dma_be_addr_set(dma_addr, pa) \
drivers/scsi/bfa/bfa_ioc.h
176
__bfa_dma_be_addr_set(&dma_addr, (u64)pa)
drivers/scsi/bfa/bfa_ioc.h
178
__bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa)
drivers/scsi/bfa/bfa_ioc.h
180
dma_addr->a32.addr_lo = cpu_to_be32(pa);
drivers/scsi/bfa/bfa_ioc.h
181
dma_addr->a32.addr_hi = cpu_to_be32(pa >> 32);
drivers/scsi/bfa/bfa_ioc.h
431
struct bfa_dma_s dma_addr;
drivers/scsi/bfa/bfa_port.c
352
bfa_dma_be_addr_set(m->dma_addr, port->stats_dma.pa);
drivers/scsi/bfa/bfa_port.c
667
bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa);
drivers/scsi/bfa/bfa_port.c
707
bfa_dma_be_addr_set(cmd->dma_addr, cee->stats_dma.pa);
drivers/scsi/bfa/bfi.h
661
union bfi_addr_u dma_addr;
drivers/scsi/bfa/bfi.h
785
union bfi_addr_u dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
791
dma_addr_t dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
802
dma_addr = ep->qp.sq_pgtbl_phys;
drivers/scsi/bnx2i/bnx2i_hwi.c
803
ofld_req1.sq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
804
ofld_req1.sq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/bnx2i/bnx2i_hwi.c
806
dma_addr = ep->qp.cq_pgtbl_phys;
drivers/scsi/bnx2i/bnx2i_hwi.c
807
ofld_req1.cq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
808
ofld_req1.cq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/bnx2i/bnx2i_hwi.c
814
dma_addr = ep->qp.rq_pgtbl_phys;
drivers/scsi/bnx2i/bnx2i_hwi.c
815
ofld_req2.rq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
816
ofld_req2.rq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/bnx2i/bnx2i_hwi.c
852
dma_addr_t dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
863
dma_addr = ep->qp.sq_pgtbl_phys + ISCSI_SQ_DB_SIZE;
drivers/scsi/bnx2i/bnx2i_hwi.c
864
ofld_req1.sq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
865
ofld_req1.sq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/bnx2i/bnx2i_hwi.c
867
dma_addr = ep->qp.cq_pgtbl_phys + ISCSI_CQ_DB_SIZE;
drivers/scsi/bnx2i/bnx2i_hwi.c
868
ofld_req1.cq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
869
ofld_req1.cq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/bnx2i/bnx2i_hwi.c
875
dma_addr = ep->qp.rq_pgtbl_phys + ISCSI_RQ_DB_SIZE;
drivers/scsi/bnx2i/bnx2i_hwi.c
876
ofld_req2.rq_page_table_addr_lo = (u32) dma_addr;
drivers/scsi/bnx2i/bnx2i_hwi.c
877
ofld_req2.rq_page_table_addr_hi = (u32) ((u64) dma_addr >> 32);
drivers/scsi/esas2r/esas2r_ioctl.c
1891
dma_addr_t dma_addr;
drivers/scsi/esas2r/esas2r_ioctl.c
1895
&dma_addr,
drivers/scsi/esas2r/esas2r_ioctl.c
1898
a->ppvda_buffer = dma_addr;
drivers/scsi/esp_scsi.c
1741
dma_addr_t dma_addr = esp_cur_dma_addr(ent, cmd);
drivers/scsi/esp_scsi.c
1753
dma_len = esp->ops->dma_length_limit(esp, dma_addr,
drivers/scsi/esp_scsi.c
1756
dma_len = esp_dma_length_limit(esp, dma_addr, dma_len);
drivers/scsi/esp_scsi.c
1772
(unsigned long long)dma_addr, dma_len, write);
drivers/scsi/esp_scsi.c
1774
esp->ops->send_dma_cmd(esp, dma_addr, dma_len, dma_len,
drivers/scsi/esp_scsi.c
522
static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len)
drivers/scsi/esp_scsi.c
543
base = dma_addr & ((1U << 24) - 1U);
drivers/scsi/esp_scsi.h
384
u32 (*dma_length_limit)(struct esp *esp, u32 dma_addr,
drivers/scsi/esp_scsi.h
411
void (*send_dma_cmd)(struct esp *esp, u32 dma_addr, u32 esp_count,
drivers/scsi/esp_scsi.h
582
extern void esp_send_pio_cmd(struct esp *esp, u32 dma_addr, u32 esp_count,
drivers/scsi/fnic/fnic_fcs.c
449
dma_unmap_single(&fnic->pdev->dev, buf->dma_addr, buf->len,
drivers/scsi/fnic/fnic_fcs.c
621
dma_unmap_single(&fnic->pdev->dev, buf->dma_addr, buf->len,
drivers/scsi/fnic/fnic_fcs.c
856
dma_unmap_single(&fnic->pdev->dev, buf->dma_addr, buf->len,
drivers/scsi/fnic/fnic_fcs.c
898
dma_unmap_single(&fnic->pdev->dev, buf->dma_addr, buf->len,
drivers/scsi/fnic/fnic_res.h
19
void *os_buf, dma_addr_t dma_addr,
drivers/scsi/fnic/fnic_res.h
212
void *os_buf, dma_addr_t dma_addr,
drivers/scsi/fnic/fnic_res.h
218
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/scsi/fnic/fnic_res.h
222
vnic_rq_post(rq, os_buf, 0, dma_addr, len);
drivers/scsi/fnic/fnic_res.h
28
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/scsi/fnic/fnic_res.h
39
vnic_wq_post(wq, os_buf, dma_addr, len, sop, eop);
drivers/scsi/fnic/fnic_res.h
43
void *os_buf, dma_addr_t dma_addr,
drivers/scsi/fnic/fnic_res.h
52
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/scsi/fnic/fnic_res.h
64
vnic_wq_post(wq, os_buf, dma_addr, len, 1, 1);
drivers/scsi/fnic/vnic_rq.h
122
dma_addr_t dma_addr, unsigned int len)
drivers/scsi/fnic/vnic_rq.h
128
buf->dma_addr = dma_addr;
drivers/scsi/fnic/vnic_rq.h
72
dma_addr_t dma_addr;
drivers/scsi/fnic/vnic_wq.h
104
void *os_buf, dma_addr_t dma_addr,
drivers/scsi/fnic/vnic_wq.h
111
buf->dma_addr = dma_addr;
drivers/scsi/fnic/vnic_wq.h
59
dma_addr_t dma_addr;
drivers/scsi/initio.c
2537
dma_addr_t dma_addr;
drivers/scsi/initio.c
2551
dma_addr = dma_map_single(&host->pci_dev->dev, cmnd->sense_buffer,
drivers/scsi/initio.c
2553
cblk->senseptr = (u32)dma_addr;
drivers/scsi/initio.c
2555
initio_priv(cmnd)->sense_dma_addr = dma_addr;
drivers/scsi/initio.c
2575
dma_addr = dma_map_single(&host->pci_dev->dev, &cblk->sglist[0],
drivers/scsi/initio.c
2578
cblk->bufptr = (u32)dma_addr;
drivers/scsi/initio.c
2579
initio_priv(cmnd)->sglist_dma_addr = dma_addr;
drivers/scsi/ipr.c
5707
ioarcb->write_ioadl_addr = cpu_to_be32((ipr_cmd->dma_addr) +
drivers/scsi/ipr.c
5790
dma_addr_t dma_addr = ipr_cmd->dma_addr;
drivers/scsi/ipr.c
5802
cpu_to_be64(dma_addr + offsetof(struct ipr_cmnd, i.ioadl64));
drivers/scsi/ipr.c
5805
cpu_to_be32(dma_addr + offsetof(struct ipr_cmnd, i.ioadl));
drivers/scsi/ipr.c
631
dma_addr_t dma_addr = ipr_cmd->dma_addr;
drivers/scsi/ipr.c
644
cpu_to_be64(dma_addr + offsetof(struct ipr_cmnd, i.ioadl64));
drivers/scsi/ipr.c
647
cpu_to_be32(dma_addr + offsetof(struct ipr_cmnd, i.ioadl));
drivers/scsi/ipr.c
6748
dma_addr_t dma_addr, u8 xfer_len)
drivers/scsi/ipr.c
6759
ipr_init_ioadl(ipr_cmd, dma_addr, xfer_len, IPR_IOADL_FLAGS_WRITE_LAST);
drivers/scsi/ipr.c
6811
u8 parm, dma_addr_t dma_addr, u8 xfer_len)
drivers/scsi/ipr.c
6821
ipr_init_ioadl(ipr_cmd, dma_addr, xfer_len, IPR_IOADL_FLAGS_READ_LAST);
drivers/scsi/ipr.c
7196
dma_addr_t dma_addr, u8 xfer_len)
drivers/scsi/ipr.c
7209
ipr_init_ioadl(ipr_cmd, dma_addr, xfer_len, IPR_IOADL_FLAGS_READ_LAST);
drivers/scsi/ipr.c
8278
cmd_pkt->cdb[10] = ((u64) hcam_cmd->dma_addr >> 56) & 0xff;
drivers/scsi/ipr.c
8279
cmd_pkt->cdb[11] = ((u64) hcam_cmd->dma_addr >> 48) & 0xff;
drivers/scsi/ipr.c
8280
cmd_pkt->cdb[12] = ((u64) hcam_cmd->dma_addr >> 40) & 0xff;
drivers/scsi/ipr.c
8281
cmd_pkt->cdb[13] = ((u64) hcam_cmd->dma_addr >> 32) & 0xff;
drivers/scsi/ipr.c
8282
cmd_pkt->cdb[2] = ((u64) hcam_cmd->dma_addr >> 24) & 0xff;
drivers/scsi/ipr.c
8283
cmd_pkt->cdb[3] = ((u64) hcam_cmd->dma_addr >> 16) & 0xff;
drivers/scsi/ipr.c
8284
cmd_pkt->cdb[4] = ((u64) hcam_cmd->dma_addr >> 8) & 0xff;
drivers/scsi/ipr.c
8285
cmd_pkt->cdb[5] = ((u64) hcam_cmd->dma_addr) & 0xff;
drivers/scsi/ipr.c
8853
dma_addr_t dma_addr;
drivers/scsi/ipr.c
8908
GFP_KERNEL, &dma_addr);
drivers/scsi/ipr.c
8916
ioa_cfg->ipr_cmnd_list_dma[i] = dma_addr;
drivers/scsi/ipr.c
8919
ipr_cmd->dma_addr = dma_addr;
drivers/scsi/ipr.c
8921
ioarcb->a.ioarcb_host_pci_addr64 = cpu_to_be64(dma_addr);
drivers/scsi/ipr.c
8923
ioarcb->a.ioarcb_host_pci_addr = cpu_to_be32(dma_addr);
drivers/scsi/ipr.c
8928
cpu_to_be64(dma_addr + offsetof(struct ipr_cmnd, i.ioadl64));
drivers/scsi/ipr.c
8930
cpu_to_be64(dma_addr + offsetof(struct ipr_cmnd, s.ioasa64));
drivers/scsi/ipr.c
8933
cpu_to_be32(dma_addr + offsetof(struct ipr_cmnd, i.ioadl));
drivers/scsi/ipr.c
8936
cpu_to_be32(dma_addr + offsetof(struct ipr_cmnd, s.ioasa));
drivers/scsi/ipr.c
8941
ipr_cmd->sense_buffer_dma = dma_addr +
drivers/scsi/ipr.c
898
dma_addr_t send_dma_addr = ipr_cmd->dma_addr;
drivers/scsi/ipr.c
975
static void ipr_init_ioadl(struct ipr_cmnd *ipr_cmd, dma_addr_t dma_addr,
drivers/scsi/ipr.c
986
ioadl64->address = cpu_to_be64(dma_addr);
drivers/scsi/ipr.c
993
ioadl->address = cpu_to_be32(dma_addr);
drivers/scsi/ipr.h
1563
dma_addr_t dma_addr;
drivers/scsi/isci/request.c
118
dma_addr_t dma_addr;
drivers/scsi/isci/request.c
137
dma_addr = to_sgl_element_pair_dma(ihost,
drivers/scsi/isci/request.c
142
upper_32_bits(dma_addr);
drivers/scsi/isci/request.c
144
lower_32_bits(dma_addr);
drivers/scsi/isci/request.c
153
dma_addr = dma_map_single(&ihost->pdev->dev,
drivers/scsi/isci/request.c
158
ireq->zero_scatter_daddr = dma_addr;
drivers/scsi/isci/request.c
161
scu_sg->A.address_upper = upper_32_bits(dma_addr);
drivers/scsi/isci/request.c
162
scu_sg->A.address_lower = lower_32_bits(dma_addr);
drivers/scsi/isci/request.c
217
dma_addr_t dma_addr;
drivers/scsi/isci/request.c
260
dma_addr = sci_io_request_get_dma_addr(ireq, &ireq->ssp.cmd);
drivers/scsi/isci/request.c
262
task_context->command_iu_upper = upper_32_bits(dma_addr);
drivers/scsi/isci/request.c
263
task_context->command_iu_lower = lower_32_bits(dma_addr);
drivers/scsi/isci/request.c
269
dma_addr = sci_io_request_get_dma_addr(ireq, &ireq->ssp.rsp);
drivers/scsi/isci/request.c
271
task_context->response_iu_upper = upper_32_bits(dma_addr);
drivers/scsi/isci/request.c
272
task_context->response_iu_lower = lower_32_bits(dma_addr);
drivers/scsi/isci/request.c
497
dma_addr_t dma_addr;
drivers/scsi/isci/request.c
544
dma_addr = sci_io_request_get_dma_addr(ireq,
drivers/scsi/isci/request.c
548
task_context->command_iu_upper = upper_32_bits(dma_addr);
drivers/scsi/isci/request.c
549
task_context->command_iu_lower = lower_32_bits(dma_addr);
drivers/scsi/lpfc/lpfc_bsg.c
1008
dma_addr = getPaddr(bde->addrHigh,
drivers/scsi/lpfc/lpfc_bsg.c
1011
pring, dma_addr);
drivers/scsi/lpfc/lpfc_bsg.c
915
dma_addr_t dma_addr;
drivers/scsi/lpfc/lpfc_ct.c
345
dma_addr_t dma_addr;
drivers/scsi/lpfc/lpfc_ct.c
438
dma_addr = getPaddr(icmd->un.cont64[i].addrHigh,
drivers/scsi/lpfc/lpfc_ct.c
441
dma_addr);
drivers/scsi/mac53c94.c
363
dma_addr_t dma_addr;
drivers/scsi/mac53c94.c
377
dma_addr = sg_dma_address(scl);
drivers/scsi/mac53c94.c
384
dcmds->phy_addr = cpu_to_le32(dma_addr);
drivers/scsi/mac53c94.c
58
dma_addr_t dma_addr;
drivers/scsi/mac_esp.c
247
static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len)
drivers/scsi/megaraid/mega_common.h
281
dma_addr_t dma_addr;
drivers/scsi/megaraid/megaraid_mbox.c
1049
ccb->mbox_dma_h = (unsigned long)mbox_pci_blk[i].dma_addr + 16;
drivers/scsi/megaraid/megaraid_mbox.c
1061
ccb->epthru_dma_h = epthru_pci_blk[i].dma_addr;
drivers/scsi/megaraid/megaraid_mbox.c
1067
ccb->sgl_dma_h = sg_pci_blk[i].dma_addr;
drivers/scsi/megaraid/megaraid_mbox.c
1160
&mbox_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_mbox.c
1186
&epthru_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_mbox.c
1209
&sg_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_mbox.c
1243
sg_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_mbox.c
1251
epthru_pci_blk[i].vaddr, epthru_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_mbox.c
1259
mbox_pci_blk[i].vaddr, mbox_pci_blk[i].dma_addr);
drivers/scsi/megaraid/megaraid_sas.h
2737
dma_addr_t dma_addr, u32 dma_len);
drivers/scsi/megaraid/megaraid_sas_base.c
288
dma_addr_t dma_addr, u32 dma_len)
drivers/scsi/megaraid/megaraid_sas_base.c
291
dcmd->sgl.sge64[0].phys_addr = cpu_to_le64(dma_addr);
drivers/scsi/megaraid/megaraid_sas_base.c
297
cpu_to_le32(lower_32_bits(dma_addr));
drivers/scsi/mesh.c
1277
u32 dma_addr = sg_dma_address(scl);
drivers/scsi/mesh.c
1289
dcmds->phy_addr = cpu_to_le32(dma_addr + off);
drivers/scsi/mpi3mr/mpi3mr.h
1468
dma_addr_t dma_addr);
drivers/scsi/mpi3mr/mpi3mr.h
953
dma_addr_t dma_addr;
drivers/scsi/mpi3mr/mpi3mr.h
970
dma_addr_t dma_addr;
drivers/scsi/mpi3mr/mpi3mr.h
983
dma_addr_t dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
1855
drv_buf_iter->dma_desc[0].dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
1898
drv_buf_iter->dma_desc[i].dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
1914
mrioc->ioctl_chain_sge.dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
1991
sgl_dma = cpu_to_le64(mrioc->ioctl_chain_sge.dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
201
&diag_buffer->dma_addr,
drivers/scsi/mpi3mr/mpi3mr_app.c
2015
sgl_dma = cpu_to_le64(drv_buf_iter->dma_desc[i].dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
2057
dma_addr_t prp_entry_dma, prp_page_dma, dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
2119
dma_addr = drv_buf_iter->dma_desc[count].dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
2120
if (dma_addr & page_mask) {
drivers/scsi/mpi3mr/mpi3mr_app.c
2123
__func__, &dma_addr, dev_pgsz);
drivers/scsi/mpi3mr/mpi3mr_app.c
2128
dma_addr = drv_buf_iter->dma_desc[0].dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
2185
offset = dma_addr & page_mask;
drivers/scsi/mpi3mr/mpi3mr_app.c
2193
*prp1_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
2242
*prp2_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
2260
*prp_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
2278
dma_addr += entry_len;
drivers/scsi/mpi3mr/mpi3mr_app.c
2289
dma_addr =
drivers/scsi/mpi3mr/mpi3mr_app.c
2290
drv_buf_iter->dma_desc[desc_count].dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
2344
drv_buf->dma_desc[i].dma_addr =
drivers/scsi/mpi3mr/mpi3mr_app.c
2345
mrioc->ioctl_sge[desc_count].dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
259
diag_buf_post_req.address = le64_to_cpu(diag_buffer->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_app.c
2624
mrioc->ioctl_chain_sge.dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
2639
mrioc->ioctl_resp_sge.dma_addr;
drivers/scsi/mpi3mr/mpi3mr_app.c
69
diag_buffer->dma_addr = diag_buffer_list_dma;
drivers/scsi/mpi3mr/mpi3mr_app.c
77
trace_size, &diag_buffer->dma_addr, GFP_KERNEL);
drivers/scsi/mpi3mr/mpi3mr_fw.c
107
dma_addr_t dma_addr)
drivers/scsi/mpi3mr/mpi3mr_fw.c
113
sgel->address = cpu_to_le64(dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1192
mem_desc->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1202
mem_desc->addr, mem_desc->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1208
mem_desc->addr, mem_desc->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1245
&mem_desc->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1254
&mem_desc->dma_addr,
drivers/scsi/mpi3mr/mpi3mr_fw.c
1263
&mem_desc->dma_addr,
drivers/scsi/mpi3mr/mpi3mr_fw.c
3882
&mrioc->chain_sgl_list[i].dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
4973
mrioc->chain_sgl_list[i].dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
5026
diag_buffer->dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
5863
mem_desc.size, &mem_desc.dma_addr, GFP_KERNEL);
drivers/scsi/mpi3mr/mpi3mr_fw.c
5869
mem_desc.dma_addr);
drivers/scsi/mpi3mr/mpi3mr_fw.c
5897
mem_desc.addr, mem_desc.dma_addr);
drivers/scsi/mpi3mr/mpi3mr_os.c
3800
chain_dma = chain_req->dma_addr;
drivers/scsi/mpi3mr/mpi3mr_transport.c
3202
dma_addr_t *dma_addr, size_t *dma_len, void **p)
drivers/scsi/mpi3mr/mpi3mr_transport.c
3206
*p = dma_alloc_coherent(dev, buf->payload_len, dma_addr,
drivers/scsi/mpi3mr/mpi3mr_transport.c
3214
*dma_addr = sg_dma_address(buf->sg_list);
drivers/scsi/mpi3mr/mpi3mr_transport.c
3233
dma_addr_t dma_addr, void *p)
drivers/scsi/mpi3mr/mpi3mr_transport.c
3236
dma_free_coherent(dev, buf->payload_len, p, dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2115
_base_add_sg_single_32(void *paddr, u32 flags_length, dma_addr_t dma_addr)
drivers/scsi/mpt3sas/mpt3sas_base.c
2122
sgel->Address = cpu_to_le32(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2133
_base_add_sg_single_64(void *paddr, u32 flags_length, dma_addr_t dma_addr)
drivers/scsi/mpt3sas/mpt3sas_base.c
2140
sgel->Address = cpu_to_le64(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2292
dma_addr_t prp_entry_dma, prp_page_dma, dma_addr;
drivers/scsi/mpt3sas/mpt3sas_base.c
2335
dma_addr = data_in_dma;
drivers/scsi/mpt3sas/mpt3sas_base.c
2338
dma_addr = data_out_dma;
drivers/scsi/mpt3sas/mpt3sas_base.c
2368
offset = dma_addr & page_mask;
drivers/scsi/mpt3sas/mpt3sas_base.c
2376
*prp1_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2408
*prp2_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2418
*prp_entry = cpu_to_le64(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.c
2427
dma_addr += entry_len;
drivers/scsi/mpt3sas/mpt3sas_base.c
2656
dma_addr_t dma_addr)
drivers/scsi/mpt3sas/mpt3sas_base.c
2663
sgel->Address = cpu_to_le64(dma_addr);
drivers/scsi/mpt3sas/mpt3sas_base.h
980
typedef void (*MPT_ADD_SGE)(void *paddr, u32 flags_length, dma_addr_t dma_addr);
drivers/scsi/mpt3sas/mpt3sas_transport.c
2007
dma_addr_t *dma_addr, size_t *dma_len, void **p)
drivers/scsi/mpt3sas/mpt3sas_transport.c
2011
*p = dma_alloc_coherent(dev, buf->payload_len, dma_addr,
drivers/scsi/mpt3sas/mpt3sas_transport.c
2019
*dma_addr = sg_dma_address(buf->sg_list);
drivers/scsi/mpt3sas/mpt3sas_transport.c
2029
dma_addr_t dma_addr, void *p)
drivers/scsi/mpt3sas/mpt3sas_transport.c
2032
dma_free_coherent(dev, buf->payload_len, p, dma_addr);
drivers/scsi/myrb.c
1320
dcdb->dma_addr = sg_dma_address(sgl);
drivers/scsi/myrb.h
550
u32 dma_addr; /* Bytes 4-7 */
drivers/scsi/myrs.c
1667
hw_sge = &mbox->SCSI_10.dma_addr;
drivers/scsi/myrs.c
1714
hw_sge = &mbox->SCSI_255.dma_addr;
drivers/scsi/myrs.c
186
sgl = &mbox->ctlr_info.dma_addr;
drivers/scsi/myrs.c
245
sgl = &mbox->ldev_info.dma_addr;
drivers/scsi/myrs.c
333
sgl = &mbox->pdev_info.dma_addr;
drivers/scsi/myrs.c
404
sgl = &mbox->pdev_info.dma_addr;
drivers/scsi/myrs.c
440
sgl = &mbox->get_event.dma_addr;
drivers/scsi/myrs.c
468
sgl = &mbox->common.dma_addr;
drivers/scsi/myrs.h
699
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
712
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
726
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
741
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
755
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
769
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
785
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
803
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
820
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/myrs.h
856
union myrs_sgl dma_addr; /* Bytes 32-63 */
drivers/scsi/pm8001/pm8001_hwi.c
4012
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm8001_hwi.c
4013
ssp_cmd.addr_low = cpu_to_le32(lower_32_bits(dma_addr));
drivers/scsi/pm8001/pm8001_hwi.c
4014
ssp_cmd.addr_high = cpu_to_le32(upper_32_bits(dma_addr));
drivers/scsi/pm8001/pm8001_hwi.c
4085
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm8001_hwi.c
4086
sata_cmd.addr_low = lower_32_bits(dma_addr);
drivers/scsi/pm8001/pm8001_hwi.c
4087
sata_cmd.addr_high = upper_32_bits(dma_addr);
drivers/scsi/pm8001/pm80xx_hwi.c
4389
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm80xx_hwi.c
4392
cpu_to_le32(lower_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4394
cpu_to_le32(upper_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4399
end_addr = dma_addr + le32_to_cpu(ssp_cmd.enc_len) - 1;
drivers/scsi/pm8001/pm80xx_hwi.c
4406
dma_addr,
drivers/scsi/pm8001/pm80xx_hwi.c
4448
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm80xx_hwi.c
4450
ssp_cmd.addr_low = cpu_to_le32(lower_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4452
cpu_to_le32(upper_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4457
end_addr = dma_addr + le32_to_cpu(ssp_cmd.len) - 1;
drivers/scsi/pm8001/pm80xx_hwi.c
4463
dma_addr,
drivers/scsi/pm8001/pm80xx_hwi.c
4559
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm80xx_hwi.c
4562
cpu_to_le32(lower_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4564
cpu_to_le32(upper_32_bits(dma_addr));
drivers/scsi/pm8001/pm80xx_hwi.c
4569
end_addr = dma_addr + le32_to_cpu(sata_cmd.enc_len) - 1;
drivers/scsi/pm8001/pm80xx_hwi.c
4575
dma_addr,
drivers/scsi/pm8001/pm80xx_hwi.c
4623
u64 dma_addr = sg_dma_address(task->scatter);
drivers/scsi/pm8001/pm80xx_hwi.c
4625
sata_cmd.addr_low = lower_32_bits(dma_addr);
drivers/scsi/pm8001/pm80xx_hwi.c
4626
sata_cmd.addr_high = upper_32_bits(dma_addr);
drivers/scsi/pm8001/pm80xx_hwi.c
4631
end_addr = dma_addr + le32_to_cpu(sata_cmd.len) - 1;
drivers/scsi/pm8001/pm80xx_hwi.c
4637
dma_addr,
drivers/scsi/pmcraid.c
300
dma_addr_t dma_addr = cmd->ioa_cb_bus_addr;
drivers/scsi/pmcraid.c
309
ioarcb->ioarcb_bus_addr = cpu_to_le64(dma_addr);
drivers/scsi/pmcraid.c
310
ioarcb->ioasa_bus_addr = cpu_to_le64(dma_addr + ioasa_offset);
drivers/scsi/qla2xxx/qla_def.h
5186
dma_addr_t dma_addr; /* OUT */
drivers/scsi/qla2xxx/qla_iocb.c
809
sgx->dma_addr = sg_dma_addr + sgx->bytes_consumed;
drivers/scsi/qla2xxx/qla_iocb.c
874
sle_dma = sgx.dma_addr;
drivers/scsi/qla4xxx/ql4_glbl.h
265
dma_addr_t dma_addr);
drivers/scsi/qla4xxx/ql4_glbl.h
28
int qla4xxx_get_flash(struct scsi_qla_host *ha, dma_addr_t dma_addr,
drivers/scsi/qla4xxx/ql4_glbl.h
79
int qla4xxx_set_flash(struct scsi_qla_host *ha, dma_addr_t dma_addr,
drivers/scsi/qla4xxx/ql4_mbx.c
1297
int qla4xxx_get_flash(struct scsi_qla_host * ha, dma_addr_t dma_addr,
drivers/scsi/qla4xxx/ql4_mbx.c
1307
mbox_cmd[1] = LSDW(dma_addr);
drivers/scsi/qla4xxx/ql4_mbx.c
1308
mbox_cmd[2] = MSDW(dma_addr);
drivers/scsi/qla4xxx/ql4_mbx.c
1396
dma_addr_t dma_addr)
drivers/scsi/qla4xxx/ql4_mbx.c
1406
mbox_cmd[2] = LSDW(dma_addr);
drivers/scsi/qla4xxx/ql4_mbx.c
1407
mbox_cmd[3] = MSDW(dma_addr);
drivers/scsi/qla4xxx/ql4_mbx.c
1464
int qla4xxx_set_flash(struct scsi_qla_host *ha, dma_addr_t dma_addr,
drivers/scsi/qla4xxx/ql4_mbx.c
1475
mbox_cmd[1] = LSDW(dma_addr);
drivers/scsi/qla4xxx/ql4_mbx.c
1476
mbox_cmd[2] = MSDW(dma_addr);
drivers/scsi/snic/snic_io.c
91
dma_unmap_single(&snic->pdev->dev, buf->dma_addr, buf->len,
drivers/scsi/snic/snic_res.h
51
dma_addr_t dma_addr,
drivers/scsi/snic/snic_res.h
60
(u64)dma_addr | VNIC_PADDR_TARGET,
drivers/scsi/snic/snic_res.h
72
svnic_wq_post(wq, os_buf, dma_addr, len, 1, 1);
drivers/scsi/snic/vnic_wq.h
38
dma_addr_t dma_addr;
drivers/scsi/snic/vnic_wq.h
89
void *os_buf, dma_addr_t dma_addr,
drivers/scsi/snic/vnic_wq.h
96
buf->dma_addr = dma_addr;
drivers/scsi/zorro_esp.c
218
static u32 zorro_esp_dma_length_limit(struct esp *esp, u32 dma_addr,
drivers/scsi/zorro_esp.c
224
static u32 fastlane_esp_dma_length_limit(struct esp *esp, u32 dma_addr,
drivers/scsi/zorro_esp.c
296
writeb((addr >> 24) & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
297
writeb((addr >> 16) & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
298
writeb((addr >> 8) & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
299
writeb(addr & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
342
writeb((addr >> 16) & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
343
writeb((addr >> 8) & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
344
writeb(addr & 0xff, &dregs->dma_addr);
drivers/scsi/zorro_esp.c
60
unsigned char dma_addr; /* DMA address [0x0000] */
drivers/scsi/zorro_esp.c
68
unsigned char dma_addr; /* DMA address [0x0000] */
drivers/soc/mediatek/mtk-cmdq-helper.c
150
dma_addr_t dma_addr;
drivers/soc/mediatek/mtk-cmdq-helper.c
159
dma_addr = dma_map_single(dev, pkt->va_base, pkt->buf_size,
drivers/soc/mediatek/mtk-cmdq-helper.c
161
if (dma_mapping_error(dev, dma_addr)) {
drivers/soc/mediatek/mtk-cmdq-helper.c
167
pkt->pa_base = dma_addr;
drivers/soc/ti/knav_qmss_queue.c
695
dma_addr_t dma_addr;
drivers/soc/ti/knav_qmss_queue.c
697
dma_addr = region->dma_start + (region->desc_size * index);
drivers/soc/ti/knav_qmss_queue.c
699
dma_sync_single_for_device(pool->dev, dma_addr, dma_size,
drivers/soc/ti/knav_qmss_queue.c
701
knav_queue_push(pool->queue, dma_addr, dma_size, 0);
drivers/spi/spi-airoha-snfi.c
561
dma_addr_t dma_addr;
drivers/spi/spi-airoha-snfi.c
644
dma_addr = dma_map_single(as_ctrl->dev, txrx_buf, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-airoha-snfi.c
646
err = dma_mapping_error(as_ctrl->dev, dma_addr);
drivers/spi/spi-airoha-snfi.c
652
dma_addr);
drivers/spi/spi-airoha-snfi.c
730
dma_unmap_single(as_ctrl->dev, dma_addr, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-airoha-snfi.c
741
dma_unmap_single(as_ctrl->dev, dma_addr, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-airoha-snfi.c
754
dma_addr_t dma_addr;
drivers/spi/spi-airoha-snfi.c
833
dma_addr = dma_map_single(as_ctrl->dev, txrx_buf, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-airoha-snfi.c
835
err = dma_mapping_error(as_ctrl->dev, dma_addr);
drivers/spi/spi-airoha-snfi.c
841
dma_addr);
drivers/spi/spi-airoha-snfi.c
916
dma_unmap_single(as_ctrl->dev, dma_addr, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-airoha-snfi.c
925
dma_unmap_single(as_ctrl->dev, dma_addr, SPI_NAND_CACHE_SIZE,
drivers/spi/spi-atcspi200.c
116
dma_addr_t dma_addr;
drivers/spi/spi-atcspi200.c
278
conf.src_addr = spi->dma_addr;
drivers/spi/spi-atcspi200.c
282
conf.dst_addr = spi->dma_addr;
drivers/spi/spi-atcspi200.c
544
spi->dma_addr = (dma_addr_t)(mem_res->start + ATCSPI_DATA);
drivers/spi/spi-cadence-quadspi.c
1006
dma_unmap_single(dev, dma_addr, bytes_to_dma, DMA_FROM_DEVICE);
drivers/spi/spi-cadence-quadspi.c
894
dma_addr_t dma_addr;
drivers/spi/spi-cadence-quadspi.c
916
dma_addr = dma_map_single(dev, rxbuf, bytes_to_dma, DMA_FROM_DEVICE);
drivers/spi/spi-cadence-quadspi.c
917
if (dma_mapping_error(dev, dma_addr)) {
drivers/spi/spi-cadence-quadspi.c
938
writel(lower_32_bits(dma_addr),
drivers/spi/spi-cadence-quadspi.c
940
writel(upper_32_bits(dma_addr),
drivers/spi/spi-cadence-quadspi.c
971
dma_unmap_single(dev, dma_addr, bytes_to_dma, DMA_FROM_DEVICE);
drivers/spi/spi-dw-core.c
940
dws->dma_addr = (dma_addr_t)(dws->paddr + DW_SPI_DR);
drivers/spi/spi-dw-dma.c
330
txconf.dst_addr = dws->dma_addr;
drivers/spi/spi-dw-dma.c
431
rxconf.src_addr = dws->dma_addr;
drivers/spi/spi-dw.h
195
dma_addr_t dma_addr; /* phy address of the Data register */
drivers/spi/spi-mtk-nor.c
367
dma_addr_t dma_addr)
drivers/spi/spi-mtk-nor.c
374
writel(dma_addr, sp->base + MTK_NOR_REG_DMA_DADR);
drivers/spi/spi-mtk-nor.c
375
writel(dma_addr + length, sp->base + MTK_NOR_REG_DMA_END_DADR);
drivers/spi/spi-mtk-nor.c
378
writel(upper_32_bits(dma_addr),
drivers/spi/spi-mtk-nor.c
380
writel(upper_32_bits(dma_addr + length),
drivers/spi/spi-mtk-nor.c
431
dma_addr_t dma_addr;
drivers/spi/spi-mtk-nor.c
436
dma_addr = dma_map_single(sp->dev, op->data.buf.in,
drivers/spi/spi-mtk-nor.c
439
if (dma_mapping_error(sp->dev, dma_addr))
drivers/spi/spi-mtk-nor.c
442
ret = mtk_nor_dma_exec(sp, op->addr.val, op->data.nbytes, dma_addr);
drivers/spi/spi-mtk-nor.c
444
dma_unmap_single(sp->dev, dma_addr, op->data.nbytes, DMA_FROM_DEVICE);
drivers/spi/spi-pci1xxxx.c
410
dma_addr_t dma_addr, u32 len)
drivers/spi/spi-pci1xxxx.c
421
writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_SAR_LO_OFFSET);
drivers/spi/spi-pci1xxxx.c
422
writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_SAR_HI_OFFSET);
drivers/spi/spi-pci1xxxx.c
431
dma_addr_t dma_addr, u32 len)
drivers/spi/spi-pci1xxxx.c
442
writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_DAR_LO_OFFSET);
drivers/spi/spi-pci1xxxx.c
443
writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_DAR_HI_OFFSET);
drivers/spi/spi-zynqmp-gqspi.c
204
dma_addr_t dma_addr;
drivers/spi/spi-zynqmp-gqspi.c
758
dma_unmap_single(xqspi->dev, xqspi->dma_addr,
drivers/spi/spi-zynqmp-gqspi.c
874
xqspi->dma_addr = addr;
drivers/staging/media/ipu7/ipu7-isys-queue.c
251
set->output_pins[aq->fw_output].addr = ivb->dma_addr;
drivers/staging/media/ipu7/ipu7-isys-queue.c
329
msg->dma_addr, sizeof(*buf),
drivers/staging/media/ipu7/ipu7-isys-queue.c
369
dma = ivb->dma_addr;
drivers/staging/media/ipu7/ipu7-isys-queue.c
428
buf, msg->dma_addr, sizeof(*buf),
drivers/staging/media/ipu7/ipu7-isys-queue.c
45
ivb->dma_addr = sg_dma_address(sg->sgl);
drivers/staging/media/ipu7/ipu7-isys-queue.c
58
ivb->dma_addr = 0;
drivers/staging/media/ipu7/ipu7-isys-queue.c
758
addr = ivb->dma_addr;
drivers/staging/media/ipu7/ipu7-isys-queue.h
39
dma_addr_t dma_addr;
drivers/staging/media/ipu7/ipu7-isys-video.c
478
stream_cfg, msg->dma_addr,
drivers/staging/media/ipu7/ipu7-isys-video.c
522
msg->dma_addr, sizeof(*buf),
drivers/staging/media/ipu7/ipu7-isys.c
584
fwmsg, fwmsg->dma_addr, 0);
drivers/staging/media/ipu7/ipu7-isys.c
588
fwmsg, fwmsg->dma_addr, 0);
drivers/staging/media/ipu7/ipu7-isys.c
603
dma_addr_t dma_addr;
drivers/staging/media/ipu7/ipu7-isys.c
609
&dma_addr, GFP_KERNEL, 0);
drivers/staging/media/ipu7/ipu7-isys.c
612
addr->dma_addr = dma_addr;
drivers/staging/media/ipu7/ipu7-isys.c
629
addr, addr->dma_addr, 0);
drivers/staging/media/ipu7/ipu7-isys.h
121
dma_addr_t dma_addr;
drivers/tee/tee_shm.c
22
dma_addr_t dma_addr;
drivers/tee/tee_shm.c
289
dma_addr_t dma_addr;
drivers/tee/tee_shm.c
296
&dma_addr, DMA_BIDIRECTIONAL, GFP_KERNEL);
drivers/tee/tee_shm.c
307
dma_mem->dma_addr = dma_addr;
drivers/tee/tee_shm.c
317
dma_free_pages(&teedev->dev, page_count * PAGE_SIZE, page, dma_addr,
drivers/tee/tee_shm.c
47
dma_mem->dma_addr, DMA_BIDIRECTIONAL);
drivers/thunderbolt/dma_test.c
268
dma_addr_t dma_addr;
drivers/thunderbolt/dma_test.c
280
dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE,
drivers/thunderbolt/dma_test.c
282
if (dma_mapping_error(dma_dev, dma_addr)) {
drivers/thunderbolt/dma_test.c
288
tf->frame.buffer_phy = dma_addr;
drivers/thunderbolt/dma_test.c
319
dma_addr_t dma_addr;
drivers/thunderbolt/dma_test.c
334
dma_addr = dma_map_single(dma_dev, tf->data, DMA_TEST_FRAME_SIZE,
drivers/thunderbolt/dma_test.c
336
if (dma_mapping_error(dma_dev, dma_addr)) {
drivers/thunderbolt/dma_test.c
342
tf->frame.buffer_phy = dma_addr;
drivers/tty/serial/atmel_serial.c
1434
pdc->dma_addr,
drivers/tty/serial/atmel_serial.c
1463
pdc->dma_addr,
drivers/tty/serial/atmel_serial.c
1471
atmel_uart_writel(port, ATMEL_PDC_TPR, pdc->dma_addr + tail);
drivers/tty/serial/atmel_serial.c
1496
pdc->dma_addr = dma_map_single(port->dev,
drivers/tty/serial/atmel_serial.c
1577
pdc->dma_addr,
drivers/tty/serial/atmel_serial.c
1599
head = atmel_uart_readl(port, ATMEL_PDC_RPR) - pdc->dma_addr;
drivers/tty/serial/atmel_serial.c
1615
dma_sync_single_for_cpu(port->dev, pdc->dma_addr,
drivers/tty/serial/atmel_serial.c
1629
dma_sync_single_for_device(port->dev, pdc->dma_addr,
drivers/tty/serial/atmel_serial.c
1642
atmel_uart_writel(port, ATMEL_PDC_RNPR, pdc->dma_addr);
drivers/tty/serial/atmel_serial.c
1668
atmel_port->pdc_rx[0].dma_addr,
drivers/tty/serial/atmel_serial.c
1676
pdc->dma_addr = dma_map_single(port->dev,
drivers/tty/serial/atmel_serial.c
1686
atmel_uart_writel(port, ATMEL_PDC_RPR, atmel_port->pdc_rx[0].dma_addr);
drivers/tty/serial/atmel_serial.c
1690
atmel_port->pdc_rx[1].dma_addr);
drivers/tty/serial/atmel_serial.c
83
dma_addr_t dma_addr;
drivers/tty/serial/cpm_uart.c
877
dma_addr_t dma_addr = 0;
drivers/tty/serial/cpm_uart.c
896
dma_addr = cpm_muram_dma((void __iomem *)mem_addr);
drivers/tty/serial/cpm_uart.c
899
dma_addr = virt_to_bus(mem_addr);
drivers/tty/serial/cpm_uart.c
901
mem_addr = dma_alloc_coherent(pinfo->port.dev, memsz, &dma_addr,
drivers/tty/serial/cpm_uart.c
913
pinfo->dma_addr = dma_addr;
drivers/tty/serial/cpm_uart.c
932
pinfo->dma_addr);
drivers/tty/serial/cpm_uart.h
102
u32 dma = (u32)pinfo->dma_addr;
drivers/tty/serial/cpm_uart.h
69
dma_addr_t dma_addr;
drivers/tty/serial/cpm_uart.h
90
return pinfo->dma_addr + offset;
drivers/uio/uio.c
793
if (mem->dma_addr & ~PAGE_MASK)
drivers/uio/uio.c
813
mem->dma_addr,
drivers/uio/uio_dmem_genirq.c
62
&uiomem->dma_addr, GFP_KERNEL);
drivers/uio/uio_dmem_genirq.c
93
uiomem->dma_addr);
drivers/usb/cdns3/cdns3-ep0.c
35
dma_addr_t dma_addr,
drivers/usb/cdns3/cdns3-ep0.c
41
priv_ep->trb_pool[0].buffer = cpu_to_le32(TRB_BUFFER(dma_addr));
drivers/usb/cdns3/cdns3-ep0.c
46
priv_ep->trb_pool[1].buffer = cpu_to_le32(TRB_BUFFER(dma_addr));
drivers/usb/dwc2/gadget.c
1008
dma_addr_t dma_addr = hs_req->req.dma;
drivers/usb/dwc2/gadget.c
1012
dma_addr = sg_dma_address(hs_req->req.sg);
drivers/usb/dwc2/gadget.c
1014
ret = dwc2_gadget_fill_isoc_desc(hs_ep, dma_addr,
drivers/usb/dwc2/gadget.c
1487
dma_addr_t dma_addr = hs_req->req.dma;
drivers/usb/dwc2/gadget.c
1491
dma_addr = sg_dma_address(hs_req->req.sg);
drivers/usb/dwc2/gadget.c
1493
dwc2_gadget_fill_isoc_desc(hs_ep, dma_addr,
drivers/usb/dwc2/hcd.c
1336
dma_addr_t dma_addr;
drivers/usb/dwc2/hcd.c
1341
dma_addr = chan->align_buf;
drivers/usb/dwc2/hcd.c
1343
dma_addr = chan->xfer_dma;
drivers/usb/dwc2/hcd.c
1345
dwc2_writel(hsotg, (u32)dma_addr, HCDMA(chan->hc_num));
drivers/usb/dwc2/hcd.c
1349
(unsigned long)dma_addr, chan->hc_num);
drivers/usb/early/xhci-dbc.c
101
if (dma_addr)
drivers/usb/early/xhci-dbc.c
102
*dma_addr = (dma_addr_t)__pa(virt);
drivers/usb/early/xhci-dbc.c
93
static void * __init xdbc_get_page(dma_addr_t *dma_addr)
drivers/usb/gadget/udc/bdc/bdc.h
382
dma_addr_t dma_addr;
drivers/usb/gadget/udc/bdc/bdc_cmd.c
222
int bdc_ep_bla(struct bdc *bdc, struct bdc_ep *ep, dma_addr_t dma_addr)
drivers/usb/gadget/udc/bdc/bdc_cmd.c
228
(unsigned long long)(dma_addr));
drivers/usb/gadget/udc/bdc/bdc_cmd.c
229
param0 = lower_32_bits(dma_addr);
drivers/usb/gadget/udc/bdc/bdc_cmd.c
230
param1 = upper_32_bits(dma_addr);
drivers/usb/gadget/udc/bdc/bdc_cmd.h
19
int bdc_ep_bla(struct bdc *bdc, struct bdc_ep *ep, dma_addr_t dma_addr);
drivers/usb/gadget/udc/bdc/bdc_core.c
204
&bdc->srr.dma_addr, GFP_KERNEL);
drivers/usb/gadget/udc/bdc/bdc_core.c
234
temp = lower_32_bits(bdc->srr.dma_addr);
drivers/usb/gadget/udc/bdc/bdc_core.c
241
upp32 = upper_32_bits(bdc->srr.dma_addr);
drivers/usb/gadget/udc/bdc/bdc_core.c
319
bdc->srr.sr_bds, bdc->srr.dma_addr);
drivers/usb/gadget/udc/bdc/bdc_dbg.c
70
addr = bdc->srr.dma_addr;
drivers/usb/gadget/udc/cdns2/cdns2-ep0.c
41
static void cdns2_ep0_enqueue(struct cdns2_device *pdev, dma_addr_t dma_addr,
drivers/usb/gadget/udc/cdns2/cdns2-ep0.c
48
ring->trbs[0].buffer = cpu_to_le32(TRB_BUFFER(dma_addr));
drivers/usb/gadget/udc/cdns2/cdns2-ep0.c
54
ring->trbs[1].buffer = cpu_to_le32(TRB_BUFFER(dma_addr));
drivers/usb/gadget/udc/pch_udc.c
1510
dma_addr_t dma_addr;
drivers/usb/gadget/udc/pch_udc.c
1525
&dma_addr);
drivers/usb/gadget/udc/pch_udc.c
1530
last->next = dma_addr;
drivers/usb/gadget/udc/pch_udc.c
1725
if (!ep->dev->dma_addr)
drivers/usb/gadget/udc/pch_udc.c
2462
ep->td_data->dataptr = dev->dma_addr;
drivers/usb/gadget/udc/pch_udc.c
2899
dev->dma_addr = dma_map_single(&dev->pdev->dev, ep0out_buf,
drivers/usb/gadget/udc/pch_udc.c
2902
return dma_mapping_error(&dev->pdev->dev, dev->dma_addr);
drivers/usb/gadget/udc/pch_udc.c
3010
if (dev->dma_addr)
drivers/usb/gadget/udc/pch_udc.c
3011
dma_unmap_single(&dev->pdev->dev, dev->dma_addr,
drivers/usb/gadget/udc/pch_udc.c
354
dma_addr_t dma_addr;
drivers/usb/gadget/udc/snps_udc_core.c
762
dma_addr_t dma_addr;
drivers/usb/gadget/udc/snps_udc_core.c
771
dma_addr = DMA_DONT_USE;
drivers/usb/gadget/udc/snps_udc_core.c
796
gfp_flags, &dma_addr);
drivers/usb/gadget/udc/snps_udc_core.c
827
req->td_data->next = dma_addr;
drivers/usb/gadget/udc/snps_udc_core.c
846
last->next = dma_addr;
drivers/usb/mtu3/mtu3_qmu.c
120
dma_addr_t dma_addr)
drivers/usb/mtu3/mtu3_qmu.c
124
u32 offset = (dma_addr - dma_base) / sizeof(*gpd_head);
drivers/usb/musb/musb_cppi41.c
413
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/musb_cppi41.c
423
cppi41_channel->buf_addr = dma_addr;
drivers/usb/musb/musb_cppi41.c
465
dma_desc = dmaengine_prep_slave_single(dc, dma_addr, len, direction,
drivers/usb/musb/musb_cppi41.c
525
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/musb_cppi41.c
547
ret = cppi41_configure_channel(channel, packet_sz, mode, dma_addr, len);
drivers/usb/musb/musb_dma.h
160
dma_addr_t dma_addr,
drivers/usb/musb/musb_gadget.c
56
dma_addr_t dma_addr;
drivers/usb/musb/musb_gadget.c
59
dma_addr = dma_map_single(
drivers/usb/musb/musb_gadget.c
66
ret = dma_mapping_error(musb->controller, dma_addr);
drivers/usb/musb/musb_gadget.c
70
request->request.dma = dma_addr;
drivers/usb/musb/musb_gadget.c
737
u32 dma_addr = request->dma + request->actual;
drivers/usb/musb/musb_gadget.c
743
dma_addr,
drivers/usb/musb/musbhsdma.c
145
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/musbhsdma.c
155
channel, packet_sz, &dma_addr, len, mode);
drivers/usb/musb/musbhsdma.c
172
musb_write_hsdma_addr(mbase, bchannel, dma_addr);
drivers/usb/musb/musbhsdma.c
183
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/musbhsdma.c
192
packet_sz, &dma_addr, len, mode);
drivers/usb/musb/musbhsdma.c
206
if ((musb->hwvers >= MUSB_HWVERS_1800) && (dma_addr % 4))
drivers/usb/musb/musbhsdma.c
210
musb_channel->start_addr = dma_addr;
drivers/usb/musb/musbhsdma.c
215
configure_channel(channel, packet_sz, mode, dma_addr, len);
drivers/usb/musb/tusb6010_omap.c
138
buf = phys_to_virt((u32)chdat->dma_addr) + chdat->transfer_len;
drivers/usb/musb/tusb6010_omap.c
140
dma_unmap_single(dev, chdat->dma_addr,
drivers/usb/musb/tusb6010_omap.c
145
dma_unmap_single(dev, chdat->dma_addr,
drivers/usb/musb/tusb6010_omap.c
182
u8 rndis_mode, dma_addr_t dma_addr, u32 len)
drivers/usb/musb/tusb6010_omap.c
202
if (unlikely(dma_addr & 0x1) || (len < 32) || (len > packet_sz))
drivers/usb/musb/tusb6010_omap.c
211
if (dma_addr & 0x2)
drivers/usb/musb/tusb6010_omap.c
256
chdat->dma_addr = dma_addr;
drivers/usb/musb/tusb6010_omap.c
262
dma_map_single(dev, phys_to_virt(dma_addr), len,
drivers/usb/musb/tusb6010_omap.c
266
dma_map_single(dev, phys_to_virt(dma_addr), len,
drivers/usb/musb/tusb6010_omap.c
273
if ((dma_addr & 0x3) == 0) {
drivers/usb/musb/tusb6010_omap.c
287
chdat->epnum, chdat->tx ? "tx" : "rx", &dma_addr,
drivers/usb/musb/tusb6010_omap.c
303
dma_desc = dmaengine_prep_slave_single(dma_data->chan, dma_addr,
drivers/usb/musb/tusb6010_omap.c
319
((dma_addr & 0x3) == 0) ? "sync" : "async",
drivers/usb/musb/tusb6010_omap.c
320
(dma_dir == DMA_MEM_TO_DEV) ? &dma_addr : &fifo_addr,
drivers/usb/musb/tusb6010_omap.c
321
(dma_dir == DMA_MEM_TO_DEV) ? &fifo_addr : &dma_addr);
drivers/usb/musb/tusb6010_omap.c
41
dma_addr_t dma_addr;
drivers/usb/musb/ux500_dma.c
191
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/ux500_dma.c
200
ret = ux500_configure_channel(channel, packet_sz, mode, dma_addr, len);
drivers/usb/musb/ux500_dma.c
72
dma_addr_t dma_addr, u32 len)
drivers/usb/musb/ux500_dma.c
88
packet_sz, mode, (unsigned long long) dma_addr,
drivers/usb/musb/ux500_dma.c
94
sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len,
drivers/usb/musb/ux500_dma.c
95
offset_in_page(dma_addr));
drivers/usb/musb/ux500_dma.c
96
sg_dma_address(&sg) = dma_addr;
drivers/vdpa/mlx5/core/mr.c
39
u64 dma_addr;
drivers/vdpa/mlx5/core/mr.c
45
for (dma_addr = sg_dma_address(sg), dma_len = sg_dma_len(sg);
drivers/vdpa/mlx5/core/mr.c
47
nsg--, dma_addr += BIT(mr->log_size), dma_len -= BIT(mr->log_size))
drivers/vdpa/mlx5/core/mr.c
48
mtt[j++] = cpu_to_be64(dma_addr);
drivers/vdpa/vdpa_user/iova_domain.c
428
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.c
433
vduse_domain_bounce(domain, dma_addr, size, DMA_TO_DEVICE);
drivers/vdpa/vdpa_user/iova_domain.c
438
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.c
443
vduse_domain_bounce(domain, dma_addr, size, DMA_FROM_DEVICE);
drivers/vdpa/vdpa_user/iova_domain.c
482
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.c
489
vduse_domain_bounce(domain, dma_addr, size, DMA_FROM_DEVICE);
drivers/vdpa/vdpa_user/iova_domain.c
491
vduse_domain_unmap_bounce_page(domain, (u64)dma_addr, (u64)size);
drivers/vdpa/vdpa_user/iova_domain.c
493
vduse_domain_free_iova(iovad, dma_addr, size);
drivers/vdpa/vdpa_user/iova_domain.c
524
dma_addr_t dma_addr, unsigned long attrs)
drivers/vdpa/vdpa_user/iova_domain.c
531
map = vhost_iotlb_itree_first(domain->iotlb, (u64)dma_addr,
drivers/vdpa/vdpa_user/iova_domain.c
532
(u64)dma_addr + size - 1);
drivers/vdpa/vdpa_user/iova_domain.c
543
vduse_domain_free_iova(iovad, dma_addr, size);
drivers/vdpa/vdpa_user/iova_domain.h
52
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.h
56
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.h
65
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/iova_domain.h
72
dma_addr_t dma_addr, unsigned long attrs);
drivers/vdpa/vdpa_user/vduse_dev.c
1001
void *vaddr, dma_addr_t dma_addr,
drivers/vdpa/vdpa_user/vduse_dev.c
1012
vduse_domain_free_coherent(domain, size, dma_addr, attrs);
drivers/vdpa/vdpa_user/vduse_dev.c
1018
static bool vduse_dev_need_sync(union virtio_map token, dma_addr_t dma_addr)
drivers/vdpa/vdpa_user/vduse_dev.c
1024
return dma_addr < token.group->as->domain->bounce_size;
drivers/vdpa/vdpa_user/vduse_dev.c
1027
static int vduse_dev_mapping_error(union virtio_map token, dma_addr_t dma_addr)
drivers/vdpa/vdpa_user/vduse_dev.c
1029
if (unlikely(dma_addr == DMA_MAPPING_ERROR))
drivers/vdpa/vdpa_user/vduse_dev.c
914
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/vduse_dev.c
924
vduse_domain_sync_single_for_device(domain, dma_addr, size, dir);
drivers/vdpa/vdpa_user/vduse_dev.c
928
dma_addr_t dma_addr, size_t size,
drivers/vdpa/vdpa_user/vduse_dev.c
938
vduse_domain_sync_single_for_cpu(domain, dma_addr, size, dir);
drivers/vdpa/vdpa_user/vduse_dev.c
956
static void vduse_dev_unmap_page(union virtio_map token, dma_addr_t dma_addr,
drivers/vdpa/vdpa_user/vduse_dev.c
967
vduse_domain_unmap_page(domain, dma_addr, size, dir, attrs);
drivers/vdpa/vdpa_user/vduse_dev.c
971
dma_addr_t *dma_addr, gfp_t flag)
drivers/vdpa/vdpa_user/vduse_dev.c
975
*dma_addr = DMA_MAPPING_ERROR;
drivers/vdpa/vdpa_user/vduse_dev.c
988
*dma_addr = vduse_domain_alloc_coherent(domain, size, addr);
drivers/vdpa/vdpa_user/vduse_dev.c
989
if (*dma_addr == DMA_MAPPING_ERROR)
drivers/video/fbdev/atmel_lcdfb.c
297
unsigned long dma_addr;
drivers/video/fbdev/atmel_lcdfb.c
299
dma_addr = (fix->smem_start + var->yoffset * fix->line_length
drivers/video/fbdev/atmel_lcdfb.c
302
dma_addr &= ~3UL;
drivers/video/fbdev/atmel_lcdfb.c
305
lcdc_writel(sinfo, ATMEL_LCDC_DMABADDR1, dma_addr);
drivers/video/fbdev/fsl-diu-fb.c
1687
dma_addr_t dma_addr; /* DMA addr of fsl_diu_data struct */
drivers/video/fbdev/fsl-diu-fb.c
1693
&dma_addr, GFP_DMA | __GFP_ZERO);
drivers/video/fbdev/fsl-diu-fb.c
1696
data->dma_addr = dma_addr;
drivers/video/fbdev/fsl-diu-fb.c
367
dma_addr_t dma_addr;
drivers/video/fbdev/fsl-diu-fb.c
390
#define DMA_ADDR(p, f) ((p)->dma_addr + offsetof(struct fsl_diu_data, f))
drivers/virtio/virtio_ring.c
1253
dma_addr_t dma_addr;
drivers/virtio/virtio_ring.c
1264
&dma_addr,
drivers/virtio/virtio_ring.c
1279
&dma_addr, GFP_KERNEL | __GFP_ZERO,
drivers/virtio/virtio_ring.c
1287
vring_split->queue_dma_addr = dma_addr;
drivers/xen/swiotlb-xen.c
100
dma_addr_t dma_addr)
drivers/xen/swiotlb-xen.c
102
unsigned long bfn = XEN_PFN_DOWN(dma_to_phys(dev, dma_addr));
drivers/xen/swiotlb-xen.c
304
xen_swiotlb_sync_single_for_cpu(struct device *dev, dma_addr_t dma_addr,
drivers/xen/swiotlb-xen.c
307
phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr);
drivers/xen/swiotlb-xen.c
311
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
drivers/xen/swiotlb-xen.c
314
xen_dma_sync_for_cpu(dev, dma_addr, size, dir);
drivers/xen/swiotlb-xen.c
317
pool = xen_swiotlb_find_pool(dev, dma_addr);
drivers/xen/swiotlb-xen.c
323
xen_swiotlb_sync_single_for_device(struct device *dev, dma_addr_t dma_addr,
drivers/xen/swiotlb-xen.c
326
phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr);
drivers/xen/swiotlb-xen.c
329
pool = xen_swiotlb_find_pool(dev, dma_addr);
drivers/xen/swiotlb-xen.c
334
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
drivers/xen/swiotlb-xen.c
337
xen_dma_sync_for_device(dev, dma_addr, size, dir);
drivers/xen/swiotlb-xen.c
72
dma_addr_t dma_addr)
drivers/xen/swiotlb-xen.c
74
return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr));
include/drm/drm_gem_dma_helper.h
23
dma_addr_t dma_addr;
include/drm/drm_gpusvm.h
148
struct drm_pagemap_addr *dma_addr;
include/linux/comedi/comedidev.h
222
dma_addr_t dma_addr;
include/linux/dibs.h
67
dma_addr_t dma_addr;
include/linux/dma-direct.h
104
static inline phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
include/linux/dma-direct.h
108
dma_addr = dma_addr_canonical(dma_addr);
include/linux/dma-direct.h
110
paddr = translate_dma_to_phys(dev, dma_addr);
include/linux/dma-direct.h
112
paddr = dma_addr;
include/linux/dma-direct.h
145
dma_addr_t dma_addr, unsigned long attrs);
include/linux/dma-direct.h
149
struct page *page, dma_addr_t dma_addr,
include/linux/dma-direct.h
43
dma_addr_t dma_addr)
include/linux/dma-direct.h
48
u64 offset = dma_addr - m->dma_start;
include/linux/dma-direct.h
50
if (dma_addr >= m->dma_start && offset < m->size)
include/linux/dma-map-ops.h
202
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-map-ops.h
205
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-map-ops.h
31
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-map-ops.h
316
dma_addr_t dma_addr, unsigned long attrs);
include/linux/dma-mapping.h
127
void debug_dma_mapping_error(struct device *dev, dma_addr_t dma_addr);
include/linux/dma-mapping.h
132
dma_addr_t dma_addr)
include/linux/dma-mapping.h
142
static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
include/linux/dma-mapping.h
144
debug_dma_mapping_error(dev, dma_addr);
include/linux/dma-mapping.h
146
if (unlikely(dma_addr == DMA_MAPPING_ERROR))
include/linux/dma-mapping.h
180
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-mapping.h
183
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-mapping.h
249
static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
include/linux/dma-mapping.h
272
struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr,
include/linux/dma-mapping.h
278
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/dma-mapping.h
417
bool __dma_need_sync(struct device *dev, dma_addr_t dma_addr);
include/linux/dma-mapping.h
453
static inline bool dma_need_sync(struct device *dev, dma_addr_t dma_addr)
include/linux/dma-mapping.h
455
return dma_dev_need_sync(dev) ? __dma_need_sync(dev, dma_addr) : false;
include/linux/dma-mapping.h
479
static inline bool dma_need_sync(struct device *dev, dma_addr_t dma_addr)
include/linux/dma-mapping.h
741
dma_addr_t *dma_addr, gfp_t gfp)
include/linux/dma-mapping.h
748
return dma_alloc_attrs(dev, size, dma_addr, gfp, attrs);
include/linux/dma-mapping.h
752
void *cpu_addr, dma_addr_t dma_addr)
include/linux/dma-mapping.h
754
return dma_free_attrs(dev, size, cpu_addr, dma_addr,
include/linux/dma-mapping.h
760
void *cpu_addr, dma_addr_t dma_addr,
include/linux/dma-mapping.h
763
return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size,
include/linux/hisi_acc_qm.h
572
int hisi_qm_mb(struct hisi_qm *qm, u8 cmd, dma_addr_t dma_addr, u16 queue,
include/linux/iommu-dma.h
35
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/iommu-dma.h
38
void *cpu_addr, dma_addr_t dma_addr, size_t size,
include/linux/mhi.h
503
dma_addr_t dma_addr;
include/linux/mm_types.h
125
unsigned long dma_addr;
include/linux/raspberrypi/vchiq_core.h
120
dma_addr_t dma_addr;
include/linux/raspberrypi/vchiq_core.h
480
dma_addr_t dma_addr;
include/linux/skbuff.h
3782
return netmem_to_net_iov(frag->netmem)->desc.dma_addr +
include/linux/switchtec.h
54
u64 dma_addr;
include/linux/uio_driver.h
45
dma_addr_t dma_addr;
include/media/vsp1.h
142
dma_addr_t dma_addr;
include/net/netmem.h
109
unsigned long dma_addr;
include/net/netmem.h
142
NET_IOV_ASSERT_OFFSET(dma_addr, dma_addr);
include/net/netmem.h
37
unsigned long dma_addr;
include/net/netmem.h
389
return netmem_to_nmdesc(netmem)->dma_addr;
include/net/netmem.h
48
NETMEM_DESC_ASSERT_OFFSET(dma_addr, dma_addr);
include/net/page_pool/helpers.h
451
const dma_addr_t dma_addr,
include/net/page_pool/helpers.h
454
dma_sync_single_range_for_cpu(pool->p.dev, dma_addr,
include/rdma/ib_umem.h
142
dma_addr_t dma_addr;
include/rdma/ib_umem.h
144
dma_addr = ib_umem_start_dma_addr(umem);
include/rdma/ib_umem.h
146
dma_addr & pgoff_bitmask);
include/rdma/ib_umem.h
151
dma_addr_t dma_addr;
include/rdma/ib_umem.h
158
dma_addr = ib_umem_start_dma_addr(umem);
include/rdma/ib_umem.h
159
pgsz = roundup_pow_of_two((dma_addr ^ (umem->length - 1 + dma_addr)) + 1);
include/rdma/ib_verbs.h
4184
static inline void *ib_virt_dma_to_ptr(u64 dma_addr)
include/rdma/ib_verbs.h
4187
return (void *)(uintptr_t)dma_addr;
include/rdma/ib_verbs.h
4197
static inline struct page *ib_virt_dma_to_page(u64 dma_addr)
include/rdma/ib_verbs.h
4199
return virt_to_page(ib_virt_dma_to_ptr(dma_addr));
include/rdma/ib_verbs.h
4207
static inline int ib_dma_mapping_error(struct ib_device *dev, u64 dma_addr)
include/rdma/ib_verbs.h
4211
return dma_mapping_error(dev->dma_device, dma_addr);
include/sound/compress_driver.h
274
runtime->dma_addr = bufp->addr;
include/sound/compress_driver.h
279
runtime->dma_addr = 0;
include/sound/compress_driver.h
82
dma_addr_t dma_addr;
include/sound/pcm.h
1270
runtime->dma_addr = bufp->addr;
include/sound/pcm.h
1275
runtime->dma_addr = 0;
include/sound/pcm.h
438
dma_addr_t dma_addr; /* physical bus address (not accessible from main CPU) */
include/trace/events/dma.h
117
TP_PROTO(struct device *dev, void *virt_addr, dma_addr_t dma_addr,
include/trace/events/dma.h
120
TP_ARGS(dev, virt_addr, dma_addr, size, dir, flags, attrs),
include/trace/events/dma.h
125
__field(u64, dma_addr)
include/trace/events/dma.h
135
__entry->dma_addr = dma_addr;
include/trace/events/dma.h
145
__entry->dma_addr,
include/trace/events/dma.h
154
TP_PROTO(struct device *dev, void *virt_addr, dma_addr_t dma_addr, \
include/trace/events/dma.h
157
TP_ARGS(dev, virt_addr, dma_addr, size, dir, flags, attrs))
include/trace/events/dma.h
171
__field(u64, dma_addr)
include/trace/events/dma.h
185
__entry->dma_addr = sg_dma_address(sgt->sgl);
include/trace/events/dma.h
195
__entry->dma_addr,
include/trace/events/dma.h
205
TP_PROTO(struct device *dev, void *virt_addr, dma_addr_t dma_addr,
include/trace/events/dma.h
207
TP_ARGS(dev, virt_addr, dma_addr, size, dir, attrs),
include/trace/events/dma.h
212
__field(u64, dma_addr)
include/trace/events/dma.h
221
__entry->dma_addr = dma_addr;
include/trace/events/dma.h
230
__entry->dma_addr,
include/trace/events/dma.h
238
TP_PROTO(struct device *dev, void *virt_addr, dma_addr_t dma_addr, \
include/trace/events/dma.h
240
TP_ARGS(dev, virt_addr, dma_addr, size, dir, attrs))
include/trace/events/dma.h
253
__field(u64, dma_addr)
include/trace/events/dma.h
265
__entry->dma_addr = sg_dma_address(sgt->sgl);
include/trace/events/dma.h
273
__entry->dma_addr,
include/trace/events/dma.h
40
TP_PROTO(struct device *dev, phys_addr_t phys_addr, dma_addr_t dma_addr,
include/trace/events/dma.h
407
TP_PROTO(struct device *dev, dma_addr_t dma_addr, size_t size,
include/trace/events/dma.h
409
TP_ARGS(dev, dma_addr, size, dir),
include/trace/events/dma.h
413
__field(u64, dma_addr)
include/trace/events/dma.h
42
TP_ARGS(dev, phys_addr, dma_addr, size, dir, attrs),
include/trace/events/dma.h
420
__entry->dma_addr = dma_addr;
include/trace/events/dma.h
428
__entry->dma_addr,
include/trace/events/dma.h
434
TP_PROTO(struct device *dev, dma_addr_t dma_addr, size_t size, \
include/trace/events/dma.h
436
TP_ARGS(dev, dma_addr, size, dir))
include/trace/events/dma.h
47
__field(u64, dma_addr)
include/trace/events/dma.h
56
__entry->dma_addr = dma_addr;
include/trace/events/dma.h
65
__entry->dma_addr,
include/trace/events/dma.h
73
TP_PROTO(struct device *dev, phys_addr_t phys_addr, dma_addr_t dma_addr, \
include/trace/events/dma.h
75
TP_ARGS(dev, phys_addr, dma_addr, size, dir, attrs))
include/trace/events/habanalabs.h
108
__entry->dma_addr = dma_addr;
include/trace/events/habanalabs.h
117
__entry->dma_addr,
include/trace/events/habanalabs.h
124
TP_PROTO(struct device *dev, u64 phys_addr, u64 dma_addr, size_t len,
include/trace/events/habanalabs.h
126
TP_ARGS(dev, phys_addr, dma_addr, len, dir, caller));
include/trace/events/habanalabs.h
129
TP_PROTO(struct device *dev, u64 phys_addr, u64 dma_addr, size_t len,
include/trace/events/habanalabs.h
131
TP_ARGS(dev, phys_addr, dma_addr, len, dir, caller));
include/trace/events/habanalabs.h
54
TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
include/trace/events/habanalabs.h
56
TP_ARGS(dev, cpu_addr, dma_addr, size, caller),
include/trace/events/habanalabs.h
61
__field(u64, dma_addr)
include/trace/events/habanalabs.h
69
__entry->dma_addr = dma_addr;
include/trace/events/habanalabs.h
77
__entry->dma_addr,
include/trace/events/habanalabs.h
83
TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
include/trace/events/habanalabs.h
84
TP_ARGS(dev, cpu_addr, dma_addr, size, caller));
include/trace/events/habanalabs.h
87
TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
include/trace/events/habanalabs.h
88
TP_ARGS(dev, cpu_addr, dma_addr, size, caller));
include/trace/events/habanalabs.h
91
TP_PROTO(struct device *dev, u64 phys_addr, u64 dma_addr, size_t len,
include/trace/events/habanalabs.h
94
TP_ARGS(dev, phys_addr, dma_addr, len, dir, caller),
include/trace/events/habanalabs.h
99
__field(u64, dma_addr)
include/trace/events/rpcrdma.h
1789
u64 dma_addr,
include/trace/events/rpcrdma.h
1793
TP_ARGS(cid, dma_addr, length),
include/trace/events/rpcrdma.h
1798
__field(u64, dma_addr)
include/trace/events/rpcrdma.h
1805
__entry->dma_addr = dma_addr;
include/trace/events/rpcrdma.h
1811
__entry->dma_addr, __entry->length
include/trace/events/rpcrdma.h
1819
u64 dma_addr, \
include/trace/events/rpcrdma.h
1822
TP_ARGS(cid, dma_addr, length) \
io_uring/zcrx.c
315
dma_addr_t dma_addr;
io_uring/zcrx.c
323
dma_addr = page_pool_get_dma_addr_netmem(net_iov_to_netmem(niov));
io_uring/zcrx.c
324
__dma_sync_single_for_device(pool->p.dev, dma_addr + pool->p.offset,
kernel/dma/debug.c
1231
int direction, dma_addr_t dma_addr, unsigned long attrs)
kernel/dma/debug.c
1238
if (dma_mapping_error(dev, dma_addr))
kernel/dma/debug.c
1248
entry->dev_addr = dma_addr;
kernel/dma/debug.c
1263
void debug_dma_mapping_error(struct device *dev, dma_addr_t dma_addr)
kernel/dma/debug.c
1274
ref.dev_addr = dma_addr;
kernel/dma/debug.c
1301
void debug_dma_unmap_phys(struct device *dev, dma_addr_t dma_addr,
kernel/dma/debug.c
1307
.dev_addr = dma_addr,
kernel/dma/debug.c
1417
dma_addr_t dma_addr, void *virt,
kernel/dma/debug.c
1440
entry->dev_addr = dma_addr;
kernel/dma/debug.c
1447
void *virt, dma_addr_t dma_addr)
kernel/dma/debug.c
1452
.dev_addr = dma_addr,
kernel/dma/debug.c
1569
dma_addr_t dma_addr,
kernel/dma/debug.c
1585
entry->dev_addr = dma_addr;
kernel/dma/debug.c
1593
dma_addr_t dma_addr)
kernel/dma/debug.c
1599
.dev_addr = dma_addr,
kernel/dma/debug.h
116
dma_addr_t dma_addr,
kernel/dma/debug.h
123
dma_addr_t dma_addr)
kernel/dma/debug.h
13
size_t size, int direction, dma_addr_t dma_addr,
kernel/dma/debug.h
27
dma_addr_t dma_addr, void *virt,
kernel/dma/debug.h
50
dma_addr_t dma_addr,
kernel/dma/debug.h
54
dma_addr_t dma_addr);
kernel/dma/debug.h
58
dma_addr_t dma_addr, unsigned long attrs)
kernel/dma/debug.h
80
dma_addr_t dma_addr, void *virt,
kernel/dma/direct.c
311
void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs)
kernel/dma/direct.c
325
arch_dma_free(dev, size, cpu_addr, dma_addr, attrs);
kernel/dma/direct.c
35
dma_addr_t dma_addr)
kernel/dma/direct.c
350
__dma_direct_free_pages(dev, dma_direct_to_page(dev, dma_addr), size);
kernel/dma/direct.c
37
return pfn_to_page(PHYS_PFN(dma_to_phys(dev, dma_addr)));
kernel/dma/direct.c
377
struct page *page, dma_addr_t dma_addr,
kernel/dma/direct.c
499
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/direct.c
502
struct page *page = dma_direct_to_page(dev, dma_addr);
kernel/dma/direct.c
518
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/direct.c
523
unsigned long pfn = PHYS_PFN(dma_to_phys(dev, dma_addr));
kernel/dma/direct.c
621
bool dma_direct_need_sync(struct device *dev, dma_addr_t dma_addr)
kernel/dma/direct.c
624
swiotlb_find_pool(dev, dma_to_phys(dev, dma_addr));
kernel/dma/direct.c
72
dma_addr_t dma_addr = phys_to_dma_direct(dev, phys);
kernel/dma/direct.c
74
if (dma_addr == DMA_MAPPING_ERROR)
kernel/dma/direct.c
76
return dma_addr + size - 1 <=
kernel/dma/direct.h
112
return dma_addr;
kernel/dma/direct.h
118
&dma_addr, size, *dev->dma_mask, dev->bus_dma_limit);
kernel/dma/direct.h
14
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/direct.h
18
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/direct.h
20
bool dma_direct_need_sync(struct device *dev, dma_addr_t dma_addr);
kernel/dma/direct.h
84
dma_addr_t dma_addr;
kernel/dma/direct.h
94
dma_addr = phys;
kernel/dma/direct.h
95
if (unlikely(!dma_capable(dev, dma_addr, size, false)))
kernel/dma/direct.h
98
dma_addr = phys_to_dma(dev, phys);
kernel/dma/direct.h
99
if (unlikely(!dma_capable(dev, dma_addr, size, true)) ||
kernel/dma/dummy.c
8
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/map_benchmark.c
37
dma_addr_t dma_addr;
kernel/dma/map_benchmark.c
62
dma_addr = dma_map_single(map->dev, buf, size, map->dir);
kernel/dma/map_benchmark.c
63
if (unlikely(dma_mapping_error(map->dev, dma_addr))) {
kernel/dma/map_benchmark.c
76
dma_unmap_single(map->dev, dma_addr, size, map->dir);
kernel/dma/mapping.c
443
bool __dma_need_sync(struct device *dev, dma_addr_t dma_addr)
kernel/dma/mapping.c
453
return dma_direct_need_sync(dev, dma_addr);
kernel/dma/mapping.c
513
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/mapping.c
519
return dma_direct_get_sgtable(dev, sgt, cpu_addr, dma_addr,
kernel/dma/mapping.c
522
return iommu_dma_get_sgtable(dev, sgt, cpu_addr, dma_addr,
kernel/dma/mapping.c
526
return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs);
kernel/dma/mapping.c
580
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/mapping.c
586
return dma_direct_mmap(dev, vma, cpu_addr, dma_addr, size,
kernel/dma/mapping.c
589
return iommu_dma_mmap(dev, vma, cpu_addr, dma_addr, size,
kernel/dma/mapping.c
593
return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs);
kernel/dma/ops_helpers.c
20
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/ops_helpers.c
36
void *cpu_addr, dma_addr_t dma_addr, size_t size,
kernel/dma/swiotlb.c
1595
dma_addr_t dma_addr;
kernel/dma/swiotlb.c
1604
dma_addr = phys_to_dma_unencrypted(dev, swiotlb_addr);
kernel/dma/swiotlb.c
1605
if (unlikely(!dma_capable(dev, dma_addr, size, true))) {
kernel/dma/swiotlb.c
1611
&dma_addr, size, *dev->dma_mask, dev->bus_dma_limit);
kernel/dma/swiotlb.c
1617
return dma_addr;
mm/hmm.c
782
dma_addr_t dma_addr;
mm/hmm.c
838
dma_addr = state->addr + offset;
mm/hmm.c
843
dma_addr = dma_map_phys(dev, paddr, map->dma_entry_size,
mm/hmm.c
845
if (dma_mapping_error(dev, dma_addr))
mm/hmm.c
849
dma_addrs[idx] = dma_addr;
mm/hmm.c
852
return dma_addr;
net/core/devmem.c
101
offset = dma_addr - owner->base_dma_addr;
net/core/devmem.c
115
unsigned long dma_addr = net_devmem_get_dma_addr(niov);
net/core/devmem.c
117
if (WARN_ON(!gen_pool_has_addr(binding->chunk_pool, dma_addr,
net/core/devmem.c
121
gen_pool_free(binding->chunk_pool, dma_addr, PAGE_SIZE);
net/core/devmem.c
265
dma_addr_t dma_addr = sg_dma_address(sg);
net/core/devmem.c
278
owner->base_dma_addr = dma_addr;
net/core/devmem.c
282
err = gen_pool_add_owner(binding->chunk_pool, dma_addr,
net/core/devmem.c
283
dma_addr, len, dev_to_node(&dev->dev),
net/core/devmem.c
91
unsigned long dma_addr;
net/core/devmem.c
96
dma_addr = gen_pool_alloc_owner(binding->chunk_pool, PAGE_SIZE,
net/core/devmem.c
98
if (!dma_addr)
net/core/netmem_priv.h
34
unsigned long dma_addr)
net/core/netmem_priv.h
36
netmem_to_nmdesc(netmem)->dma_addr = dma_addr;
net/core/page_pool.c
452
dma_addr_t dma_addr = page_pool_get_dma_addr_netmem(netmem);
net/core/page_pool.c
455
__dma_sync_single_for_device(pool->p.dev, dma_addr + pool->p.offset,
net/rds/ib_cm.c
410
dma_addr_t dma_addr, enum dma_data_direction dir)
net/rds/ib_cm.c
412
ib_dma_unmap_single(dev, dma_addr, sizeof(*hdr), dir);
net/rds/ib_cm.c
417
dma_addr_t *dma_addr, enum dma_data_direction dir)
net/rds/ib_cm.c
425
*dma_addr = ib_dma_map_single(dev, hdr, sizeof(*hdr),
net/rds/ib_cm.c
427
if (ib_dma_mapping_error(dev, *dma_addr)) {
net/rds/ib_frmr.c
226
u64 dma_addr = sg_dma_address(&ibmr->sg[i]);
net/rds/ib_frmr.c
229
if (dma_addr & ~PAGE_MASK) {
net/rds/ib_frmr.c
236
if ((dma_addr + dma_len) & ~PAGE_MASK) {
net/rds/ib_recv.c
856
dma_addr_t dma_addr = ic->i_recv_hdrs_dma[recv - ic->i_recvs];
net/rds/ib_recv.c
875
ib_dma_sync_single_for_cpu(ic->rds_ibdev->dev, dma_addr,
net/rds/ib_recv.c
980
ib_dma_sync_single_for_device(ic->rds_ibdev->dev, dma_addr,
net/smc/smc_core.c
1345
lgr->rtokens[i][lnk->link_idx].dma_addr = 0;
net/smc/smc_core.c
2642
lgr->rtokens[rtok_idx][link_idx_new].dma_addr = be64_to_cpu(nw_vaddr);
net/smc/smc_core.c
2649
u64 dma_addr = be64_to_cpu(nw_vaddr);
net/smc/smc_core.c
2663
lgr->rtokens[rtok_idx][link_idx].dma_addr = dma_addr;
net/smc/smc_core.c
2670
u64 dma_addr = be64_to_cpu(nw_vaddr);
net/smc/smc_core.c
2676
lgr->rtokens[i][lnk->link_idx].dma_addr == dma_addr &&
net/smc/smc_core.c
2686
lgr->rtokens[i][lnk->link_idx].dma_addr = dma_addr;
net/smc/smc_core.c
2702
lgr->rtokens[i][j].dma_addr = 0;
net/smc/smc_core.h
237
dma_addr_t dma_addr;
net/smc/smc_core.h
244
u64 dma_addr;
net/smc/smc_ism.c
220
if (!dmb_desc->dma_addr)
net/smc/smc_ism.c
227
dmb.dma_addr = dmb_desc->dma_addr;
net/smc/smc_ism.c
254
dmb_desc->dma_addr = dmb.dma_addr;
net/smc/smc_ism.c
286
dmb_desc->dma_addr = dmb.dma_addr;
net/smc/smc_tx.c
327
lgr->rtokens[conn->rtoken_idx][link->link_idx].dma_addr +
net/smc/smc_tx.c
363
dma_addr_t dma_addr =
net/smc/smc_tx.c
375
u64 base_addr = dma_addr;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
579
dma_addr_t dma_addr;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
583
dma_addr = ib_dma_map_page(dev, page, offset, len, DMA_TO_DEVICE);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
584
if (ib_dma_mapping_error(dev, dma_addr))
net/sunrpc/xprtrdma/svc_rdma_sendto.c
587
trace_svcrdma_dma_map_page(&ctxt->sc_cid, dma_addr, len);
net/sunrpc/xprtrdma/svc_rdma_sendto.c
588
ctxt->sc_sges[ctxt->sc_cur_sge_no].addr = dma_addr;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
594
trace_svcrdma_dma_map_err(&ctxt->sc_cid, dma_addr, len);
sound/aoa/soundbus/i2sbus/pcm.c
367
offset = runtime->dma_addr;
sound/atmel/ac97c.c
254
writel(runtime->dma_addr, chip->regs + ATMEL_PDC_TPR);
sound/atmel/ac97c.c
256
writel(runtime->dma_addr + block_size, chip->regs + ATMEL_PDC_TNPR);
sound/atmel/ac97c.c
336
writel(runtime->dma_addr, chip->regs + ATMEL_PDC_RPR);
sound/atmel/ac97c.c
338
writel(runtime->dma_addr + block_size, chip->regs + ATMEL_PDC_RNPR);
sound/atmel/ac97c.c
416
bytes -= runtime->dma_addr;
sound/atmel/ac97c.c
433
bytes -= runtime->dma_addr;
sound/atmel/ac97c.c
491
writel(runtime->dma_addr + offset, chip->regs + ATMEL_PDC_TNPR);
sound/atmel/ac97c.c
509
writel(runtime->dma_addr + offset, chip->regs + ATMEL_PDC_RNPR);
sound/core/pcm_dmaengine.c
165
substream->runtime->dma_addr,
sound/core/pcm_native.c
3961
return vm_iomap_memory(area, runtime->dma_addr, runtime->dma_bytes);
sound/isa/ad1816a/ad1816a_lib.c
213
snd_dma_program(chip->dma1, runtime->dma_addr, size,
sound/isa/ad1816a/ad1816a_lib.c
242
snd_dma_program(chip->dma2, runtime->dma_addr, size,
sound/isa/es1688/es1688_lib.c
377
snd_dma_program(chip->dma8, runtime->dma_addr, size, DMA_MODE_WRITE | DMA_AUTOINIT);
sound/isa/es1688/es1688_lib.c
433
snd_dma_program(chip->dma8, runtime->dma_addr, size, DMA_MODE_READ | DMA_AUTOINIT);
sound/isa/es18xx.c
462
snd_dma_program(chip->dma2, runtime->dma_addr, size, DMA_MODE_WRITE | DMA_AUTOINIT);
sound/isa/es18xx.c
570
snd_dma_program(chip->dma1, runtime->dma_addr, size, DMA_MODE_READ | DMA_AUTOINIT);
sound/isa/es18xx.c
634
snd_dma_program(chip->dma1, runtime->dma_addr, size, DMA_MODE_WRITE | DMA_AUTOINIT);
sound/isa/gus/gus_pcm.c
549
snd_dma_program(gus->gf1.dma2, runtime->dma_addr, gus->c_period_size, DMA_MODE_READ);
sound/isa/gus/gus_pcm.c
78
block.buf_addr = runtime->dma_addr + offset;
sound/isa/msnd/msnd.c
453
runtime->dma_addr = chip->base;
sound/isa/msnd/msnd.c
558
runtime->dma_addr = chip->base + 0x3000;
sound/isa/sb/sb16_main.c
248
snd_dma_program(dma, runtime->dma_addr, size, DMA_MODE_WRITE | DMA_AUTOINIT);
sound/isa/sb/sb16_main.c
313
snd_dma_program(dma, runtime->dma_addr, size, DMA_MODE_READ | DMA_AUTOINIT);
sound/isa/sb/sb8_main.c
158
snd_dma_program(dma, runtime->dma_addr, 1, DMA_MODE_WRITE);
sound/isa/sb/sb8_main.c
186
snd_dma_program(dma, runtime->dma_addr,
sound/isa/sb/sb8_main.c
307
snd_dma_program(dma, runtime->dma_addr,
sound/isa/wss/wss_lib.c
1010
snd_dma_program(chip->dma2, runtime->dma_addr, size, DMA_MODE_READ | DMA_AUTOINIT);
sound/isa/wss/wss_lib.c
977
snd_dma_program(chip->dma1, runtime->dma_addr, size, DMA_MODE_WRITE | DMA_AUTOINIT);
sound/parisc/harmony.c
406
h->pbuf.addr = rt->dma_addr;
sound/parisc/harmony.c
436
h->cbuf.addr = rt->dma_addr;
sound/pci/ad1889.c
360
chip->wave.addr = rt->dma_addr;
sound/pci/ad1889.c
408
chip->ramc.addr = rt->dma_addr;
sound/pci/ali5451/ali5451.c
1205
LBA = runtime->dma_addr;
sound/pci/ali5451/ali5451.c
1321
LBA = runtime->dma_addr;
sound/pci/als300.c
417
runtime->dma_addr);
sound/pci/als300.c
419
runtime->dma_addr + buffer_bytes - 1);
sound/pci/als300.c
445
runtime->dma_addr);
sound/pci/als300.c
447
runtime->dma_addr + buffer_bytes - 1);
sound/pci/als4000.c
374
snd_als4000_set_capture_dma(chip, runtime->dma_addr, size);
sound/pci/als4000.c
407
snd_als4000_set_playback_dma(chip, runtime->dma_addr, size);
sound/pci/asihpi/asihpi.c
455
params_buffer_bytes(params), runtime->dma_addr);
sound/pci/asihpi/asihpi.c
460
(unsigned long)runtime->dma_addr);
sound/pci/atiixp.c
372
addr = (u32)substream->runtime->dma_addr;
sound/pci/atiixp.c
947
dma->buf_addr = substream->runtime->dma_addr;
sound/pci/atiixp_modem.c
339
addr = (u32)substream->runtime->dma_addr;
sound/pci/atiixp_modem.c
780
dma->buf_addr = substream->runtime->dma_addr;
sound/pci/aw2/aw2-alsa.c
357
runtime->dma_addr, period_size,
sound/pci/aw2/aw2-alsa.c
384
runtime->dma_addr, period_size,
sound/pci/aw2/aw2-saa7146.c
143
unsigned long dma_addr,
sound/pci/aw2/aw2-saa7146.c
178
WRITEREG(dma_addr, BaseA2_out);
sound/pci/aw2/aw2-saa7146.c
181
WRITEREG(dma_addr + buffer_size, ProtA2_out);
sound/pci/aw2/aw2-saa7146.c
189
WRITEREG(dma_addr, BaseA1_out);
sound/pci/aw2/aw2-saa7146.c
192
WRITEREG(dma_addr + buffer_size, ProtA1_out);
sound/pci/aw2/aw2-saa7146.c
200
int stream_number, unsigned long dma_addr,
sound/pci/aw2/aw2-saa7146.c
235
WRITEREG(dma_addr, BaseA1_in);
sound/pci/aw2/aw2-saa7146.c
238
WRITEREG(dma_addr + buffer_size, ProtA1_in);
sound/pci/aw2/aw2-saa7146.h
42
unsigned long dma_addr,
sound/pci/aw2/aw2-saa7146.h
47
unsigned long dma_addr,
sound/pci/azt3328.c
1420
codec->dma_base = runtime->dma_addr;
sound/pci/azt3328.c
1428
runtime->dma_addr, count, size);
sound/pci/azt3328.c
1473
snd_azf3328_codec_setdmaa(chip, codec, runtime->dma_addr,
sound/pci/ca0106/ca0106_main.c
737
runtime->dma_addr, runtime->dma_area, table_base);
sound/pci/ca0106/ca0106_main.c
792
table_base[i*2] = runtime->dma_addr + (i * period_size_bytes);
sound/pci/ca0106/ca0106_main.c
799
snd_ca0106_ptr_write(emu, PLAYBACK_DMA_ADDR, channel, runtime->dma_addr);
sound/pci/ca0106/ca0106_main.c
844
runtime->dma_addr, runtime->dma_area, table_base);
sound/pci/ca0106/ca0106_main.c
902
snd_ca0106_ptr_write(emu, CAPTURE_DMA_ADDR, channel, runtime->dma_addr);
sound/pci/cmipci.c
768
rec->offset = runtime->dma_addr;
sound/pci/cs4281.c
738
snd_cs4281_pokeBA0(chip, dma->regDBA, runtime->dma_addr);
sound/pci/cs5535audio/cs5535audio_pcm.c
133
addr = (u32) substream->runtime->dma_addr;
sound/pci/cs5535audio/cs5535audio_pcm.c
237
dma->buf_addr = substream->runtime->dma_addr;
sound/pci/emu10k1/emu10k1x.c
438
*table_base++=runtime->dma_addr+(i*period_size_bytes);
sound/pci/emu10k1/emu10k1x.c
448
snd_emu10k1x_ptr_write(emu, PLAYBACK_DMA_ADDR, voice, runtime->dma_addr);
sound/pci/emu10k1/emu10k1x.c
615
snd_emu10k1x_ptr_write(emu, CAPTURE_DMA_ADDR, 0, runtime->dma_addr);
sound/pci/emu10k1/emupcm.c
528
snd_emu10k1_ptr_write(emu, epcm->capture_ba_reg, 0, runtime->dma_addr);
sound/pci/emu10k1/p16v.c
256
runtime->dma_addr, runtime->dma_area, table_base);
sound/pci/emu10k1/p16v.c
285
table_base[i*2]=runtime->dma_addr+(i*period_size_bytes);
sound/pci/emu10k1/p16v.c
292
snd_emu10k1_ptr20_write(emu, PLAYBACK_DMA_ADDR, channel, runtime->dma_addr);
sound/pci/emu10k1/p16v.c
334
snd_emu10k1_ptr20_write(emu, CAPTURE_DMA_ADDR, channel, runtime->dma_addr);
sound/pci/ens1370.c
870
outl(runtime->dma_addr, ES_REG(ensoniq, DAC1_FRAME));
sound/pci/ens1370.c
911
outl(runtime->dma_addr, ES_REG(ensoniq, DAC2_FRAME));
sound/pci/ens1370.c
951
outl(runtime->dma_addr, ES_REG(ensoniq, ADC_FRAME));
sound/pci/es1938.c
606
chip->dma1_start = runtime->dma_addr;
sound/pci/es1938.c
654
chip->dma2_start = runtime->dma_addr;
sound/pci/es1938.c
691
chip->dma1_start = runtime->dma_addr;
sound/pci/fm801.c
479
chip->ply_buffer = runtime->dma_addr;
sound/pci/fm801.c
505
chip->cap_buffer = runtime->dma_addr;
sound/pci/ice1712/ice1712.c
1006
outl(substream->runtime->dma_addr, ICEMT(ice, PLAYBACK_ADDR));
sound/pci/ice1712/ice1712.c
1028
outl(substream->runtime->dma_addr, ICEMT(ice, CAPTURE_ADDR));
sound/pci/ice1712/ice1712.c
566
outl(runtime->dma_addr, ice->ddma_port + 0);
sound/pci/ice1712/ice1712.c
595
ice->playback_con_virt_addr[substream->number] = runtime->dma_addr;
sound/pci/ice1712/ice1712.c
598
snd_ice1712_ds_write(ice, chn, ICE1712_DSC_ADDR0, runtime->dma_addr);
sound/pci/ice1712/ice1712.c
600
snd_ice1712_ds_write(ice, chn, ICE1712_DSC_ADDR1, runtime->dma_addr + (runtime->periods > 1 ? period_size + 1 : 0));
sound/pci/ice1712/ice1712.c
627
outl(ice->capture_con_virt_addr = runtime->dma_addr, ICEREG(ice, CONCAP_ADDR));
sound/pci/ice1712/ice1724.c
753
outl(substream->runtime->dma_addr, ICEMT1724(ice, PLAYBACK_ADDR));
sound/pci/ice1712/ice1724.c
784
if (ptr < substream->runtime->dma_addr) {
sound/pci/ice1712/ice1724.c
788
ptr -= substream->runtime->dma_addr;
sound/pci/ice1712/ice1724.c
818
outl(substream->runtime->dma_addr, ice->profi_port + reg->addr);
sound/pci/ice1712/ice1724.c
836
ptr -= substream->runtime->dma_addr;
sound/pci/intel8x0.c
973
ichdev->physbuf = runtime->dma_addr;
sound/pci/intel8x0m.c
558
ichdev->physbuf = runtime->dma_addr;
sound/pci/maestro3.c
1378
s->buffer_addr = substream->runtime->dma_addr;
sound/pci/mixart/mixart.c
632
bufferinfo[i].buffer_address = subs->runtime->dma_addr;
sound/pci/nm256/nm256.c
812
runtime->dma_addr = s->bufptr_addr;
sound/pci/oxygen/oxygen_pcm.c
307
(u32)substream->runtime->dma_addr);
sound/pci/oxygen/oxygen_pcm.c
602
return bytes_to_frames(runtime, curr_addr - (u32)runtime->dma_addr);
sound/pci/pcxhr/pcxhr.c
635
chip->chip_idx, (void *)(long)subs->runtime->dma_addr,
sound/pci/pcxhr/pcxhr.c
647
rmh.cmd[2] = subs->runtime->dma_addr >> 24;
sound/pci/pcxhr/pcxhr.c
651
rmh.cmd[3] = subs->runtime->dma_addr & MASK_DSP_WORD;
sound/pci/rme32.c
645
runtime->dma_addr = rme32->port + RME32_IO_DATA_BUFFER;
sound/pci/rme32.c
693
runtime->dma_addr = rme32->port + RME32_IO_DATA_BUFFER;
sound/pci/rme96.c
1024
runtime->dma_addr = rme96->port + RME96_IO_REC_BUFFER;
sound/pci/rme96.c
964
runtime->dma_addr = rme96->port + RME96_IO_PLAY_BUFFER;
sound/pci/sis7019.c
493
u32 format, dma_addr, control, sso_eso, delta, reg;
sound/pci/sis7019.c
510
dma_addr = runtime->dma_addr;
sound/pci/sis7019.c
533
writel(dma_addr, ctrl_base + SIS_PLAY_DMA_BASE);
sound/pci/sis7019.c
790
u32 format, dma_addr, control;
sound/pci/sis7019.c
804
dma_addr = runtime->dma_addr;
sound/pci/sis7019.c
821
writel(dma_addr, rec_base + SIS_CAPTURE_DMA_BASE);
sound/pci/sonicvibes.c
683
snd_sonicvibes_setdmaa(sonic, runtime->dma_addr, size);
sound/pci/sonicvibes.c
707
snd_sonicvibes_setdmac(sonic, runtime->dma_addr, size);
sound/pci/trident/trident_main.c
1017
voice->LBA = runtime->dma_addr;
sound/pci/trident/trident_main.c
1143
voice->LBA = runtime->dma_addr;
sound/pci/trident/trident_main.c
1217
voice->LBA = runtime->dma_addr;
sound/pci/trident/trident_main.c
1363
LBAO = runtime->dma_addr;
sound/pci/trident/trident_main.c
1416
voice->LBA = runtime->dma_addr;
sound/pci/trident/trident_main.c
918
voice->LBA = runtime->dma_addr;
sound/pci/trident/trident_memory.c
221
addr = runtime->dma_addr;
sound/pci/ymfpci/ymfpci_main.c
515
bank->base = cpu_to_le32(runtime->dma_addr);
sound/pci/ymfpci/ymfpci_main.c
703
bank->base = cpu_to_le32(runtime->dma_addr);
sound/ppc/pmac.c
220
offset = runtime->dma_addr;
sound/ppc/snd_ps3.c
220
uint32_t dma_addr;
sound/ppc/snd_ps3.c
249
dma_addr = card->null_buffer_start_dma_addr;
sound/ppc/snd_ps3.c
251
dma_addr =
sound/ppc/snd_ps3.c
258
dma_addr));
sound/ppc/snd_ps3.c
585
card->dma_start_bus_addr[SND_PS3_CH_L] = runtime->dma_addr;
sound/ppc/snd_ps3.c
592
runtime->dma_addr + (runtime->dma_bytes / 2);
sound/soc/amd/acp-pcm-dma.c
1007
rtd->dma_addr = runtime->dma_addr;
sound/soc/amd/acp-pcm-dma.c
336
acp_pte_config(acp_mmio, rtd->dma_addr, rtd->num_of_pages,
sound/soc/amd/acp.h
138
dma_addr_t dma_addr;
sound/soc/amd/ps/acp63.h
241
dma_addr_t dma_addr;
sound/soc/amd/ps/acp63.h
268
dma_addr_t dma_addr;
sound/soc/amd/ps/ps-pdm-dma.c
163
addr = rtd->dma_addr;
sound/soc/amd/ps/ps-pdm-dma.c
229
rtd->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/ps/ps-sdw-dma.c
214
addr = stream->dma_addr;
sound/soc/amd/ps/ps-sdw-dma.c
416
stream->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/raven/acp3x-pcm-dma.c
114
addr = rtd->dma_addr;
sound/soc/amd/raven/acp3x-pcm-dma.c
289
rtd->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/raven/acp3x.h
109
dma_addr_t dma_addr;
sound/soc/amd/renoir/acp3x-pdm-dma.c
182
addr = rtd->dma_addr;
sound/soc/amd/renoir/acp3x-pdm-dma.c
251
rtd->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/renoir/rn_acp3x.h
69
dma_addr_t dma_addr;
sound/soc/amd/vangogh/acp5x-pcm-dma.c
110
addr = rtd->dma_addr;
sound/soc/amd/vangogh/acp5x-pcm-dma.c
284
rtd->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/vangogh/acp5x.h
105
dma_addr_t dma_addr;
sound/soc/amd/yc/acp6x-pdm-dma.c
159
addr = rtd->dma_addr;
sound/soc/amd/yc/acp6x-pdm-dma.c
227
rtd->dma_addr = substream->runtime->dma_addr;
sound/soc/amd/yc/acp6x.h
86
dma_addr_t dma_addr;
sound/soc/atmel/atmel-pcm-pdc.c
151
prtd->dma_buffer = runtime->dma_addr;
sound/soc/atmel/atmel-pcm-pdc.c
152
prtd->dma_buffer_end = runtime->dma_addr + runtime->dma_bytes;
sound/soc/au1x/dbdma2.c
206
(unsigned long)runtime->dma_addr, runtime->dma_bytes,
sound/soc/au1x/dbdma2.c
221
pcd->dma_area_s = pcd->dma_area = runtime->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
153
dma_desc->dma_addr = runtime->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
168
regmap_write(regmap_i2s, regaddr_descaddr, dma_desc->dma_addr);
sound/soc/bcm/bcm63xx-pcm-whistler.c
181
prtd->dma_addr_next = substream->runtime->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
184
prtd->dma_addr_next - substream->runtime->dma_addr);
sound/soc/bcm/bcm63xx-pcm-whistler.c
20
dma_addr_t dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
26
dma_addr_t dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
274
dma_desc->dma_addr +=
sound/soc/bcm/bcm63xx-pcm-whistler.c
278
if (dma_desc->dma_addr - runtime->dma_addr >=
sound/soc/bcm/bcm63xx-pcm-whistler.c
280
dma_desc->dma_addr = runtime->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
284
prtd->dma_addr = dma_desc->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
288
dma_desc->dma_addr);
sound/soc/bcm/bcm63xx-pcm-whistler.c
323
dma_desc->dma_addr +=
sound/soc/bcm/bcm63xx-pcm-whistler.c
328
if (dma_desc->dma_addr - runtime->dma_addr >=
sound/soc/bcm/bcm63xx-pcm-whistler.c
330
dma_desc->dma_addr = runtime->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
334
prtd->dma_addr = dma_desc->dma_addr;
sound/soc/bcm/bcm63xx-pcm-whistler.c
338
dma_desc->dma_addr);
sound/soc/bcm/cygnus-pcm.c
653
start = runtime->dma_addr;
sound/soc/fsl/fsl_asrc_dma.c
70
pair->dma_chan[!dir], runtime->dma_addr,
sound/soc/fsl/fsl_asrc_m2m.c
103
u32 dma_addr, dma_addr_t buf_addr, u32 buf_len,
sound/soc/fsl/fsl_asrc_m2m.c
135
slave_config.dst_addr = dma_addr;
sound/soc/fsl/fsl_asrc_m2m.c
140
slave_config.src_addr = dma_addr;
sound/soc/fsl/fsl_qmc_audio.c
105
prtd->ch_dma_addr_start = runtime->dma_addr;
sound/soc/fsl/fsl_qmc_audio.c
107
prtd->ch_dma_addr_end = runtime->dma_addr + prtd->ch_dma_offset;
sound/soc/fsl/imx-pcm-rpmsg.c
365
msg->s_msg.param.buffer_addr = substream->runtime->dma_addr;
sound/soc/fsl/mpc5200_dma.c
68
bd->data[0] = s->runtime->dma_addr + (s->period_next * s->period_bytes);
sound/soc/intel/atom/sst-mfld-platform-pcm.c
130
u32 buffer_addr = substream->runtime->dma_addr;
sound/soc/kirkwood/kirkwood-dma.c
183
unsigned long addr = substream->runtime->dma_addr;
sound/soc/kirkwood/kirkwood-dma.c
211
writel(runtime->dma_addr, priv->io + KIRKWOOD_PLAY_BUF_ADDR);
sound/soc/kirkwood/kirkwood-dma.c
215
writel(runtime->dma_addr, priv->io + KIRKWOOD_REC_BUF_ADDR);
sound/soc/loongson/loongson_dma.c
163
mem_addr = runtime->dma_addr;
sound/soc/loongson/loongson_dma.c
210
x = bytes_to_frames(runtime, addr - runtime->dma_addr);
sound/soc/mediatek/common/mtk-afe-fe-dai.c
138
&substream->runtime->dma_addr,
sound/soc/mediatek/common/mtk-afe-fe-dai.c
148
substream->runtime->dma_addr,
sound/soc/mediatek/common/mtk-afe-fe-dai.c
403
dma_addr_t dma_addr,
sound/soc/mediatek/common/mtk-afe-fe-dai.c
407
int msb_at_bit33 = upper_32_bits(dma_addr) ? 1 : 0;
sound/soc/mediatek/common/mtk-afe-fe-dai.c
408
unsigned int phys_buf_addr = lower_32_bits(dma_addr);
sound/soc/mediatek/common/mtk-afe-fe-dai.c
409
unsigned int phys_buf_addr_upper_32 = upper_32_bits(dma_addr);
sound/soc/mediatek/common/mtk-afe-fe-dai.c
412
memif->dma_addr = dma_addr;
sound/soc/mediatek/common/mtk-afe-fe-dai.h
41
dma_addr_t dma_addr,
sound/soc/mediatek/common/mtk-base-afe.h
149
dma_addr_t dma_addr;
sound/soc/mediatek/mt8365/mt8365-afe-pcm.c
605
memif->phys_buf_addr = lower_32_bits(substream->runtime->dma_addr);
sound/soc/meson/aiu-fifo.c
105
end = runtime->dma_addr + runtime->dma_bytes - fifo->fifo_block;
sound/soc/meson/aiu-fifo.c
107
runtime->dma_addr);
sound/soc/meson/aiu-fifo.c
109
runtime->dma_addr);
sound/soc/meson/aiu-fifo.c
43
return bytes_to_frames(runtime, addr - (unsigned int)runtime->dma_addr);
sound/soc/meson/axg-fifo.c
107
return bytes_to_frames(runtime, addr - (unsigned int)runtime->dma_addr);
sound/soc/meson/axg-fifo.c
123
end_ptr = runtime->dma_addr + runtime->dma_bytes - AXG_FIFO_BURST;
sound/soc/meson/axg-fifo.c
124
regmap_write(fifo->map, FIFO_START_ADDR, runtime->dma_addr);
sound/soc/meson/axg-fifo.c
169
regmap_write(fifo->map, FIFO_INIT_ADDR, runtime->dma_addr);
sound/soc/qcom/lpass-platform.c
618
runtime->dma_addr);
sound/soc/qcom/lpass-platform.c
905
(runtime->dma_addr + offset) >> PAGE_SHIFT,
sound/soc/renesas/dma-sh7760.c
198
pr_debug("PCM data: addr %pad len %zu\n", &runtime->dma_addr,
sound/soc/renesas/fsi.c
1297
substream->runtime->dma_addr,
sound/soc/renesas/rz-ssi.c
731
dma_paddr = runtime->dma_addr + frames_to_bytes(runtime,
sound/soc/renesas/siu_pcm.c
221
buff = (dma_addr_t)PERIOD_OFFSET(rt->dma_addr,
sound/soc/renesas/siu_pcm.c
230
(dma_addr_t)PERIOD_OFFSET(rt->dma_addr,
sound/soc/renesas/siu_pcm.c
396
(unsigned long)rt->dma_addr, siu_stream->buf_bytes,
sound/soc/renesas/siu_pcm.c
467
ptr = PERIOD_OFFSET(rt->dma_addr,
sound/soc/renesas/siu_pcm.c
469
siu_stream->period_bytes) - rt->dma_addr;
sound/soc/samsung/idma.c
157
prtd->start = prtd->pos = runtime->dma_addr;
sound/soc/samsung/idma.c
160
prtd->end = runtime->dma_addr + runtime->dma_bytes;
sound/soc/samsung/idma.c
253
(runtime->dma_addr + offset) >> PAGE_SHIFT,
sound/soc/sof/amd/acp-loader.c
133
addr = adata->dma_addr;
sound/soc/sof/amd/acp-loader.c
237
adata->dma_addr);
sound/soc/sof/amd/acp-loader.c
79
&adata->dma_addr,
sound/soc/sof/amd/acp.h
254
dma_addr_t dma_addr;
sound/soc/sprd/sprd-pcm-dma.c
239
config.src_addr = runtime->dma_addr +
sound/soc/sprd/sprd-pcm-dma.c
245
config.dst_addr = runtime->dma_addr +
sound/soc/sprd/sprd-pcm-dma.c
255
sg_dma_address(sgt) = runtime->dma_addr +
sound/soc/sprd/sprd-pcm-dma.c
405
pointer[i] = state.residue - runtime->dma_addr -
sound/soc/uniphier/aio-dma.c
124
ret = aiodma_rb_set_buffer(sub, runtime->dma_addr,
sound/soc/uniphier/aio-dma.c
125
runtime->dma_addr + runtime->dma_bytes,
sound/soc/uniphier/aio-dma.c
149
aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes,
sound/soc/uniphier/aio-dma.c
183
aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes, bytes);
sound/soc/uniphier/aio-dma.c
201
substream->runtime->dma_addr >> PAGE_SHIFT,
sound/soc/uniphier/aio-dma.c
41
aiodma_rb_sync(sub, runtime->dma_addr, runtime->dma_bytes, bytes);
sound/soc/ux500/ux500_pcm.c
37
dma_addr_t dma_addr;
sound/soc/ux500/ux500_pcm.c
41
dma_addr = snd_dma_params->addr;
sound/soc/ux500/ux500_pcm.c
54
slave_config->dst_addr = dma_addr;
sound/soc/ux500/ux500_pcm.c
56
slave_config->src_addr = dma_addr;
sound/soc/xilinx/xlnx_formatter_pcm.c
500
low = lower_32_bits(runtime->dma_addr);
sound/soc/xilinx/xlnx_formatter_pcm.c
501
high = upper_32_bits(runtime->dma_addr);
sound/sparc/cs4231.c
438
runtime->dma_addr + offset, period_size))
sound/sparc/cs4231.c
962
ptr -= substream->runtime->dma_addr;
sound/sparc/cs4231.c
978
ptr -= substream->runtime->dma_addr;
sound/sparc/dbri.c
760
u32 dma_addr;
sound/sparc/dbri.c
774
dma_addr = dvma_addr + dbri_dma_off(intr, 0);
sound/sparc/dbri.c
775
dbri->dma->intr[0] = dma_addr;
sound/sparc/dbri.c
783
*(cmd++) = dma_addr;
sound/sparc/dbri.c
788
dma_addr = dvma_addr + dbri_dma_off(cmd, 0);
sound/sparc/dbri.c
789
sbus_writel(dma_addr, dbri->regs + REG8);
sound/spi/at73c213.c
261
(long)runtime->dma_addr);
sound/spi/at73c213.c
265
(long)runtime->dma_addr + block_size);
sound/spi/at73c213.c
306
- (unsigned long)runtime->dma_addr;
sound/spi/at73c213.c
373
(long)runtime->dma_addr + offset);
sound/x86/intel_hdmi_audio.c
870
u32 addr = substream->runtime->dma_addr + ofs;