start_dma
u64 start_dma; /* Start of available DMA addresses */
dma_start = PAGE_ALIGN(zdev->start_dma);
zdev->start_dma = response->sdma;
int start_dma;
start_dma = 1;
start_dma = 0;
if (start_dma) {
zdev->end_dma - zdev->start_dma + 1);
if (aperture_size <= (ZPCI_TABLE_SIZE_RT - zdev->start_dma)) {
} else if (aperture_size <= (ZPCI_TABLE_SIZE_RS - zdev->start_dma) &&
aperture_size = ZPCI_TABLE_SIZE_RT - zdev->start_dma;
zdev->end_dma = zdev->start_dma + aperture_size - 1;
rc = zpci_register_ioat(zdev, 0, zdev->start_dma,
rc = zpci_register_ioat(zdev, 0, zdev->start_dma,
domain->geometry.aperture_end < zdev->start_dma))
if (zdev->start_dma) {
region = iommu_alloc_resv_region(0, zdev->start_dma, 0,
if (zdev->start_dma > zdev->end_dma)
return zpci_refresh_trans((u64)zdev->fh << 32, zdev->start_dma,
zdev->end_dma - zdev->start_dma + 1);
start_dma(isi, buf);
start_dma(isi, isi->active);
start_dma(stream, buf);
start_dma(stream, buf);
dma_addr_t start_dma;
start_dma = cdnsp_trb_virt_to_dma(start_seg, start_trb);
if (start_dma == 0)
trace_cdnsp_looking_trb_in_td(suspect_dma, start_dma,
if (start_dma <= end_trb_dma) {
if (suspect_dma >= start_dma &&
if ((suspect_dma >= start_dma &&
if (suspect_dma >= start_dma && suspect_dma <= end_seg_dma)
start_dma = cdnsp_trb_virt_to_dma(cur_seg, &cur_seg->trbs[0]);
(void) start_dma(ep, req);
status = start_dma(ep, req);
(void) start_dma(ep, req);
start_dma(ep, req);
start_dma(ep, req);
start_dma(ep, list_entry(ep->queue.next,
.start_dma = zdev->start_dma,
__u64 start_dma; /* Start of available DMA addresses */
start_dma(stream->dma);