device_iommu_mapped
if (device_iommu_mapped(dev))
if (device_iommu_mapped(dev)) {
if (device_iommu_mapped(dev)) {
if (!user_address || device_iommu_mapped(&hdev->pdev->dev)) {
if (!device_iommu_mapped(&hdev->pdev->dev))
device_iommu_mapped(&intel_private.pcidev->dev));
if (!device_iommu_mapped(&GET_DEV(accel_dev))) {
if (device_iommu_mapped(&pdev->dev))
if (device_iommu_mapped(display->drm->dev))
if (device_iommu_mapped(i915->drm.dev))
if (!device_iommu_mapped(dev))
if (device_iommu_mapped(mdp_dev))
else if (mdss_dev && device_iommu_mapped(mdss_dev))
if (device_iommu_mapped(dev)) {
if (!device_iommu_mapped(dev))
if (host1x_drm_wants_iommu(dev) && device_iommu_mapped(dma_dev)) {
if (device_iommu_mapped(client->base.dev) && client->ops->can_use_memory_ctx) {
!device_iommu_mapped(&ctx->dev)) {
bool need_bounce = device_iommu_mapped(dev) && xfer->rnw && (xfer->data_len & 3);
if (!device_iommu_mapped(&pdev->dev)) {
if (!device_iommu_mapped(tqspi->dev)) {
if (of_dma_is_coherent(dev->of_node) || device_iommu_mapped(dev)) {