device_iommu_capable
!device_iommu_capable(&pdev->dev, IOMMU_CAP_CACHE_COHERENCY))
if (!device_iommu_capable(dev, IOMMU_CAP_CACHE_COHERENCY)) {
(!device_iommu_capable(dev, IOMMU_CAP_DEFERRED_FLUSH) || iommu_dma_init_fq(domain)))
EXPORT_SYMBOL_GPL(device_iommu_capable);
if (device_iommu_capable(idev->dev, IOMMU_CAP_DIRTY_TRACKING))
if (!device_iommu_capable(dev, IOMMU_CAP_CACHE_COHERENCY))
device_iommu_capable(dev, IOMMU_CAP_ENFORCE_CACHE_COHERENCY);
!device_iommu_capable(idev->dev, IOMMU_CAP_DIRTY_TRACKING))
!device_iommu_capable(&pdev->dev, IOMMU_CAP_PRE_BOOT_PROTECTION))
if (!device_iommu_capable(device->dev,
return device_iommu_capable(device->dev,
!device_iommu_capable(device->dev, IOMMU_CAP_CACHE_COHERENCY)) {
if (!device_iommu_capable(dma_dev, IOMMU_CAP_CACHE_COHERENCY)) {
extern bool device_iommu_capable(struct device *dev, enum iommu_cap cap);