GEM_WARN_ON
GEM_WARN_ON(pe[n].type != I915_GEM_ENGINE_TYPE_INVALID);
GEM_WARN_ON(ctx && pc);
GEM_WARN_ON(1);
GEM_WARN_ON(pin_count < 0);
if (GEM_WARN_ON(obj->mm.region != mr))
if (GEM_WARN_ON(type == I915_MAP_WC && !pat_enabled()))
if (GEM_WARN_ON(i915_gem_object_needs_bit17_swizzle(obj)))
GEM_WARN_ON(ret);
GEM_WARN_ON(ret);
if (GEM_WARN_ON(!IS_ALIGNED(size, mem->min_page_size)) ||
GEM_WARN_ON(!IS_ALIGNED(offset, mem->min_page_size)))
GEM_WARN_ON(apply->ww);
GEM_WARN_ON(IS_DGFX(i915));
if (GEM_WARN_ON(resource_size(&mem->region) == 0))
GEM_WARN_ON(!i915_ttm_cpu_maps_iomem(bo->resource));
GEM_WARN_ON(obj->userfault_count);
GEM_WARN_ON(st->sgl);
GEM_WARN_ON(!i915_ttm_cpu_maps_iomem(res));
GEM_WARN_ON(ret);
GEM_WARN_ON(obj->ttm.cached_io_rsgt);
GEM_WARN_ON(bo->ttm);
if (GEM_WARN_ON(i915_ttm_is_ghost_object(bo))) {
if (GEM_WARN_ON(!src_bo->resource || !dst_bo->resource))
GEM_WARN_ON(ret);
GEM_WARN_ON(err);
if (GEM_WARN_ON(i915_gem_object_has_pinned_pages(obj)))
GEM_WARN_ON(obj->userptr.page_ref);
if (GEM_WARN_ON(!wakeref))
GEM_WARN_ON(!i915_active_acquire_if_busy(&ce->ring->vma->active));
GEM_WARN_ON(!i915_active_acquire_if_busy(&ce->state->active));
GEM_WARN_ON(instance != OTHER_GSC_INSTANCE);
GEM_WARN_ON(snprintf(engine->name, sizeof(engine->name), "%s'%u",
if (GEM_WARN_ON(!*execlists->pending)) {
if (GEM_WARN_ON(!*execlists->active)) {
if (GEM_WARN_ON(ppgtt->vm.total < ggtt->vm.total)) {
GEM_WARN_ON(kref_read(&ggtt->vm.resv_ref) != 1);
GEM_WARN_ON(pci_resource_len(pdev, GEN4_GTTMMADR_BAR) != gen6_gttmmadr_size(i915));
GEM_WARN_ON(i915_vma_unbind(vma));
if (GEM_WARN_ON(range_overflows_t(resource_size_t,
GEM_WARN_ON(!node->pinned);
GEM_WARN_ON(!HAS_MSLICE_STEERING(gt->i915));
GEM_WARN_ON(!HAS_MSLICE_STEERING(gt->i915));
if (GEM_WARN_ON(lmem_size < flat_ccs_base))
if (GEM_WARN_ON(&target->link == &tl->requests))
GEM_WARN_ON(eu_mask && __fls(eu_mask) >= sseu->max_eus_per_subslice);
GEM_WARN_ON(slice > 0);
if (GEM_WARN_ON(show_unknown))
if (GEM_WARN_ON(len >= PAGE_SIZE))
if (GEM_WARN_ON(!intel_uc_fw_is_loaded(gsc_fw)))
GEM_WARN_ON(!guc_to_gt(guc)->uc.reset_in_progress);
GEM_WARN_ON(ret); /* Unexpected */
GEM_WARN_ON(ret); /* Unexpected */
GEM_WARN_ON(ret); /* Unexpected */
GEM_WARN_ON(!ce->guc_state.prio_count[guc_prio]);
GEM_WARN_ON(!ce->guc_state.prio_count[guc_prio]);
GEM_WARN_ON(intel_uc_fw_is_loaded(&huc->fw));
GEM_WARN_ON(atomic_read(&guc->outstanding_submission_g2h));
if (GEM_WARN_ON(atomic_read(&mm->pincount)))
if (GEM_WARN_ON(!list_empty(&vgpu->gtt.ppgtt_mm_list_head)))
if (GEM_WARN_ON(!radix_tree_empty(&vgpu->gtt.spt_tree))) {
if (GEM_WARN_ON(e->type != GTT_TYPE_PPGTT_PDE_ENTRY))
if (GEM_WARN_ON(e->type != GTT_TYPE_PPGTT_PDE_ENTRY))
GEM_WARN_ON(obj->client);
GEM_WARN_ON(!list_empty(&obj->client_link));
GEM_WARN_ON(!ww);
#define GEM_DEBUG_WARN_ON(expr) GEM_WARN_ON(expr)
if (GEM_WARN_ON(!i915_vma_is_pinned(vma)))
if (GEM_WARN_ON(!r->sgt.sgp))
GEM_WARN_ON(vma->resource);
GEM_WARN_ON(async && !vma->resource->bi.pages_rsgt);
GEM_WARN_ON(!vma_flags);
GEM_WARN_ON(!list_empty_careful(&mem->objects.list));
GEM_WARN_ON(ret);
if (GEM_WARN_ON(!intel_pxp_is_enabled(pxp)))
GEM_WARN_ON(intel_pxp_is_active(pxp));
GEM_WARN_ON(pxp->arb_is_valid);
if (GEM_WARN_ON(!node)) {
GEM_WARN_ON(msm_obj->vaddr);
GEM_WARN_ON(!msm_obj->pages);
GEM_WARN_ON(msm_obj->madv != MSM_MADV_DONTNEED);
GEM_WARN_ON(msm_obj->pin_count);
if (GEM_WARN_ON(msm_obj->madv != MSM_MADV_WILLNEED)) {
GEM_WARN_ON(!list_is_singular(&vm_bo->list.gpuva));
GEM_WARN_ON(vma->va.addr < range_start);
GEM_WARN_ON((vma->va.addr + obj->size) > range_end);
GEM_WARN_ON(msm_obj->pin_count < 0);
GEM_WARN_ON(msm_obj->pin_count < 0);
} else if (GEM_WARN_ON(vma->va.addr != iova)) {
GEM_WARN_ON(msm_obj->vmap_count < 1);
GEM_WARN_ON(!is_purgeable(msm_obj));
GEM_WARN_ON(is_unevictable(msm_obj));
if (!msm_obj->vaddr || GEM_WARN_ON(!is_vunmapable(msm_obj)))
if (GEM_WARN_ON(ret))
if (GEM_WARN_ON(!vma->va.addr))
GEM_WARN_ON(msm_vma->mapped);
GEM_WARN_ON((range_end - range_start) > obj->size);
GEM_WARN_ON(drm_gpuvm_bo_put(vm_bo));