drm_mm_node_allocated
return drm_mm_node_allocated(&node->mm_nodes[0]);
if (drm_mm_node_allocated(&node->mm_nodes[0]))
if (drm_mm_node_allocated(mm_node))
if (drm_mm_node_allocated(hole_node)) {
DRM_MM_BUG_ON(!drm_mm_node_allocated(node));
DRM_MM_BUG_ON(!drm_mm_node_allocated(node));
if (!drm_mm_node_allocated(&node->vm_node))
if (drm_mm_node_allocated(&node->vm_node)) {
if (drm_mm_node_allocated(&vma->node)) {
GEM_BUG_ON(drm_mm_node_allocated(&vma->node) &&
if (!drm_mm_node_allocated(&cache->node))
if (drm_mm_node_allocated(&cache->node)) {
if (drm_mm_node_allocated(&cache->node)) {
if (drm_mm_node_allocated(&vma->node) &&
return drm_mm_node_allocated(&node->node);
if (!drm_mm_node_allocated(&ggtt->error_capture))
GEM_BUG_ON(!drm_mm_node_allocated(&obj->base.vma_node.vm_node));
if (drm_mm_node_allocated(&vma->node) && i915_vma_offset(vma) != addr) {
if (drm_mm_node_allocated(&ggtt->error_capture))
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
if (drm_mm_node_allocated(&ggtt->error_capture))
if (drm_mm_node_allocated(&ggtt->uc_fw))
if (drm_mm_node_allocated(&ggtt->error_capture))
if (drm_mm_node_allocated(&ggtt->error_capture)) {
if (!drm_mm_node_allocated(&ggtt->error_capture))
GEM_BUG_ON(!drm_mm_node_allocated(node));
if (!drm_mm_node_allocated(&vma->node))
GEM_BUG_ON(!drm_mm_node_allocated(node));
if (drm_mm_node_allocated(node)) {
if (drm_mm_node_allocated(node)) {
if (drm_mm_node_allocated(&node)) {
if (drm_mm_node_allocated(&node)) {
GEM_BUG_ON(!drm_mm_node_allocated(node));
GEM_BUG_ON(drm_mm_node_allocated(node));
GEM_BUG_ON(drm_mm_node_allocated(node));
if (drm_mm_node_allocated(&ggtt->error_capture)) {
if (drm_mm_node_allocated(&orig_mm)) {
if (drm_mm_node_allocated(&orig_mm))
if (drm_mm_node_allocated(&orig_mm))
if (!drm_mm_node_allocated(node))
if (!drm_mm_node_allocated(&vma->node))
GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
if (!drm_mm_node_allocated(&vma->node))
if (!drm_mm_node_allocated(&vma->node))
if (!drm_mm_node_allocated(&vma->node))
if (!drm_mm_node_allocated(&vma->node))
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
if (!drm_mm_node_allocated(&vma->node))
GEM_BUG_ON(!drm_mm_node_allocated(node));
GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
return drm_mm_node_allocated(node) && node->color != color;
if (drm_mm_node_allocated(&hole))
if (!drm_mm_node_allocated(&vma->node)) {
GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
if (!drm_mm_node_allocated(&vma->node)) {
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
GEM_BUG_ON(!drm_mm_node_allocated(&vma->node));
if (!drm_mm_node_allocated(&vma->node) ||
__func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
if (!drm_mm_node_allocated(&vma->node) ||
if (!drm_mm_node_allocated(&vma->node) ||
__func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
if (!drm_mm_node_allocated(&vma->node) ||
__func__, p->name, vma->node.start, vma->node.size, drm_mm_node_allocated(&vma->node),
if (!drm_mm_node_allocated(&vma->node) ||
GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
if (!drm_mm_node_allocated(&vma->node) ||
if (!drm_mm_node_allocated(&vma->node) ||
if (!drm_mm_node_allocated(&vma->node) ||
if (drm_mm_node_allocated(&fw_obj->fw_mm_node)) {
if (!drm_mm_node_allocated(&fw_obj->fw_mm_node)) {
if (drm_mm_node_allocated(&fw_obj->fw_mm_node)) {
if (drm_mm_node_allocated(&mapping->mmnode))
if (!drm_mm_node_allocated(node) || node->mm != mm) {
drm_mm_node_allocated(&bo->base.vma_node.vm_node));
drm_mm_node_allocated(&bo->base.vma_node.vm_node));
if (drm_mm_node_allocated(&vc4_state->mm)) {
if (drm_mm_node_allocated(&vc4->hvs->mitchell_netravali_filter))
if (drm_mm_node_allocated(&vc4_state->lbm)) {
if (!drm_mm_node_allocated(&vc4_state->lbm)) {
if (drm_mm_node_allocated(&vram->vram_node))
unmap = drm_mm_node_allocated(&vram->vram_node);
if (!drm_mm_node_allocated(node))
xe_tile_assert(ggtt->tile, !drm_mm_node_allocated(&node->base));
xe_tile_assert(tile, drm_mm_node_allocated(node));
return drm_mm_node_allocated(&node->base);
if (drm_mm_node_allocated(&node->vm_node))