drivers/gpu/drm/i915/gt/intel_gt.h
124
return guc_to_gt(guc)->i915;
drivers/gpu/drm/i915/gt/uc/intel_guc.c
111
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
136
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
145
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
156
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
167
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
181
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
244
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
294
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
385
struct intel_uncore *uncore = guc_to_gt(guc)->uncore;
drivers/gpu/drm/i915/gt/uc/intel_guc.c
405
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
46
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
514
struct intel_uncore *uncore = guc_to_gt(guc)->uncore;
drivers/gpu/drm/i915/gt/uc/intel_guc.c
68
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
779
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
845
intel_gt_coherent_map_type(guc_to_gt(guc),
drivers/gpu/drm/i915/gt/uc/intel_guc.c
85
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
916
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
953
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc.c
96
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
163
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
206
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
435
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
467
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
531
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
621
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
706
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
838
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
887
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
916
guc_mapping_table_init(guc_to_gt(guc), &info_map);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
295
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
613
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1329
GEM_WARN_ON(!guc_to_gt(guc)->uc.reset_in_progress);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1397
intel_klog_error_capture(guc_to_gt(guc), (intel_engine_mask_t)~0U);
drivers/gpu/drm/i915/gt/uc/intel_guc_fw.c
166
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_fw.c
299
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_log.c
789
with_intel_runtime_pm(guc_to_gt(guc)->uncore->rpm, wakeref)
drivers/gpu/drm/i915/gt/uc/intel_guc_print.h
13
gt_##_level(guc_to_gt(_guc), "GUC: " _fmt, ##__VA_ARGS__)
drivers/gpu/drm/i915/gt/uc/intel_guc_print.h
40
gt_WARN(guc_to_gt(_guc), _cond, "GUC: " _fmt, ##__VA_ARGS__)
drivers/gpu/drm/i915/gt/uc/intel_guc_print.h
43
gt_WARN_ONCE(guc_to_gt(_guc), _cond, "GUC: " _fmt, ##__VA_ARGS__)
drivers/gpu/drm/i915/gt/uc/intel_guc_print.h
46
gt_WARN(guc_to_gt(_guc), _cond, "%s(%s)", "guc_WARN_ON", __stringify(_cond))
drivers/gpu/drm/i915/gt/uc/intel_guc_print.h
49
gt_WARN_ONCE(guc_to_gt(_guc), _cond, "%s(%s)", "guc_WARN_ON_ONCE", __stringify(_cond))
drivers/gpu/drm/i915/gt/uc/intel_guc_rc.c
52
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_slpc.c
56
return guc_to_gt(slpc_to_guc(slpc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1123
intel_gt_pm_put_async_untracked(guc_to_gt(guc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1298
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1430
if (mutex_is_locked(&guc_to_gt(guc)->reset.mutex) ||
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1431
test_bit(I915_RESET_BACKOFF, &guc_to_gt(guc)->reset.flags))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1439
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1471
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1484
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1525
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1587
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1605
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1677
intel_gt_is_wedged(guc_to_gt(guc)));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1732
intel_gt_park_heartbeats(guc_to_gt(guc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1738
spin_lock_irq(guc_to_gt(guc)->irq_lock);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
1739
spin_unlock_irq(guc_to_gt(guc)->irq_lock);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2067
intel_gt_is_wedged(guc_to_gt(guc)))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2085
intel_gt_unpark_heartbeats(guc_to_gt(guc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2099
return HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915) &&
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2108
if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2134
if (!HAS_GUC_TLB_INVALIDATION(guc_to_gt(guc)->i915))
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2151
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2431
intel_gt_retire_requests(guc_to_gt(guc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3412
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3562
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4481
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4740
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4767
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4966
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5092
intel_gt_pm_put_async_untracked(guc_to_gt(guc));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5183
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5303
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5316
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
685
if (!intel_uc_uses_guc_submission(&guc_to_gt(guc)->uc))
drivers/gpu/drm/i915/gt/uc/intel_uc.c
181
intel_uncore_write(guc_to_gt(guc)->uncore, SOFT_SCRATCH(15), 0);
drivers/gpu/drm/i915/gt/uc/intel_uc.c
190
val = intel_uncore_read(guc_to_gt(guc)->uncore, SOFT_SCRATCH(15));
drivers/gpu/drm/i915/gt/uc/intel_uc.c
218
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/i915/gt/uc/intel_uc.c
697
struct intel_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
101
struct xe_gt *rx_gt = guc_to_gt(guc), *test_gt, *tx_gt = NULL;
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
300
struct xe_gt *remote_gt, *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
539
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
56
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/tests/xe_guc_g2g_test.c
566
struct xe_gt *remote_gt, *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
417
xe_gt_WARN_ON(guc_to_gt(guc), value_len > 3);
drivers/gpu/drm/xe/xe_guc.c
1123
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1210
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1275
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1287
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1309
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1331
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1377
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1392
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1417
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1597
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1635
xe_gt_WARN(guc_to_gt(guc), err, "Failed to stop GuC PC: %pe\n",
drivers/gpu/drm/xe/xe_guc.c
1688
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
1736
xe_gt_assert(guc_to_gt(guc), guc_to_xe(guc)->wedged.mode);
drivers/gpu/drm/xe/xe_guc.c
1757
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
187
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
235
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
281
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
295
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
315
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
405
struct xe_gt *near_gt = guc_to_gt(near_guc);
drivers/gpu/drm/xe/xe_guc.c
443
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
465
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
519
struct xe_gt *far_gt, *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
595
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
62
addr = __xe_bo_ggtt_addr(bo, gt_to_tile(guc_to_gt(guc))->id);
drivers/gpu/drm/xe/xe_guc.c
650
xe_gt_err(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc.c
663
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
666
xe_uc_sanitize_reset(&guc_to_gt(guc)->uc);
drivers/gpu/drm/xe/xe_guc.c
679
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
689
struct xe_tile *tile = gt_to_tile(guc_to_gt(guc));
drivers/gpu/drm/xe/xe_guc.c
713
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
730
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
759
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
816
err = xe_guc_submit_init(guc, xe_gt_sriov_vf_guc_ids(guc_to_gt(guc)));
drivers/gpu/drm/xe/xe_guc.c
905
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
940
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.c
969
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc.h
98
return gt_to_xe(guc_to_gt(guc));
drivers/gpu/drm/xe/xe_guc_buf.c
24
return guc_to_gt(cache_to_guc(cache));
drivers/gpu/drm/xe/xe_guc_capture.c
1055
xe_gt_warn(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_capture.c
1171
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_capture.c
1413
xe_gt_err(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_capture.c
1430
xe_gt_dbg(guc_to_gt(guc), "Capture extraction failed:%d\n", ret);
drivers/gpu/drm/xe/xe_guc_capture.c
1502
xe_gt_warn(guc_to_gt(guc), "Register capture pre-alloc-cache failure\n");
drivers/gpu/drm/xe/xe_guc_capture.c
423
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_capture.c
497
xe_gt_dbg(guc_to_gt(guc), "capture found %d ext-regs.\n", total);
drivers/gpu/drm/xe/xe_guc_capture.c
539
xe_gt_dbg(guc_to_gt(guc), "Got short capture reglist init: %d out-of %d.\n",
drivers/gpu/drm/xe/xe_guc_capture.c
583
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_capture.c
791
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_capture.c
856
xe_gt_dbg(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_capture.c
860
xe_gt_dbg(guc_to_gt(guc), "Error state capture buffer maybe small: %d < %d\n",
drivers/gpu/drm/xe/xe_guc_capture.c
863
xe_gt_dbg(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_db_mgr.c
38
return guc_to_gt(dbm_to_guc(dbm));
drivers/gpu/drm/xe/xe_guc_engine_activity.c
129
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
178
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
308
xe_gt_assert(guc_to_gt(guc), index < engine_activity->num_activity_group);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
476
xe_gt_err(guc_to_gt(guc), "failed to enable activity stats%d\n", ret);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
498
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_engine_activity.c
93
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_hwconfig.c
55
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_id_mgr.c
24
return guc_to_gt(idm_to_guc(idm));
drivers/gpu/drm/xe/xe_guc_pagefault.c
67
pf.gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_pc.c
109
return guc_to_gt(pc_to_guc(pc));
drivers/gpu/drm/xe/xe_guc_relay.c
50
return guc_to_gt(relay_to_guc(relay));
drivers/gpu/drm/xe/xe_guc_submit.c
1042
xe_gt_assert(guc_to_gt(guc), FIELD_FIT(WQ_LEN_MASK, len_dw));
drivers/gpu/drm/xe/xe_guc_submit.c
1082
xe_gt_assert(guc_to_gt(guc), i == wqi_size / sizeof(u32));
drivers/gpu/drm/xe/xe_guc_submit.c
1088
xe_gt_assert(guc_to_gt(guc), q->guc->wqi_tail <= WQ_SIZE);
drivers/gpu/drm/xe/xe_guc_submit.c
1107
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1166
xe_gt_assert(guc_to_gt(guc), !(exec_queue_destroyed(q) || exec_queue_pending_disable(q)) ||
drivers/gpu/drm/xe/xe_guc_submit.c
1275
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
1280
xe_gt_assert(guc_to_gt(guc), guc_to_xe(guc)->wedged.mode);
drivers/gpu/drm/xe/xe_guc_submit.c
1332
struct xe_gt *gt = guc_to_gt(exec_queue_to_guc(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1391
xe_gt_assert(guc_to_gt(guc), !exec_queue_destroyed(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1392
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1393
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_disable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1394
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_enable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1411
xe_gt_warn(guc_to_gt(guc), "Schedule enable failed to respond");
drivers/gpu/drm/xe/xe_guc_submit.c
1423
xe_gt_assert(guc_to_gt(guc), !exec_queue_destroyed(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1424
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1425
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_disable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1454
xe_gt_assert(guc_to_gt(guc), !exec_queue_destroyed(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1480
xe_gt_assert(guc_to_gt(guc), skip_timeout_check);
drivers/gpu/drm/xe/xe_guc_submit.c
1559
xe_gt_warn(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
1577
xe_gt_notice(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
1706
xe_gt_assert(guc_to_gt(guc), !(q->flags & EXEC_QUEUE_FLAG_PERMANENT));
drivers/gpu/drm/xe/xe_guc_submit.c
1766
xe_gt_assert(guc_to_gt(guc), exec_queue_suspended(q) || exec_queue_killed(q) ||
drivers/gpu/drm/xe/xe_guc_submit.c
1768
xe_gt_assert(guc_to_gt(guc), q->guc->suspend_pending);
drivers/gpu/drm/xe/xe_guc_submit.c
1815
xe_gt_assert(guc_to_gt(guc), !xe_exec_queue_is_parallel(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1816
xe_gt_assert(guc_to_gt(guc), !exec_queue_destroyed(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1817
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1818
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_disable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
1866
xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_CGP_SYNC_SIZE);
drivers/gpu/drm/xe/xe_guc_submit.c
1932
xe_gt_assert(guc_to_gt(guc), xe_device_uc_enabled(guc_to_xe(guc)));
drivers/gpu/drm/xe/xe_guc_submit.c
1962
timeout, guc_to_gt(guc)->ordered_wq, NULL,
drivers/gpu/drm/xe/xe_guc_submit.c
2140
xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)), xe_exec_queue_is_multi_queue(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2200
xe_gt_warn(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
2221
xe_gt_assert(guc_to_gt(guc), !q->guc->suspend_pending);
drivers/gpu/drm/xe/xe_guc_submit.c
2330
if (xe_gt_WARN_ON(guc_to_gt(guc), vf_recovery(guc)))
drivers/gpu/drm/xe/xe_guc_submit.c
2350
xe_gt_assert(guc_to_gt(guc), xe_guc_read_stopped(guc) == 1);
drivers/gpu/drm/xe/xe_guc_submit.c
2381
xe_gt_dbg(guc_to_gt(guc), "Replay RESUME - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2387
xe_gt_dbg(guc_to_gt(guc), "Replay REGISTER - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2395
xe_gt_dbg(guc_to_gt(guc), "Replay ENABLE - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2402
xe_gt_dbg(guc_to_gt(guc), "Replay CLEANUP - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2411
xe_gt_dbg(guc_to_gt(guc), "Replay SUSPEND - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2419
xe_gt_dbg(guc_to_gt(guc), "Replay DISABLE - guc_id=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
248
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
2508
xe_gt_assert(guc_to_gt(guc), IS_SRIOV_VF(guc_to_xe(guc)));
drivers/gpu/drm/xe/xe_guc_submit.c
2509
xe_gt_assert(guc_to_gt(guc), vf_recovery(guc));
drivers/gpu/drm/xe/xe_guc_submit.c
2558
xe_gt_assert(guc_to_gt(guc), xe_guc_read_stopped(guc) == 1);
drivers/gpu/drm/xe/xe_guc_submit.c
2588
xe_gt_dbg(guc_to_gt(guc), "Replay JOB - guc_id=%d, seqno=%d",
drivers/gpu/drm/xe/xe_guc_submit.c
2609
xe_gt_assert(guc_to_gt(guc), IS_SRIOV_VF(guc_to_xe(guc)));
drivers/gpu/drm/xe/xe_guc_submit.c
2610
xe_gt_assert(guc_to_gt(guc), vf_recovery(guc));
drivers/gpu/drm/xe/xe_guc_submit.c
2700
xe_gt_assert(guc_to_gt(guc), IS_SRIOV_VF(guc_to_xe(guc)));
drivers/gpu/drm/xe/xe_guc_submit.c
2743
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
2757
xe_gt_assert(guc_to_gt(guc), guc_id >= q->guc->id);
drivers/gpu/drm/xe/xe_guc_submit.c
2758
xe_gt_assert(guc_to_gt(guc), guc_id < (q->guc->id + q->width));
drivers/gpu/drm/xe/xe_guc_submit.c
2770
xe_gt_assert(guc_to_gt(guc), exec_queue_destroyed(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2771
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2772
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_disable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2773
xe_gt_assert(guc_to_gt(guc), !exec_queue_pending_enable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2790
xe_gt_assert(guc_to_gt(guc), exec_queue_pending_enable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2798
xe_gt_assert(guc_to_gt(guc), runnable_state == 0);
drivers/gpu/drm/xe/xe_guc_submit.c
2799
xe_gt_assert(guc_to_gt(guc), exec_queue_pending_disable(q));
drivers/gpu/drm/xe/xe_guc_submit.c
2855
xe_gt_err(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
2891
xe_gt_err(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
2904
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
2955
xe_gt_warn(guc_to_gt(guc), "G2H-Error capture no space");
drivers/gpu/drm/xe/xe_guc_submit.c
2965
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
3015
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
3038
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
323
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
3319
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_submit.c
379
if (CCS_INSTANCES(guc_to_gt(guc)))
drivers/gpu/drm/xe/xe_guc_submit.c
388
xe_gt_err(guc_to_gt(guc),
drivers/gpu/drm/xe/xe_guc_submit.c
532
xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q) &&
drivers/gpu/drm/xe/xe_guc_submit.c
564
return xe_gt_recovery_pending(guc_to_gt(guc));
drivers/gpu/drm/xe/xe_guc_submit.c
584
xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)),
drivers/gpu/drm/xe/xe_guc_submit.c
632
xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)),
drivers/gpu/drm/xe/xe_guc_submit.c
776
xe_gt_warn(guc_to_gt(guc), "Wait for CGP_SYNC_DONE response failed!\n");
drivers/gpu/drm/xe/xe_guc_submit.c
807
xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_REG_SIZE);
drivers/gpu/drm/xe/xe_guc_submit.c
825
xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_multi_queue_secondary(q));
drivers/gpu/drm/xe/xe_guc_submit.c
830
xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_CGP_SYNC_SIZE);
drivers/gpu/drm/xe/xe_guc_submit.c
850
xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_parallel(q));
drivers/gpu/drm/xe/xe_guc_submit.c
874
xe_gt_assert(guc_to_gt(guc), info->wq_desc_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
876
xe_gt_assert(guc_to_gt(guc), info->wq_base_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
878
xe_gt_assert(guc_to_gt(guc), q->width ==
drivers/gpu/drm/xe/xe_guc_submit.c
880
xe_gt_assert(guc_to_gt(guc), info->hwlrca_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
882
xe_gt_assert(guc_to_gt(guc), len <= MAX_MLRC_REG_SIZE);
drivers/gpu/drm/xe/xe_guc_submit.c
907
xe_gt_assert(guc_to_gt(guc), info->wq_desc_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
909
xe_gt_assert(guc_to_gt(guc), info->wq_base_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
911
xe_gt_assert(guc_to_gt(guc), info->hwlrca_lo ==
drivers/gpu/drm/xe/xe_guc_submit.c
924
xe_gt_assert(guc_to_gt(guc), !exec_queue_registered(q));
drivers/gpu/drm/xe/xe_guc_submit.c
925
xe_gt_assert(guc_to_gt(guc), ctx_type < GUC_CONTEXT_COUNT);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
100
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
127
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
258
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
29
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
60
struct xe_gt *gt = guc_to_gt(guc);
drivers/gpu/drm/xe/xe_memirq.c
336
bool is_media = xe_gt_is_media_type(guc_to_gt(guc));
drivers/gpu/drm/xe/xe_memirq.c
47
return xe_gt_is_media_type(guc_to_gt(guc)) ? "media GuC" : "GuC";
drivers/gpu/drm/xe/xe_memirq.c
492
struct xe_gt *gt = guc_to_gt(guc);