Symbol: pc_to_xe
drivers/gpu/drm/xe/xe_gt_idle.c
267
guard(xe_pm_runtime)(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_gt_idle.c
280
scoped_guard(xe_pm_runtime, pc_to_xe(pc))
drivers/gpu/drm/xe/xe_gt_idle.c
309
scoped_guard(xe_pm_runtime, pc_to_xe(pc))
drivers/gpu/drm/xe/xe_guc_pc.c
1041
if (IS_SRIOV_VF(pc_to_xe(pc)) || pc_to_xe(pc)->info.skip_guc_pc)
drivers/gpu/drm/xe/xe_guc_pc.c
1063
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
1086
guard(xe_pm_runtime)(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_guc_pc.c
1098
guard(xe_pm_runtime)(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_guc_pc.c
1187
guard(xe_pm_runtime_noresume)(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_guc_pc.c
1218
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
123
xe_map_rd_field(pc_to_xe(pc_), pc_to_maps(pc_), 0, \
drivers/gpu/drm/xe/xe_guc_pc.c
1237
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
127
xe_map_wr_field(pc_to_xe(pc_), pc_to_maps(pc_), 0, \
drivers/gpu/drm/xe/xe_guc_pc.c
1320
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
1341
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
140
xe_device_assert_mem_access(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_guc_pc.c
180
if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED))
drivers/gpu/drm/xe/xe_guc_pc.c
204
if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED))
drivers/gpu/drm/xe/xe_guc_pc.c
227
if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED))
drivers/gpu/drm/xe/xe_guc_pc.c
249
if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED))
drivers/gpu/drm/xe/xe_guc_pc.c
266
if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED))
drivers/gpu/drm/xe/xe_guc_pc.c
554
struct xe_device *xe = pc_to_xe(pc);
drivers/gpu/drm/xe/xe_guc_pc.c
781
xe_device_assert_mem_access(pc_to_xe(pc));
drivers/gpu/drm/xe/xe_guc_pc.c
799
xe_device_assert_mem_access(pc_to_xe(pc));