#include <drm/drm_managed.h>
#include <drm/gpu_scheduler.h>
#include "pvr_cccb.h"
#include "pvr_context.h"
#include "pvr_device.h"
#include "pvr_drv.h"
#include "pvr_job.h"
#include "pvr_queue.h"
#include "pvr_vm.h"
#include "pvr_rogue_fwif_client.h"
#define MAX_DEADLINE_MS 30000
#define CTX_COMPUTE_CCCB_SIZE_LOG2 15
#define CTX_FRAG_CCCB_SIZE_LOG2 15
#define CTX_GEOM_CCCB_SIZE_LOG2 15
#define CTX_TRANSFER_CCCB_SIZE_LOG2 15
static int get_xfer_ctx_state_size(struct pvr_device *pvr_dev)
{
u32 num_isp_store_registers;
if (PVR_HAS_FEATURE(pvr_dev, xe_memory_hierarchy)) {
num_isp_store_registers = 1;
} else {
int err;
err = PVR_FEATURE_VALUE(pvr_dev, num_isp_ipp_pipes, &num_isp_store_registers);
if (WARN_ON(err))
return err;
}
return sizeof(struct rogue_fwif_frag_ctx_state) +
(num_isp_store_registers *
sizeof(((struct rogue_fwif_frag_ctx_state *)0)->frag_reg_isp_store[0]));
}
static int get_frag_ctx_state_size(struct pvr_device *pvr_dev)
{
u32 num_isp_store_registers;
int err;
if (PVR_HAS_FEATURE(pvr_dev, xe_memory_hierarchy)) {
err = PVR_FEATURE_VALUE(pvr_dev, num_raster_pipes, &num_isp_store_registers);
if (WARN_ON(err))
return err;
if (PVR_HAS_FEATURE(pvr_dev, gpu_multicore_support)) {
u32 xpu_max_slaves;
err = PVR_FEATURE_VALUE(pvr_dev, xpu_max_slaves, &xpu_max_slaves);
if (WARN_ON(err))
return err;
num_isp_store_registers *= (1 + xpu_max_slaves);
}
} else {
err = PVR_FEATURE_VALUE(pvr_dev, num_isp_ipp_pipes, &num_isp_store_registers);
if (WARN_ON(err))
return err;
}
return sizeof(struct rogue_fwif_frag_ctx_state) +
(num_isp_store_registers *
sizeof(((struct rogue_fwif_frag_ctx_state *)0)->frag_reg_isp_store[0]));
}
static int get_ctx_state_size(struct pvr_device *pvr_dev, enum drm_pvr_job_type type)
{
switch (type) {
case DRM_PVR_JOB_TYPE_GEOMETRY:
return sizeof(struct rogue_fwif_geom_ctx_state);
case DRM_PVR_JOB_TYPE_FRAGMENT:
return get_frag_ctx_state_size(pvr_dev);
case DRM_PVR_JOB_TYPE_COMPUTE:
return sizeof(struct rogue_fwif_compute_ctx_state);
case DRM_PVR_JOB_TYPE_TRANSFER_FRAG:
return get_xfer_ctx_state_size(pvr_dev);
}
WARN(1, "Invalid queue type");
return -EINVAL;
}
static u32 get_ctx_offset(enum drm_pvr_job_type type)
{
switch (type) {
case DRM_PVR_JOB_TYPE_GEOMETRY:
return offsetof(struct rogue_fwif_fwrendercontext, geom_context);
case DRM_PVR_JOB_TYPE_FRAGMENT:
return offsetof(struct rogue_fwif_fwrendercontext, frag_context);
case DRM_PVR_JOB_TYPE_COMPUTE:
return offsetof(struct rogue_fwif_fwcomputecontext, cdm_context);
case DRM_PVR_JOB_TYPE_TRANSFER_FRAG:
return offsetof(struct rogue_fwif_fwtransfercontext, tq_context);
}
return 0;
}
static const char *
pvr_queue_fence_get_driver_name(struct dma_fence *f)
{
return PVR_DRIVER_NAME;
}
static void pvr_queue_fence_release_work(struct work_struct *w)
{
struct pvr_queue_fence *fence = container_of(w, struct pvr_queue_fence, release_work);
pvr_context_put(fence->queue->ctx);
dma_fence_free(&fence->base);
}
static void pvr_queue_fence_release(struct dma_fence *f)
{
struct pvr_queue_fence *fence = container_of(f, struct pvr_queue_fence, base);
struct pvr_device *pvr_dev = fence->queue->ctx->pvr_dev;
queue_work(pvr_dev->sched_wq, &fence->release_work);
}
static const char *
pvr_queue_job_fence_get_timeline_name(struct dma_fence *f)
{
struct pvr_queue_fence *fence = container_of(f, struct pvr_queue_fence, base);
switch (fence->queue->type) {
case DRM_PVR_JOB_TYPE_GEOMETRY:
return "geometry";
case DRM_PVR_JOB_TYPE_FRAGMENT:
return "fragment";
case DRM_PVR_JOB_TYPE_COMPUTE:
return "compute";
case DRM_PVR_JOB_TYPE_TRANSFER_FRAG:
return "transfer";
}
WARN(1, "Invalid queue type");
return "invalid";
}
static const char *
pvr_queue_cccb_fence_get_timeline_name(struct dma_fence *f)
{
struct pvr_queue_fence *fence = container_of(f, struct pvr_queue_fence, base);
switch (fence->queue->type) {
case DRM_PVR_JOB_TYPE_GEOMETRY:
return "geometry-cccb";
case DRM_PVR_JOB_TYPE_FRAGMENT:
return "fragment-cccb";
case DRM_PVR_JOB_TYPE_COMPUTE:
return "compute-cccb";
case DRM_PVR_JOB_TYPE_TRANSFER_FRAG:
return "transfer-cccb";
}
WARN(1, "Invalid queue type");
return "invalid";
}
static const struct dma_fence_ops pvr_queue_job_fence_ops = {
.get_driver_name = pvr_queue_fence_get_driver_name,
.get_timeline_name = pvr_queue_job_fence_get_timeline_name,
.release = pvr_queue_fence_release,
};
static struct pvr_queue_fence *
to_pvr_queue_job_fence(struct dma_fence *f)
{
struct drm_sched_fence *sched_fence = to_drm_sched_fence(f);
if (sched_fence)
f = sched_fence->parent;
if (f && f->ops == &pvr_queue_job_fence_ops)
return container_of(f, struct pvr_queue_fence, base);
return NULL;
}
static const struct dma_fence_ops pvr_queue_cccb_fence_ops = {
.get_driver_name = pvr_queue_fence_get_driver_name,
.get_timeline_name = pvr_queue_cccb_fence_get_timeline_name,
.release = pvr_queue_fence_release,
};
static void pvr_queue_fence_put(struct dma_fence *f)
{
if (!f)
return;
if (WARN_ON(f->ops &&
f->ops != &pvr_queue_cccb_fence_ops &&
f->ops != &pvr_queue_job_fence_ops))
return;
if (f->ops)
dma_fence_put(f);
else
dma_fence_free(f);
}
static struct dma_fence *
pvr_queue_fence_alloc(void)
{
struct pvr_queue_fence *fence;
fence = kzalloc_obj(*fence);
if (!fence)
return NULL;
return &fence->base;
}
static void
pvr_queue_fence_init(struct dma_fence *f,
struct pvr_queue *queue,
const struct dma_fence_ops *fence_ops,
struct pvr_queue_fence_ctx *fence_ctx)
{
struct pvr_queue_fence *fence = container_of(f, struct pvr_queue_fence, base);
pvr_context_get(queue->ctx);
fence->queue = queue;
INIT_WORK(&fence->release_work, pvr_queue_fence_release_work);
dma_fence_init(&fence->base, fence_ops,
&fence_ctx->lock, fence_ctx->id,
atomic_inc_return(&fence_ctx->seqno));
}
static void
pvr_queue_cccb_fence_init(struct dma_fence *fence, struct pvr_queue *queue)
{
pvr_queue_fence_init(fence, queue, &pvr_queue_cccb_fence_ops,
&queue->cccb_fence_ctx.base);
}
static void
pvr_queue_job_fence_init(struct dma_fence *fence, struct pvr_queue *queue)
{
if (!fence->ops)
pvr_queue_fence_init(fence, queue, &pvr_queue_job_fence_ops,
&queue->job_fence_ctx);
}
static void
pvr_queue_fence_ctx_init(struct pvr_queue_fence_ctx *fence_ctx)
{
spin_lock_init(&fence_ctx->lock);
fence_ctx->id = dma_fence_context_alloc(1);
atomic_set(&fence_ctx->seqno, 0);
}
static u32 ufo_cmds_size(u32 elem_count)
{
u32 full_cmd_count = elem_count / ROGUE_FWIF_CCB_CMD_MAX_UFOS;
u32 remaining_elems = elem_count % ROGUE_FWIF_CCB_CMD_MAX_UFOS;
u32 size = full_cmd_count *
pvr_cccb_get_size_of_cmd_with_hdr(ROGUE_FWIF_CCB_CMD_MAX_UFOS *
sizeof(struct rogue_fwif_ufo));
if (remaining_elems) {
size += pvr_cccb_get_size_of_cmd_with_hdr(remaining_elems *
sizeof(struct rogue_fwif_ufo));
}
return size;
}
static u32 job_cmds_size(struct pvr_job *job, u32 ufo_wait_count)
{
return ufo_cmds_size(1) + ufo_cmds_size(ufo_wait_count) +
pvr_cccb_get_size_of_cmd_with_hdr(job->cmd_len);
}
static unsigned long job_count_remaining_native_deps(struct pvr_job *job)
{
unsigned long remaining_count = 0;
struct dma_fence *fence = NULL;
unsigned long index;
xa_for_each(&job->base.dependencies, index, fence) {
struct pvr_queue_fence *jfence;
jfence = to_pvr_queue_job_fence(fence);
if (!jfence)
continue;
if (!dma_fence_is_signaled(&jfence->base))
remaining_count++;
}
return remaining_count;
}
static struct dma_fence *
pvr_queue_get_job_cccb_fence(struct pvr_queue *queue, struct pvr_job *job)
{
struct pvr_queue_fence *cccb_fence;
unsigned int native_deps_remaining;
if (!job->cccb_fence)
return NULL;
mutex_lock(&queue->cccb_fence_ctx.job_lock);
native_deps_remaining = job_count_remaining_native_deps(job);
if (pvr_cccb_cmdseq_fits(&queue->cccb, job_cmds_size(job, native_deps_remaining))) {
pvr_queue_fence_put(job->cccb_fence);
job->cccb_fence = NULL;
goto out_unlock;
}
if (WARN_ON(queue->cccb_fence_ctx.job))
pvr_job_put(queue->cccb_fence_ctx.job);
queue->cccb_fence_ctx.job = pvr_job_get(job);
cccb_fence = container_of(job->cccb_fence, struct pvr_queue_fence, base);
if (!WARN_ON(cccb_fence->queue))
pvr_queue_cccb_fence_init(job->cccb_fence, queue);
out_unlock:
mutex_unlock(&queue->cccb_fence_ctx.job_lock);
return dma_fence_get(job->cccb_fence);
}
static struct dma_fence *
pvr_queue_get_job_kccb_fence(struct pvr_queue *queue, struct pvr_job *job)
{
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
struct dma_fence *kccb_fence = NULL;
if (!job->kccb_fence)
return NULL;
if (!WARN_ON(job->kccb_fence->ops)) {
kccb_fence = pvr_kccb_reserve_slot(pvr_dev, job->kccb_fence);
job->kccb_fence = NULL;
}
return kccb_fence;
}
static struct dma_fence *
pvr_queue_get_paired_frag_job_dep(struct pvr_queue *queue, struct pvr_job *job)
{
struct pvr_job *frag_job = job->type == DRM_PVR_JOB_TYPE_GEOMETRY ?
job->paired_job : NULL;
struct dma_fence *f;
unsigned long index;
if (!frag_job)
return NULL;
xa_for_each(&frag_job->base.dependencies, index, f) {
if (dma_fence_is_signaled(f))
continue;
if (f == &job->base.s_fence->scheduled)
continue;
return dma_fence_get(f);
}
return frag_job->base.sched->ops->prepare_job(&frag_job->base, &queue->entity);
}
static struct dma_fence *
pvr_queue_prepare_job(struct drm_sched_job *sched_job,
struct drm_sched_entity *s_entity)
{
struct pvr_job *job = container_of(sched_job, struct pvr_job, base);
struct pvr_queue *queue = container_of(s_entity, struct pvr_queue, entity);
struct dma_fence *internal_dep = NULL;
if (job->type == DRM_PVR_JOB_TYPE_FRAGMENT && job->paired_job) {
if (job->paired_job->has_pm_ref)
return NULL;
pvr_queue_job_fence_init(job->done_fence,
job->ctx->queues.fragment);
} else {
pvr_queue_job_fence_init(job->done_fence, queue);
}
internal_dep = pvr_queue_get_job_cccb_fence(queue, job);
if (!internal_dep)
internal_dep = pvr_queue_get_job_kccb_fence(queue, job);
if (!internal_dep)
internal_dep = pvr_queue_get_paired_frag_job_dep(queue, job);
return internal_dep;
}
static void pvr_queue_update_active_state_locked(struct pvr_queue *queue)
{
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
lockdep_assert_held(&pvr_dev->queues.lock);
if (list_empty(&queue->node))
return;
if (!atomic_read(&queue->in_flight_job_count))
list_move_tail(&queue->node, &pvr_dev->queues.idle);
else
list_move_tail(&queue->node, &pvr_dev->queues.active);
}
static void pvr_queue_update_active_state(struct pvr_queue *queue)
{
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
mutex_lock(&pvr_dev->queues.lock);
pvr_queue_update_active_state_locked(queue);
mutex_unlock(&pvr_dev->queues.lock);
}
static void pvr_queue_submit_job_to_cccb(struct pvr_job *job)
{
struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler);
struct rogue_fwif_ufo ufos[ROGUE_FWIF_CCB_CMD_MAX_UFOS];
struct pvr_cccb *cccb = &queue->cccb;
struct pvr_queue_fence *jfence;
struct dma_fence *fence;
unsigned long index;
u32 ufo_count = 0;
atomic_inc(&queue->in_flight_job_count);
pvr_queue_update_active_state(queue);
xa_for_each(&job->base.dependencies, index, fence) {
jfence = to_pvr_queue_job_fence(fence);
if (!jfence)
continue;
if (job->type == DRM_PVR_JOB_TYPE_FRAGMENT && job->paired_job &&
&job->paired_job->base.s_fence->scheduled == fence)
continue;
if (dma_fence_is_signaled(&jfence->base))
continue;
pvr_fw_object_get_fw_addr(jfence->queue->timeline_ufo.fw_obj,
&ufos[ufo_count].addr);
ufos[ufo_count++].value = jfence->base.seqno;
if (ufo_count == ARRAY_SIZE(ufos)) {
pvr_cccb_write_command_with_header(cccb, ROGUE_FWIF_CCB_CMD_TYPE_FENCE_PR,
sizeof(ufos), ufos, 0, 0);
ufo_count = 0;
}
}
if (job->type == DRM_PVR_JOB_TYPE_FRAGMENT && job->paired_job) {
jfence = to_pvr_queue_job_fence(job->paired_job->done_fence);
if (!WARN_ON(!jfence)) {
pvr_fw_object_get_fw_addr(jfence->queue->timeline_ufo.fw_obj,
&ufos[ufo_count].addr);
ufos[ufo_count++].value = job->paired_job->done_fence->seqno;
}
}
if (ufo_count) {
pvr_cccb_write_command_with_header(cccb, ROGUE_FWIF_CCB_CMD_TYPE_FENCE_PR,
sizeof(ufos[0]) * ufo_count, ufos, 0, 0);
}
if (job->type == DRM_PVR_JOB_TYPE_GEOMETRY && job->paired_job) {
struct rogue_fwif_cmd_geom *cmd = job->cmd;
pvr_fw_object_get_fw_addr(queue->timeline_ufo.fw_obj,
&cmd->partial_render_geom_frag_fence.addr);
cmd->partial_render_geom_frag_fence.value = job->done_fence->seqno - 1;
}
pvr_cccb_write_command_with_header(cccb, job->fw_ccb_cmd_type, job->cmd_len, job->cmd,
job->id, job->id);
pvr_fw_object_get_fw_addr(queue->timeline_ufo.fw_obj, &ufos[0].addr);
ufos[0].value = job->done_fence->seqno;
pvr_cccb_write_command_with_header(cccb, ROGUE_FWIF_CCB_CMD_TYPE_UPDATE,
sizeof(ufos[0]), ufos, 0, 0);
}
static struct dma_fence *pvr_queue_run_job(struct drm_sched_job *sched_job)
{
struct pvr_job *job = container_of(sched_job, struct pvr_job, base);
struct pvr_device *pvr_dev = job->pvr_dev;
int err;
if (job->paired_job && job->type == DRM_PVR_JOB_TYPE_FRAGMENT &&
job->done_fence->ops) {
return dma_fence_get(job->done_fence);
}
if (WARN_ON(job->paired_job &&
(job->type != DRM_PVR_JOB_TYPE_GEOMETRY ||
job->paired_job->type != DRM_PVR_JOB_TYPE_FRAGMENT ||
job->hwrt != job->paired_job->hwrt ||
job->ctx != job->paired_job->ctx)))
return ERR_PTR(-EINVAL);
err = pvr_job_get_pm_ref(job);
if (WARN_ON(err))
return ERR_PTR(err);
if (job->paired_job) {
err = pvr_job_get_pm_ref(job->paired_job);
if (WARN_ON(err))
return ERR_PTR(err);
}
pvr_queue_submit_job_to_cccb(job);
if (job->paired_job) {
struct pvr_job *geom_job = job;
struct pvr_job *frag_job = job->paired_job;
struct pvr_queue *geom_queue = job->ctx->queues.geometry;
struct pvr_queue *frag_queue = job->ctx->queues.fragment;
pvr_queue_submit_job_to_cccb(frag_job);
pvr_cccb_send_kccb_combined_kick(pvr_dev,
&geom_queue->cccb, &frag_queue->cccb,
pvr_context_get_fw_addr(geom_job->ctx) +
geom_queue->ctx_offset,
pvr_context_get_fw_addr(frag_job->ctx) +
frag_queue->ctx_offset,
job->hwrt,
frag_job->fw_ccb_cmd_type ==
ROGUE_FWIF_CCB_CMD_TYPE_FRAG_PR);
} else {
struct pvr_queue *queue = container_of(job->base.sched,
struct pvr_queue, scheduler);
pvr_cccb_send_kccb_kick(pvr_dev, &queue->cccb,
pvr_context_get_fw_addr(job->ctx) + queue->ctx_offset,
job->hwrt);
}
return dma_fence_get(job->done_fence);
}
static void pvr_queue_stop(struct pvr_queue *queue, struct pvr_job *bad_job)
{
drm_sched_stop(&queue->scheduler, bad_job ? &bad_job->base : NULL);
}
static void pvr_queue_start(struct pvr_queue *queue)
{
struct pvr_job *job;
*queue->timeline_ufo.value = atomic_read(&queue->job_fence_ctx.seqno);
list_for_each_entry(job, &queue->scheduler.pending_list, base.list) {
if (dma_fence_is_signaled(job->done_fence)) {
WARN_ON(job->base.s_fence->parent);
job->base.s_fence->parent = dma_fence_get(job->done_fence);
} else {
atomic_set(&queue->ctx->faulty, 1);
}
}
drm_sched_start(&queue->scheduler, 0);
}
static enum drm_gpu_sched_stat
pvr_queue_timedout_job(struct drm_sched_job *s_job)
{
struct drm_gpu_scheduler *sched = s_job->sched;
struct pvr_queue *queue = container_of(sched, struct pvr_queue, scheduler);
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
struct pvr_job *job;
u32 job_count = 0;
dev_err(sched->dev, "Job timeout\n");
mutex_lock(&pvr_dev->queues.lock);
list_del_init(&queue->node);
mutex_unlock(&pvr_dev->queues.lock);
drm_sched_stop(sched, s_job);
list_for_each_entry(job, &sched->pending_list, base.list) {
job->base.s_fence->parent = dma_fence_get(job->done_fence);
job_count++;
}
WARN_ON(atomic_read(&queue->in_flight_job_count) != job_count);
mutex_lock(&pvr_dev->queues.lock);
if (!job_count) {
list_move_tail(&queue->node, &pvr_dev->queues.idle);
} else {
atomic_set(&queue->in_flight_job_count, job_count);
list_move_tail(&queue->node, &pvr_dev->queues.active);
pvr_queue_process(queue);
}
mutex_unlock(&pvr_dev->queues.lock);
drm_sched_start(sched, 0);
return DRM_GPU_SCHED_STAT_RESET;
}
static void pvr_queue_free_job(struct drm_sched_job *sched_job)
{
struct pvr_job *job = container_of(sched_job, struct pvr_job, base);
drm_sched_job_cleanup(sched_job);
if (job->type == DRM_PVR_JOB_TYPE_FRAGMENT && job->paired_job)
pvr_job_put(job->paired_job);
job->paired_job = NULL;
pvr_job_put(job);
}
static const struct drm_sched_backend_ops pvr_queue_sched_ops = {
.prepare_job = pvr_queue_prepare_job,
.run_job = pvr_queue_run_job,
.timedout_job = pvr_queue_timedout_job,
.free_job = pvr_queue_free_job,
};
bool pvr_queue_fence_is_ufo_backed(struct dma_fence *f)
{
struct drm_sched_fence *sched_fence = f ? to_drm_sched_fence(f) : NULL;
if (sched_fence &&
sched_fence->sched->ops == &pvr_queue_sched_ops)
return true;
if (f && f->ops == &pvr_queue_job_fence_ops)
return true;
return false;
}
static void
pvr_queue_signal_done_fences(struct pvr_queue *queue)
{
struct pvr_job *job, *tmp_job;
u32 cur_seqno;
spin_lock(&queue->scheduler.job_list_lock);
cur_seqno = *queue->timeline_ufo.value;
list_for_each_entry_safe(job, tmp_job, &queue->scheduler.pending_list, base.list) {
if ((int)(cur_seqno - lower_32_bits(job->done_fence->seqno)) < 0)
break;
if (!dma_fence_is_signaled(job->done_fence)) {
dma_fence_signal(job->done_fence);
pvr_job_release_pm_ref(job);
atomic_dec(&queue->in_flight_job_count);
}
}
spin_unlock(&queue->scheduler.job_list_lock);
}
static void
pvr_queue_check_job_waiting_for_cccb_space(struct pvr_queue *queue)
{
struct pvr_queue_fence *cccb_fence;
u32 native_deps_remaining;
struct pvr_job *job;
mutex_lock(&queue->cccb_fence_ctx.job_lock);
job = queue->cccb_fence_ctx.job;
if (!job)
goto out_unlock;
if (WARN_ON(!job->cccb_fence)) {
job = NULL;
goto out_unlock;
}
cccb_fence = container_of(job->cccb_fence, struct pvr_queue_fence, base);
if (WARN_ON(!cccb_fence->queue)) {
job = NULL;
goto out_unlock;
}
native_deps_remaining = job_count_remaining_native_deps(job);
if (!pvr_cccb_cmdseq_fits(&queue->cccb, job_cmds_size(job, native_deps_remaining))) {
job = NULL;
goto out_unlock;
}
dma_fence_signal(job->cccb_fence);
pvr_queue_fence_put(job->cccb_fence);
job->cccb_fence = NULL;
queue->cccb_fence_ctx.job = NULL;
out_unlock:
mutex_unlock(&queue->cccb_fence_ctx.job_lock);
pvr_job_put(job);
}
void pvr_queue_process(struct pvr_queue *queue)
{
lockdep_assert_held(&queue->ctx->pvr_dev->queues.lock);
pvr_queue_check_job_waiting_for_cccb_space(queue);
pvr_queue_signal_done_fences(queue);
pvr_queue_update_active_state_locked(queue);
}
static u32 get_dm_type(struct pvr_queue *queue)
{
switch (queue->type) {
case DRM_PVR_JOB_TYPE_GEOMETRY:
return PVR_FWIF_DM_GEOM;
case DRM_PVR_JOB_TYPE_TRANSFER_FRAG:
case DRM_PVR_JOB_TYPE_FRAGMENT:
return PVR_FWIF_DM_FRAG;
case DRM_PVR_JOB_TYPE_COMPUTE:
return PVR_FWIF_DM_CDM;
}
return ~0;
}
static void init_fw_context(struct pvr_queue *queue, void *fw_ctx_map)
{
struct pvr_context *ctx = queue->ctx;
struct pvr_fw_object *fw_mem_ctx_obj = pvr_vm_get_fw_mem_context(ctx->vm_ctx);
struct rogue_fwif_fwcommoncontext *cctx_fw;
struct pvr_cccb *cccb = &queue->cccb;
cctx_fw = fw_ctx_map + queue->ctx_offset;
cctx_fw->ccbctl_fw_addr = cccb->ctrl_fw_addr;
cctx_fw->ccb_fw_addr = cccb->cccb_fw_addr;
cctx_fw->dm = get_dm_type(queue);
cctx_fw->priority = ctx->priority;
cctx_fw->priority_seq_num = 0;
cctx_fw->max_deadline_ms = MAX_DEADLINE_MS;
cctx_fw->pid = task_tgid_nr(current);
cctx_fw->server_common_context_id = ctx->ctx_id;
pvr_fw_object_get_fw_addr(fw_mem_ctx_obj, &cctx_fw->fw_mem_context_fw_addr);
pvr_fw_object_get_fw_addr(queue->reg_state_obj, &cctx_fw->context_state_addr);
}
static int pvr_queue_cleanup_fw_context(struct pvr_queue *queue)
{
if (!queue->ctx->fw_obj)
return 0;
return pvr_fw_structure_cleanup(queue->ctx->pvr_dev,
ROGUE_FWIF_CLEANUP_FWCOMMONCONTEXT,
queue->ctx->fw_obj, queue->ctx_offset);
}
int pvr_queue_job_init(struct pvr_job *job, u64 drm_client_id)
{
u32 min_native_dep_count = job->type == DRM_PVR_JOB_TYPE_FRAGMENT ? 1 : 0;
struct pvr_queue *queue;
int err;
if (atomic_read(&job->ctx->faulty))
return -EIO;
queue = pvr_context_get_queue_for_job(job->ctx, job->type);
if (!queue)
return -EINVAL;
if (!pvr_cccb_cmdseq_can_fit(&queue->cccb, job_cmds_size(job, min_native_dep_count)))
return -E2BIG;
err = drm_sched_job_init(&job->base, &queue->entity, 1, THIS_MODULE, drm_client_id);
if (err)
return err;
job->cccb_fence = pvr_queue_fence_alloc();
job->kccb_fence = pvr_kccb_fence_alloc();
job->done_fence = pvr_queue_fence_alloc();
if (!job->cccb_fence || !job->kccb_fence || !job->done_fence)
return -ENOMEM;
return 0;
}
struct dma_fence *pvr_queue_job_arm(struct pvr_job *job)
{
drm_sched_job_arm(&job->base);
return &job->base.s_fence->finished;
}
void pvr_queue_job_cleanup(struct pvr_job *job)
{
pvr_queue_fence_put(job->done_fence);
pvr_queue_fence_put(job->cccb_fence);
pvr_kccb_fence_put(job->kccb_fence);
if (job->base.s_fence)
drm_sched_job_cleanup(&job->base);
}
void pvr_queue_job_push(struct pvr_job *job)
{
struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler);
dma_fence_put(queue->last_queued_job_scheduled_fence);
queue->last_queued_job_scheduled_fence = dma_fence_get(&job->base.s_fence->scheduled);
pvr_job_get(job);
drm_sched_entity_push_job(&job->base);
}
static void reg_state_init(void *cpu_ptr, void *priv)
{
struct pvr_queue *queue = priv;
if (queue->type == DRM_PVR_JOB_TYPE_GEOMETRY) {
struct rogue_fwif_geom_ctx_state *geom_ctx_state_fw = cpu_ptr;
geom_ctx_state_fw->geom_core[0].geom_reg_vdm_call_stack_pointer_init =
queue->callstack_addr;
}
}
struct pvr_queue *pvr_queue_create(struct pvr_context *ctx,
enum drm_pvr_job_type type,
struct drm_pvr_ioctl_create_context_args *args,
void *fw_ctx_map)
{
static const struct {
u32 cccb_size;
const char *name;
} props[] = {
[DRM_PVR_JOB_TYPE_GEOMETRY] = {
.cccb_size = CTX_GEOM_CCCB_SIZE_LOG2,
.name = "geometry",
},
[DRM_PVR_JOB_TYPE_FRAGMENT] = {
.cccb_size = CTX_FRAG_CCCB_SIZE_LOG2,
.name = "fragment"
},
[DRM_PVR_JOB_TYPE_COMPUTE] = {
.cccb_size = CTX_COMPUTE_CCCB_SIZE_LOG2,
.name = "compute"
},
[DRM_PVR_JOB_TYPE_TRANSFER_FRAG] = {
.cccb_size = CTX_TRANSFER_CCCB_SIZE_LOG2,
.name = "transfer_frag"
},
};
struct pvr_device *pvr_dev = ctx->pvr_dev;
const struct drm_sched_init_args sched_args = {
.ops = &pvr_queue_sched_ops,
.submit_wq = pvr_dev->sched_wq,
.num_rqs = 1,
.credit_limit = 64 * 1024,
.hang_limit = 1,
.timeout = msecs_to_jiffies(500),
.timeout_wq = pvr_dev->sched_wq,
.name = "pvr-queue",
.dev = pvr_dev->base.dev,
};
struct drm_gpu_scheduler *sched;
struct pvr_queue *queue;
int ctx_state_size, err;
void *cpu_map;
if (WARN_ON(type >= sizeof(props)))
return ERR_PTR(-EINVAL);
switch (ctx->type) {
case DRM_PVR_CTX_TYPE_RENDER:
if (type != DRM_PVR_JOB_TYPE_GEOMETRY &&
type != DRM_PVR_JOB_TYPE_FRAGMENT)
return ERR_PTR(-EINVAL);
break;
case DRM_PVR_CTX_TYPE_COMPUTE:
if (type != DRM_PVR_JOB_TYPE_COMPUTE)
return ERR_PTR(-EINVAL);
break;
case DRM_PVR_CTX_TYPE_TRANSFER_FRAG:
if (type != DRM_PVR_JOB_TYPE_TRANSFER_FRAG)
return ERR_PTR(-EINVAL);
break;
default:
return ERR_PTR(-EINVAL);
}
ctx_state_size = get_ctx_state_size(pvr_dev, type);
if (ctx_state_size < 0)
return ERR_PTR(ctx_state_size);
queue = kzalloc_obj(*queue);
if (!queue)
return ERR_PTR(-ENOMEM);
queue->type = type;
queue->ctx_offset = get_ctx_offset(type);
queue->ctx = ctx;
queue->callstack_addr = args->callstack_addr;
sched = &queue->scheduler;
INIT_LIST_HEAD(&queue->node);
mutex_init(&queue->cccb_fence_ctx.job_lock);
pvr_queue_fence_ctx_init(&queue->cccb_fence_ctx.base);
pvr_queue_fence_ctx_init(&queue->job_fence_ctx);
err = pvr_cccb_init(pvr_dev, &queue->cccb, props[type].cccb_size, props[type].name);
if (err)
goto err_free_queue;
err = pvr_fw_object_create(pvr_dev, ctx_state_size,
PVR_BO_FW_FLAGS_DEVICE_UNCACHED,
reg_state_init, queue, &queue->reg_state_obj);
if (err)
goto err_cccb_fini;
init_fw_context(queue, fw_ctx_map);
if (type != DRM_PVR_JOB_TYPE_GEOMETRY && type != DRM_PVR_JOB_TYPE_FRAGMENT &&
args->callstack_addr) {
err = -EINVAL;
goto err_release_reg_state;
}
cpu_map = pvr_fw_object_create_and_map(pvr_dev, sizeof(*queue->timeline_ufo.value),
PVR_BO_FW_FLAGS_DEVICE_UNCACHED,
NULL, NULL, &queue->timeline_ufo.fw_obj);
if (IS_ERR(cpu_map)) {
err = PTR_ERR(cpu_map);
goto err_release_reg_state;
}
queue->timeline_ufo.value = cpu_map;
err = drm_sched_init(&queue->scheduler, &sched_args);
if (err)
goto err_release_ufo;
err = drm_sched_entity_init(&queue->entity,
DRM_SCHED_PRIORITY_KERNEL,
&sched, 1, &ctx->faulty);
if (err)
goto err_sched_fini;
mutex_lock(&pvr_dev->queues.lock);
list_add_tail(&queue->node, &pvr_dev->queues.idle);
mutex_unlock(&pvr_dev->queues.lock);
return queue;
err_sched_fini:
drm_sched_fini(&queue->scheduler);
err_release_ufo:
pvr_fw_object_unmap_and_destroy(queue->timeline_ufo.fw_obj);
err_release_reg_state:
pvr_fw_object_destroy(queue->reg_state_obj);
err_cccb_fini:
pvr_cccb_fini(&queue->cccb);
err_free_queue:
mutex_destroy(&queue->cccb_fence_ctx.job_lock);
kfree(queue);
return ERR_PTR(err);
}
void pvr_queue_device_pre_reset(struct pvr_device *pvr_dev)
{
struct pvr_queue *queue;
mutex_lock(&pvr_dev->queues.lock);
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
pvr_queue_stop(queue, NULL);
list_for_each_entry(queue, &pvr_dev->queues.active, node)
pvr_queue_stop(queue, NULL);
mutex_unlock(&pvr_dev->queues.lock);
}
void pvr_queue_device_post_reset(struct pvr_device *pvr_dev)
{
struct pvr_queue *queue;
mutex_lock(&pvr_dev->queues.lock);
list_for_each_entry(queue, &pvr_dev->queues.active, node)
pvr_queue_start(queue);
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
pvr_queue_start(queue);
mutex_unlock(&pvr_dev->queues.lock);
}
void pvr_queue_kill(struct pvr_queue *queue)
{
drm_sched_entity_destroy(&queue->entity);
dma_fence_put(queue->last_queued_job_scheduled_fence);
queue->last_queued_job_scheduled_fence = NULL;
}
void pvr_queue_destroy(struct pvr_queue *queue)
{
if (!queue)
return;
mutex_lock(&queue->ctx->pvr_dev->queues.lock);
list_del_init(&queue->node);
mutex_unlock(&queue->ctx->pvr_dev->queues.lock);
drm_sched_fini(&queue->scheduler);
drm_sched_entity_fini(&queue->entity);
if (WARN_ON(queue->last_queued_job_scheduled_fence))
dma_fence_put(queue->last_queued_job_scheduled_fence);
pvr_queue_cleanup_fw_context(queue);
pvr_fw_object_unmap_and_destroy(queue->timeline_ufo.fw_obj);
pvr_fw_object_destroy(queue->reg_state_obj);
pvr_cccb_fini(&queue->cccb);
mutex_destroy(&queue->cccb_fence_ctx.job_lock);
kfree(queue);
}
int pvr_queue_device_init(struct pvr_device *pvr_dev)
{
int err;
INIT_LIST_HEAD(&pvr_dev->queues.active);
INIT_LIST_HEAD(&pvr_dev->queues.idle);
err = drmm_mutex_init(from_pvr_device(pvr_dev), &pvr_dev->queues.lock);
if (err)
return err;
pvr_dev->sched_wq = alloc_workqueue("powervr-sched", WQ_UNBOUND, 0);
if (!pvr_dev->sched_wq)
return -ENOMEM;
return 0;
}
void pvr_queue_device_fini(struct pvr_device *pvr_dev)
{
destroy_workqueue(pvr_dev->sched_wq);
}