|
@@ -140,28 +140,38 @@ static const struct drm_sched_backend_ops etnaviv_sched_ops = {
|
|
|
int etnaviv_sched_push_job(struct drm_sched_entity *sched_entity,
|
|
|
struct etnaviv_gem_submit *submit)
|
|
|
{
|
|
|
- int ret;
|
|
|
+ int ret = 0;
|
|
|
+
|
|
|
+ /*
|
|
|
+ * Hold the fence lock across the whole operation to avoid jobs being
|
|
|
+ * pushed out of order with regard to their sched fence seqnos as
|
|
|
+ * allocated in drm_sched_job_init.
|
|
|
+ */
|
|
|
+ mutex_lock(&submit->gpu->fence_lock);
|
|
|
|
|
|
ret = drm_sched_job_init(&submit->sched_job, &submit->gpu->sched,
|
|
|
sched_entity, submit->cmdbuf.ctx);
|
|
|
if (ret)
|
|
|
- return ret;
|
|
|
+ goto out_unlock;
|
|
|
|
|
|
submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished);
|
|
|
- mutex_lock(&submit->gpu->fence_idr_lock);
|
|
|
submit->out_fence_id = idr_alloc_cyclic(&submit->gpu->fence_idr,
|
|
|
submit->out_fence, 0,
|
|
|
INT_MAX, GFP_KERNEL);
|
|
|
- mutex_unlock(&submit->gpu->fence_idr_lock);
|
|
|
- if (submit->out_fence_id < 0)
|
|
|
- return -ENOMEM;
|
|
|
+ if (submit->out_fence_id < 0) {
|
|
|
+ ret = -ENOMEM;
|
|
|
+ goto out_unlock;
|
|
|
+ }
|
|
|
|
|
|
/* the scheduler holds on to the job now */
|
|
|
kref_get(&submit->refcount);
|
|
|
|
|
|
drm_sched_entity_push_job(&submit->sched_job, sched_entity);
|
|
|
|
|
|
- return 0;
|
|
|
+out_unlock:
|
|
|
+ mutex_unlock(&submit->gpu->fence_lock);
|
|
|
+
|
|
|
+ return ret;
|
|
|
}
|
|
|
|
|
|
int etnaviv_sched_init(struct etnaviv_gpu *gpu)
|