|
@@ -628,6 +628,7 @@ static int execlists_move_to_gpu(struct intel_ringbuffer *ringbuf,
|
|
|
struct list_head *vmas)
|
|
|
{
|
|
|
struct intel_engine_cs *ring = ringbuf->ring;
|
|
|
+ const unsigned other_rings = ~intel_ring_flag(ring);
|
|
|
struct i915_vma *vma;
|
|
|
uint32_t flush_domains = 0;
|
|
|
bool flush_chipset = false;
|
|
@@ -636,9 +637,11 @@ static int execlists_move_to_gpu(struct intel_ringbuffer *ringbuf,
|
|
|
list_for_each_entry(vma, vmas, exec_list) {
|
|
|
struct drm_i915_gem_object *obj = vma->obj;
|
|
|
|
|
|
- ret = i915_gem_object_sync(obj, ring);
|
|
|
- if (ret)
|
|
|
- return ret;
|
|
|
+ if (obj->active & other_rings) {
|
|
|
+ ret = i915_gem_object_sync(obj, ring);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
|
|
|
if (obj->base.write_domain & I915_GEM_DOMAIN_CPU)
|
|
|
flush_chipset |= i915_gem_clflush_object(obj, false);
|