|
@@ -943,8 +943,6 @@ i915_gem_execbuffer_move_to_gpu(struct drm_i915_gem_request *req,
|
|
|
{
|
|
|
const unsigned other_rings = ~intel_engine_flag(req->engine);
|
|
|
struct i915_vma *vma;
|
|
|
- uint32_t flush_domains = 0;
|
|
|
- bool flush_chipset = false;
|
|
|
int ret;
|
|
|
|
|
|
list_for_each_entry(vma, vmas, exec_list) {
|
|
@@ -957,16 +955,11 @@ i915_gem_execbuffer_move_to_gpu(struct drm_i915_gem_request *req,
|
|
|
}
|
|
|
|
|
|
if (obj->base.write_domain & I915_GEM_DOMAIN_CPU)
|
|
|
- flush_chipset |= i915_gem_clflush_object(obj, false);
|
|
|
-
|
|
|
- flush_domains |= obj->base.write_domain;
|
|
|
+ i915_gem_clflush_object(obj, false);
|
|
|
}
|
|
|
|
|
|
- if (flush_chipset)
|
|
|
- i915_gem_chipset_flush(req->engine->i915);
|
|
|
-
|
|
|
- if (flush_domains & I915_GEM_DOMAIN_GTT)
|
|
|
- wmb();
|
|
|
+ /* Unconditionally flush any chipset caches (for streaming writes). */
|
|
|
+ i915_gem_chipset_flush(req->engine->i915);
|
|
|
|
|
|
/* Unconditionally invalidate gpu caches and ensure that we do flush
|
|
|
* any residual writes from the previous batch.
|