|
@@ -2260,6 +2260,17 @@ int i915_gem_context_create_ioctl(struct drm_device *dev, void *data,
|
|
int i915_gem_context_destroy_ioctl(struct drm_device *dev, void *data,
|
|
int i915_gem_context_destroy_ioctl(struct drm_device *dev, void *data,
|
|
struct drm_file *file);
|
|
struct drm_file *file);
|
|
|
|
|
|
|
|
+/* i915_gem_evict.c */
|
|
|
|
+int __must_check i915_gem_evict_something(struct drm_device *dev,
|
|
|
|
+ struct i915_address_space *vm,
|
|
|
|
+ int min_size,
|
|
|
|
+ unsigned alignment,
|
|
|
|
+ unsigned cache_level,
|
|
|
|
+ bool mappable,
|
|
|
|
+ bool nonblock);
|
|
|
|
+int i915_gem_evict_vm(struct i915_address_space *vm, bool do_idle);
|
|
|
|
+int i915_gem_evict_everything(struct drm_device *dev);
|
|
|
|
+
|
|
/* i915_gem_gtt.c */
|
|
/* i915_gem_gtt.c */
|
|
void i915_check_and_clear_faults(struct drm_device *dev);
|
|
void i915_check_and_clear_faults(struct drm_device *dev);
|
|
void i915_gem_suspend_gtt_mappings(struct drm_device *dev);
|
|
void i915_gem_suspend_gtt_mappings(struct drm_device *dev);
|
|
@@ -2297,22 +2308,39 @@ static inline bool intel_enable_ppgtt(struct drm_device *dev, bool full)
|
|
static inline void ppgtt_release(struct kref *kref)
|
|
static inline void ppgtt_release(struct kref *kref)
|
|
{
|
|
{
|
|
struct i915_hw_ppgtt *ppgtt = container_of(kref, struct i915_hw_ppgtt, ref);
|
|
struct i915_hw_ppgtt *ppgtt = container_of(kref, struct i915_hw_ppgtt, ref);
|
|
|
|
+ struct drm_device *dev = ppgtt->base.dev;
|
|
|
|
+ struct drm_i915_private *dev_priv = dev->dev_private;
|
|
|
|
+ struct i915_address_space *vm = &ppgtt->base;
|
|
|
|
+
|
|
|
|
+ if (ppgtt == dev_priv->mm.aliasing_ppgtt ||
|
|
|
|
+ (list_empty(&vm->active_list) && list_empty(&vm->inactive_list))) {
|
|
|
|
+ ppgtt->base.cleanup(&ppgtt->base);
|
|
|
|
+ return;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * Make sure vmas are unbound before we take down the drm_mm
|
|
|
|
+ *
|
|
|
|
+ * FIXME: Proper refcounting should take care of this, this shouldn't be
|
|
|
|
+ * needed at all.
|
|
|
|
+ */
|
|
|
|
+ if (!list_empty(&vm->active_list)) {
|
|
|
|
+ struct i915_vma *vma;
|
|
|
|
+
|
|
|
|
+ list_for_each_entry(vma, &vm->active_list, mm_list)
|
|
|
|
+ if (WARN_ON(list_empty(&vma->vma_link) ||
|
|
|
|
+ list_is_singular(&vma->vma_link)))
|
|
|
|
+ break;
|
|
|
|
+
|
|
|
|
+ i915_gem_evict_vm(&ppgtt->base, true);
|
|
|
|
+ } else {
|
|
|
|
+ i915_gem_retire_requests(dev);
|
|
|
|
+ i915_gem_evict_vm(&ppgtt->base, false);
|
|
|
|
+ }
|
|
|
|
|
|
ppgtt->base.cleanup(&ppgtt->base);
|
|
ppgtt->base.cleanup(&ppgtt->base);
|
|
}
|
|
}
|
|
|
|
|
|
-
|
|
|
|
-/* i915_gem_evict.c */
|
|
|
|
-int __must_check i915_gem_evict_something(struct drm_device *dev,
|
|
|
|
- struct i915_address_space *vm,
|
|
|
|
- int min_size,
|
|
|
|
- unsigned alignment,
|
|
|
|
- unsigned cache_level,
|
|
|
|
- bool mappable,
|
|
|
|
- bool nonblock);
|
|
|
|
-int i915_gem_evict_vm(struct i915_address_space *vm, bool do_idle);
|
|
|
|
-int i915_gem_evict_everything(struct drm_device *dev);
|
|
|
|
-
|
|
|
|
/* i915_gem_stolen.c */
|
|
/* i915_gem_stolen.c */
|
|
int i915_gem_init_stolen(struct drm_device *dev);
|
|
int i915_gem_init_stolen(struct drm_device *dev);
|
|
int i915_gem_stolen_setup_compression(struct drm_device *dev, int size);
|
|
int i915_gem_stolen_setup_compression(struct drm_device *dev, int size);
|