Merge drm/drm-next into drm-intel-gt-next
[platform/kernel/linux-rpi.git] / drivers / gpu / drm / i915 / i915_vma.c
index bef795e..e2f2c4c 100644 (file)
 
 static struct kmem_cache *slab_vmas;
 
-struct i915_vma *i915_vma_alloc(void)
+static struct i915_vma *i915_vma_alloc(void)
 {
        return kmem_cache_zalloc(slab_vmas, GFP_KERNEL);
 }
 
-void i915_vma_free(struct i915_vma *vma)
+static void i915_vma_free(struct i915_vma *vma)
 {
        return kmem_cache_free(slab_vmas, vma);
 }
@@ -113,7 +113,6 @@ vma_create(struct drm_i915_gem_object *obj,
        vma->vm = i915_vm_get(vm);
        vma->ops = &vm->vma_ops;
        vma->obj = obj;
-       vma->resv = obj->base.resv;
        vma->size = obj->base.size;
        vma->display_alignment = I915_GTT_MIN_ALIGNMENT;
 
@@ -346,7 +345,7 @@ int i915_vma_wait_for_bind(struct i915_vma *vma)
                fence = dma_fence_get_rcu_safe(&vma->active.excl.fence);
                rcu_read_unlock();
                if (fence) {
-                       err = dma_fence_wait(fence, MAX_SCHEDULE_TIMEOUT);
+                       err = dma_fence_wait(fence, true);
                        dma_fence_put(fence);
                }
        }
@@ -423,10 +422,8 @@ int i915_vma_bind(struct i915_vma *vma,
 
                work->base.dma.error = 0; /* enable the queue_work() */
 
-               if (vma->obj) {
-                       __i915_gem_object_pin_pages(vma->obj);
-                       work->pinned = i915_gem_object_get(vma->obj);
-               }
+               __i915_gem_object_pin_pages(vma->obj);
+               work->pinned = i915_gem_object_get(vma->obj);
        } else {
                vma->ops->bind_vma(vma->vm, NULL, vma, cache_level, bind_flags);
        }
@@ -667,7 +664,7 @@ i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags)
        }
 
        color = 0;
-       if (vma->obj && i915_vm_has_cache_coloring(vma->vm))
+       if (i915_vm_has_cache_coloring(vma->vm))
                color = vma->obj->cache_level;
 
        if (flags & PIN_OFFSET_FIXED) {
@@ -792,17 +789,14 @@ unpinned:
 static int vma_get_pages(struct i915_vma *vma)
 {
        int err = 0;
-       bool pinned_pages = false;
+       bool pinned_pages = true;
 
        if (atomic_add_unless(&vma->pages_count, 1, 0))
                return 0;
 
-       if (vma->obj) {
-               err = i915_gem_object_pin_pages(vma->obj);
-               if (err)
-                       return err;
-               pinned_pages = true;
-       }
+       err = i915_gem_object_pin_pages(vma->obj);
+       if (err)
+               return err;
 
        /* Allocations ahoy! */
        if (mutex_lock_interruptible(&vma->pages_mutex)) {
@@ -835,8 +829,8 @@ static void __vma_put_pages(struct i915_vma *vma, unsigned int count)
        if (atomic_sub_return(count, &vma->pages_count) == 0) {
                vma->ops->clear_pages(vma);
                GEM_BUG_ON(vma->pages);
-               if (vma->obj)
-                       i915_gem_object_unpin_pages(vma->obj);
+
+               i915_gem_object_unpin_pages(vma->obj);
        }
        mutex_unlock(&vma->pages_mutex);
 }
@@ -872,7 +866,7 @@ int i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
        int err;
 
 #ifdef CONFIG_PROVE_LOCKING
-       if (debug_locks && !WARN_ON(!ww) && vma->resv)
+       if (debug_locks && !WARN_ON(!ww))
                assert_vma_held(vma);
 #endif
 
@@ -980,7 +974,7 @@ int i915_vma_pin_ww(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
 
        GEM_BUG_ON(!vma->pages);
        err = i915_vma_bind(vma,
-                           vma->obj ? vma->obj->cache_level : 0,
+                           vma->obj->cache_level,
                            flags, work);
        if (err)
                goto err_remove;
@@ -1034,7 +1028,7 @@ int i915_ggtt_pin(struct i915_vma *vma, struct i915_gem_ww_ctx *ww,
        GEM_BUG_ON(!i915_vma_is_ggtt(vma));
 
 #ifdef CONFIG_LOCKDEP
-       WARN_ON(!ww && vma->resv && dma_resv_held(vma->resv));
+       WARN_ON(!ww && dma_resv_held(vma->obj->base.resv));
 #endif
 
        do {
@@ -1113,6 +1107,7 @@ void i915_vma_reopen(struct i915_vma *vma)
 void i915_vma_release(struct kref *ref)
 {
        struct i915_vma *vma = container_of(ref, typeof(*vma), ref);
+       struct drm_i915_gem_object *obj = vma->obj;
 
        if (drm_mm_node_allocated(&vma->node)) {
                mutex_lock(&vma->vm->mutex);
@@ -1123,15 +1118,11 @@ void i915_vma_release(struct kref *ref)
        }
        GEM_BUG_ON(i915_vma_is_active(vma));
 
-       if (vma->obj) {
-               struct drm_i915_gem_object *obj = vma->obj;
-
-               spin_lock(&obj->vma.lock);
-               list_del(&vma->obj_link);
-               if (!RB_EMPTY_NODE(&vma->obj_node))
-                       rb_erase(&vma->obj_node, &obj->vma.tree);
-               spin_unlock(&obj->vma.lock);
-       }
+       spin_lock(&obj->vma.lock);
+       list_del(&vma->obj_link);
+       if (!RB_EMPTY_NODE(&vma->obj_node))
+               rb_erase(&vma->obj_node, &obj->vma.tree);
+       spin_unlock(&obj->vma.lock);
 
        __i915_vma_remove_closed(vma);
        i915_vm_put(vma->vm);
@@ -1256,19 +1247,19 @@ int _i915_vma_move_to_active(struct i915_vma *vma,
                }
 
                if (fence) {
-                       dma_resv_add_excl_fence(vma->resv, fence);
+                       dma_resv_add_excl_fence(vma->obj->base.resv, fence);
                        obj->write_domain = I915_GEM_DOMAIN_RENDER;
                        obj->read_domains = 0;
                }
        } else {
                if (!(flags & __EXEC_OBJECT_NO_RESERVE)) {
-                       err = dma_resv_reserve_shared(vma->resv, 1);
+                       err = dma_resv_reserve_shared(vma->obj->base.resv, 1);
                        if (unlikely(err))
                                return err;
                }
 
                if (fence) {
-                       dma_resv_add_shared_fence(vma->resv, fence);
+                       dma_resv_add_shared_fence(vma->obj->base.resv, fence);
                        obj->write_domain = 0;
                }
        }