drm/ttm: drop special pipeline accel cleanup function.
authorDave Airlie <airlied@redhat.com>
Thu, 17 Sep 2020 06:36:14 +0000 (16:36 +1000)
committerDave Airlie <airlied@redhat.com>
Thu, 17 Sep 2020 20:23:06 +0000 (06:23 +1000)
The two accel cleanup paths were mostly the same once refactored.

Just pass a bool to say if the evictions are to be pipelined.

Signed-off-by: Dave Airlie <airlied@redhat.com>
Reviewed-by: Christian König <christian.koenig@amd.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20200917064132.148521-2-airlied@gmail.com
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
drivers/gpu/drm/nouveau/nouveau_bo.c
drivers/gpu/drm/radeon/radeon_ttm.c
drivers/gpu/drm/ttm/ttm_bo_util.c
include/drm/ttm/ttm_bo_driver.h

index a57aaf66634038c75e7d6d33fb87eb19de0a31cd..e7b67dc330a4a160aaf233796ece6074ce0e2d4c 100644 (file)
@@ -500,9 +500,9 @@ static int amdgpu_move_blit(struct ttm_buffer_object *bo,
 
        /* Always block for VM page tables before committing the new location */
        if (bo->type == ttm_bo_type_kernel)
-               r = ttm_bo_move_accel_cleanup(bo, fence, true, new_mem);
+               r = ttm_bo_move_accel_cleanup(bo, fence, true, false, new_mem);
        else
-               r = ttm_bo_pipeline_move(bo, fence, evict, new_mem);
+               r = ttm_bo_move_accel_cleanup(bo, fence, evict, true, new_mem);
        dma_fence_put(fence);
        return r;
 
index 56f974c28eb510b970714fa3d88e3f71e9c6fc06..2ee75646ad6fcb44e33435981d92e2147885d9b8 100644 (file)
@@ -824,7 +824,7 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
                        if (ret == 0) {
                                ret = ttm_bo_move_accel_cleanup(bo,
                                                                &fence->base,
-                                                               evict,
+                                                               evict, false,
                                                                new_reg);
                                nouveau_fence_unref(&fence);
                        }
index 99d9ca1087b7759b0e610150f9fc5e1489f550dd..36150b7f31a90aa1eece8a9f172b2b7897ab3193 100644 (file)
@@ -200,7 +200,7 @@ static int radeon_move_blit(struct ttm_buffer_object *bo,
        if (IS_ERR(fence))
                return PTR_ERR(fence);
 
-       r = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, new_mem);
+       r = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, false, new_mem);
        radeon_fence_unref(&fence);
        return r;
 }
index ee4484ccbd1e5bdc5b2895ef66c3f6452b112c8d..fb2a25f8408fc8c1fb8da9b7829c100400c262a6 100644 (file)
@@ -580,81 +580,56 @@ static int ttm_bo_move_to_ghost(struct ttm_buffer_object *bo,
        return 0;
 }
 
-int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
-                             struct dma_fence *fence,
-                             bool evict,
-                             struct ttm_resource *new_mem)
+static void ttm_bo_move_pipeline_evict(struct ttm_buffer_object *bo,
+                                      struct dma_fence *fence)
 {
        struct ttm_bo_device *bdev = bo->bdev;
-       struct ttm_resource_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
-       int ret;
+       struct ttm_resource_manager *from = ttm_manager_type(bdev, bo->mem.mem_type);
 
-       dma_resv_add_excl_fence(bo->base.resv, fence);
-       if (evict)
-               ret = ttm_bo_wait_free_node(bo, man->use_tt);
-       else
-               ret = ttm_bo_move_to_ghost(bo, fence, man->use_tt);
-       if (ret)
-               return ret;
+       /**
+        * BO doesn't have a TTM we need to bind/unbind. Just remember
+        * this eviction and free up the allocation
+        */
+       spin_lock(&from->move_lock);
+       if (!from->move || dma_fence_is_later(fence, from->move)) {
+               dma_fence_put(from->move);
+               from->move = dma_fence_get(fence);
+       }
+       spin_unlock(&from->move_lock);
 
-       ttm_bo_assign_mem(bo, new_mem);
+       ttm_bo_free_old_node(bo);
 
-       return 0;
+       dma_fence_put(bo->moving);
+       bo->moving = dma_fence_get(fence);
 }
-EXPORT_SYMBOL(ttm_bo_move_accel_cleanup);
 
-int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
-                        struct dma_fence *fence, bool evict,
-                        struct ttm_resource *new_mem)
+int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
+                             struct dma_fence *fence,
+                             bool evict,
+                             bool pipeline,
+                             struct ttm_resource *new_mem)
 {
        struct ttm_bo_device *bdev = bo->bdev;
-
        struct ttm_resource_manager *from = ttm_manager_type(bdev, bo->mem.mem_type);
-       struct ttm_resource_manager *to = ttm_manager_type(bdev, new_mem->mem_type);
-
-       int ret;
+       struct ttm_resource_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
+       int ret = 0;
 
        dma_resv_add_excl_fence(bo->base.resv, fence);
+       if (!evict)
+               ret = ttm_bo_move_to_ghost(bo, fence, man->use_tt);
+       else if (!from->use_tt && pipeline)
+               ttm_bo_move_pipeline_evict(bo, fence);
+       else
+               ret = ttm_bo_wait_free_node(bo, man->use_tt);
 
-       if (!evict) {
-               ret = ttm_bo_move_to_ghost(bo, fence, to->use_tt);
-               if (ret)
-                       return ret;
-       } else if (!from->use_tt) {
-
-               /**
-                * BO doesn't have a TTM we need to bind/unbind. Just remember
-                * this eviction and free up the allocation
-                */
-
-               spin_lock(&from->move_lock);
-               if (!from->move || dma_fence_is_later(fence, from->move)) {
-                       dma_fence_put(from->move);
-                       from->move = dma_fence_get(fence);
-               }
-               spin_unlock(&from->move_lock);
-
-               ttm_bo_free_old_node(bo);
-
-               dma_fence_put(bo->moving);
-               bo->moving = dma_fence_get(fence);
-
-       } else {
-               /**
-                * Last resort, wait for the move to be completed.
-                *
-                * Should never happen in pratice.
-                */
-               ret = ttm_bo_wait_free_node(bo, to->use_tt);
-               if (ret)
-                       return ret;
-       }
+       if (ret)
+               return ret;
 
        ttm_bo_assign_mem(bo, new_mem);
 
        return 0;
 }
-EXPORT_SYMBOL(ttm_bo_pipeline_move);
+EXPORT_SYMBOL(ttm_bo_move_accel_cleanup);
 
 int ttm_bo_pipeline_gutting(struct ttm_buffer_object *bo)
 {
index 25cc932d63f1d5bacb7597b6351fd3f7ee5a3c21..864afa8f6f18f4eb80057757e49866c4e8150aba 100644 (file)
@@ -642,6 +642,7 @@ void ttm_bo_free_old_node(struct ttm_buffer_object *bo);
  * @bo: A pointer to a struct ttm_buffer_object.
  * @fence: A fence object that signals when moving is complete.
  * @evict: This is an evict move. Don't return until the buffer is idle.
+ * @pipeline: evictions are to be pipelined.
  * @new_mem: struct ttm_resource indicating where to move.
  *
  * Accelerated move function to be called when an accelerated move
@@ -653,23 +654,9 @@ void ttm_bo_free_old_node(struct ttm_buffer_object *bo);
  */
 int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
                              struct dma_fence *fence, bool evict,
+                             bool pipeline,
                              struct ttm_resource *new_mem);
 
-/**
- * ttm_bo_pipeline_move.
- *
- * @bo: A pointer to a struct ttm_buffer_object.
- * @fence: A fence object that signals when moving is complete.
- * @evict: This is an evict move. Don't return until the buffer is idle.
- * @new_mem: struct ttm_resource indicating where to move.
- *
- * Function for pipelining accelerated moves. Either free the memory
- * immediately or hang it on a temporary buffer object.
- */
-int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
-                        struct dma_fence *fence, bool evict,
-                        struct ttm_resource *new_mem);
-
 /**
  * ttm_bo_pipeline_gutting.
  *