drm/vmwgfx: switch driver from bo->resv to bo->base.resv
authorGerd Hoffmann <kraxel@redhat.com>
Mon, 5 Aug 2019 14:01:14 +0000 (16:01 +0200)
committerGerd Hoffmann <kraxel@redhat.com>
Tue, 6 Aug 2019 06:21:54 +0000 (08:21 +0200)
Signed-off-by: Gerd Hoffmann <kraxel@redhat.com>
Acked-by: Thomas Hellstrom <thellstrom@vmware.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20190805140119.7337-13-kraxel@redhat.com
drivers/gpu/drm/vmwgfx/vmwgfx_blit.c
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c

index 703786e3d5796366e48104a8725b5f3c21c2dfb7..6c01ad2785dd80a106e796c93f3f951aed1da101 100644 (file)
@@ -459,9 +459,9 @@ int vmw_bo_cpu_blit(struct ttm_buffer_object *dst,
 
        /* Buffer objects need to be either pinned or reserved: */
        if (!(dst->mem.placement & TTM_PL_FLAG_NO_EVICT))
-               reservation_object_assert_held(dst->resv);
+               reservation_object_assert_held(dst->base.resv);
        if (!(src->mem.placement & TTM_PL_FLAG_NO_EVICT))
-               reservation_object_assert_held(src->resv);
+               reservation_object_assert_held(src->base.resv);
 
        if (dst->ttm->state == tt_unpopulated) {
                ret = dst->ttm->bdev->driver->ttm_tt_populate(dst->ttm, &ctx);
index 5739c6c49c9953415286bf705cf41c92be78b59a..369034c0de311388d00159e27ea35608f4f3c8c3 100644 (file)
@@ -342,7 +342,7 @@ void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin)
        uint32_t old_mem_type = bo->mem.mem_type;
        int ret;
 
-       reservation_object_assert_held(bo->resv);
+       reservation_object_assert_held(bo->base.resv);
 
        if (pin) {
                if (vbo->pin_count++ > 0)
@@ -690,7 +690,7 @@ static int vmw_user_bo_synccpu_grab(struct vmw_user_buffer_object *user_bo,
                long lret;
 
                lret = reservation_object_wait_timeout_rcu
-                       (bo->resv, true, true,
+                       (bo->base.resv, true, true,
                         nonblock ? 0 : MAX_SCHEDULE_TIMEOUT);
                if (!lret)
                        return -EBUSY;
@@ -1007,10 +1007,10 @@ void vmw_bo_fence_single(struct ttm_buffer_object *bo,
 
        if (fence == NULL) {
                vmw_execbuf_fence_commands(NULL, dev_priv, &fence, NULL);
-               reservation_object_add_excl_fence(bo->resv, &fence->base);
+               reservation_object_add_excl_fence(bo->base.resv, &fence->base);
                dma_fence_put(&fence->base);
        } else
-               reservation_object_add_excl_fence(bo->resv, &fence->base);
+               reservation_object_add_excl_fence(bo->base.resv, &fence->base);
 }
 
 
index 71e901bbed68c58d49022e992e0cf51ea3296cd5..7984f172ec4a6f88271ce1bd91874d4dab17da48 100644 (file)
@@ -169,7 +169,7 @@ static int vmw_cotable_unscrub(struct vmw_resource *res)
        } *cmd;
 
        WARN_ON_ONCE(bo->mem.mem_type != VMW_PL_MOB);
-       reservation_object_assert_held(bo->resv);
+       reservation_object_assert_held(bo->base.resv);
 
        cmd = VMW_FIFO_RESERVE(dev_priv, sizeof(*cmd));
        if (!cmd)
@@ -311,7 +311,7 @@ static int vmw_cotable_unbind(struct vmw_resource *res,
                return 0;
 
        WARN_ON_ONCE(bo->mem.mem_type != VMW_PL_MOB);
-       reservation_object_assert_held(bo->resv);
+       reservation_object_assert_held(bo->base.resv);
 
        mutex_lock(&dev_priv->binding_mutex);
        if (!vcotbl->scrubbed)
index 303d2c7d9ab370ba718ca58a73d78c45ff115743..701643b7b0c4b175361eb9169c90c4307606290a 100644 (file)
@@ -402,14 +402,14 @@ void vmw_resource_unreserve(struct vmw_resource *res,
 
        if (switch_backup && new_backup != res->backup) {
                if (res->backup) {
-                       reservation_object_assert_held(res->backup->base.resv);
+                       reservation_object_assert_held(res->backup->base.base.resv);
                        list_del_init(&res->mob_head);
                        vmw_bo_unreference(&res->backup);
                }
 
                if (new_backup) {
                        res->backup = vmw_bo_reference(new_backup);
-                       reservation_object_assert_held(new_backup->base.resv);
+                       reservation_object_assert_held(new_backup->base.base.resv);
                        list_add_tail(&res->mob_head, &new_backup->res_list);
                } else {
                        res->backup = NULL;
@@ -691,7 +691,7 @@ void vmw_resource_unbind_list(struct vmw_buffer_object *vbo)
                .num_shared = 0
        };
 
-       reservation_object_assert_held(vbo->base.resv);
+       reservation_object_assert_held(vbo->base.base.resv);
        list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) {
                if (!res->func->unbind)
                        continue;