2019-06-20 06:08:14

by Gerd Hoffmann

[permalink] [raw]
Subject: [PATCH v4 08/12] drm/virtio: rework virtio_gpu_execbuffer_ioctl fencing

Use gem reservation helpers and direct reservation_object_* calls
instead of ttm.

v3: Also attach the array of gem objects to the virtio command buffer,
so we can drop the object references in the completion callback. Needed
because ttm fence helpers grab a reference for us, but gem helpers
don't.

Signed-off-by: Gerd Hoffmann <[email protected]>
Acked-by: Daniel Vetter <[email protected]>
---
drivers/gpu/drm/virtio/virtgpu_drv.h | 5 ++-
drivers/gpu/drm/virtio/virtgpu_ioctl.c | 62 +++++++++++---------------
drivers/gpu/drm/virtio/virtgpu_vq.c | 16 ++++---
3 files changed, 40 insertions(+), 43 deletions(-)

diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h
index 98d646789d23..65f5ce41c341 100644
--- a/drivers/gpu/drm/virtio/virtgpu_drv.h
+++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
@@ -120,9 +120,9 @@ struct virtio_gpu_vbuffer {

char *resp_buf;
int resp_size;
-
virtio_gpu_resp_cb resp_cb;

+ struct virtio_gpu_object_array *objs;
struct list_head list;
};

@@ -311,7 +311,8 @@ void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
uint32_t resource_id);
void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
void *data, uint32_t data_size,
- uint32_t ctx_id, struct virtio_gpu_fence *fence);
+ uint32_t ctx_id, struct virtio_gpu_fence *fence,
+ struct virtio_gpu_object_array *objs);
void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
uint32_t resource_id, uint32_t ctx_id,
uint64_t offset, uint32_t level,
diff --git a/drivers/gpu/drm/virtio/virtgpu_ioctl.c b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
index 5cffd2e54c04..c3901c397b8a 100644
--- a/drivers/gpu/drm/virtio/virtgpu_ioctl.c
+++ b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
@@ -105,14 +105,11 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
struct drm_virtgpu_execbuffer *exbuf = data;
struct virtio_gpu_device *vgdev = dev->dev_private;
struct virtio_gpu_fpriv *vfpriv = drm_file->driver_priv;
- struct drm_gem_object *gobj;
struct virtio_gpu_fence *out_fence;
- struct virtio_gpu_object *qobj;
int ret;
uint32_t *bo_handles = NULL;
void __user *user_bo_handles = NULL;
- struct list_head validate_list;
- struct ttm_validate_buffer *buflist = NULL;
+ struct virtio_gpu_object_array *buflist = NULL;
int i;
struct ww_acquire_ctx ticket;
struct sync_file *sync_file;
@@ -155,15 +152,10 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
return out_fence_fd;
}

- INIT_LIST_HEAD(&validate_list);
if (exbuf->num_bo_handles) {
-
bo_handles = kvmalloc_array(exbuf->num_bo_handles,
- sizeof(uint32_t), GFP_KERNEL);
- buflist = kvmalloc_array(exbuf->num_bo_handles,
- sizeof(struct ttm_validate_buffer),
- GFP_KERNEL | __GFP_ZERO);
- if (!bo_handles || !buflist) {
+ sizeof(uint32_t), GFP_KERNEL);
+ if (!bo_handles) {
ret = -ENOMEM;
goto out_unused_fd;
}
@@ -175,25 +167,22 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
goto out_unused_fd;
}

- for (i = 0; i < exbuf->num_bo_handles; i++) {
- gobj = drm_gem_object_lookup(drm_file, bo_handles[i]);
- if (!gobj) {
- ret = -ENOENT;
- goto out_unused_fd;
- }
-
- qobj = gem_to_virtio_gpu_obj(gobj);
- buflist[i].bo = &qobj->tbo;
-
- list_add(&buflist[i].head, &validate_list);
+ buflist = virtio_gpu_array_from_handles(drm_file, bo_handles,
+ exbuf->num_bo_handles);
+ if (!buflist) {
+ ret = -ENOENT;
+ goto out_unused_fd;
}
kvfree(bo_handles);
bo_handles = NULL;
}

- ret = virtio_gpu_object_list_validate(&ticket, &validate_list);
- if (ret)
- goto out_free;
+ if (buflist) {
+ ret = drm_gem_lock_reservations(buflist->objs, buflist->nents,
+ &ticket);
+ if (ret)
+ goto out_unused_fd;
+ }

buf = memdup_user(u64_to_user_ptr(exbuf->command), exbuf->size);
if (IS_ERR(buf)) {
@@ -219,25 +208,26 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
fd_install(out_fence_fd, sync_file->file);
}

+ if (buflist) {
+ for (i = 0; i < exbuf->num_bo_handles; i++)
+ reservation_object_add_excl_fence(buflist->objs[i]->resv,
+ &out_fence->f);
+ drm_gem_unlock_reservations(buflist->objs, buflist->nents,
+ &ticket);
+ }
+
virtio_gpu_cmd_submit(vgdev, buf, exbuf->size,
- vfpriv->ctx_id, out_fence);
-
- ttm_eu_fence_buffer_objects(&ticket, &validate_list, &out_fence->f);
-
- /* fence the command bo */
- virtio_gpu_unref_list(&validate_list);
- kvfree(buflist);
+ vfpriv->ctx_id, out_fence, buflist);
return 0;

out_memdup:
kfree(buf);
out_unresv:
- ttm_eu_backoff_reservation(&ticket, &validate_list);
-out_free:
- virtio_gpu_unref_list(&validate_list);
+ drm_gem_unlock_reservations(buflist->objs, buflist->nents, &ticket);
out_unused_fd:
kvfree(bo_handles);
- kvfree(buflist);
+ if (buflist)
+ virtio_gpu_array_put_free(buflist);

if (out_fence_fd >= 0)
put_unused_fd(out_fence_fd);
diff --git a/drivers/gpu/drm/virtio/virtgpu_vq.c b/drivers/gpu/drm/virtio/virtgpu_vq.c
index 6c1a90717535..dc2c2c003200 100644
--- a/drivers/gpu/drm/virtio/virtgpu_vq.c
+++ b/drivers/gpu/drm/virtio/virtgpu_vq.c
@@ -191,7 +191,7 @@ void virtio_gpu_dequeue_ctrl_func(struct work_struct *work)
} while (!virtqueue_enable_cb(vgdev->ctrlq.vq));
spin_unlock(&vgdev->ctrlq.qlock);

- list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
+ list_for_each_entry(entry, &reclaim_list, list) {
resp = (struct virtio_gpu_ctrl_hdr *)entry->resp_buf;

trace_virtio_gpu_cmd_response(vgdev->ctrlq.vq, resp);
@@ -218,14 +218,18 @@ void virtio_gpu_dequeue_ctrl_func(struct work_struct *work)
}
if (entry->resp_cb)
entry->resp_cb(vgdev, entry);
-
- list_del(&entry->list);
- free_vbuf(vgdev, entry);
}
wake_up(&vgdev->ctrlq.ack_queue);

if (fence_id)
virtio_gpu_fence_event_process(vgdev, fence_id);
+
+ list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
+ if (entry->objs)
+ virtio_gpu_array_put_free(entry->objs);
+ list_del(&entry->list);
+ free_vbuf(vgdev, entry);
+ }
}

void virtio_gpu_dequeue_cursor_func(struct work_struct *work)
@@ -939,7 +943,8 @@ void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,

void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
void *data, uint32_t data_size,
- uint32_t ctx_id, struct virtio_gpu_fence *fence)
+ uint32_t ctx_id, struct virtio_gpu_fence *fence,
+ struct virtio_gpu_object_array *objs)
{
struct virtio_gpu_cmd_submit *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
@@ -949,6 +954,7 @@ void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,

vbuf->data_buf = data;
vbuf->data_size = data_size;
+ vbuf->objs = objs;

cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_SUBMIT_3D);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
--
2.18.1


2019-06-27 01:33:42

by Chia-I Wu

[permalink] [raw]
Subject: Re: [PATCH v4 08/12] drm/virtio: rework virtio_gpu_execbuffer_ioctl fencing

On Wed, Jun 19, 2019 at 11:08 PM Gerd Hoffmann <[email protected]> wrote:
>
> Use gem reservation helpers and direct reservation_object_* calls
> instead of ttm.
>
> v3: Also attach the array of gem objects to the virtio command buffer,
> so we can drop the object references in the completion callback. Needed
> because ttm fence helpers grab a reference for us, but gem helpers
> don't.
>
> Signed-off-by: Gerd Hoffmann <[email protected]>
> Acked-by: Daniel Vetter <[email protected]>
> ---
> drivers/gpu/drm/virtio/virtgpu_drv.h | 5 ++-
> drivers/gpu/drm/virtio/virtgpu_ioctl.c | 62 +++++++++++---------------
> drivers/gpu/drm/virtio/virtgpu_vq.c | 16 ++++---
> 3 files changed, 40 insertions(+), 43 deletions(-)
>
> diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h
> index 98d646789d23..65f5ce41c341 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_drv.h
> +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
> @@ -120,9 +120,9 @@ struct virtio_gpu_vbuffer {
>
> char *resp_buf;
> int resp_size;
> -
> virtio_gpu_resp_cb resp_cb;
>
> + struct virtio_gpu_object_array *objs;
This can use a comment (e.g., objects referenced by the vbuffer)
> struct list_head list;
> };
>
> @@ -311,7 +311,8 @@ void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
> uint32_t resource_id);
> void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
> void *data, uint32_t data_size,
> - uint32_t ctx_id, struct virtio_gpu_fence *fence);
> + uint32_t ctx_id, struct virtio_gpu_fence *fence,
> + struct virtio_gpu_object_array *objs);
Can we keep fence, which is updated, as the last parameter?
> void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
> uint32_t resource_id, uint32_t ctx_id,
> uint64_t offset, uint32_t level,
> diff --git a/drivers/gpu/drm/virtio/virtgpu_ioctl.c b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> index 5cffd2e54c04..c3901c397b8a 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
> @@ -105,14 +105,11 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
> struct drm_virtgpu_execbuffer *exbuf = data;
> struct virtio_gpu_device *vgdev = dev->dev_private;
> struct virtio_gpu_fpriv *vfpriv = drm_file->driver_priv;
> - struct drm_gem_object *gobj;
> struct virtio_gpu_fence *out_fence;
> - struct virtio_gpu_object *qobj;
> int ret;
> uint32_t *bo_handles = NULL;
> void __user *user_bo_handles = NULL;
> - struct list_head validate_list;
> - struct ttm_validate_buffer *buflist = NULL;
> + struct virtio_gpu_object_array *buflist = NULL;
> int i;
> struct ww_acquire_ctx ticket;
> struct sync_file *sync_file;
> @@ -155,15 +152,10 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
> return out_fence_fd;
> }
>
> - INIT_LIST_HEAD(&validate_list);
> if (exbuf->num_bo_handles) {
> -
> bo_handles = kvmalloc_array(exbuf->num_bo_handles,
> - sizeof(uint32_t), GFP_KERNEL);
> - buflist = kvmalloc_array(exbuf->num_bo_handles,
> - sizeof(struct ttm_validate_buffer),
> - GFP_KERNEL | __GFP_ZERO);
> - if (!bo_handles || !buflist) {
> + sizeof(uint32_t), GFP_KERNEL);
> + if (!bo_handles) {
> ret = -ENOMEM;
> goto out_unused_fd;
> }
> @@ -175,25 +167,22 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
> goto out_unused_fd;
> }
>
> - for (i = 0; i < exbuf->num_bo_handles; i++) {
> - gobj = drm_gem_object_lookup(drm_file, bo_handles[i]);
> - if (!gobj) {
> - ret = -ENOENT;
> - goto out_unused_fd;
> - }
> -
> - qobj = gem_to_virtio_gpu_obj(gobj);
> - buflist[i].bo = &qobj->tbo;
> -
> - list_add(&buflist[i].head, &validate_list);
> + buflist = virtio_gpu_array_from_handles(drm_file, bo_handles,
> + exbuf->num_bo_handles);
> + if (!buflist) {
> + ret = -ENOENT;
> + goto out_unused_fd;
> }
> kvfree(bo_handles);
> bo_handles = NULL;
> }
>
> - ret = virtio_gpu_object_list_validate(&ticket, &validate_list);
> - if (ret)
> - goto out_free;
> + if (buflist) {
> + ret = drm_gem_lock_reservations(buflist->objs, buflist->nents,
> + &ticket);
> + if (ret)
> + goto out_unused_fd;
> + }
>
> buf = memdup_user(u64_to_user_ptr(exbuf->command), exbuf->size);
> if (IS_ERR(buf)) {
> @@ -219,25 +208,26 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
> fd_install(out_fence_fd, sync_file->file);
> }
>
> + if (buflist) {
> + for (i = 0; i < exbuf->num_bo_handles; i++)
> + reservation_object_add_excl_fence(buflist->objs[i]->resv,
> + &out_fence->f);
> + drm_gem_unlock_reservations(buflist->objs, buflist->nents,
> + &ticket);
> + }
We used to unlock after virtio_gpu_cmd_submit.

I guess, the fence is considered signaled (because its seqno is still
0) until after virtio_gpu_cmd_submit. We probably don't want other
processes to see the semi-initialized fence.


> +
> virtio_gpu_cmd_submit(vgdev, buf, exbuf->size,
> - vfpriv->ctx_id, out_fence);
> -
> - ttm_eu_fence_buffer_objects(&ticket, &validate_list, &out_fence->f);
> -
> - /* fence the command bo */
> - virtio_gpu_unref_list(&validate_list);
> - kvfree(buflist);
> + vfpriv->ctx_id, out_fence, buflist);
> return 0;
>
> out_memdup:
> kfree(buf);
> out_unresv:
> - ttm_eu_backoff_reservation(&ticket, &validate_list);
> -out_free:
> - virtio_gpu_unref_list(&validate_list);
Keeping out_free to free buflist seems just fine.
> + drm_gem_unlock_reservations(buflist->objs, buflist->nents, &ticket);
> out_unused_fd:
> kvfree(bo_handles);
> - kvfree(buflist);
> + if (buflist)
> + virtio_gpu_array_put_free(buflist);
>
> if (out_fence_fd >= 0)
> put_unused_fd(out_fence_fd);
> diff --git a/drivers/gpu/drm/virtio/virtgpu_vq.c b/drivers/gpu/drm/virtio/virtgpu_vq.c
> index 6c1a90717535..dc2c2c003200 100644
> --- a/drivers/gpu/drm/virtio/virtgpu_vq.c
> +++ b/drivers/gpu/drm/virtio/virtgpu_vq.c
> @@ -191,7 +191,7 @@ void virtio_gpu_dequeue_ctrl_func(struct work_struct *work)
> } while (!virtqueue_enable_cb(vgdev->ctrlq.vq));
> spin_unlock(&vgdev->ctrlq.qlock);
>
> - list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
> + list_for_each_entry(entry, &reclaim_list, list) {
> resp = (struct virtio_gpu_ctrl_hdr *)entry->resp_buf;
>
> trace_virtio_gpu_cmd_response(vgdev->ctrlq.vq, resp);
> @@ -218,14 +218,18 @@ void virtio_gpu_dequeue_ctrl_func(struct work_struct *work)
> }
> if (entry->resp_cb)
> entry->resp_cb(vgdev, entry);
> -
> - list_del(&entry->list);
> - free_vbuf(vgdev, entry);
> }
> wake_up(&vgdev->ctrlq.ack_queue);
>
> if (fence_id)
> virtio_gpu_fence_event_process(vgdev, fence_id);
> +
> + list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
> + if (entry->objs)
> + virtio_gpu_array_put_free(entry->objs);
> + list_del(&entry->list);
We are clearing the list. I guess list_del is not needed.
> + free_vbuf(vgdev, entry);
> + }
> }
>
> void virtio_gpu_dequeue_cursor_func(struct work_struct *work)
> @@ -939,7 +943,8 @@ void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
>
> void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
> void *data, uint32_t data_size,
> - uint32_t ctx_id, struct virtio_gpu_fence *fence)
> + uint32_t ctx_id, struct virtio_gpu_fence *fence,
> + struct virtio_gpu_object_array *objs)
> {
> struct virtio_gpu_cmd_submit *cmd_p;
> struct virtio_gpu_vbuffer *vbuf;
> @@ -949,6 +954,7 @@ void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
>
> vbuf->data_buf = data;
> vbuf->data_size = data_size;
> + vbuf->objs = objs;
>
> cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_SUBMIT_3D);
> cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
> --
> 2.18.1
>
> _______________________________________________
> dri-devel mailing list
> [email protected]
> https://lists.freedesktop.org/mailman/listinfo/dri-devel

2019-06-28 10:36:07

by Gerd Hoffmann

[permalink] [raw]
Subject: Re: [PATCH v4 08/12] drm/virtio: rework virtio_gpu_execbuffer_ioctl fencing

Hi,

> > --- a/drivers/gpu/drm/virtio/virtgpu_drv.h
> > +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
> > @@ -120,9 +120,9 @@ struct virtio_gpu_vbuffer {
> >
> > char *resp_buf;
> > int resp_size;
> > -
> > virtio_gpu_resp_cb resp_cb;
> >
> > + struct virtio_gpu_object_array *objs;
> This can use a comment (e.g., objects referenced by the vbuffer)

IMHO this is obvious ...

> > void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
> > void *data, uint32_t data_size,
> > - uint32_t ctx_id, struct virtio_gpu_fence *fence);
> > + uint32_t ctx_id, struct virtio_gpu_fence *fence,
> > + struct virtio_gpu_object_array *objs);
> Can we keep fence, which is updated, as the last parameter?

Fixed.

> > + if (buflist) {
> > + for (i = 0; i < exbuf->num_bo_handles; i++)
> > + reservation_object_add_excl_fence(buflist->objs[i]->resv,
> > + &out_fence->f);
> > + drm_gem_unlock_reservations(buflist->objs, buflist->nents,
> > + &ticket);
> > + }
> We used to unlock after virtio_gpu_cmd_submit.
>
> I guess, the fence is considered signaled (because its seqno is still
> 0) until after virtio_gpu_cmd_submit. We probably don't want other
> processes to see the semi-initialized fence.

Good point. Fixed.

> > out_memdup:
> > kfree(buf);
> > out_unresv:
> > - ttm_eu_backoff_reservation(&ticket, &validate_list);
> > -out_free:
> > - virtio_gpu_unref_list(&validate_list);
> Keeping out_free to free buflist seems just fine.

We don't need the separate label though ...

> > + drm_gem_unlock_reservations(buflist->objs, buflist->nents, &ticket);
> > out_unused_fd:
> > kvfree(bo_handles);
> > - kvfree(buflist);
> > + if (buflist)
> > + virtio_gpu_array_put_free(buflist);

... and the buflist is released here if needed.

But we need if (buflist) for drm_gem_unlock_reservations too. Fixed.

> > -
> > - list_del(&entry->list);
> > - free_vbuf(vgdev, entry);
> > }
> > wake_up(&vgdev->ctrlq.ack_queue);
> >
> > if (fence_id)
> > virtio_gpu_fence_event_process(vgdev, fence_id);
> > +
> > + list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
> > + if (entry->objs)
> > + virtio_gpu_array_put_free(entry->objs);
> > + list_del(&entry->list);
> We are clearing the list. I guess list_del is not needed.
> > + free_vbuf(vgdev, entry);

This just shuffles around the code. Dropping list_del() is unrelated
and should be a separate patch.

Beside that I'm not sure it actually can be dropped. free_vbuf() will
not kfree() the vbuf but keep it cached in a freelist instead.

cheers,
Gerd

2019-06-30 18:50:28

by Chia-I Wu

[permalink] [raw]
Subject: Re: [PATCH v4 08/12] drm/virtio: rework virtio_gpu_execbuffer_ioctl fencing

On Fri, Jun 28, 2019 at 3:34 AM Gerd Hoffmann <[email protected]> wrote:
>
> Hi,
>
> > > --- a/drivers/gpu/drm/virtio/virtgpu_drv.h
> > > +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
> > > @@ -120,9 +120,9 @@ struct virtio_gpu_vbuffer {
> > >
> > > char *resp_buf;
> > > int resp_size;
> > > -
> > > virtio_gpu_resp_cb resp_cb;
> > >
> > > + struct virtio_gpu_object_array *objs;
> > This can use a comment (e.g., objects referenced by the vbuffer)
>
> IMHO this is obvious ...
>
> > > void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
> > > void *data, uint32_t data_size,
> > > - uint32_t ctx_id, struct virtio_gpu_fence *fence);
> > > + uint32_t ctx_id, struct virtio_gpu_fence *fence,
> > > + struct virtio_gpu_object_array *objs);
> > Can we keep fence, which is updated, as the last parameter?
>
> Fixed.
>
> > > + if (buflist) {
> > > + for (i = 0; i < exbuf->num_bo_handles; i++)
> > > + reservation_object_add_excl_fence(buflist->objs[i]->resv,
> > > + &out_fence->f);
> > > + drm_gem_unlock_reservations(buflist->objs, buflist->nents,
> > > + &ticket);
> > > + }
> > We used to unlock after virtio_gpu_cmd_submit.
> >
> > I guess, the fence is considered signaled (because its seqno is still
> > 0) until after virtio_gpu_cmd_submit. We probably don't want other
> > processes to see the semi-initialized fence.
>
> Good point. Fixed.
>
> > > out_memdup:
> > > kfree(buf);
> > > out_unresv:
> > > - ttm_eu_backoff_reservation(&ticket, &validate_list);
> > > -out_free:
> > > - virtio_gpu_unref_list(&validate_list);
> > Keeping out_free to free buflist seems just fine.
>
> We don't need the separate label though ...
>
> > > + drm_gem_unlock_reservations(buflist->objs, buflist->nents, &ticket);
> > > out_unused_fd:
> > > kvfree(bo_handles);
> > > - kvfree(buflist);
> > > + if (buflist)
> > > + virtio_gpu_array_put_free(buflist);
>
> ... and the buflist is released here if needed.
>
> But we need if (buflist) for drm_gem_unlock_reservations too. Fixed.
>
> > > -
> > > - list_del(&entry->list);
> > > - free_vbuf(vgdev, entry);
> > > }
> > > wake_up(&vgdev->ctrlq.ack_queue);
> > >
> > > if (fence_id)
> > > virtio_gpu_fence_event_process(vgdev, fence_id);
> > > +
> > > + list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
> > > + if (entry->objs)
> > > + virtio_gpu_array_put_free(entry->objs);
> > > + list_del(&entry->list);
> > We are clearing the list. I guess list_del is not needed.
> > > + free_vbuf(vgdev, entry);
>
> This just shuffles around the code. Dropping list_del() is unrelated
> and should be a separate patch.
Fair point. We now loop the list twice and I was just looking for
chances for micro-optimizations.
>
> Beside that I'm not sure it actually can be dropped. free_vbuf() will
> not kfree() the vbuf but keep it cached in a freelist instead.
vbuf is created with kmem_cache_zalloc which always zeros the struct.

>
> cheers,
> Gerd
>