Message ID | 20241113084438.3283737-1-ryasuoka@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [v4] drm/virtio: Add drm_panic support | expand |
On 11/13/24 11:44, Ryosuke Yasuoka wrote: > From: Jocelyn Falempe <jfalempe@redhat.com> > > Virtio gpu supports the drm_panic module, which displays a message to > the screen when a kernel panic occurs. > > Signed-off-by: Ryosuke Yasuoka <ryasuoka@redhat.com> > Signed-off-by: Jocelyn Falempe <jfalempe@redhat.com> > --- I'll apply this patch tomorrow with a shortened virtio_panic_buffer variable name if nobody will have more comments to add.
On 11/13/24 11:44, Ryosuke Yasuoka wrote: > From: Jocelyn Falempe <jfalempe@redhat.com> > > Virtio gpu supports the drm_panic module, which displays a message to > the screen when a kernel panic occurs. > > Signed-off-by: Ryosuke Yasuoka <ryasuoka@redhat.com> > Signed-off-by: Jocelyn Falempe <jfalempe@redhat.com> > --- On a second look, I spotted few problems, see them below: ... > +/* For drm_panic */ > +static int virtio_gpu_panic_resource_flush(struct drm_plane *plane, > + struct virtio_gpu_vbuffer *vbuf, > + uint32_t x, uint32_t y, > + uint32_t width, uint32_t height) > +{ > + int ret; > + struct drm_device *dev = plane->dev; > + struct virtio_gpu_device *vgdev = dev->dev_private; > + struct virtio_gpu_framebuffer *vgfb; > + struct virtio_gpu_object *bo; > + > + vgfb = to_virtio_gpu_framebuffer(plane->state->fb); > + bo = gem_to_virtio_gpu_obj(vgfb->base.obj[0]); > + > + ret = virtio_gpu_panic_cmd_resource_flush(vgdev, vbuf, bo->hw_res_handle, x, y, > + width, height); > + return ret; Nit: return directly directly in such cases, dummy ret variable not needed > +} > + > static void virtio_gpu_resource_flush(struct drm_plane *plane, > uint32_t x, uint32_t y, > uint32_t width, uint32_t height) > @@ -359,11 +406,128 @@ static void virtio_gpu_cursor_plane_update(struct drm_plane *plane, > virtio_gpu_cursor_ping(vgdev, output); > } > > +static int virtio_drm_get_scanout_buffer(struct drm_plane *plane, > + struct drm_scanout_buffer *sb) > +{ > + struct virtio_gpu_object *bo; > + > + if (!plane->state || !plane->state->fb || !plane->state->visible) > + return -ENODEV; > + > + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); VRAM BOs aren't vmappable and should be rejected. In the virtio_panic_flush() below, virtio_gpu_panic_cmd_transfer_to_host_2d() is invoked only for dumb BOs. Thus, only dumb BO supports panic output and should be accepted by get_scanout_buffer(), other should be rejected with ENODEV here, AFAICT. Correct? > + /* try to vmap it if possible */ > + if (!bo->base.vaddr) { > + int ret; > + > + ret = drm_gem_shmem_vmap(&bo->base, &sb->map[0]); > + if (ret) > + return ret; > + } else { > + iosys_map_set_vaddr(&sb->map[0], bo->base.vaddr); > + } > + > + sb->format = plane->state->fb->format; > + sb->height = plane->state->fb->height; > + sb->width = plane->state->fb->width; > + sb->pitch[0] = plane->state->fb->pitches[0]; > + return 0; > +} > + > +struct virtio_gpu_panic_object_array { > + struct ww_acquire_ctx ticket; > + struct list_head next; > + u32 nents, total; > + struct drm_gem_object *objs; > +}; This virtio_gpu_panic_object_array struct is a hack, use virtio_gpu_array_alloc(). Maybe add atomic variant of the array_alloc(). > +static void virtio_gpu_panic_put_vbuf(struct virtqueue *vq, > + struct virtio_gpu_vbuffer *vbuf, > + struct virtio_gpu_object_array *objs) > +{ > + unsigned int len; > + int i; > + > + /* waiting vbuf to be used */ > + for (i = 0; i < 500; i++) { > + if (vbuf == virtqueue_get_buf(vq, &len)) { Is it guaranteed that virtio_gpu_dequeue_ctrl_func() never runs in parallel here? > + if (objs != NULL && vbuf->objs) > + drm_gem_object_put(objs->objs[0]); This drm_gem_object_put(objs->objs) is difficult to follow. Why vbuf->objs can't be used directly? Better to remove all error handlings for simplicity. It's not important if a bit of memory leaked on panic. > + break; > + } > + udelay(1); > + } > +} > + > +static void virtio_panic_flush(struct drm_plane *plane) > +{ > + int ret; > + struct virtio_gpu_object *bo; > + struct drm_device *dev = plane->dev; > + struct virtio_gpu_device *vgdev = dev->dev_private; > + struct drm_rect rect; > + void *vp_buf = vgdev->virtio_panic_buffer; > + struct virtio_gpu_vbuffer *vbuf_dumb_bo = vp_buf; > + struct virtio_gpu_vbuffer *vbuf_resource_flush = vp_buf + VBUFFER_SIZE; > + > + rect.x1 = 0; > + rect.y1 = 0; > + rect.x2 = plane->state->fb->width; > + rect.y2 = plane->state->fb->height; > + > + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); > + > + struct drm_gem_object obj; > + struct virtio_gpu_panic_object_array objs = { > + .next = { NULL, NULL }, > + .nents = 0, > + .total = 1, > + .objs = &obj > + }; This &obj is unitialized stack data. The .objs points to an array of obj pointers and you pointing it to object. Like I suggested above, let's remove this hack and use proper virtio_gpu_array_alloc(). > + if (bo->dumb) { > + ret = virtio_gpu_panic_update_dumb_bo(vgdev, > + plane->state, > + &rect, > + (struct virtio_gpu_object_array *)&objs, > + vbuf_dumb_bo); > + if (ret) { > + if (vbuf_dumb_bo->objs) > + drm_gem_object_put(&objs.objs[0]); > + return; > + } > + } > + > + ret = virtio_gpu_panic_resource_flush(plane, vbuf_resource_flush, > + plane->state->src_x >> 16, > + plane->state->src_y >> 16, > + plane->state->src_w >> 16, > + plane->state->src_h >> 16); > + if (ret) { > + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, > + vbuf_dumb_bo, > + (struct virtio_gpu_object_array *)&objs); The virtio_gpu_panic_notify() hasn't been invoked here, thus this put_vbuf should always time out because vq hasn't been kicked. Again, best to leak resources on error than to have broken/untested error handling code paths. > + return; > + } > + > + virtio_gpu_panic_notify(vgdev); > + > + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, > + vbuf_dumb_bo, > + (struct virtio_gpu_object_array *)&objs); > + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, > + vbuf_resource_flush, > + NULL); > +} > + > static const struct drm_plane_helper_funcs virtio_gpu_primary_helper_funcs = { > .prepare_fb = virtio_gpu_plane_prepare_fb, > .cleanup_fb = virtio_gpu_plane_cleanup_fb, > .atomic_check = virtio_gpu_plane_atomic_check, > .atomic_update = virtio_gpu_primary_plane_update, > + .get_scanout_buffer = virtio_drm_get_scanout_buffer, > + .panic_flush = virtio_panic_flush, > }; > > static const struct drm_plane_helper_funcs virtio_gpu_cursor_helper_funcs = { > @@ -383,6 +547,13 @@ struct drm_plane *virtio_gpu_plane_init(struct virtio_gpu_device *vgdev, > const uint32_t *formats; > int nformats; > > + /* allocate panic buffers */ > + if (index == 0 && type == DRM_PLANE_TYPE_PRIMARY) { > + vgdev->virtio_panic_buffer = drmm_kzalloc(dev, 2 * VBUFFER_SIZE, GFP_KERNEL); > + if (!vgdev->virtio_panic_buffer) > + return ERR_PTR(-ENOMEM); > + } If there is more than one virtio-gpu display, then this panic buffer is reused by other displays. It seems to work okay, but potential implications are unclear. Won't it be more robust to have a panic buffer per CRTC? Also, please rename vgdev->virtio_panic_buffer to vgdev->panic_vbuf for brevity.
On 18/11/2024 17:16, Dmitry Osipenko wrote: > On 11/13/24 11:44, Ryosuke Yasuoka wrote: >> From: Jocelyn Falempe <jfalempe@redhat.com> >> >> Virtio gpu supports the drm_panic module, which displays a message to >> the screen when a kernel panic occurs. >> >> Signed-off-by: Ryosuke Yasuoka <ryasuoka@redhat.com> >> Signed-off-by: Jocelyn Falempe <jfalempe@redhat.com> >> --- > > On a second look, I spotted few problems, see them below: > > ... >> +/* For drm_panic */ >> +static int virtio_gpu_panic_resource_flush(struct drm_plane *plane, >> + struct virtio_gpu_vbuffer *vbuf, >> + uint32_t x, uint32_t y, >> + uint32_t width, uint32_t height) >> +{ >> + int ret; >> + struct drm_device *dev = plane->dev; >> + struct virtio_gpu_device *vgdev = dev->dev_private; >> + struct virtio_gpu_framebuffer *vgfb; >> + struct virtio_gpu_object *bo; >> + >> + vgfb = to_virtio_gpu_framebuffer(plane->state->fb); >> + bo = gem_to_virtio_gpu_obj(vgfb->base.obj[0]); >> + >> + ret = virtio_gpu_panic_cmd_resource_flush(vgdev, vbuf, bo->hw_res_handle, x, y, >> + width, height); >> + return ret; > > Nit: return directly directly in such cases, dummy ret variable not needed > >> +} >> + >> static void virtio_gpu_resource_flush(struct drm_plane *plane, >> uint32_t x, uint32_t y, >> uint32_t width, uint32_t height) >> @@ -359,11 +406,128 @@ static void virtio_gpu_cursor_plane_update(struct drm_plane *plane, >> virtio_gpu_cursor_ping(vgdev, output); >> } >> >> +static int virtio_drm_get_scanout_buffer(struct drm_plane *plane, >> + struct drm_scanout_buffer *sb) >> +{ >> + struct virtio_gpu_object *bo; >> + >> + if (!plane->state || !plane->state->fb || !plane->state->visible) >> + return -ENODEV; >> + >> + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); > > VRAM BOs aren't vmappable and should be rejected. > > In the virtio_panic_flush() below, > virtio_gpu_panic_cmd_transfer_to_host_2d() is invoked only for dumb BOs. > Thus, only dumb BO supports panic output and should be accepted by > get_scanout_buffer(), other should be rejected with ENODEV here, AFAICT. > Correct? Yes, it's correct > >> + /* try to vmap it if possible */ >> + if (!bo->base.vaddr) { >> + int ret; >> + >> + ret = drm_gem_shmem_vmap(&bo->base, &sb->map[0]); >> + if (ret) >> + return ret; >> + } else { >> + iosys_map_set_vaddr(&sb->map[0], bo->base.vaddr); >> + } >> + >> + sb->format = plane->state->fb->format; >> + sb->height = plane->state->fb->height; >> + sb->width = plane->state->fb->width; >> + sb->pitch[0] = plane->state->fb->pitches[0]; >> + return 0; >> +} >> + >> +struct virtio_gpu_panic_object_array { >> + struct ww_acquire_ctx ticket; >> + struct list_head next; >> + u32 nents, total; >> + struct drm_gem_object *objs; >> +}; > > This virtio_gpu_panic_object_array struct is a hack, use > virtio_gpu_array_alloc(). Maybe add atomic variant of the array_alloc(). We can't allocate memory in the panic handler, but maybe it can be pre-allocated, like the virtio_panic_buffer ? > >> +static void virtio_gpu_panic_put_vbuf(struct virtqueue *vq, >> + struct virtio_gpu_vbuffer *vbuf, >> + struct virtio_gpu_object_array *objs) >> +{ >> + unsigned int len; >> + int i; >> + >> + /* waiting vbuf to be used */ >> + for (i = 0; i < 500; i++) { >> + if (vbuf == virtqueue_get_buf(vq, &len)) { > > Is it guaranteed that virtio_gpu_dequeue_ctrl_func() never runs in > parallel here? Yes, in the panic handler, only one CPU remains active, and no other task can be scheduled. > >> + if (objs != NULL && vbuf->objs) >> + drm_gem_object_put(objs->objs[0]); > > This drm_gem_object_put(objs->objs) is difficult to follow. Why > vbuf->objs can't be used directly? > > Better to remove all error handlings for simplicity. It's not important > if a bit of memory leaked on panic. We try to reclaim, to make it easier to test with the debugfs interface. But this is a bit racy anyway, because it's not the real panic context. > >> + break; >> + } >> + udelay(1); >> + } >> +} >> + >> +static void virtio_panic_flush(struct drm_plane *plane) >> +{ >> + int ret; >> + struct virtio_gpu_object *bo; >> + struct drm_device *dev = plane->dev; >> + struct virtio_gpu_device *vgdev = dev->dev_private; >> + struct drm_rect rect; >> + void *vp_buf = vgdev->virtio_panic_buffer; >> + struct virtio_gpu_vbuffer *vbuf_dumb_bo = vp_buf; >> + struct virtio_gpu_vbuffer *vbuf_resource_flush = vp_buf + VBUFFER_SIZE; >> + >> + rect.x1 = 0; >> + rect.y1 = 0; >> + rect.x2 = plane->state->fb->width; >> + rect.y2 = plane->state->fb->height; >> + >> + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); >> + >> + struct drm_gem_object obj; >> + struct virtio_gpu_panic_object_array objs = { >> + .next = { NULL, NULL }, >> + .nents = 0, >> + .total = 1, >> + .objs = &obj >> + }; > > This &obj is unitialized stack data. The .objs points to an array of obj > pointers and you pointing it to object. Like I suggested above, let's > remove this hack and use proper virtio_gpu_array_alloc(). > >> + if (bo->dumb) { >> + ret = virtio_gpu_panic_update_dumb_bo(vgdev, >> + plane->state, >> + &rect, >> + (struct virtio_gpu_object_array *)&objs, >> + vbuf_dumb_bo); >> + if (ret) { >> + if (vbuf_dumb_bo->objs) >> + drm_gem_object_put(&objs.objs[0]); >> + return; >> + } >> + } >> + >> + ret = virtio_gpu_panic_resource_flush(plane, vbuf_resource_flush, >> + plane->state->src_x >> 16, >> + plane->state->src_y >> 16, >> + plane->state->src_w >> 16, >> + plane->state->src_h >> 16); >> + if (ret) { >> + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, >> + vbuf_dumb_bo, >> + (struct virtio_gpu_object_array *)&objs); > > The virtio_gpu_panic_notify() hasn't been invoked here, thus this > put_vbuf should always time out because vq hasn't been kicked. Again, > best to leak resources on error than to have broken/untested error > handling code paths. agreed > >> + return; >> + } >> + >> + virtio_gpu_panic_notify(vgdev); >> + >> + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, >> + vbuf_dumb_bo, >> + (struct virtio_gpu_object_array *)&objs); >> + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, >> + vbuf_resource_flush, >> + NULL); >> +} >> + >> static const struct drm_plane_helper_funcs virtio_gpu_primary_helper_funcs = { >> .prepare_fb = virtio_gpu_plane_prepare_fb, >> .cleanup_fb = virtio_gpu_plane_cleanup_fb, >> .atomic_check = virtio_gpu_plane_atomic_check, >> .atomic_update = virtio_gpu_primary_plane_update, >> + .get_scanout_buffer = virtio_drm_get_scanout_buffer, >> + .panic_flush = virtio_panic_flush, >> }; >> >> static const struct drm_plane_helper_funcs virtio_gpu_cursor_helper_funcs = { >> @@ -383,6 +547,13 @@ struct drm_plane *virtio_gpu_plane_init(struct virtio_gpu_device *vgdev, >> const uint32_t *formats; >> int nformats; >> >> + /* allocate panic buffers */ >> + if (index == 0 && type == DRM_PLANE_TYPE_PRIMARY) { >> + vgdev->virtio_panic_buffer = drmm_kzalloc(dev, 2 * VBUFFER_SIZE, GFP_KERNEL); >> + if (!vgdev->virtio_panic_buffer) >> + return ERR_PTR(-ENOMEM); >> + } > > If there is more than one virtio-gpu display, then this panic buffer is > reused by other displays. It seems to work okay, but potential > implications are unclear. Won't it be more robust to have a panic buffer > per CRTC? The drm panic handler call each plane sequentially, so it's not a problem. Having one buffer per CRTC would just add more complexity. > > Also, please rename vgdev->virtio_panic_buffer to vgdev->panic_vbuf for > brevity. > Thanks for the review.
diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h index 64c236169db8..3482f4e1057c 100644 --- a/drivers/gpu/drm/virtio/virtgpu_drv.h +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h @@ -125,6 +125,12 @@ struct virtio_gpu_object_array { struct drm_gem_object *objs[] __counted_by(total); }; +#define MAX_INLINE_CMD_SIZE 96 +#define MAX_INLINE_RESP_SIZE 24 +#define VBUFFER_SIZE (sizeof(struct virtio_gpu_vbuffer) \ + + MAX_INLINE_CMD_SIZE \ + + MAX_INLINE_RESP_SIZE) + struct virtio_gpu_vbuffer; struct virtio_gpu_device; @@ -267,6 +273,7 @@ struct virtio_gpu_device { spinlock_t resource_export_lock; /* protects map state and host_visible_mm */ spinlock_t host_visible_lock; + void *virtio_panic_buffer; }; struct virtio_gpu_fpriv { @@ -329,12 +336,23 @@ void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev, struct virtio_gpu_fence *fence); void virtio_gpu_cmd_unref_resource(struct virtio_gpu_device *vgdev, struct virtio_gpu_object *bo); +int virtio_gpu_panic_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev, + uint64_t offset, + uint32_t width, uint32_t height, + uint32_t x, uint32_t y, + struct virtio_gpu_object_array *objs, + struct virtio_gpu_vbuffer *vbuf); void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev, uint64_t offset, uint32_t width, uint32_t height, uint32_t x, uint32_t y, struct virtio_gpu_object_array *objs, struct virtio_gpu_fence *fence); +int virtio_gpu_panic_cmd_resource_flush(struct virtio_gpu_device *vgdev, + struct virtio_gpu_vbuffer *vbuf, + uint32_t resource_id, + uint32_t x, uint32_t y, + uint32_t width, uint32_t height); void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev, uint32_t resource_id, uint32_t x, uint32_t y, @@ -399,6 +417,7 @@ void virtio_gpu_ctrl_ack(struct virtqueue *vq); void virtio_gpu_cursor_ack(struct virtqueue *vq); void virtio_gpu_dequeue_ctrl_func(struct work_struct *work); void virtio_gpu_dequeue_cursor_func(struct work_struct *work); +void virtio_gpu_panic_notify(struct virtio_gpu_device *vgdev); void virtio_gpu_notify(struct virtio_gpu_device *vgdev); int diff --git a/drivers/gpu/drm/virtio/virtgpu_plane.c b/drivers/gpu/drm/virtio/virtgpu_plane.c index a72a2dbda031..a6f99886fd97 100644 --- a/drivers/gpu/drm/virtio/virtgpu_plane.c +++ b/drivers/gpu/drm/virtio/virtgpu_plane.c @@ -26,6 +26,9 @@ #include <drm/drm_atomic_helper.h> #include <drm/drm_damage_helper.h> #include <drm/drm_fourcc.h> +#include <drm/drm_managed.h> +#include <drm/drm_panic.h> +#include <linux/delay.h> #include "virtgpu_drv.h" @@ -108,6 +111,30 @@ static int virtio_gpu_plane_atomic_check(struct drm_plane *plane, return ret; } +/* For drm panic */ +static int virtio_gpu_panic_update_dumb_bo(struct virtio_gpu_device *vgdev, + struct drm_plane_state *state, + struct drm_rect *rect, + struct virtio_gpu_object_array *objs, + struct virtio_gpu_vbuffer *vbuf) +{ + int ret; + struct virtio_gpu_object *bo = + gem_to_virtio_gpu_obj(state->fb->obj[0]); + uint32_t w = rect->x2 - rect->x1; + uint32_t h = rect->y2 - rect->y1; + uint32_t x = rect->x1; + uint32_t y = rect->y1; + uint32_t off = x * state->fb->format->cpp[0] + + y * state->fb->pitches[0]; + + virtio_gpu_array_add_obj(objs, &bo->base.base); + + ret = virtio_gpu_panic_cmd_transfer_to_host_2d(vgdev, off, w, h, x, y, + objs, vbuf); + return ret; +} + static void virtio_gpu_update_dumb_bo(struct virtio_gpu_device *vgdev, struct drm_plane_state *state, struct drm_rect *rect) @@ -131,6 +158,26 @@ static void virtio_gpu_update_dumb_bo(struct virtio_gpu_device *vgdev, objs, NULL); } +/* For drm_panic */ +static int virtio_gpu_panic_resource_flush(struct drm_plane *plane, + struct virtio_gpu_vbuffer *vbuf, + uint32_t x, uint32_t y, + uint32_t width, uint32_t height) +{ + int ret; + struct drm_device *dev = plane->dev; + struct virtio_gpu_device *vgdev = dev->dev_private; + struct virtio_gpu_framebuffer *vgfb; + struct virtio_gpu_object *bo; + + vgfb = to_virtio_gpu_framebuffer(plane->state->fb); + bo = gem_to_virtio_gpu_obj(vgfb->base.obj[0]); + + ret = virtio_gpu_panic_cmd_resource_flush(vgdev, vbuf, bo->hw_res_handle, x, y, + width, height); + return ret; +} + static void virtio_gpu_resource_flush(struct drm_plane *plane, uint32_t x, uint32_t y, uint32_t width, uint32_t height) @@ -359,11 +406,128 @@ static void virtio_gpu_cursor_plane_update(struct drm_plane *plane, virtio_gpu_cursor_ping(vgdev, output); } +static int virtio_drm_get_scanout_buffer(struct drm_plane *plane, + struct drm_scanout_buffer *sb) +{ + struct virtio_gpu_object *bo; + + if (!plane->state || !plane->state->fb || !plane->state->visible) + return -ENODEV; + + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); + + /* try to vmap it if possible */ + if (!bo->base.vaddr) { + int ret; + + ret = drm_gem_shmem_vmap(&bo->base, &sb->map[0]); + if (ret) + return ret; + } else { + iosys_map_set_vaddr(&sb->map[0], bo->base.vaddr); + } + + sb->format = plane->state->fb->format; + sb->height = plane->state->fb->height; + sb->width = plane->state->fb->width; + sb->pitch[0] = plane->state->fb->pitches[0]; + return 0; +} + +struct virtio_gpu_panic_object_array { + struct ww_acquire_ctx ticket; + struct list_head next; + u32 nents, total; + struct drm_gem_object *objs; +}; + + +static void virtio_gpu_panic_put_vbuf(struct virtqueue *vq, + struct virtio_gpu_vbuffer *vbuf, + struct virtio_gpu_object_array *objs) +{ + unsigned int len; + int i; + + /* waiting vbuf to be used */ + for (i = 0; i < 500; i++) { + if (vbuf == virtqueue_get_buf(vq, &len)) { + if (objs != NULL && vbuf->objs) + drm_gem_object_put(objs->objs[0]); + break; + } + udelay(1); + } +} + +static void virtio_panic_flush(struct drm_plane *plane) +{ + int ret; + struct virtio_gpu_object *bo; + struct drm_device *dev = plane->dev; + struct virtio_gpu_device *vgdev = dev->dev_private; + struct drm_rect rect; + void *vp_buf = vgdev->virtio_panic_buffer; + struct virtio_gpu_vbuffer *vbuf_dumb_bo = vp_buf; + struct virtio_gpu_vbuffer *vbuf_resource_flush = vp_buf + VBUFFER_SIZE; + + rect.x1 = 0; + rect.y1 = 0; + rect.x2 = plane->state->fb->width; + rect.y2 = plane->state->fb->height; + + bo = gem_to_virtio_gpu_obj(plane->state->fb->obj[0]); + + struct drm_gem_object obj; + struct virtio_gpu_panic_object_array objs = { + .next = { NULL, NULL }, + .nents = 0, + .total = 1, + .objs = &obj + }; + + if (bo->dumb) { + ret = virtio_gpu_panic_update_dumb_bo(vgdev, + plane->state, + &rect, + (struct virtio_gpu_object_array *)&objs, + vbuf_dumb_bo); + if (ret) { + if (vbuf_dumb_bo->objs) + drm_gem_object_put(&objs.objs[0]); + return; + } + } + + ret = virtio_gpu_panic_resource_flush(plane, vbuf_resource_flush, + plane->state->src_x >> 16, + plane->state->src_y >> 16, + plane->state->src_w >> 16, + plane->state->src_h >> 16); + if (ret) { + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, + vbuf_dumb_bo, + (struct virtio_gpu_object_array *)&objs); + return; + } + + virtio_gpu_panic_notify(vgdev); + + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, + vbuf_dumb_bo, + (struct virtio_gpu_object_array *)&objs); + virtio_gpu_panic_put_vbuf(vgdev->ctrlq.vq, + vbuf_resource_flush, + NULL); +} + static const struct drm_plane_helper_funcs virtio_gpu_primary_helper_funcs = { .prepare_fb = virtio_gpu_plane_prepare_fb, .cleanup_fb = virtio_gpu_plane_cleanup_fb, .atomic_check = virtio_gpu_plane_atomic_check, .atomic_update = virtio_gpu_primary_plane_update, + .get_scanout_buffer = virtio_drm_get_scanout_buffer, + .panic_flush = virtio_panic_flush, }; static const struct drm_plane_helper_funcs virtio_gpu_cursor_helper_funcs = { @@ -383,6 +547,13 @@ struct drm_plane *virtio_gpu_plane_init(struct virtio_gpu_device *vgdev, const uint32_t *formats; int nformats; + /* allocate panic buffers */ + if (index == 0 && type == DRM_PLANE_TYPE_PRIMARY) { + vgdev->virtio_panic_buffer = drmm_kzalloc(dev, 2 * VBUFFER_SIZE, GFP_KERNEL); + if (!vgdev->virtio_panic_buffer) + return ERR_PTR(-ENOMEM); + } + if (type == DRM_PLANE_TYPE_CURSOR) { formats = virtio_gpu_cursor_formats; nformats = ARRAY_SIZE(virtio_gpu_cursor_formats); diff --git a/drivers/gpu/drm/virtio/virtgpu_vq.c b/drivers/gpu/drm/virtio/virtgpu_vq.c index 0d3d0d09f39b..f6e1655458dd 100644 --- a/drivers/gpu/drm/virtio/virtgpu_vq.c +++ b/drivers/gpu/drm/virtio/virtgpu_vq.c @@ -36,12 +36,6 @@ #include "virtgpu_drv.h" #include "virtgpu_trace.h" -#define MAX_INLINE_CMD_SIZE 96 -#define MAX_INLINE_RESP_SIZE 24 -#define VBUFFER_SIZE (sizeof(struct virtio_gpu_vbuffer) \ - + MAX_INLINE_CMD_SIZE \ - + MAX_INLINE_RESP_SIZE) - static void convert_to_hw_box(struct virtio_gpu_box *dst, const struct drm_virtgpu_3d_box *src) { @@ -311,6 +305,34 @@ static struct sg_table *vmalloc_to_sgt(char *data, uint32_t size, int *sg_ents) return sgt; } +/* For drm_panic */ +static int virtio_gpu_panic_queue_ctrl_sgs(struct virtio_gpu_device *vgdev, + struct virtio_gpu_vbuffer *vbuf, + int elemcnt, + struct scatterlist **sgs, + int outcnt, + int incnt) +{ + struct virtqueue *vq = vgdev->ctrlq.vq; + int ret; + + if (vgdev->has_indirect) + elemcnt = 1; + + if (vq->num_free < elemcnt) + return -ENOMEM; + + ret = virtqueue_add_sgs(vq, sgs, outcnt, incnt, vbuf, GFP_ATOMIC); + WARN_ON(ret); + + vbuf->seqno = ++vgdev->ctrlq.seqno; + trace_virtio_gpu_cmd_queue(vq, virtio_gpu_vbuf_ctrl_hdr(vbuf), vbuf->seqno); + + atomic_inc(&vgdev->pending_commands); + + return 0; +} + static int virtio_gpu_queue_ctrl_sgs(struct virtio_gpu_device *vgdev, struct virtio_gpu_vbuffer *vbuf, struct virtio_gpu_fence *fence, @@ -368,6 +390,33 @@ static int virtio_gpu_queue_ctrl_sgs(struct virtio_gpu_device *vgdev, return 0; } +/* For drm_panic */ +static int virtio_gpu_panic_queue_ctrl_buffer(struct virtio_gpu_device *vgdev, + struct virtio_gpu_vbuffer *vbuf) +{ + struct scatterlist *sgs[3], vcmd, vresp; + int elemcnt = 0, outcnt = 0, incnt = 0, ret; + + /* set up vcmd */ + sg_init_one(&vcmd, vbuf->buf, vbuf->size); + elemcnt++; + sgs[outcnt] = &vcmd; + outcnt++; + + /* set up vresp */ + if (vbuf->resp_size) { + sg_init_one(&vresp, vbuf->resp_buf, vbuf->resp_size); + elemcnt++; + sgs[outcnt + incnt] = &vresp; + incnt++; + } + + ret = virtio_gpu_panic_queue_ctrl_sgs(vgdev, vbuf, + elemcnt, sgs, + outcnt, incnt); + return ret; +} + static int virtio_gpu_queue_fenced_ctrl_buffer(struct virtio_gpu_device *vgdev, struct virtio_gpu_vbuffer *vbuf, struct virtio_gpu_fence *fence) @@ -422,6 +471,21 @@ static int virtio_gpu_queue_fenced_ctrl_buffer(struct virtio_gpu_device *vgdev, return ret; } +/* For drm_panic */ +void virtio_gpu_panic_notify(struct virtio_gpu_device *vgdev) +{ + bool notify; + + if (!atomic_read(&vgdev->pending_commands)) + return; + + atomic_set(&vgdev->pending_commands, 0); + notify = virtqueue_kick_prepare(vgdev->ctrlq.vq); + + if (notify) + virtqueue_notify(vgdev->ctrlq.vq); +} + void virtio_gpu_notify(struct virtio_gpu_device *vgdev) { bool notify; @@ -567,6 +631,44 @@ void virtio_gpu_cmd_set_scanout(struct virtio_gpu_device *vgdev, virtio_gpu_queue_ctrl_buffer(vgdev, vbuf); } +/* For drm_panic */ +static void virtio_gpu_panic_init_cmd(struct virtio_gpu_device *vgdev, + struct virtio_gpu_vbuffer *vbuf, + int cmd_size) +{ + vbuf->buf = (void *)vbuf + sizeof(*vbuf); + vbuf->size = cmd_size; + vbuf->resp_cb = NULL; + vbuf->resp_size = sizeof(struct virtio_gpu_ctrl_hdr); + vbuf->resp_buf = (void *)vbuf->buf + cmd_size; +} + +/* For drm_panic */ +int virtio_gpu_panic_cmd_resource_flush(struct virtio_gpu_device *vgdev, + struct virtio_gpu_vbuffer *vbuf, + uint32_t resource_id, + uint32_t x, uint32_t y, + uint32_t width, uint32_t height) +{ + int ret; + struct virtio_gpu_resource_flush *cmd_p; + + virtio_gpu_panic_init_cmd(vgdev, vbuf, + sizeof(struct virtio_gpu_resource_flush)); + cmd_p = (void *)vbuf->buf; + vbuf->objs = NULL; + + cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_RESOURCE_FLUSH); + cmd_p->resource_id = cpu_to_le32(resource_id); + cmd_p->r.width = cpu_to_le32(width); + cmd_p->r.height = cpu_to_le32(height); + cmd_p->r.x = cpu_to_le32(x); + cmd_p->r.y = cpu_to_le32(y); + + ret = virtio_gpu_panic_queue_ctrl_buffer(vgdev, vbuf); + return ret; +} + void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev, uint32_t resource_id, uint32_t x, uint32_t y, @@ -591,6 +693,40 @@ void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev, virtio_gpu_queue_fenced_ctrl_buffer(vgdev, vbuf, fence); } +/* For drm_panic */ +int virtio_gpu_panic_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev, + uint64_t offset, + uint32_t width, uint32_t height, + uint32_t x, uint32_t y, + struct virtio_gpu_object_array *objs, + struct virtio_gpu_vbuffer *vbuf) +{ + int ret; + struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(objs->objs[0]); + struct virtio_gpu_transfer_to_host_2d *cmd_p; + bool use_dma_api = !virtio_has_dma_quirk(vgdev->vdev); + + if (virtio_gpu_is_shmem(bo) && use_dma_api) + dma_sync_sgtable_for_device(vgdev->vdev->dev.parent, + bo->base.sgt, DMA_TO_DEVICE); + + virtio_gpu_panic_init_cmd(vgdev, vbuf, + sizeof(struct virtio_gpu_transfer_to_host_2d)); + cmd_p = (void *)vbuf->buf; + vbuf->objs = objs; + + cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D); + cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle); + cmd_p->offset = cpu_to_le64(offset); + cmd_p->r.width = cpu_to_le32(width); + cmd_p->r.height = cpu_to_le32(height); + cmd_p->r.x = cpu_to_le32(x); + cmd_p->r.y = cpu_to_le32(y); + + ret = virtio_gpu_panic_queue_ctrl_buffer(vgdev, vbuf); + return ret; +} + void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev, uint64_t offset, uint32_t width, uint32_t height,