Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
  1/*
  2 * Copyright (C) 2015 Red Hat, Inc.
  3 * All Rights Reserved.
  4 *
  5 * Permission is hereby granted, free of charge, to any person obtaining
  6 * a copy of this software and associated documentation files (the
  7 * "Software"), to deal in the Software without restriction, including
  8 * without limitation the rights to use, copy, modify, merge, publish,
  9 * distribute, sublicense, and/or sell copies of the Software, and to
 10 * permit persons to whom the Software is furnished to do so, subject to
 11 * the following conditions:
 12 *
 13 * The above copyright notice and this permission notice (including the
 14 * next paragraph) shall be included in all copies or substantial
 15 * portions of the Software.
 16 *
 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 24 */
 25
 26#include <drm/drmP.h>
 27#include <drm/drm_fb_helper.h>
 28#include "virtgpu_drv.h"
 29
 30#define VIRTIO_GPU_FBCON_POLL_PERIOD (HZ / 60)
 31
 32struct virtio_gpu_fbdev {
 33	struct drm_fb_helper           helper;
 34	struct virtio_gpu_framebuffer  vgfb;
 35	struct virtio_gpu_device       *vgdev;
 36	struct delayed_work            work;
 37};
 38
 39static int virtio_gpu_dirty_update(struct virtio_gpu_framebuffer *fb,
 40				   bool store, int x, int y,
 41				   int width, int height)
 42{
 43	struct drm_device *dev = fb->base.dev;
 44	struct virtio_gpu_device *vgdev = dev->dev_private;
 45	bool store_for_later = false;
 46	int bpp = fb->base.bits_per_pixel / 8;
 47	int x2, y2;
 48	unsigned long flags;
 49	struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(fb->obj);
 50
 51	if ((width <= 0) ||
 52	    (x + width > fb->base.width) ||
 53	    (y + height > fb->base.height)) {
 54		DRM_DEBUG("values out of range %dx%d+%d+%d, fb %dx%d\n",
 55			  width, height, x, y,
 56			  fb->base.width, fb->base.height);
 57		return -EINVAL;
 58	}
 59
 60	/*
 61	 * Can be called with pretty much any context (console output
 62	 * path).  If we are in atomic just store the dirty rect info
 63	 * to send out the update later.
 64	 *
 65	 * Can't test inside spin lock.
 66	 */
 67	if (in_atomic() || store)
 68		store_for_later = true;
 69
 70	x2 = x + width - 1;
 71	y2 = y + height - 1;
 72
 73	spin_lock_irqsave(&fb->dirty_lock, flags);
 74
 75	if (fb->y1 < y)
 76		y = fb->y1;
 77	if (fb->y2 > y2)
 78		y2 = fb->y2;
 79	if (fb->x1 < x)
 80		x = fb->x1;
 81	if (fb->x2 > x2)
 82		x2 = fb->x2;
 83
 84	if (store_for_later) {
 85		fb->x1 = x;
 86		fb->x2 = x2;
 87		fb->y1 = y;
 88		fb->y2 = y2;
 89		spin_unlock_irqrestore(&fb->dirty_lock, flags);
 90		return 0;
 91	}
 92
 93	fb->x1 = fb->y1 = INT_MAX;
 94	fb->x2 = fb->y2 = 0;
 95
 96	spin_unlock_irqrestore(&fb->dirty_lock, flags);
 97
 98	{
 99		uint32_t offset;
100		uint32_t w = x2 - x + 1;
101		uint32_t h = y2 - y + 1;
102
103		offset = (y * fb->base.pitches[0]) + x * bpp;
104
105		virtio_gpu_cmd_transfer_to_host_2d(vgdev, obj->hw_res_handle,
106						   offset,
107						   cpu_to_le32(w),
108						   cpu_to_le32(h),
109						   cpu_to_le32(x),
110						   cpu_to_le32(y),
111						   NULL);
112
113	}
114	virtio_gpu_cmd_resource_flush(vgdev, obj->hw_res_handle,
115				      x, y, x2 - x + 1, y2 - y + 1);
116	return 0;
117}
118
119int virtio_gpu_surface_dirty(struct virtio_gpu_framebuffer *vgfb,
120			     struct drm_clip_rect *clips,
121			     unsigned num_clips)
122{
123	struct virtio_gpu_device *vgdev = vgfb->base.dev->dev_private;
124	struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(vgfb->obj);
125	struct drm_clip_rect norect;
126	struct drm_clip_rect *clips_ptr;
127	int left, right, top, bottom;
128	int i;
129	int inc = 1;
130	if (!num_clips) {
131		num_clips = 1;
132		clips = &norect;
133		norect.x1 = norect.y1 = 0;
134		norect.x2 = vgfb->base.width;
135		norect.y2 = vgfb->base.height;
136	}
137	left = clips->x1;
138	right = clips->x2;
139	top = clips->y1;
140	bottom = clips->y2;
141
142	/* skip the first clip rect */
143	for (i = 1, clips_ptr = clips + inc;
144	     i < num_clips; i++, clips_ptr += inc) {
145		left = min_t(int, left, (int)clips_ptr->x1);
146		right = max_t(int, right, (int)clips_ptr->x2);
147		top = min_t(int, top, (int)clips_ptr->y1);
148		bottom = max_t(int, bottom, (int)clips_ptr->y2);
149	}
150
151	if (obj->dumb)
152		return virtio_gpu_dirty_update(vgfb, false, left, top,
153					       right - left, bottom - top);
154
155	virtio_gpu_cmd_resource_flush(vgdev, obj->hw_res_handle,
156				      left, top, right - left, bottom - top);
157	return 0;
158}
159
160static void virtio_gpu_fb_dirty_work(struct work_struct *work)
161{
162	struct delayed_work *delayed_work = to_delayed_work(work);
163	struct virtio_gpu_fbdev *vfbdev =
164		container_of(delayed_work, struct virtio_gpu_fbdev, work);
165	struct virtio_gpu_framebuffer *vgfb = &vfbdev->vgfb;
166
167	virtio_gpu_dirty_update(&vfbdev->vgfb, false, vgfb->x1, vgfb->y1,
168				vgfb->x2 - vgfb->x1, vgfb->y2 - vgfb->y1);
169}
170
171static void virtio_gpu_3d_fillrect(struct fb_info *info,
172				   const struct fb_fillrect *rect)
173{
174	struct virtio_gpu_fbdev *vfbdev = info->par;
175	drm_fb_helper_sys_fillrect(info, rect);
176	virtio_gpu_dirty_update(&vfbdev->vgfb, true, rect->dx, rect->dy,
177			     rect->width, rect->height);
178	schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD);
179}
180
181static void virtio_gpu_3d_copyarea(struct fb_info *info,
182				   const struct fb_copyarea *area)
183{
184	struct virtio_gpu_fbdev *vfbdev = info->par;
185	drm_fb_helper_sys_copyarea(info, area);
186	virtio_gpu_dirty_update(&vfbdev->vgfb, true, area->dx, area->dy,
187			   area->width, area->height);
188	schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD);
189}
190
191static void virtio_gpu_3d_imageblit(struct fb_info *info,
192				    const struct fb_image *image)
193{
194	struct virtio_gpu_fbdev *vfbdev = info->par;
195	drm_fb_helper_sys_imageblit(info, image);
196	virtio_gpu_dirty_update(&vfbdev->vgfb, true, image->dx, image->dy,
197			     image->width, image->height);
198	schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD);
199}
200
201static struct fb_ops virtio_gpufb_ops = {
202	.owner = THIS_MODULE,
203	.fb_check_var = drm_fb_helper_check_var,
204	.fb_set_par = drm_fb_helper_set_par, /* TODO: copy vmwgfx */
205	.fb_fillrect = virtio_gpu_3d_fillrect,
206	.fb_copyarea = virtio_gpu_3d_copyarea,
207	.fb_imageblit = virtio_gpu_3d_imageblit,
208	.fb_pan_display = drm_fb_helper_pan_display,
209	.fb_blank = drm_fb_helper_blank,
210	.fb_setcmap = drm_fb_helper_setcmap,
211	.fb_debug_enter = drm_fb_helper_debug_enter,
212	.fb_debug_leave = drm_fb_helper_debug_leave,
213};
214
215static int virtio_gpu_vmap_fb(struct virtio_gpu_device *vgdev,
216			      struct virtio_gpu_object *obj)
217{
218	return virtio_gpu_object_kmap(obj, NULL);
219}
220
221static int virtio_gpufb_create(struct drm_fb_helper *helper,
222			       struct drm_fb_helper_surface_size *sizes)
223{
224	struct virtio_gpu_fbdev *vfbdev =
225		container_of(helper, struct virtio_gpu_fbdev, helper);
226	struct drm_device *dev = helper->dev;
227	struct virtio_gpu_device *vgdev = dev->dev_private;
228	struct fb_info *info;
229	struct drm_framebuffer *fb;
230	struct drm_mode_fb_cmd2 mode_cmd = {};
231	struct virtio_gpu_object *obj;
232	uint32_t resid, format, size;
233	int ret;
234
235	mode_cmd.width = sizes->surface_width;
236	mode_cmd.height = sizes->surface_height;
237	mode_cmd.pitches[0] = mode_cmd.width * 4;
238	mode_cmd.pixel_format = drm_mode_legacy_fb_format(32, 24);
239
240	switch (mode_cmd.pixel_format) {
241#ifdef __BIG_ENDIAN
242	case DRM_FORMAT_XRGB8888:
243		format = VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM;
244		break;
245	case DRM_FORMAT_ARGB8888:
246		format = VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM;
247		break;
248	case DRM_FORMAT_BGRX8888:
249		format = VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM;
250		break;
251	case DRM_FORMAT_BGRA8888:
252		format = VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM;
253		break;
254	case DRM_FORMAT_RGBX8888:
255		format = VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM;
256		break;
257	case DRM_FORMAT_RGBA8888:
258		format = VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM;
259		break;
260	case DRM_FORMAT_XBGR8888:
261		format = VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM;
262		break;
263	case DRM_FORMAT_ABGR8888:
264		format = VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM;
265		break;
266#else
267	case DRM_FORMAT_XRGB8888:
268		format = VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM;
269		break;
270	case DRM_FORMAT_ARGB8888:
271		format = VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM;
272		break;
273	case DRM_FORMAT_BGRX8888:
274		format = VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM;
275		break;
276	case DRM_FORMAT_BGRA8888:
277		format = VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM;
278		break;
279	case DRM_FORMAT_RGBX8888:
280		format = VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM;
281		break;
282	case DRM_FORMAT_RGBA8888:
283		format = VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM;
284		break;
285	case DRM_FORMAT_XBGR8888:
286		format = VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM;
287		break;
288	case DRM_FORMAT_ABGR8888:
289		format = VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM;
290		break;
291#endif
292	default:
293		DRM_ERROR("failed to find virtio gpu format for %d\n",
294			  mode_cmd.pixel_format);
295		return -EINVAL;
296	}
297
298	size = mode_cmd.pitches[0] * mode_cmd.height;
299	obj = virtio_gpu_alloc_object(dev, size, false, true);
300	if (IS_ERR(obj))
301		return PTR_ERR(obj);
302
303	virtio_gpu_resource_id_get(vgdev, &resid);
304	virtio_gpu_cmd_create_resource(vgdev, resid, format,
305				       mode_cmd.width, mode_cmd.height);
306
307	ret = virtio_gpu_vmap_fb(vgdev, obj);
308	if (ret) {
309		DRM_ERROR("failed to vmap fb %d\n", ret);
310		goto err_obj_vmap;
311	}
312
313	/* attach the object to the resource */
314	ret = virtio_gpu_object_attach(vgdev, obj, resid, NULL);
315	if (ret)
316		goto err_obj_attach;
317
318	info = drm_fb_helper_alloc_fbi(helper);
319	if (IS_ERR(info)) {
320		ret = PTR_ERR(info);
321		goto err_fb_alloc;
322	}
323
324	info->par = helper;
325
326	ret = virtio_gpu_framebuffer_init(dev, &vfbdev->vgfb,
327					  &mode_cmd, &obj->gem_base);
328	if (ret)
329		goto err_fb_init;
330
331	fb = &vfbdev->vgfb.base;
332
333	vfbdev->helper.fb = fb;
334
335	strcpy(info->fix.id, "virtiodrmfb");
336	info->flags = FBINFO_DEFAULT;
337	info->fbops = &virtio_gpufb_ops;
338	info->pixmap.flags = FB_PIXMAP_SYSTEM;
339
340	info->screen_base = obj->vmap;
341	info->screen_size = obj->gem_base.size;
342	drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
343	drm_fb_helper_fill_var(info, &vfbdev->helper,
344			       sizes->fb_width, sizes->fb_height);
345
346	info->fix.mmio_start = 0;
347	info->fix.mmio_len = 0;
348	return 0;
349
350err_fb_init:
351	drm_fb_helper_release_fbi(helper);
352err_fb_alloc:
353	virtio_gpu_cmd_resource_inval_backing(vgdev, resid);
354err_obj_attach:
355err_obj_vmap:
356	virtio_gpu_gem_free_object(&obj->gem_base);
357	return ret;
358}
359
360static int virtio_gpu_fbdev_destroy(struct drm_device *dev,
361				    struct virtio_gpu_fbdev *vgfbdev)
362{
363	struct virtio_gpu_framebuffer *vgfb = &vgfbdev->vgfb;
364
365	drm_fb_helper_unregister_fbi(&vgfbdev->helper);
366	drm_fb_helper_release_fbi(&vgfbdev->helper);
367
368	if (vgfb->obj)
369		vgfb->obj = NULL;
370	drm_fb_helper_fini(&vgfbdev->helper);
371	drm_framebuffer_cleanup(&vgfb->base);
372
373	return 0;
374}
375static struct drm_fb_helper_funcs virtio_gpu_fb_helper_funcs = {
376	.fb_probe = virtio_gpufb_create,
377};
378
379int virtio_gpu_fbdev_init(struct virtio_gpu_device *vgdev)
380{
381	struct virtio_gpu_fbdev *vgfbdev;
382	int bpp_sel = 32; /* TODO: parameter from somewhere? */
383	int ret;
384
385	vgfbdev = kzalloc(sizeof(struct virtio_gpu_fbdev), GFP_KERNEL);
386	if (!vgfbdev)
387		return -ENOMEM;
388
389	vgfbdev->vgdev = vgdev;
390	vgdev->vgfbdev = vgfbdev;
391	INIT_DELAYED_WORK(&vgfbdev->work, virtio_gpu_fb_dirty_work);
392
393	drm_fb_helper_prepare(vgdev->ddev, &vgfbdev->helper,
394			      &virtio_gpu_fb_helper_funcs);
395	ret = drm_fb_helper_init(vgdev->ddev, &vgfbdev->helper,
396				 vgdev->num_scanouts,
397				 VIRTIO_GPUFB_CONN_LIMIT);
398	if (ret) {
399		kfree(vgfbdev);
400		return ret;
401	}
402
403	drm_fb_helper_single_add_all_connectors(&vgfbdev->helper);
404	drm_fb_helper_initial_config(&vgfbdev->helper, bpp_sel);
405	return 0;
406}
407
408void virtio_gpu_fbdev_fini(struct virtio_gpu_device *vgdev)
409{
410	if (!vgdev->vgfbdev)
411		return;
412
413	virtio_gpu_fbdev_destroy(vgdev->ddev, vgdev->vgfbdev);
414	kfree(vgdev->vgfbdev);
415	vgdev->vgfbdev = NULL;
416}