at v4.12 9.9 kB view raw
1/* 2 * Copyright (C) 2015 Red Hat, Inc. 3 * All Rights Reserved. 4 * 5 * Permission is hereby granted, free of charge, to any person obtaining 6 * a copy of this software and associated documentation files (the 7 * "Software"), to deal in the Software without restriction, including 8 * without limitation the rights to use, copy, modify, merge, publish, 9 * distribute, sublicense, and/or sell copies of the Software, and to 10 * permit persons to whom the Software is furnished to do so, subject to 11 * the following conditions: 12 * 13 * The above copyright notice and this permission notice (including the 14 * next paragraph) shall be included in all copies or substantial 15 * portions of the Software. 16 * 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 24 */ 25 26#include <drm/drmP.h> 27#include <drm/drm_fb_helper.h> 28#include "virtgpu_drv.h" 29 30#define VIRTIO_GPU_FBCON_POLL_PERIOD (HZ / 60) 31 32struct virtio_gpu_fbdev { 33 struct drm_fb_helper helper; 34 struct virtio_gpu_framebuffer vgfb; 35 struct virtio_gpu_device *vgdev; 36 struct delayed_work work; 37}; 38 39static int virtio_gpu_dirty_update(struct virtio_gpu_framebuffer *fb, 40 bool store, int x, int y, 41 int width, int height) 42{ 43 struct drm_device *dev = fb->base.dev; 44 struct virtio_gpu_device *vgdev = dev->dev_private; 45 bool store_for_later = false; 46 int bpp = fb->base.format->cpp[0]; 47 int x2, y2; 48 unsigned long flags; 49 struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(fb->obj); 50 51 if ((width <= 0) || 52 (x + width > fb->base.width) || 53 (y + height > fb->base.height)) { 54 DRM_DEBUG("values out of range %dx%d+%d+%d, fb %dx%d\n", 55 width, height, x, y, 56 fb->base.width, fb->base.height); 57 return -EINVAL; 58 } 59 60 /* 61 * Can be called with pretty much any context (console output 62 * path). If we are in atomic just store the dirty rect info 63 * to send out the update later. 64 * 65 * Can't test inside spin lock. 66 */ 67 if (in_atomic() || store) 68 store_for_later = true; 69 70 x2 = x + width - 1; 71 y2 = y + height - 1; 72 73 spin_lock_irqsave(&fb->dirty_lock, flags); 74 75 if (fb->y1 < y) 76 y = fb->y1; 77 if (fb->y2 > y2) 78 y2 = fb->y2; 79 if (fb->x1 < x) 80 x = fb->x1; 81 if (fb->x2 > x2) 82 x2 = fb->x2; 83 84 if (store_for_later) { 85 fb->x1 = x; 86 fb->x2 = x2; 87 fb->y1 = y; 88 fb->y2 = y2; 89 spin_unlock_irqrestore(&fb->dirty_lock, flags); 90 return 0; 91 } 92 93 fb->x1 = fb->y1 = INT_MAX; 94 fb->x2 = fb->y2 = 0; 95 96 spin_unlock_irqrestore(&fb->dirty_lock, flags); 97 98 { 99 uint32_t offset; 100 uint32_t w = x2 - x + 1; 101 uint32_t h = y2 - y + 1; 102 103 offset = (y * fb->base.pitches[0]) + x * bpp; 104 105 virtio_gpu_cmd_transfer_to_host_2d(vgdev, obj->hw_res_handle, 106 offset, 107 cpu_to_le32(w), 108 cpu_to_le32(h), 109 cpu_to_le32(x), 110 cpu_to_le32(y), 111 NULL); 112 113 } 114 virtio_gpu_cmd_resource_flush(vgdev, obj->hw_res_handle, 115 x, y, x2 - x + 1, y2 - y + 1); 116 return 0; 117} 118 119int virtio_gpu_surface_dirty(struct virtio_gpu_framebuffer *vgfb, 120 struct drm_clip_rect *clips, 121 unsigned num_clips) 122{ 123 struct virtio_gpu_device *vgdev = vgfb->base.dev->dev_private; 124 struct virtio_gpu_object *obj = gem_to_virtio_gpu_obj(vgfb->obj); 125 struct drm_clip_rect norect; 126 struct drm_clip_rect *clips_ptr; 127 int left, right, top, bottom; 128 int i; 129 int inc = 1; 130 if (!num_clips) { 131 num_clips = 1; 132 clips = &norect; 133 norect.x1 = norect.y1 = 0; 134 norect.x2 = vgfb->base.width; 135 norect.y2 = vgfb->base.height; 136 } 137 left = clips->x1; 138 right = clips->x2; 139 top = clips->y1; 140 bottom = clips->y2; 141 142 /* skip the first clip rect */ 143 for (i = 1, clips_ptr = clips + inc; 144 i < num_clips; i++, clips_ptr += inc) { 145 left = min_t(int, left, (int)clips_ptr->x1); 146 right = max_t(int, right, (int)clips_ptr->x2); 147 top = min_t(int, top, (int)clips_ptr->y1); 148 bottom = max_t(int, bottom, (int)clips_ptr->y2); 149 } 150 151 if (obj->dumb) 152 return virtio_gpu_dirty_update(vgfb, false, left, top, 153 right - left, bottom - top); 154 155 virtio_gpu_cmd_resource_flush(vgdev, obj->hw_res_handle, 156 left, top, right - left, bottom - top); 157 return 0; 158} 159 160static void virtio_gpu_fb_dirty_work(struct work_struct *work) 161{ 162 struct delayed_work *delayed_work = to_delayed_work(work); 163 struct virtio_gpu_fbdev *vfbdev = 164 container_of(delayed_work, struct virtio_gpu_fbdev, work); 165 struct virtio_gpu_framebuffer *vgfb = &vfbdev->vgfb; 166 167 virtio_gpu_dirty_update(&vfbdev->vgfb, false, vgfb->x1, vgfb->y1, 168 vgfb->x2 - vgfb->x1, vgfb->y2 - vgfb->y1); 169} 170 171static void virtio_gpu_3d_fillrect(struct fb_info *info, 172 const struct fb_fillrect *rect) 173{ 174 struct virtio_gpu_fbdev *vfbdev = info->par; 175 drm_fb_helper_sys_fillrect(info, rect); 176 virtio_gpu_dirty_update(&vfbdev->vgfb, true, rect->dx, rect->dy, 177 rect->width, rect->height); 178 schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD); 179} 180 181static void virtio_gpu_3d_copyarea(struct fb_info *info, 182 const struct fb_copyarea *area) 183{ 184 struct virtio_gpu_fbdev *vfbdev = info->par; 185 drm_fb_helper_sys_copyarea(info, area); 186 virtio_gpu_dirty_update(&vfbdev->vgfb, true, area->dx, area->dy, 187 area->width, area->height); 188 schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD); 189} 190 191static void virtio_gpu_3d_imageblit(struct fb_info *info, 192 const struct fb_image *image) 193{ 194 struct virtio_gpu_fbdev *vfbdev = info->par; 195 drm_fb_helper_sys_imageblit(info, image); 196 virtio_gpu_dirty_update(&vfbdev->vgfb, true, image->dx, image->dy, 197 image->width, image->height); 198 schedule_delayed_work(&vfbdev->work, VIRTIO_GPU_FBCON_POLL_PERIOD); 199} 200 201static struct fb_ops virtio_gpufb_ops = { 202 .owner = THIS_MODULE, 203 DRM_FB_HELPER_DEFAULT_OPS, 204 .fb_fillrect = virtio_gpu_3d_fillrect, 205 .fb_copyarea = virtio_gpu_3d_copyarea, 206 .fb_imageblit = virtio_gpu_3d_imageblit, 207}; 208 209static int virtio_gpu_vmap_fb(struct virtio_gpu_device *vgdev, 210 struct virtio_gpu_object *obj) 211{ 212 return virtio_gpu_object_kmap(obj, NULL); 213} 214 215static int virtio_gpufb_create(struct drm_fb_helper *helper, 216 struct drm_fb_helper_surface_size *sizes) 217{ 218 struct virtio_gpu_fbdev *vfbdev = 219 container_of(helper, struct virtio_gpu_fbdev, helper); 220 struct drm_device *dev = helper->dev; 221 struct virtio_gpu_device *vgdev = dev->dev_private; 222 struct fb_info *info; 223 struct drm_framebuffer *fb; 224 struct drm_mode_fb_cmd2 mode_cmd = {}; 225 struct virtio_gpu_object *obj; 226 uint32_t resid, format, size; 227 int ret; 228 229 mode_cmd.width = sizes->surface_width; 230 mode_cmd.height = sizes->surface_height; 231 mode_cmd.pitches[0] = mode_cmd.width * 4; 232 mode_cmd.pixel_format = drm_mode_legacy_fb_format(32, 24); 233 234 format = virtio_gpu_translate_format(mode_cmd.pixel_format); 235 if (format == 0) 236 return -EINVAL; 237 238 size = mode_cmd.pitches[0] * mode_cmd.height; 239 obj = virtio_gpu_alloc_object(dev, size, false, true); 240 if (IS_ERR(obj)) 241 return PTR_ERR(obj); 242 243 virtio_gpu_resource_id_get(vgdev, &resid); 244 virtio_gpu_cmd_create_resource(vgdev, resid, format, 245 mode_cmd.width, mode_cmd.height); 246 247 ret = virtio_gpu_vmap_fb(vgdev, obj); 248 if (ret) { 249 DRM_ERROR("failed to vmap fb %d\n", ret); 250 goto err_obj_vmap; 251 } 252 253 /* attach the object to the resource */ 254 ret = virtio_gpu_object_attach(vgdev, obj, resid, NULL); 255 if (ret) 256 goto err_obj_attach; 257 258 info = drm_fb_helper_alloc_fbi(helper); 259 if (IS_ERR(info)) { 260 ret = PTR_ERR(info); 261 goto err_fb_alloc; 262 } 263 264 info->par = helper; 265 266 ret = virtio_gpu_framebuffer_init(dev, &vfbdev->vgfb, 267 &mode_cmd, &obj->gem_base); 268 if (ret) 269 goto err_fb_alloc; 270 271 fb = &vfbdev->vgfb.base; 272 273 vfbdev->helper.fb = fb; 274 275 strcpy(info->fix.id, "virtiodrmfb"); 276 info->flags = FBINFO_DEFAULT; 277 info->fbops = &virtio_gpufb_ops; 278 info->pixmap.flags = FB_PIXMAP_SYSTEM; 279 280 info->screen_buffer = obj->vmap; 281 info->screen_size = obj->gem_base.size; 282 drm_fb_helper_fill_fix(info, fb->pitches[0], fb->format->depth); 283 drm_fb_helper_fill_var(info, &vfbdev->helper, 284 sizes->fb_width, sizes->fb_height); 285 286 info->fix.mmio_start = 0; 287 info->fix.mmio_len = 0; 288 return 0; 289 290err_fb_alloc: 291 virtio_gpu_cmd_resource_inval_backing(vgdev, resid); 292err_obj_attach: 293err_obj_vmap: 294 virtio_gpu_gem_free_object(&obj->gem_base); 295 return ret; 296} 297 298static int virtio_gpu_fbdev_destroy(struct drm_device *dev, 299 struct virtio_gpu_fbdev *vgfbdev) 300{ 301 struct virtio_gpu_framebuffer *vgfb = &vgfbdev->vgfb; 302 303 drm_fb_helper_unregister_fbi(&vgfbdev->helper); 304 305 if (vgfb->obj) 306 vgfb->obj = NULL; 307 drm_fb_helper_fini(&vgfbdev->helper); 308 drm_framebuffer_cleanup(&vgfb->base); 309 310 return 0; 311} 312static struct drm_fb_helper_funcs virtio_gpu_fb_helper_funcs = { 313 .fb_probe = virtio_gpufb_create, 314}; 315 316int virtio_gpu_fbdev_init(struct virtio_gpu_device *vgdev) 317{ 318 struct virtio_gpu_fbdev *vgfbdev; 319 int bpp_sel = 32; /* TODO: parameter from somewhere? */ 320 int ret; 321 322 vgfbdev = kzalloc(sizeof(struct virtio_gpu_fbdev), GFP_KERNEL); 323 if (!vgfbdev) 324 return -ENOMEM; 325 326 vgfbdev->vgdev = vgdev; 327 vgdev->vgfbdev = vgfbdev; 328 INIT_DELAYED_WORK(&vgfbdev->work, virtio_gpu_fb_dirty_work); 329 330 drm_fb_helper_prepare(vgdev->ddev, &vgfbdev->helper, 331 &virtio_gpu_fb_helper_funcs); 332 ret = drm_fb_helper_init(vgdev->ddev, &vgfbdev->helper, 333 VIRTIO_GPUFB_CONN_LIMIT); 334 if (ret) { 335 kfree(vgfbdev); 336 return ret; 337 } 338 339 drm_fb_helper_single_add_all_connectors(&vgfbdev->helper); 340 drm_fb_helper_initial_config(&vgfbdev->helper, bpp_sel); 341 return 0; 342} 343 344void virtio_gpu_fbdev_fini(struct virtio_gpu_device *vgdev) 345{ 346 if (!vgdev->vgfbdev) 347 return; 348 349 virtio_gpu_fbdev_destroy(vgdev->ddev, vgdev->vgfbdev); 350 kfree(vgdev->vgfbdev); 351 vgdev->vgfbdev = NULL; 352}