Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v4.5-rc6 655 lines 16 kB view raw
1/* 2 * Copyright (C) 2012 Red Hat 3 * 4 * based in parts on udlfb.c: 5 * Copyright (C) 2009 Roberto De Ioris <roberto@unbit.it> 6 * Copyright (C) 2009 Jaya Kumar <jayakumar.lkml@gmail.com> 7 * Copyright (C) 2009 Bernie Thompson <bernie@plugable.com> 8 * 9 * This file is subject to the terms and conditions of the GNU General Public 10 * License v2. See the file COPYING in the main directory of this archive for 11 * more details. 12 */ 13#include <linux/module.h> 14#include <linux/slab.h> 15#include <linux/fb.h> 16#include <linux/dma-buf.h> 17 18#include <drm/drmP.h> 19#include <drm/drm_crtc.h> 20#include <drm/drm_crtc_helper.h> 21#include "udl_drv.h" 22 23#include <drm/drm_fb_helper.h> 24 25#define DL_DEFIO_WRITE_DELAY (HZ/20) /* fb_deferred_io.delay in jiffies */ 26 27static int fb_defio = 0; /* Optionally enable experimental fb_defio mmap support */ 28static int fb_bpp = 16; 29 30module_param(fb_bpp, int, S_IWUSR | S_IRUSR | S_IWGRP | S_IRGRP); 31module_param(fb_defio, int, S_IWUSR | S_IRUSR | S_IWGRP | S_IRGRP); 32 33struct udl_fbdev { 34 struct drm_fb_helper helper; 35 struct udl_framebuffer ufb; 36 int fb_count; 37}; 38 39#define DL_ALIGN_UP(x, a) ALIGN(x, a) 40#define DL_ALIGN_DOWN(x, a) ALIGN(x-(a-1), a) 41 42/** Read the red component (0..255) of a 32 bpp colour. */ 43#define DLO_RGB_GETRED(col) (uint8_t)((col) & 0xFF) 44 45/** Read the green component (0..255) of a 32 bpp colour. */ 46#define DLO_RGB_GETGRN(col) (uint8_t)(((col) >> 8) & 0xFF) 47 48/** Read the blue component (0..255) of a 32 bpp colour. */ 49#define DLO_RGB_GETBLU(col) (uint8_t)(((col) >> 16) & 0xFF) 50 51/** Return red/green component of a 16 bpp colour number. */ 52#define DLO_RG16(red, grn) (uint8_t)((((red) & 0xF8) | ((grn) >> 5)) & 0xFF) 53 54/** Return green/blue component of a 16 bpp colour number. */ 55#define DLO_GB16(grn, blu) (uint8_t)(((((grn) & 0x1C) << 3) | ((blu) >> 3)) & 0xFF) 56 57/** Return 8 bpp colour number from red, green and blue components. */ 58#define DLO_RGB8(red, grn, blu) ((((red) << 5) | (((grn) & 3) << 3) | ((blu) & 7)) & 0xFF) 59 60#if 0 61static uint8_t rgb8(uint32_t col) 62{ 63 uint8_t red = DLO_RGB_GETRED(col); 64 uint8_t grn = DLO_RGB_GETGRN(col); 65 uint8_t blu = DLO_RGB_GETBLU(col); 66 67 return DLO_RGB8(red, grn, blu); 68} 69 70static uint16_t rgb16(uint32_t col) 71{ 72 uint8_t red = DLO_RGB_GETRED(col); 73 uint8_t grn = DLO_RGB_GETGRN(col); 74 uint8_t blu = DLO_RGB_GETBLU(col); 75 76 return (DLO_RG16(red, grn) << 8) + DLO_GB16(grn, blu); 77} 78#endif 79 80/* 81 * NOTE: fb_defio.c is holding info->fbdefio.mutex 82 * Touching ANY framebuffer memory that triggers a page fault 83 * in fb_defio will cause a deadlock, when it also tries to 84 * grab the same mutex. 85 */ 86static void udlfb_dpy_deferred_io(struct fb_info *info, 87 struct list_head *pagelist) 88{ 89 struct page *cur; 90 struct fb_deferred_io *fbdefio = info->fbdefio; 91 struct udl_fbdev *ufbdev = info->par; 92 struct drm_device *dev = ufbdev->ufb.base.dev; 93 struct udl_device *udl = dev->dev_private; 94 struct urb *urb; 95 char *cmd; 96 cycles_t start_cycles, end_cycles; 97 int bytes_sent = 0; 98 int bytes_identical = 0; 99 int bytes_rendered = 0; 100 101 if (!fb_defio) 102 return; 103 104 start_cycles = get_cycles(); 105 106 urb = udl_get_urb(dev); 107 if (!urb) 108 return; 109 110 cmd = urb->transfer_buffer; 111 112 /* walk the written page list and render each to device */ 113 list_for_each_entry(cur, &fbdefio->pagelist, lru) { 114 115 if (udl_render_hline(dev, (ufbdev->ufb.base.bits_per_pixel / 8), 116 &urb, (char *) info->fix.smem_start, 117 &cmd, cur->index << PAGE_SHIFT, 118 cur->index << PAGE_SHIFT, 119 PAGE_SIZE, &bytes_identical, &bytes_sent)) 120 goto error; 121 bytes_rendered += PAGE_SIZE; 122 } 123 124 if (cmd > (char *) urb->transfer_buffer) { 125 /* Send partial buffer remaining before exiting */ 126 int len = cmd - (char *) urb->transfer_buffer; 127 udl_submit_urb(dev, urb, len); 128 bytes_sent += len; 129 } else 130 udl_urb_completion(urb); 131 132error: 133 atomic_add(bytes_sent, &udl->bytes_sent); 134 atomic_add(bytes_identical, &udl->bytes_identical); 135 atomic_add(bytes_rendered, &udl->bytes_rendered); 136 end_cycles = get_cycles(); 137 atomic_add(((unsigned int) ((end_cycles - start_cycles) 138 >> 10)), /* Kcycles */ 139 &udl->cpu_kcycles_used); 140} 141 142int udl_handle_damage(struct udl_framebuffer *fb, int x, int y, 143 int width, int height) 144{ 145 struct drm_device *dev = fb->base.dev; 146 struct udl_device *udl = dev->dev_private; 147 int i, ret; 148 char *cmd; 149 cycles_t start_cycles, end_cycles; 150 int bytes_sent = 0; 151 int bytes_identical = 0; 152 struct urb *urb; 153 int aligned_x; 154 int bpp = (fb->base.bits_per_pixel / 8); 155 int x2, y2; 156 bool store_for_later = false; 157 unsigned long flags; 158 159 if (!fb->active_16) 160 return 0; 161 162 if (!fb->obj->vmapping) { 163 ret = udl_gem_vmap(fb->obj); 164 if (ret == -ENOMEM) { 165 DRM_ERROR("failed to vmap fb\n"); 166 return 0; 167 } 168 if (!fb->obj->vmapping) { 169 DRM_ERROR("failed to vmapping\n"); 170 return 0; 171 } 172 } 173 174 aligned_x = DL_ALIGN_DOWN(x, sizeof(unsigned long)); 175 width = DL_ALIGN_UP(width + (x-aligned_x), sizeof(unsigned long)); 176 x = aligned_x; 177 178 if ((width <= 0) || 179 (x + width > fb->base.width) || 180 (y + height > fb->base.height)) 181 return -EINVAL; 182 183 /* if we are in atomic just store the info 184 can't test inside spin lock */ 185 if (in_atomic()) 186 store_for_later = true; 187 188 x2 = x + width - 1; 189 y2 = y + height - 1; 190 191 spin_lock_irqsave(&fb->dirty_lock, flags); 192 193 if (fb->y1 < y) 194 y = fb->y1; 195 if (fb->y2 > y2) 196 y2 = fb->y2; 197 if (fb->x1 < x) 198 x = fb->x1; 199 if (fb->x2 > x2) 200 x2 = fb->x2; 201 202 if (store_for_later) { 203 fb->x1 = x; 204 fb->x2 = x2; 205 fb->y1 = y; 206 fb->y2 = y2; 207 spin_unlock_irqrestore(&fb->dirty_lock, flags); 208 return 0; 209 } 210 211 fb->x1 = fb->y1 = INT_MAX; 212 fb->x2 = fb->y2 = 0; 213 214 spin_unlock_irqrestore(&fb->dirty_lock, flags); 215 start_cycles = get_cycles(); 216 217 urb = udl_get_urb(dev); 218 if (!urb) 219 return 0; 220 cmd = urb->transfer_buffer; 221 222 for (i = y; i <= y2 ; i++) { 223 const int line_offset = fb->base.pitches[0] * i; 224 const int byte_offset = line_offset + (x * bpp); 225 const int dev_byte_offset = (fb->base.width * bpp * i) + (x * bpp); 226 if (udl_render_hline(dev, bpp, &urb, 227 (char *) fb->obj->vmapping, 228 &cmd, byte_offset, dev_byte_offset, 229 (x2 - x + 1) * bpp, 230 &bytes_identical, &bytes_sent)) 231 goto error; 232 } 233 234 if (cmd > (char *) urb->transfer_buffer) { 235 /* Send partial buffer remaining before exiting */ 236 int len = cmd - (char *) urb->transfer_buffer; 237 ret = udl_submit_urb(dev, urb, len); 238 bytes_sent += len; 239 } else 240 udl_urb_completion(urb); 241 242error: 243 atomic_add(bytes_sent, &udl->bytes_sent); 244 atomic_add(bytes_identical, &udl->bytes_identical); 245 atomic_add(width*height*bpp, &udl->bytes_rendered); 246 end_cycles = get_cycles(); 247 atomic_add(((unsigned int) ((end_cycles - start_cycles) 248 >> 10)), /* Kcycles */ 249 &udl->cpu_kcycles_used); 250 251 return 0; 252} 253 254static int udl_fb_mmap(struct fb_info *info, struct vm_area_struct *vma) 255{ 256 unsigned long start = vma->vm_start; 257 unsigned long size = vma->vm_end - vma->vm_start; 258 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; 259 unsigned long page, pos; 260 261 if (offset + size > info->fix.smem_len) 262 return -EINVAL; 263 264 pos = (unsigned long)info->fix.smem_start + offset; 265 266 pr_notice("mmap() framebuffer addr:%lu size:%lu\n", 267 pos, size); 268 269 while (size > 0) { 270 page = vmalloc_to_pfn((void *)pos); 271 if (remap_pfn_range(vma, start, page, PAGE_SIZE, PAGE_SHARED)) 272 return -EAGAIN; 273 274 start += PAGE_SIZE; 275 pos += PAGE_SIZE; 276 if (size > PAGE_SIZE) 277 size -= PAGE_SIZE; 278 else 279 size = 0; 280 } 281 282 /* VM_IO | VM_DONTEXPAND | VM_DONTDUMP are set by remap_pfn_range() */ 283 return 0; 284} 285 286static void udl_fb_fillrect(struct fb_info *info, const struct fb_fillrect *rect) 287{ 288 struct udl_fbdev *ufbdev = info->par; 289 290 drm_fb_helper_sys_fillrect(info, rect); 291 292 udl_handle_damage(&ufbdev->ufb, rect->dx, rect->dy, rect->width, 293 rect->height); 294} 295 296static void udl_fb_copyarea(struct fb_info *info, const struct fb_copyarea *region) 297{ 298 struct udl_fbdev *ufbdev = info->par; 299 300 drm_fb_helper_sys_copyarea(info, region); 301 302 udl_handle_damage(&ufbdev->ufb, region->dx, region->dy, region->width, 303 region->height); 304} 305 306static void udl_fb_imageblit(struct fb_info *info, const struct fb_image *image) 307{ 308 struct udl_fbdev *ufbdev = info->par; 309 310 drm_fb_helper_sys_imageblit(info, image); 311 312 udl_handle_damage(&ufbdev->ufb, image->dx, image->dy, image->width, 313 image->height); 314} 315 316/* 317 * It's common for several clients to have framebuffer open simultaneously. 318 * e.g. both fbcon and X. Makes things interesting. 319 * Assumes caller is holding info->lock (for open and release at least) 320 */ 321static int udl_fb_open(struct fb_info *info, int user) 322{ 323 struct udl_fbdev *ufbdev = info->par; 324 struct drm_device *dev = ufbdev->ufb.base.dev; 325 struct udl_device *udl = dev->dev_private; 326 327 /* If the USB device is gone, we don't accept new opens */ 328 if (drm_device_is_unplugged(udl->ddev)) 329 return -ENODEV; 330 331 ufbdev->fb_count++; 332 333 if (fb_defio && (info->fbdefio == NULL)) { 334 /* enable defio at last moment if not disabled by client */ 335 336 struct fb_deferred_io *fbdefio; 337 338 fbdefio = kmalloc(sizeof(struct fb_deferred_io), GFP_KERNEL); 339 340 if (fbdefio) { 341 fbdefio->delay = DL_DEFIO_WRITE_DELAY; 342 fbdefio->deferred_io = udlfb_dpy_deferred_io; 343 } 344 345 info->fbdefio = fbdefio; 346 fb_deferred_io_init(info); 347 } 348 349 pr_notice("open /dev/fb%d user=%d fb_info=%p count=%d\n", 350 info->node, user, info, ufbdev->fb_count); 351 352 return 0; 353} 354 355 356/* 357 * Assumes caller is holding info->lock mutex (for open and release at least) 358 */ 359static int udl_fb_release(struct fb_info *info, int user) 360{ 361 struct udl_fbdev *ufbdev = info->par; 362 363 ufbdev->fb_count--; 364 365 if ((ufbdev->fb_count == 0) && (info->fbdefio)) { 366 fb_deferred_io_cleanup(info); 367 kfree(info->fbdefio); 368 info->fbdefio = NULL; 369 info->fbops->fb_mmap = udl_fb_mmap; 370 } 371 372 pr_warn("released /dev/fb%d user=%d count=%d\n", 373 info->node, user, ufbdev->fb_count); 374 375 return 0; 376} 377 378static struct fb_ops udlfb_ops = { 379 .owner = THIS_MODULE, 380 .fb_check_var = drm_fb_helper_check_var, 381 .fb_set_par = drm_fb_helper_set_par, 382 .fb_fillrect = udl_fb_fillrect, 383 .fb_copyarea = udl_fb_copyarea, 384 .fb_imageblit = udl_fb_imageblit, 385 .fb_pan_display = drm_fb_helper_pan_display, 386 .fb_blank = drm_fb_helper_blank, 387 .fb_setcmap = drm_fb_helper_setcmap, 388 .fb_debug_enter = drm_fb_helper_debug_enter, 389 .fb_debug_leave = drm_fb_helper_debug_leave, 390 .fb_mmap = udl_fb_mmap, 391 .fb_open = udl_fb_open, 392 .fb_release = udl_fb_release, 393}; 394 395static int udl_user_framebuffer_dirty(struct drm_framebuffer *fb, 396 struct drm_file *file, 397 unsigned flags, unsigned color, 398 struct drm_clip_rect *clips, 399 unsigned num_clips) 400{ 401 struct udl_framebuffer *ufb = to_udl_fb(fb); 402 int i; 403 int ret = 0; 404 405 drm_modeset_lock_all(fb->dev); 406 407 if (!ufb->active_16) 408 goto unlock; 409 410 if (ufb->obj->base.import_attach) { 411 ret = dma_buf_begin_cpu_access(ufb->obj->base.import_attach->dmabuf, 412 0, ufb->obj->base.size, 413 DMA_FROM_DEVICE); 414 if (ret) 415 goto unlock; 416 } 417 418 for (i = 0; i < num_clips; i++) { 419 ret = udl_handle_damage(ufb, clips[i].x1, clips[i].y1, 420 clips[i].x2 - clips[i].x1, 421 clips[i].y2 - clips[i].y1); 422 if (ret) 423 break; 424 } 425 426 if (ufb->obj->base.import_attach) { 427 dma_buf_end_cpu_access(ufb->obj->base.import_attach->dmabuf, 428 0, ufb->obj->base.size, 429 DMA_FROM_DEVICE); 430 } 431 432 unlock: 433 drm_modeset_unlock_all(fb->dev); 434 435 return ret; 436} 437 438static void udl_user_framebuffer_destroy(struct drm_framebuffer *fb) 439{ 440 struct udl_framebuffer *ufb = to_udl_fb(fb); 441 442 if (ufb->obj) 443 drm_gem_object_unreference_unlocked(&ufb->obj->base); 444 445 drm_framebuffer_cleanup(fb); 446 kfree(ufb); 447} 448 449static const struct drm_framebuffer_funcs udlfb_funcs = { 450 .destroy = udl_user_framebuffer_destroy, 451 .dirty = udl_user_framebuffer_dirty, 452}; 453 454 455static int 456udl_framebuffer_init(struct drm_device *dev, 457 struct udl_framebuffer *ufb, 458 const struct drm_mode_fb_cmd2 *mode_cmd, 459 struct udl_gem_object *obj) 460{ 461 int ret; 462 463 spin_lock_init(&ufb->dirty_lock); 464 ufb->obj = obj; 465 drm_helper_mode_fill_fb_struct(&ufb->base, mode_cmd); 466 ret = drm_framebuffer_init(dev, &ufb->base, &udlfb_funcs); 467 return ret; 468} 469 470 471static int udlfb_create(struct drm_fb_helper *helper, 472 struct drm_fb_helper_surface_size *sizes) 473{ 474 struct udl_fbdev *ufbdev = 475 container_of(helper, struct udl_fbdev, helper); 476 struct drm_device *dev = ufbdev->helper.dev; 477 struct fb_info *info; 478 struct drm_framebuffer *fb; 479 struct drm_mode_fb_cmd2 mode_cmd; 480 struct udl_gem_object *obj; 481 uint32_t size; 482 int ret = 0; 483 484 if (sizes->surface_bpp == 24) 485 sizes->surface_bpp = 32; 486 487 mode_cmd.width = sizes->surface_width; 488 mode_cmd.height = sizes->surface_height; 489 mode_cmd.pitches[0] = mode_cmd.width * ((sizes->surface_bpp + 7) / 8); 490 491 mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp, 492 sizes->surface_depth); 493 494 size = mode_cmd.pitches[0] * mode_cmd.height; 495 size = ALIGN(size, PAGE_SIZE); 496 497 obj = udl_gem_alloc_object(dev, size); 498 if (!obj) 499 goto out; 500 501 ret = udl_gem_vmap(obj); 502 if (ret) { 503 DRM_ERROR("failed to vmap fb\n"); 504 goto out_gfree; 505 } 506 507 info = drm_fb_helper_alloc_fbi(helper); 508 if (IS_ERR(info)) { 509 ret = PTR_ERR(info); 510 goto out_gfree; 511 } 512 info->par = ufbdev; 513 514 ret = udl_framebuffer_init(dev, &ufbdev->ufb, &mode_cmd, obj); 515 if (ret) 516 goto out_destroy_fbi; 517 518 fb = &ufbdev->ufb.base; 519 520 ufbdev->helper.fb = fb; 521 522 strcpy(info->fix.id, "udldrmfb"); 523 524 info->screen_base = ufbdev->ufb.obj->vmapping; 525 info->fix.smem_len = size; 526 info->fix.smem_start = (unsigned long)ufbdev->ufb.obj->vmapping; 527 528 info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT; 529 info->fbops = &udlfb_ops; 530 drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth); 531 drm_fb_helper_fill_var(info, &ufbdev->helper, sizes->fb_width, sizes->fb_height); 532 533 DRM_DEBUG_KMS("allocated %dx%d vmal %p\n", 534 fb->width, fb->height, 535 ufbdev->ufb.obj->vmapping); 536 537 return ret; 538out_destroy_fbi: 539 drm_fb_helper_release_fbi(helper); 540out_gfree: 541 drm_gem_object_unreference(&ufbdev->ufb.obj->base); 542out: 543 return ret; 544} 545 546static const struct drm_fb_helper_funcs udl_fb_helper_funcs = { 547 .fb_probe = udlfb_create, 548}; 549 550static void udl_fbdev_destroy(struct drm_device *dev, 551 struct udl_fbdev *ufbdev) 552{ 553 drm_fb_helper_unregister_fbi(&ufbdev->helper); 554 drm_fb_helper_release_fbi(&ufbdev->helper); 555 drm_fb_helper_fini(&ufbdev->helper); 556 drm_framebuffer_unregister_private(&ufbdev->ufb.base); 557 drm_framebuffer_cleanup(&ufbdev->ufb.base); 558 drm_gem_object_unreference_unlocked(&ufbdev->ufb.obj->base); 559} 560 561int udl_fbdev_init(struct drm_device *dev) 562{ 563 struct udl_device *udl = dev->dev_private; 564 int bpp_sel = fb_bpp; 565 struct udl_fbdev *ufbdev; 566 int ret; 567 568 ufbdev = kzalloc(sizeof(struct udl_fbdev), GFP_KERNEL); 569 if (!ufbdev) 570 return -ENOMEM; 571 572 udl->fbdev = ufbdev; 573 574 drm_fb_helper_prepare(dev, &ufbdev->helper, &udl_fb_helper_funcs); 575 576 ret = drm_fb_helper_init(dev, &ufbdev->helper, 577 1, 1); 578 if (ret) 579 goto free; 580 581 ret = drm_fb_helper_single_add_all_connectors(&ufbdev->helper); 582 if (ret) 583 goto fini; 584 585 /* disable all the possible outputs/crtcs before entering KMS mode */ 586 drm_helper_disable_unused_functions(dev); 587 588 ret = drm_fb_helper_initial_config(&ufbdev->helper, bpp_sel); 589 if (ret) 590 goto fini; 591 592 return 0; 593 594fini: 595 drm_fb_helper_fini(&ufbdev->helper); 596free: 597 kfree(ufbdev); 598 return ret; 599} 600 601void udl_fbdev_cleanup(struct drm_device *dev) 602{ 603 struct udl_device *udl = dev->dev_private; 604 if (!udl->fbdev) 605 return; 606 607 udl_fbdev_destroy(dev, udl->fbdev); 608 kfree(udl->fbdev); 609 udl->fbdev = NULL; 610} 611 612void udl_fbdev_unplug(struct drm_device *dev) 613{ 614 struct udl_device *udl = dev->dev_private; 615 struct udl_fbdev *ufbdev; 616 if (!udl->fbdev) 617 return; 618 619 ufbdev = udl->fbdev; 620 drm_fb_helper_unlink_fbi(&ufbdev->helper); 621} 622 623struct drm_framebuffer * 624udl_fb_user_fb_create(struct drm_device *dev, 625 struct drm_file *file, 626 const struct drm_mode_fb_cmd2 *mode_cmd) 627{ 628 struct drm_gem_object *obj; 629 struct udl_framebuffer *ufb; 630 int ret; 631 uint32_t size; 632 633 obj = drm_gem_object_lookup(dev, file, mode_cmd->handles[0]); 634 if (obj == NULL) 635 return ERR_PTR(-ENOENT); 636 637 size = mode_cmd->pitches[0] * mode_cmd->height; 638 size = ALIGN(size, PAGE_SIZE); 639 640 if (size > obj->size) { 641 DRM_ERROR("object size not sufficient for fb %d %zu %d %d\n", size, obj->size, mode_cmd->pitches[0], mode_cmd->height); 642 return ERR_PTR(-ENOMEM); 643 } 644 645 ufb = kzalloc(sizeof(*ufb), GFP_KERNEL); 646 if (ufb == NULL) 647 return ERR_PTR(-ENOMEM); 648 649 ret = udl_framebuffer_init(dev, ufb, mode_cmd, to_udl_bo(obj)); 650 if (ret) { 651 kfree(ufb); 652 return ERR_PTR(-EINVAL); 653 } 654 return &ufb->base; 655}