Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

drm/via: fixed coding style issues, simplified return

Fixed brace, macro and spacing coding style issues.
Simplified
-if (ret) return ret;
-return 0;
+return ret;

Signed-off-by: Nicolas Kaiser <nikai@nikai.net>
Signed-off-by: Dave Airlie <airlied@redhat.com>

authored by

Nicolas Kaiser and committed by
Dave Airlie
58c1e85a 56499113

+136 -166
+52 -68
drivers/gpu/drm/via/via_dma.c
··· 58 58 *((uint32_t *)(vb)) = ((nReg) >> 2) | HALCYON_HEADER1; \ 59 59 *((uint32_t *)(vb) + 1) = (nData); \ 60 60 vb = ((uint32_t *)vb) + 2; \ 61 - dev_priv->dma_low +=8; \ 61 + dev_priv->dma_low += 8; \ 62 62 } 63 63 64 64 #define via_flush_write_combine() DRM_MEMORYBARRIER() 65 65 66 - #define VIA_OUT_RING_QW(w1,w2) \ 66 + #define VIA_OUT_RING_QW(w1, w2) do { \ 67 67 *vb++ = (w1); \ 68 68 *vb++ = (w2); \ 69 - dev_priv->dma_low += 8; 69 + dev_priv->dma_low += 8; \ 70 + } while (0) 70 71 71 - static void via_cmdbuf_start(drm_via_private_t * dev_priv); 72 - static void via_cmdbuf_pause(drm_via_private_t * dev_priv); 73 - static void via_cmdbuf_reset(drm_via_private_t * dev_priv); 74 - static void via_cmdbuf_rewind(drm_via_private_t * dev_priv); 75 - static int via_wait_idle(drm_via_private_t * dev_priv); 76 - static void via_pad_cache(drm_via_private_t * dev_priv, int qwords); 72 + static void via_cmdbuf_start(drm_via_private_t *dev_priv); 73 + static void via_cmdbuf_pause(drm_via_private_t *dev_priv); 74 + static void via_cmdbuf_reset(drm_via_private_t *dev_priv); 75 + static void via_cmdbuf_rewind(drm_via_private_t *dev_priv); 76 + static int via_wait_idle(drm_via_private_t *dev_priv); 77 + static void via_pad_cache(drm_via_private_t *dev_priv, int qwords); 77 78 78 79 /* 79 80 * Free space in command buffer. 80 81 */ 81 82 82 - static uint32_t via_cmdbuf_space(drm_via_private_t * dev_priv) 83 + static uint32_t via_cmdbuf_space(drm_via_private_t *dev_priv) 83 84 { 84 85 uint32_t agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr; 85 86 uint32_t hw_addr = *(dev_priv->hw_addr_ptr) - agp_base; ··· 94 93 * How much does the command regulator lag behind? 95 94 */ 96 95 97 - static uint32_t via_cmdbuf_lag(drm_via_private_t * dev_priv) 96 + static uint32_t via_cmdbuf_lag(drm_via_private_t *dev_priv) 98 97 { 99 98 uint32_t agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr; 100 99 uint32_t hw_addr = *(dev_priv->hw_addr_ptr) - agp_base; ··· 109 108 */ 110 109 111 110 static inline int 112 - via_cmdbuf_wait(drm_via_private_t * dev_priv, unsigned int size) 111 + via_cmdbuf_wait(drm_via_private_t *dev_priv, unsigned int size) 113 112 { 114 113 uint32_t agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr; 115 114 uint32_t cur_addr, hw_addr, next_addr; ··· 147 146 dev_priv->dma_high) { 148 147 via_cmdbuf_rewind(dev_priv); 149 148 } 150 - if (via_cmdbuf_wait(dev_priv, size) != 0) { 149 + if (via_cmdbuf_wait(dev_priv, size) != 0) 151 150 return NULL; 152 - } 153 151 154 152 return (uint32_t *) (dev_priv->dma_ptr + dev_priv->dma_low); 155 153 } 156 154 157 - int via_dma_cleanup(struct drm_device * dev) 155 + int via_dma_cleanup(struct drm_device *dev) 158 156 { 159 157 if (dev->dev_private) { 160 158 drm_via_private_t *dev_priv = ··· 171 171 return 0; 172 172 } 173 173 174 - static int via_initialize(struct drm_device * dev, 175 - drm_via_private_t * dev_priv, 176 - drm_via_dma_init_t * init) 174 + static int via_initialize(struct drm_device *dev, 175 + drm_via_private_t *dev_priv, 176 + drm_via_dma_init_t *init) 177 177 { 178 178 if (!dev_priv || !dev_priv->mmio) { 179 179 DRM_ERROR("via_dma_init called before via_map_init\n"); ··· 258 258 return retcode; 259 259 } 260 260 261 - static int via_dispatch_cmdbuffer(struct drm_device * dev, drm_via_cmdbuffer_t * cmd) 261 + static int via_dispatch_cmdbuffer(struct drm_device *dev, drm_via_cmdbuffer_t *cmd) 262 262 { 263 263 drm_via_private_t *dev_priv; 264 264 uint32_t *vb; ··· 271 271 return -EFAULT; 272 272 } 273 273 274 - if (cmd->size > VIA_PCI_BUF_SIZE) { 274 + if (cmd->size > VIA_PCI_BUF_SIZE) 275 275 return -ENOMEM; 276 - } 277 276 278 277 if (DRM_COPY_FROM_USER(dev_priv->pci_buf, cmd->buf, cmd->size)) 279 278 return -EFAULT; ··· 290 291 } 291 292 292 293 vb = via_check_dma(dev_priv, (cmd->size < 0x100) ? 0x102 : cmd->size); 293 - if (vb == NULL) { 294 + if (vb == NULL) 294 295 return -EAGAIN; 295 - } 296 296 297 297 memcpy(vb, dev_priv->pci_buf, cmd->size); 298 298 ··· 309 311 return 0; 310 312 } 311 313 312 - int via_driver_dma_quiescent(struct drm_device * dev) 314 + int via_driver_dma_quiescent(struct drm_device *dev) 313 315 { 314 316 drm_via_private_t *dev_priv = dev->dev_private; 315 317 316 - if (!via_wait_idle(dev_priv)) { 318 + if (!via_wait_idle(dev_priv)) 317 319 return -EBUSY; 318 - } 319 320 return 0; 320 321 } 321 322 ··· 336 339 DRM_DEBUG("buf %p size %lu\n", cmdbuf->buf, cmdbuf->size); 337 340 338 341 ret = via_dispatch_cmdbuffer(dev, cmdbuf); 339 - if (ret) { 340 - return ret; 341 - } 342 - 343 - return 0; 342 + return ret; 344 343 } 345 344 346 - static int via_dispatch_pci_cmdbuffer(struct drm_device * dev, 347 - drm_via_cmdbuffer_t * cmd) 345 + static int via_dispatch_pci_cmdbuffer(struct drm_device *dev, 346 + drm_via_cmdbuffer_t *cmd) 348 347 { 349 348 drm_via_private_t *dev_priv = dev->dev_private; 350 349 int ret; 351 350 352 - if (cmd->size > VIA_PCI_BUF_SIZE) { 351 + if (cmd->size > VIA_PCI_BUF_SIZE) 353 352 return -ENOMEM; 354 - } 355 353 if (DRM_COPY_FROM_USER(dev_priv->pci_buf, cmd->buf, cmd->size)) 356 354 return -EFAULT; 357 355 ··· 372 380 DRM_DEBUG("buf %p size %lu\n", cmdbuf->buf, cmdbuf->size); 373 381 374 382 ret = via_dispatch_pci_cmdbuffer(dev, cmdbuf); 375 - if (ret) { 376 - return ret; 377 - } 378 - 379 - return 0; 383 + return ret; 380 384 } 381 385 382 - static inline uint32_t *via_align_buffer(drm_via_private_t * dev_priv, 386 + static inline uint32_t *via_align_buffer(drm_via_private_t *dev_priv, 383 387 uint32_t * vb, int qw_count) 384 388 { 385 - for (; qw_count > 0; --qw_count) { 389 + for (; qw_count > 0; --qw_count) 386 390 VIA_OUT_RING_QW(HC_DUMMY, HC_DUMMY); 387 - } 388 391 return vb; 389 392 } 390 393 ··· 388 401 * 389 402 * Returns virtual pointer to ring buffer. 390 403 */ 391 - static inline uint32_t *via_get_dma(drm_via_private_t * dev_priv) 404 + static inline uint32_t *via_get_dma(drm_via_private_t *dev_priv) 392 405 { 393 406 return (uint32_t *) (dev_priv->dma_ptr + dev_priv->dma_low); 394 407 } ··· 398 411 * modifying the pause address stored in the buffer itself. If 399 412 * the regulator has already paused, restart it. 400 413 */ 401 - static int via_hook_segment(drm_via_private_t * dev_priv, 414 + static int via_hook_segment(drm_via_private_t *dev_priv, 402 415 uint32_t pause_addr_hi, uint32_t pause_addr_lo, 403 416 int no_pci_fire) 404 417 { 405 418 int paused, count; 406 419 volatile uint32_t *paused_at = dev_priv->last_pause_ptr; 407 - uint32_t reader,ptr; 420 + uint32_t reader, ptr; 408 421 uint32_t diff; 409 422 410 423 paused = 0; 411 424 via_flush_write_combine(); 412 - (void) *(volatile uint32_t *)(via_get_dma(dev_priv) -1); 425 + (void) *(volatile uint32_t *)(via_get_dma(dev_priv) - 1); 413 426 414 427 *paused_at = pause_addr_lo; 415 428 via_flush_write_combine(); ··· 422 435 dev_priv->last_pause_ptr = via_get_dma(dev_priv) - 1; 423 436 424 437 /* 425 - * If there is a possibility that the command reader will 438 + * If there is a possibility that the command reader will 426 439 * miss the new pause address and pause on the old one, 427 440 * In that case we need to program the new start address 428 441 * using PCI. ··· 430 443 431 444 diff = (uint32_t) (ptr - reader) - dev_priv->dma_diff; 432 445 count = 10000000; 433 - while(diff == 0 && count--) { 446 + while (diff == 0 && count--) { 434 447 paused = (VIA_READ(0x41c) & 0x80000000); 435 - if (paused) 448 + if (paused) 436 449 break; 437 450 reader = *(dev_priv->hw_addr_ptr); 438 451 diff = (uint32_t) (ptr - reader) - dev_priv->dma_diff; ··· 464 477 return paused; 465 478 } 466 479 467 - static int via_wait_idle(drm_via_private_t * dev_priv) 480 + static int via_wait_idle(drm_via_private_t *dev_priv) 468 481 { 469 482 int count = 10000000; 470 483 ··· 478 491 return count; 479 492 } 480 493 481 - static uint32_t *via_align_cmd(drm_via_private_t * dev_priv, uint32_t cmd_type, 482 - uint32_t addr, uint32_t * cmd_addr_hi, 483 - uint32_t * cmd_addr_lo, int skip_wait) 494 + static uint32_t *via_align_cmd(drm_via_private_t *dev_priv, uint32_t cmd_type, 495 + uint32_t addr, uint32_t *cmd_addr_hi, 496 + uint32_t *cmd_addr_lo, int skip_wait) 484 497 { 485 498 uint32_t agp_base; 486 499 uint32_t cmd_addr, addr_lo, addr_hi; ··· 508 521 return vb; 509 522 } 510 523 511 - static void via_cmdbuf_start(drm_via_private_t * dev_priv) 524 + static void via_cmdbuf_start(drm_via_private_t *dev_priv) 512 525 { 513 526 uint32_t pause_addr_lo, pause_addr_hi; 514 527 uint32_t start_addr, start_addr_lo; ··· 567 580 dev_priv->dma_diff = ptr - reader; 568 581 } 569 582 570 - static void via_pad_cache(drm_via_private_t * dev_priv, int qwords) 583 + static void via_pad_cache(drm_via_private_t *dev_priv, int qwords) 571 584 { 572 585 uint32_t *vb; 573 586 ··· 577 590 via_align_buffer(dev_priv, vb, qwords); 578 591 } 579 592 580 - static inline void via_dummy_bitblt(drm_via_private_t * dev_priv) 593 + static inline void via_dummy_bitblt(drm_via_private_t *dev_priv) 581 594 { 582 595 uint32_t *vb = via_get_dma(dev_priv); 583 596 SetReg2DAGP(0x0C, (0 | (0 << 16))); ··· 585 598 SetReg2DAGP(0x0, 0x1 | 0x2000 | 0xAA000000); 586 599 } 587 600 588 - static void via_cmdbuf_jump(drm_via_private_t * dev_priv) 601 + static void via_cmdbuf_jump(drm_via_private_t *dev_priv) 589 602 { 590 603 uint32_t agp_base; 591 604 uint32_t pause_addr_lo, pause_addr_hi; ··· 604 617 */ 605 618 606 619 dev_priv->dma_low = 0; 607 - if (via_cmdbuf_wait(dev_priv, CMDBUF_ALIGNMENT_SIZE) != 0) { 620 + if (via_cmdbuf_wait(dev_priv, CMDBUF_ALIGNMENT_SIZE) != 0) 608 621 DRM_ERROR("via_cmdbuf_jump failed\n"); 609 - } 610 622 611 623 via_dummy_bitblt(dev_priv); 612 624 via_dummy_bitblt(dev_priv); ··· 643 657 } 644 658 645 659 646 - static void via_cmdbuf_rewind(drm_via_private_t * dev_priv) 660 + static void via_cmdbuf_rewind(drm_via_private_t *dev_priv) 647 661 { 648 662 via_cmdbuf_jump(dev_priv); 649 663 } 650 664 651 - static void via_cmdbuf_flush(drm_via_private_t * dev_priv, uint32_t cmd_type) 665 + static void via_cmdbuf_flush(drm_via_private_t *dev_priv, uint32_t cmd_type) 652 666 { 653 667 uint32_t pause_addr_lo, pause_addr_hi; 654 668 ··· 656 670 via_hook_segment(dev_priv, pause_addr_hi, pause_addr_lo, 0); 657 671 } 658 672 659 - static void via_cmdbuf_pause(drm_via_private_t * dev_priv) 673 + static void via_cmdbuf_pause(drm_via_private_t *dev_priv) 660 674 { 661 675 via_cmdbuf_flush(dev_priv, HC_HAGPBpID_PAUSE); 662 676 } 663 677 664 - static void via_cmdbuf_reset(drm_via_private_t * dev_priv) 678 + static void via_cmdbuf_reset(drm_via_private_t *dev_priv) 665 679 { 666 680 via_cmdbuf_flush(dev_priv, HC_HAGPBpID_STOP); 667 681 via_wait_idle(dev_priv); ··· 694 708 case VIA_CMDBUF_SPACE: 695 709 while (((tmp_size = via_cmdbuf_space(dev_priv)) < d_siz->size) 696 710 && --count) { 697 - if (!d_siz->wait) { 711 + if (!d_siz->wait) 698 712 break; 699 - } 700 713 } 701 714 if (!count) { 702 715 DRM_ERROR("VIA_CMDBUF_SPACE timed out.\n"); ··· 705 720 case VIA_CMDBUF_LAG: 706 721 while (((tmp_size = via_cmdbuf_lag(dev_priv)) > d_siz->size) 707 722 && --count) { 708 - if (!d_siz->wait) { 723 + if (!d_siz->wait) 709 724 break; 710 - } 711 725 } 712 726 if (!count) { 713 727 DRM_ERROR("VIA_CMDBUF_LAG timed out.\n");
+32 -39
drivers/gpu/drm/via/via_dmablit.c
··· 70 70 descriptor_this_page; 71 71 dma_addr_t next = vsg->chain_start; 72 72 73 - while(num_desc--) { 73 + while (num_desc--) { 74 74 if (descriptor_this_page-- == 0) { 75 75 cur_descriptor_page--; 76 76 descriptor_this_page = vsg->descriptors_per_page - 1; ··· 174 174 struct page *page; 175 175 int i; 176 176 177 - switch(vsg->state) { 177 + switch (vsg->state) { 178 178 case dr_via_device_mapped: 179 179 via_unmap_blit_from_device(pdev, vsg); 180 180 case dr_via_desc_pages_alloc: 181 - for (i=0; i<vsg->num_desc_pages; ++i) { 181 + for (i = 0; i < vsg->num_desc_pages; ++i) { 182 182 if (vsg->desc_pages[i] != NULL) 183 - free_page((unsigned long)vsg->desc_pages[i]); 183 + free_page((unsigned long)vsg->desc_pages[i]); 184 184 } 185 185 kfree(vsg->desc_pages); 186 186 case dr_via_pages_locked: 187 - for (i=0; i<vsg->num_pages; ++i) { 188 - if ( NULL != (page = vsg->pages[i])) { 189 - if (! PageReserved(page) && (DMA_FROM_DEVICE == vsg->direction)) 187 + for (i = 0; i < vsg->num_pages; ++i) { 188 + if (NULL != (page = vsg->pages[i])) { 189 + if (!PageReserved(page) && (DMA_FROM_DEVICE == vsg->direction)) 190 190 SetPageDirty(page); 191 191 page_cache_release(page); 192 192 } ··· 232 232 { 233 233 int ret; 234 234 unsigned long first_pfn = VIA_PFN(xfer->mem_addr); 235 - vsg->num_pages = VIA_PFN(xfer->mem_addr + (xfer->num_lines * xfer->mem_stride -1)) - 235 + vsg->num_pages = VIA_PFN(xfer->mem_addr + (xfer->num_lines * xfer->mem_stride - 1)) - 236 236 first_pfn + 1; 237 237 238 238 if (NULL == (vsg->pages = vmalloc(sizeof(struct page *) * vsg->num_pages))) ··· 268 268 { 269 269 int i; 270 270 271 - vsg->descriptors_per_page = PAGE_SIZE / sizeof( drm_via_descriptor_t); 271 + vsg->descriptors_per_page = PAGE_SIZE / sizeof(drm_via_descriptor_t); 272 272 vsg->num_desc_pages = (vsg->num_desc + vsg->descriptors_per_page - 1) / 273 273 vsg->descriptors_per_page; 274 274 ··· 276 276 return -ENOMEM; 277 277 278 278 vsg->state = dr_via_desc_pages_alloc; 279 - for (i=0; i<vsg->num_desc_pages; ++i) { 279 + for (i = 0; i < vsg->num_desc_pages; ++i) { 280 280 if (NULL == (vsg->desc_pages[i] = 281 281 (drm_via_descriptor_t *) __get_free_page(GFP_KERNEL))) 282 282 return -ENOMEM; ··· 318 318 drm_via_blitq_t *blitq = dev_priv->blit_queues + engine; 319 319 int cur; 320 320 int done_transfer; 321 - unsigned long irqsave=0; 321 + unsigned long irqsave = 0; 322 322 uint32_t status = 0; 323 323 324 324 DRM_DEBUG("DMA blit handler called. engine = %d, from_irq = %d, blitq = 0x%lx\n", 325 325 engine, from_irq, (unsigned long) blitq); 326 326 327 - if (from_irq) { 327 + if (from_irq) 328 328 spin_lock(&blitq->blit_lock); 329 - } else { 329 + else 330 330 spin_lock_irqsave(&blitq->blit_lock, irqsave); 331 - } 332 331 333 332 done_transfer = blitq->is_active && 334 - (( status = VIA_READ(VIA_PCI_DMA_CSR0 + engine*0x04)) & VIA_DMA_CSR_TD); 335 - done_transfer = done_transfer || ( blitq->aborting && !(status & VIA_DMA_CSR_DE)); 333 + ((status = VIA_READ(VIA_PCI_DMA_CSR0 + engine*0x04)) & VIA_DMA_CSR_TD); 334 + done_transfer = done_transfer || (blitq->aborting && !(status & VIA_DMA_CSR_DE)); 336 335 337 336 cur = blitq->cur; 338 337 if (done_transfer) { ··· 376 377 if (!timer_pending(&blitq->poll_timer)) 377 378 mod_timer(&blitq->poll_timer, jiffies + 1); 378 379 } else { 379 - if (timer_pending(&blitq->poll_timer)) { 380 + if (timer_pending(&blitq->poll_timer)) 380 381 del_timer(&blitq->poll_timer); 381 - } 382 382 via_dmablit_engine_off(dev, engine); 383 383 } 384 384 } 385 385 386 - if (from_irq) { 386 + if (from_irq) 387 387 spin_unlock(&blitq->blit_lock); 388 - } else { 388 + else 389 389 spin_unlock_irqrestore(&blitq->blit_lock, irqsave); 390 - } 391 390 } 392 391 393 392 ··· 411 414 ((blitq->cur_blit_handle - handle) <= (1 << 23)); 412 415 413 416 if (queue && active) { 414 - slot = handle - blitq->done_blit_handle + blitq->cur -1; 415 - if (slot >= VIA_NUM_BLIT_SLOTS) { 417 + slot = handle - blitq->done_blit_handle + blitq->cur - 1; 418 + if (slot >= VIA_NUM_BLIT_SLOTS) 416 419 slot -= VIA_NUM_BLIT_SLOTS; 417 - } 418 420 *queue = blitq->blit_queue + slot; 419 421 } 420 422 ··· 502 506 int cur_released; 503 507 504 508 505 - DRM_DEBUG("Workqueue task called for blit engine %ld\n",(unsigned long) 509 + DRM_DEBUG("Workqueue task called for blit engine %ld\n", (unsigned long) 506 510 (blitq - ((drm_via_private_t *)dev->dev_private)->blit_queues)); 507 511 508 512 spin_lock_irqsave(&blitq->blit_lock, irqsave); 509 513 510 - while(blitq->serviced != blitq->cur) { 514 + while (blitq->serviced != blitq->cur) { 511 515 512 516 cur_released = blitq->serviced++; 513 517 ··· 541 545 void 542 546 via_init_dmablit(struct drm_device *dev) 543 547 { 544 - int i,j; 548 + int i, j; 545 549 drm_via_private_t *dev_priv = (drm_via_private_t *)dev->dev_private; 546 550 drm_via_blitq_t *blitq; 547 551 548 552 pci_set_master(dev->pdev); 549 553 550 - for (i=0; i< VIA_NUM_BLIT_ENGINES; ++i) { 554 + for (i = 0; i < VIA_NUM_BLIT_ENGINES; ++i) { 551 555 blitq = dev_priv->blit_queues + i; 552 556 blitq->dev = dev; 553 557 blitq->cur_blit_handle = 0; ··· 560 564 blitq->is_active = 0; 561 565 blitq->aborting = 0; 562 566 spin_lock_init(&blitq->blit_lock); 563 - for (j=0; j<VIA_NUM_BLIT_SLOTS; ++j) { 567 + for (j = 0; j < VIA_NUM_BLIT_SLOTS; ++j) 564 568 DRM_INIT_WAITQUEUE(blitq->blit_queue + j); 565 - } 566 569 DRM_INIT_WAITQUEUE(&blitq->busy_queue); 567 570 INIT_WORK(&blitq->wq, via_dmablit_workqueue); 568 571 setup_timer(&blitq->poll_timer, via_dmablit_timer, ··· 680 685 static int 681 686 via_dmablit_grab_slot(drm_via_blitq_t *blitq, int engine) 682 687 { 683 - int ret=0; 688 + int ret = 0; 684 689 unsigned long irqsave; 685 690 686 691 DRM_DEBUG("Num free is %d\n", blitq->num_free); 687 692 spin_lock_irqsave(&blitq->blit_lock, irqsave); 688 - while(blitq->num_free == 0) { 693 + while (blitq->num_free == 0) { 689 694 spin_unlock_irqrestore(&blitq->blit_lock, irqsave); 690 695 691 696 DRM_WAIT_ON(ret, blitq->busy_queue, DRM_HZ, blitq->num_free > 0); 692 - if (ret) { 697 + if (ret) 693 698 return (-EINTR == ret) ? -EAGAIN : ret; 694 - } 695 699 696 700 spin_lock_irqsave(&blitq->blit_lock, irqsave); 697 701 } ··· 713 719 spin_lock_irqsave(&blitq->blit_lock, irqsave); 714 720 blitq->num_free++; 715 721 spin_unlock_irqrestore(&blitq->blit_lock, irqsave); 716 - DRM_WAKEUP( &blitq->busy_queue ); 722 + DRM_WAKEUP(&blitq->busy_queue); 717 723 } 718 724 719 725 /* ··· 738 744 739 745 engine = (xfer->to_fb) ? 0 : 1; 740 746 blitq = dev_priv->blit_queues + engine; 741 - if (0 != (ret = via_dmablit_grab_slot(blitq, engine))) { 747 + if (0 != (ret = via_dmablit_grab_slot(blitq, engine))) 742 748 return ret; 743 - } 744 749 if (NULL == (vsg = kmalloc(sizeof(*vsg), GFP_KERNEL))) { 745 750 via_dmablit_release_slot(blitq); 746 751 return -ENOMEM; ··· 773 780 */ 774 781 775 782 int 776 - via_dma_blit_sync( struct drm_device *dev, void *data, struct drm_file *file_priv ) 783 + via_dma_blit_sync(struct drm_device *dev, void *data, struct drm_file *file_priv) 777 784 { 778 785 drm_via_blitsync_t *sync = data; 779 786 int err; ··· 797 804 */ 798 805 799 806 int 800 - via_dma_blit( struct drm_device *dev, void *data, struct drm_file *file_priv ) 807 + via_dma_blit(struct drm_device *dev, void *data, struct drm_file *file_priv) 801 808 { 802 809 drm_via_dmablit_t *xfer = data; 803 810 int err;
+4 -4
drivers/gpu/drm/via/via_dmablit.h
··· 45 45 int num_desc; 46 46 enum dma_data_direction direction; 47 47 unsigned char *bounce_buffer; 48 - dma_addr_t chain_start; 48 + dma_addr_t chain_start; 49 49 uint32_t free_on_sequence; 50 - unsigned int descriptors_per_page; 50 + unsigned int descriptors_per_page; 51 51 int aborted; 52 52 enum { 53 - dr_via_device_mapped, 53 + dr_via_device_mapped, 54 54 dr_via_desc_pages_alloc, 55 55 dr_via_pages_locked, 56 56 dr_via_pages_alloc, ··· 68 68 unsigned num_free; 69 69 unsigned num_outstanding; 70 70 unsigned long end; 71 - int aborting; 71 + int aborting; 72 72 int is_active; 73 73 drm_via_sg_info_t *blits[VIA_NUM_BLIT_SLOTS]; 74 74 spinlock_t blit_lock;
+11 -11
drivers/gpu/drm/via/via_drv.h
··· 107 107 #define VIA_BASE ((dev_priv->mmio)) 108 108 109 109 #define VIA_READ(reg) DRM_READ32(VIA_BASE, reg) 110 - #define VIA_WRITE(reg,val) DRM_WRITE32(VIA_BASE, reg, val) 110 + #define VIA_WRITE(reg, val) DRM_WRITE32(VIA_BASE, reg, val) 111 111 #define VIA_READ8(reg) DRM_READ8(VIA_BASE, reg) 112 - #define VIA_WRITE8(reg,val) DRM_WRITE8(VIA_BASE, reg, val) 112 + #define VIA_WRITE8(reg, val) DRM_WRITE8(VIA_BASE, reg, val) 113 113 114 114 extern struct drm_ioctl_desc via_ioctls[]; 115 115 extern int via_max_ioctl; ··· 121 121 extern int via_map_init(struct drm_device *dev, void *data, struct drm_file *file_priv); 122 122 extern int via_decoder_futex(struct drm_device *dev, void *data, struct drm_file *file_priv); 123 123 extern int via_wait_irq(struct drm_device *dev, void *data, struct drm_file *file_priv); 124 - extern int via_dma_blit_sync( struct drm_device *dev, void *data, struct drm_file *file_priv ); 125 - extern int via_dma_blit( struct drm_device *dev, void *data, struct drm_file *file_priv ); 124 + extern int via_dma_blit_sync(struct drm_device *dev, void *data, struct drm_file *file_priv); 125 + extern int via_dma_blit(struct drm_device *dev, void *data, struct drm_file *file_priv); 126 126 127 127 extern int via_driver_load(struct drm_device *dev, unsigned long chipset); 128 128 extern int via_driver_unload(struct drm_device *dev); 129 129 130 - extern int via_init_context(struct drm_device * dev, int context); 131 - extern int via_final_context(struct drm_device * dev, int context); 130 + extern int via_init_context(struct drm_device *dev, int context); 131 + extern int via_final_context(struct drm_device *dev, int context); 132 132 133 - extern int via_do_cleanup_map(struct drm_device * dev); 133 + extern int via_do_cleanup_map(struct drm_device *dev); 134 134 extern u32 via_get_vblank_counter(struct drm_device *dev, int crtc); 135 135 extern int via_enable_vblank(struct drm_device *dev, int crtc); 136 136 extern void via_disable_vblank(struct drm_device *dev, int crtc); 137 137 138 138 extern irqreturn_t via_driver_irq_handler(DRM_IRQ_ARGS); 139 - extern void via_driver_irq_preinstall(struct drm_device * dev); 139 + extern void via_driver_irq_preinstall(struct drm_device *dev); 140 140 extern int via_driver_irq_postinstall(struct drm_device *dev); 141 - extern void via_driver_irq_uninstall(struct drm_device * dev); 141 + extern void via_driver_irq_uninstall(struct drm_device *dev); 142 142 143 - extern int via_dma_cleanup(struct drm_device * dev); 143 + extern int via_dma_cleanup(struct drm_device *dev); 144 144 extern void via_init_command_verifier(void); 145 - extern int via_driver_dma_quiescent(struct drm_device * dev); 145 + extern int via_driver_dma_quiescent(struct drm_device *dev); 146 146 extern void via_init_futex(drm_via_private_t *dev_priv); 147 147 extern void via_cleanup_futex(drm_via_private_t *dev_priv); 148 148 extern void via_release_futex(drm_via_private_t *dev_priv, int context);
+6 -7
drivers/gpu/drm/via/via_irq.c
··· 141 141 atomic_inc(&cur_irq->irq_received); 142 142 DRM_WAKEUP(&cur_irq->irq_queue); 143 143 handled = 1; 144 - if (dev_priv->irq_map[drm_via_irq_dma0_td] == i) { 144 + if (dev_priv->irq_map[drm_via_irq_dma0_td] == i) 145 145 via_dmablit_handler(dev, 0, 1); 146 - } else if (dev_priv->irq_map[drm_via_irq_dma1_td] == i) { 146 + else if (dev_priv->irq_map[drm_via_irq_dma1_td] == i) 147 147 via_dmablit_handler(dev, 1, 1); 148 - } 149 148 } 150 149 cur_irq++; 151 150 } ··· 159 160 return IRQ_NONE; 160 161 } 161 162 162 - static __inline__ void viadrv_acknowledge_irqs(drm_via_private_t * dev_priv) 163 + static __inline__ void viadrv_acknowledge_irqs(drm_via_private_t *dev_priv) 163 164 { 164 165 u32 status; 165 166 ··· 206 207 } 207 208 208 209 static int 209 - via_driver_irq_wait(struct drm_device * dev, unsigned int irq, int force_sequence, 210 + via_driver_irq_wait(struct drm_device *dev, unsigned int irq, int force_sequence, 210 211 unsigned int *sequence) 211 212 { 212 213 drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private; ··· 259 260 * drm_dma.h hooks 260 261 */ 261 262 262 - void via_driver_irq_preinstall(struct drm_device * dev) 263 + void via_driver_irq_preinstall(struct drm_device *dev) 263 264 { 264 265 drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private; 265 266 u32 status; ··· 328 329 return 0; 329 330 } 330 331 331 - void via_driver_irq_uninstall(struct drm_device * dev) 332 + void via_driver_irq_uninstall(struct drm_device *dev) 332 333 { 333 334 drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private; 334 335 u32 status;
+2 -2
drivers/gpu/drm/via/via_map.c
··· 25 25 #include "via_drm.h" 26 26 #include "via_drv.h" 27 27 28 - static int via_do_init_map(struct drm_device * dev, drm_via_init_t * init) 28 + static int via_do_init_map(struct drm_device *dev, drm_via_init_t *init) 29 29 { 30 30 drm_via_private_t *dev_priv = dev->dev_private; 31 31 ··· 68 68 return 0; 69 69 } 70 70 71 - int via_do_cleanup_map(struct drm_device * dev) 71 + int via_do_cleanup_map(struct drm_device *dev) 72 72 { 73 73 via_dma_cleanup(dev); 74 74
+3 -4
drivers/gpu/drm/via/via_mm.c
··· 31 31 #include "drm_sman.h" 32 32 33 33 #define VIA_MM_ALIGN_SHIFT 4 34 - #define VIA_MM_ALIGN_MASK ( (1 << VIA_MM_ALIGN_SHIFT) - 1) 34 + #define VIA_MM_ALIGN_MASK ((1 << VIA_MM_ALIGN_SHIFT) - 1) 35 35 36 36 int via_agp_init(struct drm_device *dev, void *data, struct drm_file *file_priv) 37 37 { ··· 172 172 } 173 173 174 174 175 - void via_reclaim_buffers_locked(struct drm_device * dev, 175 + void via_reclaim_buffers_locked(struct drm_device *dev, 176 176 struct drm_file *file_priv) 177 177 { 178 178 drm_via_private_t *dev_priv = dev->dev_private; ··· 183 183 return; 184 184 } 185 185 186 - if (dev->driver->dma_quiescent) { 186 + if (dev->driver->dma_quiescent) 187 187 dev->driver->dma_quiescent(dev); 188 - } 189 188 190 189 drm_sman_owner_cleanup(&dev_priv->sman, (unsigned long)file_priv); 191 190 mutex_unlock(&dev->struct_mutex);
+21 -26
drivers/gpu/drm/via/via_verifier.c
··· 235 235 static hazard_t table3[256]; 236 236 237 237 static __inline__ int 238 - eat_words(const uint32_t ** buf, const uint32_t * buf_end, unsigned num_words) 238 + eat_words(const uint32_t **buf, const uint32_t *buf_end, unsigned num_words) 239 239 { 240 240 if ((buf_end - *buf) >= num_words) { 241 241 *buf += num_words; ··· 252 252 static __inline__ drm_local_map_t *via_drm_lookup_agp_map(drm_via_state_t *seq, 253 253 unsigned long offset, 254 254 unsigned long size, 255 - struct drm_device * dev) 255 + struct drm_device *dev) 256 256 { 257 257 struct drm_map_list *r_list; 258 258 drm_local_map_t *map = seq->map_cache; ··· 344 344 } 345 345 346 346 static __inline__ int 347 - investigate_hazard(uint32_t cmd, hazard_t hz, drm_via_state_t * cur_seq) 347 + investigate_hazard(uint32_t cmd, hazard_t hz, drm_via_state_t *cur_seq) 348 348 { 349 349 register uint32_t tmp, *tmp_addr; 350 350 ··· 518 518 519 519 static __inline__ int 520 520 via_check_prim_list(uint32_t const **buffer, const uint32_t * buf_end, 521 - drm_via_state_t * cur_seq) 521 + drm_via_state_t *cur_seq) 522 522 { 523 523 drm_via_private_t *dev_priv = 524 524 (drm_via_private_t *) cur_seq->dev->dev_private; ··· 621 621 } 622 622 623 623 static __inline__ verifier_state_t 624 - via_check_header2(uint32_t const **buffer, const uint32_t * buf_end, 625 - drm_via_state_t * hc_state) 624 + via_check_header2(uint32_t const **buffer, const uint32_t *buf_end, 625 + drm_via_state_t *hc_state) 626 626 { 627 627 uint32_t cmd; 628 628 int hz_mode; ··· 706 706 return state_error; 707 707 } 708 708 } 709 - if (hc_state->unfinished && finish_current_sequence(hc_state)) { 709 + if (hc_state->unfinished && finish_current_sequence(hc_state)) 710 710 return state_error; 711 - } 712 711 *buffer = buf; 713 712 return state_command; 714 713 } 715 714 716 715 static __inline__ verifier_state_t 717 - via_parse_header2(drm_via_private_t * dev_priv, uint32_t const **buffer, 718 - const uint32_t * buf_end, int *fire_count) 716 + via_parse_header2(drm_via_private_t *dev_priv, uint32_t const **buffer, 717 + const uint32_t *buf_end, int *fire_count) 719 718 { 720 719 uint32_t cmd; 721 720 const uint32_t *buf = *buffer; ··· 832 833 } 833 834 834 835 static __inline__ verifier_state_t 835 - via_parse_header1(drm_via_private_t * dev_priv, uint32_t const **buffer, 836 - const uint32_t * buf_end) 836 + via_parse_header1(drm_via_private_t *dev_priv, uint32_t const **buffer, 837 + const uint32_t *buf_end) 837 838 { 838 839 register uint32_t cmd; 839 840 const uint32_t *buf = *buffer; ··· 850 851 } 851 852 852 853 static __inline__ verifier_state_t 853 - via_check_vheader5(uint32_t const **buffer, const uint32_t * buf_end) 854 + via_check_vheader5(uint32_t const **buffer, const uint32_t *buf_end) 854 855 { 855 856 uint32_t data; 856 857 const uint32_t *buf = *buffer; ··· 883 884 } 884 885 885 886 static __inline__ verifier_state_t 886 - via_parse_vheader5(drm_via_private_t * dev_priv, uint32_t const **buffer, 887 - const uint32_t * buf_end) 887 + via_parse_vheader5(drm_via_private_t *dev_priv, uint32_t const **buffer, 888 + const uint32_t *buf_end) 888 889 { 889 890 uint32_t addr, count, i; 890 891 const uint32_t *buf = *buffer; ··· 892 893 addr = *buf++ & ~VIA_VIDEOMASK; 893 894 i = count = *buf; 894 895 buf += 3; 895 - while (i--) { 896 + while (i--) 896 897 VIA_WRITE(addr, *buf++); 897 - } 898 898 if (count & 3) 899 899 buf += 4 - (count & 3); 900 900 *buffer = buf; ··· 938 940 } 939 941 940 942 static __inline__ verifier_state_t 941 - via_parse_vheader6(drm_via_private_t * dev_priv, uint32_t const **buffer, 942 - const uint32_t * buf_end) 943 + via_parse_vheader6(drm_via_private_t *dev_priv, uint32_t const **buffer, 944 + const uint32_t *buf_end) 943 945 { 944 946 945 947 uint32_t addr, count, i; ··· 1035 1037 } 1036 1038 1037 1039 int 1038 - via_parse_command_stream(struct drm_device * dev, const uint32_t * buf, 1040 + via_parse_command_stream(struct drm_device *dev, const uint32_t *buf, 1039 1041 unsigned int size) 1040 1042 { 1041 1043 ··· 1083 1085 return -EINVAL; 1084 1086 } 1085 1087 } 1086 - if (state == state_error) { 1088 + if (state == state_error) 1087 1089 return -EINVAL; 1088 - } 1089 1090 return 0; 1090 1091 } 1091 1092 ··· 1093 1096 { 1094 1097 int i; 1095 1098 1096 - for (i = 0; i < 256; ++i) { 1099 + for (i = 0; i < 256; ++i) 1097 1100 table[i] = forbidden_command; 1098 - } 1099 1101 1100 - for (i = 0; i < size; ++i) { 1102 + for (i = 0; i < size; ++i) 1101 1103 table[init_table[i].code] = init_table[i].hz; 1102 - } 1103 1104 } 1104 1105 1105 1106 void via_init_command_verifier(void)
+2 -2
drivers/gpu/drm/via/via_verifier.h
··· 54 54 const uint32_t *buf_start; 55 55 } drm_via_state_t; 56 56 57 - extern int via_verify_command_stream(const uint32_t * buf, unsigned int size, 58 - struct drm_device * dev, int agp); 57 + extern int via_verify_command_stream(const uint32_t *buf, unsigned int size, 58 + struct drm_device *dev, int agp); 59 59 extern int via_parse_command_stream(struct drm_device *dev, const uint32_t *buf, 60 60 unsigned int size); 61 61
+3 -3
drivers/gpu/drm/via/via_video.c
··· 29 29 #include "via_drm.h" 30 30 #include "via_drv.h" 31 31 32 - void via_init_futex(drm_via_private_t * dev_priv) 32 + void via_init_futex(drm_via_private_t *dev_priv) 33 33 { 34 34 unsigned int i; 35 35 ··· 41 41 } 42 42 } 43 43 44 - void via_cleanup_futex(drm_via_private_t * dev_priv) 44 + void via_cleanup_futex(drm_via_private_t *dev_priv) 45 45 { 46 46 } 47 47 48 - void via_release_futex(drm_via_private_t * dev_priv, int context) 48 + void via_release_futex(drm_via_private_t *dev_priv, int context) 49 49 { 50 50 unsigned int i; 51 51 volatile int *lock;