Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

[AGPGART] Lots of CodingStyle/whitespace cleanups.

Eliminate trailing whitespace.
s/if(/if (/
s/for(/for (/

Signed-off-by: Dave Jones <davej@redhat.com>

+98 -105
+1 -1
drivers/char/agp/Kconfig
··· 98 98 tristate "Serverworks LE/HE chipset support" 99 99 depends on AGP && X86_32 100 100 help 101 - Say Y here to support the Serverworks AGP card. See 101 + Say Y here to support the Serverworks AGP card. See 102 102 <http://www.serverworks.com/> for product descriptions and images. 103 103 104 104 config AGP_VIA
+5 -6
drivers/char/agp/agp.h
··· 19 19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS 20 20 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 21 21 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 22 - * JEFF HARTMANN, OR ANY OTHER CONTRIBUTORS BE LIABLE FOR ANY CLAIM, 23 - * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 - * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE 22 + * JEFF HARTMANN, OR ANY OTHER CONTRIBUTORS BE LIABLE FOR ANY CLAIM, 23 + * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 + * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE 25 25 * OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 26 26 * 27 27 */ ··· 53 53 struct gatt_mask { 54 54 unsigned long mask; 55 55 u32 type; 56 - /* totally device specific, for integrated chipsets that 56 + /* totally device specific, for integrated chipsets that 57 57 * might have different types of memory masks. For other 58 58 * devices this will probably be ignored */ 59 59 }; ··· 104 104 void (*agp_enable)(struct agp_bridge_data *, u32); 105 105 void (*cleanup)(void); 106 106 void (*tlb_flush)(struct agp_memory *); 107 - unsigned long (*mask_memory)(struct agp_bridge_data *, 108 - unsigned long, int); 107 + unsigned long (*mask_memory)(struct agp_bridge_data *, unsigned long, int); 109 108 void (*cache_flush)(void); 110 109 int (*create_gatt_table)(struct agp_bridge_data *); 111 110 int (*free_gatt_table)(struct agp_bridge_data *);
+1 -1
drivers/char/agp/ali-agp.c
··· 147 147 148 148 if (!addr) 149 149 return NULL; 150 - 150 + 151 151 pci_read_config_dword(agp_bridge->dev, ALI_CACHE_FLUSH_CTRL, &temp); 152 152 pci_write_config_dword(agp_bridge->dev, ALI_CACHE_FLUSH_CTRL, 153 153 (((temp & ALI_CACHE_FLUSH_ADDR_MASK) |
+8 -6
drivers/char/agp/alpha-agp.c
··· 23 23 dma_addr = address - vma->vm_start + agp->aperture.bus_base; 24 24 pa = agp->ops->translate(agp, dma_addr); 25 25 26 - if (pa == (unsigned long)-EINVAL) return NULL; /* no translation */ 27 - 26 + if (pa == (unsigned long)-EINVAL) 27 + return NULL; /* no translation */ 28 + 28 29 /* 29 30 * Get the page, inc the use count, and return it 30 31 */ ··· 90 89 agp_device_command(agp->mode.lw, 0); 91 90 } 92 91 93 - static int alpha_core_agp_insert_memory(struct agp_memory *mem, off_t pg_start, 92 + static int alpha_core_agp_insert_memory(struct agp_memory *mem, off_t pg_start, 94 93 int type) 95 94 { 96 95 alpha_agp_info *agp = agp_bridge->dev_private_data; ··· 99 98 100 99 temp = agp_bridge->current_size; 101 100 num_entries = A_SIZE_FIX(temp)->num_entries; 102 - if ((pg_start + mem->page_count) > num_entries) return -EINVAL; 101 + if ((pg_start + mem->page_count) > num_entries) 102 + return -EINVAL; 103 103 104 104 status = agp->ops->bind(agp, pg_start, mem); 105 105 mb(); ··· 109 107 return status; 110 108 } 111 109 112 - static int alpha_core_agp_remove_memory(struct agp_memory *mem, off_t pg_start, 110 + static int alpha_core_agp_remove_memory(struct agp_memory *mem, off_t pg_start, 113 111 int type) 114 112 { 115 113 alpha_agp_info *agp = agp_bridge->dev_private_data; ··· 127 125 .size_type = FIXED_APER_SIZE, 128 126 .cant_use_aperture = 1, 129 127 .masks = NULL, 130 - 128 + 131 129 .fetch_size = alpha_core_agp_fetch_size, 132 130 .configure = alpha_core_agp_configure, 133 131 .agp_enable = alpha_core_agp_enable,
+1 -1
drivers/char/agp/amd64-agp.c
··· 216 216 {256, 65536, 6, 0x00000700 }, /* 1 1 1 0 0 0 */ 217 217 {128, 32768, 5, 0x00000720 }, /* 1 1 1 1 0 0 */ 218 218 {64, 16384, 4, 0x00000730 }, /* 1 1 1 1 1 0 */ 219 - {32, 8192, 3, 0x00000738 } /* 1 1 1 1 1 1 */ 219 + {32, 8192, 3, 0x00000738 } /* 1 1 1 1 1 1 */ 220 220 }; 221 221 222 222 static int amd_8151_configure(void)
+3 -3
drivers/char/agp/ati-agp.c
··· 74 74 /*CACHE_FLUSH();*/ 75 75 global_cache_flush(); 76 76 77 - for(i = 0; i < PAGE_SIZE / sizeof(unsigned long); i++) { 77 + for (i = 0; i < PAGE_SIZE / sizeof(unsigned long); i++) { 78 78 writel(agp_bridge->scratch_page, page_map->remapped+i); 79 79 readl(page_map->remapped+i); /* PCI Posting. */ 80 80 } ··· 99 99 ati_page_map *entry; 100 100 101 101 tables = ati_generic_private.gatt_pages; 102 - for(i = 0; i < ati_generic_private.num_tables; i++) { 102 + for (i = 0; i < ati_generic_private.num_tables; i++) { 103 103 entry = tables[i]; 104 104 if (entry != NULL) { 105 105 if (entry->real != NULL) ··· 387 387 agp_bridge->gart_bus_addr = addr; 388 388 389 389 /* Calculate the agp offset */ 390 - for(i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) { 390 + for (i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) { 391 391 writel(virt_to_gart(ati_generic_private.gatt_pages[i]->real) | 1, 392 392 page_dir.remapped+GET_PAGE_DIR_OFF(addr)); 393 393 readl(page_dir.remapped+GET_PAGE_DIR_OFF(addr)); /* PCI Posting. */
+1 -1
drivers/char/agp/backend.c
··· 228 228 struct agp_bridge_data *agp_alloc_bridge(void) 229 229 { 230 230 struct agp_bridge_data *bridge; 231 - 231 + 232 232 bridge = kzalloc(sizeof(*bridge), GFP_KERNEL); 233 233 if (!bridge) 234 234 return NULL;
+8 -8
drivers/char/agp/efficeon-agp.c
··· 1 1 /* 2 2 * Transmeta's Efficeon AGPGART driver. 3 - * 3 + * 4 4 * Based upon a diff by Linus around November '02. 5 5 * 6 6 * Ported to the 2.6 kernel by Carlos Puchol <cpglinux@puchol.com> ··· 9 9 10 10 /* 11 11 * NOTE-cpg-040217: 12 - * 12 + * 13 13 * - when compiled as a module, after loading the module, 14 14 * it will refuse to unload, indicating it is in use, 15 15 * when it is not. ··· 45 45 * 8: Present 46 46 * 7:6: reserved, write as zero 47 47 * 5:0: GATT directory index: which 1st-level entry 48 - * 48 + * 49 49 * The Efficeon AGP spec requires pages to be WB-cacheable 50 50 * but to be explicitly CLFLUSH'd after any changes. 51 51 */ ··· 125 125 struct aper_size_info_lvl2 *current_size; 126 126 127 127 printk(KERN_DEBUG PFX "efficeon_configure()\n"); 128 - 128 + 129 129 current_size = A_SIZE_LVL2(agp_bridge->current_size); 130 130 131 131 /* aperture size */ ··· 190 190 const int present = EFFICEON_PRESENT; 191 191 const int clflush_chunk = ((cpuid_ebx(1) >> 8) & 0xff) << 3; 192 192 int num_entries, l1_pages; 193 - 193 + 194 194 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; 195 195 196 196 printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries); ··· 257 257 258 258 if (!page) 259 259 continue; 260 - 260 + 261 261 page += (index & 0x3ff); 262 262 *page = insert; 263 263 264 264 /* clflush is slow, so don't clflush until we have to */ 265 - if ( last_page && 265 + if ( last_page && 266 266 ((unsigned long)page^(unsigned long)last_page) & clflush_mask ) 267 267 asm volatile("clflush %0" : : "m" (*last_page)); 268 268 ··· 373 373 */ 374 374 r = &pdev->resource[0]; 375 375 if (!r->start && r->end) { 376 - if(pci_assign_resource(pdev, 0)) { 376 + if (pci_assign_resource(pdev, 0)) { 377 377 printk(KERN_ERR PFX "could not assign resource 0\n"); 378 378 return -ENODEV; 379 379 }
+1 -1
drivers/char/agp/frontend.c
··· 781 781 if (agp_fe.current_controller != NULL) 782 782 return -EBUSY; 783 783 784 - if(!agp_bridge) 784 + if (!agp_bridge) 785 785 return -ENODEV; 786 786 787 787 if (atomic_read(&agp_bridge->agp_in_use))
+1 -1
drivers/char/agp/generic.c
··· 299 299 /** 300 300 * agp_copy_info - copy bridge state information 301 301 * 302 - * @info: agp_kern_info pointer. The caller should insure that this pointer is valid. 302 + * @info: agp_kern_info pointer. The caller should insure that this pointer is valid. 303 303 * 304 304 * This function copies information about the agp bridge device and the state of 305 305 * the agp backend into an agp_kern_info pointer.
+3 -3
drivers/char/agp/hp-agp.c
··· 85 85 /* 86 86 * IOC already configured by sba_iommu module; just use 87 87 * its setup. We assume: 88 - * - IOVA space is 1Gb in size 89 - * - first 512Mb is IOMMU, second 512Mb is GART 88 + * - IOVA space is 1Gb in size 89 + * - first 512Mb is IOMMU, second 512Mb is GART 90 90 */ 91 91 hp->io_tlb_ps = readq(hp->ioc_regs+HP_ZX1_TCNFG); 92 92 switch (hp->io_tlb_ps) { ··· 115 115 116 116 if (hp->gatt[0] != HP_ZX1_SBA_IOMMU_COOKIE) { 117 117 /* Normal case when no AGP device in system */ 118 - hp->gatt = NULL; 118 + hp->gatt = NULL; 119 119 hp->gatt_entries = 0; 120 120 printk(KERN_ERR PFX "No reserved IO PDIR entry found; " 121 121 "GART disabled\n");
+8 -8
drivers/char/agp/i460-agp.c
··· 400 400 num_entries = A_SIZE_8(temp)->num_entries; 401 401 402 402 /* Figure out what pg_start means in terms of our large GART pages */ 403 - start = &i460.lp_desc[pg_start / I460_KPAGES_PER_IOPAGE]; 404 - end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 405 - start_offset = pg_start % I460_KPAGES_PER_IOPAGE; 406 - end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; 403 + start = &i460.lp_desc[pg_start / I460_KPAGES_PER_IOPAGE]; 404 + end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 405 + start_offset = pg_start % I460_KPAGES_PER_IOPAGE; 406 + end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; 407 407 408 408 if (end > i460.lp_desc + num_entries) { 409 409 printk(KERN_ERR PFX "Looks like we're out of AGP memory\n"); ··· 458 458 num_entries = A_SIZE_8(temp)->num_entries; 459 459 460 460 /* Figure out what pg_start means in terms of our large GART pages */ 461 - start = &i460.lp_desc[pg_start / I460_KPAGES_PER_IOPAGE]; 462 - end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 463 - start_offset = pg_start % I460_KPAGES_PER_IOPAGE; 464 - end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; 461 + start = &i460.lp_desc[pg_start / I460_KPAGES_PER_IOPAGE]; 462 + end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; 463 + start_offset = pg_start % I460_KPAGES_PER_IOPAGE; 464 + end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; 465 465 466 466 for (i = 0, lp = start; lp <= end; ++lp) { 467 467 for (idx = ((lp == start) ? start_offset : 0);
+19 -26
drivers/char/agp/intel-agp.c
··· 201 201 temp = agp_bridge->current_size; 202 202 num_entries = A_SIZE_FIX(temp)->num_entries; 203 203 204 - if ((pg_start + mem->page_count) > num_entries) { 204 + if ((pg_start + mem->page_count) > num_entries) 205 205 return -EINVAL; 206 - } 206 + 207 207 for (j = pg_start; j < (pg_start + mem->page_count); j++) { 208 208 if (!PGE_EMPTY(agp_bridge, readl(agp_bridge->gatt_table+j))) 209 209 return -EBUSY; ··· 221 221 agp_bridge->driver->tlb_flush(mem); 222 222 return 0; 223 223 } 224 - if((type == AGP_PHYS_MEMORY) && (mem->type == AGP_PHYS_MEMORY)) 224 + if ((type == AGP_PHYS_MEMORY) && (mem->type == AGP_PHYS_MEMORY)) 225 225 goto insert; 226 226 return -EINVAL; 227 227 } ··· 328 328 static void intel_i810_free_by_type(struct agp_memory *curr) 329 329 { 330 330 agp_free_key(curr->key); 331 - if(curr->type == AGP_PHYS_MEMORY) { 331 + if (curr->type == AGP_PHYS_MEMORY) { 332 332 if (curr->page_count == 4) 333 333 i8xx_destroy_pages(gart_to_virt(curr->memory[0])); 334 334 else { ··· 1603 1603 name = "i820"; 1604 1604 break; 1605 1605 case PCI_DEVICE_ID_INTEL_82830_HB: 1606 - if (find_i830(PCI_DEVICE_ID_INTEL_82830_CGC)) { 1606 + if (find_i830(PCI_DEVICE_ID_INTEL_82830_CGC)) 1607 1607 bridge->driver = &intel_830_driver; 1608 - } else { 1608 + else 1609 1609 bridge->driver = &intel_830mp_driver; 1610 - } 1611 1610 name = "830M"; 1612 1611 break; 1613 1612 case PCI_DEVICE_ID_INTEL_82840_HB: ··· 1618 1619 name = "i845"; 1619 1620 break; 1620 1621 case PCI_DEVICE_ID_INTEL_82845G_HB: 1621 - if (find_i830(PCI_DEVICE_ID_INTEL_82845G_IG)) { 1622 + if (find_i830(PCI_DEVICE_ID_INTEL_82845G_IG)) 1622 1623 bridge->driver = &intel_830_driver; 1623 - } else { 1624 + else 1624 1625 bridge->driver = &intel_845_driver; 1625 - } 1626 1626 name = "845G"; 1627 1627 break; 1628 1628 case PCI_DEVICE_ID_INTEL_82850_HB: ··· 1646 1648 name = "i860"; 1647 1649 break; 1648 1650 case PCI_DEVICE_ID_INTEL_82865_HB: 1649 - if (find_i830(PCI_DEVICE_ID_INTEL_82865_IG)) { 1651 + if (find_i830(PCI_DEVICE_ID_INTEL_82865_IG)) 1650 1652 bridge->driver = &intel_830_driver; 1651 - } else { 1653 + else 1652 1654 bridge->driver = &intel_845_driver; 1653 - } 1654 1655 name = "865"; 1655 1656 break; 1656 1657 case PCI_DEVICE_ID_INTEL_82875_HB: ··· 1657 1660 name = "i875"; 1658 1661 break; 1659 1662 case PCI_DEVICE_ID_INTEL_82915G_HB: 1660 - if (find_i830(PCI_DEVICE_ID_INTEL_82915G_IG)) { 1663 + if (find_i830(PCI_DEVICE_ID_INTEL_82915G_IG)) 1661 1664 bridge->driver = &intel_915_driver; 1662 - } else { 1665 + else 1663 1666 bridge->driver = &intel_845_driver; 1664 - } 1665 1667 name = "915G"; 1666 1668 break; 1667 1669 case PCI_DEVICE_ID_INTEL_82915GM_HB: 1668 - if (find_i830(PCI_DEVICE_ID_INTEL_82915GM_IG)) { 1670 + if (find_i830(PCI_DEVICE_ID_INTEL_82915GM_IG)) 1669 1671 bridge->driver = &intel_915_driver; 1670 - } else { 1672 + else 1671 1673 bridge->driver = &intel_845_driver; 1672 - } 1673 1674 name = "915GM"; 1674 1675 break; 1675 1676 case PCI_DEVICE_ID_INTEL_82945G_HB: 1676 - if (find_i830(PCI_DEVICE_ID_INTEL_82945G_IG)) { 1677 + if (find_i830(PCI_DEVICE_ID_INTEL_82945G_IG)) 1677 1678 bridge->driver = &intel_915_driver; 1678 - } else { 1679 + else 1679 1680 bridge->driver = &intel_845_driver; 1680 - } 1681 1681 name = "945G"; 1682 1682 break; 1683 1683 case PCI_DEVICE_ID_INTEL_82945GM_HB: 1684 - if (find_i830(PCI_DEVICE_ID_INTEL_82945GM_IG)) { 1684 + if (find_i830(PCI_DEVICE_ID_INTEL_82945GM_IG)) 1685 1685 bridge->driver = &intel_915_driver; 1686 - } else { 1686 + else 1687 1687 bridge->driver = &intel_845_driver; 1688 - } 1689 1688 name = "945GM"; 1690 1689 break; 1691 1690 case PCI_DEVICE_ID_INTEL_7505_0: ··· 1717 1724 */ 1718 1725 r = &pdev->resource[0]; 1719 1726 if (!r->start && r->end) { 1720 - if(pci_assign_resource(pdev, 0)) { 1727 + if (pci_assign_resource(pdev, 0)) { 1721 1728 printk(KERN_ERR PFX "could not assign resource 0\n"); 1722 1729 agp_put_bridge(bridge); 1723 1730 return -ENODEV;
+10 -10
drivers/char/agp/isoch.c
··· 26 26 27 27 list_for_each(pos, head) { 28 28 cur = list_entry(pos, struct agp_3_5_dev, list); 29 - if(cur->maxbw > n->maxbw) 29 + if (cur->maxbw > n->maxbw) 30 30 break; 31 31 } 32 32 list_add_tail(new, pos); ··· 54 54 } 55 55 } 56 56 57 - /* 58 - * Initialize all isochronous transfer parameters for an AGP 3.0 59 - * node (i.e. a host bridge in combination with the adapters 57 + /* 58 + * Initialize all isochronous transfer parameters for an AGP 3.0 59 + * node (i.e. a host bridge in combination with the adapters 60 60 * lying behind it...) 61 61 */ 62 62 ··· 200 200 * this to the hungriest device (as per the spec) */ 201 201 rem = target.n - tot_n; 202 202 203 - /* 203 + /* 204 204 * Calculate the minimum isochronous RQ depth needed by each master. 205 205 * Along the way, distribute the extra ISOCH_N capability calculated 206 206 * above. ··· 214 214 * many writes on the AGP bus). 215 215 */ 216 216 master[cdev].rq = master[cdev].n; 217 - if(master[cdev].y > 0x1) 217 + if (master[cdev].y > 0x1) 218 218 master[cdev].rq *= (1 << (master[cdev].y - 1)); 219 219 220 220 tot_rq += master[cdev].rq; ··· 334 334 335 335 arqsz = (tstatus >> 13) & 0x7; 336 336 337 - /* 337 + /* 338 338 * Allocate a head for our AGP 3.5 device list 339 - * (multiple AGP v3 devices are allowed behind a single bridge). 339 + * (multiple AGP v3 devices are allowed behind a single bridge). 340 340 */ 341 341 if ((dev_list = kmalloc(sizeof(*dev_list), GFP_KERNEL)) == NULL) { 342 342 ret = -ENOMEM; ··· 366 366 367 367 case 0x0300: /* Display controller */ 368 368 case 0x0400: /* Multimedia controller */ 369 - if((cur = kmalloc(sizeof(*cur), GFP_KERNEL)) == NULL) { 369 + if ((cur = kmalloc(sizeof(*cur), GFP_KERNEL)) == NULL) { 370 370 ret = -ENOMEM; 371 371 goto free_and_exit; 372 372 } ··· 391 391 list_for_each(pos, head) { 392 392 cur = list_entry(pos, struct agp_3_5_dev, list); 393 393 dev = cur->dev; 394 - 394 + 395 395 pci_read_config_word(dev, PCI_STATUS, &mpstat); 396 396 if ((mpstat & PCI_STATUS_CAP_LIST) == 0) 397 397 continue;
+10 -10
drivers/char/agp/nvidia-agp.c
··· 72 72 /* Find the iorr that is already used for the base */ 73 73 /* If not found, determine the uppermost available iorr */ 74 74 free_iorr_addr = AMD_K7_NUM_IORR; 75 - for(iorr_addr = 0; iorr_addr < AMD_K7_NUM_IORR; iorr_addr++) { 75 + for (iorr_addr = 0; iorr_addr < AMD_K7_NUM_IORR; iorr_addr++) { 76 76 rdmsr(IORR_BASE0 + 2 * iorr_addr, base_lo, base_hi); 77 77 rdmsr(IORR_MASK0 + 2 * iorr_addr, mask_lo, mask_hi); 78 78 ··· 82 82 if ((mask_lo & 0x00000800) == 0) 83 83 free_iorr_addr = iorr_addr; 84 84 } 85 - 85 + 86 86 if (iorr_addr >= AMD_K7_NUM_IORR) { 87 87 iorr_addr = free_iorr_addr; 88 88 if (iorr_addr >= AMD_K7_NUM_IORR) ··· 139 139 } 140 140 141 141 /* attbase */ 142 - for(i = 0; i < 8; i++) { 142 + for (i = 0; i < 8; i++) { 143 143 pci_write_config_dword(nvidia_private.dev_2, NVIDIA_2_ATTBASE(i), 144 144 (agp_bridge->gatt_bus_addr + (i % num_dirs) * 64 * 1024) | 1); 145 145 } ··· 197 197 static int nvidia_insert_memory(struct agp_memory *mem, off_t pg_start, int type) 198 198 { 199 199 int i, j; 200 - 200 + 201 201 if ((type != 0) || (mem->type != 0)) 202 202 return -EINVAL; 203 - 203 + 204 204 if ((pg_start + mem->page_count) > 205 205 (nvidia_private.num_active_entries - agp_memory_reserved/PAGE_SIZE)) 206 206 return -EINVAL; 207 - 208 - for(j = pg_start; j < (pg_start + mem->page_count); j++) { 207 + 208 + for (j = pg_start; j < (pg_start + mem->page_count); j++) { 209 209 if (!PGE_EMPTY(agp_bridge, readl(agp_bridge->gatt_table+nvidia_private.pg_offset+j))) 210 210 return -EBUSY; 211 211 } ··· 264 264 } 265 265 266 266 /* flush TLB entries */ 267 - for(i = 0; i < 32 + 1; i++) 267 + for (i = 0; i < 32 + 1; i++) 268 268 temp = readl(nvidia_private.aperture+(i * PAGE_SIZE / sizeof(u32))); 269 - for(i = 0; i < 32 + 1; i++) 269 + for (i = 0; i < 32 + 1; i++) 270 270 temp = readl(nvidia_private.aperture+(i * PAGE_SIZE / sizeof(u32))); 271 271 } 272 272 ··· 323 323 pci_find_slot((unsigned int)pdev->bus->number, PCI_DEVFN(0, 2)); 324 324 nvidia_private.dev_3 = 325 325 pci_find_slot((unsigned int)pdev->bus->number, PCI_DEVFN(30, 0)); 326 - 326 + 327 327 if (!nvidia_private.dev_1 || !nvidia_private.dev_2 || !nvidia_private.dev_3) { 328 328 printk(KERN_INFO PFX "Detected an NVIDIA nForce/nForce2 " 329 329 "chipset, but could not find the secondary devices.\n");
+1 -1
drivers/char/agp/sgi-agp.c
··· 329 329 330 330 static void __devexit agp_sgi_cleanup(void) 331 331 { 332 - if(sgi_tioca_agp_bridges) 332 + if (sgi_tioca_agp_bridges) 333 333 kfree(sgi_tioca_agp_bridges); 334 334 sgi_tioca_agp_bridges=NULL; 335 335 }
+4 -4
drivers/char/agp/sis-agp.c
··· 121 121 122 122 static struct agp_bridge_driver sis_driver = { 123 123 .owner = THIS_MODULE, 124 - .aperture_sizes = sis_generic_sizes, 124 + .aperture_sizes = sis_generic_sizes, 125 125 .size_type = U8_APER_SIZE, 126 126 .num_aperture_sizes = 7, 127 127 .configure = sis_configure, ··· 243 243 { 244 244 int i; 245 245 246 - for(i=0; sis_broken_chipsets[i]!=0; ++i) 247 - if(bridge->dev->device==sis_broken_chipsets[i]) 246 + for (i=0; sis_broken_chipsets[i]!=0; ++i) 247 + if (bridge->dev->device==sis_broken_chipsets[i]) 248 248 break; 249 249 250 - if(sis_broken_chipsets[i] || agp_sis_force_delay) 250 + if (sis_broken_chipsets[i] || agp_sis_force_delay) 251 251 sis_driver.agp_enable=sis_delayed_enable; 252 252 253 253 // sis chipsets that indicate less than agp3.5
+7 -8
drivers/char/agp/sworks-agp.c
··· 64 64 } 65 65 global_cache_flush(); 66 66 67 - for(i = 0; i < PAGE_SIZE / sizeof(unsigned long); i++) 67 + for (i = 0; i < PAGE_SIZE / sizeof(unsigned long); i++) 68 68 writel(agp_bridge->scratch_page, page_map->remapped+i); 69 69 70 70 return 0; ··· 84 84 struct serverworks_page_map *entry; 85 85 86 86 tables = serverworks_private.gatt_pages; 87 - for(i = 0; i < serverworks_private.num_tables; i++) { 87 + for (i = 0; i < serverworks_private.num_tables; i++) { 88 88 entry = tables[i]; 89 89 if (entry != NULL) { 90 90 if (entry->real != NULL) { ··· 103 103 int retval = 0; 104 104 int i; 105 105 106 - tables = kzalloc((nr_tables + 1) * sizeof(struct serverworks_page_map *), 106 + tables = kzalloc((nr_tables + 1) * sizeof(struct serverworks_page_map *), 107 107 GFP_KERNEL); 108 108 if (tables == NULL) 109 109 return -ENOMEM; ··· 161 161 return retval; 162 162 } 163 163 /* Create a fake scratch directory */ 164 - for(i = 0; i < 1024; i++) { 164 + for (i = 0; i < 1024; i++) { 165 165 writel(agp_bridge->scratch_page, serverworks_private.scratch_dir.remapped+i); 166 166 writel(virt_to_gart(serverworks_private.scratch_dir.real) | 1, page_dir.remapped+i); 167 167 } ··· 185 185 pci_read_config_dword(agp_bridge->dev,serverworks_private.gart_addr_ofs,&temp); 186 186 agp_bridge->gart_bus_addr = (temp & PCI_BASE_ADDRESS_MEM_MASK); 187 187 188 - /* Calculate the agp offset */ 189 - 190 - for(i = 0; i < value->num_entries / 1024; i++) 188 + /* Calculate the agp offset */ 189 + for (i = 0; i < value->num_entries / 1024; i++) 191 190 writel(virt_to_gart(serverworks_private.gatt_pages[i]->real)|1, page_dir.remapped+i); 192 191 193 192 return 0; ··· 195 196 static int serverworks_free_gatt_table(struct agp_bridge_data *bridge) 196 197 { 197 198 struct serverworks_page_map page_dir; 198 - 199 + 199 200 page_dir.real = (unsigned long *)agp_bridge->gatt_table_real; 200 201 page_dir.remapped = (unsigned long __iomem *)agp_bridge->gatt_table; 201 202
+6 -6
drivers/char/agp/uninorth-agp.c
··· 95 95 static int uninorth_configure(void) 96 96 { 97 97 struct aper_size_info_32 *current_size; 98 - 98 + 99 99 current_size = A_SIZE_32(agp_bridge->current_size); 100 100 101 101 printk(KERN_INFO PFX "configuring for size idx: %d\n", 102 102 current_size->size_value); 103 - 103 + 104 104 /* aperture size and gatt addr */ 105 105 pci_write_config_dword(agp_bridge->dev, 106 106 UNI_N_CFG_GART_BASE, ··· 127 127 UNI_N_CFG_GART_DUMMY_PAGE, 128 128 agp_bridge->scratch_page_real >> 12); 129 129 } 130 - 130 + 131 131 return 0; 132 132 } 133 133 ··· 162 162 } 163 163 (void)in_le32((volatile u32*)&agp_bridge->gatt_table[pg_start]); 164 164 mb(); 165 - flush_dcache_range((unsigned long)&agp_bridge->gatt_table[pg_start], 165 + flush_dcache_range((unsigned long)&agp_bridge->gatt_table[pg_start], 166 166 (unsigned long)&agp_bridge->gatt_table[pg_start + mem->page_count]); 167 167 168 168 uninorth_tlbflush(mem); ··· 235 235 236 236 command = agp_collect_device_status(bridge, mode, status); 237 237 command |= PCI_AGP_COMMAND_AGP; 238 - 238 + 239 239 if (uninorth_rev == 0x21) { 240 240 /* 241 241 * Darwin disable AGP 4x on this revision, thus we ··· 456 456 {256, 65536, 6, 64}, 457 457 {128, 32768, 5, 32}, 458 458 {64, 16384, 4, 16}, 459 - #endif 459 + #endif 460 460 {32, 8192, 3, 8}, 461 461 {16, 4096, 2, 4}, 462 462 {8, 2048, 1, 2},