Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
at v2.6.23-rc1 1753 lines 44 kB view raw
1/* 2 * rrunner.c: Linux driver for the Essential RoadRunner HIPPI board. 3 * 4 * Copyright (C) 1998-2002 by Jes Sorensen, <jes@wildopensource.com>. 5 * 6 * Thanks to Essential Communication for providing us with hardware 7 * and very comprehensive documentation without which I would not have 8 * been able to write this driver. A special thank you to John Gibbon 9 * for sorting out the legal issues, with the NDA, allowing the code to 10 * be released under the GPL. 11 * 12 * This program is free software; you can redistribute it and/or modify 13 * it under the terms of the GNU General Public License as published by 14 * the Free Software Foundation; either version 2 of the License, or 15 * (at your option) any later version. 16 * 17 * Thanks to Jayaram Bhat from ODS/Essential for fixing some of the 18 * stupid bugs in my code. 19 * 20 * Softnet support and various other patches from Val Henson of 21 * ODS/Essential. 22 * 23 * PCI DMA mapping code partly based on work by Francois Romieu. 24 */ 25 26 27#define DEBUG 1 28#define RX_DMA_SKBUFF 1 29#define PKT_COPY_THRESHOLD 512 30 31#include <linux/module.h> 32#include <linux/types.h> 33#include <linux/errno.h> 34#include <linux/ioport.h> 35#include <linux/pci.h> 36#include <linux/kernel.h> 37#include <linux/netdevice.h> 38#include <linux/hippidevice.h> 39#include <linux/skbuff.h> 40#include <linux/init.h> 41#include <linux/delay.h> 42#include <linux/mm.h> 43#include <net/sock.h> 44 45#include <asm/system.h> 46#include <asm/cache.h> 47#include <asm/byteorder.h> 48#include <asm/io.h> 49#include <asm/irq.h> 50#include <asm/uaccess.h> 51 52#define rr_if_busy(dev) netif_queue_stopped(dev) 53#define rr_if_running(dev) netif_running(dev) 54 55#include "rrunner.h" 56 57#define RUN_AT(x) (jiffies + (x)) 58 59 60MODULE_AUTHOR("Jes Sorensen <jes@wildopensource.com>"); 61MODULE_DESCRIPTION("Essential RoadRunner HIPPI driver"); 62MODULE_LICENSE("GPL"); 63 64static char version[] __devinitdata = "rrunner.c: v0.50 11/11/2002 Jes Sorensen (jes@wildopensource.com)\n"; 65 66/* 67 * Implementation notes: 68 * 69 * The DMA engine only allows for DMA within physical 64KB chunks of 70 * memory. The current approach of the driver (and stack) is to use 71 * linear blocks of memory for the skbuffs. However, as the data block 72 * is always the first part of the skb and skbs are 2^n aligned so we 73 * are guarantted to get the whole block within one 64KB align 64KB 74 * chunk. 75 * 76 * On the long term, relying on being able to allocate 64KB linear 77 * chunks of memory is not feasible and the skb handling code and the 78 * stack will need to know about I/O vectors or something similar. 79 */ 80 81/* 82 * These are checked at init time to see if they are at least 256KB 83 * and increased to 256KB if they are not. This is done to avoid ending 84 * up with socket buffers smaller than the MTU size, 85 */ 86extern __u32 sysctl_wmem_max; 87extern __u32 sysctl_rmem_max; 88 89static int __devinit rr_init_one(struct pci_dev *pdev, 90 const struct pci_device_id *ent) 91{ 92 struct net_device *dev; 93 static int version_disp; 94 u8 pci_latency; 95 struct rr_private *rrpriv; 96 void *tmpptr; 97 dma_addr_t ring_dma; 98 int ret = -ENOMEM; 99 100 dev = alloc_hippi_dev(sizeof(struct rr_private)); 101 if (!dev) 102 goto out3; 103 104 ret = pci_enable_device(pdev); 105 if (ret) { 106 ret = -ENODEV; 107 goto out2; 108 } 109 110 rrpriv = netdev_priv(dev); 111 112 SET_MODULE_OWNER(dev); 113 SET_NETDEV_DEV(dev, &pdev->dev); 114 115 if (pci_request_regions(pdev, "rrunner")) { 116 ret = -EIO; 117 goto out; 118 } 119 120 pci_set_drvdata(pdev, dev); 121 122 rrpriv->pci_dev = pdev; 123 124 spin_lock_init(&rrpriv->lock); 125 126 dev->irq = pdev->irq; 127 dev->open = &rr_open; 128 dev->hard_start_xmit = &rr_start_xmit; 129 dev->stop = &rr_close; 130 dev->get_stats = &rr_get_stats; 131 dev->do_ioctl = &rr_ioctl; 132 133 dev->base_addr = pci_resource_start(pdev, 0); 134 135 /* display version info if adapter is found */ 136 if (!version_disp) { 137 /* set display flag to TRUE so that */ 138 /* we only display this string ONCE */ 139 version_disp = 1; 140 printk(version); 141 } 142 143 pci_read_config_byte(pdev, PCI_LATENCY_TIMER, &pci_latency); 144 if (pci_latency <= 0x58){ 145 pci_latency = 0x58; 146 pci_write_config_byte(pdev, PCI_LATENCY_TIMER, pci_latency); 147 } 148 149 pci_set_master(pdev); 150 151 printk(KERN_INFO "%s: Essential RoadRunner serial HIPPI " 152 "at 0x%08lx, irq %i, PCI latency %i\n", dev->name, 153 dev->base_addr, dev->irq, pci_latency); 154 155 /* 156 * Remap the regs into kernel space. 157 */ 158 159 rrpriv->regs = ioremap(dev->base_addr, 0x1000); 160 161 if (!rrpriv->regs){ 162 printk(KERN_ERR "%s: Unable to map I/O register, " 163 "RoadRunner will be disabled.\n", dev->name); 164 ret = -EIO; 165 goto out; 166 } 167 168 tmpptr = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma); 169 rrpriv->tx_ring = tmpptr; 170 rrpriv->tx_ring_dma = ring_dma; 171 172 if (!tmpptr) { 173 ret = -ENOMEM; 174 goto out; 175 } 176 177 tmpptr = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma); 178 rrpriv->rx_ring = tmpptr; 179 rrpriv->rx_ring_dma = ring_dma; 180 181 if (!tmpptr) { 182 ret = -ENOMEM; 183 goto out; 184 } 185 186 tmpptr = pci_alloc_consistent(pdev, EVT_RING_SIZE, &ring_dma); 187 rrpriv->evt_ring = tmpptr; 188 rrpriv->evt_ring_dma = ring_dma; 189 190 if (!tmpptr) { 191 ret = -ENOMEM; 192 goto out; 193 } 194 195 /* 196 * Don't access any register before this point! 197 */ 198#ifdef __BIG_ENDIAN 199 writel(readl(&rrpriv->regs->HostCtrl) | NO_SWAP, 200 &rrpriv->regs->HostCtrl); 201#endif 202 /* 203 * Need to add a case for little-endian 64-bit hosts here. 204 */ 205 206 rr_init(dev); 207 208 dev->base_addr = 0; 209 210 ret = register_netdev(dev); 211 if (ret) 212 goto out; 213 return 0; 214 215 out: 216 if (rrpriv->rx_ring) 217 pci_free_consistent(pdev, RX_TOTAL_SIZE, rrpriv->rx_ring, 218 rrpriv->rx_ring_dma); 219 if (rrpriv->tx_ring) 220 pci_free_consistent(pdev, TX_TOTAL_SIZE, rrpriv->tx_ring, 221 rrpriv->tx_ring_dma); 222 if (rrpriv->regs) 223 iounmap(rrpriv->regs); 224 if (pdev) { 225 pci_release_regions(pdev); 226 pci_set_drvdata(pdev, NULL); 227 } 228 out2: 229 free_netdev(dev); 230 out3: 231 return ret; 232} 233 234static void __devexit rr_remove_one (struct pci_dev *pdev) 235{ 236 struct net_device *dev = pci_get_drvdata(pdev); 237 238 if (dev) { 239 struct rr_private *rr = netdev_priv(dev); 240 241 if (!(readl(&rr->regs->HostCtrl) & NIC_HALTED)){ 242 printk(KERN_ERR "%s: trying to unload running NIC\n", 243 dev->name); 244 writel(HALT_NIC, &rr->regs->HostCtrl); 245 } 246 247 pci_free_consistent(pdev, EVT_RING_SIZE, rr->evt_ring, 248 rr->evt_ring_dma); 249 pci_free_consistent(pdev, RX_TOTAL_SIZE, rr->rx_ring, 250 rr->rx_ring_dma); 251 pci_free_consistent(pdev, TX_TOTAL_SIZE, rr->tx_ring, 252 rr->tx_ring_dma); 253 unregister_netdev(dev); 254 iounmap(rr->regs); 255 free_netdev(dev); 256 pci_release_regions(pdev); 257 pci_disable_device(pdev); 258 pci_set_drvdata(pdev, NULL); 259 } 260} 261 262 263/* 264 * Commands are considered to be slow, thus there is no reason to 265 * inline this. 266 */ 267static void rr_issue_cmd(struct rr_private *rrpriv, struct cmd *cmd) 268{ 269 struct rr_regs __iomem *regs; 270 u32 idx; 271 272 regs = rrpriv->regs; 273 /* 274 * This is temporary - it will go away in the final version. 275 * We probably also want to make this function inline. 276 */ 277 if (readl(&regs->HostCtrl) & NIC_HALTED){ 278 printk("issuing command for halted NIC, code 0x%x, " 279 "HostCtrl %08x\n", cmd->code, readl(&regs->HostCtrl)); 280 if (readl(&regs->Mode) & FATAL_ERR) 281 printk("error codes Fail1 %02x, Fail2 %02x\n", 282 readl(&regs->Fail1), readl(&regs->Fail2)); 283 } 284 285 idx = rrpriv->info->cmd_ctrl.pi; 286 287 writel(*(u32*)(cmd), &regs->CmdRing[idx]); 288 wmb(); 289 290 idx = (idx - 1) % CMD_RING_ENTRIES; 291 rrpriv->info->cmd_ctrl.pi = idx; 292 wmb(); 293 294 if (readl(&regs->Mode) & FATAL_ERR) 295 printk("error code %02x\n", readl(&regs->Fail1)); 296} 297 298 299/* 300 * Reset the board in a sensible manner. The NIC is already halted 301 * when we get here and a spin-lock is held. 302 */ 303static int rr_reset(struct net_device *dev) 304{ 305 struct rr_private *rrpriv; 306 struct rr_regs __iomem *regs; 307 struct eeprom *hw = NULL; 308 u32 start_pc; 309 int i; 310 311 rrpriv = netdev_priv(dev); 312 regs = rrpriv->regs; 313 314 rr_load_firmware(dev); 315 316 writel(0x01000000, &regs->TX_state); 317 writel(0xff800000, &regs->RX_state); 318 writel(0, &regs->AssistState); 319 writel(CLEAR_INTA, &regs->LocalCtrl); 320 writel(0x01, &regs->BrkPt); 321 writel(0, &regs->Timer); 322 writel(0, &regs->TimerRef); 323 writel(RESET_DMA, &regs->DmaReadState); 324 writel(RESET_DMA, &regs->DmaWriteState); 325 writel(0, &regs->DmaWriteHostHi); 326 writel(0, &regs->DmaWriteHostLo); 327 writel(0, &regs->DmaReadHostHi); 328 writel(0, &regs->DmaReadHostLo); 329 writel(0, &regs->DmaReadLen); 330 writel(0, &regs->DmaWriteLen); 331 writel(0, &regs->DmaWriteLcl); 332 writel(0, &regs->DmaWriteIPchecksum); 333 writel(0, &regs->DmaReadLcl); 334 writel(0, &regs->DmaReadIPchecksum); 335 writel(0, &regs->PciState); 336#if (BITS_PER_LONG == 64) && defined __LITTLE_ENDIAN 337 writel(SWAP_DATA | PTR64BIT | PTR_WD_SWAP, &regs->Mode); 338#elif (BITS_PER_LONG == 64) 339 writel(SWAP_DATA | PTR64BIT | PTR_WD_NOSWAP, &regs->Mode); 340#else 341 writel(SWAP_DATA | PTR32BIT | PTR_WD_NOSWAP, &regs->Mode); 342#endif 343 344#if 0 345 /* 346 * Don't worry, this is just black magic. 347 */ 348 writel(0xdf000, &regs->RxBase); 349 writel(0xdf000, &regs->RxPrd); 350 writel(0xdf000, &regs->RxCon); 351 writel(0xce000, &regs->TxBase); 352 writel(0xce000, &regs->TxPrd); 353 writel(0xce000, &regs->TxCon); 354 writel(0, &regs->RxIndPro); 355 writel(0, &regs->RxIndCon); 356 writel(0, &regs->RxIndRef); 357 writel(0, &regs->TxIndPro); 358 writel(0, &regs->TxIndCon); 359 writel(0, &regs->TxIndRef); 360 writel(0xcc000, &regs->pad10[0]); 361 writel(0, &regs->DrCmndPro); 362 writel(0, &regs->DrCmndCon); 363 writel(0, &regs->DwCmndPro); 364 writel(0, &regs->DwCmndCon); 365 writel(0, &regs->DwCmndRef); 366 writel(0, &regs->DrDataPro); 367 writel(0, &regs->DrDataCon); 368 writel(0, &regs->DrDataRef); 369 writel(0, &regs->DwDataPro); 370 writel(0, &regs->DwDataCon); 371 writel(0, &regs->DwDataRef); 372#endif 373 374 writel(0xffffffff, &regs->MbEvent); 375 writel(0, &regs->Event); 376 377 writel(0, &regs->TxPi); 378 writel(0, &regs->IpRxPi); 379 380 writel(0, &regs->EvtCon); 381 writel(0, &regs->EvtPrd); 382 383 rrpriv->info->evt_ctrl.pi = 0; 384 385 for (i = 0; i < CMD_RING_ENTRIES; i++) 386 writel(0, &regs->CmdRing[i]); 387 388/* 389 * Why 32 ? is this not cache line size dependent? 390 */ 391 writel(RBURST_64|WBURST_64, &regs->PciState); 392 wmb(); 393 394 start_pc = rr_read_eeprom_word(rrpriv, &hw->rncd_info.FwStart); 395 396#if (DEBUG > 1) 397 printk("%s: Executing firmware at address 0x%06x\n", 398 dev->name, start_pc); 399#endif 400 401 writel(start_pc + 0x800, &regs->Pc); 402 wmb(); 403 udelay(5); 404 405 writel(start_pc, &regs->Pc); 406 wmb(); 407 408 return 0; 409} 410 411 412/* 413 * Read a string from the EEPROM. 414 */ 415static unsigned int rr_read_eeprom(struct rr_private *rrpriv, 416 unsigned long offset, 417 unsigned char *buf, 418 unsigned long length) 419{ 420 struct rr_regs __iomem *regs = rrpriv->regs; 421 u32 misc, io, host, i; 422 423 io = readl(&regs->ExtIo); 424 writel(0, &regs->ExtIo); 425 misc = readl(&regs->LocalCtrl); 426 writel(0, &regs->LocalCtrl); 427 host = readl(&regs->HostCtrl); 428 writel(host | HALT_NIC, &regs->HostCtrl); 429 mb(); 430 431 for (i = 0; i < length; i++){ 432 writel((EEPROM_BASE + ((offset+i) << 3)), &regs->WinBase); 433 mb(); 434 buf[i] = (readl(&regs->WinData) >> 24) & 0xff; 435 mb(); 436 } 437 438 writel(host, &regs->HostCtrl); 439 writel(misc, &regs->LocalCtrl); 440 writel(io, &regs->ExtIo); 441 mb(); 442 return i; 443} 444 445 446/* 447 * Shortcut to read one word (4 bytes) out of the EEPROM and convert 448 * it to our CPU byte-order. 449 */ 450static u32 rr_read_eeprom_word(struct rr_private *rrpriv, 451 void * offset) 452{ 453 u32 word; 454 455 if ((rr_read_eeprom(rrpriv, (unsigned long)offset, 456 (char *)&word, 4) == 4)) 457 return be32_to_cpu(word); 458 return 0; 459} 460 461 462/* 463 * Write a string to the EEPROM. 464 * 465 * This is only called when the firmware is not running. 466 */ 467static unsigned int write_eeprom(struct rr_private *rrpriv, 468 unsigned long offset, 469 unsigned char *buf, 470 unsigned long length) 471{ 472 struct rr_regs __iomem *regs = rrpriv->regs; 473 u32 misc, io, data, i, j, ready, error = 0; 474 475 io = readl(&regs->ExtIo); 476 writel(0, &regs->ExtIo); 477 misc = readl(&regs->LocalCtrl); 478 writel(ENABLE_EEPROM_WRITE, &regs->LocalCtrl); 479 mb(); 480 481 for (i = 0; i < length; i++){ 482 writel((EEPROM_BASE + ((offset+i) << 3)), &regs->WinBase); 483 mb(); 484 data = buf[i] << 24; 485 /* 486 * Only try to write the data if it is not the same 487 * value already. 488 */ 489 if ((readl(&regs->WinData) & 0xff000000) != data){ 490 writel(data, &regs->WinData); 491 ready = 0; 492 j = 0; 493 mb(); 494 while(!ready){ 495 udelay(20); 496 if ((readl(&regs->WinData) & 0xff000000) == 497 data) 498 ready = 1; 499 mb(); 500 if (j++ > 5000){ 501 printk("data mismatch: %08x, " 502 "WinData %08x\n", data, 503 readl(&regs->WinData)); 504 ready = 1; 505 error = 1; 506 } 507 } 508 } 509 } 510 511 writel(misc, &regs->LocalCtrl); 512 writel(io, &regs->ExtIo); 513 mb(); 514 515 return error; 516} 517 518 519static int __devinit rr_init(struct net_device *dev) 520{ 521 struct rr_private *rrpriv; 522 struct rr_regs __iomem *regs; 523 struct eeprom *hw = NULL; 524 u32 sram_size, rev; 525 int i; 526 527 rrpriv = netdev_priv(dev); 528 regs = rrpriv->regs; 529 530 rev = readl(&regs->FwRev); 531 rrpriv->fw_rev = rev; 532 if (rev > 0x00020024) 533 printk(" Firmware revision: %i.%i.%i\n", (rev >> 16), 534 ((rev >> 8) & 0xff), (rev & 0xff)); 535 else if (rev >= 0x00020000) { 536 printk(" Firmware revision: %i.%i.%i (2.0.37 or " 537 "later is recommended)\n", (rev >> 16), 538 ((rev >> 8) & 0xff), (rev & 0xff)); 539 }else{ 540 printk(" Firmware revision too old: %i.%i.%i, please " 541 "upgrade to 2.0.37 or later.\n", 542 (rev >> 16), ((rev >> 8) & 0xff), (rev & 0xff)); 543 } 544 545#if (DEBUG > 2) 546 printk(" Maximum receive rings %i\n", readl(&regs->MaxRxRng)); 547#endif 548 549 /* 550 * Read the hardware address from the eeprom. The HW address 551 * is not really necessary for HIPPI but awfully convenient. 552 * The pointer arithmetic to put it in dev_addr is ugly, but 553 * Donald Becker does it this way for the GigE version of this 554 * card and it's shorter and more portable than any 555 * other method I've seen. -VAL 556 */ 557 558 *(u16 *)(dev->dev_addr) = 559 htons(rr_read_eeprom_word(rrpriv, &hw->manf.BoardULA)); 560 *(u32 *)(dev->dev_addr+2) = 561 htonl(rr_read_eeprom_word(rrpriv, &hw->manf.BoardULA[4])); 562 563 printk(" MAC: "); 564 565 for (i = 0; i < 5; i++) 566 printk("%2.2x:", dev->dev_addr[i]); 567 printk("%2.2x\n", dev->dev_addr[i]); 568 569 sram_size = rr_read_eeprom_word(rrpriv, (void *)8); 570 printk(" SRAM size 0x%06x\n", sram_size); 571 572 if (sysctl_rmem_max < 262144){ 573 printk(" Receive socket buffer limit too low (%i), " 574 "setting to 262144\n", sysctl_rmem_max); 575 sysctl_rmem_max = 262144; 576 } 577 578 if (sysctl_wmem_max < 262144){ 579 printk(" Transmit socket buffer limit too low (%i), " 580 "setting to 262144\n", sysctl_wmem_max); 581 sysctl_wmem_max = 262144; 582 } 583 584 return 0; 585} 586 587 588static int rr_init1(struct net_device *dev) 589{ 590 struct rr_private *rrpriv; 591 struct rr_regs __iomem *regs; 592 unsigned long myjif, flags; 593 struct cmd cmd; 594 u32 hostctrl; 595 int ecode = 0; 596 short i; 597 598 rrpriv = netdev_priv(dev); 599 regs = rrpriv->regs; 600 601 spin_lock_irqsave(&rrpriv->lock, flags); 602 603 hostctrl = readl(&regs->HostCtrl); 604 writel(hostctrl | HALT_NIC | RR_CLEAR_INT, &regs->HostCtrl); 605 wmb(); 606 607 if (hostctrl & PARITY_ERR){ 608 printk("%s: Parity error halting NIC - this is serious!\n", 609 dev->name); 610 spin_unlock_irqrestore(&rrpriv->lock, flags); 611 ecode = -EFAULT; 612 goto error; 613 } 614 615 set_rxaddr(regs, rrpriv->rx_ctrl_dma); 616 set_infoaddr(regs, rrpriv->info_dma); 617 618 rrpriv->info->evt_ctrl.entry_size = sizeof(struct event); 619 rrpriv->info->evt_ctrl.entries = EVT_RING_ENTRIES; 620 rrpriv->info->evt_ctrl.mode = 0; 621 rrpriv->info->evt_ctrl.pi = 0; 622 set_rraddr(&rrpriv->info->evt_ctrl.rngptr, rrpriv->evt_ring_dma); 623 624 rrpriv->info->cmd_ctrl.entry_size = sizeof(struct cmd); 625 rrpriv->info->cmd_ctrl.entries = CMD_RING_ENTRIES; 626 rrpriv->info->cmd_ctrl.mode = 0; 627 rrpriv->info->cmd_ctrl.pi = 15; 628 629 for (i = 0; i < CMD_RING_ENTRIES; i++) { 630 writel(0, &regs->CmdRing[i]); 631 } 632 633 for (i = 0; i < TX_RING_ENTRIES; i++) { 634 rrpriv->tx_ring[i].size = 0; 635 set_rraddr(&rrpriv->tx_ring[i].addr, 0); 636 rrpriv->tx_skbuff[i] = NULL; 637 } 638 rrpriv->info->tx_ctrl.entry_size = sizeof(struct tx_desc); 639 rrpriv->info->tx_ctrl.entries = TX_RING_ENTRIES; 640 rrpriv->info->tx_ctrl.mode = 0; 641 rrpriv->info->tx_ctrl.pi = 0; 642 set_rraddr(&rrpriv->info->tx_ctrl.rngptr, rrpriv->tx_ring_dma); 643 644 /* 645 * Set dirty_tx before we start receiving interrupts, otherwise 646 * the interrupt handler might think it is supposed to process 647 * tx ints before we are up and running, which may cause a null 648 * pointer access in the int handler. 649 */ 650 rrpriv->tx_full = 0; 651 rrpriv->cur_rx = 0; 652 rrpriv->dirty_rx = rrpriv->dirty_tx = 0; 653 654 rr_reset(dev); 655 656 /* Tuning values */ 657 writel(0x5000, &regs->ConRetry); 658 writel(0x100, &regs->ConRetryTmr); 659 writel(0x500000, &regs->ConTmout); 660 writel(0x60, &regs->IntrTmr); 661 writel(0x500000, &regs->TxDataMvTimeout); 662 writel(0x200000, &regs->RxDataMvTimeout); 663 writel(0x80, &regs->WriteDmaThresh); 664 writel(0x80, &regs->ReadDmaThresh); 665 666 rrpriv->fw_running = 0; 667 wmb(); 668 669 hostctrl &= ~(HALT_NIC | INVALID_INST_B | PARITY_ERR); 670 writel(hostctrl, &regs->HostCtrl); 671 wmb(); 672 673 spin_unlock_irqrestore(&rrpriv->lock, flags); 674 675 for (i = 0; i < RX_RING_ENTRIES; i++) { 676 struct sk_buff *skb; 677 dma_addr_t addr; 678 679 rrpriv->rx_ring[i].mode = 0; 680 skb = alloc_skb(dev->mtu + HIPPI_HLEN, GFP_ATOMIC); 681 if (!skb) { 682 printk(KERN_WARNING "%s: Unable to allocate memory " 683 "for receive ring - halting NIC\n", dev->name); 684 ecode = -ENOMEM; 685 goto error; 686 } 687 rrpriv->rx_skbuff[i] = skb; 688 addr = pci_map_single(rrpriv->pci_dev, skb->data, 689 dev->mtu + HIPPI_HLEN, PCI_DMA_FROMDEVICE); 690 /* 691 * Sanity test to see if we conflict with the DMA 692 * limitations of the Roadrunner. 693 */ 694 if ((((unsigned long)skb->data) & 0xfff) > ~65320) 695 printk("skb alloc error\n"); 696 697 set_rraddr(&rrpriv->rx_ring[i].addr, addr); 698 rrpriv->rx_ring[i].size = dev->mtu + HIPPI_HLEN; 699 } 700 701 rrpriv->rx_ctrl[4].entry_size = sizeof(struct rx_desc); 702 rrpriv->rx_ctrl[4].entries = RX_RING_ENTRIES; 703 rrpriv->rx_ctrl[4].mode = 8; 704 rrpriv->rx_ctrl[4].pi = 0; 705 wmb(); 706 set_rraddr(&rrpriv->rx_ctrl[4].rngptr, rrpriv->rx_ring_dma); 707 708 udelay(1000); 709 710 /* 711 * Now start the FirmWare. 712 */ 713 cmd.code = C_START_FW; 714 cmd.ring = 0; 715 cmd.index = 0; 716 717 rr_issue_cmd(rrpriv, &cmd); 718 719 /* 720 * Give the FirmWare time to chew on the `get running' command. 721 */ 722 myjif = jiffies + 5 * HZ; 723 while (time_before(jiffies, myjif) && !rrpriv->fw_running) 724 cpu_relax(); 725 726 netif_start_queue(dev); 727 728 return ecode; 729 730 error: 731 /* 732 * We might have gotten here because we are out of memory, 733 * make sure we release everything we allocated before failing 734 */ 735 for (i = 0; i < RX_RING_ENTRIES; i++) { 736 struct sk_buff *skb = rrpriv->rx_skbuff[i]; 737 738 if (skb) { 739 pci_unmap_single(rrpriv->pci_dev, 740 rrpriv->rx_ring[i].addr.addrlo, 741 dev->mtu + HIPPI_HLEN, 742 PCI_DMA_FROMDEVICE); 743 rrpriv->rx_ring[i].size = 0; 744 set_rraddr(&rrpriv->rx_ring[i].addr, 0); 745 dev_kfree_skb(skb); 746 rrpriv->rx_skbuff[i] = NULL; 747 } 748 } 749 return ecode; 750} 751 752 753/* 754 * All events are considered to be slow (RX/TX ints do not generate 755 * events) and are handled here, outside the main interrupt handler, 756 * to reduce the size of the handler. 757 */ 758static u32 rr_handle_event(struct net_device *dev, u32 prodidx, u32 eidx) 759{ 760 struct rr_private *rrpriv; 761 struct rr_regs __iomem *regs; 762 u32 tmp; 763 764 rrpriv = netdev_priv(dev); 765 regs = rrpriv->regs; 766 767 while (prodidx != eidx){ 768 switch (rrpriv->evt_ring[eidx].code){ 769 case E_NIC_UP: 770 tmp = readl(&regs->FwRev); 771 printk(KERN_INFO "%s: Firmware revision %i.%i.%i " 772 "up and running\n", dev->name, 773 (tmp >> 16), ((tmp >> 8) & 0xff), (tmp & 0xff)); 774 rrpriv->fw_running = 1; 775 writel(RX_RING_ENTRIES - 1, &regs->IpRxPi); 776 wmb(); 777 break; 778 case E_LINK_ON: 779 printk(KERN_INFO "%s: Optical link ON\n", dev->name); 780 break; 781 case E_LINK_OFF: 782 printk(KERN_INFO "%s: Optical link OFF\n", dev->name); 783 break; 784 case E_RX_IDLE: 785 printk(KERN_WARNING "%s: RX data not moving\n", 786 dev->name); 787 goto drop; 788 case E_WATCHDOG: 789 printk(KERN_INFO "%s: The watchdog is here to see " 790 "us\n", dev->name); 791 break; 792 case E_INTERN_ERR: 793 printk(KERN_ERR "%s: HIPPI Internal NIC error\n", 794 dev->name); 795 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 796 &regs->HostCtrl); 797 wmb(); 798 break; 799 case E_HOST_ERR: 800 printk(KERN_ERR "%s: Host software error\n", 801 dev->name); 802 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 803 &regs->HostCtrl); 804 wmb(); 805 break; 806 /* 807 * TX events. 808 */ 809 case E_CON_REJ: 810 printk(KERN_WARNING "%s: Connection rejected\n", 811 dev->name); 812 rrpriv->stats.tx_aborted_errors++; 813 break; 814 case E_CON_TMOUT: 815 printk(KERN_WARNING "%s: Connection timeout\n", 816 dev->name); 817 break; 818 case E_DISC_ERR: 819 printk(KERN_WARNING "%s: HIPPI disconnect error\n", 820 dev->name); 821 rrpriv->stats.tx_aborted_errors++; 822 break; 823 case E_INT_PRTY: 824 printk(KERN_ERR "%s: HIPPI Internal Parity error\n", 825 dev->name); 826 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 827 &regs->HostCtrl); 828 wmb(); 829 break; 830 case E_TX_IDLE: 831 printk(KERN_WARNING "%s: Transmitter idle\n", 832 dev->name); 833 break; 834 case E_TX_LINK_DROP: 835 printk(KERN_WARNING "%s: Link lost during transmit\n", 836 dev->name); 837 rrpriv->stats.tx_aborted_errors++; 838 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 839 &regs->HostCtrl); 840 wmb(); 841 break; 842 case E_TX_INV_RNG: 843 printk(KERN_ERR "%s: Invalid send ring block\n", 844 dev->name); 845 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 846 &regs->HostCtrl); 847 wmb(); 848 break; 849 case E_TX_INV_BUF: 850 printk(KERN_ERR "%s: Invalid send buffer address\n", 851 dev->name); 852 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 853 &regs->HostCtrl); 854 wmb(); 855 break; 856 case E_TX_INV_DSC: 857 printk(KERN_ERR "%s: Invalid descriptor address\n", 858 dev->name); 859 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 860 &regs->HostCtrl); 861 wmb(); 862 break; 863 /* 864 * RX events. 865 */ 866 case E_RX_RNG_OUT: 867 printk(KERN_INFO "%s: Receive ring full\n", dev->name); 868 break; 869 870 case E_RX_PAR_ERR: 871 printk(KERN_WARNING "%s: Receive parity error\n", 872 dev->name); 873 goto drop; 874 case E_RX_LLRC_ERR: 875 printk(KERN_WARNING "%s: Receive LLRC error\n", 876 dev->name); 877 goto drop; 878 case E_PKT_LN_ERR: 879 printk(KERN_WARNING "%s: Receive packet length " 880 "error\n", dev->name); 881 goto drop; 882 case E_DTA_CKSM_ERR: 883 printk(KERN_WARNING "%s: Data checksum error\n", 884 dev->name); 885 goto drop; 886 case E_SHT_BST: 887 printk(KERN_WARNING "%s: Unexpected short burst " 888 "error\n", dev->name); 889 goto drop; 890 case E_STATE_ERR: 891 printk(KERN_WARNING "%s: Recv. state transition" 892 " error\n", dev->name); 893 goto drop; 894 case E_UNEXP_DATA: 895 printk(KERN_WARNING "%s: Unexpected data error\n", 896 dev->name); 897 goto drop; 898 case E_LST_LNK_ERR: 899 printk(KERN_WARNING "%s: Link lost error\n", 900 dev->name); 901 goto drop; 902 case E_FRM_ERR: 903 printk(KERN_WARNING "%s: Framming Error\n", 904 dev->name); 905 goto drop; 906 case E_FLG_SYN_ERR: 907 printk(KERN_WARNING "%s: Flag sync. lost during" 908 "packet\n", dev->name); 909 goto drop; 910 case E_RX_INV_BUF: 911 printk(KERN_ERR "%s: Invalid receive buffer " 912 "address\n", dev->name); 913 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 914 &regs->HostCtrl); 915 wmb(); 916 break; 917 case E_RX_INV_DSC: 918 printk(KERN_ERR "%s: Invalid receive descriptor " 919 "address\n", dev->name); 920 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 921 &regs->HostCtrl); 922 wmb(); 923 break; 924 case E_RNG_BLK: 925 printk(KERN_ERR "%s: Invalid ring block\n", 926 dev->name); 927 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 928 &regs->HostCtrl); 929 wmb(); 930 break; 931 drop: 932 /* Label packet to be dropped. 933 * Actual dropping occurs in rx 934 * handling. 935 * 936 * The index of packet we get to drop is 937 * the index of the packet following 938 * the bad packet. -kbf 939 */ 940 { 941 u16 index = rrpriv->evt_ring[eidx].index; 942 index = (index + (RX_RING_ENTRIES - 1)) % 943 RX_RING_ENTRIES; 944 rrpriv->rx_ring[index].mode |= 945 (PACKET_BAD | PACKET_END); 946 } 947 break; 948 default: 949 printk(KERN_WARNING "%s: Unhandled event 0x%02x\n", 950 dev->name, rrpriv->evt_ring[eidx].code); 951 } 952 eidx = (eidx + 1) % EVT_RING_ENTRIES; 953 } 954 955 rrpriv->info->evt_ctrl.pi = eidx; 956 wmb(); 957 return eidx; 958} 959 960 961static void rx_int(struct net_device *dev, u32 rxlimit, u32 index) 962{ 963 struct rr_private *rrpriv = netdev_priv(dev); 964 struct rr_regs __iomem *regs = rrpriv->regs; 965 966 do { 967 struct rx_desc *desc; 968 u32 pkt_len; 969 970 desc = &(rrpriv->rx_ring[index]); 971 pkt_len = desc->size; 972#if (DEBUG > 2) 973 printk("index %i, rxlimit %i\n", index, rxlimit); 974 printk("len %x, mode %x\n", pkt_len, desc->mode); 975#endif 976 if ( (rrpriv->rx_ring[index].mode & PACKET_BAD) == PACKET_BAD){ 977 rrpriv->stats.rx_dropped++; 978 goto defer; 979 } 980 981 if (pkt_len > 0){ 982 struct sk_buff *skb, *rx_skb; 983 984 rx_skb = rrpriv->rx_skbuff[index]; 985 986 if (pkt_len < PKT_COPY_THRESHOLD) { 987 skb = alloc_skb(pkt_len, GFP_ATOMIC); 988 if (skb == NULL){ 989 printk(KERN_WARNING "%s: Unable to allocate skb (%i bytes), deferring packet\n", dev->name, pkt_len); 990 rrpriv->stats.rx_dropped++; 991 goto defer; 992 } else { 993 pci_dma_sync_single_for_cpu(rrpriv->pci_dev, 994 desc->addr.addrlo, 995 pkt_len, 996 PCI_DMA_FROMDEVICE); 997 998 memcpy(skb_put(skb, pkt_len), 999 rx_skb->data, pkt_len); 1000 1001 pci_dma_sync_single_for_device(rrpriv->pci_dev, 1002 desc->addr.addrlo, 1003 pkt_len, 1004 PCI_DMA_FROMDEVICE); 1005 } 1006 }else{ 1007 struct sk_buff *newskb; 1008 1009 newskb = alloc_skb(dev->mtu + HIPPI_HLEN, 1010 GFP_ATOMIC); 1011 if (newskb){ 1012 dma_addr_t addr; 1013 1014 pci_unmap_single(rrpriv->pci_dev, 1015 desc->addr.addrlo, dev->mtu + 1016 HIPPI_HLEN, PCI_DMA_FROMDEVICE); 1017 skb = rx_skb; 1018 skb_put(skb, pkt_len); 1019 rrpriv->rx_skbuff[index] = newskb; 1020 addr = pci_map_single(rrpriv->pci_dev, 1021 newskb->data, 1022 dev->mtu + HIPPI_HLEN, 1023 PCI_DMA_FROMDEVICE); 1024 set_rraddr(&desc->addr, addr); 1025 } else { 1026 printk("%s: Out of memory, deferring " 1027 "packet\n", dev->name); 1028 rrpriv->stats.rx_dropped++; 1029 goto defer; 1030 } 1031 } 1032 skb->protocol = hippi_type_trans(skb, dev); 1033 1034 netif_rx(skb); /* send it up */ 1035 1036 dev->last_rx = jiffies; 1037 rrpriv->stats.rx_packets++; 1038 rrpriv->stats.rx_bytes += pkt_len; 1039 } 1040 defer: 1041 desc->mode = 0; 1042 desc->size = dev->mtu + HIPPI_HLEN; 1043 1044 if ((index & 7) == 7) 1045 writel(index, &regs->IpRxPi); 1046 1047 index = (index + 1) % RX_RING_ENTRIES; 1048 } while(index != rxlimit); 1049 1050 rrpriv->cur_rx = index; 1051 wmb(); 1052} 1053 1054 1055static irqreturn_t rr_interrupt(int irq, void *dev_id) 1056{ 1057 struct rr_private *rrpriv; 1058 struct rr_regs __iomem *regs; 1059 struct net_device *dev = (struct net_device *)dev_id; 1060 u32 prodidx, rxindex, eidx, txcsmr, rxlimit, txcon; 1061 1062 rrpriv = netdev_priv(dev); 1063 regs = rrpriv->regs; 1064 1065 if (!(readl(&regs->HostCtrl) & RR_INT)) 1066 return IRQ_NONE; 1067 1068 spin_lock(&rrpriv->lock); 1069 1070 prodidx = readl(&regs->EvtPrd); 1071 txcsmr = (prodidx >> 8) & 0xff; 1072 rxlimit = (prodidx >> 16) & 0xff; 1073 prodidx &= 0xff; 1074 1075#if (DEBUG > 2) 1076 printk("%s: interrupt, prodidx = %i, eidx = %i\n", dev->name, 1077 prodidx, rrpriv->info->evt_ctrl.pi); 1078#endif 1079 /* 1080 * Order here is important. We must handle events 1081 * before doing anything else in order to catch 1082 * such things as LLRC errors, etc -kbf 1083 */ 1084 1085 eidx = rrpriv->info->evt_ctrl.pi; 1086 if (prodidx != eidx) 1087 eidx = rr_handle_event(dev, prodidx, eidx); 1088 1089 rxindex = rrpriv->cur_rx; 1090 if (rxindex != rxlimit) 1091 rx_int(dev, rxlimit, rxindex); 1092 1093 txcon = rrpriv->dirty_tx; 1094 if (txcsmr != txcon) { 1095 do { 1096 /* Due to occational firmware TX producer/consumer out 1097 * of sync. error need to check entry in ring -kbf 1098 */ 1099 if(rrpriv->tx_skbuff[txcon]){ 1100 struct tx_desc *desc; 1101 struct sk_buff *skb; 1102 1103 desc = &(rrpriv->tx_ring[txcon]); 1104 skb = rrpriv->tx_skbuff[txcon]; 1105 1106 rrpriv->stats.tx_packets++; 1107 rrpriv->stats.tx_bytes += skb->len; 1108 1109 pci_unmap_single(rrpriv->pci_dev, 1110 desc->addr.addrlo, skb->len, 1111 PCI_DMA_TODEVICE); 1112 dev_kfree_skb_irq(skb); 1113 1114 rrpriv->tx_skbuff[txcon] = NULL; 1115 desc->size = 0; 1116 set_rraddr(&rrpriv->tx_ring[txcon].addr, 0); 1117 desc->mode = 0; 1118 } 1119 txcon = (txcon + 1) % TX_RING_ENTRIES; 1120 } while (txcsmr != txcon); 1121 wmb(); 1122 1123 rrpriv->dirty_tx = txcon; 1124 if (rrpriv->tx_full && rr_if_busy(dev) && 1125 (((rrpriv->info->tx_ctrl.pi + 1) % TX_RING_ENTRIES) 1126 != rrpriv->dirty_tx)){ 1127 rrpriv->tx_full = 0; 1128 netif_wake_queue(dev); 1129 } 1130 } 1131 1132 eidx |= ((txcsmr << 8) | (rxlimit << 16)); 1133 writel(eidx, &regs->EvtCon); 1134 wmb(); 1135 1136 spin_unlock(&rrpriv->lock); 1137 return IRQ_HANDLED; 1138} 1139 1140static inline void rr_raz_tx(struct rr_private *rrpriv, 1141 struct net_device *dev) 1142{ 1143 int i; 1144 1145 for (i = 0; i < TX_RING_ENTRIES; i++) { 1146 struct sk_buff *skb = rrpriv->tx_skbuff[i]; 1147 1148 if (skb) { 1149 struct tx_desc *desc = &(rrpriv->tx_ring[i]); 1150 1151 pci_unmap_single(rrpriv->pci_dev, desc->addr.addrlo, 1152 skb->len, PCI_DMA_TODEVICE); 1153 desc->size = 0; 1154 set_rraddr(&desc->addr, 0); 1155 dev_kfree_skb(skb); 1156 rrpriv->tx_skbuff[i] = NULL; 1157 } 1158 } 1159} 1160 1161 1162static inline void rr_raz_rx(struct rr_private *rrpriv, 1163 struct net_device *dev) 1164{ 1165 int i; 1166 1167 for (i = 0; i < RX_RING_ENTRIES; i++) { 1168 struct sk_buff *skb = rrpriv->rx_skbuff[i]; 1169 1170 if (skb) { 1171 struct rx_desc *desc = &(rrpriv->rx_ring[i]); 1172 1173 pci_unmap_single(rrpriv->pci_dev, desc->addr.addrlo, 1174 dev->mtu + HIPPI_HLEN, PCI_DMA_FROMDEVICE); 1175 desc->size = 0; 1176 set_rraddr(&desc->addr, 0); 1177 dev_kfree_skb(skb); 1178 rrpriv->rx_skbuff[i] = NULL; 1179 } 1180 } 1181} 1182 1183static void rr_timer(unsigned long data) 1184{ 1185 struct net_device *dev = (struct net_device *)data; 1186 struct rr_private *rrpriv = netdev_priv(dev); 1187 struct rr_regs __iomem *regs = rrpriv->regs; 1188 unsigned long flags; 1189 1190 if (readl(&regs->HostCtrl) & NIC_HALTED){ 1191 printk("%s: Restarting nic\n", dev->name); 1192 memset(rrpriv->rx_ctrl, 0, 256 * sizeof(struct ring_ctrl)); 1193 memset(rrpriv->info, 0, sizeof(struct rr_info)); 1194 wmb(); 1195 1196 rr_raz_tx(rrpriv, dev); 1197 rr_raz_rx(rrpriv, dev); 1198 1199 if (rr_init1(dev)) { 1200 spin_lock_irqsave(&rrpriv->lock, flags); 1201 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, 1202 &regs->HostCtrl); 1203 spin_unlock_irqrestore(&rrpriv->lock, flags); 1204 } 1205 } 1206 rrpriv->timer.expires = RUN_AT(5*HZ); 1207 add_timer(&rrpriv->timer); 1208} 1209 1210 1211static int rr_open(struct net_device *dev) 1212{ 1213 struct rr_private *rrpriv = netdev_priv(dev); 1214 struct pci_dev *pdev = rrpriv->pci_dev; 1215 struct rr_regs __iomem *regs; 1216 int ecode = 0; 1217 unsigned long flags; 1218 dma_addr_t dma_addr; 1219 1220 regs = rrpriv->regs; 1221 1222 if (rrpriv->fw_rev < 0x00020000) { 1223 printk(KERN_WARNING "%s: trying to configure device with " 1224 "obsolete firmware\n", dev->name); 1225 ecode = -EBUSY; 1226 goto error; 1227 } 1228 1229 rrpriv->rx_ctrl = pci_alloc_consistent(pdev, 1230 256 * sizeof(struct ring_ctrl), 1231 &dma_addr); 1232 if (!rrpriv->rx_ctrl) { 1233 ecode = -ENOMEM; 1234 goto error; 1235 } 1236 rrpriv->rx_ctrl_dma = dma_addr; 1237 memset(rrpriv->rx_ctrl, 0, 256*sizeof(struct ring_ctrl)); 1238 1239 rrpriv->info = pci_alloc_consistent(pdev, sizeof(struct rr_info), 1240 &dma_addr); 1241 if (!rrpriv->info) { 1242 ecode = -ENOMEM; 1243 goto error; 1244 } 1245 rrpriv->info_dma = dma_addr; 1246 memset(rrpriv->info, 0, sizeof(struct rr_info)); 1247 wmb(); 1248 1249 spin_lock_irqsave(&rrpriv->lock, flags); 1250 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, &regs->HostCtrl); 1251 readl(&regs->HostCtrl); 1252 spin_unlock_irqrestore(&rrpriv->lock, flags); 1253 1254 if (request_irq(dev->irq, rr_interrupt, IRQF_SHARED, dev->name, dev)) { 1255 printk(KERN_WARNING "%s: Requested IRQ %d is busy\n", 1256 dev->name, dev->irq); 1257 ecode = -EAGAIN; 1258 goto error; 1259 } 1260 1261 if ((ecode = rr_init1(dev))) 1262 goto error; 1263 1264 /* Set the timer to switch to check for link beat and perhaps switch 1265 to an alternate media type. */ 1266 init_timer(&rrpriv->timer); 1267 rrpriv->timer.expires = RUN_AT(5*HZ); /* 5 sec. watchdog */ 1268 rrpriv->timer.data = (unsigned long)dev; 1269 rrpriv->timer.function = &rr_timer; /* timer handler */ 1270 add_timer(&rrpriv->timer); 1271 1272 netif_start_queue(dev); 1273 1274 return ecode; 1275 1276 error: 1277 spin_lock_irqsave(&rrpriv->lock, flags); 1278 writel(readl(&regs->HostCtrl)|HALT_NIC|RR_CLEAR_INT, &regs->HostCtrl); 1279 spin_unlock_irqrestore(&rrpriv->lock, flags); 1280 1281 if (rrpriv->info) { 1282 pci_free_consistent(pdev, sizeof(struct rr_info), rrpriv->info, 1283 rrpriv->info_dma); 1284 rrpriv->info = NULL; 1285 } 1286 if (rrpriv->rx_ctrl) { 1287 pci_free_consistent(pdev, sizeof(struct ring_ctrl), 1288 rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma); 1289 rrpriv->rx_ctrl = NULL; 1290 } 1291 1292 netif_stop_queue(dev); 1293 1294 return ecode; 1295} 1296 1297 1298static void rr_dump(struct net_device *dev) 1299{ 1300 struct rr_private *rrpriv; 1301 struct rr_regs __iomem *regs; 1302 u32 index, cons; 1303 short i; 1304 int len; 1305 1306 rrpriv = netdev_priv(dev); 1307 regs = rrpriv->regs; 1308 1309 printk("%s: dumping NIC TX rings\n", dev->name); 1310 1311 printk("RxPrd %08x, TxPrd %02x, EvtPrd %08x, TxPi %02x, TxCtrlPi %02x\n", 1312 readl(&regs->RxPrd), readl(&regs->TxPrd), 1313 readl(&regs->EvtPrd), readl(&regs->TxPi), 1314 rrpriv->info->tx_ctrl.pi); 1315 1316 printk("Error code 0x%x\n", readl(&regs->Fail1)); 1317 1318 index = (((readl(&regs->EvtPrd) >> 8) & 0xff ) - 1) % EVT_RING_ENTRIES; 1319 cons = rrpriv->dirty_tx; 1320 printk("TX ring index %i, TX consumer %i\n", 1321 index, cons); 1322 1323 if (rrpriv->tx_skbuff[index]){ 1324 len = min_t(int, 0x80, rrpriv->tx_skbuff[index]->len); 1325 printk("skbuff for index %i is valid - dumping data (0x%x bytes - DMA len 0x%x)\n", index, len, rrpriv->tx_ring[index].size); 1326 for (i = 0; i < len; i++){ 1327 if (!(i & 7)) 1328 printk("\n"); 1329 printk("%02x ", (unsigned char) rrpriv->tx_skbuff[index]->data[i]); 1330 } 1331 printk("\n"); 1332 } 1333 1334 if (rrpriv->tx_skbuff[cons]){ 1335 len = min_t(int, 0x80, rrpriv->tx_skbuff[cons]->len); 1336 printk("skbuff for cons %i is valid - dumping data (0x%x bytes - skbuff len 0x%x)\n", cons, len, rrpriv->tx_skbuff[cons]->len); 1337 printk("mode 0x%x, size 0x%x,\n phys %08Lx, skbuff-addr %08lx, truesize 0x%x\n", 1338 rrpriv->tx_ring[cons].mode, 1339 rrpriv->tx_ring[cons].size, 1340 (unsigned long long) rrpriv->tx_ring[cons].addr.addrlo, 1341 (unsigned long)rrpriv->tx_skbuff[cons]->data, 1342 (unsigned int)rrpriv->tx_skbuff[cons]->truesize); 1343 for (i = 0; i < len; i++){ 1344 if (!(i & 7)) 1345 printk("\n"); 1346 printk("%02x ", (unsigned char)rrpriv->tx_ring[cons].size); 1347 } 1348 printk("\n"); 1349 } 1350 1351 printk("dumping TX ring info:\n"); 1352 for (i = 0; i < TX_RING_ENTRIES; i++) 1353 printk("mode 0x%x, size 0x%x, phys-addr %08Lx\n", 1354 rrpriv->tx_ring[i].mode, 1355 rrpriv->tx_ring[i].size, 1356 (unsigned long long) rrpriv->tx_ring[i].addr.addrlo); 1357 1358} 1359 1360 1361static int rr_close(struct net_device *dev) 1362{ 1363 struct rr_private *rrpriv; 1364 struct rr_regs __iomem *regs; 1365 unsigned long flags; 1366 u32 tmp; 1367 short i; 1368 1369 netif_stop_queue(dev); 1370 1371 rrpriv = netdev_priv(dev); 1372 regs = rrpriv->regs; 1373 1374 /* 1375 * Lock to make sure we are not cleaning up while another CPU 1376 * is handling interrupts. 1377 */ 1378 spin_lock_irqsave(&rrpriv->lock, flags); 1379 1380 tmp = readl(&regs->HostCtrl); 1381 if (tmp & NIC_HALTED){ 1382 printk("%s: NIC already halted\n", dev->name); 1383 rr_dump(dev); 1384 }else{ 1385 tmp |= HALT_NIC | RR_CLEAR_INT; 1386 writel(tmp, &regs->HostCtrl); 1387 readl(&regs->HostCtrl); 1388 } 1389 1390 rrpriv->fw_running = 0; 1391 1392 del_timer_sync(&rrpriv->timer); 1393 1394 writel(0, &regs->TxPi); 1395 writel(0, &regs->IpRxPi); 1396 1397 writel(0, &regs->EvtCon); 1398 writel(0, &regs->EvtPrd); 1399 1400 for (i = 0; i < CMD_RING_ENTRIES; i++) 1401 writel(0, &regs->CmdRing[i]); 1402 1403 rrpriv->info->tx_ctrl.entries = 0; 1404 rrpriv->info->cmd_ctrl.pi = 0; 1405 rrpriv->info->evt_ctrl.pi = 0; 1406 rrpriv->rx_ctrl[4].entries = 0; 1407 1408 rr_raz_tx(rrpriv, dev); 1409 rr_raz_rx(rrpriv, dev); 1410 1411 pci_free_consistent(rrpriv->pci_dev, 256 * sizeof(struct ring_ctrl), 1412 rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma); 1413 rrpriv->rx_ctrl = NULL; 1414 1415 pci_free_consistent(rrpriv->pci_dev, sizeof(struct rr_info), 1416 rrpriv->info, rrpriv->info_dma); 1417 rrpriv->info = NULL; 1418 1419 free_irq(dev->irq, dev); 1420 spin_unlock_irqrestore(&rrpriv->lock, flags); 1421 1422 return 0; 1423} 1424 1425 1426static int rr_start_xmit(struct sk_buff *skb, struct net_device *dev) 1427{ 1428 struct rr_private *rrpriv = netdev_priv(dev); 1429 struct rr_regs __iomem *regs = rrpriv->regs; 1430 struct hippi_cb *hcb = (struct hippi_cb *) skb->cb; 1431 struct ring_ctrl *txctrl; 1432 unsigned long flags; 1433 u32 index, len = skb->len; 1434 u32 *ifield; 1435 struct sk_buff *new_skb; 1436 1437 if (readl(&regs->Mode) & FATAL_ERR) 1438 printk("error codes Fail1 %02x, Fail2 %02x\n", 1439 readl(&regs->Fail1), readl(&regs->Fail2)); 1440 1441 /* 1442 * We probably need to deal with tbusy here to prevent overruns. 1443 */ 1444 1445 if (skb_headroom(skb) < 8){ 1446 printk("incoming skb too small - reallocating\n"); 1447 if (!(new_skb = dev_alloc_skb(len + 8))) { 1448 dev_kfree_skb(skb); 1449 netif_wake_queue(dev); 1450 return -EBUSY; 1451 } 1452 skb_reserve(new_skb, 8); 1453 skb_put(new_skb, len); 1454 skb_copy_from_linear_data(skb, new_skb->data, len); 1455 dev_kfree_skb(skb); 1456 skb = new_skb; 1457 } 1458 1459 ifield = (u32 *)skb_push(skb, 8); 1460 1461 ifield[0] = 0; 1462 ifield[1] = hcb->ifield; 1463 1464 /* 1465 * We don't need the lock before we are actually going to start 1466 * fiddling with the control blocks. 1467 */ 1468 spin_lock_irqsave(&rrpriv->lock, flags); 1469 1470 txctrl = &rrpriv->info->tx_ctrl; 1471 1472 index = txctrl->pi; 1473 1474 rrpriv->tx_skbuff[index] = skb; 1475 set_rraddr(&rrpriv->tx_ring[index].addr, pci_map_single( 1476 rrpriv->pci_dev, skb->data, len + 8, PCI_DMA_TODEVICE)); 1477 rrpriv->tx_ring[index].size = len + 8; /* include IFIELD */ 1478 rrpriv->tx_ring[index].mode = PACKET_START | PACKET_END; 1479 txctrl->pi = (index + 1) % TX_RING_ENTRIES; 1480 wmb(); 1481 writel(txctrl->pi, &regs->TxPi); 1482 1483 if (txctrl->pi == rrpriv->dirty_tx){ 1484 rrpriv->tx_full = 1; 1485 netif_stop_queue(dev); 1486 } 1487 1488 spin_unlock_irqrestore(&rrpriv->lock, flags); 1489 1490 dev->trans_start = jiffies; 1491 return 0; 1492} 1493 1494 1495static struct net_device_stats *rr_get_stats(struct net_device *dev) 1496{ 1497 struct rr_private *rrpriv; 1498 1499 rrpriv = netdev_priv(dev); 1500 1501 return(&rrpriv->stats); 1502} 1503 1504 1505/* 1506 * Read the firmware out of the EEPROM and put it into the SRAM 1507 * (or from user space - later) 1508 * 1509 * This operation requires the NIC to be halted and is performed with 1510 * interrupts disabled and with the spinlock hold. 1511 */ 1512static int rr_load_firmware(struct net_device *dev) 1513{ 1514 struct rr_private *rrpriv; 1515 struct rr_regs __iomem *regs; 1516 unsigned long eptr, segptr; 1517 int i, j; 1518 u32 localctrl, sptr, len, tmp; 1519 u32 p2len, p2size, nr_seg, revision, io, sram_size; 1520 struct eeprom *hw = NULL; 1521 1522 rrpriv = netdev_priv(dev); 1523 regs = rrpriv->regs; 1524 1525 if (dev->flags & IFF_UP) 1526 return -EBUSY; 1527 1528 if (!(readl(&regs->HostCtrl) & NIC_HALTED)){ 1529 printk("%s: Trying to load firmware to a running NIC.\n", 1530 dev->name); 1531 return -EBUSY; 1532 } 1533 1534 localctrl = readl(&regs->LocalCtrl); 1535 writel(0, &regs->LocalCtrl); 1536 1537 writel(0, &regs->EvtPrd); 1538 writel(0, &regs->RxPrd); 1539 writel(0, &regs->TxPrd); 1540 1541 /* 1542 * First wipe the entire SRAM, otherwise we might run into all 1543 * kinds of trouble ... sigh, this took almost all afternoon 1544 * to track down ;-( 1545 */ 1546 io = readl(&regs->ExtIo); 1547 writel(0, &regs->ExtIo); 1548 sram_size = rr_read_eeprom_word(rrpriv, (void *)8); 1549 1550 for (i = 200; i < sram_size / 4; i++){ 1551 writel(i * 4, &regs->WinBase); 1552 mb(); 1553 writel(0, &regs->WinData); 1554 mb(); 1555 } 1556 writel(io, &regs->ExtIo); 1557 mb(); 1558 1559 eptr = (unsigned long)rr_read_eeprom_word(rrpriv, 1560 &hw->rncd_info.AddrRunCodeSegs); 1561 eptr = ((eptr & 0x1fffff) >> 3); 1562 1563 p2len = rr_read_eeprom_word(rrpriv, (void *)(0x83*4)); 1564 p2len = (p2len << 2); 1565 p2size = rr_read_eeprom_word(rrpriv, (void *)(0x84*4)); 1566 p2size = ((p2size & 0x1fffff) >> 3); 1567 1568 if ((eptr < p2size) || (eptr > (p2size + p2len))){ 1569 printk("%s: eptr is invalid\n", dev->name); 1570 goto out; 1571 } 1572 1573 revision = rr_read_eeprom_word(rrpriv, &hw->manf.HeaderFmt); 1574 1575 if (revision != 1){ 1576 printk("%s: invalid firmware format (%i)\n", 1577 dev->name, revision); 1578 goto out; 1579 } 1580 1581 nr_seg = rr_read_eeprom_word(rrpriv, (void *)eptr); 1582 eptr +=4; 1583#if (DEBUG > 1) 1584 printk("%s: nr_seg %i\n", dev->name, nr_seg); 1585#endif 1586 1587 for (i = 0; i < nr_seg; i++){ 1588 sptr = rr_read_eeprom_word(rrpriv, (void *)eptr); 1589 eptr += 4; 1590 len = rr_read_eeprom_word(rrpriv, (void *)eptr); 1591 eptr += 4; 1592 segptr = (unsigned long)rr_read_eeprom_word(rrpriv, (void *)eptr); 1593 segptr = ((segptr & 0x1fffff) >> 3); 1594 eptr += 4; 1595#if (DEBUG > 1) 1596 printk("%s: segment %i, sram address %06x, length %04x, segptr %06x\n", 1597 dev->name, i, sptr, len, segptr); 1598#endif 1599 for (j = 0; j < len; j++){ 1600 tmp = rr_read_eeprom_word(rrpriv, (void *)segptr); 1601 writel(sptr, &regs->WinBase); 1602 mb(); 1603 writel(tmp, &regs->WinData); 1604 mb(); 1605 segptr += 4; 1606 sptr += 4; 1607 } 1608 } 1609 1610out: 1611 writel(localctrl, &regs->LocalCtrl); 1612 mb(); 1613 return 0; 1614} 1615 1616 1617static int rr_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) 1618{ 1619 struct rr_private *rrpriv; 1620 unsigned char *image, *oldimage; 1621 unsigned long flags; 1622 unsigned int i; 1623 int error = -EOPNOTSUPP; 1624 1625 rrpriv = netdev_priv(dev); 1626 1627 switch(cmd){ 1628 case SIOCRRGFW: 1629 if (!capable(CAP_SYS_RAWIO)){ 1630 return -EPERM; 1631 } 1632 1633 image = kmalloc(EEPROM_WORDS * sizeof(u32), GFP_KERNEL); 1634 if (!image){ 1635 printk(KERN_ERR "%s: Unable to allocate memory " 1636 "for EEPROM image\n", dev->name); 1637 return -ENOMEM; 1638 } 1639 1640 1641 if (rrpriv->fw_running){ 1642 printk("%s: Firmware already running\n", dev->name); 1643 error = -EPERM; 1644 goto gf_out; 1645 } 1646 1647 spin_lock_irqsave(&rrpriv->lock, flags); 1648 i = rr_read_eeprom(rrpriv, 0, image, EEPROM_BYTES); 1649 spin_unlock_irqrestore(&rrpriv->lock, flags); 1650 if (i != EEPROM_BYTES){ 1651 printk(KERN_ERR "%s: Error reading EEPROM\n", 1652 dev->name); 1653 error = -EFAULT; 1654 goto gf_out; 1655 } 1656 error = copy_to_user(rq->ifr_data, image, EEPROM_BYTES); 1657 if (error) 1658 error = -EFAULT; 1659 gf_out: 1660 kfree(image); 1661 return error; 1662 1663 case SIOCRRPFW: 1664 if (!capable(CAP_SYS_RAWIO)){ 1665 return -EPERM; 1666 } 1667 1668 image = kmalloc(EEPROM_WORDS * sizeof(u32), GFP_KERNEL); 1669 oldimage = kmalloc(EEPROM_WORDS * sizeof(u32), GFP_KERNEL); 1670 if (!image || !oldimage) { 1671 printk(KERN_ERR "%s: Unable to allocate memory " 1672 "for EEPROM image\n", dev->name); 1673 error = -ENOMEM; 1674 goto wf_out; 1675 } 1676 1677 error = copy_from_user(image, rq->ifr_data, EEPROM_BYTES); 1678 if (error) { 1679 error = -EFAULT; 1680 goto wf_out; 1681 } 1682 1683 if (rrpriv->fw_running){ 1684 printk("%s: Firmware already running\n", dev->name); 1685 error = -EPERM; 1686 goto wf_out; 1687 } 1688 1689 printk("%s: Updating EEPROM firmware\n", dev->name); 1690 1691 spin_lock_irqsave(&rrpriv->lock, flags); 1692 error = write_eeprom(rrpriv, 0, image, EEPROM_BYTES); 1693 if (error) 1694 printk(KERN_ERR "%s: Error writing EEPROM\n", 1695 dev->name); 1696 1697 i = rr_read_eeprom(rrpriv, 0, oldimage, EEPROM_BYTES); 1698 spin_unlock_irqrestore(&rrpriv->lock, flags); 1699 1700 if (i != EEPROM_BYTES) 1701 printk(KERN_ERR "%s: Error reading back EEPROM " 1702 "image\n", dev->name); 1703 1704 error = memcmp(image, oldimage, EEPROM_BYTES); 1705 if (error){ 1706 printk(KERN_ERR "%s: Error verifying EEPROM image\n", 1707 dev->name); 1708 error = -EFAULT; 1709 } 1710 wf_out: 1711 kfree(oldimage); 1712 kfree(image); 1713 return error; 1714 1715 case SIOCRRID: 1716 return put_user(0x52523032, (int __user *)rq->ifr_data); 1717 default: 1718 return error; 1719 } 1720} 1721 1722static struct pci_device_id rr_pci_tbl[] = { 1723 { PCI_VENDOR_ID_ESSENTIAL, PCI_DEVICE_ID_ESSENTIAL_ROADRUNNER, 1724 PCI_ANY_ID, PCI_ANY_ID, }, 1725 { 0,} 1726}; 1727MODULE_DEVICE_TABLE(pci, rr_pci_tbl); 1728 1729static struct pci_driver rr_driver = { 1730 .name = "rrunner", 1731 .id_table = rr_pci_tbl, 1732 .probe = rr_init_one, 1733 .remove = __devexit_p(rr_remove_one), 1734}; 1735 1736static int __init rr_init_module(void) 1737{ 1738 return pci_register_driver(&rr_driver); 1739} 1740 1741static void __exit rr_cleanup_module(void) 1742{ 1743 pci_unregister_driver(&rr_driver); 1744} 1745 1746module_init(rr_init_module); 1747module_exit(rr_cleanup_module); 1748 1749/* 1750 * Local variables: 1751 * compile-command: "gcc -D__KERNEL__ -I../../include -Wall -Wstrict-prototypes -O2 -pipe -fomit-frame-pointer -fno-strength-reduce -m486 -malign-loops=2 -malign-jumps=2 -malign-functions=2 -DMODULE -DMODVERSIONS -include ../../include/linux/modversions.h -c rrunner.c" 1752 * End: 1753 */