···631631 return status;632632}633633634634-int myri10ge_get_firmware_capabilities(struct myri10ge_priv *mgp)634634+static int myri10ge_get_firmware_capabilities(struct myri10ge_priv *mgp)635635{636636 struct myri10ge_cmd cmd;637637 int status;
+3-1
drivers/net/pcmcia/fmvj18x_cs.c
···391391 cardtype = CONTEC;392392 break;393393 case MANFID_FUJITSU:394394- if (link->card_id == PRODID_FUJITSU_MBH10302)394394+ if (link->conf.ConfigBase == 0x0fe0)395395+ cardtype = MBH10302;396396+ else if (link->card_id == PRODID_FUJITSU_MBH10302) 395397 /* RATOC REX-5588/9822/4886's PRODID are 0004(=MBH10302),396398 but these are MBH10304 based card. */ 397399 cardtype = MBH10304;
+8-4
drivers/net/pcmcia/xirc2ps_cs.c
···14611461set_multicast_list(struct net_device *dev)14621462{14631463 unsigned int ioaddr = dev->base_addr;14641464+ unsigned value;1464146514651466 SelectPage(0x42);14671467+ value = GetByte(XIRCREG42_SWC1) & 0xC0;14681468+14661469 if (dev->flags & IFF_PROMISC) { /* snoop */14671467- PutByte(XIRCREG42_SWC1, 0x06); /* set MPE and PME */14701470+ PutByte(XIRCREG42_SWC1, value | 0x06); /* set MPE and PME */14681471 } else if (dev->mc_count > 9 || (dev->flags & IFF_ALLMULTI)) {14691469- PutByte(XIRCREG42_SWC1, 0x02); /* set MPE */14721472+ PutByte(XIRCREG42_SWC1, value | 0x02); /* set MPE */14701473 } else if (dev->mc_count) {14711474 /* the chip can filter 9 addresses perfectly */14721472- PutByte(XIRCREG42_SWC1, 0x01);14751475+ PutByte(XIRCREG42_SWC1, value | 0x01);14731476 SelectPage(0x40);14741477 PutByte(XIRCREG40_CMD0, Offline);14751478 set_addresses(dev);14761479 SelectPage(0x40);14771480 PutByte(XIRCREG40_CMD0, EnableRecv | Online);14781481 } else { /* standard usage */14791479- PutByte(XIRCREG42_SWC1, 0x00);14821482+ PutByte(XIRCREG42_SWC1, value | 0x00);14801483 }14811484 SelectPage(0);14821485}···1725172217261723 /* enable receiver and put the mac online */17271724 if (full) {17251725+ set_multicast_list(dev);17281726 SelectPage(0x40);17291727 PutByte(XIRCREG40_CMD0, EnableRecv | Online);17301728 }
+2-2
drivers/net/pcnet32.c
···325325static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs,326326 void *ptr);327327static void pcnet32_purge_tx_ring(struct net_device *dev);328328-static int pcnet32_alloc_ring(struct net_device *dev, char *name);328328+static int pcnet32_alloc_ring(struct net_device *dev, const char *name);329329static void pcnet32_free_ring(struct net_device *dev);330330static void pcnet32_check_media(struct net_device *dev, int verbose);331331···19831983}1984198419851985/* if any allocation fails, caller must also call pcnet32_free_ring */19861986-static int pcnet32_alloc_ring(struct net_device *dev, char *name)19861986+static int pcnet32_alloc_ring(struct net_device *dev, const char *name)19871987{19881988 struct pcnet32_private *lp = netdev_priv(dev);19891989
+1-1
drivers/net/phy/Kconfig
···55menuconfig PHYLIB66 tristate "PHY Device support and infrastructure"77 depends on !S39088- depends on NET_ETHERNET && (BROKEN || !S390)88+ depends on NET_ETHERNET99 help1010 Ethernet controllers are usually attached to PHY1111 devices. This option provides infrastructure for
+1
drivers/net/phy/phy_device.c
···207207208208 return 0;209209}210210+EXPORT_SYMBOL(get_phy_id);210211211212/**212213 * get_phy_device - reads the specified PHY device and returns its @phy_device struct
···199199 */200200static inline void efx_channel_processed(struct efx_channel *channel)201201{202202- /* Write to EVQ_RPTR_REG. If a new event arrived in a race203203- * with finishing processing, a new interrupt will be raised.204204- */202202+ /* The interrupt handler for this channel may set work_pending203203+ * as soon as we acknowledge the events we've seen. Make sure204204+ * it's cleared before then. */205205 channel->work_pending = 0;206206- smp_wmb(); /* Ensure channel updated before any new interrupt. */206206+ smp_wmb();207207+207208 falcon_eventq_read_ack(channel);208209}209210···266265 napi_disable(&channel->napi_str);267266268267 /* Poll the channel */269269- (void) efx_process_channel(channel, efx->type->evq_size);268268+ efx_process_channel(channel, efx->type->evq_size);270269271270 /* Ack the eventq. This may cause an interrupt to be generated272271 * when they are reenabled */···318317 *319318 *************************************************************************/320319321321-/* Setup per-NIC RX buffer parameters.322322- * Calculate the rx buffer allocation parameters required to support323323- * the current MTU, including padding for header alignment and overruns.324324- */325325-static void efx_calc_rx_buffer_params(struct efx_nic *efx)326326-{327327- unsigned int order, len;328328-329329- len = (max(EFX_PAGE_IP_ALIGN, NET_IP_ALIGN) +330330- EFX_MAX_FRAME_LEN(efx->net_dev->mtu) +331331- efx->type->rx_buffer_padding);332332-333333- /* Calculate page-order */334334- for (order = 0; ((1u << order) * PAGE_SIZE) < len; ++order)335335- ;336336-337337- efx->rx_buffer_len = len;338338- efx->rx_buffer_order = order;339339-}340340-341320static int efx_probe_channel(struct efx_channel *channel)342321{343322 struct efx_tx_queue *tx_queue;···368387 struct efx_channel *channel;369388 int rc = 0;370389371371- efx_calc_rx_buffer_params(efx);390390+ /* Calculate the rx buffer allocation parameters required to391391+ * support the current MTU, including padding for header392392+ * alignment and overruns.393393+ */394394+ efx->rx_buffer_len = (max(EFX_PAGE_IP_ALIGN, NET_IP_ALIGN) +395395+ EFX_MAX_FRAME_LEN(efx->net_dev->mtu) +396396+ efx->type->rx_buffer_padding);397397+ efx->rx_buffer_order = get_order(efx->rx_buffer_len);372398373399 /* Initialise the channels */374400 efx_for_each_channel(channel, efx) {···428440 netif_napi_add(channel->napi_dev, &channel->napi_str,429441 efx_poll, napi_weight);430442443443+ /* The interrupt handler for this channel may set work_pending444444+ * as soon as we enable it. Make sure it's cleared before445445+ * then. Similarly, make sure it sees the enabled flag set. */431446 channel->work_pending = 0;432447 channel->enabled = 1;433433- smp_wmb(); /* ensure channel updated before first interrupt */448448+ smp_wmb();434449435450 napi_enable(&channel->napi_str);436451···695704 mutex_unlock(&efx->mac_lock);696705697706 /* Serialise against efx_set_multicast_list() */698698- if (NET_DEV_REGISTERED(efx)) {707707+ if (efx_dev_registered(efx)) {699708 netif_tx_lock_bh(efx->net_dev);700709 netif_tx_unlock_bh(efx->net_dev);701710 }···782791 efx->membase = ioremap_nocache(efx->membase_phys,783792 efx->type->mem_map_size);784793 if (!efx->membase) {785785- EFX_ERR(efx, "could not map memory BAR %d at %lx+%x\n",786786- efx->type->mem_bar, efx->membase_phys,794794+ EFX_ERR(efx, "could not map memory BAR %d at %llx+%x\n",795795+ efx->type->mem_bar,796796+ (unsigned long long)efx->membase_phys,787797 efx->type->mem_map_size);788798 rc = -ENOMEM;789799 goto fail4;790800 }791791- EFX_LOG(efx, "memory BAR %u at %lx+%x (virtual %p)\n",792792- efx->type->mem_bar, efx->membase_phys, efx->type->mem_map_size,793793- efx->membase);801801+ EFX_LOG(efx, "memory BAR %u at %llx+%x (virtual %p)\n",802802+ efx->type->mem_bar, (unsigned long long)efx->membase_phys,803803+ efx->type->mem_map_size, efx->membase);794804795805 return 0;796806797807 fail4:798808 release_mem_region(efx->membase_phys, efx->type->mem_map_size);799809 fail3:800800- efx->membase_phys = 0UL;810810+ efx->membase_phys = 0;801811 fail2:802812 pci_disable_device(efx->pci_dev);803813 fail1:···816824817825 if (efx->membase_phys) {818826 pci_release_region(efx->pci_dev, efx->type->mem_bar);819819- efx->membase_phys = 0UL;827827+ efx->membase_phys = 0;820828 }821829822830 pci_disable_device(efx->pci_dev);···10351043 return;10361044 if ((efx->state != STATE_RUNNING) && (efx->state != STATE_INIT))10371045 return;10381038- if (NET_DEV_REGISTERED(efx) && !netif_running(efx->net_dev))10461046+ if (efx_dev_registered(efx) && !netif_running(efx->net_dev))10391047 return;1040104810411049 /* Mark the port as enabled so port reconfigurations can start, then···10651073 cancel_delayed_work_sync(&efx->monitor_work);1066107410671075 /* Ensure that all RX slow refills are complete. */10681068- efx_for_each_rx_queue(rx_queue, efx) {10761076+ efx_for_each_rx_queue(rx_queue, efx)10691077 cancel_delayed_work_sync(&rx_queue->work);10701070- }1071107810721079 /* Stop scheduled port reconfigurations */10731080 cancel_work_sync(&efx->reconfigure_work);···10921101 falcon_disable_interrupts(efx);10931102 if (efx->legacy_irq)10941103 synchronize_irq(efx->legacy_irq);10951095- efx_for_each_channel_with_interrupt(channel, efx)11041104+ efx_for_each_channel_with_interrupt(channel, efx) {10961105 if (channel->irq)10971106 synchronize_irq(channel->irq);11071107+ }1098110810991109 /* Stop all NAPI processing and synchronous rx refills */11001110 efx_for_each_channel(channel, efx)···11171125 /* Stop the kernel transmit interface late, so the watchdog11181126 * timer isn't ticking over the flush */11191127 efx_stop_queue(efx);11201120- if (NET_DEV_REGISTERED(efx)) {11281128+ if (efx_dev_registered(efx)) {11211129 netif_tx_lock_bh(efx->net_dev);11221130 netif_tx_unlock_bh(efx->net_dev);11231131 }···13361344 return 0;13371345}1338134613391339-/* Context: process, dev_base_lock held, non-blocking. */13471347+/* Context: process, dev_base_lock or RTNL held, non-blocking. */13401348static struct net_device_stats *efx_net_stats(struct net_device *net_dev)13411349{13421350 struct efx_nic *efx = net_dev->priv;13431351 struct efx_mac_stats *mac_stats = &efx->mac_stats;13441352 struct net_device_stats *stats = &net_dev->stats;1345135313541354+ /* Update stats if possible, but do not wait if another thread13551355+ * is updating them (or resetting the NIC); slightly stale13561356+ * stats are acceptable.13571357+ */13461358 if (!spin_trylock(&efx->stats_lock))13471359 return stats;13481360 if (efx->state == STATE_RUNNING) {···14901494static int efx_netdev_event(struct notifier_block *this,14911495 unsigned long event, void *ptr)14921496{14931493- struct net_device *net_dev = (struct net_device *)ptr;14971497+ struct net_device *net_dev = ptr;1494149814951499 if (net_dev->open == efx_net_open && event == NETDEV_CHANGENAME) {14961500 struct efx_nic *efx = net_dev->priv;···15591563 efx_for_each_tx_queue(tx_queue, efx)15601564 efx_release_tx_buffers(tx_queue);1561156515621562- if (NET_DEV_REGISTERED(efx)) {15661566+ if (efx_dev_registered(efx)) {15631567 strlcpy(efx->name, pci_name(efx->pci_dev), sizeof(efx->name));15641568 unregister_netdev(efx->net_dev);15651569 }···16841688 if (method == RESET_TYPE_DISABLE) {16851689 /* Reinitialise the device anyway so the driver unload sequence16861690 * can talk to the external SRAM */16871687- (void) falcon_init_nic(efx);16911691+ falcon_init_nic(efx);16881692 rc = -EIO;16891693 goto fail4;16901694 }
+38-49
drivers/net/sfc/falcon.c
···116116 **************************************************************************117117 */118118119119-/* DMA address mask (up to 46-bit, avoiding compiler warnings)120120- *121121- * Note that it is possible to have a platform with 64-bit longs and122122- * 32-bit DMA addresses, or vice versa. EFX_DMA_MASK takes care of the123123- * platform DMA mask.124124- */125125-#if BITS_PER_LONG == 64126126-#define FALCON_DMA_MASK EFX_DMA_MASK(0x00003fffffffffffUL)127127-#else128128-#define FALCON_DMA_MASK EFX_DMA_MASK(0x00003fffffffffffULL)129129-#endif119119+/* DMA address mask */120120+#define FALCON_DMA_MASK DMA_BIT_MASK(46)130121131122/* TX DMA length mask (13-bit) */132123#define FALCON_TX_DMA_MASK (4096 - 1)···136145#define PCI_EXP_LNKSTA_LNK_WID_LBN 4137146138147#define FALCON_IS_DUAL_FUNC(efx) \139139- (FALCON_REV(efx) < FALCON_REV_B0)148148+ (falcon_rev(efx) < FALCON_REV_B0)140149141150/**************************************************************************142151 *···456465 TX_DESCQ_TYPE, 0,457466 TX_NON_IP_DROP_DIS_B0, 1);458467459459- if (FALCON_REV(efx) >= FALCON_REV_B0) {468468+ if (falcon_rev(efx) >= FALCON_REV_B0) {460469 int csum = !(efx->net_dev->features & NETIF_F_IP_CSUM);461470 EFX_SET_OWORD_FIELD(tx_desc_ptr, TX_IP_CHKSM_DIS_B0, csum);462471 EFX_SET_OWORD_FIELD(tx_desc_ptr, TX_TCP_CHKSM_DIS_B0, csum);···465474 falcon_write_table(efx, &tx_desc_ptr, efx->type->txd_ptr_tbl_base,466475 tx_queue->queue);467476468468- if (FALCON_REV(efx) < FALCON_REV_B0) {477477+ if (falcon_rev(efx) < FALCON_REV_B0) {469478 efx_oword_t reg;470479471480 BUG_ON(tx_queue->queue >= 128); /* HW limit */···626635 efx_oword_t rx_desc_ptr;627636 struct efx_nic *efx = rx_queue->efx;628637 int rc;629629- int is_b0 = FALCON_REV(efx) >= FALCON_REV_B0;638638+ int is_b0 = falcon_rev(efx) >= FALCON_REV_B0;630639 int iscsi_digest_en = is_b0;631640632641 EFX_LOG(efx, "RX queue %d ring in special buffers %d-%d\n",···813822 tx_ev_q_label = EFX_QWORD_FIELD(*event, TX_EV_Q_LABEL);814823 tx_queue = &efx->tx_queue[tx_ev_q_label];815824816816- if (NET_DEV_REGISTERED(efx))825825+ if (efx_dev_registered(efx))817826 netif_tx_lock(efx->net_dev);818827 falcon_notify_tx_desc(tx_queue);819819- if (NET_DEV_REGISTERED(efx))828828+ if (efx_dev_registered(efx))820829 netif_tx_unlock(efx->net_dev);821830 } else if (EFX_QWORD_FIELD(*event, TX_EV_PKT_ERR) &&822831 EFX_WORKAROUND_10727(efx)) {···875884 RX_EV_TCP_UDP_CHKSUM_ERR);876885 rx_ev_eth_crc_err = EFX_QWORD_FIELD(*event, RX_EV_ETH_CRC_ERR);877886 rx_ev_frm_trunc = EFX_QWORD_FIELD(*event, RX_EV_FRM_TRUNC);878878- rx_ev_drib_nib = ((FALCON_REV(efx) >= FALCON_REV_B0) ?887887+ rx_ev_drib_nib = ((falcon_rev(efx) >= FALCON_REV_B0) ?879888 0 : EFX_QWORD_FIELD(*event, RX_EV_DRIB_NIB));880889 rx_ev_pause_frm = EFX_QWORD_FIELD(*event, RX_EV_PAUSE_FRM_ERR);881890···10561065 EFX_QWORD_FIELD(*event, XG_PHY_INTR))10571066 is_phy_event = 1;1058106710591059- if ((FALCON_REV(efx) >= FALCON_REV_B0) &&10681068+ if ((falcon_rev(efx) >= FALCON_REV_B0) &&10601069 EFX_OWORD_FIELD(*event, XG_MNT_INTR_B0))10611070 is_phy_event = 1;10621071···13961405static irqreturn_t falcon_fatal_interrupt(struct efx_nic *efx)13971406{13981407 struct falcon_nic_data *nic_data = efx->nic_data;13991399- efx_oword_t *int_ker = (efx_oword_t *) efx->irq_status.addr;14081408+ efx_oword_t *int_ker = efx->irq_status.addr;14001409 efx_oword_t fatal_intr;14011410 int error, mem_perr;14021411 static int n_int_errors;···14421451 */14431452static irqreturn_t falcon_legacy_interrupt_b0(int irq, void *dev_id)14441453{14451445- struct efx_nic *efx = (struct efx_nic *)dev_id;14461446- efx_oword_t *int_ker = (efx_oword_t *) efx->irq_status.addr;14541454+ struct efx_nic *efx = dev_id;14551455+ efx_oword_t *int_ker = efx->irq_status.addr;14471456 struct efx_channel *channel;14481457 efx_dword_t reg;14491458 u32 queues;···1480148914811490static irqreturn_t falcon_legacy_interrupt_a1(int irq, void *dev_id)14821491{14831483- struct efx_nic *efx = (struct efx_nic *)dev_id;14841484- efx_oword_t *int_ker = (efx_oword_t *) efx->irq_status.addr;14921492+ struct efx_nic *efx = dev_id;14931493+ efx_oword_t *int_ker = efx->irq_status.addr;14851494 struct efx_channel *channel;14861495 int syserr;14871496 int queues;···15331542 */15341543static irqreturn_t falcon_msi_interrupt(int irq, void *dev_id)15351544{15361536- struct efx_channel *channel = (struct efx_channel *)dev_id;15451545+ struct efx_channel *channel = dev_id;15371546 struct efx_nic *efx = channel->efx;15381538- efx_oword_t *int_ker = (efx_oword_t *) efx->irq_status.addr;15471547+ efx_oword_t *int_ker = efx->irq_status.addr;15391548 int syserr;1540154915411550 efx->last_irq_cpu = raw_smp_processor_id();···15631572 unsigned long offset;15641573 efx_dword_t dword;1565157415661566- if (FALCON_REV(efx) < FALCON_REV_B0)15751575+ if (falcon_rev(efx) < FALCON_REV_B0)15671576 return;1568157715691578 for (offset = RX_RSS_INDIR_TBL_B0;···1586159515871596 if (!EFX_INT_MODE_USE_MSI(efx)) {15881597 irq_handler_t handler;15891589- if (FALCON_REV(efx) >= FALCON_REV_B0)15981598+ if (falcon_rev(efx) >= FALCON_REV_B0)15901599 handler = falcon_legacy_interrupt_b0;15911600 else15921601 handler = falcon_legacy_interrupt_a1;···16271636 efx_oword_t reg;1628163716291638 /* Disable MSI/MSI-X interrupts */16301630- efx_for_each_channel_with_interrupt(channel, efx)16391639+ efx_for_each_channel_with_interrupt(channel, efx) {16311640 if (channel->irq)16321641 free_irq(channel->irq, channel);16421642+ }1633164316341644 /* ACK legacy interrupt */16351635- if (FALCON_REV(efx) >= FALCON_REV_B0)16451645+ if (falcon_rev(efx) >= FALCON_REV_B0)16361646 falcon_read(efx, ®, INT_ISR0_B0);16371647 else16381648 falcon_irq_ack_a1(efx);···17241732 efx_oword_t temp;17251733 int count;1726173417271727- if ((FALCON_REV(efx) < FALCON_REV_B0) ||17351735+ if ((falcon_rev(efx) < FALCON_REV_B0) ||17281736 (efx->loopback_mode != LOOPBACK_NONE))17291737 return;17301738···17771785{17781786 efx_oword_t temp;1779178717801780- if (FALCON_REV(efx) < FALCON_REV_B0)17881788+ if (falcon_rev(efx) < FALCON_REV_B0)17811789 return;1782179017831791 /* Isolate the MAC -> RX */···18151823 MAC_SPEED, link_speed);18161824 /* On B0, MAC backpressure can be disabled and packets get18171825 * discarded. */18181818- if (FALCON_REV(efx) >= FALCON_REV_B0) {18261826+ if (falcon_rev(efx) >= FALCON_REV_B0) {18191827 EFX_SET_OWORD_FIELD(reg, TXFIFO_DRAIN_EN_B0,18201828 !efx->link_up);18211829 }···18331841 EFX_SET_OWORD_FIELD_VER(efx, reg, RX_XOFF_MAC_EN, tx_fc);1834184218351843 /* Unisolate the MAC -> RX */18361836- if (FALCON_REV(efx) >= FALCON_REV_B0)18441844+ if (falcon_rev(efx) >= FALCON_REV_B0)18371845 EFX_SET_OWORD_FIELD(reg, RX_INGR_EN_B0, 1);18381846 falcon_write(efx, ®, RX_CFG_REG_KER);18391847}···18481856 return 0;1849185718501858 /* Statistics fetch will fail if the MAC is in TX drain */18511851- if (FALCON_REV(efx) >= FALCON_REV_B0) {18591859+ if (falcon_rev(efx) >= FALCON_REV_B0) {18521860 efx_oword_t temp;18531861 falcon_read(efx, &temp, MAC0_CTRL_REG_KER);18541862 if (EFX_OWORD_FIELD(temp, TXFIFO_DRAIN_EN_B0))···19321940static void falcon_mdio_write(struct net_device *net_dev, int phy_id,19331941 int addr, int value)19341942{19351935- struct efx_nic *efx = (struct efx_nic *)net_dev->priv;19431943+ struct efx_nic *efx = net_dev->priv;19361944 unsigned int phy_id2 = phy_id & FALCON_PHY_ID_ID_MASK;19371945 efx_oword_t reg;19381946···20002008 * could be read, -1 will be returned. */20012009static int falcon_mdio_read(struct net_device *net_dev, int phy_id, int addr)20022010{20032003- struct efx_nic *efx = (struct efx_nic *)net_dev->priv;20112011+ struct efx_nic *efx = net_dev->priv;20042012 unsigned int phy_addr = phy_id & FALCON_PHY_ID_ID_MASK;20052013 efx_oword_t reg;20062014 int value = -1;···21052113 falcon_init_mdio(&efx->mii);2106211421072115 /* Hardware flow ctrl. FalconA RX FIFO too small for pause generation */21082108- if (FALCON_REV(efx) >= FALCON_REV_B0)21162116+ if (falcon_rev(efx) >= FALCON_REV_B0)21092117 efx->flow_control = EFX_FC_RX | EFX_FC_TX;21102118 else21112119 efx->flow_control = EFX_FC_RX;···23652373 return -ENODEV;23662374 }2367237523682368- switch (FALCON_REV(efx)) {23762376+ switch (falcon_rev(efx)) {23692377 case FALCON_REV_A0:23702378 case 0xff:23712379 EFX_ERR(efx, "Falcon rev A0 not supported\n");···23912399 break;2392240023932401 default:23942394- EFX_ERR(efx, "Unknown Falcon rev %d\n", FALCON_REV(efx));24022402+ EFX_ERR(efx, "Unknown Falcon rev %d\n", falcon_rev(efx));23952403 return -ENODEV;23962404 }23972405···2411241924122420 /* Allocate storage for hardware specific data */24132421 nic_data = kzalloc(sizeof(*nic_data), GFP_KERNEL);24142414- efx->nic_data = (void *) nic_data;24222422+ efx->nic_data = nic_data;2415242324162424 /* Determine number of ports etc. */24172425 rc = falcon_probe_nic_variant(efx);···24812489 */24822490int falcon_init_nic(struct efx_nic *efx)24832491{24842484- struct falcon_nic_data *data;24852492 efx_oword_t temp;24862493 unsigned thresh;24872494 int rc;24882488-24892489- data = (struct falcon_nic_data *)efx->nic_data;2490249524912496 /* Set up the address region register. This is only needed24922497 * for the B0 FPGA, but since we are just pushing in the···2551256225522563 /* Set number of RSS queues for receive path. */25532564 falcon_read(efx, &temp, RX_FILTER_CTL_REG);25542554- if (FALCON_REV(efx) >= FALCON_REV_B0)25652565+ if (falcon_rev(efx) >= FALCON_REV_B0)25552566 EFX_SET_OWORD_FIELD(temp, NUM_KER, 0);25562567 else25572568 EFX_SET_OWORD_FIELD(temp, NUM_KER, efx->rss_queues - 1);···25892600 /* Prefetch threshold 2 => fetch when descriptor cache half empty */25902601 EFX_SET_OWORD_FIELD(temp, TX_PREF_THRESHOLD, 2);25912602 /* Squash TX of packets of 16 bytes or less */25922592- if (FALCON_REV(efx) >= FALCON_REV_B0 && EFX_WORKAROUND_9141(efx))26032603+ if (falcon_rev(efx) >= FALCON_REV_B0 && EFX_WORKAROUND_9141(efx))25932604 EFX_SET_OWORD_FIELD(temp, TX_FLUSH_MIN_LEN_EN_B0, 1);25942605 falcon_write(efx, &temp, TX_CFG2_REG_KER);25952606···26062617 if (EFX_WORKAROUND_7575(efx))26072618 EFX_SET_OWORD_FIELD_VER(efx, temp, RX_USR_BUF_SIZE,26082619 (3 * 4096) / 32);26092609- if (FALCON_REV(efx) >= FALCON_REV_B0)26202620+ if (falcon_rev(efx) >= FALCON_REV_B0)26102621 EFX_SET_OWORD_FIELD(temp, RX_INGR_EN_B0, 1);2611262226122623 /* RX FIFO flow control thresholds */···26222633 falcon_write(efx, &temp, RX_CFG_REG_KER);2623263426242635 /* Set destination of both TX and RX Flush events */26252625- if (FALCON_REV(efx) >= FALCON_REV_B0) {26362636+ if (falcon_rev(efx) >= FALCON_REV_B0) {26262637 EFX_POPULATE_OWORD_1(temp, FLS_EVQ_ID, 0);26272638 falcon_write(efx, &temp, DP_CTRL_REG);26282639 }···2636264726372648 falcon_free_buffer(efx, &efx->irq_status);2638264926392639- (void) falcon_reset_hw(efx, RESET_TYPE_ALL);26502650+ falcon_reset_hw(efx, RESET_TYPE_ALL);2640265126412652 /* Release the second function after the reset */26422653 if (nic_data->pci_dev2) {