Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

net: bcmgenet: re-align multiple lines correctly

checkpatch.pl flagged a lot of "CHECK: Alignment should match open
parenthesis" checks, fix all of them to make the driver neater. While
at it fix some obvious typos and re-arrange some of the lines to avoid
going over 80 columns.

Signed-off-by: Florian Fainelli <f.fainelli@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>

authored by

Florian Fainelli and committed by
David S. Miller
c91b7f66 5e811b39

+142 -147
+126 -131
drivers/net/ethernet/broadcom/genet/bcmgenet.c
··· 70 70 TOTAL_DESC * DMA_DESC_SIZE) 71 71 72 72 static inline void dmadesc_set_length_status(struct bcmgenet_priv *priv, 73 - void __iomem *d, u32 value) 73 + void __iomem *d, u32 value) 74 74 { 75 75 __raw_writel(value, d + DMA_DESC_LENGTH_STATUS); 76 76 } 77 77 78 78 static inline u32 dmadesc_get_length_status(struct bcmgenet_priv *priv, 79 - void __iomem *d) 79 + void __iomem *d) 80 80 { 81 81 return __raw_readl(d + DMA_DESC_LENGTH_STATUS); 82 82 } ··· 99 99 100 100 /* Combined address + length/status setter */ 101 101 static inline void dmadesc_set(struct bcmgenet_priv *priv, 102 - void __iomem *d, dma_addr_t addr, u32 val) 102 + void __iomem *d, dma_addr_t addr, u32 val) 103 103 { 104 104 dmadesc_set_length_status(priv, d, val); 105 105 dmadesc_set_addr(priv, d, addr); ··· 233 233 } 234 234 235 235 static inline u32 bcmgenet_tdma_readl(struct bcmgenet_priv *priv, 236 - enum dma_reg r) 236 + enum dma_reg r) 237 237 { 238 238 return __raw_readl(priv->base + GENET_TDMA_REG_OFF + 239 239 DMA_RINGS_SIZE + bcmgenet_dma_regs[r]); ··· 247 247 } 248 248 249 249 static inline u32 bcmgenet_rdma_readl(struct bcmgenet_priv *priv, 250 - enum dma_reg r) 250 + enum dma_reg r) 251 251 { 252 252 return __raw_readl(priv->base + GENET_RDMA_REG_OFF + 253 253 DMA_RINGS_SIZE + bcmgenet_dma_regs[r]); ··· 324 324 static const u8 *genet_dma_ring_regs; 325 325 326 326 static inline u32 bcmgenet_tdma_ring_readl(struct bcmgenet_priv *priv, 327 - unsigned int ring, 328 - enum dma_ring_reg r) 327 + unsigned int ring, 328 + enum dma_ring_reg r) 329 329 { 330 330 return __raw_readl(priv->base + GENET_TDMA_REG_OFF + 331 331 (DMA_RING_SIZE * ring) + ··· 333 333 } 334 334 335 335 static inline void bcmgenet_tdma_ring_writel(struct bcmgenet_priv *priv, 336 - unsigned int ring, 337 - u32 val, 338 - enum dma_ring_reg r) 336 + unsigned int ring, u32 val, 337 + enum dma_ring_reg r) 339 338 { 340 339 __raw_writel(val, priv->base + GENET_TDMA_REG_OFF + 341 340 (DMA_RING_SIZE * ring) + ··· 342 343 } 343 344 344 345 static inline u32 bcmgenet_rdma_ring_readl(struct bcmgenet_priv *priv, 345 - unsigned int ring, 346 - enum dma_ring_reg r) 346 + unsigned int ring, 347 + enum dma_ring_reg r) 347 348 { 348 349 return __raw_readl(priv->base + GENET_RDMA_REG_OFF + 349 350 (DMA_RING_SIZE * ring) + ··· 351 352 } 352 353 353 354 static inline void bcmgenet_rdma_ring_writel(struct bcmgenet_priv *priv, 354 - unsigned int ring, 355 - u32 val, 356 - enum dma_ring_reg r) 355 + unsigned int ring, u32 val, 356 + enum dma_ring_reg r) 357 357 { 358 358 __raw_writel(val, priv->base + GENET_RDMA_REG_OFF + 359 359 (DMA_RING_SIZE * ring) + ··· 360 362 } 361 363 362 364 static int bcmgenet_get_settings(struct net_device *dev, 363 - struct ethtool_cmd *cmd) 365 + struct ethtool_cmd *cmd) 364 366 { 365 367 struct bcmgenet_priv *priv = netdev_priv(dev); 366 368 ··· 374 376 } 375 377 376 378 static int bcmgenet_set_settings(struct net_device *dev, 377 - struct ethtool_cmd *cmd) 379 + struct ethtool_cmd *cmd) 378 380 { 379 381 struct bcmgenet_priv *priv = netdev_priv(dev); 380 382 ··· 447 449 } 448 450 449 451 static int bcmgenet_set_features(struct net_device *dev, 450 - netdev_features_t features) 452 + netdev_features_t features) 451 453 { 452 454 netdev_features_t changed = features ^ dev->features; 453 455 netdev_features_t wanted = dev->wanted_features; ··· 614 616 #define BCMGENET_STATS_LEN ARRAY_SIZE(bcmgenet_gstrings_stats) 615 617 616 618 static void bcmgenet_get_drvinfo(struct net_device *dev, 617 - struct ethtool_drvinfo *info) 619 + struct ethtool_drvinfo *info) 618 620 { 619 621 strlcpy(info->driver, "bcmgenet", sizeof(info->driver)); 620 622 strlcpy(info->version, "v2.0", sizeof(info->version)); ··· 632 634 } 633 635 } 634 636 635 - static void bcmgenet_get_strings(struct net_device *dev, 636 - u32 stringset, u8 *data) 637 + static void bcmgenet_get_strings(struct net_device *dev, u32 stringset, 638 + u8 *data) 637 639 { 638 640 int i; 639 641 ··· 641 643 case ETH_SS_STATS: 642 644 for (i = 0; i < BCMGENET_STATS_LEN; i++) { 643 645 memcpy(data + i * ETH_GSTRING_LEN, 644 - bcmgenet_gstrings_stats[i].stat_string, 645 - ETH_GSTRING_LEN); 646 + bcmgenet_gstrings_stats[i].stat_string, 647 + ETH_GSTRING_LEN); 646 648 } 647 649 break; 648 650 } ··· 667 669 case BCMGENET_STAT_RUNT: 668 670 if (s->type != BCMGENET_STAT_MIB_RX) 669 671 offset = BCMGENET_STAT_OFFSET; 670 - val = bcmgenet_umac_readl(priv, UMAC_MIB_START + 671 - j + offset); 672 + val = bcmgenet_umac_readl(priv, 673 + UMAC_MIB_START + j + offset); 672 674 break; 673 675 case BCMGENET_STAT_MISC: 674 676 val = bcmgenet_umac_readl(priv, s->reg_offset); ··· 685 687 } 686 688 687 689 static void bcmgenet_get_ethtool_stats(struct net_device *dev, 688 - struct ethtool_stats *stats, 689 - u64 *data) 690 + struct ethtool_stats *stats, 691 + u64 *data) 690 692 { 691 693 struct bcmgenet_priv *priv = netdev_priv(dev); 692 694 int i; ··· 754 756 } 755 757 756 758 static void bcmgenet_power_up(struct bcmgenet_priv *priv, 757 - enum bcmgenet_power_mode mode) 759 + enum bcmgenet_power_mode mode) 758 760 { 759 761 u32 reg; 760 762 ··· 839 841 struct bcmgenet_tx_ring *ring) 840 842 { 841 843 bcmgenet_intrl2_0_writel(priv, 842 - UMAC_IRQ_TXDMA_BDONE | UMAC_IRQ_TXDMA_PDONE, 843 - INTRL2_CPU_MASK_SET); 844 + UMAC_IRQ_TXDMA_BDONE | UMAC_IRQ_TXDMA_PDONE, 845 + INTRL2_CPU_MASK_SET); 844 846 } 845 847 846 848 static inline void bcmgenet_tx_ring16_int_enable(struct bcmgenet_priv *priv, 847 849 struct bcmgenet_tx_ring *ring) 848 850 { 849 851 bcmgenet_intrl2_0_writel(priv, 850 - UMAC_IRQ_TXDMA_BDONE | UMAC_IRQ_TXDMA_PDONE, 851 - INTRL2_CPU_MASK_CLEAR); 852 + UMAC_IRQ_TXDMA_BDONE | UMAC_IRQ_TXDMA_PDONE, 853 + INTRL2_CPU_MASK_CLEAR); 852 854 } 853 855 854 856 static inline void bcmgenet_tx_ring_int_enable(struct bcmgenet_priv *priv, 855 - struct bcmgenet_tx_ring *ring) 857 + struct bcmgenet_tx_ring *ring) 856 858 { 857 - bcmgenet_intrl2_1_writel(priv, 858 - (1 << ring->index), INTRL2_CPU_MASK_CLEAR); 859 + bcmgenet_intrl2_1_writel(priv, (1 << ring->index), 860 + INTRL2_CPU_MASK_CLEAR); 859 861 priv->int1_mask &= ~(1 << ring->index); 860 862 } 861 863 862 864 static inline void bcmgenet_tx_ring_int_disable(struct bcmgenet_priv *priv, 863 865 struct bcmgenet_tx_ring *ring) 864 866 { 865 - bcmgenet_intrl2_1_writel(priv, 866 - (1 << ring->index), INTRL2_CPU_MASK_SET); 867 + bcmgenet_intrl2_1_writel(priv, (1 << ring->index), 868 + INTRL2_CPU_MASK_SET); 867 869 priv->int1_mask |= (1 << ring->index); 868 870 } 869 871 870 872 /* Unlocked version of the reclaim routine */ 871 873 static void __bcmgenet_tx_reclaim(struct net_device *dev, 872 - struct bcmgenet_tx_ring *ring) 874 + struct bcmgenet_tx_ring *ring) 873 875 { 874 876 struct bcmgenet_priv *priv = netdev_priv(dev); 875 877 int last_tx_cn, last_c_index, num_tx_bds; ··· 892 894 last_tx_cn = num_tx_bds - last_c_index + c_index; 893 895 894 896 netif_dbg(priv, tx_done, dev, 895 - "%s ring=%d index=%d last_tx_cn=%d last_index=%d\n", 896 - __func__, ring->index, 897 - c_index, last_tx_cn, last_c_index); 897 + "%s ring=%d index=%d last_tx_cn=%d last_index=%d\n", 898 + __func__, ring->index, 899 + c_index, last_tx_cn, last_c_index); 898 900 899 901 /* Reclaim transmitted buffers */ 900 902 while (last_tx_cn-- > 0) { ··· 902 904 if (tx_cb_ptr->skb) { 903 905 dev->stats.tx_bytes += tx_cb_ptr->skb->len; 904 906 dma_unmap_single(&dev->dev, 905 - dma_unmap_addr(tx_cb_ptr, dma_addr), 906 - tx_cb_ptr->skb->len, 907 - DMA_TO_DEVICE); 907 + dma_unmap_addr(tx_cb_ptr, dma_addr), 908 + tx_cb_ptr->skb->len, 909 + DMA_TO_DEVICE); 908 910 bcmgenet_free_cb(tx_cb_ptr); 909 911 } else if (dma_unmap_addr(tx_cb_ptr, dma_addr)) { 910 912 dev->stats.tx_bytes += 911 913 dma_unmap_len(tx_cb_ptr, dma_len); 912 914 dma_unmap_page(&dev->dev, 913 - dma_unmap_addr(tx_cb_ptr, dma_addr), 914 - dma_unmap_len(tx_cb_ptr, dma_len), 915 - DMA_TO_DEVICE); 915 + dma_unmap_addr(tx_cb_ptr, dma_addr), 916 + dma_unmap_len(tx_cb_ptr, dma_len), 917 + DMA_TO_DEVICE); 916 918 dma_unmap_addr_set(tx_cb_ptr, dma_addr, 0); 917 919 } 918 920 dev->stats.tx_packets++; ··· 932 934 } 933 935 934 936 static void bcmgenet_tx_reclaim(struct net_device *dev, 935 - struct bcmgenet_tx_ring *ring) 937 + struct bcmgenet_tx_ring *ring) 936 938 { 937 939 unsigned long flags; 938 940 ··· 1008 1010 1009 1011 /* Transmit a SKB fragement */ 1010 1012 static int bcmgenet_xmit_frag(struct net_device *dev, 1011 - skb_frag_t *frag, 1012 - u16 dma_desc_flags, 1013 - struct bcmgenet_tx_ring *ring) 1013 + skb_frag_t *frag, 1014 + u16 dma_desc_flags, 1015 + struct bcmgenet_tx_ring *ring) 1014 1016 { 1015 1017 struct bcmgenet_priv *priv = netdev_priv(dev); 1016 1018 struct device *kdev = &priv->pdev->dev; ··· 1025 1027 tx_cb_ptr->skb = NULL; 1026 1028 1027 1029 mapping = skb_frag_dma_map(kdev, frag, 0, 1028 - skb_frag_size(frag), DMA_TO_DEVICE); 1030 + skb_frag_size(frag), DMA_TO_DEVICE); 1029 1031 ret = dma_mapping_error(kdev, mapping); 1030 1032 if (ret) { 1031 1033 netif_err(priv, tx_err, dev, "%s: Tx DMA map failed\n", 1032 - __func__); 1034 + __func__); 1033 1035 return ret; 1034 1036 } 1035 1037 ··· 1037 1039 dma_unmap_len_set(tx_cb_ptr, dma_len, frag->size); 1038 1040 1039 1041 dmadesc_set(priv, tx_cb_ptr->bd_addr, mapping, 1040 - (frag->size << DMA_BUFLENGTH_SHIFT) | dma_desc_flags | 1041 - (priv->hw_params->qtag_mask << DMA_TX_QTAG_SHIFT)); 1042 + (frag->size << DMA_BUFLENGTH_SHIFT) | dma_desc_flags | 1043 + (priv->hw_params->qtag_mask << DMA_TX_QTAG_SHIFT)); 1042 1044 1043 1045 1044 1046 ring->free_bds -= 1; ··· 1142 1144 if (ring->free_bds <= nr_frags + 1) { 1143 1145 netif_tx_stop_queue(txq); 1144 1146 netdev_err(dev, "%s: tx ring %d full when queue %d awake\n", 1145 - __func__, index, ring->queue); 1147 + __func__, index, ring->queue); 1146 1148 ret = NETDEV_TX_BUSY; 1147 1149 goto out; 1148 1150 } ··· 1170 1172 /* xmit fragment */ 1171 1173 for (i = 0; i < nr_frags; i++) { 1172 1174 ret = bcmgenet_xmit_frag(dev, 1173 - &skb_shinfo(skb)->frags[i], 1174 - (i == nr_frags - 1) ? DMA_EOP : 0, ring); 1175 + &skb_shinfo(skb)->frags[i], 1176 + (i == nr_frags - 1) ? DMA_EOP : 0, 1177 + ring); 1175 1178 if (ret) { 1176 1179 ret = NETDEV_TX_OK; 1177 1180 goto out; ··· 1185 1186 * producer index, now write it down to the hardware 1186 1187 */ 1187 1188 bcmgenet_tdma_ring_writel(priv, ring->index, 1188 - ring->prod_index, TDMA_PROD_INDEX); 1189 + ring->prod_index, TDMA_PROD_INDEX); 1189 1190 1190 1191 if (ring->free_bds <= (MAX_SKB_FRAGS + 1)) { 1191 1192 netif_tx_stop_queue(txq); ··· 1199 1200 } 1200 1201 1201 1202 1202 - static int bcmgenet_rx_refill(struct bcmgenet_priv *priv, 1203 - struct enet_cb *cb) 1203 + static int bcmgenet_rx_refill(struct bcmgenet_priv *priv, struct enet_cb *cb) 1204 1204 { 1205 1205 struct device *kdev = &priv->pdev->dev; 1206 1206 struct sk_buff *skb; 1207 1207 dma_addr_t mapping; 1208 1208 int ret; 1209 1209 1210 - skb = netdev_alloc_skb(priv->dev, 1211 - priv->rx_buf_len + SKB_ALIGNMENT); 1210 + skb = netdev_alloc_skb(priv->dev, priv->rx_buf_len + SKB_ALIGNMENT); 1212 1211 if (!skb) 1213 1212 return -ENOMEM; 1214 1213 ··· 1214 1217 WARN_ON(cb->skb != NULL); 1215 1218 cb->skb = skb; 1216 1219 mapping = dma_map_single(kdev, skb->data, 1217 - priv->rx_buf_len, DMA_FROM_DEVICE); 1220 + priv->rx_buf_len, DMA_FROM_DEVICE); 1218 1221 ret = dma_mapping_error(kdev, mapping); 1219 1222 if (ret) { 1220 1223 bcmgenet_free_cb(cb); 1221 1224 netif_err(priv, rx_err, priv->dev, 1222 - "%s DMA map failed\n", __func__); 1225 + "%s DMA map failed\n", __func__); 1223 1226 return ret; 1224 1227 } 1225 1228 ··· 1254 1257 unsigned int p_index; 1255 1258 unsigned int chksum_ok = 0; 1256 1259 1257 - p_index = bcmgenet_rdma_ring_readl(priv, 1258 - DESC_INDEX, RDMA_PROD_INDEX); 1260 + p_index = bcmgenet_rdma_ring_readl(priv, DESC_INDEX, RDMA_PROD_INDEX); 1259 1261 p_index &= DMA_P_INDEX_MASK; 1260 1262 1261 1263 if (p_index < priv->rx_c_index) ··· 1264 1268 rxpkttoprocess = p_index - priv->rx_c_index; 1265 1269 1266 1270 netif_dbg(priv, rx_status, dev, 1267 - "RDMA: rxpkttoprocess=%d\n", rxpkttoprocess); 1271 + "RDMA: rxpkttoprocess=%d\n", rxpkttoprocess); 1268 1272 1269 1273 while ((rxpktprocessed < rxpkttoprocess) && 1270 - (rxpktprocessed < budget)) { 1274 + (rxpktprocessed < budget)) { 1271 1275 1272 1276 /* Unmap the packet contents such that we can use the 1273 1277 * RSV from the 64 bytes descriptor when enabled and save ··· 1276 1280 cb = &priv->rx_cbs[priv->rx_read_ptr]; 1277 1281 skb = cb->skb; 1278 1282 dma_unmap_single(&dev->dev, dma_unmap_addr(cb, dma_addr), 1279 - priv->rx_buf_len, DMA_FROM_DEVICE); 1283 + priv->rx_buf_len, DMA_FROM_DEVICE); 1280 1284 1281 1285 if (!priv->desc_64b_en) { 1282 - dma_length_status = dmadesc_get_length_status(priv, 1283 - priv->rx_bds + 1284 - (priv->rx_read_ptr * 1285 - DMA_DESC_SIZE)); 1286 + dma_length_status = 1287 + dmadesc_get_length_status(priv, 1288 + priv->rx_bds + 1289 + (priv->rx_read_ptr * 1290 + DMA_DESC_SIZE)); 1286 1291 } else { 1287 1292 struct status_64 *status; 1288 1293 status = (struct status_64 *)skb->data; ··· 1297 1300 len = dma_length_status >> DMA_BUFLENGTH_SHIFT; 1298 1301 1299 1302 netif_dbg(priv, rx_status, dev, 1300 - "%s: p_ind=%d c_ind=%d read_ptr=%d len_stat=0x%08x\n", 1301 - __func__, p_index, priv->rx_c_index, priv->rx_read_ptr, 1302 - dma_length_status); 1303 + "%s:p_ind=%d c_ind=%d read_ptr=%d len_stat=0x%08x\n", 1304 + __func__, p_index, priv->rx_c_index, 1305 + priv->rx_read_ptr, dma_length_status); 1303 1306 1304 1307 rxpktprocessed++; 1305 1308 ··· 1315 1318 1316 1319 if (unlikely(!(dma_flag & DMA_EOP) || !(dma_flag & DMA_SOP))) { 1317 1320 netif_err(priv, rx_status, dev, 1318 - "Droping fragmented packet!\n"); 1321 + "dropping fragmented packet!\n"); 1319 1322 dev->stats.rx_dropped++; 1320 1323 dev->stats.rx_errors++; 1321 1324 dev_kfree_skb_any(cb->skb); ··· 1329 1332 DMA_RX_LG | 1330 1333 DMA_RX_RXER))) { 1331 1334 netif_err(priv, rx_status, dev, "dma_flag=0x%x\n", 1332 - (unsigned int)dma_flag); 1335 + (unsigned int)dma_flag); 1333 1336 if (dma_flag & DMA_RX_CRC_ERROR) 1334 1337 dev->stats.rx_crc_errors++; 1335 1338 if (dma_flag & DMA_RX_OV) ··· 1348 1351 } /* error packet */ 1349 1352 1350 1353 chksum_ok = (dma_flag & priv->dma_rx_chk_bit) && 1351 - priv->desc_rxchk_en; 1354 + priv->desc_rxchk_en; 1352 1355 1353 1356 skb_put(skb, len); 1354 1357 if (priv->desc_64b_en) { ··· 1424 1427 1425 1428 if (dma_unmap_addr(cb, dma_addr)) { 1426 1429 dma_unmap_single(&priv->dev->dev, 1427 - dma_unmap_addr(cb, dma_addr), 1428 - priv->rx_buf_len, DMA_FROM_DEVICE); 1430 + dma_unmap_addr(cb, dma_addr), 1431 + priv->rx_buf_len, DMA_FROM_DEVICE); 1429 1432 dma_unmap_addr_set(cb, dma_addr, 0); 1430 1433 } 1431 1434 ··· 1434 1437 } 1435 1438 } 1436 1439 1437 - static void umac_enable_set(struct bcmgenet_priv *priv, u32 mask, 1438 - bool enable) 1440 + static void umac_enable_set(struct bcmgenet_priv *priv, u32 mask, bool enable) 1439 1441 { 1440 1442 u32 reg; 1441 1443 ··· 1510 1514 bcmgenet_umac_writel(priv, 0, UMAC_CMD); 1511 1515 /* clear tx/rx counter */ 1512 1516 bcmgenet_umac_writel(priv, 1513 - MIB_RESET_RX | MIB_RESET_TX | MIB_RESET_RUNT, UMAC_MIB_CTRL); 1517 + MIB_RESET_RX | MIB_RESET_TX | MIB_RESET_RUNT, 1518 + UMAC_MIB_CTRL); 1514 1519 bcmgenet_umac_writel(priv, 0, UMAC_MIB_CTRL); 1515 1520 1516 1521 bcmgenet_umac_writel(priv, ENET_MAX_MTU_SIZE, UMAC_MAX_FRAME_LEN); ··· 1551 1554 if (priv->hw_params->flags & GENET_HAS_MDIO_INTR) 1552 1555 cpu_mask_clear |= UMAC_IRQ_MDIO_DONE | UMAC_IRQ_MDIO_ERROR; 1553 1556 1554 - bcmgenet_intrl2_0_writel(priv, cpu_mask_clear, 1555 - INTRL2_CPU_MASK_CLEAR); 1557 + bcmgenet_intrl2_0_writel(priv, cpu_mask_clear, INTRL2_CPU_MASK_CLEAR); 1556 1558 1557 1559 /* Enable rx/tx engine.*/ 1558 1560 dev_dbg(kdev, "done init umac\n"); ··· 1600 1604 bcmgenet_tdma_ring_writel(priv, index, 1, DMA_MBUF_DONE_THRESH); 1601 1605 /* Disable rate control for now */ 1602 1606 bcmgenet_tdma_ring_writel(priv, index, flow_period_val, 1603 - TDMA_FLOW_PERIOD); 1607 + TDMA_FLOW_PERIOD); 1604 1608 /* Unclassified traffic goes to ring 16 */ 1605 1609 bcmgenet_tdma_ring_writel(priv, index, 1606 - ((size << DMA_RING_SIZE_SHIFT) | RX_BUF_LENGTH), 1607 - DMA_RING_BUF_SIZE); 1610 + ((size << DMA_RING_SIZE_SHIFT) | 1611 + RX_BUF_LENGTH), DMA_RING_BUF_SIZE); 1608 1612 1609 1613 first_bd = write_ptr; 1610 1614 1611 1615 /* Set start and end address, read and write pointers */ 1612 1616 bcmgenet_tdma_ring_writel(priv, index, first_bd * words_per_bd, 1613 - DMA_START_ADDR); 1617 + DMA_START_ADDR); 1614 1618 bcmgenet_tdma_ring_writel(priv, index, first_bd * words_per_bd, 1615 - TDMA_READ_PTR); 1619 + TDMA_READ_PTR); 1616 1620 bcmgenet_tdma_ring_writel(priv, index, first_bd, 1617 - TDMA_WRITE_PTR); 1621 + TDMA_WRITE_PTR); 1618 1622 bcmgenet_tdma_ring_writel(priv, index, end_ptr * words_per_bd - 1, 1619 - DMA_END_ADDR); 1623 + DMA_END_ADDR); 1620 1624 } 1621 1625 1622 1626 /* Initialize a RDMA ring */ 1623 1627 static int bcmgenet_init_rx_ring(struct bcmgenet_priv *priv, 1624 - unsigned int index, unsigned int size) 1628 + unsigned int index, unsigned int size) 1625 1629 { 1626 1630 u32 words_per_bd = WORDS_PER_BD(priv); 1627 1631 int ret; ··· 1647 1651 bcmgenet_rdma_ring_writel(priv, index, 0, RDMA_PROD_INDEX); 1648 1652 bcmgenet_rdma_ring_writel(priv, index, 0, RDMA_CONS_INDEX); 1649 1653 bcmgenet_rdma_ring_writel(priv, index, 1650 - ((size << DMA_RING_SIZE_SHIFT) | RX_BUF_LENGTH), 1651 - DMA_RING_BUF_SIZE); 1654 + ((size << DMA_RING_SIZE_SHIFT) | 1655 + RX_BUF_LENGTH), DMA_RING_BUF_SIZE); 1652 1656 bcmgenet_rdma_ring_writel(priv, index, 0, DMA_START_ADDR); 1653 1657 bcmgenet_rdma_ring_writel(priv, index, 1654 - words_per_bd * size - 1, DMA_END_ADDR); 1658 + words_per_bd * size - 1, DMA_END_ADDR); 1655 1659 bcmgenet_rdma_ring_writel(priv, index, 1656 - (DMA_FC_THRESH_LO << DMA_XOFF_THRESHOLD_SHIFT) | 1657 - DMA_FC_THRESH_HI, RDMA_XON_XOFF_THRESH); 1660 + (DMA_FC_THRESH_LO << 1661 + DMA_XOFF_THRESHOLD_SHIFT) | 1662 + DMA_FC_THRESH_HI, RDMA_XON_XOFF_THRESH); 1658 1663 bcmgenet_rdma_ring_writel(priv, index, 0, RDMA_READ_PTR); 1659 1664 1660 1665 return ret; ··· 1702 1705 * (ring 16) 1703 1706 */ 1704 1707 bcmgenet_init_tx_ring(priv, i, priv->hw_params->bds_cnt, 1705 - i * priv->hw_params->bds_cnt, 1706 - (i + 1) * priv->hw_params->bds_cnt); 1708 + i * priv->hw_params->bds_cnt, 1709 + (i + 1) * priv->hw_params->bds_cnt); 1707 1710 1708 1711 /* Configure ring as decriptor ring and setup priority */ 1709 1712 ring_cfg |= 1 << i; ··· 1775 1778 priv->tx_bds = priv->base + priv->hw_params->tdma_offset; 1776 1779 priv->num_tx_bds = TOTAL_DESC; 1777 1780 priv->tx_cbs = kzalloc(priv->num_tx_bds * sizeof(struct enet_cb), 1778 - GFP_KERNEL); 1781 + GFP_KERNEL); 1779 1782 if (!priv->tx_cbs) { 1780 1783 bcmgenet_fini_dma(priv); 1781 1784 return -ENOMEM; ··· 1786 1789 1787 1790 /* initialize special ring 16 */ 1788 1791 bcmgenet_init_tx_ring(priv, DESC_INDEX, GENET_DEFAULT_BD_CNT, 1789 - priv->hw_params->tx_queues * priv->hw_params->bds_cnt, 1790 - TOTAL_DESC); 1792 + priv->hw_params->tx_queues * 1793 + priv->hw_params->bds_cnt, 1794 + TOTAL_DESC); 1791 1795 1792 1796 return 0; 1793 1797 } ··· 1809 1811 priv->rx_c_index += work_done; 1810 1812 priv->rx_c_index &= DMA_C_INDEX_MASK; 1811 1813 bcmgenet_rdma_ring_writel(priv, DESC_INDEX, 1812 - priv->rx_c_index, RDMA_CONS_INDEX); 1814 + priv->rx_c_index, RDMA_CONS_INDEX); 1813 1815 if (work_done < budget) { 1814 1816 napi_complete(napi); 1815 - bcmgenet_intrl2_0_writel(priv, 1816 - UMAC_IRQ_RXDMA_BDONE, INTRL2_CPU_MASK_CLEAR); 1817 + bcmgenet_intrl2_0_writel(priv, UMAC_IRQ_RXDMA_BDONE, 1818 + INTRL2_CPU_MASK_CLEAR); 1817 1819 } 1818 1820 1819 1821 return work_done; ··· 1836 1838 1837 1839 /* Link UP/DOWN event */ 1838 1840 if ((priv->hw_params->flags & GENET_HAS_MDIO_INTR) && 1839 - (priv->irq0_stat & (UMAC_IRQ_LINK_UP|UMAC_IRQ_LINK_DOWN))) { 1841 + (priv->irq0_stat & (UMAC_IRQ_LINK_UP|UMAC_IRQ_LINK_DOWN))) { 1840 1842 phy_mac_interrupt(priv->phydev, 1841 - priv->irq0_stat & UMAC_IRQ_LINK_UP); 1843 + priv->irq0_stat & UMAC_IRQ_LINK_UP); 1842 1844 priv->irq0_stat &= ~(UMAC_IRQ_LINK_UP|UMAC_IRQ_LINK_DOWN); 1843 1845 } 1844 1846 } ··· 1857 1859 bcmgenet_intrl2_1_writel(priv, priv->irq1_stat, INTRL2_CPU_CLEAR); 1858 1860 1859 1861 netif_dbg(priv, intr, priv->dev, 1860 - "%s: IRQ=0x%x\n", __func__, priv->irq1_stat); 1862 + "%s: IRQ=0x%x\n", __func__, priv->irq1_stat); 1861 1863 /* Check the MBDONE interrupts. 1862 1864 * packet is done, reclaim descriptors 1863 1865 */ ··· 1866 1868 for (index = 0; index < 16; index++) { 1867 1869 if (priv->irq1_stat & (1 << index)) 1868 1870 bcmgenet_tx_reclaim(priv->dev, 1869 - &priv->tx_rings[index]); 1871 + &priv->tx_rings[index]); 1870 1872 } 1871 1873 } 1872 1874 return IRQ_HANDLED; ··· 1885 1887 bcmgenet_intrl2_0_writel(priv, priv->irq0_stat, INTRL2_CPU_CLEAR); 1886 1888 1887 1889 netif_dbg(priv, intr, priv->dev, 1888 - "IRQ=0x%x\n", priv->irq0_stat); 1890 + "IRQ=0x%x\n", priv->irq0_stat); 1889 1891 1890 1892 if (priv->irq0_stat & (UMAC_IRQ_RXDMA_BDONE | UMAC_IRQ_RXDMA_PDONE)) { 1891 1893 /* We use NAPI(software interrupt throttling, if ··· 1893 1895 * Disable interrupt, will be enabled in the poll method. 1894 1896 */ 1895 1897 if (likely(napi_schedule_prep(&priv->napi))) { 1896 - bcmgenet_intrl2_0_writel(priv, 1897 - UMAC_IRQ_RXDMA_BDONE, INTRL2_CPU_MASK_SET); 1898 + bcmgenet_intrl2_0_writel(priv, UMAC_IRQ_RXDMA_BDONE, 1899 + INTRL2_CPU_MASK_SET); 1898 1900 __napi_schedule(&priv->napi); 1899 1901 } 1900 1902 } ··· 1915 1917 } 1916 1918 1917 1919 if ((priv->hw_params->flags & GENET_HAS_MDIO_INTR) && 1918 - priv->irq0_stat & (UMAC_IRQ_MDIO_DONE | UMAC_IRQ_MDIO_ERROR)) { 1920 + priv->irq0_stat & (UMAC_IRQ_MDIO_DONE | UMAC_IRQ_MDIO_ERROR)) { 1919 1921 priv->irq0_stat &= ~(UMAC_IRQ_MDIO_DONE | UMAC_IRQ_MDIO_ERROR); 1920 1922 wake_up(&priv->wq); 1921 1923 } ··· 1947 1949 } 1948 1950 1949 1951 static void bcmgenet_set_hw_addr(struct bcmgenet_priv *priv, 1950 - unsigned char *addr) 1952 + unsigned char *addr) 1951 1953 { 1952 1954 bcmgenet_umac_writel(priv, (addr[0] << 24) | (addr[1] << 16) | 1953 1955 (addr[2] << 8) | addr[3], UMAC_MAC0); ··· 2068 2070 bcmgenet_enable_dma(priv, dma_ctrl); 2069 2071 2070 2072 ret = request_irq(priv->irq0, bcmgenet_isr0, IRQF_SHARED, 2071 - dev->name, priv); 2073 + dev->name, priv); 2072 2074 if (ret < 0) { 2073 2075 netdev_err(dev, "can't request IRQ %d\n", priv->irq0); 2074 2076 goto err_fini_dma; 2075 2077 } 2076 2078 2077 2079 ret = request_irq(priv->irq1, bcmgenet_isr1, IRQF_SHARED, 2078 - dev->name, priv); 2080 + dev->name, priv); 2079 2081 if (ret < 0) { 2080 2082 netdev_err(dev, "can't request IRQ %d\n", priv->irq1); 2081 2083 goto err_irq0; ··· 2116 2118 } 2117 2119 2118 2120 if (timeout == DMA_TIMEOUT_VAL) { 2119 - netdev_warn(priv->dev, 2120 - "Timed out while disabling TX DMA\n"); 2121 + netdev_warn(priv->dev, "Timed out while disabling TX DMA\n"); 2121 2122 ret = -ETIMEDOUT; 2122 2123 } 2123 2124 ··· 2139 2142 } 2140 2143 2141 2144 if (timeout == DMA_TIMEOUT_VAL) { 2142 - netdev_warn(priv->dev, 2143 - "Timed out while disabling RX DMA\n"); 2144 - ret = -ETIMEDOUT; 2145 + netdev_warn(priv->dev, "Timed out while disabling RX DMA\n"); 2146 + ret = -ETIMEDOUT; 2145 2147 } 2146 2148 2147 2149 return ret; ··· 2219 2223 { 2220 2224 u32 reg; 2221 2225 2222 - bcmgenet_umac_writel(priv, 2223 - addr[0] << 8 | addr[1], UMAC_MDF_ADDR + (*i * 4)); 2224 - bcmgenet_umac_writel(priv, 2225 - addr[2] << 24 | addr[3] << 16 | 2226 - addr[4] << 8 | addr[5], 2227 - UMAC_MDF_ADDR + ((*i + 1) * 4)); 2226 + bcmgenet_umac_writel(priv, addr[0] << 8 | addr[1], 2227 + UMAC_MDF_ADDR + (*i * 4)); 2228 + bcmgenet_umac_writel(priv, addr[2] << 24 | addr[3] << 16 | 2229 + addr[4] << 8 | addr[5], 2230 + UMAC_MDF_ADDR + ((*i + 1) * 4)); 2228 2231 reg = bcmgenet_umac_readl(priv, UMAC_MDF_CTRL); 2229 2232 reg |= (1 << (MAX_MC_COUNT - *mc)); 2230 2233 bcmgenet_umac_writel(priv, reg, UMAC_MDF_CTRL); ··· 2420 2425 2421 2426 /* Print the GENET core version */ 2422 2427 dev_info(&priv->pdev->dev, "GENET " GENET_VER_FMT, 2423 - major, (reg >> 16) & 0x0f, reg & 0xffff); 2428 + major, (reg >> 16) & 0x0f, reg & 0xffff); 2424 2429 2425 2430 #ifdef CONFIG_PHYS_ADDR_T_64BIT 2426 2431 if (!(params->flags & GENET_HAS_40BITS))
+2 -2
drivers/net/ethernet/broadcom/genet/bcmgenet.h
··· 504 504 unsigned int cb_ptr; /* Tx ring initial CB ptr */ 505 505 unsigned int end_ptr; /* Tx ring end CB ptr */ 506 506 void (*int_enable)(struct bcmgenet_priv *priv, 507 - struct bcmgenet_tx_ring *); 507 + struct bcmgenet_tx_ring *); 508 508 void (*int_disable)(struct bcmgenet_priv *priv, 509 - struct bcmgenet_tx_ring *); 509 + struct bcmgenet_tx_ring *); 510 510 }; 511 511 512 512 /* device context */
+14 -14
drivers/net/ethernet/broadcom/genet/bcmmii.c
··· 35 35 u32 reg; 36 36 37 37 bcmgenet_umac_writel(priv, (MDIO_RD | (phy_id << MDIO_PMD_SHIFT) | 38 - (location << MDIO_REG_SHIFT)), UMAC_MDIO_CMD); 38 + (location << MDIO_REG_SHIFT)), UMAC_MDIO_CMD); 39 39 /* Start MDIO transaction*/ 40 40 reg = bcmgenet_umac_readl(priv, UMAC_MDIO_CMD); 41 41 reg |= MDIO_START_BUSY; 42 42 bcmgenet_umac_writel(priv, reg, UMAC_MDIO_CMD); 43 43 wait_event_timeout(priv->wq, 44 - !(bcmgenet_umac_readl(priv, UMAC_MDIO_CMD) 45 - & MDIO_START_BUSY), 46 - HZ / 100); 44 + !(bcmgenet_umac_readl(priv, UMAC_MDIO_CMD) 45 + & MDIO_START_BUSY), 46 + HZ / 100); 47 47 ret = bcmgenet_umac_readl(priv, UMAC_MDIO_CMD); 48 48 49 49 if (ret & MDIO_READ_FAIL) ··· 54 54 55 55 /* write a value to the MII */ 56 56 static int bcmgenet_mii_write(struct mii_bus *bus, int phy_id, 57 - int location, u16 val) 57 + int location, u16 val) 58 58 { 59 59 struct net_device *dev = bus->priv; 60 60 struct bcmgenet_priv *priv = netdev_priv(dev); 61 61 u32 reg; 62 62 63 63 bcmgenet_umac_writel(priv, (MDIO_WR | (phy_id << MDIO_PMD_SHIFT) | 64 - (location << MDIO_REG_SHIFT) | (0xffff & val)), 65 - UMAC_MDIO_CMD); 64 + (location << MDIO_REG_SHIFT) | (0xffff & val)), 65 + UMAC_MDIO_CMD); 66 66 reg = bcmgenet_umac_readl(priv, UMAC_MDIO_CMD); 67 67 reg |= MDIO_START_BUSY; 68 68 bcmgenet_umac_writel(priv, reg, UMAC_MDIO_CMD); 69 69 wait_event_timeout(priv->wq, 70 - !(bcmgenet_umac_readl(priv, UMAC_MDIO_CMD) & 71 - MDIO_START_BUSY), 72 - HZ / 100); 70 + !(bcmgenet_umac_readl(priv, UMAC_MDIO_CMD) & 71 + MDIO_START_BUSY), 72 + HZ / 100); 73 73 74 74 return 0; 75 75 } ··· 239 239 phy_name = "external MII"; 240 240 phydev->supported &= PHY_BASIC_FEATURES; 241 241 bcmgenet_sys_writel(priv, 242 - PORT_MODE_EXT_EPHY, SYS_PORT_CTRL); 242 + PORT_MODE_EXT_EPHY, SYS_PORT_CTRL); 243 243 break; 244 244 245 245 case PHY_INTERFACE_MODE_REVMII: ··· 275 275 reg |= RGMII_MODE_EN | id_mode_dis; 276 276 bcmgenet_ext_writel(priv, reg, EXT_RGMII_OOB_CTRL); 277 277 bcmgenet_sys_writel(priv, 278 - PORT_MODE_EXT_GPHY, SYS_PORT_CTRL); 278 + PORT_MODE_EXT_GPHY, SYS_PORT_CTRL); 279 279 break; 280 280 default: 281 281 dev_err(kdev, "unknown phy mode: %d\n", priv->phy_interface); ··· 354 354 priv->mii_bus->irq[phydev->addr] = PHY_POLL; 355 355 356 356 pr_info("attached PHY at address %d [%s]\n", 357 - phydev->addr, phydev->drv->name); 357 + phydev->addr, phydev->drv->name); 358 358 359 359 return 0; 360 360 } ··· 379 379 bus->read = bcmgenet_mii_read; 380 380 bus->write = bcmgenet_mii_write; 381 381 snprintf(bus->id, MII_BUS_ID_SIZE, "%s-%d", 382 - priv->pdev->name, priv->pdev->id); 382 + priv->pdev->name, priv->pdev->id); 383 383 384 384 bus->irq = kzalloc(sizeof(int) * PHY_MAX_ADDR, GFP_KERNEL); 385 385 if (!bus->irq) {