Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'dpaa2-rx-copybreak'

Ioana Ciornei says:

====================
dpaa2-eth: add rx copybreak support

DMA unmapping, allocating a new buffer and DMA mapping it back on the
refill path is really not that efficient. Proper buffer recycling (page
pool, flipping the page and using the other half) cannot be done for
DPAA2 since it's not a ring based controller but it rather deals with
multiple queues which all get their buffers from the same buffer pool on
Rx.

To circumvent these limitations, add support for Rx copybreak in
dpaa2-eth.

Below you can find a summary of the tests that were run to end up
with the default rx copybreak value of 512.
A bit about the setup - a LS2088A SoC, 8 x Cortex A72 @ 1.8GHz, IPfwd
zero loss test @ 20Gbit/s throughput. I tested multiple frame sizes to
get an idea where is the break even point.

Here are 2 sets of results, (1) is the baseline and (2) is just
allocating a new skb for all frames sizes received (as if the copybreak
was even to the MTU). All numbers are in Mpps.

64 128 256 512 640 768 896

(1) 3.23 3.23 3.24 3.21 3.1 2.76 2.71
(2) 3.95 3.88 3.79 3.62 3.3 3.02 2.65

It seems that even for 512 bytes frame sizes it's comfortably better when
allocating a new skb. After that, we see diminishing rewards or even worse.

Changes in v2:
- properly marked dpaa2_eth_copybreak as static
====================

Signed-off-by: David S. Miller <davem@davemloft.net>

+98 -18
+50 -16
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
··· 223 223 } 224 224 } 225 225 226 - static void dpaa2_eth_xdp_release_buf(struct dpaa2_eth_priv *priv, 227 - struct dpaa2_eth_channel *ch, 228 - dma_addr_t addr) 226 + static void dpaa2_eth_recycle_buf(struct dpaa2_eth_priv *priv, 227 + struct dpaa2_eth_channel *ch, 228 + dma_addr_t addr) 229 229 { 230 230 int retries = 0; 231 231 int err; 232 232 233 - ch->xdp.drop_bufs[ch->xdp.drop_cnt++] = addr; 234 - if (ch->xdp.drop_cnt < DPAA2_ETH_BUFS_PER_CMD) 233 + ch->recycled_bufs[ch->recycled_bufs_cnt++] = addr; 234 + if (ch->recycled_bufs_cnt < DPAA2_ETH_BUFS_PER_CMD) 235 235 return; 236 236 237 237 while ((err = dpaa2_io_service_release(ch->dpio, priv->bpid, 238 - ch->xdp.drop_bufs, 239 - ch->xdp.drop_cnt)) == -EBUSY) { 238 + ch->recycled_bufs, 239 + ch->recycled_bufs_cnt)) == -EBUSY) { 240 240 if (retries++ >= DPAA2_ETH_SWP_BUSY_RETRIES) 241 241 break; 242 242 cpu_relax(); 243 243 } 244 244 245 245 if (err) { 246 - dpaa2_eth_free_bufs(priv, ch->xdp.drop_bufs, ch->xdp.drop_cnt); 247 - ch->buf_count -= ch->xdp.drop_cnt; 246 + dpaa2_eth_free_bufs(priv, ch->recycled_bufs, ch->recycled_bufs_cnt); 247 + ch->buf_count -= ch->recycled_bufs_cnt; 248 248 } 249 249 250 - ch->xdp.drop_cnt = 0; 250 + ch->recycled_bufs_cnt = 0; 251 251 } 252 252 253 253 static int dpaa2_eth_xdp_flush(struct dpaa2_eth_priv *priv, ··· 300 300 ch->stats.xdp_tx++; 301 301 } 302 302 for (i = enqueued; i < fq->xdp_tx_fds.num; i++) { 303 - dpaa2_eth_xdp_release_buf(priv, ch, dpaa2_fd_get_addr(&fds[i])); 303 + dpaa2_eth_recycle_buf(priv, ch, dpaa2_fd_get_addr(&fds[i])); 304 304 percpu_stats->tx_errors++; 305 305 ch->stats.xdp_tx_err++; 306 306 } ··· 382 382 trace_xdp_exception(priv->net_dev, xdp_prog, xdp_act); 383 383 fallthrough; 384 384 case XDP_DROP: 385 - dpaa2_eth_xdp_release_buf(priv, ch, addr); 385 + dpaa2_eth_recycle_buf(priv, ch, addr); 386 386 ch->stats.xdp_drop++; 387 387 break; 388 388 case XDP_REDIRECT: ··· 403 403 free_pages((unsigned long)vaddr, 0); 404 404 } else { 405 405 ch->buf_count++; 406 - dpaa2_eth_xdp_release_buf(priv, ch, addr); 406 + dpaa2_eth_recycle_buf(priv, ch, addr); 407 407 } 408 408 ch->stats.xdp_drop++; 409 409 } else { ··· 416 416 out: 417 417 rcu_read_unlock(); 418 418 return xdp_act; 419 + } 420 + 421 + static struct sk_buff *dpaa2_eth_copybreak(struct dpaa2_eth_channel *ch, 422 + const struct dpaa2_fd *fd, 423 + void *fd_vaddr) 424 + { 425 + u16 fd_offset = dpaa2_fd_get_offset(fd); 426 + struct dpaa2_eth_priv *priv = ch->priv; 427 + u32 fd_length = dpaa2_fd_get_len(fd); 428 + struct sk_buff *skb = NULL; 429 + unsigned int skb_len; 430 + 431 + if (fd_length > priv->rx_copybreak) 432 + return NULL; 433 + 434 + skb_len = fd_length + dpaa2_eth_needed_headroom(NULL); 435 + 436 + skb = napi_alloc_skb(&ch->napi, skb_len); 437 + if (!skb) 438 + return NULL; 439 + 440 + skb_reserve(skb, dpaa2_eth_needed_headroom(NULL)); 441 + skb_put(skb, fd_length); 442 + 443 + memcpy(skb->data, fd_vaddr + fd_offset, fd_length); 444 + 445 + dpaa2_eth_recycle_buf(priv, ch, dpaa2_fd_get_addr(fd)); 446 + 447 + return skb; 419 448 } 420 449 421 450 /* Main Rx frame processing routine */ ··· 488 459 return; 489 460 } 490 461 491 - dma_unmap_page(dev, addr, priv->rx_buf_size, 492 - DMA_BIDIRECTIONAL); 493 - skb = dpaa2_eth_build_linear_skb(ch, fd, vaddr); 462 + skb = dpaa2_eth_copybreak(ch, fd, vaddr); 463 + if (!skb) { 464 + dma_unmap_page(dev, addr, priv->rx_buf_size, 465 + DMA_BIDIRECTIONAL); 466 + skb = dpaa2_eth_build_linear_skb(ch, fd, vaddr); 467 + } 494 468 } else if (fd_format == dpaa2_fd_sg) { 495 469 WARN_ON(priv->xdp_prog); 496 470 ··· 4333 4301 INIT_WORK(&priv->tx_onestep_tstamp, dpaa2_eth_tx_onestep_tstamp); 4334 4302 4335 4303 skb_queue_head_init(&priv->tx_skbs); 4304 + 4305 + priv->rx_copybreak = DPAA2_ETH_DEFAULT_COPYBREAK; 4336 4306 4337 4307 /* Obtain a MC portal */ 4338 4308 err = fsl_mc_portal_allocate(dpni_dev, FSL_MC_IO_ATOMIC_CONTEXT_PORTAL,
+8 -2
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h
··· 438 438 439 439 struct dpaa2_eth_ch_xdp { 440 440 struct bpf_prog *prog; 441 - u64 drop_bufs[DPAA2_ETH_BUFS_PER_CMD]; 442 - int drop_cnt; 443 441 unsigned int res; 444 442 }; 445 443 ··· 455 457 struct dpaa2_eth_ch_xdp xdp; 456 458 struct xdp_rxq_info xdp_rxq; 457 459 struct list_head *rx_list; 460 + 461 + /* Buffers to be recycled back in the buffer pool */ 462 + u64 recycled_bufs[DPAA2_ETH_BUFS_PER_CMD]; 463 + int recycled_bufs_cnt; 458 464 }; 459 465 460 466 struct dpaa2_eth_dist_fields { ··· 488 486 struct dpaa2_eth_trap_item *trap_items_arr; 489 487 struct dpaa2_eth_priv *priv; 490 488 }; 489 + 490 + #define DPAA2_ETH_DEFAULT_COPYBREAK 512 491 491 492 492 /* Driver private data */ 493 493 struct dpaa2_eth_priv { ··· 571 567 struct devlink *devlink; 572 568 struct dpaa2_eth_trap_data *trap_data; 573 569 struct devlink_port devlink_port; 570 + 571 + u32 rx_copybreak; 574 572 }; 575 573 576 574 struct dpaa2_eth_devlink_priv {
+40
drivers/net/ethernet/freescale/dpaa2/dpaa2-ethtool.c
··· 782 782 return 0; 783 783 } 784 784 785 + static int dpaa2_eth_get_tunable(struct net_device *net_dev, 786 + const struct ethtool_tunable *tuna, 787 + void *data) 788 + { 789 + struct dpaa2_eth_priv *priv = netdev_priv(net_dev); 790 + int err = 0; 791 + 792 + switch (tuna->id) { 793 + case ETHTOOL_RX_COPYBREAK: 794 + *(u32 *)data = priv->rx_copybreak; 795 + break; 796 + default: 797 + err = -EOPNOTSUPP; 798 + break; 799 + } 800 + 801 + return err; 802 + } 803 + 804 + static int dpaa2_eth_set_tunable(struct net_device *net_dev, 805 + const struct ethtool_tunable *tuna, 806 + const void *data) 807 + { 808 + struct dpaa2_eth_priv *priv = netdev_priv(net_dev); 809 + int err = 0; 810 + 811 + switch (tuna->id) { 812 + case ETHTOOL_RX_COPYBREAK: 813 + priv->rx_copybreak = *(u32 *)data; 814 + break; 815 + default: 816 + err = -EOPNOTSUPP; 817 + break; 818 + } 819 + 820 + return err; 821 + } 822 + 785 823 const struct ethtool_ops dpaa2_ethtool_ops = { 786 824 .get_drvinfo = dpaa2_eth_get_drvinfo, 787 825 .nway_reset = dpaa2_eth_nway_reset, ··· 834 796 .get_rxnfc = dpaa2_eth_get_rxnfc, 835 797 .set_rxnfc = dpaa2_eth_set_rxnfc, 836 798 .get_ts_info = dpaa2_eth_get_ts_info, 799 + .get_tunable = dpaa2_eth_get_tunable, 800 + .set_tunable = dpaa2_eth_set_tunable, 837 801 };