Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
at v2.6.39-rc7 1097 lines 32 kB view raw
1/**************************************************************************** 2 * Driver for Solarflare Solarstorm network controllers and boards 3 * Copyright 2005-2006 Fen Systems Ltd. 4 * Copyright 2006-2010 Solarflare Communications Inc. 5 * 6 * This program is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License version 2 as published 8 * by the Free Software Foundation, incorporated herein by reference. 9 */ 10 11#include <linux/netdevice.h> 12#include <linux/ethtool.h> 13#include <linux/rtnetlink.h> 14#include <linux/in.h> 15#include "net_driver.h" 16#include "workarounds.h" 17#include "selftest.h" 18#include "efx.h" 19#include "filter.h" 20#include "nic.h" 21 22struct ethtool_string { 23 char name[ETH_GSTRING_LEN]; 24}; 25 26struct efx_ethtool_stat { 27 const char *name; 28 enum { 29 EFX_ETHTOOL_STAT_SOURCE_mac_stats, 30 EFX_ETHTOOL_STAT_SOURCE_nic, 31 EFX_ETHTOOL_STAT_SOURCE_channel, 32 EFX_ETHTOOL_STAT_SOURCE_tx_queue 33 } source; 34 unsigned offset; 35 u64(*get_stat) (void *field); /* Reader function */ 36}; 37 38/* Initialiser for a struct #efx_ethtool_stat with type-checking */ 39#define EFX_ETHTOOL_STAT(stat_name, source_name, field, field_type, \ 40 get_stat_function) { \ 41 .name = #stat_name, \ 42 .source = EFX_ETHTOOL_STAT_SOURCE_##source_name, \ 43 .offset = ((((field_type *) 0) == \ 44 &((struct efx_##source_name *)0)->field) ? \ 45 offsetof(struct efx_##source_name, field) : \ 46 offsetof(struct efx_##source_name, field)), \ 47 .get_stat = get_stat_function, \ 48} 49 50static u64 efx_get_uint_stat(void *field) 51{ 52 return *(unsigned int *)field; 53} 54 55static u64 efx_get_ulong_stat(void *field) 56{ 57 return *(unsigned long *)field; 58} 59 60static u64 efx_get_u64_stat(void *field) 61{ 62 return *(u64 *) field; 63} 64 65static u64 efx_get_atomic_stat(void *field) 66{ 67 return atomic_read((atomic_t *) field); 68} 69 70#define EFX_ETHTOOL_ULONG_MAC_STAT(field) \ 71 EFX_ETHTOOL_STAT(field, mac_stats, field, \ 72 unsigned long, efx_get_ulong_stat) 73 74#define EFX_ETHTOOL_U64_MAC_STAT(field) \ 75 EFX_ETHTOOL_STAT(field, mac_stats, field, \ 76 u64, efx_get_u64_stat) 77 78#define EFX_ETHTOOL_UINT_NIC_STAT(name) \ 79 EFX_ETHTOOL_STAT(name, nic, n_##name, \ 80 unsigned int, efx_get_uint_stat) 81 82#define EFX_ETHTOOL_ATOMIC_NIC_ERROR_STAT(field) \ 83 EFX_ETHTOOL_STAT(field, nic, field, \ 84 atomic_t, efx_get_atomic_stat) 85 86#define EFX_ETHTOOL_UINT_CHANNEL_STAT(field) \ 87 EFX_ETHTOOL_STAT(field, channel, n_##field, \ 88 unsigned int, efx_get_uint_stat) 89 90#define EFX_ETHTOOL_UINT_TXQ_STAT(field) \ 91 EFX_ETHTOOL_STAT(tx_##field, tx_queue, field, \ 92 unsigned int, efx_get_uint_stat) 93 94static struct efx_ethtool_stat efx_ethtool_stats[] = { 95 EFX_ETHTOOL_U64_MAC_STAT(tx_bytes), 96 EFX_ETHTOOL_U64_MAC_STAT(tx_good_bytes), 97 EFX_ETHTOOL_U64_MAC_STAT(tx_bad_bytes), 98 EFX_ETHTOOL_ULONG_MAC_STAT(tx_packets), 99 EFX_ETHTOOL_ULONG_MAC_STAT(tx_bad), 100 EFX_ETHTOOL_ULONG_MAC_STAT(tx_pause), 101 EFX_ETHTOOL_ULONG_MAC_STAT(tx_control), 102 EFX_ETHTOOL_ULONG_MAC_STAT(tx_unicast), 103 EFX_ETHTOOL_ULONG_MAC_STAT(tx_multicast), 104 EFX_ETHTOOL_ULONG_MAC_STAT(tx_broadcast), 105 EFX_ETHTOOL_ULONG_MAC_STAT(tx_lt64), 106 EFX_ETHTOOL_ULONG_MAC_STAT(tx_64), 107 EFX_ETHTOOL_ULONG_MAC_STAT(tx_65_to_127), 108 EFX_ETHTOOL_ULONG_MAC_STAT(tx_128_to_255), 109 EFX_ETHTOOL_ULONG_MAC_STAT(tx_256_to_511), 110 EFX_ETHTOOL_ULONG_MAC_STAT(tx_512_to_1023), 111 EFX_ETHTOOL_ULONG_MAC_STAT(tx_1024_to_15xx), 112 EFX_ETHTOOL_ULONG_MAC_STAT(tx_15xx_to_jumbo), 113 EFX_ETHTOOL_ULONG_MAC_STAT(tx_gtjumbo), 114 EFX_ETHTOOL_ULONG_MAC_STAT(tx_collision), 115 EFX_ETHTOOL_ULONG_MAC_STAT(tx_single_collision), 116 EFX_ETHTOOL_ULONG_MAC_STAT(tx_multiple_collision), 117 EFX_ETHTOOL_ULONG_MAC_STAT(tx_excessive_collision), 118 EFX_ETHTOOL_ULONG_MAC_STAT(tx_deferred), 119 EFX_ETHTOOL_ULONG_MAC_STAT(tx_late_collision), 120 EFX_ETHTOOL_ULONG_MAC_STAT(tx_excessive_deferred), 121 EFX_ETHTOOL_ULONG_MAC_STAT(tx_non_tcpudp), 122 EFX_ETHTOOL_ULONG_MAC_STAT(tx_mac_src_error), 123 EFX_ETHTOOL_ULONG_MAC_STAT(tx_ip_src_error), 124 EFX_ETHTOOL_UINT_TXQ_STAT(tso_bursts), 125 EFX_ETHTOOL_UINT_TXQ_STAT(tso_long_headers), 126 EFX_ETHTOOL_UINT_TXQ_STAT(tso_packets), 127 EFX_ETHTOOL_UINT_TXQ_STAT(pushes), 128 EFX_ETHTOOL_U64_MAC_STAT(rx_bytes), 129 EFX_ETHTOOL_U64_MAC_STAT(rx_good_bytes), 130 EFX_ETHTOOL_U64_MAC_STAT(rx_bad_bytes), 131 EFX_ETHTOOL_ULONG_MAC_STAT(rx_packets), 132 EFX_ETHTOOL_ULONG_MAC_STAT(rx_good), 133 EFX_ETHTOOL_ULONG_MAC_STAT(rx_bad), 134 EFX_ETHTOOL_ULONG_MAC_STAT(rx_pause), 135 EFX_ETHTOOL_ULONG_MAC_STAT(rx_control), 136 EFX_ETHTOOL_ULONG_MAC_STAT(rx_unicast), 137 EFX_ETHTOOL_ULONG_MAC_STAT(rx_multicast), 138 EFX_ETHTOOL_ULONG_MAC_STAT(rx_broadcast), 139 EFX_ETHTOOL_ULONG_MAC_STAT(rx_lt64), 140 EFX_ETHTOOL_ULONG_MAC_STAT(rx_64), 141 EFX_ETHTOOL_ULONG_MAC_STAT(rx_65_to_127), 142 EFX_ETHTOOL_ULONG_MAC_STAT(rx_128_to_255), 143 EFX_ETHTOOL_ULONG_MAC_STAT(rx_256_to_511), 144 EFX_ETHTOOL_ULONG_MAC_STAT(rx_512_to_1023), 145 EFX_ETHTOOL_ULONG_MAC_STAT(rx_1024_to_15xx), 146 EFX_ETHTOOL_ULONG_MAC_STAT(rx_15xx_to_jumbo), 147 EFX_ETHTOOL_ULONG_MAC_STAT(rx_gtjumbo), 148 EFX_ETHTOOL_ULONG_MAC_STAT(rx_bad_lt64), 149 EFX_ETHTOOL_ULONG_MAC_STAT(rx_bad_64_to_15xx), 150 EFX_ETHTOOL_ULONG_MAC_STAT(rx_bad_15xx_to_jumbo), 151 EFX_ETHTOOL_ULONG_MAC_STAT(rx_bad_gtjumbo), 152 EFX_ETHTOOL_ULONG_MAC_STAT(rx_overflow), 153 EFX_ETHTOOL_ULONG_MAC_STAT(rx_missed), 154 EFX_ETHTOOL_ULONG_MAC_STAT(rx_false_carrier), 155 EFX_ETHTOOL_ULONG_MAC_STAT(rx_symbol_error), 156 EFX_ETHTOOL_ULONG_MAC_STAT(rx_align_error), 157 EFX_ETHTOOL_ULONG_MAC_STAT(rx_length_error), 158 EFX_ETHTOOL_ULONG_MAC_STAT(rx_internal_error), 159 EFX_ETHTOOL_UINT_NIC_STAT(rx_nodesc_drop_cnt), 160 EFX_ETHTOOL_ATOMIC_NIC_ERROR_STAT(rx_reset), 161 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_tobe_disc), 162 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_ip_hdr_chksum_err), 163 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_tcp_udp_chksum_err), 164 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_mcast_mismatch), 165 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_frm_trunc), 166}; 167 168/* Number of ethtool statistics */ 169#define EFX_ETHTOOL_NUM_STATS ARRAY_SIZE(efx_ethtool_stats) 170 171#define EFX_ETHTOOL_EEPROM_MAGIC 0xEFAB 172 173/************************************************************************** 174 * 175 * Ethtool operations 176 * 177 ************************************************************************** 178 */ 179 180/* Identify device by flashing LEDs */ 181static int efx_ethtool_phys_id(struct net_device *net_dev, u32 count) 182{ 183 struct efx_nic *efx = netdev_priv(net_dev); 184 185 do { 186 efx->type->set_id_led(efx, EFX_LED_ON); 187 schedule_timeout_interruptible(HZ / 2); 188 189 efx->type->set_id_led(efx, EFX_LED_OFF); 190 schedule_timeout_interruptible(HZ / 2); 191 } while (!signal_pending(current) && --count != 0); 192 193 efx->type->set_id_led(efx, EFX_LED_DEFAULT); 194 return 0; 195} 196 197/* This must be called with rtnl_lock held. */ 198static int efx_ethtool_get_settings(struct net_device *net_dev, 199 struct ethtool_cmd *ecmd) 200{ 201 struct efx_nic *efx = netdev_priv(net_dev); 202 struct efx_link_state *link_state = &efx->link_state; 203 204 mutex_lock(&efx->mac_lock); 205 efx->phy_op->get_settings(efx, ecmd); 206 mutex_unlock(&efx->mac_lock); 207 208 /* GMAC does not support 1000Mbps HD */ 209 ecmd->supported &= ~SUPPORTED_1000baseT_Half; 210 /* Both MACs support pause frames (bidirectional and respond-only) */ 211 ecmd->supported |= SUPPORTED_Pause | SUPPORTED_Asym_Pause; 212 213 if (LOOPBACK_INTERNAL(efx)) { 214 ecmd->speed = link_state->speed; 215 ecmd->duplex = link_state->fd ? DUPLEX_FULL : DUPLEX_HALF; 216 } 217 218 return 0; 219} 220 221/* This must be called with rtnl_lock held. */ 222static int efx_ethtool_set_settings(struct net_device *net_dev, 223 struct ethtool_cmd *ecmd) 224{ 225 struct efx_nic *efx = netdev_priv(net_dev); 226 int rc; 227 228 /* GMAC does not support 1000Mbps HD */ 229 if (ecmd->speed == SPEED_1000 && ecmd->duplex != DUPLEX_FULL) { 230 netif_dbg(efx, drv, efx->net_dev, 231 "rejecting unsupported 1000Mbps HD setting\n"); 232 return -EINVAL; 233 } 234 235 mutex_lock(&efx->mac_lock); 236 rc = efx->phy_op->set_settings(efx, ecmd); 237 mutex_unlock(&efx->mac_lock); 238 return rc; 239} 240 241static void efx_ethtool_get_drvinfo(struct net_device *net_dev, 242 struct ethtool_drvinfo *info) 243{ 244 struct efx_nic *efx = netdev_priv(net_dev); 245 246 strlcpy(info->driver, KBUILD_MODNAME, sizeof(info->driver)); 247 strlcpy(info->version, EFX_DRIVER_VERSION, sizeof(info->version)); 248 if (efx_nic_rev(efx) >= EFX_REV_SIENA_A0) 249 efx_mcdi_print_fwver(efx, info->fw_version, 250 sizeof(info->fw_version)); 251 strlcpy(info->bus_info, pci_name(efx->pci_dev), sizeof(info->bus_info)); 252} 253 254static int efx_ethtool_get_regs_len(struct net_device *net_dev) 255{ 256 return efx_nic_get_regs_len(netdev_priv(net_dev)); 257} 258 259static void efx_ethtool_get_regs(struct net_device *net_dev, 260 struct ethtool_regs *regs, void *buf) 261{ 262 struct efx_nic *efx = netdev_priv(net_dev); 263 264 regs->version = efx->type->revision; 265 efx_nic_get_regs(efx, buf); 266} 267 268static u32 efx_ethtool_get_msglevel(struct net_device *net_dev) 269{ 270 struct efx_nic *efx = netdev_priv(net_dev); 271 return efx->msg_enable; 272} 273 274static void efx_ethtool_set_msglevel(struct net_device *net_dev, u32 msg_enable) 275{ 276 struct efx_nic *efx = netdev_priv(net_dev); 277 efx->msg_enable = msg_enable; 278} 279 280/** 281 * efx_fill_test - fill in an individual self-test entry 282 * @test_index: Index of the test 283 * @strings: Ethtool strings, or %NULL 284 * @data: Ethtool test results, or %NULL 285 * @test: Pointer to test result (used only if data != %NULL) 286 * @unit_format: Unit name format (e.g. "chan\%d") 287 * @unit_id: Unit id (e.g. 0 for "chan0") 288 * @test_format: Test name format (e.g. "loopback.\%s.tx.sent") 289 * @test_id: Test id (e.g. "PHYXS" for "loopback.PHYXS.tx_sent") 290 * 291 * Fill in an individual self-test entry. 292 */ 293static void efx_fill_test(unsigned int test_index, 294 struct ethtool_string *strings, u64 *data, 295 int *test, const char *unit_format, int unit_id, 296 const char *test_format, const char *test_id) 297{ 298 struct ethtool_string unit_str, test_str; 299 300 /* Fill data value, if applicable */ 301 if (data) 302 data[test_index] = *test; 303 304 /* Fill string, if applicable */ 305 if (strings) { 306 if (strchr(unit_format, '%')) 307 snprintf(unit_str.name, sizeof(unit_str.name), 308 unit_format, unit_id); 309 else 310 strcpy(unit_str.name, unit_format); 311 snprintf(test_str.name, sizeof(test_str.name), 312 test_format, test_id); 313 snprintf(strings[test_index].name, 314 sizeof(strings[test_index].name), 315 "%-6s %-24s", unit_str.name, test_str.name); 316 } 317} 318 319#define EFX_CHANNEL_NAME(_channel) "chan%d", _channel->channel 320#define EFX_TX_QUEUE_NAME(_tx_queue) "txq%d", _tx_queue->queue 321#define EFX_RX_QUEUE_NAME(_rx_queue) "rxq%d", _rx_queue->queue 322#define EFX_LOOPBACK_NAME(_mode, _counter) \ 323 "loopback.%s." _counter, STRING_TABLE_LOOKUP(_mode, efx_loopback_mode) 324 325/** 326 * efx_fill_loopback_test - fill in a block of loopback self-test entries 327 * @efx: Efx NIC 328 * @lb_tests: Efx loopback self-test results structure 329 * @mode: Loopback test mode 330 * @test_index: Starting index of the test 331 * @strings: Ethtool strings, or %NULL 332 * @data: Ethtool test results, or %NULL 333 */ 334static int efx_fill_loopback_test(struct efx_nic *efx, 335 struct efx_loopback_self_tests *lb_tests, 336 enum efx_loopback_mode mode, 337 unsigned int test_index, 338 struct ethtool_string *strings, u64 *data) 339{ 340 struct efx_channel *channel = efx_get_channel(efx, 0); 341 struct efx_tx_queue *tx_queue; 342 343 efx_for_each_channel_tx_queue(tx_queue, channel) { 344 efx_fill_test(test_index++, strings, data, 345 &lb_tests->tx_sent[tx_queue->queue], 346 EFX_TX_QUEUE_NAME(tx_queue), 347 EFX_LOOPBACK_NAME(mode, "tx_sent")); 348 efx_fill_test(test_index++, strings, data, 349 &lb_tests->tx_done[tx_queue->queue], 350 EFX_TX_QUEUE_NAME(tx_queue), 351 EFX_LOOPBACK_NAME(mode, "tx_done")); 352 } 353 efx_fill_test(test_index++, strings, data, 354 &lb_tests->rx_good, 355 "rx", 0, 356 EFX_LOOPBACK_NAME(mode, "rx_good")); 357 efx_fill_test(test_index++, strings, data, 358 &lb_tests->rx_bad, 359 "rx", 0, 360 EFX_LOOPBACK_NAME(mode, "rx_bad")); 361 362 return test_index; 363} 364 365/** 366 * efx_ethtool_fill_self_tests - get self-test details 367 * @efx: Efx NIC 368 * @tests: Efx self-test results structure, or %NULL 369 * @strings: Ethtool strings, or %NULL 370 * @data: Ethtool test results, or %NULL 371 */ 372static int efx_ethtool_fill_self_tests(struct efx_nic *efx, 373 struct efx_self_tests *tests, 374 struct ethtool_string *strings, 375 u64 *data) 376{ 377 struct efx_channel *channel; 378 unsigned int n = 0, i; 379 enum efx_loopback_mode mode; 380 381 efx_fill_test(n++, strings, data, &tests->phy_alive, 382 "phy", 0, "alive", NULL); 383 efx_fill_test(n++, strings, data, &tests->nvram, 384 "core", 0, "nvram", NULL); 385 efx_fill_test(n++, strings, data, &tests->interrupt, 386 "core", 0, "interrupt", NULL); 387 388 /* Event queues */ 389 efx_for_each_channel(channel, efx) { 390 efx_fill_test(n++, strings, data, 391 &tests->eventq_dma[channel->channel], 392 EFX_CHANNEL_NAME(channel), 393 "eventq.dma", NULL); 394 efx_fill_test(n++, strings, data, 395 &tests->eventq_int[channel->channel], 396 EFX_CHANNEL_NAME(channel), 397 "eventq.int", NULL); 398 efx_fill_test(n++, strings, data, 399 &tests->eventq_poll[channel->channel], 400 EFX_CHANNEL_NAME(channel), 401 "eventq.poll", NULL); 402 } 403 404 efx_fill_test(n++, strings, data, &tests->registers, 405 "core", 0, "registers", NULL); 406 407 if (efx->phy_op->run_tests != NULL) { 408 EFX_BUG_ON_PARANOID(efx->phy_op->test_name == NULL); 409 410 for (i = 0; true; ++i) { 411 const char *name; 412 413 EFX_BUG_ON_PARANOID(i >= EFX_MAX_PHY_TESTS); 414 name = efx->phy_op->test_name(efx, i); 415 if (name == NULL) 416 break; 417 418 efx_fill_test(n++, strings, data, &tests->phy_ext[i], 419 "phy", 0, name, NULL); 420 } 421 } 422 423 /* Loopback tests */ 424 for (mode = LOOPBACK_NONE; mode <= LOOPBACK_TEST_MAX; mode++) { 425 if (!(efx->loopback_modes & (1 << mode))) 426 continue; 427 n = efx_fill_loopback_test(efx, 428 &tests->loopback[mode], mode, n, 429 strings, data); 430 } 431 432 return n; 433} 434 435static int efx_ethtool_get_sset_count(struct net_device *net_dev, 436 int string_set) 437{ 438 switch (string_set) { 439 case ETH_SS_STATS: 440 return EFX_ETHTOOL_NUM_STATS; 441 case ETH_SS_TEST: 442 return efx_ethtool_fill_self_tests(netdev_priv(net_dev), 443 NULL, NULL, NULL); 444 default: 445 return -EINVAL; 446 } 447} 448 449static void efx_ethtool_get_strings(struct net_device *net_dev, 450 u32 string_set, u8 *strings) 451{ 452 struct efx_nic *efx = netdev_priv(net_dev); 453 struct ethtool_string *ethtool_strings = 454 (struct ethtool_string *)strings; 455 int i; 456 457 switch (string_set) { 458 case ETH_SS_STATS: 459 for (i = 0; i < EFX_ETHTOOL_NUM_STATS; i++) 460 strncpy(ethtool_strings[i].name, 461 efx_ethtool_stats[i].name, 462 sizeof(ethtool_strings[i].name)); 463 break; 464 case ETH_SS_TEST: 465 efx_ethtool_fill_self_tests(efx, NULL, 466 ethtool_strings, NULL); 467 break; 468 default: 469 /* No other string sets */ 470 break; 471 } 472} 473 474static void efx_ethtool_get_stats(struct net_device *net_dev, 475 struct ethtool_stats *stats, 476 u64 *data) 477{ 478 struct efx_nic *efx = netdev_priv(net_dev); 479 struct efx_mac_stats *mac_stats = &efx->mac_stats; 480 struct efx_ethtool_stat *stat; 481 struct efx_channel *channel; 482 struct efx_tx_queue *tx_queue; 483 struct rtnl_link_stats64 temp; 484 int i; 485 486 EFX_BUG_ON_PARANOID(stats->n_stats != EFX_ETHTOOL_NUM_STATS); 487 488 /* Update MAC and NIC statistics */ 489 dev_get_stats(net_dev, &temp); 490 491 /* Fill detailed statistics buffer */ 492 for (i = 0; i < EFX_ETHTOOL_NUM_STATS; i++) { 493 stat = &efx_ethtool_stats[i]; 494 switch (stat->source) { 495 case EFX_ETHTOOL_STAT_SOURCE_mac_stats: 496 data[i] = stat->get_stat((void *)mac_stats + 497 stat->offset); 498 break; 499 case EFX_ETHTOOL_STAT_SOURCE_nic: 500 data[i] = stat->get_stat((void *)efx + stat->offset); 501 break; 502 case EFX_ETHTOOL_STAT_SOURCE_channel: 503 data[i] = 0; 504 efx_for_each_channel(channel, efx) 505 data[i] += stat->get_stat((void *)channel + 506 stat->offset); 507 break; 508 case EFX_ETHTOOL_STAT_SOURCE_tx_queue: 509 data[i] = 0; 510 efx_for_each_channel(channel, efx) { 511 efx_for_each_channel_tx_queue(tx_queue, channel) 512 data[i] += 513 stat->get_stat((void *)tx_queue 514 + stat->offset); 515 } 516 break; 517 } 518 } 519} 520 521static int efx_ethtool_set_tso(struct net_device *net_dev, u32 enable) 522{ 523 struct efx_nic *efx __attribute__ ((unused)) = netdev_priv(net_dev); 524 u32 features; 525 526 features = NETIF_F_TSO; 527 if (efx->type->offload_features & NETIF_F_V6_CSUM) 528 features |= NETIF_F_TSO6; 529 530 if (enable) 531 net_dev->features |= features; 532 else 533 net_dev->features &= ~features; 534 535 return 0; 536} 537 538static int efx_ethtool_set_tx_csum(struct net_device *net_dev, u32 enable) 539{ 540 struct efx_nic *efx = netdev_priv(net_dev); 541 u32 features = efx->type->offload_features & NETIF_F_ALL_CSUM; 542 543 if (enable) 544 net_dev->features |= features; 545 else 546 net_dev->features &= ~features; 547 548 return 0; 549} 550 551static int efx_ethtool_set_rx_csum(struct net_device *net_dev, u32 enable) 552{ 553 struct efx_nic *efx = netdev_priv(net_dev); 554 555 /* No way to stop the hardware doing the checks; we just 556 * ignore the result. 557 */ 558 efx->rx_checksum_enabled = !!enable; 559 560 return 0; 561} 562 563static u32 efx_ethtool_get_rx_csum(struct net_device *net_dev) 564{ 565 struct efx_nic *efx = netdev_priv(net_dev); 566 567 return efx->rx_checksum_enabled; 568} 569 570static int efx_ethtool_set_flags(struct net_device *net_dev, u32 data) 571{ 572 struct efx_nic *efx = netdev_priv(net_dev); 573 u32 supported = (efx->type->offload_features & 574 (ETH_FLAG_RXHASH | ETH_FLAG_NTUPLE)); 575 int rc; 576 577 rc = ethtool_op_set_flags(net_dev, data, supported); 578 if (rc) 579 return rc; 580 581 if (!(data & ETH_FLAG_NTUPLE)) 582 efx_filter_clear_rx(efx, EFX_FILTER_PRI_MANUAL); 583 584 return 0; 585} 586 587static void efx_ethtool_self_test(struct net_device *net_dev, 588 struct ethtool_test *test, u64 *data) 589{ 590 struct efx_nic *efx = netdev_priv(net_dev); 591 struct efx_self_tests *efx_tests; 592 int already_up; 593 int rc = -ENOMEM; 594 595 efx_tests = kzalloc(sizeof(*efx_tests), GFP_KERNEL); 596 if (!efx_tests) 597 goto fail; 598 599 600 ASSERT_RTNL(); 601 if (efx->state != STATE_RUNNING) { 602 rc = -EIO; 603 goto fail1; 604 } 605 606 netif_info(efx, drv, efx->net_dev, "starting %sline testing\n", 607 (test->flags & ETH_TEST_FL_OFFLINE) ? "off" : "on"); 608 609 /* We need rx buffers and interrupts. */ 610 already_up = (efx->net_dev->flags & IFF_UP); 611 if (!already_up) { 612 rc = dev_open(efx->net_dev); 613 if (rc) { 614 netif_err(efx, drv, efx->net_dev, 615 "failed opening device.\n"); 616 goto fail1; 617 } 618 } 619 620 rc = efx_selftest(efx, efx_tests, test->flags); 621 622 if (!already_up) 623 dev_close(efx->net_dev); 624 625 netif_info(efx, drv, efx->net_dev, "%s %sline self-tests\n", 626 rc == 0 ? "passed" : "failed", 627 (test->flags & ETH_TEST_FL_OFFLINE) ? "off" : "on"); 628 629fail1: 630 /* Fill ethtool results structures */ 631 efx_ethtool_fill_self_tests(efx, efx_tests, NULL, data); 632 kfree(efx_tests); 633fail: 634 if (rc) 635 test->flags |= ETH_TEST_FL_FAILED; 636} 637 638/* Restart autonegotiation */ 639static int efx_ethtool_nway_reset(struct net_device *net_dev) 640{ 641 struct efx_nic *efx = netdev_priv(net_dev); 642 643 return mdio45_nway_restart(&efx->mdio); 644} 645 646static int efx_ethtool_get_coalesce(struct net_device *net_dev, 647 struct ethtool_coalesce *coalesce) 648{ 649 struct efx_nic *efx = netdev_priv(net_dev); 650 struct efx_channel *channel; 651 652 memset(coalesce, 0, sizeof(*coalesce)); 653 654 /* Find lowest IRQ moderation across all used TX queues */ 655 coalesce->tx_coalesce_usecs_irq = ~((u32) 0); 656 efx_for_each_channel(channel, efx) { 657 if (!efx_channel_has_tx_queues(channel)) 658 continue; 659 if (channel->irq_moderation < coalesce->tx_coalesce_usecs_irq) { 660 if (channel->channel < efx->n_rx_channels) 661 coalesce->tx_coalesce_usecs_irq = 662 channel->irq_moderation; 663 else 664 coalesce->tx_coalesce_usecs_irq = 0; 665 } 666 } 667 668 coalesce->use_adaptive_rx_coalesce = efx->irq_rx_adaptive; 669 coalesce->rx_coalesce_usecs_irq = efx->irq_rx_moderation; 670 671 coalesce->tx_coalesce_usecs_irq *= EFX_IRQ_MOD_RESOLUTION; 672 coalesce->rx_coalesce_usecs_irq *= EFX_IRQ_MOD_RESOLUTION; 673 674 return 0; 675} 676 677/* Set coalescing parameters 678 * The difficulties occur for shared channels 679 */ 680static int efx_ethtool_set_coalesce(struct net_device *net_dev, 681 struct ethtool_coalesce *coalesce) 682{ 683 struct efx_nic *efx = netdev_priv(net_dev); 684 struct efx_channel *channel; 685 unsigned tx_usecs, rx_usecs, adaptive; 686 687 if (coalesce->use_adaptive_tx_coalesce) 688 return -EOPNOTSUPP; 689 690 if (coalesce->rx_coalesce_usecs || coalesce->tx_coalesce_usecs) { 691 netif_err(efx, drv, efx->net_dev, "invalid coalescing setting. " 692 "Only rx/tx_coalesce_usecs_irq are supported\n"); 693 return -EOPNOTSUPP; 694 } 695 696 rx_usecs = coalesce->rx_coalesce_usecs_irq; 697 tx_usecs = coalesce->tx_coalesce_usecs_irq; 698 adaptive = coalesce->use_adaptive_rx_coalesce; 699 700 /* If the channel is shared only allow RX parameters to be set */ 701 efx_for_each_channel(channel, efx) { 702 if (efx_channel_has_rx_queue(channel) && 703 efx_channel_has_tx_queues(channel) && 704 tx_usecs) { 705 netif_err(efx, drv, efx->net_dev, "Channel is shared. " 706 "Only RX coalescing may be set\n"); 707 return -EOPNOTSUPP; 708 } 709 } 710 711 efx_init_irq_moderation(efx, tx_usecs, rx_usecs, adaptive); 712 efx_for_each_channel(channel, efx) 713 efx->type->push_irq_moderation(channel); 714 715 return 0; 716} 717 718static void efx_ethtool_get_ringparam(struct net_device *net_dev, 719 struct ethtool_ringparam *ring) 720{ 721 struct efx_nic *efx = netdev_priv(net_dev); 722 723 ring->rx_max_pending = EFX_MAX_DMAQ_SIZE; 724 ring->tx_max_pending = EFX_MAX_DMAQ_SIZE; 725 ring->rx_mini_max_pending = 0; 726 ring->rx_jumbo_max_pending = 0; 727 ring->rx_pending = efx->rxq_entries; 728 ring->tx_pending = efx->txq_entries; 729 ring->rx_mini_pending = 0; 730 ring->rx_jumbo_pending = 0; 731} 732 733static int efx_ethtool_set_ringparam(struct net_device *net_dev, 734 struct ethtool_ringparam *ring) 735{ 736 struct efx_nic *efx = netdev_priv(net_dev); 737 738 if (ring->rx_mini_pending || ring->rx_jumbo_pending || 739 ring->rx_pending > EFX_MAX_DMAQ_SIZE || 740 ring->tx_pending > EFX_MAX_DMAQ_SIZE) 741 return -EINVAL; 742 743 if (ring->rx_pending < EFX_MIN_RING_SIZE || 744 ring->tx_pending < EFX_MIN_RING_SIZE) { 745 netif_err(efx, drv, efx->net_dev, 746 "TX and RX queues cannot be smaller than %ld\n", 747 EFX_MIN_RING_SIZE); 748 return -EINVAL; 749 } 750 751 return efx_realloc_channels(efx, ring->rx_pending, ring->tx_pending); 752} 753 754static int efx_ethtool_set_pauseparam(struct net_device *net_dev, 755 struct ethtool_pauseparam *pause) 756{ 757 struct efx_nic *efx = netdev_priv(net_dev); 758 enum efx_fc_type wanted_fc, old_fc; 759 u32 old_adv; 760 bool reset; 761 int rc = 0; 762 763 mutex_lock(&efx->mac_lock); 764 765 wanted_fc = ((pause->rx_pause ? EFX_FC_RX : 0) | 766 (pause->tx_pause ? EFX_FC_TX : 0) | 767 (pause->autoneg ? EFX_FC_AUTO : 0)); 768 769 if ((wanted_fc & EFX_FC_TX) && !(wanted_fc & EFX_FC_RX)) { 770 netif_dbg(efx, drv, efx->net_dev, 771 "Flow control unsupported: tx ON rx OFF\n"); 772 rc = -EINVAL; 773 goto out; 774 } 775 776 if ((wanted_fc & EFX_FC_AUTO) && !efx->link_advertising) { 777 netif_dbg(efx, drv, efx->net_dev, 778 "Autonegotiation is disabled\n"); 779 rc = -EINVAL; 780 goto out; 781 } 782 783 /* TX flow control may automatically turn itself off if the 784 * link partner (intermittently) stops responding to pause 785 * frames. There isn't any indication that this has happened, 786 * so the best we do is leave it up to the user to spot this 787 * and fix it be cycling transmit flow control on this end. */ 788 reset = (wanted_fc & EFX_FC_TX) && !(efx->wanted_fc & EFX_FC_TX); 789 if (EFX_WORKAROUND_11482(efx) && reset) { 790 if (efx_nic_rev(efx) == EFX_REV_FALCON_B0) { 791 /* Recover by resetting the EM block */ 792 falcon_stop_nic_stats(efx); 793 falcon_drain_tx_fifo(efx); 794 efx->mac_op->reconfigure(efx); 795 falcon_start_nic_stats(efx); 796 } else { 797 /* Schedule a reset to recover */ 798 efx_schedule_reset(efx, RESET_TYPE_INVISIBLE); 799 } 800 } 801 802 old_adv = efx->link_advertising; 803 old_fc = efx->wanted_fc; 804 efx_link_set_wanted_fc(efx, wanted_fc); 805 if (efx->link_advertising != old_adv || 806 (efx->wanted_fc ^ old_fc) & EFX_FC_AUTO) { 807 rc = efx->phy_op->reconfigure(efx); 808 if (rc) { 809 netif_err(efx, drv, efx->net_dev, 810 "Unable to advertise requested flow " 811 "control setting\n"); 812 goto out; 813 } 814 } 815 816 /* Reconfigure the MAC. The PHY *may* generate a link state change event 817 * if the user just changed the advertised capabilities, but there's no 818 * harm doing this twice */ 819 efx->mac_op->reconfigure(efx); 820 821out: 822 mutex_unlock(&efx->mac_lock); 823 824 return rc; 825} 826 827static void efx_ethtool_get_pauseparam(struct net_device *net_dev, 828 struct ethtool_pauseparam *pause) 829{ 830 struct efx_nic *efx = netdev_priv(net_dev); 831 832 pause->rx_pause = !!(efx->wanted_fc & EFX_FC_RX); 833 pause->tx_pause = !!(efx->wanted_fc & EFX_FC_TX); 834 pause->autoneg = !!(efx->wanted_fc & EFX_FC_AUTO); 835} 836 837 838static void efx_ethtool_get_wol(struct net_device *net_dev, 839 struct ethtool_wolinfo *wol) 840{ 841 struct efx_nic *efx = netdev_priv(net_dev); 842 return efx->type->get_wol(efx, wol); 843} 844 845 846static int efx_ethtool_set_wol(struct net_device *net_dev, 847 struct ethtool_wolinfo *wol) 848{ 849 struct efx_nic *efx = netdev_priv(net_dev); 850 return efx->type->set_wol(efx, wol->wolopts); 851} 852 853static int efx_ethtool_reset(struct net_device *net_dev, u32 *flags) 854{ 855 struct efx_nic *efx = netdev_priv(net_dev); 856 enum reset_type method; 857 enum { 858 ETH_RESET_EFX_INVISIBLE = (ETH_RESET_DMA | ETH_RESET_FILTER | 859 ETH_RESET_OFFLOAD | ETH_RESET_MAC) 860 }; 861 862 /* Check for minimal reset flags */ 863 if ((*flags & ETH_RESET_EFX_INVISIBLE) != ETH_RESET_EFX_INVISIBLE) 864 return -EINVAL; 865 *flags ^= ETH_RESET_EFX_INVISIBLE; 866 method = RESET_TYPE_INVISIBLE; 867 868 if (*flags & ETH_RESET_PHY) { 869 *flags ^= ETH_RESET_PHY; 870 method = RESET_TYPE_ALL; 871 } 872 873 if ((*flags & efx->type->reset_world_flags) == 874 efx->type->reset_world_flags) { 875 *flags ^= efx->type->reset_world_flags; 876 method = RESET_TYPE_WORLD; 877 } 878 879 return efx_reset(efx, method); 880} 881 882static int 883efx_ethtool_get_rxnfc(struct net_device *net_dev, 884 struct ethtool_rxnfc *info, void *rules __always_unused) 885{ 886 struct efx_nic *efx = netdev_priv(net_dev); 887 888 switch (info->cmd) { 889 case ETHTOOL_GRXRINGS: 890 info->data = efx->n_rx_channels; 891 return 0; 892 893 case ETHTOOL_GRXFH: { 894 unsigned min_revision = 0; 895 896 info->data = 0; 897 switch (info->flow_type) { 898 case TCP_V4_FLOW: 899 info->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; 900 /* fall through */ 901 case UDP_V4_FLOW: 902 case SCTP_V4_FLOW: 903 case AH_ESP_V4_FLOW: 904 case IPV4_FLOW: 905 info->data |= RXH_IP_SRC | RXH_IP_DST; 906 min_revision = EFX_REV_FALCON_B0; 907 break; 908 case TCP_V6_FLOW: 909 info->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; 910 /* fall through */ 911 case UDP_V6_FLOW: 912 case SCTP_V6_FLOW: 913 case AH_ESP_V6_FLOW: 914 case IPV6_FLOW: 915 info->data |= RXH_IP_SRC | RXH_IP_DST; 916 min_revision = EFX_REV_SIENA_A0; 917 break; 918 default: 919 break; 920 } 921 if (efx_nic_rev(efx) < min_revision) 922 info->data = 0; 923 return 0; 924 } 925 926 default: 927 return -EOPNOTSUPP; 928 } 929} 930 931static int efx_ethtool_set_rx_ntuple(struct net_device *net_dev, 932 struct ethtool_rx_ntuple *ntuple) 933{ 934 struct efx_nic *efx = netdev_priv(net_dev); 935 struct ethtool_tcpip4_spec *ip_entry = &ntuple->fs.h_u.tcp_ip4_spec; 936 struct ethtool_tcpip4_spec *ip_mask = &ntuple->fs.m_u.tcp_ip4_spec; 937 struct ethhdr *mac_entry = &ntuple->fs.h_u.ether_spec; 938 struct ethhdr *mac_mask = &ntuple->fs.m_u.ether_spec; 939 struct efx_filter_spec filter; 940 int rc; 941 942 /* Range-check action */ 943 if (ntuple->fs.action < ETHTOOL_RXNTUPLE_ACTION_CLEAR || 944 ntuple->fs.action >= (s32)efx->n_rx_channels) 945 return -EINVAL; 946 947 if (~ntuple->fs.data_mask) 948 return -EINVAL; 949 950 efx_filter_init_rx(&filter, EFX_FILTER_PRI_MANUAL, 0, 951 (ntuple->fs.action == ETHTOOL_RXNTUPLE_ACTION_DROP) ? 952 0xfff : ntuple->fs.action); 953 954 switch (ntuple->fs.flow_type) { 955 case TCP_V4_FLOW: 956 case UDP_V4_FLOW: { 957 u8 proto = (ntuple->fs.flow_type == TCP_V4_FLOW ? 958 IPPROTO_TCP : IPPROTO_UDP); 959 960 /* Must match all of destination, */ 961 if (ip_mask->ip4dst | ip_mask->pdst) 962 return -EINVAL; 963 /* all or none of source, */ 964 if ((ip_mask->ip4src | ip_mask->psrc) && 965 ((__force u32)~ip_mask->ip4src | 966 (__force u16)~ip_mask->psrc)) 967 return -EINVAL; 968 /* and nothing else */ 969 if ((u8)~ip_mask->tos | (u16)~ntuple->fs.vlan_tag_mask) 970 return -EINVAL; 971 972 if (!ip_mask->ip4src) 973 rc = efx_filter_set_ipv4_full(&filter, proto, 974 ip_entry->ip4dst, 975 ip_entry->pdst, 976 ip_entry->ip4src, 977 ip_entry->psrc); 978 else 979 rc = efx_filter_set_ipv4_local(&filter, proto, 980 ip_entry->ip4dst, 981 ip_entry->pdst); 982 if (rc) 983 return rc; 984 break; 985 } 986 987 case ETHER_FLOW: 988 /* Must match all of destination, */ 989 if (!is_zero_ether_addr(mac_mask->h_dest)) 990 return -EINVAL; 991 /* all or none of VID, */ 992 if (ntuple->fs.vlan_tag_mask != 0xf000 && 993 ntuple->fs.vlan_tag_mask != 0xffff) 994 return -EINVAL; 995 /* and nothing else */ 996 if (!is_broadcast_ether_addr(mac_mask->h_source) || 997 mac_mask->h_proto != htons(0xffff)) 998 return -EINVAL; 999 1000 rc = efx_filter_set_eth_local( 1001 &filter, 1002 (ntuple->fs.vlan_tag_mask == 0xf000) ? 1003 ntuple->fs.vlan_tag : EFX_FILTER_VID_UNSPEC, 1004 mac_entry->h_dest); 1005 if (rc) 1006 return rc; 1007 break; 1008 1009 default: 1010 return -EINVAL; 1011 } 1012 1013 if (ntuple->fs.action == ETHTOOL_RXNTUPLE_ACTION_CLEAR) 1014 return efx_filter_remove_filter(efx, &filter); 1015 else 1016 return efx_filter_insert_filter(efx, &filter, true); 1017} 1018 1019static int efx_ethtool_get_rxfh_indir(struct net_device *net_dev, 1020 struct ethtool_rxfh_indir *indir) 1021{ 1022 struct efx_nic *efx = netdev_priv(net_dev); 1023 size_t copy_size = 1024 min_t(size_t, indir->size, ARRAY_SIZE(efx->rx_indir_table)); 1025 1026 if (efx_nic_rev(efx) < EFX_REV_FALCON_B0) 1027 return -EOPNOTSUPP; 1028 1029 indir->size = ARRAY_SIZE(efx->rx_indir_table); 1030 memcpy(indir->ring_index, efx->rx_indir_table, 1031 copy_size * sizeof(indir->ring_index[0])); 1032 return 0; 1033} 1034 1035static int efx_ethtool_set_rxfh_indir(struct net_device *net_dev, 1036 const struct ethtool_rxfh_indir *indir) 1037{ 1038 struct efx_nic *efx = netdev_priv(net_dev); 1039 size_t i; 1040 1041 if (efx_nic_rev(efx) < EFX_REV_FALCON_B0) 1042 return -EOPNOTSUPP; 1043 1044 /* Validate size and indices */ 1045 if (indir->size != ARRAY_SIZE(efx->rx_indir_table)) 1046 return -EINVAL; 1047 for (i = 0; i < ARRAY_SIZE(efx->rx_indir_table); i++) 1048 if (indir->ring_index[i] >= efx->n_rx_channels) 1049 return -EINVAL; 1050 1051 memcpy(efx->rx_indir_table, indir->ring_index, 1052 sizeof(efx->rx_indir_table)); 1053 efx_nic_push_rx_indir_table(efx); 1054 return 0; 1055} 1056 1057const struct ethtool_ops efx_ethtool_ops = { 1058 .get_settings = efx_ethtool_get_settings, 1059 .set_settings = efx_ethtool_set_settings, 1060 .get_drvinfo = efx_ethtool_get_drvinfo, 1061 .get_regs_len = efx_ethtool_get_regs_len, 1062 .get_regs = efx_ethtool_get_regs, 1063 .get_msglevel = efx_ethtool_get_msglevel, 1064 .set_msglevel = efx_ethtool_set_msglevel, 1065 .nway_reset = efx_ethtool_nway_reset, 1066 .get_link = ethtool_op_get_link, 1067 .get_coalesce = efx_ethtool_get_coalesce, 1068 .set_coalesce = efx_ethtool_set_coalesce, 1069 .get_ringparam = efx_ethtool_get_ringparam, 1070 .set_ringparam = efx_ethtool_set_ringparam, 1071 .get_pauseparam = efx_ethtool_get_pauseparam, 1072 .set_pauseparam = efx_ethtool_set_pauseparam, 1073 .get_rx_csum = efx_ethtool_get_rx_csum, 1074 .set_rx_csum = efx_ethtool_set_rx_csum, 1075 .get_tx_csum = ethtool_op_get_tx_csum, 1076 /* Need to enable/disable IPv6 too */ 1077 .set_tx_csum = efx_ethtool_set_tx_csum, 1078 .get_sg = ethtool_op_get_sg, 1079 .set_sg = ethtool_op_set_sg, 1080 .get_tso = ethtool_op_get_tso, 1081 /* Need to enable/disable TSO-IPv6 too */ 1082 .set_tso = efx_ethtool_set_tso, 1083 .get_flags = ethtool_op_get_flags, 1084 .set_flags = efx_ethtool_set_flags, 1085 .get_sset_count = efx_ethtool_get_sset_count, 1086 .self_test = efx_ethtool_self_test, 1087 .get_strings = efx_ethtool_get_strings, 1088 .phys_id = efx_ethtool_phys_id, 1089 .get_ethtool_stats = efx_ethtool_get_stats, 1090 .get_wol = efx_ethtool_get_wol, 1091 .set_wol = efx_ethtool_set_wol, 1092 .reset = efx_ethtool_reset, 1093 .get_rxnfc = efx_ethtool_get_rxnfc, 1094 .set_rx_ntuple = efx_ethtool_set_rx_ntuple, 1095 .get_rxfh_indir = efx_ethtool_get_rxfh_indir, 1096 .set_rxfh_indir = efx_ethtool_set_rxfh_indir, 1097};