Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.5.6.
   1/*******************************************************************************
   2 *
   3 * Intel Ethernet Controller XL710 Family Linux Driver
   4 * Copyright(c) 2013 - 2014 Intel Corporation.
   5 *
   6 * This program is free software; you can redistribute it and/or modify it
   7 * under the terms and conditions of the GNU General Public License,
   8 * version 2, as published by the Free Software Foundation.
   9 *
  10 * This program is distributed in the hope it will be useful, but WITHOUT
  11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  12 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
  13 * more details.
  14 *
  15 * You should have received a copy of the GNU General Public License along
  16 * with this program.  If not, see <http://www.gnu.org/licenses/>.
  17 *
  18 * The full GNU General Public License is included in this distribution in
  19 * the file called "COPYING".
  20 *
  21 * Contact Information:
  22 * e1000-devel Mailing List <e1000-devel@lists.sourceforge.net>
  23 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
  24 *
  25 ******************************************************************************/
  26
  27/* ethtool support for i40e */
  28
  29#include "i40e.h"
  30#include "i40e_diag.h"
  31
  32struct i40e_stats {
  33	char stat_string[ETH_GSTRING_LEN];
  34	int sizeof_stat;
  35	int stat_offset;
  36};
  37
  38#define I40E_STAT(_type, _name, _stat) { \
  39	.stat_string = _name, \
  40	.sizeof_stat = FIELD_SIZEOF(_type, _stat), \
  41	.stat_offset = offsetof(_type, _stat) \
  42}
  43#define I40E_NETDEV_STAT(_net_stat) \
  44		I40E_STAT(struct net_device_stats, #_net_stat, _net_stat)
  45#define I40E_PF_STAT(_name, _stat) \
  46		I40E_STAT(struct i40e_pf, _name, _stat)
  47#define I40E_VSI_STAT(_name, _stat) \
  48		I40E_STAT(struct i40e_vsi, _name, _stat)
  49
  50static const struct i40e_stats i40e_gstrings_net_stats[] = {
  51	I40E_NETDEV_STAT(rx_packets),
  52	I40E_NETDEV_STAT(tx_packets),
  53	I40E_NETDEV_STAT(rx_bytes),
  54	I40E_NETDEV_STAT(tx_bytes),
  55	I40E_NETDEV_STAT(rx_errors),
  56	I40E_NETDEV_STAT(tx_errors),
  57	I40E_NETDEV_STAT(rx_dropped),
  58	I40E_NETDEV_STAT(tx_dropped),
  59	I40E_NETDEV_STAT(multicast),
  60	I40E_NETDEV_STAT(collisions),
  61	I40E_NETDEV_STAT(rx_length_errors),
  62	I40E_NETDEV_STAT(rx_crc_errors),
  63};
  64
  65static int i40e_add_fdir_ethtool(struct i40e_vsi *vsi,
  66				 struct ethtool_rxnfc *cmd);
  67
  68/* These PF_STATs might look like duplicates of some NETDEV_STATs,
  69 * but they are separate.  This device supports Virtualization, and
  70 * as such might have several netdevs supporting VMDq and FCoE going
  71 * through a single port.  The NETDEV_STATs are for individual netdevs
  72 * seen at the top of the stack, and the PF_STATs are for the physical
  73 * function at the bottom of the stack hosting those netdevs.
  74 *
  75 * The PF_STATs are appended to the netdev stats only when ethtool -S
  76 * is queried on the base PF netdev, not on the VMDq or FCoE netdev.
  77 */
  78static struct i40e_stats i40e_gstrings_stats[] = {
  79	I40E_PF_STAT("rx_bytes", stats.eth.rx_bytes),
  80	I40E_PF_STAT("tx_bytes", stats.eth.tx_bytes),
  81	I40E_PF_STAT("rx_errors", stats.eth.rx_errors),
  82	I40E_PF_STAT("tx_errors", stats.eth.tx_errors),
  83	I40E_PF_STAT("rx_dropped", stats.eth.rx_discards),
  84	I40E_PF_STAT("tx_dropped", stats.eth.tx_discards),
  85	I40E_PF_STAT("tx_dropped_link_down", stats.tx_dropped_link_down),
  86	I40E_PF_STAT("crc_errors", stats.crc_errors),
  87	I40E_PF_STAT("illegal_bytes", stats.illegal_bytes),
  88	I40E_PF_STAT("mac_local_faults", stats.mac_local_faults),
  89	I40E_PF_STAT("mac_remote_faults", stats.mac_remote_faults),
  90	I40E_PF_STAT("tx_timeout", tx_timeout_count),
  91	I40E_PF_STAT("rx_length_errors", stats.rx_length_errors),
  92	I40E_PF_STAT("link_xon_rx", stats.link_xon_rx),
  93	I40E_PF_STAT("link_xoff_rx", stats.link_xoff_rx),
  94	I40E_PF_STAT("link_xon_tx", stats.link_xon_tx),
  95	I40E_PF_STAT("link_xoff_tx", stats.link_xoff_tx),
  96	I40E_PF_STAT("rx_size_64", stats.rx_size_64),
  97	I40E_PF_STAT("rx_size_127", stats.rx_size_127),
  98	I40E_PF_STAT("rx_size_255", stats.rx_size_255),
  99	I40E_PF_STAT("rx_size_511", stats.rx_size_511),
 100	I40E_PF_STAT("rx_size_1023", stats.rx_size_1023),
 101	I40E_PF_STAT("rx_size_1522", stats.rx_size_1522),
 102	I40E_PF_STAT("rx_size_big", stats.rx_size_big),
 103	I40E_PF_STAT("tx_size_64", stats.tx_size_64),
 104	I40E_PF_STAT("tx_size_127", stats.tx_size_127),
 105	I40E_PF_STAT("tx_size_255", stats.tx_size_255),
 106	I40E_PF_STAT("tx_size_511", stats.tx_size_511),
 107	I40E_PF_STAT("tx_size_1023", stats.tx_size_1023),
 108	I40E_PF_STAT("tx_size_1522", stats.tx_size_1522),
 109	I40E_PF_STAT("tx_size_big", stats.tx_size_big),
 110	I40E_PF_STAT("rx_undersize", stats.rx_undersize),
 111	I40E_PF_STAT("rx_fragments", stats.rx_fragments),
 112	I40E_PF_STAT("rx_oversize", stats.rx_oversize),
 113	I40E_PF_STAT("rx_jabber", stats.rx_jabber),
 114	I40E_PF_STAT("VF_admin_queue_requests", vf_aq_requests),
 115	I40E_PF_STAT("tx_hwtstamp_timeouts", tx_hwtstamp_timeouts),
 116	I40E_PF_STAT("rx_hwtstamp_cleared", rx_hwtstamp_cleared),
 117	/* LPI stats */
 118	I40E_PF_STAT("tx_lpi_status", stats.tx_lpi_status),
 119	I40E_PF_STAT("rx_lpi_status", stats.rx_lpi_status),
 120	I40E_PF_STAT("tx_lpi_count", stats.tx_lpi_count),
 121	I40E_PF_STAT("rx_lpi_count", stats.rx_lpi_count),
 122};
 123
 124#define I40E_QUEUE_STATS_LEN(n) \
 125  ((((struct i40e_netdev_priv *)netdev_priv((n)))->vsi->num_queue_pairs + \
 126    ((struct i40e_netdev_priv *)netdev_priv((n)))->vsi->num_queue_pairs) * 2)
 127#define I40E_GLOBAL_STATS_LEN	ARRAY_SIZE(i40e_gstrings_stats)
 128#define I40E_NETDEV_STATS_LEN   ARRAY_SIZE(i40e_gstrings_net_stats)
 129#define I40E_VSI_STATS_LEN(n)   (I40E_NETDEV_STATS_LEN + \
 130				 I40E_QUEUE_STATS_LEN((n)))
 131#define I40E_PFC_STATS_LEN ( \
 132		(FIELD_SIZEOF(struct i40e_pf, stats.priority_xoff_rx) + \
 133		 FIELD_SIZEOF(struct i40e_pf, stats.priority_xon_rx) + \
 134		 FIELD_SIZEOF(struct i40e_pf, stats.priority_xoff_tx) + \
 135		 FIELD_SIZEOF(struct i40e_pf, stats.priority_xon_tx) + \
 136		 FIELD_SIZEOF(struct i40e_pf, stats.priority_xon_2_xoff)) \
 137		 / sizeof(u64))
 138#define I40E_PF_STATS_LEN(n)	(I40E_GLOBAL_STATS_LEN + \
 139				 I40E_PFC_STATS_LEN + \
 140				 I40E_VSI_STATS_LEN((n)))
 141
 142enum i40e_ethtool_test_id {
 143	I40E_ETH_TEST_REG = 0,
 144	I40E_ETH_TEST_EEPROM,
 145	I40E_ETH_TEST_INTR,
 146	I40E_ETH_TEST_LOOPBACK,
 147	I40E_ETH_TEST_LINK,
 148};
 149
 150static const char i40e_gstrings_test[][ETH_GSTRING_LEN] = {
 151	"Register test  (offline)",
 152	"Eeprom test    (offline)",
 153	"Interrupt test (offline)",
 154	"Loopback test  (offline)",
 155	"Link test   (on/offline)"
 156};
 157
 158#define I40E_TEST_LEN (sizeof(i40e_gstrings_test) / ETH_GSTRING_LEN)
 159
 160/**
 161 * i40e_get_settings - Get Link Speed and Duplex settings
 162 * @netdev: network interface device structure
 163 * @ecmd: ethtool command
 164 *
 165 * Reports speed/duplex settings based on media_type
 166 **/
 167static int i40e_get_settings(struct net_device *netdev,
 168			     struct ethtool_cmd *ecmd)
 169{
 170	struct i40e_netdev_priv *np = netdev_priv(netdev);
 171	struct i40e_pf *pf = np->vsi->back;
 172	struct i40e_hw *hw = &pf->hw;
 173	struct i40e_link_status *hw_link_info = &hw->phy.link_info;
 174	bool link_up = hw_link_info->link_info & I40E_AQ_LINK_UP;
 175	u32 link_speed = hw_link_info->link_speed;
 176
 177	/* hardware is either in 40G mode or 10G mode
 178	 * NOTE: this section initializes supported and advertising
 179	 */
 180	switch (hw_link_info->phy_type) {
 181	case I40E_PHY_TYPE_40GBASE_CR4:
 182	case I40E_PHY_TYPE_40GBASE_CR4_CU:
 183		ecmd->supported = SUPPORTED_40000baseCR4_Full;
 184		ecmd->advertising = ADVERTISED_40000baseCR4_Full;
 185		break;
 186	case I40E_PHY_TYPE_40GBASE_KR4:
 187		ecmd->supported = SUPPORTED_40000baseKR4_Full;
 188		ecmd->advertising = ADVERTISED_40000baseKR4_Full;
 189		break;
 190	case I40E_PHY_TYPE_40GBASE_SR4:
 191		ecmd->supported = SUPPORTED_40000baseSR4_Full;
 192		ecmd->advertising = ADVERTISED_40000baseSR4_Full;
 193		break;
 194	case I40E_PHY_TYPE_40GBASE_LR4:
 195		ecmd->supported = SUPPORTED_40000baseLR4_Full;
 196		ecmd->advertising = ADVERTISED_40000baseLR4_Full;
 197		break;
 198	case I40E_PHY_TYPE_10GBASE_KX4:
 199		ecmd->supported = SUPPORTED_10000baseKX4_Full;
 200		ecmd->advertising = ADVERTISED_10000baseKX4_Full;
 201		break;
 202	case I40E_PHY_TYPE_10GBASE_KR:
 203		ecmd->supported = SUPPORTED_10000baseKR_Full;
 204		ecmd->advertising = ADVERTISED_10000baseKR_Full;
 205		break;
 206	default:
 207		if (i40e_is_40G_device(hw->device_id)) {
 208			ecmd->supported = SUPPORTED_40000baseSR4_Full;
 209			ecmd->advertising = ADVERTISED_40000baseSR4_Full;
 210		} else {
 211			ecmd->supported = SUPPORTED_10000baseT_Full;
 212			ecmd->advertising = ADVERTISED_10000baseT_Full;
 213		}
 214		break;
 215	}
 216
 217	ecmd->supported |= SUPPORTED_Autoneg;
 218	ecmd->advertising |= ADVERTISED_Autoneg;
 219	ecmd->autoneg = ((hw_link_info->an_info & I40E_AQ_AN_COMPLETED) ?
 220			  AUTONEG_ENABLE : AUTONEG_DISABLE);
 221
 222	switch (hw->phy.media_type) {
 223	case I40E_MEDIA_TYPE_BACKPLANE:
 224		ecmd->supported |= SUPPORTED_Backplane;
 225		ecmd->advertising |= ADVERTISED_Backplane;
 226		ecmd->port = PORT_NONE;
 227		break;
 228	case I40E_MEDIA_TYPE_BASET:
 229		ecmd->supported |= SUPPORTED_TP;
 230		ecmd->advertising |= ADVERTISED_TP;
 231		ecmd->port = PORT_TP;
 232		break;
 233	case I40E_MEDIA_TYPE_DA:
 234	case I40E_MEDIA_TYPE_CX4:
 235		ecmd->supported |= SUPPORTED_FIBRE;
 236		ecmd->advertising |= ADVERTISED_FIBRE;
 237		ecmd->port = PORT_DA;
 238		break;
 239	case I40E_MEDIA_TYPE_FIBER:
 240		ecmd->supported |= SUPPORTED_FIBRE;
 241		ecmd->advertising |= ADVERTISED_FIBRE;
 242		ecmd->port = PORT_FIBRE;
 243		break;
 244	case I40E_MEDIA_TYPE_UNKNOWN:
 245	default:
 246		ecmd->port = PORT_OTHER;
 247		break;
 248	}
 249
 250	ecmd->transceiver = XCVR_EXTERNAL;
 251
 252	if (link_up) {
 253		switch (link_speed) {
 254		case I40E_LINK_SPEED_40GB:
 255			/* need a SPEED_40000 in ethtool.h */
 256			ethtool_cmd_speed_set(ecmd, 40000);
 257			break;
 258		case I40E_LINK_SPEED_10GB:
 259			ethtool_cmd_speed_set(ecmd, SPEED_10000);
 260			break;
 261		default:
 262			break;
 263		}
 264		ecmd->duplex = DUPLEX_FULL;
 265	} else {
 266		ethtool_cmd_speed_set(ecmd, SPEED_UNKNOWN);
 267		ecmd->duplex = DUPLEX_UNKNOWN;
 268	}
 269
 270	return 0;
 271}
 272
 273/**
 274 * i40e_get_pauseparam -  Get Flow Control status
 275 * Return tx/rx-pause status
 276 **/
 277static void i40e_get_pauseparam(struct net_device *netdev,
 278				struct ethtool_pauseparam *pause)
 279{
 280	struct i40e_netdev_priv *np = netdev_priv(netdev);
 281	struct i40e_pf *pf = np->vsi->back;
 282	struct i40e_hw *hw = &pf->hw;
 283	struct i40e_link_status *hw_link_info = &hw->phy.link_info;
 284
 285	pause->autoneg =
 286		((hw_link_info->an_info & I40E_AQ_AN_COMPLETED) ?
 287		  AUTONEG_ENABLE : AUTONEG_DISABLE);
 288
 289	if (hw->fc.current_mode == I40E_FC_RX_PAUSE) {
 290		pause->rx_pause = 1;
 291	} else if (hw->fc.current_mode == I40E_FC_TX_PAUSE) {
 292		pause->tx_pause = 1;
 293	} else if (hw->fc.current_mode == I40E_FC_FULL) {
 294		pause->rx_pause = 1;
 295		pause->tx_pause = 1;
 296	}
 297}
 298
 299static u32 i40e_get_msglevel(struct net_device *netdev)
 300{
 301	struct i40e_netdev_priv *np = netdev_priv(netdev);
 302	struct i40e_pf *pf = np->vsi->back;
 303
 304	return pf->msg_enable;
 305}
 306
 307static void i40e_set_msglevel(struct net_device *netdev, u32 data)
 308{
 309	struct i40e_netdev_priv *np = netdev_priv(netdev);
 310	struct i40e_pf *pf = np->vsi->back;
 311
 312	if (I40E_DEBUG_USER & data)
 313		pf->hw.debug_mask = data;
 314	pf->msg_enable = data;
 315}
 316
 317static int i40e_get_regs_len(struct net_device *netdev)
 318{
 319	int reg_count = 0;
 320	int i;
 321
 322	for (i = 0; i40e_reg_list[i].offset != 0; i++)
 323		reg_count += i40e_reg_list[i].elements;
 324
 325	return reg_count * sizeof(u32);
 326}
 327
 328static void i40e_get_regs(struct net_device *netdev, struct ethtool_regs *regs,
 329			  void *p)
 330{
 331	struct i40e_netdev_priv *np = netdev_priv(netdev);
 332	struct i40e_pf *pf = np->vsi->back;
 333	struct i40e_hw *hw = &pf->hw;
 334	u32 *reg_buf = p;
 335	int i, j, ri;
 336	u32 reg;
 337
 338	/* Tell ethtool which driver-version-specific regs output we have.
 339	 *
 340	 * At some point, if we have ethtool doing special formatting of
 341	 * this data, it will rely on this version number to know how to
 342	 * interpret things.  Hence, this needs to be updated if/when the
 343	 * diags register table is changed.
 344	 */
 345	regs->version = 1;
 346
 347	/* loop through the diags reg table for what to print */
 348	ri = 0;
 349	for (i = 0; i40e_reg_list[i].offset != 0; i++) {
 350		for (j = 0; j < i40e_reg_list[i].elements; j++) {
 351			reg = i40e_reg_list[i].offset
 352				+ (j * i40e_reg_list[i].stride);
 353			reg_buf[ri++] = rd32(hw, reg);
 354		}
 355	}
 356
 357}
 358
 359static int i40e_get_eeprom(struct net_device *netdev,
 360			   struct ethtool_eeprom *eeprom, u8 *bytes)
 361{
 362	struct i40e_netdev_priv *np = netdev_priv(netdev);
 363	struct i40e_hw *hw = &np->vsi->back->hw;
 364	struct i40e_pf *pf = np->vsi->back;
 365	int ret_val = 0, len;
 366	u8 *eeprom_buff;
 367	u16 i, sectors;
 368	bool last;
 369#define I40E_NVM_SECTOR_SIZE  4096
 370	if (eeprom->len == 0)
 371		return -EINVAL;
 372
 373	eeprom->magic = hw->vendor_id | (hw->device_id << 16);
 374
 375	eeprom_buff = kzalloc(eeprom->len, GFP_KERNEL);
 376	if (!eeprom_buff)
 377		return -ENOMEM;
 378
 379	ret_val = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
 380	if (ret_val) {
 381		dev_info(&pf->pdev->dev,
 382			 "Failed Acquiring NVM resource for read err=%d status=0x%x\n",
 383			 ret_val, hw->aq.asq_last_status);
 384		goto free_buff;
 385	}
 386
 387	sectors = eeprom->len / I40E_NVM_SECTOR_SIZE;
 388	sectors += (eeprom->len % I40E_NVM_SECTOR_SIZE) ? 1 : 0;
 389	len = I40E_NVM_SECTOR_SIZE;
 390	last = false;
 391	for (i = 0; i < sectors; i++) {
 392		if (i == (sectors - 1)) {
 393			len = eeprom->len - (I40E_NVM_SECTOR_SIZE * i);
 394			last = true;
 395		}
 396		ret_val = i40e_aq_read_nvm(hw, 0x0,
 397				eeprom->offset + (I40E_NVM_SECTOR_SIZE * i),
 398				len,
 399				eeprom_buff + (I40E_NVM_SECTOR_SIZE * i),
 400				last, NULL);
 401		if (ret_val) {
 402			dev_info(&pf->pdev->dev,
 403				 "read NVM failed err=%d status=0x%x\n",
 404				 ret_val, hw->aq.asq_last_status);
 405			goto release_nvm;
 406		}
 407	}
 408
 409release_nvm:
 410	i40e_release_nvm(hw);
 411	memcpy(bytes, eeprom_buff, eeprom->len);
 412free_buff:
 413	kfree(eeprom_buff);
 414	return ret_val;
 415}
 416
 417static int i40e_get_eeprom_len(struct net_device *netdev)
 418{
 419	struct i40e_netdev_priv *np = netdev_priv(netdev);
 420	struct i40e_hw *hw = &np->vsi->back->hw;
 421	u32 val;
 422
 423	val = (rd32(hw, I40E_GLPCI_LBARCTRL)
 424		& I40E_GLPCI_LBARCTRL_FL_SIZE_MASK)
 425		>> I40E_GLPCI_LBARCTRL_FL_SIZE_SHIFT;
 426	/* register returns value in power of 2, 64Kbyte chunks. */
 427	val = (64 * 1024) * (1 << val);
 428	return val;
 429}
 430
 431static void i40e_get_drvinfo(struct net_device *netdev,
 432			     struct ethtool_drvinfo *drvinfo)
 433{
 434	struct i40e_netdev_priv *np = netdev_priv(netdev);
 435	struct i40e_vsi *vsi = np->vsi;
 436	struct i40e_pf *pf = vsi->back;
 437
 438	strlcpy(drvinfo->driver, i40e_driver_name, sizeof(drvinfo->driver));
 439	strlcpy(drvinfo->version, i40e_driver_version_str,
 440		sizeof(drvinfo->version));
 441	strlcpy(drvinfo->fw_version, i40e_fw_version_str(&pf->hw),
 442		sizeof(drvinfo->fw_version));
 443	strlcpy(drvinfo->bus_info, pci_name(pf->pdev),
 444		sizeof(drvinfo->bus_info));
 445}
 446
 447static void i40e_get_ringparam(struct net_device *netdev,
 448			       struct ethtool_ringparam *ring)
 449{
 450	struct i40e_netdev_priv *np = netdev_priv(netdev);
 451	struct i40e_pf *pf = np->vsi->back;
 452	struct i40e_vsi *vsi = pf->vsi[pf->lan_vsi];
 453
 454	ring->rx_max_pending = I40E_MAX_NUM_DESCRIPTORS;
 455	ring->tx_max_pending = I40E_MAX_NUM_DESCRIPTORS;
 456	ring->rx_mini_max_pending = 0;
 457	ring->rx_jumbo_max_pending = 0;
 458	ring->rx_pending = vsi->rx_rings[0]->count;
 459	ring->tx_pending = vsi->tx_rings[0]->count;
 460	ring->rx_mini_pending = 0;
 461	ring->rx_jumbo_pending = 0;
 462}
 463
 464static int i40e_set_ringparam(struct net_device *netdev,
 465			      struct ethtool_ringparam *ring)
 466{
 467	struct i40e_ring *tx_rings = NULL, *rx_rings = NULL;
 468	struct i40e_netdev_priv *np = netdev_priv(netdev);
 469	struct i40e_vsi *vsi = np->vsi;
 470	struct i40e_pf *pf = vsi->back;
 471	u32 new_rx_count, new_tx_count;
 472	int i, err = 0;
 473
 474	if ((ring->rx_mini_pending) || (ring->rx_jumbo_pending))
 475		return -EINVAL;
 476
 477	if (ring->tx_pending > I40E_MAX_NUM_DESCRIPTORS ||
 478	    ring->tx_pending < I40E_MIN_NUM_DESCRIPTORS ||
 479	    ring->rx_pending > I40E_MAX_NUM_DESCRIPTORS ||
 480	    ring->rx_pending < I40E_MIN_NUM_DESCRIPTORS) {
 481		netdev_info(netdev,
 482			    "Descriptors requested (Tx: %d / Rx: %d) out of range [%d-%d]\n",
 483			    ring->tx_pending, ring->rx_pending,
 484			    I40E_MIN_NUM_DESCRIPTORS, I40E_MAX_NUM_DESCRIPTORS);
 485		return -EINVAL;
 486	}
 487
 488	new_tx_count = ALIGN(ring->tx_pending, I40E_REQ_DESCRIPTOR_MULTIPLE);
 489	new_rx_count = ALIGN(ring->rx_pending, I40E_REQ_DESCRIPTOR_MULTIPLE);
 490
 491	/* if nothing to do return success */
 492	if ((new_tx_count == vsi->tx_rings[0]->count) &&
 493	    (new_rx_count == vsi->rx_rings[0]->count))
 494		return 0;
 495
 496	while (test_and_set_bit(__I40E_CONFIG_BUSY, &pf->state))
 497		usleep_range(1000, 2000);
 498
 499	if (!netif_running(vsi->netdev)) {
 500		/* simple case - set for the next time the netdev is started */
 501		for (i = 0; i < vsi->num_queue_pairs; i++) {
 502			vsi->tx_rings[i]->count = new_tx_count;
 503			vsi->rx_rings[i]->count = new_rx_count;
 504		}
 505		goto done;
 506	}
 507
 508	/* We can't just free everything and then setup again,
 509	 * because the ISRs in MSI-X mode get passed pointers
 510	 * to the Tx and Rx ring structs.
 511	 */
 512
 513	/* alloc updated Tx resources */
 514	if (new_tx_count != vsi->tx_rings[0]->count) {
 515		netdev_info(netdev,
 516			    "Changing Tx descriptor count from %d to %d.\n",
 517			    vsi->tx_rings[0]->count, new_tx_count);
 518		tx_rings = kcalloc(vsi->alloc_queue_pairs,
 519				   sizeof(struct i40e_ring), GFP_KERNEL);
 520		if (!tx_rings) {
 521			err = -ENOMEM;
 522			goto done;
 523		}
 524
 525		for (i = 0; i < vsi->num_queue_pairs; i++) {
 526			/* clone ring and setup updated count */
 527			tx_rings[i] = *vsi->tx_rings[i];
 528			tx_rings[i].count = new_tx_count;
 529			err = i40e_setup_tx_descriptors(&tx_rings[i]);
 530			if (err) {
 531				while (i) {
 532					i--;
 533					i40e_free_tx_resources(&tx_rings[i]);
 534				}
 535				kfree(tx_rings);
 536				tx_rings = NULL;
 537
 538				goto done;
 539			}
 540		}
 541	}
 542
 543	/* alloc updated Rx resources */
 544	if (new_rx_count != vsi->rx_rings[0]->count) {
 545		netdev_info(netdev,
 546			    "Changing Rx descriptor count from %d to %d\n",
 547			    vsi->rx_rings[0]->count, new_rx_count);
 548		rx_rings = kcalloc(vsi->alloc_queue_pairs,
 549				   sizeof(struct i40e_ring), GFP_KERNEL);
 550		if (!rx_rings) {
 551			err = -ENOMEM;
 552			goto free_tx;
 553		}
 554
 555		for (i = 0; i < vsi->num_queue_pairs; i++) {
 556			/* clone ring and setup updated count */
 557			rx_rings[i] = *vsi->rx_rings[i];
 558			rx_rings[i].count = new_rx_count;
 559			err = i40e_setup_rx_descriptors(&rx_rings[i]);
 560			if (err) {
 561				while (i) {
 562					i--;
 563					i40e_free_rx_resources(&rx_rings[i]);
 564				}
 565				kfree(rx_rings);
 566				rx_rings = NULL;
 567
 568				goto free_tx;
 569			}
 570		}
 571	}
 572
 573	/* Bring interface down, copy in the new ring info,
 574	 * then restore the interface
 575	 */
 576	i40e_down(vsi);
 577
 578	if (tx_rings) {
 579		for (i = 0; i < vsi->num_queue_pairs; i++) {
 580			i40e_free_tx_resources(vsi->tx_rings[i]);
 581			*vsi->tx_rings[i] = tx_rings[i];
 582		}
 583		kfree(tx_rings);
 584		tx_rings = NULL;
 585	}
 586
 587	if (rx_rings) {
 588		for (i = 0; i < vsi->num_queue_pairs; i++) {
 589			i40e_free_rx_resources(vsi->rx_rings[i]);
 590			*vsi->rx_rings[i] = rx_rings[i];
 591		}
 592		kfree(rx_rings);
 593		rx_rings = NULL;
 594	}
 595
 596	i40e_up(vsi);
 597
 598free_tx:
 599	/* error cleanup if the Rx allocations failed after getting Tx */
 600	if (tx_rings) {
 601		for (i = 0; i < vsi->num_queue_pairs; i++)
 602			i40e_free_tx_resources(&tx_rings[i]);
 603		kfree(tx_rings);
 604		tx_rings = NULL;
 605	}
 606
 607done:
 608	clear_bit(__I40E_CONFIG_BUSY, &pf->state);
 609
 610	return err;
 611}
 612
 613static int i40e_get_sset_count(struct net_device *netdev, int sset)
 614{
 615	struct i40e_netdev_priv *np = netdev_priv(netdev);
 616	struct i40e_vsi *vsi = np->vsi;
 617	struct i40e_pf *pf = vsi->back;
 618
 619	switch (sset) {
 620	case ETH_SS_TEST:
 621		return I40E_TEST_LEN;
 622	case ETH_SS_STATS:
 623		if (vsi == pf->vsi[pf->lan_vsi])
 624			return I40E_PF_STATS_LEN(netdev);
 625		else
 626			return I40E_VSI_STATS_LEN(netdev);
 627	default:
 628		return -EOPNOTSUPP;
 629	}
 630}
 631
 632static void i40e_get_ethtool_stats(struct net_device *netdev,
 633				   struct ethtool_stats *stats, u64 *data)
 634{
 635	struct i40e_netdev_priv *np = netdev_priv(netdev);
 636	struct i40e_vsi *vsi = np->vsi;
 637	struct i40e_pf *pf = vsi->back;
 638	int i = 0;
 639	char *p;
 640	int j;
 641	struct rtnl_link_stats64 *net_stats = i40e_get_vsi_stats_struct(vsi);
 642	unsigned int start;
 643
 644	i40e_update_stats(vsi);
 645
 646	for (j = 0; j < I40E_NETDEV_STATS_LEN; j++) {
 647		p = (char *)net_stats + i40e_gstrings_net_stats[j].stat_offset;
 648		data[i++] = (i40e_gstrings_net_stats[j].sizeof_stat ==
 649			sizeof(u64)) ? *(u64 *)p : *(u32 *)p;
 650	}
 651	rcu_read_lock();
 652	for (j = 0; j < vsi->num_queue_pairs; j++, i += 4) {
 653		struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[j]);
 654		struct i40e_ring *rx_ring;
 655
 656		if (!tx_ring)
 657			continue;
 658
 659		/* process Tx ring statistics */
 660		do {
 661			start = u64_stats_fetch_begin_irq(&tx_ring->syncp);
 662			data[i] = tx_ring->stats.packets;
 663			data[i + 1] = tx_ring->stats.bytes;
 664		} while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start));
 665
 666		/* Rx ring is the 2nd half of the queue pair */
 667		rx_ring = &tx_ring[1];
 668		do {
 669			start = u64_stats_fetch_begin_irq(&rx_ring->syncp);
 670			data[i + 2] = rx_ring->stats.packets;
 671			data[i + 3] = rx_ring->stats.bytes;
 672		} while (u64_stats_fetch_retry_irq(&rx_ring->syncp, start));
 673	}
 674	rcu_read_unlock();
 675	if (vsi == pf->vsi[pf->lan_vsi]) {
 676		for (j = 0; j < I40E_GLOBAL_STATS_LEN; j++) {
 677			p = (char *)pf + i40e_gstrings_stats[j].stat_offset;
 678			data[i++] = (i40e_gstrings_stats[j].sizeof_stat ==
 679				   sizeof(u64)) ? *(u64 *)p : *(u32 *)p;
 680		}
 681		for (j = 0; j < I40E_MAX_USER_PRIORITY; j++) {
 682			data[i++] = pf->stats.priority_xon_tx[j];
 683			data[i++] = pf->stats.priority_xoff_tx[j];
 684		}
 685		for (j = 0; j < I40E_MAX_USER_PRIORITY; j++) {
 686			data[i++] = pf->stats.priority_xon_rx[j];
 687			data[i++] = pf->stats.priority_xoff_rx[j];
 688		}
 689		for (j = 0; j < I40E_MAX_USER_PRIORITY; j++)
 690			data[i++] = pf->stats.priority_xon_2_xoff[j];
 691	}
 692}
 693
 694static void i40e_get_strings(struct net_device *netdev, u32 stringset,
 695			     u8 *data)
 696{
 697	struct i40e_netdev_priv *np = netdev_priv(netdev);
 698	struct i40e_vsi *vsi = np->vsi;
 699	struct i40e_pf *pf = vsi->back;
 700	char *p = (char *)data;
 701	int i;
 702
 703	switch (stringset) {
 704	case ETH_SS_TEST:
 705		for (i = 0; i < I40E_TEST_LEN; i++) {
 706			memcpy(data, i40e_gstrings_test[i], ETH_GSTRING_LEN);
 707			data += ETH_GSTRING_LEN;
 708		}
 709		break;
 710	case ETH_SS_STATS:
 711		for (i = 0; i < I40E_NETDEV_STATS_LEN; i++) {
 712			snprintf(p, ETH_GSTRING_LEN, "%s",
 713				 i40e_gstrings_net_stats[i].stat_string);
 714			p += ETH_GSTRING_LEN;
 715		}
 716		for (i = 0; i < vsi->num_queue_pairs; i++) {
 717			snprintf(p, ETH_GSTRING_LEN, "tx-%u.tx_packets", i);
 718			p += ETH_GSTRING_LEN;
 719			snprintf(p, ETH_GSTRING_LEN, "tx-%u.tx_bytes", i);
 720			p += ETH_GSTRING_LEN;
 721			snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_packets", i);
 722			p += ETH_GSTRING_LEN;
 723			snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_bytes", i);
 724			p += ETH_GSTRING_LEN;
 725		}
 726		if (vsi == pf->vsi[pf->lan_vsi]) {
 727			for (i = 0; i < I40E_GLOBAL_STATS_LEN; i++) {
 728				snprintf(p, ETH_GSTRING_LEN, "port.%s",
 729					 i40e_gstrings_stats[i].stat_string);
 730				p += ETH_GSTRING_LEN;
 731			}
 732			for (i = 0; i < I40E_MAX_USER_PRIORITY; i++) {
 733				snprintf(p, ETH_GSTRING_LEN,
 734					 "port.tx_priority_%u_xon", i);
 735				p += ETH_GSTRING_LEN;
 736				snprintf(p, ETH_GSTRING_LEN,
 737					 "port.tx_priority_%u_xoff", i);
 738				p += ETH_GSTRING_LEN;
 739			}
 740			for (i = 0; i < I40E_MAX_USER_PRIORITY; i++) {
 741				snprintf(p, ETH_GSTRING_LEN,
 742					 "port.rx_priority_%u_xon", i);
 743				p += ETH_GSTRING_LEN;
 744				snprintf(p, ETH_GSTRING_LEN,
 745					 "port.rx_priority_%u_xoff", i);
 746				p += ETH_GSTRING_LEN;
 747			}
 748			for (i = 0; i < I40E_MAX_USER_PRIORITY; i++) {
 749				snprintf(p, ETH_GSTRING_LEN,
 750					 "port.rx_priority_%u_xon_2_xoff", i);
 751				p += ETH_GSTRING_LEN;
 752			}
 753		}
 754		/* BUG_ON(p - data != I40E_STATS_LEN * ETH_GSTRING_LEN); */
 755		break;
 756	}
 757}
 758
 759static int i40e_get_ts_info(struct net_device *dev,
 760			    struct ethtool_ts_info *info)
 761{
 762	struct i40e_pf *pf = i40e_netdev_to_pf(dev);
 763
 764	info->so_timestamping = SOF_TIMESTAMPING_TX_SOFTWARE |
 765				SOF_TIMESTAMPING_RX_SOFTWARE |
 766				SOF_TIMESTAMPING_SOFTWARE |
 767				SOF_TIMESTAMPING_TX_HARDWARE |
 768				SOF_TIMESTAMPING_RX_HARDWARE |
 769				SOF_TIMESTAMPING_RAW_HARDWARE;
 770
 771	if (pf->ptp_clock)
 772		info->phc_index = ptp_clock_index(pf->ptp_clock);
 773	else
 774		info->phc_index = -1;
 775
 776	info->tx_types = (1 << HWTSTAMP_TX_OFF) | (1 << HWTSTAMP_TX_ON);
 777
 778	info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
 779			   (1 << HWTSTAMP_FILTER_PTP_V1_L4_SYNC) |
 780			   (1 << HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ) |
 781			   (1 << HWTSTAMP_FILTER_PTP_V2_EVENT) |
 782			   (1 << HWTSTAMP_FILTER_PTP_V2_L2_EVENT) |
 783			   (1 << HWTSTAMP_FILTER_PTP_V2_L4_EVENT) |
 784			   (1 << HWTSTAMP_FILTER_PTP_V2_SYNC) |
 785			   (1 << HWTSTAMP_FILTER_PTP_V2_L2_SYNC) |
 786			   (1 << HWTSTAMP_FILTER_PTP_V2_L4_SYNC) |
 787			   (1 << HWTSTAMP_FILTER_PTP_V2_DELAY_REQ) |
 788			   (1 << HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ) |
 789			   (1 << HWTSTAMP_FILTER_PTP_V2_L4_DELAY_REQ);
 790
 791	return 0;
 792}
 793
 794static int i40e_link_test(struct net_device *netdev, u64 *data)
 795{
 796	struct i40e_netdev_priv *np = netdev_priv(netdev);
 797	struct i40e_pf *pf = np->vsi->back;
 798
 799	netif_info(pf, hw, netdev, "link test\n");
 800	if (i40e_get_link_status(&pf->hw))
 801		*data = 0;
 802	else
 803		*data = 1;
 804
 805	return *data;
 806}
 807
 808static int i40e_reg_test(struct net_device *netdev, u64 *data)
 809{
 810	struct i40e_netdev_priv *np = netdev_priv(netdev);
 811	struct i40e_pf *pf = np->vsi->back;
 812
 813	netif_info(pf, hw, netdev, "register test\n");
 814	*data = i40e_diag_reg_test(&pf->hw);
 815
 816	return *data;
 817}
 818
 819static int i40e_eeprom_test(struct net_device *netdev, u64 *data)
 820{
 821	struct i40e_netdev_priv *np = netdev_priv(netdev);
 822	struct i40e_pf *pf = np->vsi->back;
 823
 824	netif_info(pf, hw, netdev, "eeprom test\n");
 825	*data = i40e_diag_eeprom_test(&pf->hw);
 826
 827	return *data;
 828}
 829
 830static int i40e_intr_test(struct net_device *netdev, u64 *data)
 831{
 832	struct i40e_netdev_priv *np = netdev_priv(netdev);
 833	struct i40e_pf *pf = np->vsi->back;
 834	u16 swc_old = pf->sw_int_count;
 835
 836	netif_info(pf, hw, netdev, "interrupt test\n");
 837	wr32(&pf->hw, I40E_PFINT_DYN_CTL0,
 838	     (I40E_PFINT_DYN_CTL0_INTENA_MASK |
 839	      I40E_PFINT_DYN_CTL0_SWINT_TRIG_MASK));
 840	usleep_range(1000, 2000);
 841	*data = (swc_old == pf->sw_int_count);
 842
 843	return *data;
 844}
 845
 846static int i40e_loopback_test(struct net_device *netdev, u64 *data)
 847{
 848	struct i40e_netdev_priv *np = netdev_priv(netdev);
 849	struct i40e_pf *pf = np->vsi->back;
 850
 851	netif_info(pf, hw, netdev, "loopback test not implemented\n");
 852	*data = 0;
 853
 854	return *data;
 855}
 856
 857static void i40e_diag_test(struct net_device *netdev,
 858			   struct ethtool_test *eth_test, u64 *data)
 859{
 860	struct i40e_netdev_priv *np = netdev_priv(netdev);
 861	struct i40e_pf *pf = np->vsi->back;
 862
 863	if (eth_test->flags == ETH_TEST_FL_OFFLINE) {
 864		/* Offline tests */
 865		netif_info(pf, drv, netdev, "offline testing starting\n");
 866
 867		set_bit(__I40E_TESTING, &pf->state);
 868
 869		/* Link test performed before hardware reset
 870		 * so autoneg doesn't interfere with test result
 871		 */
 872		if (i40e_link_test(netdev, &data[I40E_ETH_TEST_LINK]))
 873			eth_test->flags |= ETH_TEST_FL_FAILED;
 874
 875		if (i40e_eeprom_test(netdev, &data[I40E_ETH_TEST_EEPROM]))
 876			eth_test->flags |= ETH_TEST_FL_FAILED;
 877
 878		if (i40e_intr_test(netdev, &data[I40E_ETH_TEST_INTR]))
 879			eth_test->flags |= ETH_TEST_FL_FAILED;
 880
 881		if (i40e_loopback_test(netdev, &data[I40E_ETH_TEST_LOOPBACK]))
 882			eth_test->flags |= ETH_TEST_FL_FAILED;
 883
 884		/* run reg test last, a reset is required after it */
 885		if (i40e_reg_test(netdev, &data[I40E_ETH_TEST_REG]))
 886			eth_test->flags |= ETH_TEST_FL_FAILED;
 887
 888		clear_bit(__I40E_TESTING, &pf->state);
 889		i40e_do_reset(pf, (1 << __I40E_PF_RESET_REQUESTED));
 890	} else {
 891		/* Online tests */
 892		netif_info(pf, drv, netdev, "online testing starting\n");
 893
 894		if (i40e_link_test(netdev, &data[I40E_ETH_TEST_LINK]))
 895			eth_test->flags |= ETH_TEST_FL_FAILED;
 896
 897		/* Offline only tests, not run in online; pass by default */
 898		data[I40E_ETH_TEST_REG] = 0;
 899		data[I40E_ETH_TEST_EEPROM] = 0;
 900		data[I40E_ETH_TEST_INTR] = 0;
 901		data[I40E_ETH_TEST_LOOPBACK] = 0;
 902	}
 903
 904	netif_info(pf, drv, netdev, "testing finished\n");
 905}
 906
 907static void i40e_get_wol(struct net_device *netdev,
 908			 struct ethtool_wolinfo *wol)
 909{
 910	struct i40e_netdev_priv *np = netdev_priv(netdev);
 911	struct i40e_pf *pf = np->vsi->back;
 912	struct i40e_hw *hw = &pf->hw;
 913	u16 wol_nvm_bits;
 914
 915	/* NVM bit on means WoL disabled for the port */
 916	i40e_read_nvm_word(hw, I40E_SR_NVM_WAKE_ON_LAN, &wol_nvm_bits);
 917	if ((1 << hw->port) & wol_nvm_bits) {
 918		wol->supported = 0;
 919		wol->wolopts = 0;
 920	} else {
 921		wol->supported = WAKE_MAGIC;
 922		wol->wolopts = (pf->wol_en ? WAKE_MAGIC : 0);
 923	}
 924}
 925
 926static int i40e_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
 927{
 928	struct i40e_netdev_priv *np = netdev_priv(netdev);
 929	struct i40e_pf *pf = np->vsi->back;
 930	struct i40e_hw *hw = &pf->hw;
 931	u16 wol_nvm_bits;
 932
 933	/* NVM bit on means WoL disabled for the port */
 934	i40e_read_nvm_word(hw, I40E_SR_NVM_WAKE_ON_LAN, &wol_nvm_bits);
 935	if (((1 << hw->port) & wol_nvm_bits))
 936		return -EOPNOTSUPP;
 937
 938	/* only magic packet is supported */
 939	if (wol->wolopts && (wol->wolopts != WAKE_MAGIC))
 940		return -EOPNOTSUPP;
 941
 942	/* is this a new value? */
 943	if (pf->wol_en != !!wol->wolopts) {
 944		pf->wol_en = !!wol->wolopts;
 945		device_set_wakeup_enable(&pf->pdev->dev, pf->wol_en);
 946	}
 947
 948	return 0;
 949}
 950
 951static int i40e_nway_reset(struct net_device *netdev)
 952{
 953	/* restart autonegotiation */
 954	struct i40e_netdev_priv *np = netdev_priv(netdev);
 955	struct i40e_pf *pf = np->vsi->back;
 956	struct i40e_hw *hw = &pf->hw;
 957	i40e_status ret = 0;
 958
 959	ret = i40e_aq_set_link_restart_an(hw, NULL);
 960	if (ret) {
 961		netdev_info(netdev, "link restart failed, aq_err=%d\n",
 962			    pf->hw.aq.asq_last_status);
 963		return -EIO;
 964	}
 965
 966	return 0;
 967}
 968
 969static int i40e_set_phys_id(struct net_device *netdev,
 970			    enum ethtool_phys_id_state state)
 971{
 972	struct i40e_netdev_priv *np = netdev_priv(netdev);
 973	struct i40e_pf *pf = np->vsi->back;
 974	struct i40e_hw *hw = &pf->hw;
 975	int blink_freq = 2;
 976
 977	switch (state) {
 978	case ETHTOOL_ID_ACTIVE:
 979		pf->led_status = i40e_led_get(hw);
 980		return blink_freq;
 981	case ETHTOOL_ID_ON:
 982		i40e_led_set(hw, 0xF, false);
 983		break;
 984	case ETHTOOL_ID_OFF:
 985		i40e_led_set(hw, 0x0, false);
 986		break;
 987	case ETHTOOL_ID_INACTIVE:
 988		i40e_led_set(hw, pf->led_status, false);
 989		break;
 990	}
 991
 992	return 0;
 993}
 994
 995/* NOTE: i40e hardware uses a conversion factor of 2 for Interrupt
 996 * Throttle Rate (ITR) ie. ITR(1) = 2us ITR(10) = 20 us, and also
 997 * 125us (8000 interrupts per second) == ITR(62)
 998 */
 999
1000static int i40e_get_coalesce(struct net_device *netdev,
1001			     struct ethtool_coalesce *ec)
1002{
1003	struct i40e_netdev_priv *np = netdev_priv(netdev);
1004	struct i40e_vsi *vsi = np->vsi;
1005
1006	ec->tx_max_coalesced_frames_irq = vsi->work_limit;
1007	ec->rx_max_coalesced_frames_irq = vsi->work_limit;
1008
1009	if (ITR_IS_DYNAMIC(vsi->rx_itr_setting))
1010		ec->rx_coalesce_usecs = 1;
1011	else
1012		ec->rx_coalesce_usecs = vsi->rx_itr_setting;
1013
1014	if (ITR_IS_DYNAMIC(vsi->tx_itr_setting))
1015		ec->tx_coalesce_usecs = 1;
1016	else
1017		ec->tx_coalesce_usecs = vsi->tx_itr_setting;
1018
1019	return 0;
1020}
1021
1022static int i40e_set_coalesce(struct net_device *netdev,
1023			     struct ethtool_coalesce *ec)
1024{
1025	struct i40e_netdev_priv *np = netdev_priv(netdev);
1026	struct i40e_q_vector *q_vector;
1027	struct i40e_vsi *vsi = np->vsi;
1028	struct i40e_pf *pf = vsi->back;
1029	struct i40e_hw *hw = &pf->hw;
1030	u16 vector;
1031	int i;
1032
1033	if (ec->tx_max_coalesced_frames_irq || ec->rx_max_coalesced_frames_irq)
1034		vsi->work_limit = ec->tx_max_coalesced_frames_irq;
1035
1036	switch (ec->rx_coalesce_usecs) {
1037	case 0:
1038		vsi->rx_itr_setting = 0;
1039		break;
1040	case 1:
1041		vsi->rx_itr_setting = (I40E_ITR_DYNAMIC |
1042				       ITR_REG_TO_USEC(I40E_ITR_RX_DEF));
1043		break;
1044	default:
1045		if ((ec->rx_coalesce_usecs < (I40E_MIN_ITR << 1)) ||
1046		    (ec->rx_coalesce_usecs > (I40E_MAX_ITR << 1)))
1047			return -EINVAL;
1048		vsi->rx_itr_setting = ec->rx_coalesce_usecs;
1049		break;
1050	}
1051
1052	switch (ec->tx_coalesce_usecs) {
1053	case 0:
1054		vsi->tx_itr_setting = 0;
1055		break;
1056	case 1:
1057		vsi->tx_itr_setting = (I40E_ITR_DYNAMIC |
1058				       ITR_REG_TO_USEC(I40E_ITR_TX_DEF));
1059		break;
1060	default:
1061		if ((ec->tx_coalesce_usecs < (I40E_MIN_ITR << 1)) ||
1062		    (ec->tx_coalesce_usecs > (I40E_MAX_ITR << 1)))
1063			return -EINVAL;
1064		vsi->tx_itr_setting = ec->tx_coalesce_usecs;
1065		break;
1066	}
1067
1068	vector = vsi->base_vector;
1069	for (i = 0; i < vsi->num_q_vectors; i++, vector++) {
1070		q_vector = vsi->q_vectors[i];
1071		q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting);
1072		wr32(hw, I40E_PFINT_ITRN(0, vector - 1), q_vector->rx.itr);
1073		q_vector->tx.itr = ITR_TO_REG(vsi->tx_itr_setting);
1074		wr32(hw, I40E_PFINT_ITRN(1, vector - 1), q_vector->tx.itr);
1075		i40e_flush(hw);
1076	}
1077
1078	return 0;
1079}
1080
1081/**
1082 * i40e_get_rss_hash_opts - Get RSS hash Input Set for each flow type
1083 * @pf: pointer to the physical function struct
1084 * @cmd: ethtool rxnfc command
1085 *
1086 * Returns Success if the flow is supported, else Invalid Input.
1087 **/
1088static int i40e_get_rss_hash_opts(struct i40e_pf *pf, struct ethtool_rxnfc *cmd)
1089{
1090	cmd->data = 0;
1091
1092	/* Report default options for RSS on i40e */
1093	switch (cmd->flow_type) {
1094	case TCP_V4_FLOW:
1095	case UDP_V4_FLOW:
1096		cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
1097	/* fall through to add IP fields */
1098	case SCTP_V4_FLOW:
1099	case AH_ESP_V4_FLOW:
1100	case AH_V4_FLOW:
1101	case ESP_V4_FLOW:
1102	case IPV4_FLOW:
1103		cmd->data |= RXH_IP_SRC | RXH_IP_DST;
1104		break;
1105	case TCP_V6_FLOW:
1106	case UDP_V6_FLOW:
1107		cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
1108	/* fall through to add IP fields */
1109	case SCTP_V6_FLOW:
1110	case AH_ESP_V6_FLOW:
1111	case AH_V6_FLOW:
1112	case ESP_V6_FLOW:
1113	case IPV6_FLOW:
1114		cmd->data |= RXH_IP_SRC | RXH_IP_DST;
1115		break;
1116	default:
1117		return -EINVAL;
1118	}
1119
1120	return 0;
1121}
1122
1123/**
1124 * i40e_get_ethtool_fdir_all - Populates the rule count of a command
1125 * @pf: Pointer to the physical function struct
1126 * @cmd: The command to get or set Rx flow classification rules
1127 * @rule_locs: Array of used rule locations
1128 *
1129 * This function populates both the total and actual rule count of
1130 * the ethtool flow classification command
1131 *
1132 * Returns 0 on success or -EMSGSIZE if entry not found
1133 **/
1134static int i40e_get_ethtool_fdir_all(struct i40e_pf *pf,
1135				     struct ethtool_rxnfc *cmd,
1136				     u32 *rule_locs)
1137{
1138	struct i40e_fdir_filter *rule;
1139	struct hlist_node *node2;
1140	int cnt = 0;
1141
1142	/* report total rule count */
1143	cmd->data = pf->hw.fdir_shared_filter_count +
1144		    pf->fdir_pf_filter_count;
1145
1146	hlist_for_each_entry_safe(rule, node2,
1147				  &pf->fdir_filter_list, fdir_node) {
1148		if (cnt == cmd->rule_cnt)
1149			return -EMSGSIZE;
1150
1151		rule_locs[cnt] = rule->fd_id;
1152		cnt++;
1153	}
1154
1155	cmd->rule_cnt = cnt;
1156
1157	return 0;
1158}
1159
1160/**
1161 * i40e_get_ethtool_fdir_entry - Look up a filter based on Rx flow
1162 * @pf: Pointer to the physical function struct
1163 * @cmd: The command to get or set Rx flow classification rules
1164 *
1165 * This function looks up a filter based on the Rx flow classification
1166 * command and fills the flow spec info for it if found
1167 *
1168 * Returns 0 on success or -EINVAL if filter not found
1169 **/
1170static int i40e_get_ethtool_fdir_entry(struct i40e_pf *pf,
1171				       struct ethtool_rxnfc *cmd)
1172{
1173	struct ethtool_rx_flow_spec *fsp =
1174			(struct ethtool_rx_flow_spec *)&cmd->fs;
1175	struct i40e_fdir_filter *rule = NULL;
1176	struct hlist_node *node2;
1177
1178	/* report total rule count */
1179	cmd->data = pf->hw.fdir_shared_filter_count +
1180		    pf->fdir_pf_filter_count;
1181
1182	hlist_for_each_entry_safe(rule, node2,
1183				  &pf->fdir_filter_list, fdir_node) {
1184		if (fsp->location <= rule->fd_id)
1185			break;
1186	}
1187
1188	if (!rule || fsp->location != rule->fd_id)
1189		return -EINVAL;
1190
1191	fsp->flow_type = rule->flow_type;
1192	fsp->h_u.tcp_ip4_spec.psrc = rule->src_port;
1193	fsp->h_u.tcp_ip4_spec.pdst = rule->dst_port;
1194	fsp->h_u.tcp_ip4_spec.ip4src = rule->src_ip[0];
1195	fsp->h_u.tcp_ip4_spec.ip4dst = rule->dst_ip[0];
1196	fsp->ring_cookie = rule->q_index;
1197
1198	return 0;
1199}
1200
1201/**
1202 * i40e_get_rxnfc - command to get RX flow classification rules
1203 * @netdev: network interface device structure
1204 * @cmd: ethtool rxnfc command
1205 *
1206 * Returns Success if the command is supported.
1207 **/
1208static int i40e_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd,
1209			  u32 *rule_locs)
1210{
1211	struct i40e_netdev_priv *np = netdev_priv(netdev);
1212	struct i40e_vsi *vsi = np->vsi;
1213	struct i40e_pf *pf = vsi->back;
1214	int ret = -EOPNOTSUPP;
1215
1216	switch (cmd->cmd) {
1217	case ETHTOOL_GRXRINGS:
1218		cmd->data = vsi->alloc_queue_pairs;
1219		ret = 0;
1220		break;
1221	case ETHTOOL_GRXFH:
1222		ret = i40e_get_rss_hash_opts(pf, cmd);
1223		break;
1224	case ETHTOOL_GRXCLSRLCNT:
1225		cmd->rule_cnt = pf->fdir_pf_active_filters;
1226		ret = 0;
1227		break;
1228	case ETHTOOL_GRXCLSRULE:
1229		ret = i40e_get_ethtool_fdir_entry(pf, cmd);
1230		break;
1231	case ETHTOOL_GRXCLSRLALL:
1232		ret = i40e_get_ethtool_fdir_all(pf, cmd, rule_locs);
1233		break;
1234	default:
1235		break;
1236	}
1237
1238	return ret;
1239}
1240
1241/**
1242 * i40e_set_rss_hash_opt - Enable/Disable flow types for RSS hash
1243 * @pf: pointer to the physical function struct
1244 * @cmd: ethtool rxnfc command
1245 *
1246 * Returns Success if the flow input set is supported.
1247 **/
1248static int i40e_set_rss_hash_opt(struct i40e_pf *pf, struct ethtool_rxnfc *nfc)
1249{
1250	struct i40e_hw *hw = &pf->hw;
1251	u64 hena = (u64)rd32(hw, I40E_PFQF_HENA(0)) |
1252		   ((u64)rd32(hw, I40E_PFQF_HENA(1)) << 32);
1253
1254	/* RSS does not support anything other than hashing
1255	 * to queues on src and dst IPs and ports
1256	 */
1257	if (nfc->data & ~(RXH_IP_SRC | RXH_IP_DST |
1258			  RXH_L4_B_0_1 | RXH_L4_B_2_3))
1259		return -EINVAL;
1260
1261	/* We need at least the IP SRC and DEST fields for hashing */
1262	if (!(nfc->data & RXH_IP_SRC) ||
1263	    !(nfc->data & RXH_IP_DST))
1264		return -EINVAL;
1265
1266	switch (nfc->flow_type) {
1267	case TCP_V4_FLOW:
1268		switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
1269		case 0:
1270			hena &= ~((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV4_TCP);
1271			break;
1272		case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
1273			hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV4_TCP);
1274			break;
1275		default:
1276			return -EINVAL;
1277		}
1278		break;
1279	case TCP_V6_FLOW:
1280		switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
1281		case 0:
1282			hena &= ~((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV6_TCP);
1283			break;
1284		case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
1285			hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV6_TCP);
1286			break;
1287		default:
1288			return -EINVAL;
1289		}
1290		break;
1291	case UDP_V4_FLOW:
1292		switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
1293		case 0:
1294			hena &=
1295			~(((u64)1 << I40E_FILTER_PCTYPE_NONF_UNICAST_IPV4_UDP) |
1296			((u64)1 << I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV4_UDP) |
1297			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV4));
1298			break;
1299		case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
1300			hena |=
1301			(((u64)1 << I40E_FILTER_PCTYPE_NONF_UNICAST_IPV4_UDP)  |
1302			((u64)1 << I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV4_UDP) |
1303			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV4));
1304			break;
1305		default:
1306			return -EINVAL;
1307		}
1308		break;
1309	case UDP_V6_FLOW:
1310		switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
1311		case 0:
1312			hena &=
1313			~(((u64)1 << I40E_FILTER_PCTYPE_NONF_UNICAST_IPV6_UDP) |
1314			((u64)1 << I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV6_UDP) |
1315			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV6));
1316			break;
1317		case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
1318			hena |=
1319			(((u64)1 << I40E_FILTER_PCTYPE_NONF_UNICAST_IPV6_UDP)  |
1320			((u64)1 << I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV6_UDP) |
1321			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV6));
1322			break;
1323		default:
1324			return -EINVAL;
1325		}
1326		break;
1327	case AH_ESP_V4_FLOW:
1328	case AH_V4_FLOW:
1329	case ESP_V4_FLOW:
1330	case SCTP_V4_FLOW:
1331		if ((nfc->data & RXH_L4_B_0_1) ||
1332		    (nfc->data & RXH_L4_B_2_3))
1333			return -EINVAL;
1334		hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV4_OTHER);
1335		break;
1336	case AH_ESP_V6_FLOW:
1337	case AH_V6_FLOW:
1338	case ESP_V6_FLOW:
1339	case SCTP_V6_FLOW:
1340		if ((nfc->data & RXH_L4_B_0_1) ||
1341		    (nfc->data & RXH_L4_B_2_3))
1342			return -EINVAL;
1343		hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV6_OTHER);
1344		break;
1345	case IPV4_FLOW:
1346		hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV4_OTHER) |
1347			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV4);
1348		break;
1349	case IPV6_FLOW:
1350		hena |= ((u64)1 << I40E_FILTER_PCTYPE_NONF_IPV6_OTHER) |
1351			((u64)1 << I40E_FILTER_PCTYPE_FRAG_IPV6);
1352		break;
1353	default:
1354		return -EINVAL;
1355	}
1356
1357	wr32(hw, I40E_PFQF_HENA(0), (u32)hena);
1358	wr32(hw, I40E_PFQF_HENA(1), (u32)(hena >> 32));
1359	i40e_flush(hw);
1360
1361	return 0;
1362}
1363
1364/**
1365 * i40e_match_fdir_input_set - Match a new filter against an existing one
1366 * @rule: The filter already added
1367 * @input: The new filter to comapre against
1368 *
1369 * Returns true if the two input set match
1370 **/
1371static bool i40e_match_fdir_input_set(struct i40e_fdir_filter *rule,
1372				      struct i40e_fdir_filter *input)
1373{
1374	if ((rule->dst_ip[0] != input->dst_ip[0]) ||
1375	    (rule->src_ip[0] != input->src_ip[0]) ||
1376	    (rule->dst_port != input->dst_port) ||
1377	    (rule->src_port != input->src_port))
1378		return false;
1379	return true;
1380}
1381
1382/**
1383 * i40e_update_ethtool_fdir_entry - Updates the fdir filter entry
1384 * @vsi: Pointer to the targeted VSI
1385 * @input: The filter to update or NULL to indicate deletion
1386 * @sw_idx: Software index to the filter
1387 * @cmd: The command to get or set Rx flow classification rules
1388 *
1389 * This function updates (or deletes) a Flow Director entry from
1390 * the hlist of the corresponding PF
1391 *
1392 * Returns 0 on success
1393 **/
1394static int i40e_update_ethtool_fdir_entry(struct i40e_vsi *vsi,
1395					  struct i40e_fdir_filter *input,
1396					  u16 sw_idx,
1397					  struct ethtool_rxnfc *cmd)
1398{
1399	struct i40e_fdir_filter *rule, *parent;
1400	struct i40e_pf *pf = vsi->back;
1401	struct hlist_node *node2;
1402	int err = -EINVAL;
1403
1404	parent = NULL;
1405	rule = NULL;
1406
1407	hlist_for_each_entry_safe(rule, node2,
1408				  &pf->fdir_filter_list, fdir_node) {
1409		/* hash found, or no matching entry */
1410		if (rule->fd_id >= sw_idx)
1411			break;
1412		parent = rule;
1413	}
1414
1415	/* if there is an old rule occupying our place remove it */
1416	if (rule && (rule->fd_id == sw_idx)) {
1417		if (input && !i40e_match_fdir_input_set(rule, input))
1418			err = i40e_add_del_fdir(vsi, rule, false);
1419		else if (!input)
1420			err = i40e_add_del_fdir(vsi, rule, false);
1421		hlist_del(&rule->fdir_node);
1422		kfree(rule);
1423		pf->fdir_pf_active_filters--;
1424	}
1425
1426	/* If no input this was a delete, err should be 0 if a rule was
1427	 * successfully found and removed from the list else -EINVAL
1428	 */
1429	if (!input)
1430		return err;
1431
1432	/* initialize node and set software index */
1433	INIT_HLIST_NODE(&input->fdir_node);
1434
1435	/* add filter to the list */
1436	if (parent)
1437		hlist_add_after(&parent->fdir_node, &input->fdir_node);
1438	else
1439		hlist_add_head(&input->fdir_node,
1440			       &pf->fdir_filter_list);
1441
1442	/* update counts */
1443	pf->fdir_pf_active_filters++;
1444
1445	return 0;
1446}
1447
1448/**
1449 * i40e_del_fdir_entry - Deletes a Flow Director filter entry
1450 * @vsi: Pointer to the targeted VSI
1451 * @cmd: The command to get or set Rx flow classification rules
1452 *
1453 * The function removes a Flow Director filter entry from the
1454 * hlist of the corresponding PF
1455 *
1456 * Returns 0 on success
1457 */
1458static int i40e_del_fdir_entry(struct i40e_vsi *vsi,
1459			       struct ethtool_rxnfc *cmd)
1460{
1461	struct ethtool_rx_flow_spec *fsp =
1462		(struct ethtool_rx_flow_spec *)&cmd->fs;
1463	struct i40e_pf *pf = vsi->back;
1464	int ret = 0;
1465
1466	ret = i40e_update_ethtool_fdir_entry(vsi, NULL, fsp->location, cmd);
1467
1468	i40e_fdir_check_and_reenable(pf);
1469	return ret;
1470}
1471
1472/**
1473 * i40e_add_fdir_ethtool - Add/Remove Flow Director filters
1474 * @vsi: pointer to the targeted VSI
1475 * @cmd: command to get or set RX flow classification rules
1476 *
1477 * Add Flow Director filters for a specific flow spec based on their
1478 * protocol.  Returns 0 if the filters were successfully added.
1479 **/
1480static int i40e_add_fdir_ethtool(struct i40e_vsi *vsi,
1481				 struct ethtool_rxnfc *cmd)
1482{
1483	struct ethtool_rx_flow_spec *fsp;
1484	struct i40e_fdir_filter *input;
1485	struct i40e_pf *pf;
1486	int ret = -EINVAL;
1487
1488	if (!vsi)
1489		return -EINVAL;
1490
1491	pf = vsi->back;
1492
1493	if (!(pf->flags & I40E_FLAG_FD_SB_ENABLED))
1494		return -EOPNOTSUPP;
1495
1496	if (pf->auto_disable_flags & I40E_FLAG_FD_SB_ENABLED)
1497		return -ENOSPC;
1498
1499	fsp = (struct ethtool_rx_flow_spec *)&cmd->fs;
1500
1501	if (fsp->location >= (pf->hw.func_caps.fd_filters_best_effort +
1502			      pf->hw.func_caps.fd_filters_guaranteed)) {
1503		return -EINVAL;
1504	}
1505
1506	if (fsp->ring_cookie >= vsi->num_queue_pairs)
1507		return -EINVAL;
1508
1509	input = kzalloc(sizeof(*input), GFP_KERNEL);
1510
1511	if (!input)
1512		return -ENOMEM;
1513
1514	input->fd_id = fsp->location;
1515
1516	if (fsp->ring_cookie == RX_CLS_FLOW_DISC)
1517		input->dest_ctl = I40E_FILTER_PROGRAM_DESC_DEST_DROP_PACKET;
1518	else
1519		input->dest_ctl =
1520			     I40E_FILTER_PROGRAM_DESC_DEST_DIRECT_PACKET_QINDEX;
1521
1522	input->q_index = fsp->ring_cookie;
1523	input->flex_off = 0;
1524	input->pctype = 0;
1525	input->dest_vsi = vsi->id;
1526	input->fd_status = I40E_FILTER_PROGRAM_DESC_FD_STATUS_FD_ID;
1527	input->cnt_index = 0;
1528	input->flow_type = fsp->flow_type;
1529	input->ip4_proto = fsp->h_u.usr_ip4_spec.proto;
1530	input->src_port = fsp->h_u.tcp_ip4_spec.psrc;
1531	input->dst_port = fsp->h_u.tcp_ip4_spec.pdst;
1532	input->src_ip[0] = fsp->h_u.tcp_ip4_spec.ip4src;
1533	input->dst_ip[0] = fsp->h_u.tcp_ip4_spec.ip4dst;
1534
1535	ret = i40e_add_del_fdir(vsi, input, true);
1536	if (ret)
1537		kfree(input);
1538	else
1539		i40e_update_ethtool_fdir_entry(vsi, input, fsp->location, NULL);
1540
1541	return ret;
1542}
1543
1544/**
1545 * i40e_set_rxnfc - command to set RX flow classification rules
1546 * @netdev: network interface device structure
1547 * @cmd: ethtool rxnfc command
1548 *
1549 * Returns Success if the command is supported.
1550 **/
1551static int i40e_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *cmd)
1552{
1553	struct i40e_netdev_priv *np = netdev_priv(netdev);
1554	struct i40e_vsi *vsi = np->vsi;
1555	struct i40e_pf *pf = vsi->back;
1556	int ret = -EOPNOTSUPP;
1557
1558	switch (cmd->cmd) {
1559	case ETHTOOL_SRXFH:
1560		ret = i40e_set_rss_hash_opt(pf, cmd);
1561		break;
1562	case ETHTOOL_SRXCLSRLINS:
1563		ret = i40e_add_fdir_ethtool(vsi, cmd);
1564		break;
1565	case ETHTOOL_SRXCLSRLDEL:
1566		ret = i40e_del_fdir_entry(vsi, cmd);
1567		break;
1568	default:
1569		break;
1570	}
1571
1572	return ret;
1573}
1574
1575/**
1576 * i40e_max_channels - get Max number of combined channels supported
1577 * @vsi: vsi pointer
1578 **/
1579static unsigned int i40e_max_channels(struct i40e_vsi *vsi)
1580{
1581	/* TODO: This code assumes DCB and FD is disabled for now. */
1582	return vsi->alloc_queue_pairs;
1583}
1584
1585/**
1586 * i40e_get_channels - Get the current channels enabled and max supported etc.
1587 * @netdev: network interface device structure
1588 * @ch: ethtool channels structure
1589 *
1590 * We don't support separate tx and rx queues as channels. The other count
1591 * represents how many queues are being used for control. max_combined counts
1592 * how many queue pairs we can support. They may not be mapped 1 to 1 with
1593 * q_vectors since we support a lot more queue pairs than q_vectors.
1594 **/
1595static void i40e_get_channels(struct net_device *dev,
1596			       struct ethtool_channels *ch)
1597{
1598	struct i40e_netdev_priv *np = netdev_priv(dev);
1599	struct i40e_vsi *vsi = np->vsi;
1600	struct i40e_pf *pf = vsi->back;
1601
1602	/* report maximum channels */
1603	ch->max_combined = i40e_max_channels(vsi);
1604
1605	/* report info for other vector */
1606	ch->other_count = (pf->flags & I40E_FLAG_FD_SB_ENABLED) ? 1 : 0;
1607	ch->max_other = ch->other_count;
1608
1609	/* Note: This code assumes DCB is disabled for now. */
1610	ch->combined_count = vsi->num_queue_pairs;
1611}
1612
1613/**
1614 * i40e_set_channels - Set the new channels count.
1615 * @netdev: network interface device structure
1616 * @ch: ethtool channels structure
1617 *
1618 * The new channels count may not be the same as requested by the user
1619 * since it gets rounded down to a power of 2 value.
1620 **/
1621static int i40e_set_channels(struct net_device *dev,
1622			      struct ethtool_channels *ch)
1623{
1624	struct i40e_netdev_priv *np = netdev_priv(dev);
1625	unsigned int count = ch->combined_count;
1626	struct i40e_vsi *vsi = np->vsi;
1627	struct i40e_pf *pf = vsi->back;
1628	int new_count;
1629
1630	/* We do not support setting channels for any other VSI at present */
1631	if (vsi->type != I40E_VSI_MAIN)
1632		return -EINVAL;
1633
1634	/* verify they are not requesting separate vectors */
1635	if (!count || ch->rx_count || ch->tx_count)
1636		return -EINVAL;
1637
1638	/* verify other_count has not changed */
1639	if (ch->other_count != ((pf->flags & I40E_FLAG_FD_SB_ENABLED) ? 1 : 0))
1640		return -EINVAL;
1641
1642	/* verify the number of channels does not exceed hardware limits */
1643	if (count > i40e_max_channels(vsi))
1644		return -EINVAL;
1645
1646	/* update feature limits from largest to smallest supported values */
1647	/* TODO: Flow director limit, DCB etc */
1648
1649	/* cap RSS limit */
1650	if (count > pf->rss_size_max)
1651		count = pf->rss_size_max;
1652
1653	/* use rss_reconfig to rebuild with new queue count and update traffic
1654	 * class queue mapping
1655	 */
1656	new_count = i40e_reconfig_rss_queues(pf, count);
1657	if (new_count > 0)
1658		return 0;
1659	else
1660		return -EINVAL;
1661}
1662
1663static const struct ethtool_ops i40e_ethtool_ops = {
1664	.get_settings		= i40e_get_settings,
1665	.get_drvinfo		= i40e_get_drvinfo,
1666	.get_regs_len		= i40e_get_regs_len,
1667	.get_regs		= i40e_get_regs,
1668	.nway_reset		= i40e_nway_reset,
1669	.get_link		= ethtool_op_get_link,
1670	.get_wol		= i40e_get_wol,
1671	.set_wol		= i40e_set_wol,
1672	.get_eeprom_len		= i40e_get_eeprom_len,
1673	.get_eeprom		= i40e_get_eeprom,
1674	.get_ringparam		= i40e_get_ringparam,
1675	.set_ringparam		= i40e_set_ringparam,
1676	.get_pauseparam		= i40e_get_pauseparam,
1677	.get_msglevel		= i40e_get_msglevel,
1678	.set_msglevel		= i40e_set_msglevel,
1679	.get_rxnfc		= i40e_get_rxnfc,
1680	.set_rxnfc		= i40e_set_rxnfc,
1681	.self_test		= i40e_diag_test,
1682	.get_strings		= i40e_get_strings,
1683	.set_phys_id		= i40e_set_phys_id,
1684	.get_sset_count		= i40e_get_sset_count,
1685	.get_ethtool_stats	= i40e_get_ethtool_stats,
1686	.get_coalesce		= i40e_get_coalesce,
1687	.set_coalesce		= i40e_set_coalesce,
1688	.get_channels		= i40e_get_channels,
1689	.set_channels		= i40e_set_channels,
1690	.get_ts_info		= i40e_get_ts_info,
1691};
1692
1693void i40e_set_ethtool_ops(struct net_device *netdev)
1694{
1695	SET_ETHTOOL_OPS(netdev, &i40e_ethtool_ops);
1696}