Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.15.
   1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause)
   2/* QLogic qed NIC Driver
   3 * Copyright (c) 2015-2017  QLogic Corporation
   4 * Copyright (c) 2019-2020 Marvell International Ltd.
   5 */
   6
   7#include <linux/stddef.h>
   8#include <linux/pci.h>
   9#include <linux/kernel.h>
  10#include <linux/slab.h>
  11#include <linux/delay.h>
  12#include <asm/byteorder.h>
  13#include <linux/dma-mapping.h>
  14#include <linux/string.h>
  15#include <linux/module.h>
  16#include <linux/interrupt.h>
  17#include <linux/workqueue.h>
  18#include <linux/ethtool.h>
  19#include <linux/etherdevice.h>
  20#include <linux/vmalloc.h>
  21#include <linux/crash_dump.h>
  22#include <linux/crc32.h>
  23#include <linux/qed/qed_if.h>
  24#include <linux/qed/qed_ll2_if.h>
  25#include <net/devlink.h>
  26#include <linux/phylink.h>
  27
  28#include "qed.h"
  29#include "qed_sriov.h"
  30#include "qed_sp.h"
  31#include "qed_dev_api.h"
  32#include "qed_ll2.h"
  33#include "qed_fcoe.h"
  34#include "qed_iscsi.h"
  35
  36#include "qed_mcp.h"
  37#include "qed_reg_addr.h"
  38#include "qed_hw.h"
  39#include "qed_selftest.h"
  40#include "qed_debug.h"
  41#include "qed_devlink.h"
  42
  43#define QED_ROCE_QPS			(8192)
  44#define QED_ROCE_DPIS			(8)
  45#define QED_RDMA_SRQS                   QED_ROCE_QPS
  46#define QED_NVM_CFG_GET_FLAGS		0xA
  47#define QED_NVM_CFG_GET_PF_FLAGS	0x1A
  48#define QED_NVM_CFG_MAX_ATTRS		50
  49
  50static char version[] =
  51	"QLogic FastLinQ 4xxxx Core Module qed\n";
  52
  53MODULE_DESCRIPTION("QLogic FastLinQ 4xxxx Core Module");
  54MODULE_LICENSE("GPL");
  55
  56#define FW_FILE_VERSION				\
  57	__stringify(FW_MAJOR_VERSION) "."	\
  58	__stringify(FW_MINOR_VERSION) "."	\
  59	__stringify(FW_REVISION_VERSION) "."	\
  60	__stringify(FW_ENGINEERING_VERSION)
  61
  62#define QED_FW_FILE_NAME	\
  63	"qed/qed_init_values_zipped-" FW_FILE_VERSION ".bin"
  64
  65MODULE_FIRMWARE(QED_FW_FILE_NAME);
  66
  67/* MFW speed capabilities maps */
  68
  69struct qed_mfw_speed_map {
  70	u32		mfw_val;
  71	__ETHTOOL_DECLARE_LINK_MODE_MASK(caps);
  72
  73	const u32	*cap_arr;
  74	u32		arr_size;
  75};
  76
  77#define QED_MFW_SPEED_MAP(type, arr)		\
  78{						\
  79	.mfw_val	= (type),		\
  80	.cap_arr	= (arr),		\
  81	.arr_size	= ARRAY_SIZE(arr),	\
  82}
  83
  84static const u32 qed_mfw_ext_1g[] __initconst = {
  85	ETHTOOL_LINK_MODE_1000baseT_Full_BIT,
  86	ETHTOOL_LINK_MODE_1000baseKX_Full_BIT,
  87	ETHTOOL_LINK_MODE_1000baseX_Full_BIT,
  88};
  89
  90static const u32 qed_mfw_ext_10g[] __initconst = {
  91	ETHTOOL_LINK_MODE_10000baseT_Full_BIT,
  92	ETHTOOL_LINK_MODE_10000baseKR_Full_BIT,
  93	ETHTOOL_LINK_MODE_10000baseKX4_Full_BIT,
  94	ETHTOOL_LINK_MODE_10000baseR_FEC_BIT,
  95	ETHTOOL_LINK_MODE_10000baseCR_Full_BIT,
  96	ETHTOOL_LINK_MODE_10000baseSR_Full_BIT,
  97	ETHTOOL_LINK_MODE_10000baseLR_Full_BIT,
  98	ETHTOOL_LINK_MODE_10000baseLRM_Full_BIT,
  99};
 100
 101static const u32 qed_mfw_ext_25g[] __initconst = {
 102	ETHTOOL_LINK_MODE_25000baseKR_Full_BIT,
 103	ETHTOOL_LINK_MODE_25000baseCR_Full_BIT,
 104	ETHTOOL_LINK_MODE_25000baseSR_Full_BIT,
 105};
 106
 107static const u32 qed_mfw_ext_40g[] __initconst = {
 108	ETHTOOL_LINK_MODE_40000baseLR4_Full_BIT,
 109	ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT,
 110	ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT,
 111	ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT,
 112};
 113
 114static const u32 qed_mfw_ext_50g_base_r[] __initconst = {
 115	ETHTOOL_LINK_MODE_50000baseKR_Full_BIT,
 116	ETHTOOL_LINK_MODE_50000baseCR_Full_BIT,
 117	ETHTOOL_LINK_MODE_50000baseSR_Full_BIT,
 118	ETHTOOL_LINK_MODE_50000baseLR_ER_FR_Full_BIT,
 119	ETHTOOL_LINK_MODE_50000baseDR_Full_BIT,
 120};
 121
 122static const u32 qed_mfw_ext_50g_base_r2[] __initconst = {
 123	ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT,
 124	ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT,
 125	ETHTOOL_LINK_MODE_50000baseSR2_Full_BIT,
 126};
 127
 128static const u32 qed_mfw_ext_100g_base_r2[] __initconst = {
 129	ETHTOOL_LINK_MODE_100000baseKR2_Full_BIT,
 130	ETHTOOL_LINK_MODE_100000baseSR2_Full_BIT,
 131	ETHTOOL_LINK_MODE_100000baseCR2_Full_BIT,
 132	ETHTOOL_LINK_MODE_100000baseDR2_Full_BIT,
 133	ETHTOOL_LINK_MODE_100000baseLR2_ER2_FR2_Full_BIT,
 134};
 135
 136static const u32 qed_mfw_ext_100g_base_r4[] __initconst = {
 137	ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT,
 138	ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT,
 139	ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT,
 140	ETHTOOL_LINK_MODE_100000baseLR4_ER4_Full_BIT,
 141};
 142
 143static struct qed_mfw_speed_map qed_mfw_ext_maps[] __ro_after_init = {
 144	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_1G, qed_mfw_ext_1g),
 145	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_10G, qed_mfw_ext_10g),
 146	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_25G, qed_mfw_ext_25g),
 147	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_40G, qed_mfw_ext_40g),
 148	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_50G_BASE_R,
 149			  qed_mfw_ext_50g_base_r),
 150	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_50G_BASE_R2,
 151			  qed_mfw_ext_50g_base_r2),
 152	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_100G_BASE_R2,
 153			  qed_mfw_ext_100g_base_r2),
 154	QED_MFW_SPEED_MAP(ETH_EXT_ADV_SPEED_100G_BASE_R4,
 155			  qed_mfw_ext_100g_base_r4),
 156};
 157
 158static const u32 qed_mfw_legacy_1g[] __initconst = {
 159	ETHTOOL_LINK_MODE_1000baseT_Full_BIT,
 160	ETHTOOL_LINK_MODE_1000baseKX_Full_BIT,
 161	ETHTOOL_LINK_MODE_1000baseX_Full_BIT,
 162};
 163
 164static const u32 qed_mfw_legacy_10g[] __initconst = {
 165	ETHTOOL_LINK_MODE_10000baseT_Full_BIT,
 166	ETHTOOL_LINK_MODE_10000baseKR_Full_BIT,
 167	ETHTOOL_LINK_MODE_10000baseKX4_Full_BIT,
 168	ETHTOOL_LINK_MODE_10000baseR_FEC_BIT,
 169	ETHTOOL_LINK_MODE_10000baseCR_Full_BIT,
 170	ETHTOOL_LINK_MODE_10000baseSR_Full_BIT,
 171	ETHTOOL_LINK_MODE_10000baseLR_Full_BIT,
 172	ETHTOOL_LINK_MODE_10000baseLRM_Full_BIT,
 173};
 174
 175static const u32 qed_mfw_legacy_20g[] __initconst = {
 176	ETHTOOL_LINK_MODE_20000baseKR2_Full_BIT,
 177};
 178
 179static const u32 qed_mfw_legacy_25g[] __initconst = {
 180	ETHTOOL_LINK_MODE_25000baseKR_Full_BIT,
 181	ETHTOOL_LINK_MODE_25000baseCR_Full_BIT,
 182	ETHTOOL_LINK_MODE_25000baseSR_Full_BIT,
 183};
 184
 185static const u32 qed_mfw_legacy_40g[] __initconst = {
 186	ETHTOOL_LINK_MODE_40000baseLR4_Full_BIT,
 187	ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT,
 188	ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT,
 189	ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT,
 190};
 191
 192static const u32 qed_mfw_legacy_50g[] __initconst = {
 193	ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT,
 194	ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT,
 195	ETHTOOL_LINK_MODE_50000baseSR2_Full_BIT,
 196};
 197
 198static const u32 qed_mfw_legacy_bb_100g[] __initconst = {
 199	ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT,
 200	ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT,
 201	ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT,
 202	ETHTOOL_LINK_MODE_100000baseLR4_ER4_Full_BIT,
 203};
 204
 205static struct qed_mfw_speed_map qed_mfw_legacy_maps[] __ro_after_init = {
 206	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G,
 207			  qed_mfw_legacy_1g),
 208	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G,
 209			  qed_mfw_legacy_10g),
 210	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_20G,
 211			  qed_mfw_legacy_20g),
 212	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_25G,
 213			  qed_mfw_legacy_25g),
 214	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_40G,
 215			  qed_mfw_legacy_40g),
 216	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_50G,
 217			  qed_mfw_legacy_50g),
 218	QED_MFW_SPEED_MAP(NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_BB_100G,
 219			  qed_mfw_legacy_bb_100g),
 220};
 221
 222static void __init qed_mfw_speed_map_populate(struct qed_mfw_speed_map *map)
 223{
 224	linkmode_set_bit_array(map->cap_arr, map->arr_size, map->caps);
 225
 226	map->cap_arr = NULL;
 227	map->arr_size = 0;
 228}
 229
 230static void __init qed_mfw_speed_maps_init(void)
 231{
 232	u32 i;
 233
 234	for (i = 0; i < ARRAY_SIZE(qed_mfw_ext_maps); i++)
 235		qed_mfw_speed_map_populate(qed_mfw_ext_maps + i);
 236
 237	for (i = 0; i < ARRAY_SIZE(qed_mfw_legacy_maps); i++)
 238		qed_mfw_speed_map_populate(qed_mfw_legacy_maps + i);
 239}
 240
 241static int __init qed_init(void)
 242{
 243	pr_info("%s", version);
 244
 245	qed_mfw_speed_maps_init();
 246
 247	return 0;
 248}
 249module_init(qed_init);
 250
 251static void __exit qed_exit(void)
 252{
 253	/* To prevent marking this module as "permanent" */
 254}
 255module_exit(qed_exit);
 256
 257static void qed_free_pci(struct qed_dev *cdev)
 258{
 259	struct pci_dev *pdev = cdev->pdev;
 260
 261	if (cdev->doorbells && cdev->db_size)
 262		iounmap(cdev->doorbells);
 263	if (cdev->regview)
 264		iounmap(cdev->regview);
 265	if (atomic_read(&pdev->enable_cnt) == 1)
 266		pci_release_regions(pdev);
 267
 268	pci_disable_device(pdev);
 269}
 270
 271#define PCI_REVISION_ID_ERROR_VAL	0xff
 272
 273/* Performs PCI initializations as well as initializing PCI-related parameters
 274 * in the device structrue. Returns 0 in case of success.
 275 */
 276static int qed_init_pci(struct qed_dev *cdev, struct pci_dev *pdev)
 277{
 278	u8 rev_id;
 279	int rc;
 280
 281	cdev->pdev = pdev;
 282
 283	rc = pci_enable_device(pdev);
 284	if (rc) {
 285		DP_NOTICE(cdev, "Cannot enable PCI device\n");
 286		goto err0;
 287	}
 288
 289	if (!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)) {
 290		DP_NOTICE(cdev, "No memory region found in bar #0\n");
 291		rc = -EIO;
 292		goto err1;
 293	}
 294
 295	if (IS_PF(cdev) && !(pci_resource_flags(pdev, 2) & IORESOURCE_MEM)) {
 296		DP_NOTICE(cdev, "No memory region found in bar #2\n");
 297		rc = -EIO;
 298		goto err1;
 299	}
 300
 301	if (atomic_read(&pdev->enable_cnt) == 1) {
 302		rc = pci_request_regions(pdev, "qed");
 303		if (rc) {
 304			DP_NOTICE(cdev,
 305				  "Failed to request PCI memory resources\n");
 306			goto err1;
 307		}
 308		pci_set_master(pdev);
 309		pci_save_state(pdev);
 310	}
 311
 312	pci_read_config_byte(pdev, PCI_REVISION_ID, &rev_id);
 313	if (rev_id == PCI_REVISION_ID_ERROR_VAL) {
 314		DP_NOTICE(cdev,
 315			  "Detected PCI device error [rev_id 0x%x]. Probably due to prior indication. Aborting.\n",
 316			  rev_id);
 317		rc = -ENODEV;
 318		goto err2;
 319	}
 320	if (!pci_is_pcie(pdev)) {
 321		DP_NOTICE(cdev, "The bus is not PCI Express\n");
 322		rc = -EIO;
 323		goto err2;
 324	}
 325
 326	cdev->pci_params.pm_cap = pci_find_capability(pdev, PCI_CAP_ID_PM);
 327	if (IS_PF(cdev) && !cdev->pci_params.pm_cap)
 328		DP_NOTICE(cdev, "Cannot find power management capability\n");
 329
 330	rc = dma_set_mask_and_coherent(&cdev->pdev->dev, DMA_BIT_MASK(64));
 331	if (rc) {
 332		DP_NOTICE(cdev, "Can't request DMA addresses\n");
 333		rc = -EIO;
 334		goto err2;
 335	}
 336
 337	cdev->pci_params.mem_start = pci_resource_start(pdev, 0);
 338	cdev->pci_params.mem_end = pci_resource_end(pdev, 0);
 339	cdev->pci_params.irq = pdev->irq;
 340
 341	cdev->regview = pci_ioremap_bar(pdev, 0);
 342	if (!cdev->regview) {
 343		DP_NOTICE(cdev, "Cannot map register space, aborting\n");
 344		rc = -ENOMEM;
 345		goto err2;
 346	}
 347
 348	cdev->db_phys_addr = pci_resource_start(cdev->pdev, 2);
 349	cdev->db_size = pci_resource_len(cdev->pdev, 2);
 350	if (!cdev->db_size) {
 351		if (IS_PF(cdev)) {
 352			DP_NOTICE(cdev, "No Doorbell bar available\n");
 353			return -EINVAL;
 354		} else {
 355			return 0;
 356		}
 357	}
 358
 359	cdev->doorbells = ioremap_wc(cdev->db_phys_addr, cdev->db_size);
 360
 361	if (!cdev->doorbells) {
 362		DP_NOTICE(cdev, "Cannot map doorbell space\n");
 363		return -ENOMEM;
 364	}
 365
 366	return 0;
 367
 368err2:
 369	pci_release_regions(pdev);
 370err1:
 371	pci_disable_device(pdev);
 372err0:
 373	return rc;
 374}
 375
 376int qed_fill_dev_info(struct qed_dev *cdev,
 377		      struct qed_dev_info *dev_info)
 378{
 379	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
 380	struct qed_hw_info *hw_info = &p_hwfn->hw_info;
 381	struct qed_tunnel_info *tun = &cdev->tunnel;
 382	struct qed_ptt  *ptt;
 383
 384	memset(dev_info, 0, sizeof(struct qed_dev_info));
 385
 386	if (tun->vxlan.tun_cls == QED_TUNN_CLSS_MAC_VLAN &&
 387	    tun->vxlan.b_mode_enabled)
 388		dev_info->vxlan_enable = true;
 389
 390	if (tun->l2_gre.b_mode_enabled && tun->ip_gre.b_mode_enabled &&
 391	    tun->l2_gre.tun_cls == QED_TUNN_CLSS_MAC_VLAN &&
 392	    tun->ip_gre.tun_cls == QED_TUNN_CLSS_MAC_VLAN)
 393		dev_info->gre_enable = true;
 394
 395	if (tun->l2_geneve.b_mode_enabled && tun->ip_geneve.b_mode_enabled &&
 396	    tun->l2_geneve.tun_cls == QED_TUNN_CLSS_MAC_VLAN &&
 397	    tun->ip_geneve.tun_cls == QED_TUNN_CLSS_MAC_VLAN)
 398		dev_info->geneve_enable = true;
 399
 400	dev_info->num_hwfns = cdev->num_hwfns;
 401	dev_info->pci_mem_start = cdev->pci_params.mem_start;
 402	dev_info->pci_mem_end = cdev->pci_params.mem_end;
 403	dev_info->pci_irq = cdev->pci_params.irq;
 404	dev_info->rdma_supported = QED_IS_RDMA_PERSONALITY(p_hwfn);
 405	dev_info->dev_type = cdev->type;
 406	ether_addr_copy(dev_info->hw_mac, hw_info->hw_mac_addr);
 407
 408	if (IS_PF(cdev)) {
 409		dev_info->fw_major = FW_MAJOR_VERSION;
 410		dev_info->fw_minor = FW_MINOR_VERSION;
 411		dev_info->fw_rev = FW_REVISION_VERSION;
 412		dev_info->fw_eng = FW_ENGINEERING_VERSION;
 413		dev_info->b_inter_pf_switch = test_bit(QED_MF_INTER_PF_SWITCH,
 414						       &cdev->mf_bits);
 415		if (!test_bit(QED_MF_DISABLE_ARFS, &cdev->mf_bits))
 416			dev_info->b_arfs_capable = true;
 417		dev_info->tx_switching = true;
 418
 419		if (hw_info->b_wol_support == QED_WOL_SUPPORT_PME)
 420			dev_info->wol_support = true;
 421
 422		dev_info->smart_an = qed_mcp_is_smart_an_supported(p_hwfn);
 423		dev_info->esl = qed_mcp_is_esl_supported(p_hwfn);
 424		dev_info->abs_pf_id = QED_LEADING_HWFN(cdev)->abs_pf_id;
 425	} else {
 426		qed_vf_get_fw_version(&cdev->hwfns[0], &dev_info->fw_major,
 427				      &dev_info->fw_minor, &dev_info->fw_rev,
 428				      &dev_info->fw_eng);
 429	}
 430
 431	if (IS_PF(cdev)) {
 432		ptt = qed_ptt_acquire(QED_LEADING_HWFN(cdev));
 433		if (ptt) {
 434			qed_mcp_get_mfw_ver(QED_LEADING_HWFN(cdev), ptt,
 435					    &dev_info->mfw_rev, NULL);
 436
 437			qed_mcp_get_mbi_ver(QED_LEADING_HWFN(cdev), ptt,
 438					    &dev_info->mbi_version);
 439
 440			qed_mcp_get_flash_size(QED_LEADING_HWFN(cdev), ptt,
 441					       &dev_info->flash_size);
 442
 443			qed_ptt_release(QED_LEADING_HWFN(cdev), ptt);
 444		}
 445	} else {
 446		qed_mcp_get_mfw_ver(QED_LEADING_HWFN(cdev), NULL,
 447				    &dev_info->mfw_rev, NULL);
 448	}
 449
 450	dev_info->mtu = hw_info->mtu;
 451	cdev->common_dev_info = *dev_info;
 452
 453	return 0;
 454}
 455
 456static void qed_free_cdev(struct qed_dev *cdev)
 457{
 458	kfree((void *)cdev);
 459}
 460
 461static struct qed_dev *qed_alloc_cdev(struct pci_dev *pdev)
 462{
 463	struct qed_dev *cdev;
 464
 465	cdev = kzalloc(sizeof(*cdev), GFP_KERNEL);
 466	if (!cdev)
 467		return cdev;
 468
 469	qed_init_struct(cdev);
 470
 471	return cdev;
 472}
 473
 474/* Sets the requested power state */
 475static int qed_set_power_state(struct qed_dev *cdev, pci_power_t state)
 476{
 477	if (!cdev)
 478		return -ENODEV;
 479
 480	DP_VERBOSE(cdev, NETIF_MSG_DRV, "Omitting Power state change\n");
 481	return 0;
 482}
 483
 484/* probing */
 485static struct qed_dev *qed_probe(struct pci_dev *pdev,
 486				 struct qed_probe_params *params)
 487{
 488	struct qed_dev *cdev;
 489	int rc;
 490
 491	cdev = qed_alloc_cdev(pdev);
 492	if (!cdev)
 493		goto err0;
 494
 495	cdev->drv_type = DRV_ID_DRV_TYPE_LINUX;
 496	cdev->protocol = params->protocol;
 497
 498	if (params->is_vf)
 499		cdev->b_is_vf = true;
 500
 501	qed_init_dp(cdev, params->dp_module, params->dp_level);
 502
 503	cdev->recov_in_prog = params->recov_in_prog;
 504
 505	rc = qed_init_pci(cdev, pdev);
 506	if (rc) {
 507		DP_ERR(cdev, "init pci failed\n");
 508		goto err1;
 509	}
 510	DP_INFO(cdev, "PCI init completed successfully\n");
 511
 512	rc = qed_hw_prepare(cdev, QED_PCI_DEFAULT);
 513	if (rc) {
 514		DP_ERR(cdev, "hw prepare failed\n");
 515		goto err2;
 516	}
 517
 518	DP_INFO(cdev, "%s completed successfully\n", __func__);
 519
 520	return cdev;
 521
 522err2:
 523	qed_free_pci(cdev);
 524err1:
 525	qed_free_cdev(cdev);
 526err0:
 527	return NULL;
 528}
 529
 530static void qed_remove(struct qed_dev *cdev)
 531{
 532	if (!cdev)
 533		return;
 534
 535	qed_hw_remove(cdev);
 536
 537	qed_free_pci(cdev);
 538
 539	qed_set_power_state(cdev, PCI_D3hot);
 540
 541	qed_free_cdev(cdev);
 542}
 543
 544static void qed_disable_msix(struct qed_dev *cdev)
 545{
 546	if (cdev->int_params.out.int_mode == QED_INT_MODE_MSIX) {
 547		pci_disable_msix(cdev->pdev);
 548		kfree(cdev->int_params.msix_table);
 549	} else if (cdev->int_params.out.int_mode == QED_INT_MODE_MSI) {
 550		pci_disable_msi(cdev->pdev);
 551	}
 552
 553	memset(&cdev->int_params.out, 0, sizeof(struct qed_int_param));
 554}
 555
 556static int qed_enable_msix(struct qed_dev *cdev,
 557			   struct qed_int_params *int_params)
 558{
 559	int i, rc, cnt;
 560
 561	cnt = int_params->in.num_vectors;
 562
 563	for (i = 0; i < cnt; i++)
 564		int_params->msix_table[i].entry = i;
 565
 566	rc = pci_enable_msix_range(cdev->pdev, int_params->msix_table,
 567				   int_params->in.min_msix_cnt, cnt);
 568	if (rc < cnt && rc >= int_params->in.min_msix_cnt &&
 569	    (rc % cdev->num_hwfns)) {
 570		pci_disable_msix(cdev->pdev);
 571
 572		/* If fastpath is initialized, we need at least one interrupt
 573		 * per hwfn [and the slow path interrupts]. New requested number
 574		 * should be a multiple of the number of hwfns.
 575		 */
 576		cnt = (rc / cdev->num_hwfns) * cdev->num_hwfns;
 577		DP_NOTICE(cdev,
 578			  "Trying to enable MSI-X with less vectors (%d out of %d)\n",
 579			  cnt, int_params->in.num_vectors);
 580		rc = pci_enable_msix_exact(cdev->pdev, int_params->msix_table,
 581					   cnt);
 582		if (!rc)
 583			rc = cnt;
 584	}
 585
 586	/* For VFs, we should return with an error in case we didn't get the
 587	 * exact number of msix vectors as we requested.
 588	 * Not doing that will lead to a crash when starting queues for
 589	 * this VF.
 590	 */
 591	if ((IS_PF(cdev) && rc > 0) || (IS_VF(cdev) && rc == cnt)) {
 592		/* MSI-x configuration was achieved */
 593		int_params->out.int_mode = QED_INT_MODE_MSIX;
 594		int_params->out.num_vectors = rc;
 595		rc = 0;
 596	} else {
 597		DP_NOTICE(cdev,
 598			  "Failed to enable MSI-X [Requested %d vectors][rc %d]\n",
 599			  cnt, rc);
 600	}
 601
 602	return rc;
 603}
 604
 605/* This function outputs the int mode and the number of enabled msix vector */
 606static int qed_set_int_mode(struct qed_dev *cdev, bool force_mode)
 607{
 608	struct qed_int_params *int_params = &cdev->int_params;
 609	struct msix_entry *tbl;
 610	int rc = 0, cnt;
 611
 612	switch (int_params->in.int_mode) {
 613	case QED_INT_MODE_MSIX:
 614		/* Allocate MSIX table */
 615		cnt = int_params->in.num_vectors;
 616		int_params->msix_table = kcalloc(cnt, sizeof(*tbl), GFP_KERNEL);
 617		if (!int_params->msix_table) {
 618			rc = -ENOMEM;
 619			goto out;
 620		}
 621
 622		/* Enable MSIX */
 623		rc = qed_enable_msix(cdev, int_params);
 624		if (!rc)
 625			goto out;
 626
 627		DP_NOTICE(cdev, "Failed to enable MSI-X\n");
 628		kfree(int_params->msix_table);
 629		if (force_mode)
 630			goto out;
 631		fallthrough;
 632
 633	case QED_INT_MODE_MSI:
 634		if (cdev->num_hwfns == 1) {
 635			rc = pci_enable_msi(cdev->pdev);
 636			if (!rc) {
 637				int_params->out.int_mode = QED_INT_MODE_MSI;
 638				goto out;
 639			}
 640
 641			DP_NOTICE(cdev, "Failed to enable MSI\n");
 642			if (force_mode)
 643				goto out;
 644		}
 645		fallthrough;
 646
 647	case QED_INT_MODE_INTA:
 648			int_params->out.int_mode = QED_INT_MODE_INTA;
 649			rc = 0;
 650			goto out;
 651	default:
 652		DP_NOTICE(cdev, "Unknown int_mode value %d\n",
 653			  int_params->in.int_mode);
 654		rc = -EINVAL;
 655	}
 656
 657out:
 658	if (!rc)
 659		DP_INFO(cdev, "Using %s interrupts\n",
 660			int_params->out.int_mode == QED_INT_MODE_INTA ?
 661			"INTa" : int_params->out.int_mode == QED_INT_MODE_MSI ?
 662			"MSI" : "MSIX");
 663	cdev->int_coalescing_mode = QED_COAL_MODE_ENABLE;
 664
 665	return rc;
 666}
 667
 668static void qed_simd_handler_config(struct qed_dev *cdev, void *token,
 669				    int index, void(*handler)(void *))
 670{
 671	struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns];
 672	int relative_idx = index / cdev->num_hwfns;
 673
 674	hwfn->simd_proto_handler[relative_idx].func = handler;
 675	hwfn->simd_proto_handler[relative_idx].token = token;
 676}
 677
 678static void qed_simd_handler_clean(struct qed_dev *cdev, int index)
 679{
 680	struct qed_hwfn *hwfn = &cdev->hwfns[index % cdev->num_hwfns];
 681	int relative_idx = index / cdev->num_hwfns;
 682
 683	memset(&hwfn->simd_proto_handler[relative_idx], 0,
 684	       sizeof(struct qed_simd_fp_handler));
 685}
 686
 687static irqreturn_t qed_msix_sp_int(int irq, void *tasklet)
 688{
 689	tasklet_schedule((struct tasklet_struct *)tasklet);
 690	return IRQ_HANDLED;
 691}
 692
 693static irqreturn_t qed_single_int(int irq, void *dev_instance)
 694{
 695	struct qed_dev *cdev = (struct qed_dev *)dev_instance;
 696	struct qed_hwfn *hwfn;
 697	irqreturn_t rc = IRQ_NONE;
 698	u64 status;
 699	int i, j;
 700
 701	for (i = 0; i < cdev->num_hwfns; i++) {
 702		status = qed_int_igu_read_sisr_reg(&cdev->hwfns[i]);
 703
 704		if (!status)
 705			continue;
 706
 707		hwfn = &cdev->hwfns[i];
 708
 709		/* Slowpath interrupt */
 710		if (unlikely(status & 0x1)) {
 711			tasklet_schedule(&hwfn->sp_dpc);
 712			status &= ~0x1;
 713			rc = IRQ_HANDLED;
 714		}
 715
 716		/* Fastpath interrupts */
 717		for (j = 0; j < 64; j++) {
 718			if ((0x2ULL << j) & status) {
 719				struct qed_simd_fp_handler *p_handler =
 720					&hwfn->simd_proto_handler[j];
 721
 722				if (p_handler->func)
 723					p_handler->func(p_handler->token);
 724				else
 725					DP_NOTICE(hwfn,
 726						  "Not calling fastpath handler as it is NULL [handler #%d, status 0x%llx]\n",
 727						  j, status);
 728
 729				status &= ~(0x2ULL << j);
 730				rc = IRQ_HANDLED;
 731			}
 732		}
 733
 734		if (unlikely(status))
 735			DP_VERBOSE(hwfn, NETIF_MSG_INTR,
 736				   "got an unknown interrupt status 0x%llx\n",
 737				   status);
 738	}
 739
 740	return rc;
 741}
 742
 743int qed_slowpath_irq_req(struct qed_hwfn *hwfn)
 744{
 745	struct qed_dev *cdev = hwfn->cdev;
 746	u32 int_mode;
 747	int rc = 0;
 748	u8 id;
 749
 750	int_mode = cdev->int_params.out.int_mode;
 751	if (int_mode == QED_INT_MODE_MSIX) {
 752		id = hwfn->my_id;
 753		snprintf(hwfn->name, NAME_SIZE, "sp-%d-%02x:%02x.%02x",
 754			 id, cdev->pdev->bus->number,
 755			 PCI_SLOT(cdev->pdev->devfn), hwfn->abs_pf_id);
 756		rc = request_irq(cdev->int_params.msix_table[id].vector,
 757				 qed_msix_sp_int, 0, hwfn->name, &hwfn->sp_dpc);
 758	} else {
 759		unsigned long flags = 0;
 760
 761		snprintf(cdev->name, NAME_SIZE, "%02x:%02x.%02x",
 762			 cdev->pdev->bus->number, PCI_SLOT(cdev->pdev->devfn),
 763			 PCI_FUNC(cdev->pdev->devfn));
 764
 765		if (cdev->int_params.out.int_mode == QED_INT_MODE_INTA)
 766			flags |= IRQF_SHARED;
 767
 768		rc = request_irq(cdev->pdev->irq, qed_single_int,
 769				 flags, cdev->name, cdev);
 770	}
 771
 772	if (rc)
 773		DP_NOTICE(cdev, "request_irq failed, rc = %d\n", rc);
 774	else
 775		DP_VERBOSE(hwfn, (NETIF_MSG_INTR | QED_MSG_SP),
 776			   "Requested slowpath %s\n",
 777			   (int_mode == QED_INT_MODE_MSIX) ? "MSI-X" : "IRQ");
 778
 779	return rc;
 780}
 781
 782static void qed_slowpath_tasklet_flush(struct qed_hwfn *p_hwfn)
 783{
 784	/* Calling the disable function will make sure that any
 785	 * currently-running function is completed. The following call to the
 786	 * enable function makes this sequence a flush-like operation.
 787	 */
 788	if (p_hwfn->b_sp_dpc_enabled) {
 789		tasklet_disable(&p_hwfn->sp_dpc);
 790		tasklet_enable(&p_hwfn->sp_dpc);
 791	}
 792}
 793
 794void qed_slowpath_irq_sync(struct qed_hwfn *p_hwfn)
 795{
 796	struct qed_dev *cdev = p_hwfn->cdev;
 797	u8 id = p_hwfn->my_id;
 798	u32 int_mode;
 799
 800	int_mode = cdev->int_params.out.int_mode;
 801	if (int_mode == QED_INT_MODE_MSIX)
 802		synchronize_irq(cdev->int_params.msix_table[id].vector);
 803	else
 804		synchronize_irq(cdev->pdev->irq);
 805
 806	qed_slowpath_tasklet_flush(p_hwfn);
 807}
 808
 809static void qed_slowpath_irq_free(struct qed_dev *cdev)
 810{
 811	int i;
 812
 813	if (cdev->int_params.out.int_mode == QED_INT_MODE_MSIX) {
 814		for_each_hwfn(cdev, i) {
 815			if (!cdev->hwfns[i].b_int_requested)
 816				break;
 817			free_irq(cdev->int_params.msix_table[i].vector,
 818				 &cdev->hwfns[i].sp_dpc);
 819		}
 820	} else {
 821		if (QED_LEADING_HWFN(cdev)->b_int_requested)
 822			free_irq(cdev->pdev->irq, cdev);
 823	}
 824	qed_int_disable_post_isr_release(cdev);
 825}
 826
 827static int qed_nic_stop(struct qed_dev *cdev)
 828{
 829	int i, rc;
 830
 831	rc = qed_hw_stop(cdev);
 832
 833	for (i = 0; i < cdev->num_hwfns; i++) {
 834		struct qed_hwfn *p_hwfn = &cdev->hwfns[i];
 835
 836		if (p_hwfn->b_sp_dpc_enabled) {
 837			tasklet_disable(&p_hwfn->sp_dpc);
 838			p_hwfn->b_sp_dpc_enabled = false;
 839			DP_VERBOSE(cdev, NETIF_MSG_IFDOWN,
 840				   "Disabled sp tasklet [hwfn %d] at %p\n",
 841				   i, &p_hwfn->sp_dpc);
 842		}
 843	}
 844
 845	qed_dbg_pf_exit(cdev);
 846
 847	return rc;
 848}
 849
 850static int qed_nic_setup(struct qed_dev *cdev)
 851{
 852	int rc, i;
 853
 854	/* Determine if interface is going to require LL2 */
 855	if (QED_LEADING_HWFN(cdev)->hw_info.personality != QED_PCI_ETH) {
 856		for (i = 0; i < cdev->num_hwfns; i++) {
 857			struct qed_hwfn *p_hwfn = &cdev->hwfns[i];
 858
 859			p_hwfn->using_ll2 = true;
 860		}
 861	}
 862
 863	rc = qed_resc_alloc(cdev);
 864	if (rc)
 865		return rc;
 866
 867	DP_INFO(cdev, "Allocated qed resources\n");
 868
 869	qed_resc_setup(cdev);
 870
 871	return rc;
 872}
 873
 874static int qed_set_int_fp(struct qed_dev *cdev, u16 cnt)
 875{
 876	int limit = 0;
 877
 878	/* Mark the fastpath as free/used */
 879	cdev->int_params.fp_initialized = cnt ? true : false;
 880
 881	if (cdev->int_params.out.int_mode != QED_INT_MODE_MSIX)
 882		limit = cdev->num_hwfns * 63;
 883	else if (cdev->int_params.fp_msix_cnt)
 884		limit = cdev->int_params.fp_msix_cnt;
 885
 886	if (!limit)
 887		return -ENOMEM;
 888
 889	return min_t(int, cnt, limit);
 890}
 891
 892static int qed_get_int_fp(struct qed_dev *cdev, struct qed_int_info *info)
 893{
 894	memset(info, 0, sizeof(struct qed_int_info));
 895
 896	if (!cdev->int_params.fp_initialized) {
 897		DP_INFO(cdev,
 898			"Protocol driver requested interrupt information, but its support is not yet configured\n");
 899		return -EINVAL;
 900	}
 901
 902	/* Need to expose only MSI-X information; Single IRQ is handled solely
 903	 * by qed.
 904	 */
 905	if (cdev->int_params.out.int_mode == QED_INT_MODE_MSIX) {
 906		int msix_base = cdev->int_params.fp_msix_base;
 907
 908		info->msix_cnt = cdev->int_params.fp_msix_cnt;
 909		info->msix = &cdev->int_params.msix_table[msix_base];
 910	}
 911
 912	return 0;
 913}
 914
 915static int qed_slowpath_setup_int(struct qed_dev *cdev,
 916				  enum qed_int_mode int_mode)
 917{
 918	struct qed_sb_cnt_info sb_cnt_info;
 919	int num_l2_queues = 0;
 920	int rc;
 921	int i;
 922
 923	if ((int_mode == QED_INT_MODE_MSI) && (cdev->num_hwfns > 1)) {
 924		DP_NOTICE(cdev, "MSI mode is not supported for CMT devices\n");
 925		return -EINVAL;
 926	}
 927
 928	memset(&cdev->int_params, 0, sizeof(struct qed_int_params));
 929	cdev->int_params.in.int_mode = int_mode;
 930	for_each_hwfn(cdev, i) {
 931		memset(&sb_cnt_info, 0, sizeof(sb_cnt_info));
 932		qed_int_get_num_sbs(&cdev->hwfns[i], &sb_cnt_info);
 933		cdev->int_params.in.num_vectors += sb_cnt_info.cnt;
 934		cdev->int_params.in.num_vectors++; /* slowpath */
 935	}
 936
 937	/* We want a minimum of one slowpath and one fastpath vector per hwfn */
 938	cdev->int_params.in.min_msix_cnt = cdev->num_hwfns * 2;
 939
 940	if (is_kdump_kernel()) {
 941		DP_INFO(cdev,
 942			"Kdump kernel: Limit the max number of requested MSI-X vectors to %hd\n",
 943			cdev->int_params.in.min_msix_cnt);
 944		cdev->int_params.in.num_vectors =
 945			cdev->int_params.in.min_msix_cnt;
 946	}
 947
 948	rc = qed_set_int_mode(cdev, false);
 949	if (rc)  {
 950		DP_ERR(cdev, "%s ERR\n", __func__);
 951		return rc;
 952	}
 953
 954	cdev->int_params.fp_msix_base = cdev->num_hwfns;
 955	cdev->int_params.fp_msix_cnt = cdev->int_params.out.num_vectors -
 956				       cdev->num_hwfns;
 957
 958	if (!IS_ENABLED(CONFIG_QED_RDMA) ||
 959	    !QED_IS_RDMA_PERSONALITY(QED_LEADING_HWFN(cdev)))
 960		return 0;
 961
 962	for_each_hwfn(cdev, i)
 963		num_l2_queues += FEAT_NUM(&cdev->hwfns[i], QED_PF_L2_QUE);
 964
 965	DP_VERBOSE(cdev, QED_MSG_RDMA,
 966		   "cdev->int_params.fp_msix_cnt=%d num_l2_queues=%d\n",
 967		   cdev->int_params.fp_msix_cnt, num_l2_queues);
 968
 969	if (cdev->int_params.fp_msix_cnt > num_l2_queues) {
 970		cdev->int_params.rdma_msix_cnt =
 971			(cdev->int_params.fp_msix_cnt - num_l2_queues)
 972			/ cdev->num_hwfns;
 973		cdev->int_params.rdma_msix_base =
 974			cdev->int_params.fp_msix_base + num_l2_queues;
 975		cdev->int_params.fp_msix_cnt = num_l2_queues;
 976	} else {
 977		cdev->int_params.rdma_msix_cnt = 0;
 978	}
 979
 980	DP_VERBOSE(cdev, QED_MSG_RDMA, "roce_msix_cnt=%d roce_msix_base=%d\n",
 981		   cdev->int_params.rdma_msix_cnt,
 982		   cdev->int_params.rdma_msix_base);
 983
 984	return 0;
 985}
 986
 987static int qed_slowpath_vf_setup_int(struct qed_dev *cdev)
 988{
 989	int rc;
 990
 991	memset(&cdev->int_params, 0, sizeof(struct qed_int_params));
 992	cdev->int_params.in.int_mode = QED_INT_MODE_MSIX;
 993
 994	qed_vf_get_num_rxqs(QED_LEADING_HWFN(cdev),
 995			    &cdev->int_params.in.num_vectors);
 996	if (cdev->num_hwfns > 1) {
 997		u8 vectors = 0;
 998
 999		qed_vf_get_num_rxqs(&cdev->hwfns[1], &vectors);
1000		cdev->int_params.in.num_vectors += vectors;
1001	}
1002
1003	/* We want a minimum of one fastpath vector per vf hwfn */
1004	cdev->int_params.in.min_msix_cnt = cdev->num_hwfns;
1005
1006	rc = qed_set_int_mode(cdev, true);
1007	if (rc)
1008		return rc;
1009
1010	cdev->int_params.fp_msix_base = 0;
1011	cdev->int_params.fp_msix_cnt = cdev->int_params.out.num_vectors;
1012
1013	return 0;
1014}
1015
1016u32 qed_unzip_data(struct qed_hwfn *p_hwfn, u32 input_len,
1017		   u8 *input_buf, u32 max_size, u8 *unzip_buf)
1018{
1019	int rc;
1020
1021	p_hwfn->stream->next_in = input_buf;
1022	p_hwfn->stream->avail_in = input_len;
1023	p_hwfn->stream->next_out = unzip_buf;
1024	p_hwfn->stream->avail_out = max_size;
1025
1026	rc = zlib_inflateInit2(p_hwfn->stream, MAX_WBITS);
1027
1028	if (rc != Z_OK) {
1029		DP_VERBOSE(p_hwfn, NETIF_MSG_DRV, "zlib init failed, rc = %d\n",
1030			   rc);
1031		return 0;
1032	}
1033
1034	rc = zlib_inflate(p_hwfn->stream, Z_FINISH);
1035	zlib_inflateEnd(p_hwfn->stream);
1036
1037	if (rc != Z_OK && rc != Z_STREAM_END) {
1038		DP_VERBOSE(p_hwfn, NETIF_MSG_DRV, "FW unzip error: %s, rc=%d\n",
1039			   p_hwfn->stream->msg, rc);
1040		return 0;
1041	}
1042
1043	return p_hwfn->stream->total_out / 4;
1044}
1045
1046static int qed_alloc_stream_mem(struct qed_dev *cdev)
1047{
1048	int i;
1049	void *workspace;
1050
1051	for_each_hwfn(cdev, i) {
1052		struct qed_hwfn *p_hwfn = &cdev->hwfns[i];
1053
1054		p_hwfn->stream = kzalloc(sizeof(*p_hwfn->stream), GFP_KERNEL);
1055		if (!p_hwfn->stream)
1056			return -ENOMEM;
1057
1058		workspace = vzalloc(zlib_inflate_workspacesize());
1059		if (!workspace)
1060			return -ENOMEM;
1061		p_hwfn->stream->workspace = workspace;
1062	}
1063
1064	return 0;
1065}
1066
1067static void qed_free_stream_mem(struct qed_dev *cdev)
1068{
1069	int i;
1070
1071	for_each_hwfn(cdev, i) {
1072		struct qed_hwfn *p_hwfn = &cdev->hwfns[i];
1073
1074		if (!p_hwfn->stream)
1075			return;
1076
1077		vfree(p_hwfn->stream->workspace);
1078		kfree(p_hwfn->stream);
1079	}
1080}
1081
1082static void qed_update_pf_params(struct qed_dev *cdev,
1083				 struct qed_pf_params *params)
1084{
1085	int i;
1086
1087	if (IS_ENABLED(CONFIG_QED_RDMA)) {
1088		params->rdma_pf_params.num_qps = QED_ROCE_QPS;
1089		params->rdma_pf_params.min_dpis = QED_ROCE_DPIS;
1090		params->rdma_pf_params.num_srqs = QED_RDMA_SRQS;
1091		/* divide by 3 the MRs to avoid MF ILT overflow */
1092		params->rdma_pf_params.gl_pi = QED_ROCE_PROTOCOL_INDEX;
1093	}
1094
1095	if (cdev->num_hwfns > 1 || IS_VF(cdev))
1096		params->eth_pf_params.num_arfs_filters = 0;
1097
1098	/* In case we might support RDMA, don't allow qede to be greedy
1099	 * with the L2 contexts. Allow for 64 queues [rx, tx cos, xdp]
1100	 * per hwfn.
1101	 */
1102	if (QED_IS_RDMA_PERSONALITY(QED_LEADING_HWFN(cdev))) {
1103		u16 *num_cons;
1104
1105		num_cons = &params->eth_pf_params.num_cons;
1106		*num_cons = min_t(u16, *num_cons, QED_MAX_L2_CONS);
1107	}
1108
1109	for (i = 0; i < cdev->num_hwfns; i++) {
1110		struct qed_hwfn *p_hwfn = &cdev->hwfns[i];
1111
1112		p_hwfn->pf_params = *params;
1113	}
1114}
1115
1116#define QED_PERIODIC_DB_REC_COUNT		10
1117#define QED_PERIODIC_DB_REC_INTERVAL_MS		100
1118#define QED_PERIODIC_DB_REC_INTERVAL \
1119	msecs_to_jiffies(QED_PERIODIC_DB_REC_INTERVAL_MS)
1120
1121static int qed_slowpath_delayed_work(struct qed_hwfn *hwfn,
1122				     enum qed_slowpath_wq_flag wq_flag,
1123				     unsigned long delay)
1124{
1125	if (!hwfn->slowpath_wq_active)
1126		return -EINVAL;
1127
1128	/* Memory barrier for setting atomic bit */
1129	smp_mb__before_atomic();
1130	set_bit(wq_flag, &hwfn->slowpath_task_flags);
1131	/* Memory barrier after setting atomic bit */
1132	smp_mb__after_atomic();
1133	queue_delayed_work(hwfn->slowpath_wq, &hwfn->slowpath_task, delay);
1134
1135	return 0;
1136}
1137
1138void qed_periodic_db_rec_start(struct qed_hwfn *p_hwfn)
1139{
1140	/* Reset periodic Doorbell Recovery counter */
1141	p_hwfn->periodic_db_rec_count = QED_PERIODIC_DB_REC_COUNT;
1142
1143	/* Don't schedule periodic Doorbell Recovery if already scheduled */
1144	if (test_bit(QED_SLOWPATH_PERIODIC_DB_REC,
1145		     &p_hwfn->slowpath_task_flags))
1146		return;
1147
1148	qed_slowpath_delayed_work(p_hwfn, QED_SLOWPATH_PERIODIC_DB_REC,
1149				  QED_PERIODIC_DB_REC_INTERVAL);
1150}
1151
1152static void qed_slowpath_wq_stop(struct qed_dev *cdev)
1153{
1154	int i;
1155
1156	if (IS_VF(cdev))
1157		return;
1158
1159	for_each_hwfn(cdev, i) {
1160		if (!cdev->hwfns[i].slowpath_wq)
1161			continue;
1162
1163		/* Stop queuing new delayed works */
1164		cdev->hwfns[i].slowpath_wq_active = false;
1165
1166		cancel_delayed_work(&cdev->hwfns[i].slowpath_task);
1167		destroy_workqueue(cdev->hwfns[i].slowpath_wq);
1168	}
1169}
1170
1171static void qed_slowpath_task(struct work_struct *work)
1172{
1173	struct qed_hwfn *hwfn = container_of(work, struct qed_hwfn,
1174					     slowpath_task.work);
1175	struct qed_ptt *ptt = qed_ptt_acquire(hwfn);
1176
1177	if (!ptt) {
1178		if (hwfn->slowpath_wq_active)
1179			queue_delayed_work(hwfn->slowpath_wq,
1180					   &hwfn->slowpath_task, 0);
1181
1182		return;
1183	}
1184
1185	if (test_and_clear_bit(QED_SLOWPATH_MFW_TLV_REQ,
1186			       &hwfn->slowpath_task_flags))
1187		qed_mfw_process_tlv_req(hwfn, ptt);
1188
1189	if (test_and_clear_bit(QED_SLOWPATH_PERIODIC_DB_REC,
1190			       &hwfn->slowpath_task_flags)) {
1191		/* skip qed_db_rec_handler during recovery/unload */
1192		if (hwfn->cdev->recov_in_prog || !hwfn->slowpath_wq_active)
1193			goto out;
1194
1195		qed_db_rec_handler(hwfn, ptt);
1196		if (hwfn->periodic_db_rec_count--)
1197			qed_slowpath_delayed_work(hwfn,
1198						  QED_SLOWPATH_PERIODIC_DB_REC,
1199						  QED_PERIODIC_DB_REC_INTERVAL);
1200	}
1201
1202out:
1203	qed_ptt_release(hwfn, ptt);
1204}
1205
1206static int qed_slowpath_wq_start(struct qed_dev *cdev)
1207{
1208	struct qed_hwfn *hwfn;
1209	char name[NAME_SIZE];
1210	int i;
1211
1212	if (IS_VF(cdev))
1213		return 0;
1214
1215	for_each_hwfn(cdev, i) {
1216		hwfn = &cdev->hwfns[i];
1217
1218		snprintf(name, NAME_SIZE, "slowpath-%02x:%02x.%02x",
1219			 cdev->pdev->bus->number,
1220			 PCI_SLOT(cdev->pdev->devfn), hwfn->abs_pf_id);
1221
1222		hwfn->slowpath_wq = alloc_workqueue(name, 0, 0);
1223		if (!hwfn->slowpath_wq) {
1224			DP_NOTICE(hwfn, "Cannot create slowpath workqueue\n");
1225			return -ENOMEM;
1226		}
1227
1228		INIT_DELAYED_WORK(&hwfn->slowpath_task, qed_slowpath_task);
1229		hwfn->slowpath_wq_active = true;
1230	}
1231
1232	return 0;
1233}
1234
1235static int qed_slowpath_start(struct qed_dev *cdev,
1236			      struct qed_slowpath_params *params)
1237{
1238	struct qed_drv_load_params drv_load_params;
1239	struct qed_hw_init_params hw_init_params;
1240	struct qed_mcp_drv_version drv_version;
1241	struct qed_tunnel_info tunn_info;
1242	const u8 *data = NULL;
1243	struct qed_hwfn *hwfn;
1244	struct qed_ptt *p_ptt;
1245	int rc = -EINVAL;
1246
1247	if (qed_iov_wq_start(cdev))
1248		goto err;
1249
1250	if (qed_slowpath_wq_start(cdev))
1251		goto err;
1252
1253	if (IS_PF(cdev)) {
1254		rc = request_firmware(&cdev->firmware, QED_FW_FILE_NAME,
1255				      &cdev->pdev->dev);
1256		if (rc) {
1257			DP_NOTICE(cdev,
1258				  "Failed to find fw file - /lib/firmware/%s\n",
1259				  QED_FW_FILE_NAME);
1260			goto err;
1261		}
1262
1263		if (cdev->num_hwfns == 1) {
1264			p_ptt = qed_ptt_acquire(QED_LEADING_HWFN(cdev));
1265			if (p_ptt) {
1266				QED_LEADING_HWFN(cdev)->p_arfs_ptt = p_ptt;
1267			} else {
1268				DP_NOTICE(cdev,
1269					  "Failed to acquire PTT for aRFS\n");
1270				rc = -EINVAL;
1271				goto err;
1272			}
1273		}
1274	}
1275
1276	cdev->rx_coalesce_usecs = QED_DEFAULT_RX_USECS;
1277	rc = qed_nic_setup(cdev);
1278	if (rc)
1279		goto err;
1280
1281	if (IS_PF(cdev))
1282		rc = qed_slowpath_setup_int(cdev, params->int_mode);
1283	else
1284		rc = qed_slowpath_vf_setup_int(cdev);
1285	if (rc)
1286		goto err1;
1287
1288	if (IS_PF(cdev)) {
1289		/* Allocate stream for unzipping */
1290		rc = qed_alloc_stream_mem(cdev);
1291		if (rc)
1292			goto err2;
1293
1294		/* First Dword used to differentiate between various sources */
1295		data = cdev->firmware->data + sizeof(u32);
1296
1297		qed_dbg_pf_init(cdev);
1298	}
1299
1300	/* Start the slowpath */
1301	memset(&hw_init_params, 0, sizeof(hw_init_params));
1302	memset(&tunn_info, 0, sizeof(tunn_info));
1303	tunn_info.vxlan.b_mode_enabled = true;
1304	tunn_info.l2_gre.b_mode_enabled = true;
1305	tunn_info.ip_gre.b_mode_enabled = true;
1306	tunn_info.l2_geneve.b_mode_enabled = true;
1307	tunn_info.ip_geneve.b_mode_enabled = true;
1308	tunn_info.vxlan.tun_cls = QED_TUNN_CLSS_MAC_VLAN;
1309	tunn_info.l2_gre.tun_cls = QED_TUNN_CLSS_MAC_VLAN;
1310	tunn_info.ip_gre.tun_cls = QED_TUNN_CLSS_MAC_VLAN;
1311	tunn_info.l2_geneve.tun_cls = QED_TUNN_CLSS_MAC_VLAN;
1312	tunn_info.ip_geneve.tun_cls = QED_TUNN_CLSS_MAC_VLAN;
1313	hw_init_params.p_tunn = &tunn_info;
1314	hw_init_params.b_hw_start = true;
1315	hw_init_params.int_mode = cdev->int_params.out.int_mode;
1316	hw_init_params.allow_npar_tx_switch = true;
1317	hw_init_params.bin_fw_data = data;
1318
1319	memset(&drv_load_params, 0, sizeof(drv_load_params));
1320	drv_load_params.is_crash_kernel = is_kdump_kernel();
1321	drv_load_params.mfw_timeout_val = QED_LOAD_REQ_LOCK_TO_DEFAULT;
1322	drv_load_params.avoid_eng_reset = false;
1323	drv_load_params.override_force_load = QED_OVERRIDE_FORCE_LOAD_NONE;
1324	hw_init_params.p_drv_load_params = &drv_load_params;
1325
1326	rc = qed_hw_init(cdev, &hw_init_params);
1327	if (rc)
1328		goto err2;
1329
1330	DP_INFO(cdev,
1331		"HW initialization and function start completed successfully\n");
1332
1333	if (IS_PF(cdev)) {
1334		cdev->tunn_feature_mask = (BIT(QED_MODE_VXLAN_TUNN) |
1335					   BIT(QED_MODE_L2GENEVE_TUNN) |
1336					   BIT(QED_MODE_IPGENEVE_TUNN) |
1337					   BIT(QED_MODE_L2GRE_TUNN) |
1338					   BIT(QED_MODE_IPGRE_TUNN));
1339	}
1340
1341	/* Allocate LL2 interface if needed */
1342	if (QED_LEADING_HWFN(cdev)->using_ll2) {
1343		rc = qed_ll2_alloc_if(cdev);
1344		if (rc)
1345			goto err3;
1346	}
1347	if (IS_PF(cdev)) {
1348		hwfn = QED_LEADING_HWFN(cdev);
1349		drv_version.version = (params->drv_major << 24) |
1350				      (params->drv_minor << 16) |
1351				      (params->drv_rev << 8) |
1352				      (params->drv_eng);
1353		strscpy(drv_version.name, params->name,
1354			MCP_DRV_VER_STR_SIZE - 4);
1355		rc = qed_mcp_send_drv_version(hwfn, hwfn->p_main_ptt,
1356					      &drv_version);
1357		if (rc) {
1358			DP_NOTICE(cdev, "Failed sending drv version command\n");
1359			goto err4;
1360		}
1361	}
1362
1363	qed_reset_vport_stats(cdev);
1364
1365	return 0;
1366
1367err4:
1368	qed_ll2_dealloc_if(cdev);
1369err3:
1370	qed_hw_stop(cdev);
1371err2:
1372	qed_hw_timers_stop_all(cdev);
1373	if (IS_PF(cdev))
1374		qed_slowpath_irq_free(cdev);
1375	qed_free_stream_mem(cdev);
1376	qed_disable_msix(cdev);
1377err1:
1378	qed_resc_free(cdev);
1379err:
1380	if (IS_PF(cdev))
1381		release_firmware(cdev->firmware);
1382
1383	if (IS_PF(cdev) && (cdev->num_hwfns == 1) &&
1384	    QED_LEADING_HWFN(cdev)->p_arfs_ptt)
1385		qed_ptt_release(QED_LEADING_HWFN(cdev),
1386				QED_LEADING_HWFN(cdev)->p_arfs_ptt);
1387
1388	qed_iov_wq_stop(cdev, false);
1389
1390	qed_slowpath_wq_stop(cdev);
1391
1392	return rc;
1393}
1394
1395static int qed_slowpath_stop(struct qed_dev *cdev)
1396{
1397	if (!cdev)
1398		return -ENODEV;
1399
1400	qed_slowpath_wq_stop(cdev);
1401
1402	qed_ll2_dealloc_if(cdev);
1403
1404	if (IS_PF(cdev)) {
1405		if (cdev->num_hwfns == 1)
1406			qed_ptt_release(QED_LEADING_HWFN(cdev),
1407					QED_LEADING_HWFN(cdev)->p_arfs_ptt);
1408		qed_free_stream_mem(cdev);
1409		if (IS_QED_ETH_IF(cdev))
1410			qed_sriov_disable(cdev, true);
1411	}
1412
1413	qed_nic_stop(cdev);
1414
1415	if (IS_PF(cdev))
1416		qed_slowpath_irq_free(cdev);
1417
1418	qed_disable_msix(cdev);
1419
1420	qed_resc_free(cdev);
1421
1422	qed_iov_wq_stop(cdev, true);
1423
1424	if (IS_PF(cdev))
1425		release_firmware(cdev->firmware);
1426
1427	return 0;
1428}
1429
1430static void qed_set_name(struct qed_dev *cdev, char name[NAME_SIZE])
1431{
1432	int i;
1433
1434	memcpy(cdev->name, name, NAME_SIZE);
1435	for_each_hwfn(cdev, i)
1436		snprintf(cdev->hwfns[i].name, NAME_SIZE, "%s-%d", name, i);
1437}
1438
1439static u32 qed_sb_init(struct qed_dev *cdev,
1440		       struct qed_sb_info *sb_info,
1441		       void *sb_virt_addr,
1442		       dma_addr_t sb_phy_addr, u16 sb_id,
1443		       enum qed_sb_type type)
1444{
1445	struct qed_hwfn *p_hwfn;
1446	struct qed_ptt *p_ptt;
1447	u16 rel_sb_id;
1448	u32 rc;
1449
1450	/* RoCE/Storage use a single engine in CMT mode while L2 uses both */
1451	if (type == QED_SB_TYPE_L2_QUEUE) {
1452		p_hwfn = &cdev->hwfns[sb_id % cdev->num_hwfns];
1453		rel_sb_id = sb_id / cdev->num_hwfns;
1454	} else {
1455		p_hwfn = QED_AFFIN_HWFN(cdev);
1456		rel_sb_id = sb_id;
1457	}
1458
1459	DP_VERBOSE(cdev, NETIF_MSG_INTR,
1460		   "hwfn [%d] <--[init]-- SB %04x [0x%04x upper]\n",
1461		   IS_LEAD_HWFN(p_hwfn) ? 0 : 1, rel_sb_id, sb_id);
1462
1463	if (IS_PF(p_hwfn->cdev)) {
1464		p_ptt = qed_ptt_acquire(p_hwfn);
1465		if (!p_ptt)
1466			return -EBUSY;
1467
1468		rc = qed_int_sb_init(p_hwfn, p_ptt, sb_info, sb_virt_addr,
1469				     sb_phy_addr, rel_sb_id);
1470		qed_ptt_release(p_hwfn, p_ptt);
1471	} else {
1472		rc = qed_int_sb_init(p_hwfn, NULL, sb_info, sb_virt_addr,
1473				     sb_phy_addr, rel_sb_id);
1474	}
1475
1476	return rc;
1477}
1478
1479static u32 qed_sb_release(struct qed_dev *cdev,
1480			  struct qed_sb_info *sb_info,
1481			  u16 sb_id,
1482			  enum qed_sb_type type)
1483{
1484	struct qed_hwfn *p_hwfn;
1485	u16 rel_sb_id;
1486	u32 rc;
1487
1488	/* RoCE/Storage use a single engine in CMT mode while L2 uses both */
1489	if (type == QED_SB_TYPE_L2_QUEUE) {
1490		p_hwfn = &cdev->hwfns[sb_id % cdev->num_hwfns];
1491		rel_sb_id = sb_id / cdev->num_hwfns;
1492	} else {
1493		p_hwfn = QED_AFFIN_HWFN(cdev);
1494		rel_sb_id = sb_id;
1495	}
1496
1497	DP_VERBOSE(cdev, NETIF_MSG_INTR,
1498		   "hwfn [%d] <--[init]-- SB %04x [0x%04x upper]\n",
1499		   IS_LEAD_HWFN(p_hwfn) ? 0 : 1, rel_sb_id, sb_id);
1500
1501	rc = qed_int_sb_release(p_hwfn, sb_info, rel_sb_id);
1502
1503	return rc;
1504}
1505
1506static bool qed_can_link_change(struct qed_dev *cdev)
1507{
1508	return true;
1509}
1510
1511static void qed_set_ext_speed_params(struct qed_mcp_link_params *link_params,
1512				     const struct qed_link_params *params)
1513{
1514	struct qed_mcp_link_speed_params *ext_speed = &link_params->ext_speed;
1515	const struct qed_mfw_speed_map *map;
1516	u32 i;
1517
1518	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_AUTONEG)
1519		ext_speed->autoneg = !!params->autoneg;
1520
1521	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_ADV_SPEEDS) {
1522		ext_speed->advertised_speeds = 0;
1523
1524		for (i = 0; i < ARRAY_SIZE(qed_mfw_ext_maps); i++) {
1525			map = qed_mfw_ext_maps + i;
1526
1527			if (linkmode_intersects(params->adv_speeds, map->caps))
1528				ext_speed->advertised_speeds |= map->mfw_val;
1529		}
1530	}
1531
1532	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_FORCED_SPEED) {
1533		switch (params->forced_speed) {
1534		case SPEED_1000:
1535			ext_speed->forced_speed = QED_EXT_SPEED_1G;
1536			break;
1537		case SPEED_10000:
1538			ext_speed->forced_speed = QED_EXT_SPEED_10G;
1539			break;
1540		case SPEED_20000:
1541			ext_speed->forced_speed = QED_EXT_SPEED_20G;
1542			break;
1543		case SPEED_25000:
1544			ext_speed->forced_speed = QED_EXT_SPEED_25G;
1545			break;
1546		case SPEED_40000:
1547			ext_speed->forced_speed = QED_EXT_SPEED_40G;
1548			break;
1549		case SPEED_50000:
1550			ext_speed->forced_speed = QED_EXT_SPEED_50G_R |
1551						  QED_EXT_SPEED_50G_R2;
1552			break;
1553		case SPEED_100000:
1554			ext_speed->forced_speed = QED_EXT_SPEED_100G_R2 |
1555						  QED_EXT_SPEED_100G_R4 |
1556						  QED_EXT_SPEED_100G_P4;
1557			break;
1558		default:
1559			break;
1560		}
1561	}
1562
1563	if (!(params->override_flags & QED_LINK_OVERRIDE_FEC_CONFIG))
1564		return;
1565
1566	switch (params->forced_speed) {
1567	case SPEED_25000:
1568		switch (params->fec) {
1569		case FEC_FORCE_MODE_NONE:
1570			link_params->ext_fec_mode = ETH_EXT_FEC_25G_NONE;
1571			break;
1572		case FEC_FORCE_MODE_FIRECODE:
1573			link_params->ext_fec_mode = ETH_EXT_FEC_25G_BASE_R;
1574			break;
1575		case FEC_FORCE_MODE_RS:
1576			link_params->ext_fec_mode = ETH_EXT_FEC_25G_RS528;
1577			break;
1578		case FEC_FORCE_MODE_AUTO:
1579			link_params->ext_fec_mode = ETH_EXT_FEC_25G_RS528 |
1580						    ETH_EXT_FEC_25G_BASE_R |
1581						    ETH_EXT_FEC_25G_NONE;
1582			break;
1583		default:
1584			break;
1585		}
1586
1587		break;
1588	case SPEED_40000:
1589		switch (params->fec) {
1590		case FEC_FORCE_MODE_NONE:
1591			link_params->ext_fec_mode = ETH_EXT_FEC_40G_NONE;
1592			break;
1593		case FEC_FORCE_MODE_FIRECODE:
1594			link_params->ext_fec_mode = ETH_EXT_FEC_40G_BASE_R;
1595			break;
1596		case FEC_FORCE_MODE_AUTO:
1597			link_params->ext_fec_mode = ETH_EXT_FEC_40G_BASE_R |
1598						    ETH_EXT_FEC_40G_NONE;
1599			break;
1600		default:
1601			break;
1602		}
1603
1604		break;
1605	case SPEED_50000:
1606		switch (params->fec) {
1607		case FEC_FORCE_MODE_NONE:
1608			link_params->ext_fec_mode = ETH_EXT_FEC_50G_NONE;
1609			break;
1610		case FEC_FORCE_MODE_FIRECODE:
1611			link_params->ext_fec_mode = ETH_EXT_FEC_50G_BASE_R;
1612			break;
1613		case FEC_FORCE_MODE_RS:
1614			link_params->ext_fec_mode = ETH_EXT_FEC_50G_RS528;
1615			break;
1616		case FEC_FORCE_MODE_AUTO:
1617			link_params->ext_fec_mode = ETH_EXT_FEC_50G_RS528 |
1618						    ETH_EXT_FEC_50G_BASE_R |
1619						    ETH_EXT_FEC_50G_NONE;
1620			break;
1621		default:
1622			break;
1623		}
1624
1625		break;
1626	case SPEED_100000:
1627		switch (params->fec) {
1628		case FEC_FORCE_MODE_NONE:
1629			link_params->ext_fec_mode = ETH_EXT_FEC_100G_NONE;
1630			break;
1631		case FEC_FORCE_MODE_FIRECODE:
1632			link_params->ext_fec_mode = ETH_EXT_FEC_100G_BASE_R;
1633			break;
1634		case FEC_FORCE_MODE_RS:
1635			link_params->ext_fec_mode = ETH_EXT_FEC_100G_RS528;
1636			break;
1637		case FEC_FORCE_MODE_AUTO:
1638			link_params->ext_fec_mode = ETH_EXT_FEC_100G_RS528 |
1639						    ETH_EXT_FEC_100G_BASE_R |
1640						    ETH_EXT_FEC_100G_NONE;
1641			break;
1642		default:
1643			break;
1644		}
1645
1646		break;
1647	default:
1648		break;
1649	}
1650}
1651
1652static int qed_set_link(struct qed_dev *cdev, struct qed_link_params *params)
1653{
1654	struct qed_mcp_link_params *link_params;
1655	struct qed_mcp_link_speed_params *speed;
1656	const struct qed_mfw_speed_map *map;
1657	struct qed_hwfn *hwfn;
1658	struct qed_ptt *ptt;
1659	int rc;
1660	u32 i;
1661
1662	if (!cdev)
1663		return -ENODEV;
1664
1665	/* The link should be set only once per PF */
1666	hwfn = &cdev->hwfns[0];
1667
1668	/* When VF wants to set link, force it to read the bulletin instead.
1669	 * This mimics the PF behavior, where a noitification [both immediate
1670	 * and possible later] would be generated when changing properties.
1671	 */
1672	if (IS_VF(cdev)) {
1673		qed_schedule_iov(hwfn, QED_IOV_WQ_VF_FORCE_LINK_QUERY_FLAG);
1674		return 0;
1675	}
1676
1677	ptt = qed_ptt_acquire(hwfn);
1678	if (!ptt)
1679		return -EBUSY;
1680
1681	link_params = qed_mcp_get_link_params(hwfn);
1682	if (!link_params)
1683		return -ENODATA;
1684
1685	speed = &link_params->speed;
1686
1687	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_AUTONEG)
1688		speed->autoneg = !!params->autoneg;
1689
1690	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_ADV_SPEEDS) {
1691		speed->advertised_speeds = 0;
1692
1693		for (i = 0; i < ARRAY_SIZE(qed_mfw_legacy_maps); i++) {
1694			map = qed_mfw_legacy_maps + i;
1695
1696			if (linkmode_intersects(params->adv_speeds, map->caps))
1697				speed->advertised_speeds |= map->mfw_val;
1698		}
1699	}
1700
1701	if (params->override_flags & QED_LINK_OVERRIDE_SPEED_FORCED_SPEED)
1702		speed->forced_speed = params->forced_speed;
1703
1704	if (qed_mcp_is_ext_speed_supported(hwfn))
1705		qed_set_ext_speed_params(link_params, params);
1706
1707	if (params->override_flags & QED_LINK_OVERRIDE_PAUSE_CONFIG) {
1708		if (params->pause_config & QED_LINK_PAUSE_AUTONEG_ENABLE)
1709			link_params->pause.autoneg = true;
1710		else
1711			link_params->pause.autoneg = false;
1712		if (params->pause_config & QED_LINK_PAUSE_RX_ENABLE)
1713			link_params->pause.forced_rx = true;
1714		else
1715			link_params->pause.forced_rx = false;
1716		if (params->pause_config & QED_LINK_PAUSE_TX_ENABLE)
1717			link_params->pause.forced_tx = true;
1718		else
1719			link_params->pause.forced_tx = false;
1720	}
1721
1722	if (params->override_flags & QED_LINK_OVERRIDE_LOOPBACK_MODE) {
1723		switch (params->loopback_mode) {
1724		case QED_LINK_LOOPBACK_INT_PHY:
1725			link_params->loopback_mode = ETH_LOOPBACK_INT_PHY;
1726			break;
1727		case QED_LINK_LOOPBACK_EXT_PHY:
1728			link_params->loopback_mode = ETH_LOOPBACK_EXT_PHY;
1729			break;
1730		case QED_LINK_LOOPBACK_EXT:
1731			link_params->loopback_mode = ETH_LOOPBACK_EXT;
1732			break;
1733		case QED_LINK_LOOPBACK_MAC:
1734			link_params->loopback_mode = ETH_LOOPBACK_MAC;
1735			break;
1736		case QED_LINK_LOOPBACK_CNIG_AH_ONLY_0123:
1737			link_params->loopback_mode =
1738				ETH_LOOPBACK_CNIG_AH_ONLY_0123;
1739			break;
1740		case QED_LINK_LOOPBACK_CNIG_AH_ONLY_2301:
1741			link_params->loopback_mode =
1742				ETH_LOOPBACK_CNIG_AH_ONLY_2301;
1743			break;
1744		case QED_LINK_LOOPBACK_PCS_AH_ONLY:
1745			link_params->loopback_mode = ETH_LOOPBACK_PCS_AH_ONLY;
1746			break;
1747		case QED_LINK_LOOPBACK_REVERSE_MAC_AH_ONLY:
1748			link_params->loopback_mode =
1749				ETH_LOOPBACK_REVERSE_MAC_AH_ONLY;
1750			break;
1751		case QED_LINK_LOOPBACK_INT_PHY_FEA_AH_ONLY:
1752			link_params->loopback_mode =
1753				ETH_LOOPBACK_INT_PHY_FEA_AH_ONLY;
1754			break;
1755		default:
1756			link_params->loopback_mode = ETH_LOOPBACK_NONE;
1757			break;
1758		}
1759	}
1760
1761	if (params->override_flags & QED_LINK_OVERRIDE_EEE_CONFIG)
1762		memcpy(&link_params->eee, &params->eee,
1763		       sizeof(link_params->eee));
1764
1765	if (params->override_flags & QED_LINK_OVERRIDE_FEC_CONFIG)
1766		link_params->fec = params->fec;
1767
1768	rc = qed_mcp_set_link(hwfn, ptt, params->link_up);
1769
1770	qed_ptt_release(hwfn, ptt);
1771
1772	return rc;
1773}
1774
1775static int qed_get_port_type(u32 media_type)
1776{
1777	int port_type;
1778
1779	switch (media_type) {
1780	case MEDIA_SFPP_10G_FIBER:
1781	case MEDIA_SFP_1G_FIBER:
1782	case MEDIA_XFP_FIBER:
1783	case MEDIA_MODULE_FIBER:
1784		port_type = PORT_FIBRE;
1785		break;
1786	case MEDIA_DA_TWINAX:
1787		port_type = PORT_DA;
1788		break;
1789	case MEDIA_BASE_T:
1790		port_type = PORT_TP;
1791		break;
1792	case MEDIA_KR:
1793	case MEDIA_NOT_PRESENT:
1794		port_type = PORT_NONE;
1795		break;
1796	case MEDIA_UNSPECIFIED:
1797	default:
1798		port_type = PORT_OTHER;
1799		break;
1800	}
1801	return port_type;
1802}
1803
1804static int qed_get_link_data(struct qed_hwfn *hwfn,
1805			     struct qed_mcp_link_params *params,
1806			     struct qed_mcp_link_state *link,
1807			     struct qed_mcp_link_capabilities *link_caps)
1808{
1809	void *p;
1810
1811	if (!IS_PF(hwfn->cdev)) {
1812		qed_vf_get_link_params(hwfn, params);
1813		qed_vf_get_link_state(hwfn, link);
1814		qed_vf_get_link_caps(hwfn, link_caps);
1815
1816		return 0;
1817	}
1818
1819	p = qed_mcp_get_link_params(hwfn);
1820	if (!p)
1821		return -ENXIO;
1822	memcpy(params, p, sizeof(*params));
1823
1824	p = qed_mcp_get_link_state(hwfn);
1825	if (!p)
1826		return -ENXIO;
1827	memcpy(link, p, sizeof(*link));
1828
1829	p = qed_mcp_get_link_capabilities(hwfn);
1830	if (!p)
1831		return -ENXIO;
1832	memcpy(link_caps, p, sizeof(*link_caps));
1833
1834	return 0;
1835}
1836
1837static void qed_fill_link_capability(struct qed_hwfn *hwfn,
1838				     struct qed_ptt *ptt, u32 capability,
1839				     unsigned long *if_caps)
1840{
1841	u32 media_type, tcvr_state, tcvr_type;
1842	u32 speed_mask, board_cfg;
1843
1844	if (qed_mcp_get_media_type(hwfn, ptt, &media_type))
1845		media_type = MEDIA_UNSPECIFIED;
1846
1847	if (qed_mcp_get_transceiver_data(hwfn, ptt, &tcvr_state, &tcvr_type))
1848		tcvr_type = ETH_TRANSCEIVER_STATE_UNPLUGGED;
1849
1850	if (qed_mcp_trans_speed_mask(hwfn, ptt, &speed_mask))
1851		speed_mask = 0xFFFFFFFF;
1852
1853	if (qed_mcp_get_board_config(hwfn, ptt, &board_cfg))
1854		board_cfg = NVM_CFG1_PORT_PORT_TYPE_UNDEFINED;
1855
1856	DP_VERBOSE(hwfn->cdev, NETIF_MSG_DRV,
1857		   "Media_type = 0x%x tcvr_state = 0x%x tcvr_type = 0x%x speed_mask = 0x%x board_cfg = 0x%x\n",
1858		   media_type, tcvr_state, tcvr_type, speed_mask, board_cfg);
1859
1860	switch (media_type) {
1861	case MEDIA_DA_TWINAX:
1862		phylink_set(if_caps, FIBRE);
1863
1864		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_20G)
1865			phylink_set(if_caps, 20000baseKR2_Full);
1866
1867		/* For DAC media multiple speed capabilities are supported */
1868		capability |= speed_mask;
1869
1870		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G)
1871			phylink_set(if_caps, 1000baseKX_Full);
1872		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G)
1873			phylink_set(if_caps, 10000baseCR_Full);
1874
1875		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_40G)
1876			switch (tcvr_type) {
1877			case ETH_TRANSCEIVER_TYPE_40G_CR4:
1878			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_40G_CR:
1879			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_CR:
1880				phylink_set(if_caps, 40000baseCR4_Full);
1881				break;
1882			default:
1883				break;
1884			}
1885
1886		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_25G)
1887			phylink_set(if_caps, 25000baseCR_Full);
1888		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_50G)
1889			phylink_set(if_caps, 50000baseCR2_Full);
1890
1891		if (capability &
1892		    NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_BB_100G)
1893			switch (tcvr_type) {
1894			case ETH_TRANSCEIVER_TYPE_100G_CR4:
1895			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_CR:
1896				phylink_set(if_caps, 100000baseCR4_Full);
1897				break;
1898			default:
1899				break;
1900			}
1901
1902		break;
1903	case MEDIA_BASE_T:
1904		phylink_set(if_caps, TP);
1905
1906		if (board_cfg & NVM_CFG1_PORT_PORT_TYPE_EXT_PHY) {
1907			if (capability &
1908			    NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G)
1909				phylink_set(if_caps, 1000baseT_Full);
1910			if (capability &
1911			    NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G)
1912				phylink_set(if_caps, 10000baseT_Full);
1913		}
1914
1915		if (board_cfg & NVM_CFG1_PORT_PORT_TYPE_MODULE) {
1916			phylink_set(if_caps, FIBRE);
1917
1918			switch (tcvr_type) {
1919			case ETH_TRANSCEIVER_TYPE_1000BASET:
1920				phylink_set(if_caps, 1000baseT_Full);
1921				break;
1922			case ETH_TRANSCEIVER_TYPE_10G_BASET:
1923				phylink_set(if_caps, 10000baseT_Full);
1924				break;
1925			default:
1926				break;
1927			}
1928		}
1929
1930		break;
1931	case MEDIA_SFP_1G_FIBER:
1932	case MEDIA_SFPP_10G_FIBER:
1933	case MEDIA_XFP_FIBER:
1934	case MEDIA_MODULE_FIBER:
1935		phylink_set(if_caps, FIBRE);
1936		capability |= speed_mask;
1937
1938		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G)
1939			switch (tcvr_type) {
1940			case ETH_TRANSCEIVER_TYPE_1G_LX:
1941			case ETH_TRANSCEIVER_TYPE_1G_SX:
1942			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_1G_10G_SR:
1943			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_1G_10G_LR:
1944				phylink_set(if_caps, 1000baseKX_Full);
1945				break;
1946			default:
1947				break;
1948			}
1949
1950		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G)
1951			switch (tcvr_type) {
1952			case ETH_TRANSCEIVER_TYPE_10G_SR:
1953			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_40G_SR:
1954			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_25G_SR:
1955			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_1G_10G_SR:
1956				phylink_set(if_caps, 10000baseSR_Full);
1957				break;
1958			case ETH_TRANSCEIVER_TYPE_10G_LR:
1959			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_40G_LR:
1960			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_25G_LR:
1961			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_1G_10G_LR:
1962				phylink_set(if_caps, 10000baseLR_Full);
1963				break;
1964			case ETH_TRANSCEIVER_TYPE_10G_LRM:
1965				phylink_set(if_caps, 10000baseLRM_Full);
1966				break;
1967			case ETH_TRANSCEIVER_TYPE_10G_ER:
1968				phylink_set(if_caps, 10000baseR_FEC);
1969				break;
1970			default:
1971				break;
1972			}
1973
1974		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_20G)
1975			phylink_set(if_caps, 20000baseKR2_Full);
1976
1977		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_25G)
1978			switch (tcvr_type) {
1979			case ETH_TRANSCEIVER_TYPE_25G_SR:
1980			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_25G_SR:
1981				phylink_set(if_caps, 25000baseSR_Full);
1982				break;
1983			default:
1984				break;
1985			}
1986
1987		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_40G)
1988			switch (tcvr_type) {
1989			case ETH_TRANSCEIVER_TYPE_40G_LR4:
1990			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_40G_LR:
1991			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_LR:
1992				phylink_set(if_caps, 40000baseLR4_Full);
1993				break;
1994			case ETH_TRANSCEIVER_TYPE_40G_SR4:
1995			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_SR:
1996			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_10G_40G_SR:
1997				phylink_set(if_caps, 40000baseSR4_Full);
1998				break;
1999			default:
2000				break;
2001			}
2002
2003		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_50G)
2004			phylink_set(if_caps, 50000baseKR2_Full);
2005
2006		if (capability &
2007		    NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_BB_100G)
2008			switch (tcvr_type) {
2009			case ETH_TRANSCEIVER_TYPE_100G_SR4:
2010			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_SR:
2011				phylink_set(if_caps, 100000baseSR4_Full);
2012				break;
2013			case ETH_TRANSCEIVER_TYPE_MULTI_RATE_40G_100G_LR:
2014				phylink_set(if_caps, 100000baseLR4_ER4_Full);
2015				break;
2016			default:
2017				break;
2018			}
2019
2020		break;
2021	case MEDIA_KR:
2022		phylink_set(if_caps, Backplane);
2023
2024		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_20G)
2025			phylink_set(if_caps, 20000baseKR2_Full);
2026		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G)
2027			phylink_set(if_caps, 1000baseKX_Full);
2028		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G)
2029			phylink_set(if_caps, 10000baseKR_Full);
2030		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_25G)
2031			phylink_set(if_caps, 25000baseKR_Full);
2032		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_40G)
2033			phylink_set(if_caps, 40000baseKR4_Full);
2034		if (capability & NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_50G)
2035			phylink_set(if_caps, 50000baseKR2_Full);
2036		if (capability &
2037		    NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_BB_100G)
2038			phylink_set(if_caps, 100000baseKR4_Full);
2039
2040		break;
2041	case MEDIA_UNSPECIFIED:
2042	case MEDIA_NOT_PRESENT:
2043	default:
2044		DP_VERBOSE(hwfn->cdev, QED_MSG_DEBUG,
2045			   "Unknown media and transceiver type;\n");
2046		break;
2047	}
2048}
2049
2050static void qed_lp_caps_to_speed_mask(u32 caps, u32 *speed_mask)
2051{
2052	*speed_mask = 0;
2053
2054	if (caps &
2055	    (QED_LINK_PARTNER_SPEED_1G_FD | QED_LINK_PARTNER_SPEED_1G_HD))
2056		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_1G;
2057	if (caps & QED_LINK_PARTNER_SPEED_10G)
2058		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_10G;
2059	if (caps & QED_LINK_PARTNER_SPEED_20G)
2060		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_20G;
2061	if (caps & QED_LINK_PARTNER_SPEED_25G)
2062		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_25G;
2063	if (caps & QED_LINK_PARTNER_SPEED_40G)
2064		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_40G;
2065	if (caps & QED_LINK_PARTNER_SPEED_50G)
2066		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_50G;
2067	if (caps & QED_LINK_PARTNER_SPEED_100G)
2068		*speed_mask |= NVM_CFG1_PORT_DRV_SPEED_CAPABILITY_MASK_BB_100G;
2069}
2070
2071static void qed_fill_link(struct qed_hwfn *hwfn,
2072			  struct qed_ptt *ptt,
2073			  struct qed_link_output *if_link)
2074{
2075	struct qed_mcp_link_capabilities link_caps;
2076	struct qed_mcp_link_params params;
2077	struct qed_mcp_link_state link;
2078	u32 media_type, speed_mask;
2079
2080	memset(if_link, 0, sizeof(*if_link));
2081
2082	/* Prepare source inputs */
2083	if (qed_get_link_data(hwfn, &params, &link, &link_caps)) {
2084		dev_warn(&hwfn->cdev->pdev->dev, "no link data available\n");
2085		return;
2086	}
2087
2088	/* Set the link parameters to pass to protocol driver */
2089	if (link.link_up)
2090		if_link->link_up = true;
2091
2092	if (IS_PF(hwfn->cdev) && qed_mcp_is_ext_speed_supported(hwfn)) {
2093		if (link_caps.default_ext_autoneg)
2094			phylink_set(if_link->supported_caps, Autoneg);
2095
2096		linkmode_copy(if_link->advertised_caps, if_link->supported_caps);
2097
2098		if (params.ext_speed.autoneg)
2099			phylink_set(if_link->advertised_caps, Autoneg);
2100		else
2101			phylink_clear(if_link->advertised_caps, Autoneg);
2102
2103		qed_fill_link_capability(hwfn, ptt,
2104					 params.ext_speed.advertised_speeds,
2105					 if_link->advertised_caps);
2106	} else {
2107		if (link_caps.default_speed_autoneg)
2108			phylink_set(if_link->supported_caps, Autoneg);
2109
2110		linkmode_copy(if_link->advertised_caps, if_link->supported_caps);
2111
2112		if (params.speed.autoneg)
2113			phylink_set(if_link->advertised_caps, Autoneg);
2114		else
2115			phylink_clear(if_link->advertised_caps, Autoneg);
2116	}
2117
2118	if (params.pause.autoneg ||
2119	    (params.pause.forced_rx && params.pause.forced_tx))
2120		phylink_set(if_link->supported_caps, Asym_Pause);
2121	if (params.pause.autoneg || params.pause.forced_rx ||
2122	    params.pause.forced_tx)
2123		phylink_set(if_link->supported_caps, Pause);
2124
2125	if_link->sup_fec = link_caps.fec_default;
2126	if_link->active_fec = params.fec;
2127
2128	/* Fill link advertised capability */
2129	qed_fill_link_capability(hwfn, ptt, params.speed.advertised_speeds,
2130				 if_link->advertised_caps);
2131
2132	/* Fill link supported capability */
2133	qed_fill_link_capability(hwfn, ptt, link_caps.speed_capabilities,
2134				 if_link->supported_caps);
2135
2136	/* Fill partner advertised capability */
2137	qed_lp_caps_to_speed_mask(link.partner_adv_speed, &speed_mask);
2138	qed_fill_link_capability(hwfn, ptt, speed_mask, if_link->lp_caps);
2139
2140	if (link.link_up)
2141		if_link->speed = link.speed;
2142
2143	/* TODO - fill duplex properly */
2144	if_link->duplex = DUPLEX_FULL;
2145	qed_mcp_get_media_type(hwfn, ptt, &media_type);
2146	if_link->port = qed_get_port_type(media_type);
2147
2148	if_link->autoneg = params.speed.autoneg;
2149
2150	if (params.pause.autoneg)
2151		if_link->pause_config |= QED_LINK_PAUSE_AUTONEG_ENABLE;
2152	if (params.pause.forced_rx)
2153		if_link->pause_config |= QED_LINK_PAUSE_RX_ENABLE;
2154	if (params.pause.forced_tx)
2155		if_link->pause_config |= QED_LINK_PAUSE_TX_ENABLE;
2156
2157	if (link.an_complete)
2158		phylink_set(if_link->lp_caps, Autoneg);
2159	if (link.partner_adv_pause)
2160		phylink_set(if_link->lp_caps, Pause);
2161	if (link.partner_adv_pause == QED_LINK_PARTNER_ASYMMETRIC_PAUSE ||
2162	    link.partner_adv_pause == QED_LINK_PARTNER_BOTH_PAUSE)
2163		phylink_set(if_link->lp_caps, Asym_Pause);
2164
2165	if (link_caps.default_eee == QED_MCP_EEE_UNSUPPORTED) {
2166		if_link->eee_supported = false;
2167	} else {
2168		if_link->eee_supported = true;
2169		if_link->eee_active = link.eee_active;
2170		if_link->sup_caps = link_caps.eee_speed_caps;
2171		/* MFW clears adv_caps on eee disable; use configured value */
2172		if_link->eee.adv_caps = link.eee_adv_caps ? link.eee_adv_caps :
2173					params.eee.adv_caps;
2174		if_link->eee.lp_adv_caps = link.eee_lp_adv_caps;
2175		if_link->eee.enable = params.eee.enable;
2176		if_link->eee.tx_lpi_enable = params.eee.tx_lpi_enable;
2177		if_link->eee.tx_lpi_timer = params.eee.tx_lpi_timer;
2178	}
2179}
2180
2181static void qed_get_current_link(struct qed_dev *cdev,
2182				 struct qed_link_output *if_link)
2183{
2184	struct qed_hwfn *hwfn;
2185	struct qed_ptt *ptt;
2186	int i;
2187
2188	hwfn = &cdev->hwfns[0];
2189	if (IS_PF(cdev)) {
2190		ptt = qed_ptt_acquire(hwfn);
2191		if (ptt) {
2192			qed_fill_link(hwfn, ptt, if_link);
2193			qed_ptt_release(hwfn, ptt);
2194		} else {
2195			DP_NOTICE(hwfn, "Failed to fill link; No PTT\n");
2196		}
2197	} else {
2198		qed_fill_link(hwfn, NULL, if_link);
2199	}
2200
2201	for_each_hwfn(cdev, i)
2202		qed_inform_vf_link_state(&cdev->hwfns[i]);
2203}
2204
2205void qed_link_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt)
2206{
2207	void *cookie = hwfn->cdev->ops_cookie;
2208	struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common;
2209	struct qed_link_output if_link;
2210
2211	qed_fill_link(hwfn, ptt, &if_link);
2212	qed_inform_vf_link_state(hwfn);
2213
2214	if (IS_LEAD_HWFN(hwfn) && cookie)
2215		op->link_update(cookie, &if_link);
2216}
2217
2218void qed_bw_update(struct qed_hwfn *hwfn, struct qed_ptt *ptt)
2219{
2220	void *cookie = hwfn->cdev->ops_cookie;
2221	struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common;
2222
2223	if (IS_LEAD_HWFN(hwfn) && cookie && op && op->bw_update)
2224		op->bw_update(cookie);
2225}
2226
2227static int qed_drain(struct qed_dev *cdev)
2228{
2229	struct qed_hwfn *hwfn;
2230	struct qed_ptt *ptt;
2231	int i, rc;
2232
2233	if (IS_VF(cdev))
2234		return 0;
2235
2236	for_each_hwfn(cdev, i) {
2237		hwfn = &cdev->hwfns[i];
2238		ptt = qed_ptt_acquire(hwfn);
2239		if (!ptt) {
2240			DP_NOTICE(hwfn, "Failed to drain NIG; No PTT\n");
2241			return -EBUSY;
2242		}
2243		rc = qed_mcp_drain(hwfn, ptt);
2244		qed_ptt_release(hwfn, ptt);
2245		if (rc)
2246			return rc;
2247	}
2248
2249	return 0;
2250}
2251
2252static u32 qed_nvm_flash_image_access_crc(struct qed_dev *cdev,
2253					  struct qed_nvm_image_att *nvm_image,
2254					  u32 *crc)
2255{
2256	u8 *buf = NULL;
2257	int rc;
2258
2259	/* Allocate a buffer for holding the nvram image */
2260	buf = kzalloc(nvm_image->length, GFP_KERNEL);
2261	if (!buf)
2262		return -ENOMEM;
2263
2264	/* Read image into buffer */
2265	rc = qed_mcp_nvm_read(cdev, nvm_image->start_addr,
2266			      buf, nvm_image->length);
2267	if (rc) {
2268		DP_ERR(cdev, "Failed reading image from nvm\n");
2269		goto out;
2270	}
2271
2272	/* Convert the buffer into big-endian format (excluding the
2273	 * closing 4 bytes of CRC).
2274	 */
2275	cpu_to_be32_array((__force __be32 *)buf, (const u32 *)buf,
2276			  DIV_ROUND_UP(nvm_image->length - 4, 4));
2277
2278	/* Calc CRC for the "actual" image buffer, i.e. not including
2279	 * the last 4 CRC bytes.
2280	 */
2281	*crc = ~crc32(~0U, buf, nvm_image->length - 4);
2282	*crc = (__force u32)cpu_to_be32p(crc);
2283
2284out:
2285	kfree(buf);
2286
2287	return rc;
2288}
2289
2290/* Binary file format -
2291 *     /----------------------------------------------------------------------\
2292 * 0B  |                       0x4 [command index]                            |
2293 * 4B  | image_type     | Options        |  Number of register settings       |
2294 * 8B  |                       Value                                          |
2295 * 12B |                       Mask                                           |
2296 * 16B |                       Offset                                         |
2297 *     \----------------------------------------------------------------------/
2298 * There can be several Value-Mask-Offset sets as specified by 'Number of...'.
2299 * Options - 0'b - Calculate & Update CRC for image
2300 */
2301static int qed_nvm_flash_image_access(struct qed_dev *cdev, const u8 **data,
2302				      bool *check_resp)
2303{
2304	struct qed_nvm_image_att nvm_image;
2305	struct qed_hwfn *p_hwfn;
2306	bool is_crc = false;
2307	u32 image_type;
2308	int rc = 0, i;
2309	u16 len;
2310
2311	*data += 4;
2312	image_type = **data;
2313	p_hwfn = QED_LEADING_HWFN(cdev);
2314	for (i = 0; i < p_hwfn->nvm_info.num_images; i++)
2315		if (image_type == p_hwfn->nvm_info.image_att[i].image_type)
2316			break;
2317	if (i == p_hwfn->nvm_info.num_images) {
2318		DP_ERR(cdev, "Failed to find nvram image of type %08x\n",
2319		       image_type);
2320		return -ENOENT;
2321	}
2322
2323	nvm_image.start_addr = p_hwfn->nvm_info.image_att[i].nvm_start_addr;
2324	nvm_image.length = p_hwfn->nvm_info.image_att[i].len;
2325
2326	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2327		   "Read image %02x; type = %08x; NVM [%08x,...,%08x]\n",
2328		   **data, image_type, nvm_image.start_addr,
2329		   nvm_image.start_addr + nvm_image.length - 1);
2330	(*data)++;
2331	is_crc = !!(**data & BIT(0));
2332	(*data)++;
2333	len = *((u16 *)*data);
2334	*data += 2;
2335	if (is_crc) {
2336		u32 crc = 0;
2337
2338		rc = qed_nvm_flash_image_access_crc(cdev, &nvm_image, &crc);
2339		if (rc) {
2340			DP_ERR(cdev, "Failed calculating CRC, rc = %d\n", rc);
2341			goto exit;
2342		}
2343
2344		rc = qed_mcp_nvm_write(cdev, QED_NVM_WRITE_NVRAM,
2345				       (nvm_image.start_addr +
2346					nvm_image.length - 4), (u8 *)&crc, 4);
2347		if (rc)
2348			DP_ERR(cdev, "Failed writing to %08x, rc = %d\n",
2349			       nvm_image.start_addr + nvm_image.length - 4, rc);
2350		goto exit;
2351	}
2352
2353	/* Iterate over the values for setting */
2354	while (len) {
2355		u32 offset, mask, value, cur_value;
2356		u8 buf[4];
2357
2358		value = *((u32 *)*data);
2359		*data += 4;
2360		mask = *((u32 *)*data);
2361		*data += 4;
2362		offset = *((u32 *)*data);
2363		*data += 4;
2364
2365		rc = qed_mcp_nvm_read(cdev, nvm_image.start_addr + offset, buf,
2366				      4);
2367		if (rc) {
2368			DP_ERR(cdev, "Failed reading from %08x\n",
2369			       nvm_image.start_addr + offset);
2370			goto exit;
2371		}
2372
2373		cur_value = le32_to_cpu(*((__le32 *)buf));
2374		DP_VERBOSE(cdev, NETIF_MSG_DRV,
2375			   "NVM %08x: %08x -> %08x [Value %08x Mask %08x]\n",
2376			   nvm_image.start_addr + offset, cur_value,
2377			   (cur_value & ~mask) | (value & mask), value, mask);
2378		value = (value & mask) | (cur_value & ~mask);
2379		rc = qed_mcp_nvm_write(cdev, QED_NVM_WRITE_NVRAM,
2380				       nvm_image.start_addr + offset,
2381				       (u8 *)&value, 4);
2382		if (rc) {
2383			DP_ERR(cdev, "Failed writing to %08x\n",
2384			       nvm_image.start_addr + offset);
2385			goto exit;
2386		}
2387
2388		len--;
2389	}
2390exit:
2391	return rc;
2392}
2393
2394/* Binary file format -
2395 *     /----------------------------------------------------------------------\
2396 * 0B  |                       0x3 [command index]                            |
2397 * 4B  | b'0: check_response?   | b'1-31  reserved                            |
2398 * 8B  | File-type |                   reserved                               |
2399 * 12B |                    Image length in bytes                             |
2400 *     \----------------------------------------------------------------------/
2401 *     Start a new file of the provided type
2402 */
2403static int qed_nvm_flash_image_file_start(struct qed_dev *cdev,
2404					  const u8 **data, bool *check_resp)
2405{
2406	u32 file_type, file_size = 0;
2407	int rc;
2408
2409	*data += 4;
2410	*check_resp = !!(**data & BIT(0));
2411	*data += 4;
2412	file_type = **data;
2413
2414	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2415		   "About to start a new file of type %02x\n", file_type);
2416	if (file_type == DRV_MB_PARAM_NVM_PUT_FILE_BEGIN_MBI) {
2417		*data += 4;
2418		file_size = *((u32 *)(*data));
2419	}
2420
2421	rc = qed_mcp_nvm_write(cdev, QED_PUT_FILE_BEGIN, file_type,
2422			       (u8 *)(&file_size), 4);
2423	*data += 4;
2424
2425	return rc;
2426}
2427
2428/* Binary file format -
2429 *     /----------------------------------------------------------------------\
2430 * 0B  |                       0x2 [command index]                            |
2431 * 4B  |                       Length in bytes                                |
2432 * 8B  | b'0: check_response?   | b'1-31  reserved                            |
2433 * 12B |                       Offset in bytes                                |
2434 * 16B |                       Data ...                                       |
2435 *     \----------------------------------------------------------------------/
2436 *     Write data as part of a file that was previously started. Data should be
2437 *     of length equal to that provided in the message
2438 */
2439static int qed_nvm_flash_image_file_data(struct qed_dev *cdev,
2440					 const u8 **data, bool *check_resp)
2441{
2442	u32 offset, len;
2443	int rc;
2444
2445	*data += 4;
2446	len = *((u32 *)(*data));
2447	*data += 4;
2448	*check_resp = !!(**data & BIT(0));
2449	*data += 4;
2450	offset = *((u32 *)(*data));
2451	*data += 4;
2452
2453	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2454		   "About to write File-data: %08x bytes to offset %08x\n",
2455		   len, offset);
2456
2457	rc = qed_mcp_nvm_write(cdev, QED_PUT_FILE_DATA, offset,
2458			       (char *)(*data), len);
2459	*data += len;
2460
2461	return rc;
2462}
2463
2464/* Binary file format [General header] -
2465 *     /----------------------------------------------------------------------\
2466 * 0B  |                       QED_NVM_SIGNATURE                              |
2467 * 4B  |                       Length in bytes                                |
2468 * 8B  | Highest command in this batchfile |          Reserved                |
2469 *     \----------------------------------------------------------------------/
2470 */
2471static int qed_nvm_flash_image_validate(struct qed_dev *cdev,
2472					const struct firmware *image,
2473					const u8 **data)
2474{
2475	u32 signature, len;
2476
2477	/* Check minimum size */
2478	if (image->size < 12) {
2479		DP_ERR(cdev, "Image is too short [%08x]\n", (u32)image->size);
2480		return -EINVAL;
2481	}
2482
2483	/* Check signature */
2484	signature = *((u32 *)(*data));
2485	if (signature != QED_NVM_SIGNATURE) {
2486		DP_ERR(cdev, "Wrong signature '%08x'\n", signature);
2487		return -EINVAL;
2488	}
2489
2490	*data += 4;
2491	/* Validate internal size equals the image-size */
2492	len = *((u32 *)(*data));
2493	if (len != image->size) {
2494		DP_ERR(cdev, "Size mismatch: internal = %08x image = %08x\n",
2495		       len, (u32)image->size);
2496		return -EINVAL;
2497	}
2498
2499	*data += 4;
2500	/* Make sure driver familiar with all commands necessary for this */
2501	if (*((u16 *)(*data)) >= QED_NVM_FLASH_CMD_NVM_MAX) {
2502		DP_ERR(cdev, "File contains unsupported commands [Need %04x]\n",
2503		       *((u16 *)(*data)));
2504		return -EINVAL;
2505	}
2506
2507	*data += 4;
2508
2509	return 0;
2510}
2511
2512/* Binary file format -
2513 *     /----------------------------------------------------------------------\
2514 * 0B  |                       0x5 [command index]                            |
2515 * 4B  | Number of config attributes     |          Reserved                  |
2516 * 4B  | Config ID                       | Entity ID      | Length            |
2517 * 4B  | Value                                                                |
2518 *     |                                                                      |
2519 *     \----------------------------------------------------------------------/
2520 * There can be several cfg_id-entity_id-Length-Value sets as specified by
2521 * 'Number of config attributes'.
2522 *
2523 * The API parses config attributes from the user provided buffer and flashes
2524 * them to the respective NVM path using Management FW inerface.
2525 */
2526static int qed_nvm_flash_cfg_write(struct qed_dev *cdev, const u8 **data)
2527{
2528	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2529	u8 entity_id, len, buf[32];
2530	bool need_nvm_init = true;
2531	struct qed_ptt *ptt;
2532	u16 cfg_id, count;
2533	int rc = 0, i;
2534	u32 flags;
2535
2536	ptt = qed_ptt_acquire(hwfn);
2537	if (!ptt)
2538		return -EAGAIN;
2539
2540	/* NVM CFG ID attribute header */
2541	*data += 4;
2542	count = *((u16 *)*data);
2543	*data += 4;
2544
2545	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2546		   "Read config ids: num_attrs = %0d\n", count);
2547	/* NVM CFG ID attributes. Start loop index from 1 to avoid additional
2548	 * arithmetic operations in the implementation.
2549	 */
2550	for (i = 1; i <= count; i++) {
2551		cfg_id = *((u16 *)*data);
2552		*data += 2;
2553		entity_id = **data;
2554		(*data)++;
2555		len = **data;
2556		(*data)++;
2557		memcpy(buf, *data, len);
2558		*data += len;
2559
2560		flags = 0;
2561		if (need_nvm_init) {
2562			flags |= QED_NVM_CFG_OPTION_INIT;
2563			need_nvm_init = false;
2564		}
2565
2566		/* Commit to flash and free the resources */
2567		if (!(i % QED_NVM_CFG_MAX_ATTRS) || i == count) {
2568			flags |= QED_NVM_CFG_OPTION_COMMIT |
2569				 QED_NVM_CFG_OPTION_FREE;
2570			need_nvm_init = true;
2571		}
2572
2573		if (entity_id)
2574			flags |= QED_NVM_CFG_OPTION_ENTITY_SEL;
2575
2576		DP_VERBOSE(cdev, NETIF_MSG_DRV,
2577			   "cfg_id = %d entity = %d len = %d\n", cfg_id,
2578			   entity_id, len);
2579		rc = qed_mcp_nvm_set_cfg(hwfn, ptt, cfg_id, entity_id, flags,
2580					 buf, len);
2581		if (rc) {
2582			DP_ERR(cdev, "Error %d configuring %d\n", rc, cfg_id);
2583			break;
2584		}
2585	}
2586
2587	qed_ptt_release(hwfn, ptt);
2588
2589	return rc;
2590}
2591
2592#define QED_MAX_NVM_BUF_LEN	32
2593static int qed_nvm_flash_cfg_len(struct qed_dev *cdev, u32 cmd)
2594{
2595	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2596	u8 buf[QED_MAX_NVM_BUF_LEN];
2597	struct qed_ptt *ptt;
2598	u32 len;
2599	int rc;
2600
2601	ptt = qed_ptt_acquire(hwfn);
2602	if (!ptt)
2603		return QED_MAX_NVM_BUF_LEN;
2604
2605	rc = qed_mcp_nvm_get_cfg(hwfn, ptt, cmd, 0, QED_NVM_CFG_GET_FLAGS, buf,
2606				 &len);
2607	if (rc || !len) {
2608		DP_ERR(cdev, "Error %d reading %d\n", rc, cmd);
2609		len = QED_MAX_NVM_BUF_LEN;
2610	}
2611
2612	qed_ptt_release(hwfn, ptt);
2613
2614	return len;
2615}
2616
2617static int qed_nvm_flash_cfg_read(struct qed_dev *cdev, u8 **data,
2618				  u32 cmd, u32 entity_id)
2619{
2620	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2621	struct qed_ptt *ptt;
2622	u32 flags, len;
2623	int rc = 0;
2624
2625	ptt = qed_ptt_acquire(hwfn);
2626	if (!ptt)
2627		return -EAGAIN;
2628
2629	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2630		   "Read config cmd = %d entity id %d\n", cmd, entity_id);
2631	flags = entity_id ? QED_NVM_CFG_GET_PF_FLAGS : QED_NVM_CFG_GET_FLAGS;
2632	rc = qed_mcp_nvm_get_cfg(hwfn, ptt, cmd, entity_id, flags, *data, &len);
2633	if (rc)
2634		DP_ERR(cdev, "Error %d reading %d\n", rc, cmd);
2635
2636	qed_ptt_release(hwfn, ptt);
2637
2638	return rc;
2639}
2640
2641static int qed_nvm_flash(struct qed_dev *cdev, const char *name)
2642{
2643	const struct firmware *image;
2644	const u8 *data, *data_end;
2645	u32 cmd_type;
2646	int rc;
2647
2648	rc = request_firmware(&image, name, &cdev->pdev->dev);
2649	if (rc) {
2650		DP_ERR(cdev, "Failed to find '%s'\n", name);
2651		return rc;
2652	}
2653
2654	DP_VERBOSE(cdev, NETIF_MSG_DRV,
2655		   "Flashing '%s' - firmware's data at %p, size is %08x\n",
2656		   name, image->data, (u32)image->size);
2657	data = image->data;
2658	data_end = data + image->size;
2659
2660	rc = qed_nvm_flash_image_validate(cdev, image, &data);
2661	if (rc)
2662		goto exit;
2663
2664	while (data < data_end) {
2665		bool check_resp = false;
2666
2667		/* Parse the actual command */
2668		cmd_type = *((u32 *)data);
2669		switch (cmd_type) {
2670		case QED_NVM_FLASH_CMD_FILE_DATA:
2671			rc = qed_nvm_flash_image_file_data(cdev, &data,
2672							   &check_resp);
2673			break;
2674		case QED_NVM_FLASH_CMD_FILE_START:
2675			rc = qed_nvm_flash_image_file_start(cdev, &data,
2676							    &check_resp);
2677			break;
2678		case QED_NVM_FLASH_CMD_NVM_CHANGE:
2679			rc = qed_nvm_flash_image_access(cdev, &data,
2680							&check_resp);
2681			break;
2682		case QED_NVM_FLASH_CMD_NVM_CFG_ID:
2683			rc = qed_nvm_flash_cfg_write(cdev, &data);
2684			break;
2685		default:
2686			DP_ERR(cdev, "Unknown command %08x\n", cmd_type);
2687			rc = -EINVAL;
2688			goto exit;
2689		}
2690
2691		if (rc) {
2692			DP_ERR(cdev, "Command %08x failed\n", cmd_type);
2693			goto exit;
2694		}
2695
2696		/* Check response if needed */
2697		if (check_resp) {
2698			u32 mcp_response = 0;
2699
2700			if (qed_mcp_nvm_resp(cdev, (u8 *)&mcp_response)) {
2701				DP_ERR(cdev, "Failed getting MCP response\n");
2702				rc = -EINVAL;
2703				goto exit;
2704			}
2705
2706			switch (mcp_response & FW_MSG_CODE_MASK) {
2707			case FW_MSG_CODE_OK:
2708			case FW_MSG_CODE_NVM_OK:
2709			case FW_MSG_CODE_NVM_PUT_FILE_FINISH_OK:
2710			case FW_MSG_CODE_PHY_OK:
2711				break;
2712			default:
2713				DP_ERR(cdev, "MFW returns error: %08x\n",
2714				       mcp_response);
2715				rc = -EINVAL;
2716				goto exit;
2717			}
2718		}
2719	}
2720
2721exit:
2722	release_firmware(image);
2723
2724	return rc;
2725}
2726
2727static int qed_nvm_get_image(struct qed_dev *cdev, enum qed_nvm_images type,
2728			     u8 *buf, u16 len)
2729{
2730	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2731
2732	return qed_mcp_get_nvm_image(hwfn, type, buf, len);
2733}
2734
2735void qed_schedule_recovery_handler(struct qed_hwfn *p_hwfn)
2736{
2737	struct qed_common_cb_ops *ops = p_hwfn->cdev->protocol_ops.common;
2738	void *cookie = p_hwfn->cdev->ops_cookie;
2739
2740	if (ops && ops->schedule_recovery_handler)
2741		ops->schedule_recovery_handler(cookie);
2742}
2743
2744static const char * const qed_hw_err_type_descr[] = {
2745	[QED_HW_ERR_FAN_FAIL]		= "Fan Failure",
2746	[QED_HW_ERR_MFW_RESP_FAIL]	= "MFW Response Failure",
2747	[QED_HW_ERR_HW_ATTN]		= "HW Attention",
2748	[QED_HW_ERR_DMAE_FAIL]		= "DMAE Failure",
2749	[QED_HW_ERR_RAMROD_FAIL]	= "Ramrod Failure",
2750	[QED_HW_ERR_FW_ASSERT]		= "FW Assertion",
2751	[QED_HW_ERR_LAST]		= "Unknown",
2752};
2753
2754void qed_hw_error_occurred(struct qed_hwfn *p_hwfn,
2755			   enum qed_hw_err_type err_type)
2756{
2757	struct qed_common_cb_ops *ops = p_hwfn->cdev->protocol_ops.common;
2758	void *cookie = p_hwfn->cdev->ops_cookie;
2759	const char *err_str;
2760
2761	if (err_type > QED_HW_ERR_LAST)
2762		err_type = QED_HW_ERR_LAST;
2763	err_str = qed_hw_err_type_descr[err_type];
2764
2765	DP_NOTICE(p_hwfn, "HW error occurred [%s]\n", err_str);
2766
2767	/* Call the HW error handler of the protocol driver.
2768	 * If it is not available - perform a minimal handling of preventing
2769	 * HW attentions from being reasserted.
2770	 */
2771	if (ops && ops->schedule_hw_err_handler)
2772		ops->schedule_hw_err_handler(cookie, err_type);
2773	else
2774		qed_int_attn_clr_enable(p_hwfn->cdev, true);
2775}
2776
2777static int qed_set_coalesce(struct qed_dev *cdev, u16 rx_coal, u16 tx_coal,
2778			    void *handle)
2779{
2780		return qed_set_queue_coalesce(rx_coal, tx_coal, handle);
2781}
2782
2783static int qed_set_led(struct qed_dev *cdev, enum qed_led_mode mode)
2784{
2785	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2786	struct qed_ptt *ptt;
2787	int status = 0;
2788
2789	ptt = qed_ptt_acquire(hwfn);
2790	if (!ptt)
2791		return -EAGAIN;
2792
2793	status = qed_mcp_set_led(hwfn, ptt, mode);
2794
2795	qed_ptt_release(hwfn, ptt);
2796
2797	return status;
2798}
2799
2800int qed_recovery_process(struct qed_dev *cdev)
2801{
2802	struct qed_hwfn *p_hwfn = QED_LEADING_HWFN(cdev);
2803	struct qed_ptt *p_ptt;
2804	int rc = 0;
2805
2806	p_ptt = qed_ptt_acquire(p_hwfn);
2807	if (!p_ptt)
2808		return -EAGAIN;
2809
2810	rc = qed_start_recovery_process(p_hwfn, p_ptt);
2811
2812	qed_ptt_release(p_hwfn, p_ptt);
2813
2814	return rc;
2815}
2816
2817static int qed_update_wol(struct qed_dev *cdev, bool enabled)
2818{
2819	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2820	struct qed_ptt *ptt;
2821	int rc = 0;
2822
2823	if (IS_VF(cdev))
2824		return 0;
2825
2826	ptt = qed_ptt_acquire(hwfn);
2827	if (!ptt)
2828		return -EAGAIN;
2829
2830	rc = qed_mcp_ov_update_wol(hwfn, ptt, enabled ? QED_OV_WOL_ENABLED
2831				   : QED_OV_WOL_DISABLED);
2832	if (rc)
2833		goto out;
2834	rc = qed_mcp_ov_update_current_config(hwfn, ptt, QED_OV_CLIENT_DRV);
2835
2836out:
2837	qed_ptt_release(hwfn, ptt);
2838	return rc;
2839}
2840
2841static int qed_update_drv_state(struct qed_dev *cdev, bool active)
2842{
2843	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2844	struct qed_ptt *ptt;
2845	int status = 0;
2846
2847	if (IS_VF(cdev))
2848		return 0;
2849
2850	ptt = qed_ptt_acquire(hwfn);
2851	if (!ptt)
2852		return -EAGAIN;
2853
2854	status = qed_mcp_ov_update_driver_state(hwfn, ptt, active ?
2855						QED_OV_DRIVER_STATE_ACTIVE :
2856						QED_OV_DRIVER_STATE_DISABLED);
2857
2858	qed_ptt_release(hwfn, ptt);
2859
2860	return status;
2861}
2862
2863static int qed_update_mac(struct qed_dev *cdev, const u8 *mac)
2864{
2865	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2866	struct qed_ptt *ptt;
2867	int status = 0;
2868
2869	if (IS_VF(cdev))
2870		return 0;
2871
2872	ptt = qed_ptt_acquire(hwfn);
2873	if (!ptt)
2874		return -EAGAIN;
2875
2876	status = qed_mcp_ov_update_mac(hwfn, ptt, mac);
2877	if (status)
2878		goto out;
2879
2880	status = qed_mcp_ov_update_current_config(hwfn, ptt, QED_OV_CLIENT_DRV);
2881
2882out:
2883	qed_ptt_release(hwfn, ptt);
2884	return status;
2885}
2886
2887static int qed_update_mtu(struct qed_dev *cdev, u16 mtu)
2888{
2889	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2890	struct qed_ptt *ptt;
2891	int status = 0;
2892
2893	if (IS_VF(cdev))
2894		return 0;
2895
2896	ptt = qed_ptt_acquire(hwfn);
2897	if (!ptt)
2898		return -EAGAIN;
2899
2900	status = qed_mcp_ov_update_mtu(hwfn, ptt, mtu);
2901	if (status)
2902		goto out;
2903
2904	status = qed_mcp_ov_update_current_config(hwfn, ptt, QED_OV_CLIENT_DRV);
2905
2906out:
2907	qed_ptt_release(hwfn, ptt);
2908	return status;
2909}
2910
2911static int
2912qed_get_sb_info(struct qed_dev *cdev, struct qed_sb_info *sb,
2913		u16 qid, struct qed_sb_info_dbg *sb_dbg)
2914{
2915	struct qed_hwfn *hwfn = &cdev->hwfns[qid % cdev->num_hwfns];
2916	struct qed_ptt *ptt;
2917	int rc;
2918
2919	if (IS_VF(cdev))
2920		return -EINVAL;
2921
2922	ptt = qed_ptt_acquire(hwfn);
2923	if (!ptt) {
2924		DP_NOTICE(hwfn, "Can't acquire PTT\n");
2925		return -EAGAIN;
2926	}
2927
2928	memset(sb_dbg, 0, sizeof(*sb_dbg));
2929	rc = qed_int_get_sb_dbg(hwfn, ptt, sb, sb_dbg);
2930
2931	qed_ptt_release(hwfn, ptt);
2932	return rc;
2933}
2934
2935static int qed_read_module_eeprom(struct qed_dev *cdev, char *buf,
2936				  u8 dev_addr, u32 offset, u32 len)
2937{
2938	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2939	struct qed_ptt *ptt;
2940	int rc = 0;
2941
2942	if (IS_VF(cdev))
2943		return 0;
2944
2945	ptt = qed_ptt_acquire(hwfn);
2946	if (!ptt)
2947		return -EAGAIN;
2948
2949	rc = qed_mcp_phy_sfp_read(hwfn, ptt, MFW_PORT(hwfn), dev_addr,
2950				  offset, len, buf);
2951
2952	qed_ptt_release(hwfn, ptt);
2953
2954	return rc;
2955}
2956
2957static int qed_set_grc_config(struct qed_dev *cdev, u32 cfg_id, u32 val)
2958{
2959	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
2960	struct qed_ptt *ptt;
2961	int rc = 0;
2962
2963	if (IS_VF(cdev))
2964		return 0;
2965
2966	ptt = qed_ptt_acquire(hwfn);
2967	if (!ptt)
2968		return -EAGAIN;
2969
2970	rc = qed_dbg_grc_config(hwfn, cfg_id, val);
2971
2972	qed_ptt_release(hwfn, ptt);
2973
2974	return rc;
2975}
2976
2977static __printf(2, 3) void qed_mfw_report(struct qed_dev *cdev, char *fmt, ...)
2978{
2979	char buf[QED_MFW_REPORT_STR_SIZE];
2980	struct qed_hwfn *p_hwfn;
2981	struct qed_ptt *p_ptt;
2982	va_list vl;
2983
2984	va_start(vl, fmt);
2985	vsnprintf(buf, QED_MFW_REPORT_STR_SIZE, fmt, vl);
2986	va_end(vl);
2987
2988	if (IS_PF(cdev)) {
2989		p_hwfn = QED_LEADING_HWFN(cdev);
2990		p_ptt = qed_ptt_acquire(p_hwfn);
2991		if (p_ptt) {
2992			qed_mcp_send_raw_debug_data(p_hwfn, p_ptt, buf, strlen(buf));
2993			qed_ptt_release(p_hwfn, p_ptt);
2994		}
2995	}
2996}
2997
2998static u8 qed_get_affin_hwfn_idx(struct qed_dev *cdev)
2999{
3000	return QED_AFFIN_HWFN_IDX(cdev);
3001}
3002
3003static int qed_get_esl_status(struct qed_dev *cdev, bool *esl_active)
3004{
3005	struct qed_hwfn *hwfn = QED_LEADING_HWFN(cdev);
3006	struct qed_ptt *ptt;
3007	int rc = 0;
3008
3009	*esl_active = false;
3010
3011	if (IS_VF(cdev))
3012		return 0;
3013
3014	ptt = qed_ptt_acquire(hwfn);
3015	if (!ptt)
3016		return -EAGAIN;
3017
3018	rc = qed_mcp_get_esl_status(hwfn, ptt, esl_active);
3019
3020	qed_ptt_release(hwfn, ptt);
3021
3022	return rc;
3023}
3024
3025static struct qed_selftest_ops qed_selftest_ops_pass = {
3026	.selftest_memory = &qed_selftest_memory,
3027	.selftest_interrupt = &qed_selftest_interrupt,
3028	.selftest_register = &qed_selftest_register,
3029	.selftest_clock = &qed_selftest_clock,
3030	.selftest_nvram = &qed_selftest_nvram,
3031};
3032
3033const struct qed_common_ops qed_common_ops_pass = {
3034	.selftest = &qed_selftest_ops_pass,
3035	.probe = &qed_probe,
3036	.remove = &qed_remove,
3037	.set_power_state = &qed_set_power_state,
3038	.set_name = &qed_set_name,
3039	.update_pf_params = &qed_update_pf_params,
3040	.slowpath_start = &qed_slowpath_start,
3041	.slowpath_stop = &qed_slowpath_stop,
3042	.set_fp_int = &qed_set_int_fp,
3043	.get_fp_int = &qed_get_int_fp,
3044	.sb_init = &qed_sb_init,
3045	.sb_release = &qed_sb_release,
3046	.simd_handler_config = &qed_simd_handler_config,
3047	.simd_handler_clean = &qed_simd_handler_clean,
3048	.dbg_grc = &qed_dbg_grc,
3049	.dbg_grc_size = &qed_dbg_grc_size,
3050	.can_link_change = &qed_can_link_change,
3051	.set_link = &qed_set_link,
3052	.get_link = &qed_get_current_link,
3053	.drain = &qed_drain,
3054	.update_msglvl = &qed_init_dp,
3055	.devlink_register = qed_devlink_register,
3056	.devlink_unregister = qed_devlink_unregister,
3057	.report_fatal_error = qed_report_fatal_error,
3058	.dbg_all_data = &qed_dbg_all_data,
3059	.dbg_all_data_size = &qed_dbg_all_data_size,
3060	.chain_alloc = &qed_chain_alloc,
3061	.chain_free = &qed_chain_free,
3062	.nvm_flash = &qed_nvm_flash,
3063	.nvm_get_image = &qed_nvm_get_image,
3064	.set_coalesce = &qed_set_coalesce,
3065	.set_led = &qed_set_led,
3066	.recovery_process = &qed_recovery_process,
3067	.recovery_prolog = &qed_recovery_prolog,
3068	.attn_clr_enable = &qed_int_attn_clr_enable,
3069	.update_drv_state = &qed_update_drv_state,
3070	.update_mac = &qed_update_mac,
3071	.update_mtu = &qed_update_mtu,
3072	.update_wol = &qed_update_wol,
3073	.db_recovery_add = &qed_db_recovery_add,
3074	.db_recovery_del = &qed_db_recovery_del,
3075	.read_module_eeprom = &qed_read_module_eeprom,
3076	.get_affin_hwfn_idx = &qed_get_affin_hwfn_idx,
3077	.read_nvm_cfg = &qed_nvm_flash_cfg_read,
3078	.read_nvm_cfg_len = &qed_nvm_flash_cfg_len,
3079	.set_grc_config = &qed_set_grc_config,
3080	.mfw_report = &qed_mfw_report,
3081	.get_sb_info = &qed_get_sb_info,
3082	.get_esl_status = &qed_get_esl_status,
3083};
3084
3085void qed_get_protocol_stats(struct qed_dev *cdev,
3086			    enum qed_mcp_protocol_type type,
3087			    union qed_mcp_protocol_stats *stats)
3088{
3089	struct qed_eth_stats eth_stats;
3090
3091	memset(stats, 0, sizeof(*stats));
3092
3093	switch (type) {
3094	case QED_MCP_LAN_STATS:
3095		qed_get_vport_stats_context(cdev, &eth_stats, true);
3096		stats->lan_stats.ucast_rx_pkts =
3097					eth_stats.common.rx_ucast_pkts;
3098		stats->lan_stats.ucast_tx_pkts =
3099					eth_stats.common.tx_ucast_pkts;
3100		stats->lan_stats.fcs_err = -1;
3101		break;
3102	case QED_MCP_FCOE_STATS:
3103		qed_get_protocol_stats_fcoe(cdev, &stats->fcoe_stats, true);
3104		break;
3105	case QED_MCP_ISCSI_STATS:
3106		qed_get_protocol_stats_iscsi(cdev, &stats->iscsi_stats, true);
3107		break;
3108	default:
3109		DP_VERBOSE(cdev, QED_MSG_SP,
3110			   "Invalid protocol type = %d\n", type);
3111		return;
3112	}
3113}
3114
3115int qed_mfw_tlv_req(struct qed_hwfn *hwfn)
3116{
3117	DP_VERBOSE(hwfn->cdev, NETIF_MSG_DRV,
3118		   "Scheduling slowpath task [Flag: %d]\n",
3119		   QED_SLOWPATH_MFW_TLV_REQ);
3120	/* Memory barrier for setting atomic bit */
3121	smp_mb__before_atomic();
3122	set_bit(QED_SLOWPATH_MFW_TLV_REQ, &hwfn->slowpath_task_flags);
3123	/* Memory barrier after setting atomic bit */
3124	smp_mb__after_atomic();
3125	queue_delayed_work(hwfn->slowpath_wq, &hwfn->slowpath_task, 0);
3126
3127	return 0;
3128}
3129
3130static void
3131qed_fill_generic_tlv_data(struct qed_dev *cdev, struct qed_mfw_tlv_generic *tlv)
3132{
3133	struct qed_common_cb_ops *op = cdev->protocol_ops.common;
3134	struct qed_eth_stats_common *p_common;
3135	struct qed_generic_tlvs gen_tlvs;
3136	struct qed_eth_stats stats;
3137	int i;
3138
3139	memset(&gen_tlvs, 0, sizeof(gen_tlvs));
3140	op->get_generic_tlv_data(cdev->ops_cookie, &gen_tlvs);
3141
3142	if (gen_tlvs.feat_flags & QED_TLV_IP_CSUM)
3143		tlv->flags.ipv4_csum_offload = true;
3144	if (gen_tlvs.feat_flags & QED_TLV_LSO)
3145		tlv->flags.lso_supported = true;
3146	tlv->flags.b_set = true;
3147
3148	for (i = 0; i < QED_TLV_MAC_COUNT; i++) {
3149		if (is_valid_ether_addr(gen_tlvs.mac[i])) {
3150			ether_addr_copy(tlv->mac[i], gen_tlvs.mac[i]);
3151			tlv->mac_set[i] = true;
3152		}
3153	}
3154
3155	qed_get_vport_stats(cdev, &stats);
3156	p_common = &stats.common;
3157	tlv->rx_frames = p_common->rx_ucast_pkts + p_common->rx_mcast_pkts +
3158			 p_common->rx_bcast_pkts;
3159	tlv->rx_frames_set = true;
3160	tlv->rx_bytes = p_common->rx_ucast_bytes + p_common->rx_mcast_bytes +
3161			p_common->rx_bcast_bytes;
3162	tlv->rx_bytes_set = true;
3163	tlv->tx_frames = p_common->tx_ucast_pkts + p_common->tx_mcast_pkts +
3164			 p_common->tx_bcast_pkts;
3165	tlv->tx_frames_set = true;
3166	tlv->tx_bytes = p_common->tx_ucast_bytes + p_common->tx_mcast_bytes +
3167			p_common->tx_bcast_bytes;
3168	tlv->rx_bytes_set = true;
3169}
3170
3171int qed_mfw_fill_tlv_data(struct qed_hwfn *hwfn, enum qed_mfw_tlv_type type,
3172			  union qed_mfw_tlv_data *tlv_buf)
3173{
3174	struct qed_dev *cdev = hwfn->cdev;
3175	struct qed_common_cb_ops *ops;
3176
3177	ops = cdev->protocol_ops.common;
3178	if (!ops || !ops->get_protocol_tlv_data || !ops->get_generic_tlv_data) {
3179		DP_NOTICE(hwfn, "Can't collect TLV management info\n");
3180		return -EINVAL;
3181	}
3182
3183	switch (type) {
3184	case QED_MFW_TLV_GENERIC:
3185		qed_fill_generic_tlv_data(hwfn->cdev, &tlv_buf->generic);
3186		break;
3187	case QED_MFW_TLV_ETH:
3188		ops->get_protocol_tlv_data(cdev->ops_cookie, &tlv_buf->eth);
3189		break;
3190	case QED_MFW_TLV_FCOE:
3191		ops->get_protocol_tlv_data(cdev->ops_cookie, &tlv_buf->fcoe);
3192		break;
3193	case QED_MFW_TLV_ISCSI:
3194		ops->get_protocol_tlv_data(cdev->ops_cookie, &tlv_buf->iscsi);
3195		break;
3196	default:
3197		break;
3198	}
3199
3200	return 0;
3201}
3202
3203unsigned long qed_get_epoch_time(void)
3204{
3205	return ktime_get_real_seconds();
3206}