Linux Audio

Check our new training course

Loading...
v4.17
   1// SPDX-License-Identifier: GPL-2.0
   2/* Copyright (c) 2018, Intel Corporation. */
   3
   4#include "ice_switch.h"
   5
   6#define ICE_ETH_DA_OFFSET		0
   7#define ICE_ETH_ETHTYPE_OFFSET		12
   8#define ICE_ETH_VLAN_TCI_OFFSET		14
   9#define ICE_MAX_VLAN_ID			0xFFF
  10
  11/* Dummy ethernet header needed in the ice_aqc_sw_rules_elem
  12 * struct to configure any switch filter rules.
  13 * {DA (6 bytes), SA(6 bytes),
  14 * Ether type (2 bytes for header without VLAN tag) OR
  15 * VLAN tag (4 bytes for header with VLAN tag) }
  16 *
  17 * Word on Hardcoded values
  18 * byte 0 = 0x2: to identify it as locally administered DA MAC
  19 * byte 6 = 0x2: to identify it as locally administered SA MAC
  20 * byte 12 = 0x81 & byte 13 = 0x00:
  21 *	In case of VLAN filter first two bytes defines ether type (0x8100)
  22 *	and remaining two bytes are placeholder for programming a given VLAN id
  23 *	In case of Ether type filter it is treated as header without VLAN tag
  24 *	and byte 12 and 13 is used to program a given Ether type instead
  25 */
  26#define DUMMY_ETH_HDR_LEN		16
  27static const u8 dummy_eth_header[DUMMY_ETH_HDR_LEN] = { 0x2, 0, 0, 0, 0, 0,
  28							0x2, 0, 0, 0, 0, 0,
  29							0x81, 0, 0, 0};
  30
  31#define ICE_SW_RULE_RX_TX_ETH_HDR_SIZE \
  32	(sizeof(struct ice_aqc_sw_rules_elem) - \
  33	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  34	 sizeof(struct ice_sw_rule_lkup_rx_tx) + DUMMY_ETH_HDR_LEN - 1)
  35#define ICE_SW_RULE_RX_TX_NO_HDR_SIZE \
  36	(sizeof(struct ice_aqc_sw_rules_elem) - \
  37	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  38	 sizeof(struct ice_sw_rule_lkup_rx_tx) - 1)
  39#define ICE_SW_RULE_LG_ACT_SIZE(n) \
  40	(sizeof(struct ice_aqc_sw_rules_elem) - \
  41	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  42	 sizeof(struct ice_sw_rule_lg_act) - \
  43	 sizeof(((struct ice_sw_rule_lg_act *)0)->act) + \
  44	 ((n) * sizeof(((struct ice_sw_rule_lg_act *)0)->act)))
  45#define ICE_SW_RULE_VSI_LIST_SIZE(n) \
  46	(sizeof(struct ice_aqc_sw_rules_elem) - \
  47	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  48	 sizeof(struct ice_sw_rule_vsi_list) - \
  49	 sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi) + \
  50	 ((n) * sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi)))
  51
  52/**
  53 * ice_aq_alloc_free_res - command to allocate/free resources
  54 * @hw: pointer to the hw struct
  55 * @num_entries: number of resource entries in buffer
  56 * @buf: Indirect buffer to hold data parameters and response
  57 * @buf_size: size of buffer for indirect commands
  58 * @opc: pass in the command opcode
  59 * @cd: pointer to command details structure or NULL
  60 *
  61 * Helper function to allocate/free resources using the admin queue commands
  62 */
  63static enum ice_status
  64ice_aq_alloc_free_res(struct ice_hw *hw, u16 num_entries,
  65		      struct ice_aqc_alloc_free_res_elem *buf, u16 buf_size,
  66		      enum ice_adminq_opc opc, struct ice_sq_cd *cd)
  67{
  68	struct ice_aqc_alloc_free_res_cmd *cmd;
  69	struct ice_aq_desc desc;
  70
  71	cmd = &desc.params.sw_res_ctrl;
  72
  73	if (!buf)
  74		return ICE_ERR_PARAM;
  75
  76	if (buf_size < (num_entries * sizeof(buf->elem[0])))
  77		return ICE_ERR_PARAM;
  78
  79	ice_fill_dflt_direct_cmd_desc(&desc, opc);
  80
  81	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
  82
  83	cmd->num_entries = cpu_to_le16(num_entries);
  84
  85	return ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
  86}
  87
  88/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  89 * ice_aq_get_sw_cfg - get switch configuration
  90 * @hw: pointer to the hardware structure
  91 * @buf: pointer to the result buffer
  92 * @buf_size: length of the buffer available for response
  93 * @req_desc: pointer to requested descriptor
  94 * @num_elems: pointer to number of elements
  95 * @cd: pointer to command details structure or NULL
  96 *
  97 * Get switch configuration (0x0200) to be placed in 'buff'.
  98 * This admin command returns information such as initial VSI/port number
  99 * and switch ID it belongs to.
 100 *
 101 * NOTE: *req_desc is both an input/output parameter.
 102 * The caller of this function first calls this function with *request_desc set
 103 * to 0.  If the response from f/w has *req_desc set to 0, all the switch
 104 * configuration information has been returned; if non-zero (meaning not all
 105 * the information was returned), the caller should call this function again
 106 * with *req_desc set to the previous value returned by f/w to get the
 107 * next block of switch configuration information.
 108 *
 109 * *num_elems is output only parameter. This reflects the number of elements
 110 * in response buffer. The caller of this function to use *num_elems while
 111 * parsing the response buffer.
 112 */
 113static enum ice_status
 114ice_aq_get_sw_cfg(struct ice_hw *hw, struct ice_aqc_get_sw_cfg_resp *buf,
 115		  u16 buf_size, u16 *req_desc, u16 *num_elems,
 116		  struct ice_sq_cd *cd)
 117{
 118	struct ice_aqc_get_sw_cfg *cmd;
 119	enum ice_status status;
 120	struct ice_aq_desc desc;
 121
 122	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_get_sw_cfg);
 123	cmd = &desc.params.get_sw_conf;
 124	cmd->element = cpu_to_le16(*req_desc);
 125
 126	status = ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
 127	if (!status) {
 128		*req_desc = le16_to_cpu(cmd->element);
 129		*num_elems = le16_to_cpu(cmd->num_elems);
 130	}
 131
 132	return status;
 133}
 134
 135/**
 136 * ice_aq_add_vsi
 137 * @hw: pointer to the hw struct
 138 * @vsi_ctx: pointer to a VSI context struct
 139 * @cd: pointer to command details structure or NULL
 140 *
 141 * Add a VSI context to the hardware (0x0210)
 142 */
 143enum ice_status
 144ice_aq_add_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 145	       struct ice_sq_cd *cd)
 146{
 147	struct ice_aqc_add_update_free_vsi_resp *res;
 148	struct ice_aqc_add_get_update_free_vsi *cmd;
 149	enum ice_status status;
 150	struct ice_aq_desc desc;
 
 151
 152	cmd = &desc.params.vsi_cmd;
 153	res = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
 154
 155	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_add_vsi);
 156
 157	if (!vsi_ctx->alloc_from_pool)
 158		cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num |
 159					   ICE_AQ_VSI_IS_VALID);
 
 160
 161	cmd->vsi_flags = cpu_to_le16(vsi_ctx->flags);
 162
 163	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 164
 165	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 166				 sizeof(vsi_ctx->info), cd);
 167
 168	if (!status) {
 169		vsi_ctx->vsi_num = le16_to_cpu(res->vsi_num) & ICE_AQ_VSI_NUM_M;
 170		vsi_ctx->vsis_allocd = le16_to_cpu(res->vsi_used);
 171		vsi_ctx->vsis_unallocated = le16_to_cpu(res->vsi_free);
 172	}
 173
 174	return status;
 175}
 176
 177/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 178 * ice_aq_update_vsi
 179 * @hw: pointer to the hw struct
 180 * @vsi_ctx: pointer to a VSI context struct
 181 * @cd: pointer to command details structure or NULL
 182 *
 183 * Update VSI context in the hardware (0x0211)
 184 */
 185enum ice_status
 186ice_aq_update_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 187		  struct ice_sq_cd *cd)
 188{
 189	struct ice_aqc_add_update_free_vsi_resp *resp;
 190	struct ice_aqc_add_get_update_free_vsi *cmd;
 191	struct ice_aq_desc desc;
 192	enum ice_status status;
 193
 194	cmd = &desc.params.vsi_cmd;
 195	resp = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
 196
 197	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_update_vsi);
 198
 199	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 200
 201	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 202
 203	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 204				 sizeof(vsi_ctx->info), cd);
 205
 206	if (!status) {
 207		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 208		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 209	}
 210
 211	return status;
 212}
 213
 214/**
 215 * ice_aq_free_vsi
 216 * @hw: pointer to the hw struct
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 217 * @vsi_ctx: pointer to a VSI context struct
 218 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 219 * @cd: pointer to command details structure or NULL
 220 *
 221 * Get VSI context info from hardware (0x0213)
 
 
 222 */
 223enum ice_status
 224ice_aq_free_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 225		bool keep_vsi_alloc, struct ice_sq_cd *cd)
 226{
 227	struct ice_aqc_add_update_free_vsi_resp *resp;
 228	struct ice_aqc_add_get_update_free_vsi *cmd;
 229	struct ice_aq_desc desc;
 230	enum ice_status status;
 231
 232	cmd = &desc.params.vsi_cmd;
 233	resp = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
 234
 235	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_free_vsi);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 236
 237	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 238	if (keep_vsi_alloc)
 239		cmd->cmd_flags = cpu_to_le16(ICE_AQ_VSI_KEEP_ALLOC);
 240
 241	status = ice_aq_send_cmd(hw, &desc, NULL, 0, cd);
 242	if (!status) {
 243		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 244		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 245	}
 
 
 
 
 
 
 
 
 
 
 246
 
 
 
 
 
 
 247	return status;
 248}
 249
 250/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 251 * ice_aq_alloc_free_vsi_list
 252 * @hw: pointer to the hw struct
 253 * @vsi_list_id: VSI list id returned or used for lookup
 254 * @lkup_type: switch rule filter lookup type
 255 * @opc: switch rules population command type - pass in the command opcode
 256 *
 257 * allocates or free a VSI list resource
 258 */
 259static enum ice_status
 260ice_aq_alloc_free_vsi_list(struct ice_hw *hw, u16 *vsi_list_id,
 261			   enum ice_sw_lkup_type lkup_type,
 262			   enum ice_adminq_opc opc)
 263{
 264	struct ice_aqc_alloc_free_res_elem *sw_buf;
 265	struct ice_aqc_res_elem *vsi_ele;
 266	enum ice_status status;
 267	u16 buf_len;
 268
 269	buf_len = sizeof(*sw_buf);
 270	sw_buf = devm_kzalloc(ice_hw_to_dev(hw), buf_len, GFP_KERNEL);
 271	if (!sw_buf)
 272		return ICE_ERR_NO_MEMORY;
 273	sw_buf->num_elems = cpu_to_le16(1);
 274
 275	if (lkup_type == ICE_SW_LKUP_MAC ||
 276	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 277	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 278	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 279	    lkup_type == ICE_SW_LKUP_PROMISC ||
 280	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN) {
 281		sw_buf->res_type = cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_REP);
 282	} else if (lkup_type == ICE_SW_LKUP_VLAN) {
 283		sw_buf->res_type =
 284			cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_PRUNE);
 285	} else {
 286		status = ICE_ERR_PARAM;
 287		goto ice_aq_alloc_free_vsi_list_exit;
 288	}
 289
 290	if (opc == ice_aqc_opc_free_res)
 291		sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
 292
 293	status = ice_aq_alloc_free_res(hw, 1, sw_buf, buf_len, opc, NULL);
 294	if (status)
 295		goto ice_aq_alloc_free_vsi_list_exit;
 296
 297	if (opc == ice_aqc_opc_alloc_res) {
 298		vsi_ele = &sw_buf->elem[0];
 299		*vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
 300	}
 301
 302ice_aq_alloc_free_vsi_list_exit:
 303	devm_kfree(ice_hw_to_dev(hw), sw_buf);
 304	return status;
 305}
 306
 307/**
 308 * ice_aq_sw_rules - add/update/remove switch rules
 309 * @hw: pointer to the hw struct
 310 * @rule_list: pointer to switch rule population list
 311 * @rule_list_sz: total size of the rule list in bytes
 312 * @num_rules: number of switch rules in the rule_list
 313 * @opc: switch rules population command type - pass in the command opcode
 314 * @cd: pointer to command details structure or NULL
 315 *
 316 * Add(0x02a0)/Update(0x02a1)/Remove(0x02a2) switch rules commands to firmware
 317 */
 318static enum ice_status
 319ice_aq_sw_rules(struct ice_hw *hw, void *rule_list, u16 rule_list_sz,
 320		u8 num_rules, enum ice_adminq_opc opc, struct ice_sq_cd *cd)
 321{
 322	struct ice_aq_desc desc;
 323
 324	if (opc != ice_aqc_opc_add_sw_rules &&
 325	    opc != ice_aqc_opc_update_sw_rules &&
 326	    opc != ice_aqc_opc_remove_sw_rules)
 327		return ICE_ERR_PARAM;
 328
 329	ice_fill_dflt_direct_cmd_desc(&desc, opc);
 330
 331	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 332	desc.params.sw_rules.num_rules_fltr_entry_index =
 333		cpu_to_le16(num_rules);
 334	return ice_aq_send_cmd(hw, &desc, rule_list, rule_list_sz, cd);
 335}
 336
 337/* ice_init_port_info - Initialize port_info with switch configuration data
 338 * @pi: pointer to port_info
 339 * @vsi_port_num: VSI number or port number
 340 * @type: Type of switch element (port or VSI)
 341 * @swid: switch ID of the switch the element is attached to
 342 * @pf_vf_num: PF or VF number
 343 * @is_vf: true if the element is a VF, false otherwise
 344 */
 345static void
 346ice_init_port_info(struct ice_port_info *pi, u16 vsi_port_num, u8 type,
 347		   u16 swid, u16 pf_vf_num, bool is_vf)
 348{
 349	switch (type) {
 350	case ICE_AQC_GET_SW_CONF_RESP_PHYS_PORT:
 351		pi->lport = (u8)(vsi_port_num & ICE_LPORT_MASK);
 352		pi->sw_id = swid;
 353		pi->pf_vf_num = pf_vf_num;
 354		pi->is_vf = is_vf;
 355		pi->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
 356		pi->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
 357		break;
 358	default:
 359		ice_debug(pi->hw, ICE_DBG_SW,
 360			  "incorrect VSI/port type received\n");
 361		break;
 362	}
 363}
 364
 365/* ice_get_initial_sw_cfg - Get initial port and default VSI data
 366 * @hw: pointer to the hardware structure
 367 */
 368enum ice_status ice_get_initial_sw_cfg(struct ice_hw *hw)
 369{
 370	struct ice_aqc_get_sw_cfg_resp *rbuf;
 371	enum ice_status status;
 372	u16 req_desc = 0;
 373	u16 num_elems;
 374	u16 i;
 375
 376	rbuf = devm_kzalloc(ice_hw_to_dev(hw), ICE_SW_CFG_MAX_BUF_LEN,
 377			    GFP_KERNEL);
 378
 379	if (!rbuf)
 380		return ICE_ERR_NO_MEMORY;
 381
 382	/* Multiple calls to ice_aq_get_sw_cfg may be required
 383	 * to get all the switch configuration information. The need
 384	 * for additional calls is indicated by ice_aq_get_sw_cfg
 385	 * writing a non-zero value in req_desc
 386	 */
 387	do {
 388		status = ice_aq_get_sw_cfg(hw, rbuf, ICE_SW_CFG_MAX_BUF_LEN,
 389					   &req_desc, &num_elems, NULL);
 390
 391		if (status)
 392			break;
 393
 394		for (i = 0; i < num_elems; i++) {
 395			struct ice_aqc_get_sw_cfg_resp_elem *ele;
 396			u16 pf_vf_num, swid, vsi_port_num;
 397			bool is_vf = false;
 398			u8 type;
 399
 400			ele = rbuf[i].elements;
 401			vsi_port_num = le16_to_cpu(ele->vsi_port_num) &
 402				ICE_AQC_GET_SW_CONF_RESP_VSI_PORT_NUM_M;
 403
 404			pf_vf_num = le16_to_cpu(ele->pf_vf_num) &
 405				ICE_AQC_GET_SW_CONF_RESP_FUNC_NUM_M;
 406
 407			swid = le16_to_cpu(ele->swid);
 408
 409			if (le16_to_cpu(ele->pf_vf_num) &
 410			    ICE_AQC_GET_SW_CONF_RESP_IS_VF)
 411				is_vf = true;
 412
 413			type = le16_to_cpu(ele->vsi_port_num) >>
 414				ICE_AQC_GET_SW_CONF_RESP_TYPE_S;
 415
 416			if (type == ICE_AQC_GET_SW_CONF_RESP_VSI) {
 417				/* FW VSI is not needed. Just continue. */
 418				continue;
 419			}
 420
 421			ice_init_port_info(hw->port_info, vsi_port_num,
 422					   type, swid, pf_vf_num, is_vf);
 423		}
 424	} while (req_desc && !status);
 425
 426	devm_kfree(ice_hw_to_dev(hw), (void *)rbuf);
 427	return status;
 428}
 429
 430/**
 431 * ice_fill_sw_info - Helper function to populate lb_en and lan_en
 432 * @hw: pointer to the hardware structure
 433 * @f_info: filter info structure to fill/update
 434 *
 435 * This helper function populates the lb_en and lan_en elements of the provided
 436 * ice_fltr_info struct using the switch's type and characteristics of the
 437 * switch rule being configured.
 438 */
 439static void ice_fill_sw_info(struct ice_hw *hw, struct ice_fltr_info *f_info)
 440{
 441	f_info->lb_en = false;
 442	f_info->lan_en = false;
 443	if ((f_info->flag & ICE_FLTR_TX) &&
 444	    (f_info->fltr_act == ICE_FWD_TO_VSI ||
 445	     f_info->fltr_act == ICE_FWD_TO_VSI_LIST ||
 446	     f_info->fltr_act == ICE_FWD_TO_Q ||
 447	     f_info->fltr_act == ICE_FWD_TO_QGRP)) {
 448		f_info->lb_en = true;
 449		if (!(hw->evb_veb && f_info->lkup_type == ICE_SW_LKUP_MAC &&
 450		      is_unicast_ether_addr(f_info->l_data.mac.mac_addr)))
 451			f_info->lan_en = true;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 452	}
 453}
 454
 455/**
 456 * ice_fill_sw_rule - Helper function to fill switch rule structure
 457 * @hw: pointer to the hardware structure
 458 * @f_info: entry containing packet forwarding information
 459 * @s_rule: switch rule structure to be filled in based on mac_entry
 460 * @opc: switch rules population command type - pass in the command opcode
 461 */
 462static void
 463ice_fill_sw_rule(struct ice_hw *hw, struct ice_fltr_info *f_info,
 464		 struct ice_aqc_sw_rules_elem *s_rule, enum ice_adminq_opc opc)
 465{
 466	u16 vlan_id = ICE_MAX_VLAN_ID + 1;
 467	u8 eth_hdr[DUMMY_ETH_HDR_LEN];
 468	void *daddr = NULL;
 
 
 469	u32 act = 0;
 470	__be16 *off;
 
 471
 472	if (opc == ice_aqc_opc_remove_sw_rules) {
 473		s_rule->pdata.lkup_tx_rx.act = 0;
 474		s_rule->pdata.lkup_tx_rx.index =
 475			cpu_to_le16(f_info->fltr_rule_id);
 476		s_rule->pdata.lkup_tx_rx.hdr_len = 0;
 477		return;
 478	}
 479
 
 
 
 480	/* initialize the ether header with a dummy header */
 481	memcpy(eth_hdr, dummy_eth_header, sizeof(dummy_eth_header));
 482	ice_fill_sw_info(hw, f_info);
 483
 484	switch (f_info->fltr_act) {
 485	case ICE_FWD_TO_VSI:
 486		act |= (f_info->fwd_id.vsi_id << ICE_SINGLE_ACT_VSI_ID_S) &
 487			ICE_SINGLE_ACT_VSI_ID_M;
 488		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 489			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 490				ICE_SINGLE_ACT_VALID_BIT;
 491		break;
 492	case ICE_FWD_TO_VSI_LIST:
 493		act |= ICE_SINGLE_ACT_VSI_LIST;
 494		act |= (f_info->fwd_id.vsi_list_id <<
 495			ICE_SINGLE_ACT_VSI_LIST_ID_S) &
 496			ICE_SINGLE_ACT_VSI_LIST_ID_M;
 497		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 498			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 499				ICE_SINGLE_ACT_VALID_BIT;
 500		break;
 501	case ICE_FWD_TO_Q:
 502		act |= ICE_SINGLE_ACT_TO_Q;
 503		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 504			ICE_SINGLE_ACT_Q_INDEX_M;
 505		break;
 
 
 
 
 506	case ICE_FWD_TO_QGRP:
 
 
 507		act |= ICE_SINGLE_ACT_TO_Q;
 508		act |= (f_info->qgrp_size << ICE_SINGLE_ACT_Q_REGION_S) &
 
 
 509			ICE_SINGLE_ACT_Q_REGION_M;
 510		break;
 511	case ICE_DROP_PACKET:
 512		act |= ICE_SINGLE_ACT_VSI_FORWARDING | ICE_SINGLE_ACT_DROP;
 513		break;
 514	default:
 515		return;
 516	}
 517
 518	if (f_info->lb_en)
 519		act |= ICE_SINGLE_ACT_LB_ENABLE;
 520	if (f_info->lan_en)
 521		act |= ICE_SINGLE_ACT_LAN_ENABLE;
 522
 523	switch (f_info->lkup_type) {
 524	case ICE_SW_LKUP_MAC:
 525		daddr = f_info->l_data.mac.mac_addr;
 526		break;
 527	case ICE_SW_LKUP_VLAN:
 528		vlan_id = f_info->l_data.vlan.vlan_id;
 529		if (f_info->fltr_act == ICE_FWD_TO_VSI ||
 530		    f_info->fltr_act == ICE_FWD_TO_VSI_LIST) {
 531			act |= ICE_SINGLE_ACT_PRUNE;
 532			act |= ICE_SINGLE_ACT_EGRESS | ICE_SINGLE_ACT_INGRESS;
 533		}
 534		break;
 535	case ICE_SW_LKUP_ETHERTYPE_MAC:
 536		daddr = f_info->l_data.ethertype_mac.mac_addr;
 537		/* fall-through */
 538	case ICE_SW_LKUP_ETHERTYPE:
 539		off = (__be16 *)&eth_hdr[ICE_ETH_ETHTYPE_OFFSET];
 540		*off = cpu_to_be16(f_info->l_data.ethertype_mac.ethertype);
 541		break;
 542	case ICE_SW_LKUP_MAC_VLAN:
 543		daddr = f_info->l_data.mac_vlan.mac_addr;
 544		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 545		break;
 546	case ICE_SW_LKUP_PROMISC_VLAN:
 547		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 548		/* fall-through */
 549	case ICE_SW_LKUP_PROMISC:
 550		daddr = f_info->l_data.mac_vlan.mac_addr;
 551		break;
 552	default:
 553		break;
 554	}
 555
 556	s_rule->type = (f_info->flag & ICE_FLTR_RX) ?
 557		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_RX) :
 558		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_TX);
 559
 560	/* Recipe set depending on lookup type */
 561	s_rule->pdata.lkup_tx_rx.recipe_id = cpu_to_le16(f_info->lkup_type);
 562	s_rule->pdata.lkup_tx_rx.src = cpu_to_le16(f_info->src);
 563	s_rule->pdata.lkup_tx_rx.act = cpu_to_le32(act);
 564
 565	if (daddr)
 566		ether_addr_copy(&eth_hdr[ICE_ETH_DA_OFFSET], daddr);
 567
 568	if (!(vlan_id > ICE_MAX_VLAN_ID)) {
 569		off = (__be16 *)&eth_hdr[ICE_ETH_VLAN_TCI_OFFSET];
 570		*off = cpu_to_be16(vlan_id);
 571	}
 572
 573	/* Create the switch rule with the final dummy Ethernet header */
 574	if (opc != ice_aqc_opc_update_sw_rules)
 575		s_rule->pdata.lkup_tx_rx.hdr_len = cpu_to_le16(sizeof(eth_hdr));
 576
 577	memcpy(s_rule->pdata.lkup_tx_rx.hdr, eth_hdr, sizeof(eth_hdr));
 578}
 579
 580/**
 581 * ice_add_marker_act
 582 * @hw: pointer to the hardware structure
 583 * @m_ent: the management entry for which sw marker needs to be added
 584 * @sw_marker: sw marker to tag the Rx descriptor with
 585 * @l_id: large action resource id
 586 *
 587 * Create a large action to hold software marker and update the switch rule
 588 * entry pointed by m_ent with newly created large action
 589 */
 590static enum ice_status
 591ice_add_marker_act(struct ice_hw *hw, struct ice_fltr_mgmt_list_entry *m_ent,
 592		   u16 sw_marker, u16 l_id)
 593{
 594	struct ice_aqc_sw_rules_elem *lg_act, *rx_tx;
 595	/* For software marker we need 3 large actions
 596	 * 1. FWD action: FWD TO VSI or VSI LIST
 597	 * 2. GENERIC VALUE action to hold the profile id
 598	 * 3. GENERIC VALUE action to hold the software marker id
 599	 */
 600	const u16 num_lg_acts = 3;
 601	enum ice_status status;
 602	u16 lg_act_size;
 603	u16 rules_size;
 604	u16 vsi_info;
 605	u32 act;
 
 606
 607	if (m_ent->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
 608		return ICE_ERR_PARAM;
 609
 610	/* Create two back-to-back switch rules and submit them to the HW using
 611	 * one memory buffer:
 612	 *    1. Large Action
 613	 *    2. Look up tx rx
 614	 */
 615	lg_act_size = (u16)ICE_SW_RULE_LG_ACT_SIZE(num_lg_acts);
 616	rules_size = lg_act_size + ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
 617	lg_act = devm_kzalloc(ice_hw_to_dev(hw), rules_size, GFP_KERNEL);
 618	if (!lg_act)
 619		return ICE_ERR_NO_MEMORY;
 620
 621	rx_tx = (struct ice_aqc_sw_rules_elem *)((u8 *)lg_act + lg_act_size);
 622
 623	/* Fill in the first switch rule i.e. large action */
 624	lg_act->type = cpu_to_le16(ICE_AQC_SW_RULES_T_LG_ACT);
 625	lg_act->pdata.lg_act.index = cpu_to_le16(l_id);
 626	lg_act->pdata.lg_act.size = cpu_to_le16(num_lg_acts);
 627
 628	/* First action VSI forwarding or VSI list forwarding depending on how
 629	 * many VSIs
 630	 */
 631	vsi_info = (m_ent->vsi_count > 1) ?
 632		m_ent->fltr_info.fwd_id.vsi_list_id :
 633		m_ent->fltr_info.fwd_id.vsi_id;
 634
 635	act = ICE_LG_ACT_VSI_FORWARDING | ICE_LG_ACT_VALID_BIT;
 636	act |= (vsi_info << ICE_LG_ACT_VSI_LIST_ID_S) &
 637		ICE_LG_ACT_VSI_LIST_ID_M;
 638	if (m_ent->vsi_count > 1)
 639		act |= ICE_LG_ACT_VSI_LIST;
 640	lg_act->pdata.lg_act.act[0] = cpu_to_le32(act);
 641
 642	/* Second action descriptor type */
 643	act = ICE_LG_ACT_GENERIC;
 644
 645	act |= (1 << ICE_LG_ACT_GENERIC_VALUE_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 646	lg_act->pdata.lg_act.act[1] = cpu_to_le32(act);
 647
 648	act = (7 << ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 
 649
 650	/* Third action Marker value */
 651	act |= ICE_LG_ACT_GENERIC;
 652	act |= (sw_marker << ICE_LG_ACT_GENERIC_VALUE_S) &
 653		ICE_LG_ACT_GENERIC_VALUE_M;
 654
 655	act |= (0 << ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 656	lg_act->pdata.lg_act.act[2] = cpu_to_le32(act);
 657
 658	/* call the fill switch rule to fill the lookup tx rx structure */
 659	ice_fill_sw_rule(hw, &m_ent->fltr_info, rx_tx,
 660			 ice_aqc_opc_update_sw_rules);
 661
 662	/* Update the action to point to the large action id */
 663	rx_tx->pdata.lkup_tx_rx.act =
 664		cpu_to_le32(ICE_SINGLE_ACT_PTR |
 665			    ((l_id << ICE_SINGLE_ACT_PTR_VAL_S) &
 666			     ICE_SINGLE_ACT_PTR_VAL_M));
 667
 668	/* Use the filter rule id of the previously created rule with single
 669	 * act. Once the update happens, hardware will treat this as large
 670	 * action
 671	 */
 672	rx_tx->pdata.lkup_tx_rx.index =
 673		cpu_to_le16(m_ent->fltr_info.fltr_rule_id);
 674
 675	status = ice_aq_sw_rules(hw, lg_act, rules_size, 2,
 676				 ice_aqc_opc_update_sw_rules, NULL);
 677	if (!status) {
 678		m_ent->lg_act_idx = l_id;
 679		m_ent->sw_marker_id = sw_marker;
 680	}
 681
 682	devm_kfree(ice_hw_to_dev(hw), lg_act);
 683	return status;
 684}
 685
 686/**
 687 * ice_create_vsi_list_map
 688 * @hw: pointer to the hardware structure
 689 * @vsi_array: array of VSIs to form a VSI list
 690 * @num_vsi: num VSI in the array
 691 * @vsi_list_id: VSI list id generated as part of allocate resource
 692 *
 693 * Helper function to create a new entry of VSI list id to VSI mapping
 694 * using the given VSI list id
 695 */
 696static struct ice_vsi_list_map_info *
 697ice_create_vsi_list_map(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
 698			u16 vsi_list_id)
 699{
 700	struct ice_switch_info *sw = hw->switch_info;
 701	struct ice_vsi_list_map_info *v_map;
 702	int i;
 703
 704	v_map = devm_kcalloc(ice_hw_to_dev(hw), 1, sizeof(*v_map), GFP_KERNEL);
 705	if (!v_map)
 706		return NULL;
 707
 708	v_map->vsi_list_id = vsi_list_id;
 709
 710	for (i = 0; i < num_vsi; i++)
 711		set_bit(vsi_array[i], v_map->vsi_map);
 712
 713	list_add(&v_map->list_entry, &sw->vsi_list_map_head);
 714	return v_map;
 715}
 716
 717/**
 718 * ice_update_vsi_list_rule
 719 * @hw: pointer to the hardware structure
 720 * @vsi_array: array of VSIs to form a VSI list
 721 * @num_vsi: num VSI in the array
 722 * @vsi_list_id: VSI list id generated as part of allocate resource
 723 * @remove: Boolean value to indicate if this is a remove action
 724 * @opc: switch rules population command type - pass in the command opcode
 725 * @lkup_type: lookup type of the filter
 726 *
 727 * Call AQ command to add a new switch rule or update existing switch rule
 728 * using the given VSI list id
 729 */
 730static enum ice_status
 731ice_update_vsi_list_rule(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
 732			 u16 vsi_list_id, bool remove, enum ice_adminq_opc opc,
 733			 enum ice_sw_lkup_type lkup_type)
 734{
 735	struct ice_aqc_sw_rules_elem *s_rule;
 736	enum ice_status status;
 737	u16 s_rule_size;
 738	u16 type;
 739	int i;
 740
 741	if (!num_vsi)
 742		return ICE_ERR_PARAM;
 743
 744	if (lkup_type == ICE_SW_LKUP_MAC ||
 745	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 746	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 747	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 748	    lkup_type == ICE_SW_LKUP_PROMISC ||
 749	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN)
 750		type = remove ? ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR :
 751				ICE_AQC_SW_RULES_T_VSI_LIST_SET;
 752	else if (lkup_type == ICE_SW_LKUP_VLAN)
 753		type = remove ? ICE_AQC_SW_RULES_T_PRUNE_LIST_CLEAR :
 754				ICE_AQC_SW_RULES_T_PRUNE_LIST_SET;
 755	else
 756		return ICE_ERR_PARAM;
 757
 758	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(num_vsi);
 759	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
 760	if (!s_rule)
 761		return ICE_ERR_NO_MEMORY;
 762
 763	for (i = 0; i < num_vsi; i++)
 764		s_rule->pdata.vsi_list.vsi[i] = cpu_to_le16(vsi_array[i]);
 
 
 
 
 
 
 765
 766	s_rule->type = cpu_to_le16(type);
 767	s_rule->pdata.vsi_list.number_vsi = cpu_to_le16(num_vsi);
 768	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
 769
 770	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opc, NULL);
 771
 
 772	devm_kfree(ice_hw_to_dev(hw), s_rule);
 773	return status;
 774}
 775
 776/**
 777 * ice_create_vsi_list_rule - Creates and populates a VSI list rule
 778 * @hw: pointer to the hw struct
 779 * @vsi_array: array of VSIs to form a VSI list
 780 * @num_vsi: number of VSIs in the array
 781 * @vsi_list_id: stores the ID of the VSI list to be created
 782 * @lkup_type: switch rule filter's lookup type
 783 */
 784static enum ice_status
 785ice_create_vsi_list_rule(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
 786			 u16 *vsi_list_id, enum ice_sw_lkup_type lkup_type)
 787{
 788	enum ice_status status;
 789	int i;
 790
 791	for (i = 0; i < num_vsi; i++)
 792		if (vsi_array[i] >= ICE_MAX_VSI)
 793			return ICE_ERR_OUT_OF_RANGE;
 794
 795	status = ice_aq_alloc_free_vsi_list(hw, vsi_list_id, lkup_type,
 796					    ice_aqc_opc_alloc_res);
 797	if (status)
 798		return status;
 799
 800	/* Update the newly created VSI list to include the specified VSIs */
 801	return ice_update_vsi_list_rule(hw, vsi_array, num_vsi, *vsi_list_id,
 802					false, ice_aqc_opc_add_sw_rules,
 803					lkup_type);
 804}
 805
 806/**
 807 * ice_create_pkt_fwd_rule
 808 * @hw: pointer to the hardware structure
 809 * @f_entry: entry containing packet forwarding information
 810 *
 811 * Create switch rule with given filter information and add an entry
 812 * to the corresponding filter management list to track this switch rule
 813 * and VSI mapping
 814 */
 815static enum ice_status
 816ice_create_pkt_fwd_rule(struct ice_hw *hw,
 817			struct ice_fltr_list_entry *f_entry)
 818{
 819	struct ice_switch_info *sw = hw->switch_info;
 820	struct ice_fltr_mgmt_list_entry *fm_entry;
 821	struct ice_aqc_sw_rules_elem *s_rule;
 822	enum ice_sw_lkup_type l_type;
 
 823	enum ice_status status;
 824
 825	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
 826			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
 827	if (!s_rule)
 828		return ICE_ERR_NO_MEMORY;
 829	fm_entry = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*fm_entry),
 830				GFP_KERNEL);
 831	if (!fm_entry) {
 832		status = ICE_ERR_NO_MEMORY;
 833		goto ice_create_pkt_fwd_rule_exit;
 834	}
 835
 836	fm_entry->fltr_info = f_entry->fltr_info;
 837
 838	/* Initialize all the fields for the management entry */
 839	fm_entry->vsi_count = 1;
 840	fm_entry->lg_act_idx = ICE_INVAL_LG_ACT_INDEX;
 841	fm_entry->sw_marker_id = ICE_INVAL_SW_MARKER_ID;
 842	fm_entry->counter_index = ICE_INVAL_COUNTER_ID;
 843
 844	ice_fill_sw_rule(hw, &fm_entry->fltr_info, s_rule,
 845			 ice_aqc_opc_add_sw_rules);
 846
 847	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
 848				 ice_aqc_opc_add_sw_rules, NULL);
 849	if (status) {
 850		devm_kfree(ice_hw_to_dev(hw), fm_entry);
 851		goto ice_create_pkt_fwd_rule_exit;
 852	}
 853
 854	f_entry->fltr_info.fltr_rule_id =
 855		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
 856	fm_entry->fltr_info.fltr_rule_id =
 857		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
 858
 859	/* The book keeping entries will get removed when base driver
 860	 * calls remove filter AQ command
 861	 */
 862	l_type = fm_entry->fltr_info.lkup_type;
 863	if (l_type == ICE_SW_LKUP_MAC) {
 864		mutex_lock(&sw->mac_list_lock);
 865		list_add(&fm_entry->list_entry, &sw->mac_list_head);
 866		mutex_unlock(&sw->mac_list_lock);
 867	} else if (l_type == ICE_SW_LKUP_VLAN) {
 868		mutex_lock(&sw->vlan_list_lock);
 869		list_add(&fm_entry->list_entry, &sw->vlan_list_head);
 870		mutex_unlock(&sw->vlan_list_lock);
 871	} else if (l_type == ICE_SW_LKUP_ETHERTYPE ||
 872		   l_type == ICE_SW_LKUP_ETHERTYPE_MAC) {
 873		mutex_lock(&sw->eth_m_list_lock);
 874		list_add(&fm_entry->list_entry, &sw->eth_m_list_head);
 875		mutex_unlock(&sw->eth_m_list_lock);
 876	} else if (l_type == ICE_SW_LKUP_PROMISC ||
 877		   l_type == ICE_SW_LKUP_PROMISC_VLAN) {
 878		mutex_lock(&sw->promisc_list_lock);
 879		list_add(&fm_entry->list_entry, &sw->promisc_list_head);
 880		mutex_unlock(&sw->promisc_list_lock);
 881	} else if (fm_entry->fltr_info.lkup_type == ICE_SW_LKUP_MAC_VLAN) {
 882		mutex_lock(&sw->mac_vlan_list_lock);
 883		list_add(&fm_entry->list_entry, &sw->mac_vlan_list_head);
 884		mutex_unlock(&sw->mac_vlan_list_lock);
 885	} else {
 886		status = ICE_ERR_NOT_IMPL;
 887	}
 888ice_create_pkt_fwd_rule_exit:
 889	devm_kfree(ice_hw_to_dev(hw), s_rule);
 890	return status;
 891}
 892
 893/**
 894 * ice_update_pkt_fwd_rule
 895 * @hw: pointer to the hardware structure
 896 * @rule_id: rule of previously created switch rule to update
 897 * @vsi_list_id: VSI list id to be updated with
 898 * @f_info: ice_fltr_info to pull other information for switch rule
 899 *
 900 * Call AQ command to update a previously created switch rule with a
 901 * VSI list id
 902 */
 903static enum ice_status
 904ice_update_pkt_fwd_rule(struct ice_hw *hw, u16 rule_id, u16 vsi_list_id,
 905			struct ice_fltr_info f_info)
 906{
 907	struct ice_aqc_sw_rules_elem *s_rule;
 908	struct ice_fltr_info tmp_fltr;
 909	enum ice_status status;
 910
 911	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
 912			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
 913	if (!s_rule)
 914		return ICE_ERR_NO_MEMORY;
 915
 916	tmp_fltr = f_info;
 917	tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
 918	tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
 919
 920	ice_fill_sw_rule(hw, &tmp_fltr, s_rule,
 921			 ice_aqc_opc_update_sw_rules);
 922
 923	s_rule->pdata.lkup_tx_rx.index = cpu_to_le16(rule_id);
 924
 925	/* Update switch rule with new rule set to forward VSI list */
 926	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
 927				 ice_aqc_opc_update_sw_rules, NULL);
 928
 929	devm_kfree(ice_hw_to_dev(hw), s_rule);
 930	return status;
 931}
 932
 933/**
 934 * ice_handle_vsi_list_mgmt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 935 * @hw: pointer to the hardware structure
 936 * @m_entry: pointer to current filter management list entry
 937 * @cur_fltr: filter information from the book keeping entry
 938 * @new_fltr: filter information with the new VSI to be added
 939 *
 940 * Call AQ command to add or update previously created VSI list with new VSI.
 941 *
 942 * Helper function to do book keeping associated with adding filter information
 943 * The algorithm to do the booking keeping is described below :
 944 * When a VSI needs to subscribe to a given filter( MAC/VLAN/Ethtype etc.)
 945 *	if only one VSI has been added till now
 946 *		Allocate a new VSI list and add two VSIs
 947 *		to this list using switch rule command
 948 *		Update the previously created switch rule with the
 949 *		newly created VSI list id
 950 *	if a VSI list was previously created
 951 *		Add the new VSI to the previously created VSI list set
 952 *		using the update switch rule command
 953 */
 954static enum ice_status
 955ice_handle_vsi_list_mgmt(struct ice_hw *hw,
 956			 struct ice_fltr_mgmt_list_entry *m_entry,
 957			 struct ice_fltr_info *cur_fltr,
 958			 struct ice_fltr_info *new_fltr)
 959{
 960	enum ice_status status = 0;
 961	u16 vsi_list_id = 0;
 962
 963	if ((cur_fltr->fltr_act == ICE_FWD_TO_Q ||
 964	     cur_fltr->fltr_act == ICE_FWD_TO_QGRP))
 965		return ICE_ERR_NOT_IMPL;
 966
 967	if ((new_fltr->fltr_act == ICE_FWD_TO_Q ||
 968	     new_fltr->fltr_act == ICE_FWD_TO_QGRP) &&
 969	    (cur_fltr->fltr_act == ICE_FWD_TO_VSI ||
 970	     cur_fltr->fltr_act == ICE_FWD_TO_VSI_LIST))
 971		return ICE_ERR_NOT_IMPL;
 972
 973	if (m_entry->vsi_count < 2 && !m_entry->vsi_list_info) {
 974		/* Only one entry existed in the mapping and it was not already
 975		 * a part of a VSI list. So, create a VSI list with the old and
 976		 * new VSIs.
 977		 */
 978		u16 vsi_id_arr[2];
 979		u16 fltr_rule;
 980
 981		/* A rule already exists with the new VSI being added */
 982		if (cur_fltr->fwd_id.vsi_id == new_fltr->fwd_id.vsi_id)
 983			return ICE_ERR_ALREADY_EXISTS;
 984
 985		vsi_id_arr[0] = cur_fltr->fwd_id.vsi_id;
 986		vsi_id_arr[1] = new_fltr->fwd_id.vsi_id;
 987		status = ice_create_vsi_list_rule(hw, &vsi_id_arr[0], 2,
 988						  &vsi_list_id,
 989						  new_fltr->lkup_type);
 990		if (status)
 991			return status;
 992
 993		fltr_rule = cur_fltr->fltr_rule_id;
 
 
 
 994		/* Update the previous switch rule of "MAC forward to VSI" to
 995		 * "MAC fwd to VSI list"
 996		 */
 997		status = ice_update_pkt_fwd_rule(hw, fltr_rule, vsi_list_id,
 998						 *new_fltr);
 999		if (status)
1000			return status;
1001
1002		cur_fltr->fwd_id.vsi_list_id = vsi_list_id;
1003		cur_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1004		m_entry->vsi_list_info =
1005			ice_create_vsi_list_map(hw, &vsi_id_arr[0], 2,
1006						vsi_list_id);
1007
1008		/* If this entry was large action then the large action needs
1009		 * to be updated to point to FWD to VSI list
1010		 */
1011		if (m_entry->sw_marker_id != ICE_INVAL_SW_MARKER_ID)
1012			status =
1013			    ice_add_marker_act(hw, m_entry,
1014					       m_entry->sw_marker_id,
1015					       m_entry->lg_act_idx);
1016	} else {
1017		u16 vsi_id = new_fltr->fwd_id.vsi_id;
1018		enum ice_adminq_opc opcode;
1019
 
 
 
1020		/* A rule already exists with the new VSI being added */
1021		if (test_bit(vsi_id, m_entry->vsi_list_info->vsi_map))
1022			return 0;
1023
1024		/* Update the previously created VSI list set with
1025		 * the new VSI id passed in
1026		 */
1027		vsi_list_id = cur_fltr->fwd_id.vsi_list_id;
1028		opcode = ice_aqc_opc_update_sw_rules;
1029
1030		status = ice_update_vsi_list_rule(hw, &vsi_id, 1, vsi_list_id,
1031						  false, opcode,
1032						  new_fltr->lkup_type);
1033		/* update VSI list mapping info with new VSI id */
1034		if (!status)
1035			set_bit(vsi_id, m_entry->vsi_list_info->vsi_map);
1036	}
1037	if (!status)
1038		m_entry->vsi_count++;
1039	return status;
1040}
1041
1042/**
1043 * ice_find_mac_entry
1044 * @hw: pointer to the hardware structure
1045 * @mac_addr: MAC address to search for
 
1046 *
1047 * Helper function to search for a MAC entry using a given MAC address
1048 * Returns pointer to the entry if found.
1049 */
1050static struct ice_fltr_mgmt_list_entry *
1051ice_find_mac_entry(struct ice_hw *hw, u8 *mac_addr)
1052{
1053	struct ice_fltr_mgmt_list_entry *m_list_itr, *mac_ret = NULL;
1054	struct ice_switch_info *sw = hw->switch_info;
 
1055
1056	mutex_lock(&sw->mac_list_lock);
1057	list_for_each_entry(m_list_itr, &sw->mac_list_head, list_entry) {
1058		u8 *buf = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1059
1060		if (ether_addr_equal(buf, mac_addr)) {
1061			mac_ret = m_list_itr;
1062			break;
1063		}
1064	}
1065	mutex_unlock(&sw->mac_list_lock);
1066	return mac_ret;
1067}
1068
1069/**
1070 * ice_add_shared_mac - Add one MAC shared filter rule
1071 * @hw: pointer to the hardware structure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1072 * @f_entry: structure containing MAC forwarding information
1073 *
1074 * Adds or updates the book keeping list for the MAC addresses
1075 */
1076static enum ice_status
1077ice_add_shared_mac(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
 
1078{
 
1079	struct ice_fltr_info *new_fltr, *cur_fltr;
1080	struct ice_fltr_mgmt_list_entry *m_entry;
 
 
1081
1082	new_fltr = &f_entry->fltr_info;
 
 
 
 
 
1083
1084	m_entry = ice_find_mac_entry(hw, &new_fltr->l_data.mac.mac_addr[0]);
1085	if (!m_entry)
 
 
 
 
 
 
 
 
1086		return ice_create_pkt_fwd_rule(hw, f_entry);
 
1087
1088	cur_fltr = &m_entry->fltr_info;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1089
1090	return ice_handle_vsi_list_mgmt(hw, m_entry, cur_fltr, new_fltr);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1091}
1092
1093/**
1094 * ice_add_mac - Add a MAC address based filter rule
1095 * @hw: pointer to the hardware structure
1096 * @m_list: list of MAC addresses and forwarding information
1097 *
1098 * IMPORTANT: When the ucast_shared flag is set to false and m_list has
1099 * multiple unicast addresses, the function assumes that all the
1100 * addresses are unique in a given add_mac call. It doesn't
1101 * check for duplicates in this case, removing duplicates from a given
1102 * list should be taken care of in the caller of this function.
1103 */
1104enum ice_status
1105ice_add_mac(struct ice_hw *hw, struct list_head *m_list)
1106{
1107	struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1108	struct ice_fltr_list_entry *m_list_itr;
 
1109	u16 elem_sent, total_elem_left;
 
 
1110	enum ice_status status = 0;
1111	u16 num_unicast = 0;
1112	u16 s_rule_size;
1113
1114	if (!m_list || !hw)
1115		return ICE_ERR_PARAM;
1116
 
 
 
1117	list_for_each_entry(m_list_itr, m_list, list_entry) {
1118		u8 *add = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
 
 
1119
1120		if (m_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
 
 
 
 
 
 
 
1121			return ICE_ERR_PARAM;
1122		if (is_zero_ether_addr(add))
 
 
1123			return ICE_ERR_PARAM;
1124		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
1125			/* Don't overwrite the unicast address */
1126			if (ice_find_mac_entry(hw, add))
 
 
 
1127				return ICE_ERR_ALREADY_EXISTS;
 
 
1128			num_unicast++;
1129		} else if (is_multicast_ether_addr(add) ||
1130			   (is_unicast_ether_addr(add) && hw->ucast_shared)) {
1131			status = ice_add_shared_mac(hw, m_list_itr);
1132			if (status) {
1133				m_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1134				return status;
1135			}
1136			m_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
1137		}
1138	}
1139
 
1140	/* Exit if no suitable entries were found for adding bulk switch rule */
1141	if (!num_unicast)
1142		return 0;
 
 
 
 
1143
1144	/* Allocate switch rule buffer for the bulk update for unicast */
1145	s_rule_size = ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
1146	s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1147			      GFP_KERNEL);
1148	if (!s_rule)
1149		return ICE_ERR_NO_MEMORY;
 
 
1150
1151	r_iter = s_rule;
1152	list_for_each_entry(m_list_itr, m_list, list_entry) {
1153		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1154		u8 *addr = &f_info->l_data.mac.mac_addr[0];
1155
1156		if (is_unicast_ether_addr(addr)) {
1157			ice_fill_sw_rule(hw, &m_list_itr->fltr_info,
1158					 r_iter, ice_aqc_opc_add_sw_rules);
1159			r_iter = (struct ice_aqc_sw_rules_elem *)
1160				((u8 *)r_iter + s_rule_size);
1161		}
1162	}
1163
1164	/* Call AQ bulk switch rule update for all unicast addresses */
1165	r_iter = s_rule;
1166	/* Call AQ switch rule in AQ_MAX chunk */
1167	for (total_elem_left = num_unicast; total_elem_left > 0;
1168	     total_elem_left -= elem_sent) {
1169		struct ice_aqc_sw_rules_elem *entry = r_iter;
1170
1171		elem_sent = min(total_elem_left,
1172				(u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1173		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1174					 elem_sent, ice_aqc_opc_add_sw_rules,
1175					 NULL);
1176		if (status)
1177			goto ice_add_mac_exit;
1178		r_iter = (struct ice_aqc_sw_rules_elem *)
1179			((u8 *)r_iter + (elem_sent * s_rule_size));
1180	}
1181
1182	/* Fill up rule id based on the value returned from FW */
1183	r_iter = s_rule;
1184	list_for_each_entry(m_list_itr, m_list, list_entry) {
1185		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1186		u8 *addr = &f_info->l_data.mac.mac_addr[0];
1187		struct ice_switch_info *sw = hw->switch_info;
1188		struct ice_fltr_mgmt_list_entry *fm_entry;
1189
1190		if (is_unicast_ether_addr(addr)) {
1191			f_info->fltr_rule_id =
1192				le16_to_cpu(r_iter->pdata.lkup_tx_rx.index);
1193			f_info->fltr_act = ICE_FWD_TO_VSI;
1194			/* Create an entry to track this MAC address */
1195			fm_entry = devm_kzalloc(ice_hw_to_dev(hw),
1196						sizeof(*fm_entry), GFP_KERNEL);
1197			if (!fm_entry) {
1198				status = ICE_ERR_NO_MEMORY;
1199				goto ice_add_mac_exit;
1200			}
1201			fm_entry->fltr_info = *f_info;
1202			fm_entry->vsi_count = 1;
1203			/* The book keeping entries will get removed when
1204			 * base driver calls remove filter AQ command
1205			 */
1206			mutex_lock(&sw->mac_list_lock);
1207			list_add(&fm_entry->list_entry, &sw->mac_list_head);
1208			mutex_unlock(&sw->mac_list_lock);
1209
 
1210			r_iter = (struct ice_aqc_sw_rules_elem *)
1211				((u8 *)r_iter + s_rule_size);
1212		}
1213	}
1214
1215ice_add_mac_exit:
1216	devm_kfree(ice_hw_to_dev(hw), s_rule);
 
 
1217	return status;
1218}
1219
1220/**
1221 * ice_find_vlan_entry
1222 * @hw: pointer to the hardware structure
1223 * @vlan_id: VLAN id to search for
1224 *
1225 * Helper function to search for a VLAN entry using a given VLAN id
1226 * Returns pointer to the entry if found.
1227 */
1228static struct ice_fltr_mgmt_list_entry *
1229ice_find_vlan_entry(struct ice_hw *hw, u16 vlan_id)
1230{
1231	struct ice_fltr_mgmt_list_entry *vlan_list_itr, *vlan_ret = NULL;
1232	struct ice_switch_info *sw = hw->switch_info;
1233
1234	mutex_lock(&sw->vlan_list_lock);
1235	list_for_each_entry(vlan_list_itr, &sw->vlan_list_head, list_entry)
1236		if (vlan_list_itr->fltr_info.l_data.vlan.vlan_id == vlan_id) {
1237			vlan_ret = vlan_list_itr;
1238			break;
1239		}
1240
1241	mutex_unlock(&sw->vlan_list_lock);
1242	return vlan_ret;
1243}
1244
1245/**
1246 * ice_add_vlan_internal - Add one VLAN based filter rule
1247 * @hw: pointer to the hardware structure
1248 * @f_entry: filter entry containing one VLAN information
1249 */
1250static enum ice_status
1251ice_add_vlan_internal(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1252{
1253	struct ice_fltr_info *new_fltr, *cur_fltr;
1254	struct ice_fltr_mgmt_list_entry *v_list_itr;
1255	u16 vlan_id;
 
 
 
 
 
 
 
1256
 
 
1257	new_fltr = &f_entry->fltr_info;
1258	/* VLAN id should only be 12 bits */
 
1259	if (new_fltr->l_data.vlan.vlan_id > ICE_MAX_VLAN_ID)
1260		return ICE_ERR_PARAM;
1261
1262	vlan_id = new_fltr->l_data.vlan.vlan_id;
1263	v_list_itr = ice_find_vlan_entry(hw, vlan_id);
 
 
 
 
 
 
 
1264	if (!v_list_itr) {
1265		u16 vsi_id = ICE_VSI_INVAL_ID;
1266		enum ice_status status;
1267		u16 vsi_list_id = 0;
1268
1269		if (new_fltr->fltr_act == ICE_FWD_TO_VSI) {
1270			enum ice_sw_lkup_type lkup_type = new_fltr->lkup_type;
1271
1272			/* All VLAN pruning rules use a VSI list.
1273			 * Convert the action to forwarding to a VSI list.
1274			 */
1275			vsi_id = new_fltr->fwd_id.vsi_id;
1276			status = ice_create_vsi_list_rule(hw, &vsi_id, 1,
1277							  &vsi_list_id,
1278							  lkup_type);
1279			if (status)
1280				return status;
 
 
 
 
 
 
 
1281			new_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1282			new_fltr->fwd_id.vsi_list_id = vsi_list_id;
1283		}
1284
1285		status = ice_create_pkt_fwd_rule(hw, f_entry);
1286		if (!status && vsi_id != ICE_VSI_INVAL_ID) {
1287			v_list_itr = ice_find_vlan_entry(hw, vlan_id);
1288			if (!v_list_itr)
1289				return ICE_ERR_DOES_NOT_EXIST;
1290			v_list_itr->vsi_list_info =
1291				ice_create_vsi_list_map(hw, &vsi_id, 1,
1292							vsi_list_id);
 
 
 
 
 
 
 
 
 
1293		}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1294
1295		return status;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1296	}
1297
1298	cur_fltr = &v_list_itr->fltr_info;
1299	return ice_handle_vsi_list_mgmt(hw, v_list_itr, cur_fltr, new_fltr);
 
1300}
1301
1302/**
1303 * ice_add_vlan - Add VLAN based filter rule
1304 * @hw: pointer to the hardware structure
1305 * @v_list: list of VLAN entries and forwarding information
1306 */
1307enum ice_status
1308ice_add_vlan(struct ice_hw *hw, struct list_head *v_list)
1309{
1310	struct ice_fltr_list_entry *v_list_itr;
1311
1312	if (!v_list || !hw)
1313		return ICE_ERR_PARAM;
1314
1315	list_for_each_entry(v_list_itr, v_list, list_entry) {
1316		enum ice_status status;
1317
1318		if (v_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_VLAN)
1319			return ICE_ERR_PARAM;
1320
1321		status = ice_add_vlan_internal(hw, v_list_itr);
1322		if (status) {
1323			v_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1324			return status;
1325		}
1326		v_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
1327	}
1328	return 0;
1329}
1330
1331/**
1332 * ice_remove_vsi_list_rule
1333 * @hw: pointer to the hardware structure
1334 * @vsi_list_id: VSI list id generated as part of allocate resource
1335 * @lkup_type: switch rule filter lookup type
 
 
 
1336 */
1337static enum ice_status
1338ice_remove_vsi_list_rule(struct ice_hw *hw, u16 vsi_list_id,
1339			 enum ice_sw_lkup_type lkup_type)
1340{
1341	struct ice_aqc_sw_rules_elem *s_rule;
1342	enum ice_status status;
1343	u16 s_rule_size;
1344
1345	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(0);
1346	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1347	if (!s_rule)
1348		return ICE_ERR_NO_MEMORY;
1349
1350	s_rule->type = cpu_to_le16(ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR);
1351	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1352	/* FW expects number of VSIs in vsi_list resource to be 0 for clear
1353	 * command. Since memory is zero'ed out during initialization, it's not
1354	 * necessary to explicitly initialize the variable to 0.
1355	 */
1356
1357	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1,
1358				 ice_aqc_opc_remove_sw_rules, NULL);
1359	if (!status)
1360		/* Free the vsi_list resource that we allocated */
1361		status = ice_aq_alloc_free_vsi_list(hw, &vsi_list_id, lkup_type,
1362						    ice_aqc_opc_free_res);
1363
1364	devm_kfree(ice_hw_to_dev(hw), s_rule);
1365	return status;
 
 
 
 
1366}
1367
1368/**
1369 * ice_handle_rem_vsi_list_mgmt
1370 * @hw: pointer to the hardware structure
1371 * @vsi_id: ID of the VSI to remove
1372 * @fm_list_itr: filter management entry for which the VSI list management
1373 * needs to be done
1374 */
1375static enum ice_status
1376ice_handle_rem_vsi_list_mgmt(struct ice_hw *hw, u16 vsi_id,
1377			     struct ice_fltr_mgmt_list_entry *fm_list_itr)
1378{
1379	struct ice_switch_info *sw = hw->switch_info;
1380	enum ice_status status = 0;
1381	enum ice_sw_lkup_type lkup_type;
1382	bool is_last_elem = true;
1383	bool conv_list = false;
1384	bool del_list = false;
1385	u16 vsi_list_id;
1386
1387	lkup_type = fm_list_itr->fltr_info.lkup_type;
1388	vsi_list_id = fm_list_itr->fltr_info.fwd_id.vsi_list_id;
1389
1390	if (fm_list_itr->vsi_count > 1) {
1391		status = ice_update_vsi_list_rule(hw, &vsi_id, 1, vsi_list_id,
1392						  true,
1393						  ice_aqc_opc_update_sw_rules,
1394						  lkup_type);
1395		if (status)
1396			return status;
1397		fm_list_itr->vsi_count--;
1398		is_last_elem = false;
1399		clear_bit(vsi_id, fm_list_itr->vsi_list_info->vsi_map);
1400	}
1401
1402	/* For non-VLAN rules that forward packets to a VSI list, convert them
1403	 * to forwarding packets to a VSI if there is only one VSI left in the
1404	 * list.  Unused lists are then removed.
1405	 * VLAN rules need to use VSI lists even with only one VSI.
1406	 */
1407	if (fm_list_itr->fltr_info.fltr_act == ICE_FWD_TO_VSI_LIST) {
1408		if (lkup_type == ICE_SW_LKUP_VLAN) {
1409			del_list = is_last_elem;
1410		} else if (fm_list_itr->vsi_count == 1) {
1411			conv_list = true;
1412			del_list = true;
1413		}
1414	}
1415
1416	if (del_list) {
1417		/* Remove the VSI list since it is no longer used */
1418		struct ice_vsi_list_map_info *vsi_list_info =
1419			fm_list_itr->vsi_list_info;
1420
1421		status = ice_remove_vsi_list_rule(hw, vsi_list_id, lkup_type);
1422		if (status)
1423			return status;
1424
1425		if (conv_list) {
1426			u16 rem_vsi_id;
1427
1428			rem_vsi_id = find_first_bit(vsi_list_info->vsi_map,
1429						    ICE_MAX_VSI);
1430
1431			/* Error out when the expected last element is not in
1432			 * the VSI list map
1433			 */
1434			if (rem_vsi_id == ICE_MAX_VSI)
1435				return ICE_ERR_OUT_OF_RANGE;
1436
1437			/* Change the list entry action from VSI_LIST to VSI */
1438			fm_list_itr->fltr_info.fltr_act = ICE_FWD_TO_VSI;
1439			fm_list_itr->fltr_info.fwd_id.vsi_id = rem_vsi_id;
1440		}
1441
1442		list_del(&vsi_list_info->list_entry);
1443		devm_kfree(ice_hw_to_dev(hw), vsi_list_info);
1444		fm_list_itr->vsi_list_info = NULL;
1445	}
1446
1447	if (conv_list) {
1448		/* Convert the rule's forward action to forwarding packets to
1449		 * a VSI
1450		 */
1451		struct ice_aqc_sw_rules_elem *s_rule;
1452
1453		s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1454				      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE,
1455				      GFP_KERNEL);
1456		if (!s_rule)
1457			return ICE_ERR_NO_MEMORY;
1458
1459		ice_fill_sw_rule(hw, &fm_list_itr->fltr_info, s_rule,
1460				 ice_aqc_opc_update_sw_rules);
1461
1462		s_rule->pdata.lkup_tx_rx.index =
1463			cpu_to_le16(fm_list_itr->fltr_info.fltr_rule_id);
1464
1465		status = ice_aq_sw_rules(hw, s_rule,
1466					 ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1467					 ice_aqc_opc_update_sw_rules, NULL);
1468		devm_kfree(ice_hw_to_dev(hw), s_rule);
1469		if (status)
1470			return status;
1471	}
1472
1473	if (is_last_elem) {
1474		/* Remove the lookup rule */
1475		struct ice_aqc_sw_rules_elem *s_rule;
1476
1477		s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1478				      ICE_SW_RULE_RX_TX_NO_HDR_SIZE,
1479				      GFP_KERNEL);
1480		if (!s_rule)
1481			return ICE_ERR_NO_MEMORY;
1482
1483		ice_fill_sw_rule(hw, &fm_list_itr->fltr_info, s_rule,
1484				 ice_aqc_opc_remove_sw_rules);
 
1485
1486		status = ice_aq_sw_rules(hw, s_rule,
1487					 ICE_SW_RULE_RX_TX_NO_HDR_SIZE, 1,
1488					 ice_aqc_opc_remove_sw_rules, NULL);
1489		if (status)
1490			return status;
1491
1492		/* Remove a book keeping entry from the MAC address list */
1493		mutex_lock(&sw->mac_list_lock);
1494		list_del(&fm_list_itr->list_entry);
1495		mutex_unlock(&sw->mac_list_lock);
1496		devm_kfree(ice_hw_to_dev(hw), fm_list_itr);
1497		devm_kfree(ice_hw_to_dev(hw), s_rule);
1498	}
1499	return status;
1500}
1501
1502/**
1503 * ice_remove_mac_entry
1504 * @hw: pointer to the hardware structure
1505 * @f_entry: structure containing MAC forwarding information
1506 */
1507static enum ice_status
1508ice_remove_mac_entry(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1509{
1510	struct ice_fltr_mgmt_list_entry *m_entry;
1511	u16 vsi_id;
1512	u8 *add;
1513
1514	add = &f_entry->fltr_info.l_data.mac.mac_addr[0];
1515
1516	m_entry = ice_find_mac_entry(hw, add);
1517	if (!m_entry)
1518		return ICE_ERR_PARAM;
1519
1520	vsi_id = f_entry->fltr_info.fwd_id.vsi_id;
1521	return ice_handle_rem_vsi_list_mgmt(hw, vsi_id, m_entry);
1522}
1523
1524/**
1525 * ice_remove_mac - remove a MAC address based filter rule
1526 * @hw: pointer to the hardware structure
1527 * @m_list: list of MAC addresses and forwarding information
1528 *
1529 * This function removes either a MAC filter rule or a specific VSI from a
1530 * VSI list for a multicast MAC address.
1531 *
1532 * Returns ICE_ERR_DOES_NOT_EXIST if a given entry was not added by
1533 * ice_add_mac. Caller should be aware that this call will only work if all
1534 * the entries passed into m_list were added previously. It will not attempt to
1535 * do a partial remove of entries that were found.
1536 */
1537enum ice_status
1538ice_remove_mac(struct ice_hw *hw, struct list_head *m_list)
1539{
1540	struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1541	u8 s_rule_size = ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
1542	struct ice_switch_info *sw = hw->switch_info;
1543	struct ice_fltr_mgmt_list_entry *m_entry;
1544	struct ice_fltr_list_entry *m_list_itr;
1545	u16 elem_sent, total_elem_left;
1546	enum ice_status status = 0;
1547	u16 num_unicast = 0;
1548
1549	if (!m_list)
1550		return ICE_ERR_PARAM;
1551
1552	list_for_each_entry(m_list_itr, m_list, list_entry) {
1553		u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1554
1555		if (is_unicast_ether_addr(addr) && !hw->ucast_shared)
1556			num_unicast++;
1557		else if (is_multicast_ether_addr(addr) ||
1558			 (is_unicast_ether_addr(addr) && hw->ucast_shared))
1559			ice_remove_mac_entry(hw, m_list_itr);
1560	}
1561
1562	/* Exit if no unicast addresses found. Multicast switch rules
1563	 * were added individually
1564	 */
1565	if (!num_unicast)
1566		return 0;
1567
1568	/* Allocate switch rule buffer for the bulk update for unicast */
1569	s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1570			      GFP_KERNEL);
1571	if (!s_rule)
1572		return ICE_ERR_NO_MEMORY;
1573
1574	r_iter = s_rule;
1575	list_for_each_entry(m_list_itr, m_list, list_entry) {
1576		u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1577
1578		if (is_unicast_ether_addr(addr)) {
1579			m_entry = ice_find_mac_entry(hw, addr);
1580			if (!m_entry) {
1581				status = ICE_ERR_DOES_NOT_EXIST;
1582				goto ice_remove_mac_exit;
1583			}
1584
1585			ice_fill_sw_rule(hw, &m_entry->fltr_info,
1586					 r_iter, ice_aqc_opc_remove_sw_rules);
1587			r_iter = (struct ice_aqc_sw_rules_elem *)
1588				((u8 *)r_iter + s_rule_size);
1589		}
1590	}
1591
1592	/* Call AQ bulk switch rule update for all unicast addresses */
1593	r_iter = s_rule;
1594	/* Call AQ switch rule in AQ_MAX chunk */
1595	for (total_elem_left = num_unicast; total_elem_left > 0;
1596	     total_elem_left -= elem_sent) {
1597		struct ice_aqc_sw_rules_elem *entry = r_iter;
1598
1599		elem_sent = min(total_elem_left,
1600				(u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1601		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1602					 elem_sent, ice_aqc_opc_remove_sw_rules,
1603					 NULL);
1604		if (status)
1605			break;
1606		r_iter = (struct ice_aqc_sw_rules_elem *)
1607			((u8 *)r_iter + s_rule_size);
1608	}
1609
1610	list_for_each_entry(m_list_itr, m_list, list_entry) {
1611		u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1612
1613		if (is_unicast_ether_addr(addr)) {
1614			m_entry = ice_find_mac_entry(hw, addr);
1615			if (!m_entry)
1616				return ICE_ERR_OUT_OF_RANGE;
1617			mutex_lock(&sw->mac_list_lock);
1618			list_del(&m_entry->list_entry);
1619			mutex_unlock(&sw->mac_list_lock);
1620			devm_kfree(ice_hw_to_dev(hw), m_entry);
1621		}
1622	}
1623
1624ice_remove_mac_exit:
1625	devm_kfree(ice_hw_to_dev(hw), s_rule);
1626	return status;
1627}
1628
1629/**
1630 * ice_cfg_dflt_vsi - add filter rule to set/unset given VSI as default
1631 * VSI for the switch (represented by swid)
1632 * @hw: pointer to the hardware structure
1633 * @vsi_id: number of VSI to set as default
1634 * @set: true to add the above mentioned switch rule, false to remove it
1635 * @direction: ICE_FLTR_RX or ICE_FLTR_TX
 
 
 
1636 */
1637enum ice_status
1638ice_cfg_dflt_vsi(struct ice_hw *hw, u16 vsi_id, bool set, u8 direction)
1639{
1640	struct ice_aqc_sw_rules_elem *s_rule;
1641	struct ice_fltr_info f_info;
1642	enum ice_adminq_opc opcode;
1643	enum ice_status status;
1644	u16 s_rule_size;
 
 
 
 
 
1645
1646	s_rule_size = set ? ICE_SW_RULE_RX_TX_ETH_HDR_SIZE :
1647			    ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
1648	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1649	if (!s_rule)
1650		return ICE_ERR_NO_MEMORY;
1651
1652	memset(&f_info, 0, sizeof(f_info));
1653
1654	f_info.lkup_type = ICE_SW_LKUP_DFLT;
1655	f_info.flag = direction;
1656	f_info.fltr_act = ICE_FWD_TO_VSI;
1657	f_info.fwd_id.vsi_id = vsi_id;
1658
1659	if (f_info.flag & ICE_FLTR_RX) {
1660		f_info.src = hw->port_info->lport;
 
1661		if (!set)
1662			f_info.fltr_rule_id =
1663				hw->port_info->dflt_rx_vsi_rule_id;
1664	} else if (f_info.flag & ICE_FLTR_TX) {
1665		f_info.src = vsi_id;
 
1666		if (!set)
1667			f_info.fltr_rule_id =
1668				hw->port_info->dflt_tx_vsi_rule_id;
1669	}
1670
1671	if (set)
1672		opcode = ice_aqc_opc_add_sw_rules;
1673	else
1674		opcode = ice_aqc_opc_remove_sw_rules;
1675
1676	ice_fill_sw_rule(hw, &f_info, s_rule, opcode);
1677
1678	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opcode, NULL);
1679	if (status || !(f_info.flag & ICE_FLTR_TX_RX))
1680		goto out;
1681	if (set) {
1682		u16 index = le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1683
1684		if (f_info.flag & ICE_FLTR_TX) {
1685			hw->port_info->dflt_tx_vsi_num = vsi_id;
1686			hw->port_info->dflt_tx_vsi_rule_id = index;
1687		} else if (f_info.flag & ICE_FLTR_RX) {
1688			hw->port_info->dflt_rx_vsi_num = vsi_id;
1689			hw->port_info->dflt_rx_vsi_rule_id = index;
1690		}
1691	} else {
1692		if (f_info.flag & ICE_FLTR_TX) {
1693			hw->port_info->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
1694			hw->port_info->dflt_tx_vsi_rule_id = ICE_INVAL_ACT;
1695		} else if (f_info.flag & ICE_FLTR_RX) {
1696			hw->port_info->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
1697			hw->port_info->dflt_rx_vsi_rule_id = ICE_INVAL_ACT;
1698		}
1699	}
1700
1701out:
1702	devm_kfree(ice_hw_to_dev(hw), s_rule);
1703	return status;
1704}
1705
1706/**
1707 * ice_remove_vlan_internal - Remove one VLAN based filter rule
1708 * @hw: pointer to the hardware structure
1709 * @f_entry: filter entry containing one VLAN information
 
 
 
 
 
 
 
1710 */
1711static enum ice_status
1712ice_remove_vlan_internal(struct ice_hw *hw,
1713			 struct ice_fltr_list_entry *f_entry)
1714{
1715	struct ice_fltr_info *new_fltr;
1716	struct ice_fltr_mgmt_list_entry *v_list_elem;
1717	u16 vsi_id;
1718
1719	new_fltr = &f_entry->fltr_info;
 
 
 
 
 
 
 
 
 
 
1720
1721	v_list_elem = ice_find_vlan_entry(hw, new_fltr->l_data.vlan.vlan_id);
1722	if (!v_list_elem)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1723		return ICE_ERR_PARAM;
1724
1725	vsi_id = f_entry->fltr_info.fwd_id.vsi_id;
1726	return ice_handle_rem_vsi_list_mgmt(hw, vsi_id, v_list_elem);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1727}
1728
1729/**
1730 * ice_remove_vlan - Remove VLAN based filter rule
1731 * @hw: pointer to the hardware structure
1732 * @v_list: list of VLAN entries and forwarding information
1733 */
1734enum ice_status
1735ice_remove_vlan(struct ice_hw *hw, struct list_head *v_list)
1736{
1737	struct ice_fltr_list_entry *v_list_itr;
1738	enum ice_status status = 0;
1739
1740	if (!v_list || !hw)
1741		return ICE_ERR_PARAM;
1742
1743	list_for_each_entry(v_list_itr, v_list, list_entry) {
1744		status = ice_remove_vlan_internal(hw, v_list_itr);
1745		if (status) {
1746			v_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1747			return status;
1748		}
1749		v_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
 
 
 
1750	}
1751	return status;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1752}
1753
1754/**
1755 * ice_add_to_vsi_fltr_list - Add VSI filters to the list
1756 * @hw: pointer to the hardware structure
1757 * @vsi_id: ID of VSI to remove filters from
1758 * @lkup_list_head: pointer to the list that has certain lookup type filters
1759 * @vsi_list_head: pointer to the list pertaining to VSI with vsi_id
 
 
 
 
 
 
1760 */
1761static enum ice_status
1762ice_add_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_id,
1763			 struct list_head *lkup_list_head,
1764			 struct list_head *vsi_list_head)
1765{
1766	struct ice_fltr_mgmt_list_entry *fm_entry;
 
1767
1768	/* check to make sure VSI id is valid and within boundary */
1769	if (vsi_id >=
1770	    (sizeof(fm_entry->vsi_list_info->vsi_map) * BITS_PER_BYTE - 1))
1771		return ICE_ERR_PARAM;
1772
1773	list_for_each_entry(fm_entry, lkup_list_head, list_entry) {
1774		struct ice_fltr_info *fi;
1775
1776		fi = &fm_entry->fltr_info;
1777		if ((fi->fltr_act == ICE_FWD_TO_VSI &&
1778		     fi->fwd_id.vsi_id == vsi_id) ||
1779		    (fi->fltr_act == ICE_FWD_TO_VSI_LIST &&
1780		     (test_bit(vsi_id, fm_entry->vsi_list_info->vsi_map)))) {
1781			struct ice_fltr_list_entry *tmp;
1782
1783			/* this memory is freed up in the caller function
1784			 * ice_remove_vsi_lkup_fltr() once filters for
1785			 * this VSI are removed
1786			 */
1787			tmp = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*tmp),
1788					   GFP_KERNEL);
1789			if (!tmp)
1790				return ICE_ERR_NO_MEMORY;
1791
1792			memcpy(&tmp->fltr_info, fi, sizeof(*fi));
1793
1794			/* Expected below fields to be set to ICE_FWD_TO_VSI and
1795			 * the particular VSI id since we are only removing this
1796			 * one VSI
1797			 */
1798			if (fi->fltr_act == ICE_FWD_TO_VSI_LIST) {
1799				tmp->fltr_info.fltr_act = ICE_FWD_TO_VSI;
1800				tmp->fltr_info.fwd_id.vsi_id = vsi_id;
1801			}
1802
1803			list_add(&tmp->list_entry, vsi_list_head);
1804		}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1805	}
1806	return 0;
1807}
1808
1809/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1810 * ice_remove_vsi_lkup_fltr - Remove lookup type filters for a VSI
1811 * @hw: pointer to the hardware structure
1812 * @vsi_id: ID of VSI to remove filters from
1813 * @lkup: switch rule filter lookup type
1814 */
1815static void
1816ice_remove_vsi_lkup_fltr(struct ice_hw *hw, u16 vsi_id,
1817			 enum ice_sw_lkup_type lkup)
1818{
1819	struct ice_switch_info *sw = hw->switch_info;
1820	struct ice_fltr_list_entry *fm_entry;
1821	struct list_head remove_list_head;
 
1822	struct ice_fltr_list_entry *tmp;
 
1823	enum ice_status status;
1824
1825	INIT_LIST_HEAD(&remove_list_head);
 
 
 
 
 
 
 
 
 
1826	switch (lkup) {
1827	case ICE_SW_LKUP_MAC:
1828		mutex_lock(&sw->mac_list_lock);
1829		status = ice_add_to_vsi_fltr_list(hw, vsi_id,
1830						  &sw->mac_list_head,
1831						  &remove_list_head);
1832		mutex_unlock(&sw->mac_list_lock);
1833		if (!status) {
1834			ice_remove_mac(hw, &remove_list_head);
1835			goto free_fltr_list;
1836		}
1837		break;
1838	case ICE_SW_LKUP_VLAN:
1839		mutex_lock(&sw->vlan_list_lock);
1840		status = ice_add_to_vsi_fltr_list(hw, vsi_id,
1841						  &sw->vlan_list_head,
1842						  &remove_list_head);
1843		mutex_unlock(&sw->vlan_list_lock);
1844		if (!status) {
1845			ice_remove_vlan(hw, &remove_list_head);
1846			goto free_fltr_list;
1847		}
1848		break;
1849	case ICE_SW_LKUP_MAC_VLAN:
1850	case ICE_SW_LKUP_ETHERTYPE:
1851	case ICE_SW_LKUP_ETHERTYPE_MAC:
1852	case ICE_SW_LKUP_PROMISC:
1853	case ICE_SW_LKUP_PROMISC_VLAN:
1854	case ICE_SW_LKUP_DFLT:
1855		ice_debug(hw, ICE_DBG_SW,
1856			  "Remove filters for this lookup type hasn't been implemented yet\n");
 
1857		break;
1858	}
1859
1860	return;
1861free_fltr_list:
1862	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
1863		list_del(&fm_entry->list_entry);
1864		devm_kfree(ice_hw_to_dev(hw), fm_entry);
1865	}
1866}
1867
1868/**
1869 * ice_remove_vsi_fltr - Remove all filters for a VSI
1870 * @hw: pointer to the hardware structure
1871 * @vsi_id: ID of VSI to remove filters from
1872 */
1873void ice_remove_vsi_fltr(struct ice_hw *hw, u16 vsi_id)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1874{
1875	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_MAC);
1876	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_MAC_VLAN);
1877	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_PROMISC);
1878	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_VLAN);
1879	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_DFLT);
1880	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_ETHERTYPE);
1881	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_ETHERTYPE_MAC);
1882	ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_PROMISC_VLAN);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1883}
v5.4
   1// SPDX-License-Identifier: GPL-2.0
   2/* Copyright (c) 2018, Intel Corporation. */
   3
   4#include "ice_switch.h"
   5
   6#define ICE_ETH_DA_OFFSET		0
   7#define ICE_ETH_ETHTYPE_OFFSET		12
   8#define ICE_ETH_VLAN_TCI_OFFSET		14
   9#define ICE_MAX_VLAN_ID			0xFFF
  10
  11/* Dummy ethernet header needed in the ice_aqc_sw_rules_elem
  12 * struct to configure any switch filter rules.
  13 * {DA (6 bytes), SA(6 bytes),
  14 * Ether type (2 bytes for header without VLAN tag) OR
  15 * VLAN tag (4 bytes for header with VLAN tag) }
  16 *
  17 * Word on Hardcoded values
  18 * byte 0 = 0x2: to identify it as locally administered DA MAC
  19 * byte 6 = 0x2: to identify it as locally administered SA MAC
  20 * byte 12 = 0x81 & byte 13 = 0x00:
  21 *	In case of VLAN filter first two bytes defines ether type (0x8100)
  22 *	and remaining two bytes are placeholder for programming a given VLAN ID
  23 *	In case of Ether type filter it is treated as header without VLAN tag
  24 *	and byte 12 and 13 is used to program a given Ether type instead
  25 */
  26#define DUMMY_ETH_HDR_LEN		16
  27static const u8 dummy_eth_header[DUMMY_ETH_HDR_LEN] = { 0x2, 0, 0, 0, 0, 0,
  28							0x2, 0, 0, 0, 0, 0,
  29							0x81, 0, 0, 0};
  30
  31#define ICE_SW_RULE_RX_TX_ETH_HDR_SIZE \
  32	(sizeof(struct ice_aqc_sw_rules_elem) - \
  33	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  34	 sizeof(struct ice_sw_rule_lkup_rx_tx) + DUMMY_ETH_HDR_LEN - 1)
  35#define ICE_SW_RULE_RX_TX_NO_HDR_SIZE \
  36	(sizeof(struct ice_aqc_sw_rules_elem) - \
  37	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  38	 sizeof(struct ice_sw_rule_lkup_rx_tx) - 1)
  39#define ICE_SW_RULE_LG_ACT_SIZE(n) \
  40	(sizeof(struct ice_aqc_sw_rules_elem) - \
  41	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  42	 sizeof(struct ice_sw_rule_lg_act) - \
  43	 sizeof(((struct ice_sw_rule_lg_act *)0)->act) + \
  44	 ((n) * sizeof(((struct ice_sw_rule_lg_act *)0)->act)))
  45#define ICE_SW_RULE_VSI_LIST_SIZE(n) \
  46	(sizeof(struct ice_aqc_sw_rules_elem) - \
  47	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  48	 sizeof(struct ice_sw_rule_vsi_list) - \
  49	 sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi) + \
  50	 ((n) * sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi)))
  51
  52/**
  53 * ice_aq_alloc_free_res - command to allocate/free resources
  54 * @hw: pointer to the HW struct
  55 * @num_entries: number of resource entries in buffer
  56 * @buf: Indirect buffer to hold data parameters and response
  57 * @buf_size: size of buffer for indirect commands
  58 * @opc: pass in the command opcode
  59 * @cd: pointer to command details structure or NULL
  60 *
  61 * Helper function to allocate/free resources using the admin queue commands
  62 */
  63static enum ice_status
  64ice_aq_alloc_free_res(struct ice_hw *hw, u16 num_entries,
  65		      struct ice_aqc_alloc_free_res_elem *buf, u16 buf_size,
  66		      enum ice_adminq_opc opc, struct ice_sq_cd *cd)
  67{
  68	struct ice_aqc_alloc_free_res_cmd *cmd;
  69	struct ice_aq_desc desc;
  70
  71	cmd = &desc.params.sw_res_ctrl;
  72
  73	if (!buf)
  74		return ICE_ERR_PARAM;
  75
  76	if (buf_size < (num_entries * sizeof(buf->elem[0])))
  77		return ICE_ERR_PARAM;
  78
  79	ice_fill_dflt_direct_cmd_desc(&desc, opc);
  80
  81	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
  82
  83	cmd->num_entries = cpu_to_le16(num_entries);
  84
  85	return ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
  86}
  87
  88/**
  89 * ice_init_def_sw_recp - initialize the recipe book keeping tables
  90 * @hw: pointer to the HW struct
  91 *
  92 * Allocate memory for the entire recipe table and initialize the structures/
  93 * entries corresponding to basic recipes.
  94 */
  95enum ice_status ice_init_def_sw_recp(struct ice_hw *hw)
  96{
  97	struct ice_sw_recipe *recps;
  98	u8 i;
  99
 100	recps = devm_kcalloc(ice_hw_to_dev(hw), ICE_MAX_NUM_RECIPES,
 101			     sizeof(*recps), GFP_KERNEL);
 102	if (!recps)
 103		return ICE_ERR_NO_MEMORY;
 104
 105	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
 106		recps[i].root_rid = i;
 107		INIT_LIST_HEAD(&recps[i].filt_rules);
 108		INIT_LIST_HEAD(&recps[i].filt_replay_rules);
 109		mutex_init(&recps[i].filt_rule_lock);
 110	}
 111
 112	hw->switch_info->recp_list = recps;
 113
 114	return 0;
 115}
 116
 117/**
 118 * ice_aq_get_sw_cfg - get switch configuration
 119 * @hw: pointer to the hardware structure
 120 * @buf: pointer to the result buffer
 121 * @buf_size: length of the buffer available for response
 122 * @req_desc: pointer to requested descriptor
 123 * @num_elems: pointer to number of elements
 124 * @cd: pointer to command details structure or NULL
 125 *
 126 * Get switch configuration (0x0200) to be placed in 'buff'.
 127 * This admin command returns information such as initial VSI/port number
 128 * and switch ID it belongs to.
 129 *
 130 * NOTE: *req_desc is both an input/output parameter.
 131 * The caller of this function first calls this function with *request_desc set
 132 * to 0. If the response from f/w has *req_desc set to 0, all the switch
 133 * configuration information has been returned; if non-zero (meaning not all
 134 * the information was returned), the caller should call this function again
 135 * with *req_desc set to the previous value returned by f/w to get the
 136 * next block of switch configuration information.
 137 *
 138 * *num_elems is output only parameter. This reflects the number of elements
 139 * in response buffer. The caller of this function to use *num_elems while
 140 * parsing the response buffer.
 141 */
 142static enum ice_status
 143ice_aq_get_sw_cfg(struct ice_hw *hw, struct ice_aqc_get_sw_cfg_resp *buf,
 144		  u16 buf_size, u16 *req_desc, u16 *num_elems,
 145		  struct ice_sq_cd *cd)
 146{
 147	struct ice_aqc_get_sw_cfg *cmd;
 148	enum ice_status status;
 149	struct ice_aq_desc desc;
 150
 151	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_get_sw_cfg);
 152	cmd = &desc.params.get_sw_conf;
 153	cmd->element = cpu_to_le16(*req_desc);
 154
 155	status = ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
 156	if (!status) {
 157		*req_desc = le16_to_cpu(cmd->element);
 158		*num_elems = le16_to_cpu(cmd->num_elems);
 159	}
 160
 161	return status;
 162}
 163
 164/**
 165 * ice_aq_add_vsi
 166 * @hw: pointer to the HW struct
 167 * @vsi_ctx: pointer to a VSI context struct
 168 * @cd: pointer to command details structure or NULL
 169 *
 170 * Add a VSI context to the hardware (0x0210)
 171 */
 172static enum ice_status
 173ice_aq_add_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 174	       struct ice_sq_cd *cd)
 175{
 176	struct ice_aqc_add_update_free_vsi_resp *res;
 177	struct ice_aqc_add_get_update_free_vsi *cmd;
 
 178	struct ice_aq_desc desc;
 179	enum ice_status status;
 180
 181	cmd = &desc.params.vsi_cmd;
 182	res = &desc.params.add_update_free_vsi_res;
 183
 184	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_add_vsi);
 185
 186	if (!vsi_ctx->alloc_from_pool)
 187		cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num |
 188					   ICE_AQ_VSI_IS_VALID);
 189	cmd->vf_id = vsi_ctx->vf_num;
 190
 191	cmd->vsi_flags = cpu_to_le16(vsi_ctx->flags);
 192
 193	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 194
 195	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 196				 sizeof(vsi_ctx->info), cd);
 197
 198	if (!status) {
 199		vsi_ctx->vsi_num = le16_to_cpu(res->vsi_num) & ICE_AQ_VSI_NUM_M;
 200		vsi_ctx->vsis_allocd = le16_to_cpu(res->vsi_used);
 201		vsi_ctx->vsis_unallocated = le16_to_cpu(res->vsi_free);
 202	}
 203
 204	return status;
 205}
 206
 207/**
 208 * ice_aq_free_vsi
 209 * @hw: pointer to the HW struct
 210 * @vsi_ctx: pointer to a VSI context struct
 211 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 212 * @cd: pointer to command details structure or NULL
 213 *
 214 * Free VSI context info from hardware (0x0213)
 215 */
 216static enum ice_status
 217ice_aq_free_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 218		bool keep_vsi_alloc, struct ice_sq_cd *cd)
 219{
 220	struct ice_aqc_add_update_free_vsi_resp *resp;
 221	struct ice_aqc_add_get_update_free_vsi *cmd;
 222	struct ice_aq_desc desc;
 223	enum ice_status status;
 224
 225	cmd = &desc.params.vsi_cmd;
 226	resp = &desc.params.add_update_free_vsi_res;
 227
 228	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_free_vsi);
 229
 230	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 231	if (keep_vsi_alloc)
 232		cmd->cmd_flags = cpu_to_le16(ICE_AQ_VSI_KEEP_ALLOC);
 233
 234	status = ice_aq_send_cmd(hw, &desc, NULL, 0, cd);
 235	if (!status) {
 236		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 237		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 238	}
 239
 240	return status;
 241}
 242
 243/**
 244 * ice_aq_update_vsi
 245 * @hw: pointer to the HW struct
 246 * @vsi_ctx: pointer to a VSI context struct
 247 * @cd: pointer to command details structure or NULL
 248 *
 249 * Update VSI context in the hardware (0x0211)
 250 */
 251static enum ice_status
 252ice_aq_update_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 253		  struct ice_sq_cd *cd)
 254{
 255	struct ice_aqc_add_update_free_vsi_resp *resp;
 256	struct ice_aqc_add_get_update_free_vsi *cmd;
 257	struct ice_aq_desc desc;
 258	enum ice_status status;
 259
 260	cmd = &desc.params.vsi_cmd;
 261	resp = &desc.params.add_update_free_vsi_res;
 262
 263	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_update_vsi);
 264
 265	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 266
 267	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 268
 269	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 270				 sizeof(vsi_ctx->info), cd);
 271
 272	if (!status) {
 273		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 274		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 275	}
 276
 277	return status;
 278}
 279
 280/**
 281 * ice_is_vsi_valid - check whether the VSI is valid or not
 282 * @hw: pointer to the HW struct
 283 * @vsi_handle: VSI handle
 284 *
 285 * check whether the VSI is valid or not
 286 */
 287bool ice_is_vsi_valid(struct ice_hw *hw, u16 vsi_handle)
 288{
 289	return vsi_handle < ICE_MAX_VSI && hw->vsi_ctx[vsi_handle];
 290}
 291
 292/**
 293 * ice_get_hw_vsi_num - return the HW VSI number
 294 * @hw: pointer to the HW struct
 295 * @vsi_handle: VSI handle
 296 *
 297 * return the HW VSI number
 298 * Caution: call this function only if VSI is valid (ice_is_vsi_valid)
 299 */
 300u16 ice_get_hw_vsi_num(struct ice_hw *hw, u16 vsi_handle)
 301{
 302	return hw->vsi_ctx[vsi_handle]->vsi_num;
 303}
 304
 305/**
 306 * ice_get_vsi_ctx - return the VSI context entry for a given VSI handle
 307 * @hw: pointer to the HW struct
 308 * @vsi_handle: VSI handle
 309 *
 310 * return the VSI context entry for a given VSI handle
 311 */
 312struct ice_vsi_ctx *ice_get_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 313{
 314	return (vsi_handle >= ICE_MAX_VSI) ? NULL : hw->vsi_ctx[vsi_handle];
 315}
 316
 317/**
 318 * ice_save_vsi_ctx - save the VSI context for a given VSI handle
 319 * @hw: pointer to the HW struct
 320 * @vsi_handle: VSI handle
 321 * @vsi: VSI context pointer
 322 *
 323 * save the VSI context entry for a given VSI handle
 324 */
 325static void
 326ice_save_vsi_ctx(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi)
 327{
 328	hw->vsi_ctx[vsi_handle] = vsi;
 329}
 330
 331/**
 332 * ice_clear_vsi_q_ctx - clear VSI queue contexts for all TCs
 333 * @hw: pointer to the HW struct
 334 * @vsi_handle: VSI handle
 335 */
 336static void ice_clear_vsi_q_ctx(struct ice_hw *hw, u16 vsi_handle)
 337{
 338	struct ice_vsi_ctx *vsi;
 339	u8 i;
 340
 341	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 342	if (!vsi)
 343		return;
 344	ice_for_each_traffic_class(i) {
 345		if (vsi->lan_q_ctx[i]) {
 346			devm_kfree(ice_hw_to_dev(hw), vsi->lan_q_ctx[i]);
 347			vsi->lan_q_ctx[i] = NULL;
 348		}
 349	}
 350}
 351
 352/**
 353 * ice_clear_vsi_ctx - clear the VSI context entry
 354 * @hw: pointer to the HW struct
 355 * @vsi_handle: VSI handle
 356 *
 357 * clear the VSI context entry
 358 */
 359static void ice_clear_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 360{
 361	struct ice_vsi_ctx *vsi;
 362
 363	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 364	if (vsi) {
 365		ice_clear_vsi_q_ctx(hw, vsi_handle);
 366		devm_kfree(ice_hw_to_dev(hw), vsi);
 367		hw->vsi_ctx[vsi_handle] = NULL;
 368	}
 369}
 370
 371/**
 372 * ice_clear_all_vsi_ctx - clear all the VSI context entries
 373 * @hw: pointer to the HW struct
 374 */
 375void ice_clear_all_vsi_ctx(struct ice_hw *hw)
 376{
 377	u16 i;
 378
 379	for (i = 0; i < ICE_MAX_VSI; i++)
 380		ice_clear_vsi_ctx(hw, i);
 381}
 382
 383/**
 384 * ice_add_vsi - add VSI context to the hardware and VSI handle list
 385 * @hw: pointer to the HW struct
 386 * @vsi_handle: unique VSI handle provided by drivers
 387 * @vsi_ctx: pointer to a VSI context struct
 
 388 * @cd: pointer to command details structure or NULL
 389 *
 390 * Add a VSI context to the hardware also add it into the VSI handle list.
 391 * If this function gets called after reset for existing VSIs then update
 392 * with the new HW VSI number in the corresponding VSI handle list entry.
 393 */
 394enum ice_status
 395ice_add_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 396	    struct ice_sq_cd *cd)
 397{
 398	struct ice_vsi_ctx *tmp_vsi_ctx;
 
 
 399	enum ice_status status;
 400
 401	if (vsi_handle >= ICE_MAX_VSI)
 402		return ICE_ERR_PARAM;
 403	status = ice_aq_add_vsi(hw, vsi_ctx, cd);
 404	if (status)
 405		return status;
 406	tmp_vsi_ctx = ice_get_vsi_ctx(hw, vsi_handle);
 407	if (!tmp_vsi_ctx) {
 408		/* Create a new VSI context */
 409		tmp_vsi_ctx = devm_kzalloc(ice_hw_to_dev(hw),
 410					   sizeof(*tmp_vsi_ctx), GFP_KERNEL);
 411		if (!tmp_vsi_ctx) {
 412			ice_aq_free_vsi(hw, vsi_ctx, false, cd);
 413			return ICE_ERR_NO_MEMORY;
 414		}
 415		*tmp_vsi_ctx = *vsi_ctx;
 416		ice_save_vsi_ctx(hw, vsi_handle, tmp_vsi_ctx);
 417	} else {
 418		/* update with new HW VSI num */
 419		if (tmp_vsi_ctx->vsi_num != vsi_ctx->vsi_num)
 420			tmp_vsi_ctx->vsi_num = vsi_ctx->vsi_num;
 421	}
 422
 423	return 0;
 424}
 
 425
 426/**
 427 * ice_free_vsi- free VSI context from hardware and VSI handle list
 428 * @hw: pointer to the HW struct
 429 * @vsi_handle: unique VSI handle
 430 * @vsi_ctx: pointer to a VSI context struct
 431 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 432 * @cd: pointer to command details structure or NULL
 433 *
 434 * Free VSI context info from hardware as well as from VSI handle list
 435 */
 436enum ice_status
 437ice_free_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 438	     bool keep_vsi_alloc, struct ice_sq_cd *cd)
 439{
 440	enum ice_status status;
 441
 442	if (!ice_is_vsi_valid(hw, vsi_handle))
 443		return ICE_ERR_PARAM;
 444	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 445	status = ice_aq_free_vsi(hw, vsi_ctx, keep_vsi_alloc, cd);
 446	if (!status)
 447		ice_clear_vsi_ctx(hw, vsi_handle);
 448	return status;
 449}
 450
 451/**
 452 * ice_update_vsi
 453 * @hw: pointer to the HW struct
 454 * @vsi_handle: unique VSI handle
 455 * @vsi_ctx: pointer to a VSI context struct
 456 * @cd: pointer to command details structure or NULL
 457 *
 458 * Update VSI context in the hardware
 459 */
 460enum ice_status
 461ice_update_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 462	       struct ice_sq_cd *cd)
 463{
 464	if (!ice_is_vsi_valid(hw, vsi_handle))
 465		return ICE_ERR_PARAM;
 466	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 467	return ice_aq_update_vsi(hw, vsi_ctx, cd);
 468}
 469
 470/**
 471 * ice_aq_alloc_free_vsi_list
 472 * @hw: pointer to the HW struct
 473 * @vsi_list_id: VSI list ID returned or used for lookup
 474 * @lkup_type: switch rule filter lookup type
 475 * @opc: switch rules population command type - pass in the command opcode
 476 *
 477 * allocates or free a VSI list resource
 478 */
 479static enum ice_status
 480ice_aq_alloc_free_vsi_list(struct ice_hw *hw, u16 *vsi_list_id,
 481			   enum ice_sw_lkup_type lkup_type,
 482			   enum ice_adminq_opc opc)
 483{
 484	struct ice_aqc_alloc_free_res_elem *sw_buf;
 485	struct ice_aqc_res_elem *vsi_ele;
 486	enum ice_status status;
 487	u16 buf_len;
 488
 489	buf_len = sizeof(*sw_buf);
 490	sw_buf = devm_kzalloc(ice_hw_to_dev(hw), buf_len, GFP_KERNEL);
 491	if (!sw_buf)
 492		return ICE_ERR_NO_MEMORY;
 493	sw_buf->num_elems = cpu_to_le16(1);
 494
 495	if (lkup_type == ICE_SW_LKUP_MAC ||
 496	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 497	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 498	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 499	    lkup_type == ICE_SW_LKUP_PROMISC ||
 500	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN) {
 501		sw_buf->res_type = cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_REP);
 502	} else if (lkup_type == ICE_SW_LKUP_VLAN) {
 503		sw_buf->res_type =
 504			cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_PRUNE);
 505	} else {
 506		status = ICE_ERR_PARAM;
 507		goto ice_aq_alloc_free_vsi_list_exit;
 508	}
 509
 510	if (opc == ice_aqc_opc_free_res)
 511		sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
 512
 513	status = ice_aq_alloc_free_res(hw, 1, sw_buf, buf_len, opc, NULL);
 514	if (status)
 515		goto ice_aq_alloc_free_vsi_list_exit;
 516
 517	if (opc == ice_aqc_opc_alloc_res) {
 518		vsi_ele = &sw_buf->elem[0];
 519		*vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
 520	}
 521
 522ice_aq_alloc_free_vsi_list_exit:
 523	devm_kfree(ice_hw_to_dev(hw), sw_buf);
 524	return status;
 525}
 526
 527/**
 528 * ice_aq_sw_rules - add/update/remove switch rules
 529 * @hw: pointer to the HW struct
 530 * @rule_list: pointer to switch rule population list
 531 * @rule_list_sz: total size of the rule list in bytes
 532 * @num_rules: number of switch rules in the rule_list
 533 * @opc: switch rules population command type - pass in the command opcode
 534 * @cd: pointer to command details structure or NULL
 535 *
 536 * Add(0x02a0)/Update(0x02a1)/Remove(0x02a2) switch rules commands to firmware
 537 */
 538static enum ice_status
 539ice_aq_sw_rules(struct ice_hw *hw, void *rule_list, u16 rule_list_sz,
 540		u8 num_rules, enum ice_adminq_opc opc, struct ice_sq_cd *cd)
 541{
 542	struct ice_aq_desc desc;
 543
 544	if (opc != ice_aqc_opc_add_sw_rules &&
 545	    opc != ice_aqc_opc_update_sw_rules &&
 546	    opc != ice_aqc_opc_remove_sw_rules)
 547		return ICE_ERR_PARAM;
 548
 549	ice_fill_dflt_direct_cmd_desc(&desc, opc);
 550
 551	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 552	desc.params.sw_rules.num_rules_fltr_entry_index =
 553		cpu_to_le16(num_rules);
 554	return ice_aq_send_cmd(hw, &desc, rule_list, rule_list_sz, cd);
 555}
 556
 557/* ice_init_port_info - Initialize port_info with switch configuration data
 558 * @pi: pointer to port_info
 559 * @vsi_port_num: VSI number or port number
 560 * @type: Type of switch element (port or VSI)
 561 * @swid: switch ID of the switch the element is attached to
 562 * @pf_vf_num: PF or VF number
 563 * @is_vf: true if the element is a VF, false otherwise
 564 */
 565static void
 566ice_init_port_info(struct ice_port_info *pi, u16 vsi_port_num, u8 type,
 567		   u16 swid, u16 pf_vf_num, bool is_vf)
 568{
 569	switch (type) {
 570	case ICE_AQC_GET_SW_CONF_RESP_PHYS_PORT:
 571		pi->lport = (u8)(vsi_port_num & ICE_LPORT_MASK);
 572		pi->sw_id = swid;
 573		pi->pf_vf_num = pf_vf_num;
 574		pi->is_vf = is_vf;
 575		pi->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
 576		pi->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
 577		break;
 578	default:
 579		ice_debug(pi->hw, ICE_DBG_SW,
 580			  "incorrect VSI/port type received\n");
 581		break;
 582	}
 583}
 584
 585/* ice_get_initial_sw_cfg - Get initial port and default VSI data
 586 * @hw: pointer to the hardware structure
 587 */
 588enum ice_status ice_get_initial_sw_cfg(struct ice_hw *hw)
 589{
 590	struct ice_aqc_get_sw_cfg_resp *rbuf;
 591	enum ice_status status;
 592	u16 req_desc = 0;
 593	u16 num_elems;
 594	u16 i;
 595
 596	rbuf = devm_kzalloc(ice_hw_to_dev(hw), ICE_SW_CFG_MAX_BUF_LEN,
 597			    GFP_KERNEL);
 598
 599	if (!rbuf)
 600		return ICE_ERR_NO_MEMORY;
 601
 602	/* Multiple calls to ice_aq_get_sw_cfg may be required
 603	 * to get all the switch configuration information. The need
 604	 * for additional calls is indicated by ice_aq_get_sw_cfg
 605	 * writing a non-zero value in req_desc
 606	 */
 607	do {
 608		status = ice_aq_get_sw_cfg(hw, rbuf, ICE_SW_CFG_MAX_BUF_LEN,
 609					   &req_desc, &num_elems, NULL);
 610
 611		if (status)
 612			break;
 613
 614		for (i = 0; i < num_elems; i++) {
 615			struct ice_aqc_get_sw_cfg_resp_elem *ele;
 616			u16 pf_vf_num, swid, vsi_port_num;
 617			bool is_vf = false;
 618			u8 type;
 619
 620			ele = rbuf[i].elements;
 621			vsi_port_num = le16_to_cpu(ele->vsi_port_num) &
 622				ICE_AQC_GET_SW_CONF_RESP_VSI_PORT_NUM_M;
 623
 624			pf_vf_num = le16_to_cpu(ele->pf_vf_num) &
 625				ICE_AQC_GET_SW_CONF_RESP_FUNC_NUM_M;
 626
 627			swid = le16_to_cpu(ele->swid);
 628
 629			if (le16_to_cpu(ele->pf_vf_num) &
 630			    ICE_AQC_GET_SW_CONF_RESP_IS_VF)
 631				is_vf = true;
 632
 633			type = le16_to_cpu(ele->vsi_port_num) >>
 634				ICE_AQC_GET_SW_CONF_RESP_TYPE_S;
 635
 636			if (type == ICE_AQC_GET_SW_CONF_RESP_VSI) {
 637				/* FW VSI is not needed. Just continue. */
 638				continue;
 639			}
 640
 641			ice_init_port_info(hw->port_info, vsi_port_num,
 642					   type, swid, pf_vf_num, is_vf);
 643		}
 644	} while (req_desc && !status);
 645
 646	devm_kfree(ice_hw_to_dev(hw), (void *)rbuf);
 647	return status;
 648}
 649
 650/**
 651 * ice_fill_sw_info - Helper function to populate lb_en and lan_en
 652 * @hw: pointer to the hardware structure
 653 * @fi: filter info structure to fill/update
 654 *
 655 * This helper function populates the lb_en and lan_en elements of the provided
 656 * ice_fltr_info struct using the switch's type and characteristics of the
 657 * switch rule being configured.
 658 */
 659static void ice_fill_sw_info(struct ice_hw *hw, struct ice_fltr_info *fi)
 660{
 661	fi->lb_en = false;
 662	fi->lan_en = false;
 663	if ((fi->flag & ICE_FLTR_TX) &&
 664	    (fi->fltr_act == ICE_FWD_TO_VSI ||
 665	     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
 666	     fi->fltr_act == ICE_FWD_TO_Q ||
 667	     fi->fltr_act == ICE_FWD_TO_QGRP)) {
 668		/* Setting LB for prune actions will result in replicated
 669		 * packets to the internal switch that will be dropped.
 670		 */
 671		if (fi->lkup_type != ICE_SW_LKUP_VLAN)
 672			fi->lb_en = true;
 673
 674		/* Set lan_en to TRUE if
 675		 * 1. The switch is a VEB AND
 676		 * 2
 677		 * 2.1 The lookup is a directional lookup like ethertype,
 678		 * promiscuous, ethertype-MAC, promiscuous-VLAN
 679		 * and default-port OR
 680		 * 2.2 The lookup is VLAN, OR
 681		 * 2.3 The lookup is MAC with mcast or bcast addr for MAC, OR
 682		 * 2.4 The lookup is MAC_VLAN with mcast or bcast addr for MAC.
 683		 *
 684		 * OR
 685		 *
 686		 * The switch is a VEPA.
 687		 *
 688		 * In all other cases, the LAN enable has to be set to false.
 689		 */
 690		if (hw->evb_veb) {
 691			if (fi->lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 692			    fi->lkup_type == ICE_SW_LKUP_PROMISC ||
 693			    fi->lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 694			    fi->lkup_type == ICE_SW_LKUP_PROMISC_VLAN ||
 695			    fi->lkup_type == ICE_SW_LKUP_DFLT ||
 696			    fi->lkup_type == ICE_SW_LKUP_VLAN ||
 697			    (fi->lkup_type == ICE_SW_LKUP_MAC &&
 698			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)) ||
 699			    (fi->lkup_type == ICE_SW_LKUP_MAC_VLAN &&
 700			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)))
 701				fi->lan_en = true;
 702		} else {
 703			fi->lan_en = true;
 704		}
 705	}
 706}
 707
 708/**
 709 * ice_fill_sw_rule - Helper function to fill switch rule structure
 710 * @hw: pointer to the hardware structure
 711 * @f_info: entry containing packet forwarding information
 712 * @s_rule: switch rule structure to be filled in based on mac_entry
 713 * @opc: switch rules population command type - pass in the command opcode
 714 */
 715static void
 716ice_fill_sw_rule(struct ice_hw *hw, struct ice_fltr_info *f_info,
 717		 struct ice_aqc_sw_rules_elem *s_rule, enum ice_adminq_opc opc)
 718{
 719	u16 vlan_id = ICE_MAX_VLAN_ID + 1;
 
 720	void *daddr = NULL;
 721	u16 eth_hdr_sz;
 722	u8 *eth_hdr;
 723	u32 act = 0;
 724	__be16 *off;
 725	u8 q_rgn;
 726
 727	if (opc == ice_aqc_opc_remove_sw_rules) {
 728		s_rule->pdata.lkup_tx_rx.act = 0;
 729		s_rule->pdata.lkup_tx_rx.index =
 730			cpu_to_le16(f_info->fltr_rule_id);
 731		s_rule->pdata.lkup_tx_rx.hdr_len = 0;
 732		return;
 733	}
 734
 735	eth_hdr_sz = sizeof(dummy_eth_header);
 736	eth_hdr = s_rule->pdata.lkup_tx_rx.hdr;
 737
 738	/* initialize the ether header with a dummy header */
 739	memcpy(eth_hdr, dummy_eth_header, eth_hdr_sz);
 740	ice_fill_sw_info(hw, f_info);
 741
 742	switch (f_info->fltr_act) {
 743	case ICE_FWD_TO_VSI:
 744		act |= (f_info->fwd_id.hw_vsi_id << ICE_SINGLE_ACT_VSI_ID_S) &
 745			ICE_SINGLE_ACT_VSI_ID_M;
 746		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 747			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 748				ICE_SINGLE_ACT_VALID_BIT;
 749		break;
 750	case ICE_FWD_TO_VSI_LIST:
 751		act |= ICE_SINGLE_ACT_VSI_LIST;
 752		act |= (f_info->fwd_id.vsi_list_id <<
 753			ICE_SINGLE_ACT_VSI_LIST_ID_S) &
 754			ICE_SINGLE_ACT_VSI_LIST_ID_M;
 755		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 756			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 757				ICE_SINGLE_ACT_VALID_BIT;
 758		break;
 759	case ICE_FWD_TO_Q:
 760		act |= ICE_SINGLE_ACT_TO_Q;
 761		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 762			ICE_SINGLE_ACT_Q_INDEX_M;
 763		break;
 764	case ICE_DROP_PACKET:
 765		act |= ICE_SINGLE_ACT_VSI_FORWARDING | ICE_SINGLE_ACT_DROP |
 766			ICE_SINGLE_ACT_VALID_BIT;
 767		break;
 768	case ICE_FWD_TO_QGRP:
 769		q_rgn = f_info->qgrp_size > 0 ?
 770			(u8)ilog2(f_info->qgrp_size) : 0;
 771		act |= ICE_SINGLE_ACT_TO_Q;
 772		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 773			ICE_SINGLE_ACT_Q_INDEX_M;
 774		act |= (q_rgn << ICE_SINGLE_ACT_Q_REGION_S) &
 775			ICE_SINGLE_ACT_Q_REGION_M;
 776		break;
 
 
 
 777	default:
 778		return;
 779	}
 780
 781	if (f_info->lb_en)
 782		act |= ICE_SINGLE_ACT_LB_ENABLE;
 783	if (f_info->lan_en)
 784		act |= ICE_SINGLE_ACT_LAN_ENABLE;
 785
 786	switch (f_info->lkup_type) {
 787	case ICE_SW_LKUP_MAC:
 788		daddr = f_info->l_data.mac.mac_addr;
 789		break;
 790	case ICE_SW_LKUP_VLAN:
 791		vlan_id = f_info->l_data.vlan.vlan_id;
 792		if (f_info->fltr_act == ICE_FWD_TO_VSI ||
 793		    f_info->fltr_act == ICE_FWD_TO_VSI_LIST) {
 794			act |= ICE_SINGLE_ACT_PRUNE;
 795			act |= ICE_SINGLE_ACT_EGRESS | ICE_SINGLE_ACT_INGRESS;
 796		}
 797		break;
 798	case ICE_SW_LKUP_ETHERTYPE_MAC:
 799		daddr = f_info->l_data.ethertype_mac.mac_addr;
 800		/* fall-through */
 801	case ICE_SW_LKUP_ETHERTYPE:
 802		off = (__force __be16 *)(eth_hdr + ICE_ETH_ETHTYPE_OFFSET);
 803		*off = cpu_to_be16(f_info->l_data.ethertype_mac.ethertype);
 804		break;
 805	case ICE_SW_LKUP_MAC_VLAN:
 806		daddr = f_info->l_data.mac_vlan.mac_addr;
 807		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 808		break;
 809	case ICE_SW_LKUP_PROMISC_VLAN:
 810		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 811		/* fall-through */
 812	case ICE_SW_LKUP_PROMISC:
 813		daddr = f_info->l_data.mac_vlan.mac_addr;
 814		break;
 815	default:
 816		break;
 817	}
 818
 819	s_rule->type = (f_info->flag & ICE_FLTR_RX) ?
 820		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_RX) :
 821		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_TX);
 822
 823	/* Recipe set depending on lookup type */
 824	s_rule->pdata.lkup_tx_rx.recipe_id = cpu_to_le16(f_info->lkup_type);
 825	s_rule->pdata.lkup_tx_rx.src = cpu_to_le16(f_info->src);
 826	s_rule->pdata.lkup_tx_rx.act = cpu_to_le32(act);
 827
 828	if (daddr)
 829		ether_addr_copy(eth_hdr + ICE_ETH_DA_OFFSET, daddr);
 830
 831	if (!(vlan_id > ICE_MAX_VLAN_ID)) {
 832		off = (__force __be16 *)(eth_hdr + ICE_ETH_VLAN_TCI_OFFSET);
 833		*off = cpu_to_be16(vlan_id);
 834	}
 835
 836	/* Create the switch rule with the final dummy Ethernet header */
 837	if (opc != ice_aqc_opc_update_sw_rules)
 838		s_rule->pdata.lkup_tx_rx.hdr_len = cpu_to_le16(eth_hdr_sz);
 
 
 839}
 840
 841/**
 842 * ice_add_marker_act
 843 * @hw: pointer to the hardware structure
 844 * @m_ent: the management entry for which sw marker needs to be added
 845 * @sw_marker: sw marker to tag the Rx descriptor with
 846 * @l_id: large action resource ID
 847 *
 848 * Create a large action to hold software marker and update the switch rule
 849 * entry pointed by m_ent with newly created large action
 850 */
 851static enum ice_status
 852ice_add_marker_act(struct ice_hw *hw, struct ice_fltr_mgmt_list_entry *m_ent,
 853		   u16 sw_marker, u16 l_id)
 854{
 855	struct ice_aqc_sw_rules_elem *lg_act, *rx_tx;
 856	/* For software marker we need 3 large actions
 857	 * 1. FWD action: FWD TO VSI or VSI LIST
 858	 * 2. GENERIC VALUE action to hold the profile ID
 859	 * 3. GENERIC VALUE action to hold the software marker ID
 860	 */
 861	const u16 num_lg_acts = 3;
 862	enum ice_status status;
 863	u16 lg_act_size;
 864	u16 rules_size;
 
 865	u32 act;
 866	u16 id;
 867
 868	if (m_ent->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
 869		return ICE_ERR_PARAM;
 870
 871	/* Create two back-to-back switch rules and submit them to the HW using
 872	 * one memory buffer:
 873	 *    1. Large Action
 874	 *    2. Look up Tx Rx
 875	 */
 876	lg_act_size = (u16)ICE_SW_RULE_LG_ACT_SIZE(num_lg_acts);
 877	rules_size = lg_act_size + ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
 878	lg_act = devm_kzalloc(ice_hw_to_dev(hw), rules_size, GFP_KERNEL);
 879	if (!lg_act)
 880		return ICE_ERR_NO_MEMORY;
 881
 882	rx_tx = (struct ice_aqc_sw_rules_elem *)((u8 *)lg_act + lg_act_size);
 883
 884	/* Fill in the first switch rule i.e. large action */
 885	lg_act->type = cpu_to_le16(ICE_AQC_SW_RULES_T_LG_ACT);
 886	lg_act->pdata.lg_act.index = cpu_to_le16(l_id);
 887	lg_act->pdata.lg_act.size = cpu_to_le16(num_lg_acts);
 888
 889	/* First action VSI forwarding or VSI list forwarding depending on how
 890	 * many VSIs
 891	 */
 892	id = (m_ent->vsi_count > 1) ? m_ent->fltr_info.fwd_id.vsi_list_id :
 893		m_ent->fltr_info.fwd_id.hw_vsi_id;
 
 894
 895	act = ICE_LG_ACT_VSI_FORWARDING | ICE_LG_ACT_VALID_BIT;
 896	act |= (id << ICE_LG_ACT_VSI_LIST_ID_S) &
 897		ICE_LG_ACT_VSI_LIST_ID_M;
 898	if (m_ent->vsi_count > 1)
 899		act |= ICE_LG_ACT_VSI_LIST;
 900	lg_act->pdata.lg_act.act[0] = cpu_to_le32(act);
 901
 902	/* Second action descriptor type */
 903	act = ICE_LG_ACT_GENERIC;
 904
 905	act |= (1 << ICE_LG_ACT_GENERIC_VALUE_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 906	lg_act->pdata.lg_act.act[1] = cpu_to_le32(act);
 907
 908	act = (ICE_LG_ACT_GENERIC_OFF_RX_DESC_PROF_IDX <<
 909	       ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_OFFSET_M;
 910
 911	/* Third action Marker value */
 912	act |= ICE_LG_ACT_GENERIC;
 913	act |= (sw_marker << ICE_LG_ACT_GENERIC_VALUE_S) &
 914		ICE_LG_ACT_GENERIC_VALUE_M;
 915
 
 916	lg_act->pdata.lg_act.act[2] = cpu_to_le32(act);
 917
 918	/* call the fill switch rule to fill the lookup Tx Rx structure */
 919	ice_fill_sw_rule(hw, &m_ent->fltr_info, rx_tx,
 920			 ice_aqc_opc_update_sw_rules);
 921
 922	/* Update the action to point to the large action ID */
 923	rx_tx->pdata.lkup_tx_rx.act =
 924		cpu_to_le32(ICE_SINGLE_ACT_PTR |
 925			    ((l_id << ICE_SINGLE_ACT_PTR_VAL_S) &
 926			     ICE_SINGLE_ACT_PTR_VAL_M));
 927
 928	/* Use the filter rule ID of the previously created rule with single
 929	 * act. Once the update happens, hardware will treat this as large
 930	 * action
 931	 */
 932	rx_tx->pdata.lkup_tx_rx.index =
 933		cpu_to_le16(m_ent->fltr_info.fltr_rule_id);
 934
 935	status = ice_aq_sw_rules(hw, lg_act, rules_size, 2,
 936				 ice_aqc_opc_update_sw_rules, NULL);
 937	if (!status) {
 938		m_ent->lg_act_idx = l_id;
 939		m_ent->sw_marker_id = sw_marker;
 940	}
 941
 942	devm_kfree(ice_hw_to_dev(hw), lg_act);
 943	return status;
 944}
 945
 946/**
 947 * ice_create_vsi_list_map
 948 * @hw: pointer to the hardware structure
 949 * @vsi_handle_arr: array of VSI handles to set in the VSI mapping
 950 * @num_vsi: number of VSI handles in the array
 951 * @vsi_list_id: VSI list ID generated as part of allocate resource
 952 *
 953 * Helper function to create a new entry of VSI list ID to VSI mapping
 954 * using the given VSI list ID
 955 */
 956static struct ice_vsi_list_map_info *
 957ice_create_vsi_list_map(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 958			u16 vsi_list_id)
 959{
 960	struct ice_switch_info *sw = hw->switch_info;
 961	struct ice_vsi_list_map_info *v_map;
 962	int i;
 963
 964	v_map = devm_kcalloc(ice_hw_to_dev(hw), 1, sizeof(*v_map), GFP_KERNEL);
 965	if (!v_map)
 966		return NULL;
 967
 968	v_map->vsi_list_id = vsi_list_id;
 969	v_map->ref_cnt = 1;
 970	for (i = 0; i < num_vsi; i++)
 971		set_bit(vsi_handle_arr[i], v_map->vsi_map);
 972
 973	list_add(&v_map->list_entry, &sw->vsi_list_map_head);
 974	return v_map;
 975}
 976
 977/**
 978 * ice_update_vsi_list_rule
 979 * @hw: pointer to the hardware structure
 980 * @vsi_handle_arr: array of VSI handles to form a VSI list
 981 * @num_vsi: number of VSI handles in the array
 982 * @vsi_list_id: VSI list ID generated as part of allocate resource
 983 * @remove: Boolean value to indicate if this is a remove action
 984 * @opc: switch rules population command type - pass in the command opcode
 985 * @lkup_type: lookup type of the filter
 986 *
 987 * Call AQ command to add a new switch rule or update existing switch rule
 988 * using the given VSI list ID
 989 */
 990static enum ice_status
 991ice_update_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 992			 u16 vsi_list_id, bool remove, enum ice_adminq_opc opc,
 993			 enum ice_sw_lkup_type lkup_type)
 994{
 995	struct ice_aqc_sw_rules_elem *s_rule;
 996	enum ice_status status;
 997	u16 s_rule_size;
 998	u16 type;
 999	int i;
1000
1001	if (!num_vsi)
1002		return ICE_ERR_PARAM;
1003
1004	if (lkup_type == ICE_SW_LKUP_MAC ||
1005	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
1006	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
1007	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
1008	    lkup_type == ICE_SW_LKUP_PROMISC ||
1009	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN)
1010		type = remove ? ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR :
1011				ICE_AQC_SW_RULES_T_VSI_LIST_SET;
1012	else if (lkup_type == ICE_SW_LKUP_VLAN)
1013		type = remove ? ICE_AQC_SW_RULES_T_PRUNE_LIST_CLEAR :
1014				ICE_AQC_SW_RULES_T_PRUNE_LIST_SET;
1015	else
1016		return ICE_ERR_PARAM;
1017
1018	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(num_vsi);
1019	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1020	if (!s_rule)
1021		return ICE_ERR_NO_MEMORY;
1022	for (i = 0; i < num_vsi; i++) {
1023		if (!ice_is_vsi_valid(hw, vsi_handle_arr[i])) {
1024			status = ICE_ERR_PARAM;
1025			goto exit;
1026		}
1027		/* AQ call requires hw_vsi_id(s) */
1028		s_rule->pdata.vsi_list.vsi[i] =
1029			cpu_to_le16(ice_get_hw_vsi_num(hw, vsi_handle_arr[i]));
1030	}
1031
1032	s_rule->type = cpu_to_le16(type);
1033	s_rule->pdata.vsi_list.number_vsi = cpu_to_le16(num_vsi);
1034	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1035
1036	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opc, NULL);
1037
1038exit:
1039	devm_kfree(ice_hw_to_dev(hw), s_rule);
1040	return status;
1041}
1042
1043/**
1044 * ice_create_vsi_list_rule - Creates and populates a VSI list rule
1045 * @hw: pointer to the HW struct
1046 * @vsi_handle_arr: array of VSI handles to form a VSI list
1047 * @num_vsi: number of VSI handles in the array
1048 * @vsi_list_id: stores the ID of the VSI list to be created
1049 * @lkup_type: switch rule filter's lookup type
1050 */
1051static enum ice_status
1052ice_create_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
1053			 u16 *vsi_list_id, enum ice_sw_lkup_type lkup_type)
1054{
1055	enum ice_status status;
 
 
 
 
 
1056
1057	status = ice_aq_alloc_free_vsi_list(hw, vsi_list_id, lkup_type,
1058					    ice_aqc_opc_alloc_res);
1059	if (status)
1060		return status;
1061
1062	/* Update the newly created VSI list to include the specified VSIs */
1063	return ice_update_vsi_list_rule(hw, vsi_handle_arr, num_vsi,
1064					*vsi_list_id, false,
1065					ice_aqc_opc_add_sw_rules, lkup_type);
1066}
1067
1068/**
1069 * ice_create_pkt_fwd_rule
1070 * @hw: pointer to the hardware structure
1071 * @f_entry: entry containing packet forwarding information
1072 *
1073 * Create switch rule with given filter information and add an entry
1074 * to the corresponding filter management list to track this switch rule
1075 * and VSI mapping
1076 */
1077static enum ice_status
1078ice_create_pkt_fwd_rule(struct ice_hw *hw,
1079			struct ice_fltr_list_entry *f_entry)
1080{
 
1081	struct ice_fltr_mgmt_list_entry *fm_entry;
1082	struct ice_aqc_sw_rules_elem *s_rule;
1083	enum ice_sw_lkup_type l_type;
1084	struct ice_sw_recipe *recp;
1085	enum ice_status status;
1086
1087	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1088			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1089	if (!s_rule)
1090		return ICE_ERR_NO_MEMORY;
1091	fm_entry = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*fm_entry),
1092				GFP_KERNEL);
1093	if (!fm_entry) {
1094		status = ICE_ERR_NO_MEMORY;
1095		goto ice_create_pkt_fwd_rule_exit;
1096	}
1097
1098	fm_entry->fltr_info = f_entry->fltr_info;
1099
1100	/* Initialize all the fields for the management entry */
1101	fm_entry->vsi_count = 1;
1102	fm_entry->lg_act_idx = ICE_INVAL_LG_ACT_INDEX;
1103	fm_entry->sw_marker_id = ICE_INVAL_SW_MARKER_ID;
1104	fm_entry->counter_index = ICE_INVAL_COUNTER_ID;
1105
1106	ice_fill_sw_rule(hw, &fm_entry->fltr_info, s_rule,
1107			 ice_aqc_opc_add_sw_rules);
1108
1109	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1110				 ice_aqc_opc_add_sw_rules, NULL);
1111	if (status) {
1112		devm_kfree(ice_hw_to_dev(hw), fm_entry);
1113		goto ice_create_pkt_fwd_rule_exit;
1114	}
1115
1116	f_entry->fltr_info.fltr_rule_id =
1117		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1118	fm_entry->fltr_info.fltr_rule_id =
1119		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1120
1121	/* The book keeping entries will get removed when base driver
1122	 * calls remove filter AQ command
1123	 */
1124	l_type = fm_entry->fltr_info.lkup_type;
1125	recp = &hw->switch_info->recp_list[l_type];
1126	list_add(&fm_entry->list_entry, &recp->filt_rules);
1127
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1128ice_create_pkt_fwd_rule_exit:
1129	devm_kfree(ice_hw_to_dev(hw), s_rule);
1130	return status;
1131}
1132
1133/**
1134 * ice_update_pkt_fwd_rule
1135 * @hw: pointer to the hardware structure
1136 * @f_info: filter information for switch rule
 
 
1137 *
1138 * Call AQ command to update a previously created switch rule with a
1139 * VSI list ID
1140 */
1141static enum ice_status
1142ice_update_pkt_fwd_rule(struct ice_hw *hw, struct ice_fltr_info *f_info)
 
1143{
1144	struct ice_aqc_sw_rules_elem *s_rule;
 
1145	enum ice_status status;
1146
1147	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1148			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1149	if (!s_rule)
1150		return ICE_ERR_NO_MEMORY;
1151
1152	ice_fill_sw_rule(hw, f_info, s_rule, ice_aqc_opc_update_sw_rules);
 
 
1153
1154	s_rule->pdata.lkup_tx_rx.index = cpu_to_le16(f_info->fltr_rule_id);
 
 
 
1155
1156	/* Update switch rule with new rule set to forward VSI list */
1157	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1158				 ice_aqc_opc_update_sw_rules, NULL);
1159
1160	devm_kfree(ice_hw_to_dev(hw), s_rule);
1161	return status;
1162}
1163
1164/**
1165 * ice_update_sw_rule_bridge_mode
1166 * @hw: pointer to the HW struct
1167 *
1168 * Updates unicast switch filter rules based on VEB/VEPA mode
1169 */
1170enum ice_status ice_update_sw_rule_bridge_mode(struct ice_hw *hw)
1171{
1172	struct ice_switch_info *sw = hw->switch_info;
1173	struct ice_fltr_mgmt_list_entry *fm_entry;
1174	enum ice_status status = 0;
1175	struct list_head *rule_head;
1176	struct mutex *rule_lock; /* Lock to protect filter rule list */
1177
1178	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1179	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1180
1181	mutex_lock(rule_lock);
1182	list_for_each_entry(fm_entry, rule_head, list_entry) {
1183		struct ice_fltr_info *fi = &fm_entry->fltr_info;
1184		u8 *addr = fi->l_data.mac.mac_addr;
1185
1186		/* Update unicast Tx rules to reflect the selected
1187		 * VEB/VEPA mode
1188		 */
1189		if ((fi->flag & ICE_FLTR_TX) && is_unicast_ether_addr(addr) &&
1190		    (fi->fltr_act == ICE_FWD_TO_VSI ||
1191		     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
1192		     fi->fltr_act == ICE_FWD_TO_Q ||
1193		     fi->fltr_act == ICE_FWD_TO_QGRP)) {
1194			status = ice_update_pkt_fwd_rule(hw, fi);
1195			if (status)
1196				break;
1197		}
1198	}
1199
1200	mutex_unlock(rule_lock);
1201
1202	return status;
1203}
1204
1205/**
1206 * ice_add_update_vsi_list
1207 * @hw: pointer to the hardware structure
1208 * @m_entry: pointer to current filter management list entry
1209 * @cur_fltr: filter information from the book keeping entry
1210 * @new_fltr: filter information with the new VSI to be added
1211 *
1212 * Call AQ command to add or update previously created VSI list with new VSI.
1213 *
1214 * Helper function to do book keeping associated with adding filter information
1215 * The algorithm to do the book keeping is described below :
1216 * When a VSI needs to subscribe to a given filter (MAC/VLAN/Ethtype etc.)
1217 *	if only one VSI has been added till now
1218 *		Allocate a new VSI list and add two VSIs
1219 *		to this list using switch rule command
1220 *		Update the previously created switch rule with the
1221 *		newly created VSI list ID
1222 *	if a VSI list was previously created
1223 *		Add the new VSI to the previously created VSI list set
1224 *		using the update switch rule command
1225 */
1226static enum ice_status
1227ice_add_update_vsi_list(struct ice_hw *hw,
1228			struct ice_fltr_mgmt_list_entry *m_entry,
1229			struct ice_fltr_info *cur_fltr,
1230			struct ice_fltr_info *new_fltr)
1231{
1232	enum ice_status status = 0;
1233	u16 vsi_list_id = 0;
1234
1235	if ((cur_fltr->fltr_act == ICE_FWD_TO_Q ||
1236	     cur_fltr->fltr_act == ICE_FWD_TO_QGRP))
1237		return ICE_ERR_NOT_IMPL;
1238
1239	if ((new_fltr->fltr_act == ICE_FWD_TO_Q ||
1240	     new_fltr->fltr_act == ICE_FWD_TO_QGRP) &&
1241	    (cur_fltr->fltr_act == ICE_FWD_TO_VSI ||
1242	     cur_fltr->fltr_act == ICE_FWD_TO_VSI_LIST))
1243		return ICE_ERR_NOT_IMPL;
1244
1245	if (m_entry->vsi_count < 2 && !m_entry->vsi_list_info) {
1246		/* Only one entry existed in the mapping and it was not already
1247		 * a part of a VSI list. So, create a VSI list with the old and
1248		 * new VSIs.
1249		 */
1250		struct ice_fltr_info tmp_fltr;
1251		u16 vsi_handle_arr[2];
1252
1253		/* A rule already exists with the new VSI being added */
1254		if (cur_fltr->fwd_id.hw_vsi_id == new_fltr->fwd_id.hw_vsi_id)
1255			return ICE_ERR_ALREADY_EXISTS;
1256
1257		vsi_handle_arr[0] = cur_fltr->vsi_handle;
1258		vsi_handle_arr[1] = new_fltr->vsi_handle;
1259		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1260						  &vsi_list_id,
1261						  new_fltr->lkup_type);
1262		if (status)
1263			return status;
1264
1265		tmp_fltr = *new_fltr;
1266		tmp_fltr.fltr_rule_id = cur_fltr->fltr_rule_id;
1267		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1268		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1269		/* Update the previous switch rule of "MAC forward to VSI" to
1270		 * "MAC fwd to VSI list"
1271		 */
1272		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
 
1273		if (status)
1274			return status;
1275
1276		cur_fltr->fwd_id.vsi_list_id = vsi_list_id;
1277		cur_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1278		m_entry->vsi_list_info =
1279			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1280						vsi_list_id);
1281
1282		/* If this entry was large action then the large action needs
1283		 * to be updated to point to FWD to VSI list
1284		 */
1285		if (m_entry->sw_marker_id != ICE_INVAL_SW_MARKER_ID)
1286			status =
1287			    ice_add_marker_act(hw, m_entry,
1288					       m_entry->sw_marker_id,
1289					       m_entry->lg_act_idx);
1290	} else {
1291		u16 vsi_handle = new_fltr->vsi_handle;
1292		enum ice_adminq_opc opcode;
1293
1294		if (!m_entry->vsi_list_info)
1295			return ICE_ERR_CFG;
1296
1297		/* A rule already exists with the new VSI being added */
1298		if (test_bit(vsi_handle, m_entry->vsi_list_info->vsi_map))
1299			return 0;
1300
1301		/* Update the previously created VSI list set with
1302		 * the new VSI ID passed in
1303		 */
1304		vsi_list_id = cur_fltr->fwd_id.vsi_list_id;
1305		opcode = ice_aqc_opc_update_sw_rules;
1306
1307		status = ice_update_vsi_list_rule(hw, &vsi_handle, 1,
1308						  vsi_list_id, false, opcode,
1309						  new_fltr->lkup_type);
1310		/* update VSI list mapping info with new VSI ID */
1311		if (!status)
1312			set_bit(vsi_handle, m_entry->vsi_list_info->vsi_map);
1313	}
1314	if (!status)
1315		m_entry->vsi_count++;
1316	return status;
1317}
1318
1319/**
1320 * ice_find_rule_entry - Search a rule entry
1321 * @hw: pointer to the hardware structure
1322 * @recp_id: lookup type for which the specified rule needs to be searched
1323 * @f_info: rule information
1324 *
1325 * Helper function to search for a given rule entry
1326 * Returns pointer to entry storing the rule if found
1327 */
1328static struct ice_fltr_mgmt_list_entry *
1329ice_find_rule_entry(struct ice_hw *hw, u8 recp_id, struct ice_fltr_info *f_info)
1330{
1331	struct ice_fltr_mgmt_list_entry *list_itr, *ret = NULL;
1332	struct ice_switch_info *sw = hw->switch_info;
1333	struct list_head *list_head;
1334
1335	list_head = &sw->recp_list[recp_id].filt_rules;
1336	list_for_each_entry(list_itr, list_head, list_entry) {
1337		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
1338			    sizeof(f_info->l_data)) &&
1339		    f_info->flag == list_itr->fltr_info.flag) {
1340			ret = list_itr;
1341			break;
1342		}
1343	}
1344	return ret;
 
1345}
1346
1347/**
1348 * ice_find_vsi_list_entry - Search VSI list map with VSI count 1
1349 * @hw: pointer to the hardware structure
1350 * @recp_id: lookup type for which VSI lists needs to be searched
1351 * @vsi_handle: VSI handle to be found in VSI list
1352 * @vsi_list_id: VSI list ID found containing vsi_handle
1353 *
1354 * Helper function to search a VSI list with single entry containing given VSI
1355 * handle element. This can be extended further to search VSI list with more
1356 * than 1 vsi_count. Returns pointer to VSI list entry if found.
1357 */
1358static struct ice_vsi_list_map_info *
1359ice_find_vsi_list_entry(struct ice_hw *hw, u8 recp_id, u16 vsi_handle,
1360			u16 *vsi_list_id)
1361{
1362	struct ice_vsi_list_map_info *map_info = NULL;
1363	struct ice_switch_info *sw = hw->switch_info;
1364	struct ice_fltr_mgmt_list_entry *list_itr;
1365	struct list_head *list_head;
1366
1367	list_head = &sw->recp_list[recp_id].filt_rules;
1368	list_for_each_entry(list_itr, list_head, list_entry) {
1369		if (list_itr->vsi_count == 1 && list_itr->vsi_list_info) {
1370			map_info = list_itr->vsi_list_info;
1371			if (test_bit(vsi_handle, map_info->vsi_map)) {
1372				*vsi_list_id = map_info->vsi_list_id;
1373				return map_info;
1374			}
1375		}
1376	}
1377	return NULL;
1378}
1379
1380/**
1381 * ice_add_rule_internal - add rule for a given lookup type
1382 * @hw: pointer to the hardware structure
1383 * @recp_id: lookup type (recipe ID) for which rule has to be added
1384 * @f_entry: structure containing MAC forwarding information
1385 *
1386 * Adds or updates the rule lists for a given recipe
1387 */
1388static enum ice_status
1389ice_add_rule_internal(struct ice_hw *hw, u8 recp_id,
1390		      struct ice_fltr_list_entry *f_entry)
1391{
1392	struct ice_switch_info *sw = hw->switch_info;
1393	struct ice_fltr_info *new_fltr, *cur_fltr;
1394	struct ice_fltr_mgmt_list_entry *m_entry;
1395	struct mutex *rule_lock; /* Lock to protect filter rule list */
1396	enum ice_status status = 0;
1397
1398	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1399		return ICE_ERR_PARAM;
1400	f_entry->fltr_info.fwd_id.hw_vsi_id =
1401		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1402
1403	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1404
1405	mutex_lock(rule_lock);
1406	new_fltr = &f_entry->fltr_info;
1407	if (new_fltr->flag & ICE_FLTR_RX)
1408		new_fltr->src = hw->port_info->lport;
1409	else if (new_fltr->flag & ICE_FLTR_TX)
1410		new_fltr->src = f_entry->fltr_info.fwd_id.hw_vsi_id;
1411
1412	m_entry = ice_find_rule_entry(hw, recp_id, new_fltr);
1413	if (!m_entry) {
1414		mutex_unlock(rule_lock);
1415		return ice_create_pkt_fwd_rule(hw, f_entry);
1416	}
1417
1418	cur_fltr = &m_entry->fltr_info;
1419	status = ice_add_update_vsi_list(hw, m_entry, cur_fltr, new_fltr);
1420	mutex_unlock(rule_lock);
1421
1422	return status;
1423}
1424
1425/**
1426 * ice_remove_vsi_list_rule
1427 * @hw: pointer to the hardware structure
1428 * @vsi_list_id: VSI list ID generated as part of allocate resource
1429 * @lkup_type: switch rule filter lookup type
1430 *
1431 * The VSI list should be emptied before this function is called to remove the
1432 * VSI list.
1433 */
1434static enum ice_status
1435ice_remove_vsi_list_rule(struct ice_hw *hw, u16 vsi_list_id,
1436			 enum ice_sw_lkup_type lkup_type)
1437{
1438	struct ice_aqc_sw_rules_elem *s_rule;
1439	enum ice_status status;
1440	u16 s_rule_size;
1441
1442	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(0);
1443	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1444	if (!s_rule)
1445		return ICE_ERR_NO_MEMORY;
1446
1447	s_rule->type = cpu_to_le16(ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR);
1448	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1449
1450	/* Free the vsi_list resource that we allocated. It is assumed that the
1451	 * list is empty at this point.
1452	 */
1453	status = ice_aq_alloc_free_vsi_list(hw, &vsi_list_id, lkup_type,
1454					    ice_aqc_opc_free_res);
1455
1456	devm_kfree(ice_hw_to_dev(hw), s_rule);
1457	return status;
1458}
1459
1460/**
1461 * ice_rem_update_vsi_list
1462 * @hw: pointer to the hardware structure
1463 * @vsi_handle: VSI handle of the VSI to remove
1464 * @fm_list: filter management entry for which the VSI list management needs to
1465 *           be done
1466 */
1467static enum ice_status
1468ice_rem_update_vsi_list(struct ice_hw *hw, u16 vsi_handle,
1469			struct ice_fltr_mgmt_list_entry *fm_list)
1470{
1471	enum ice_sw_lkup_type lkup_type;
1472	enum ice_status status = 0;
1473	u16 vsi_list_id;
1474
1475	if (fm_list->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST ||
1476	    fm_list->vsi_count == 0)
1477		return ICE_ERR_PARAM;
1478
1479	/* A rule with the VSI being removed does not exist */
1480	if (!test_bit(vsi_handle, fm_list->vsi_list_info->vsi_map))
1481		return ICE_ERR_DOES_NOT_EXIST;
1482
1483	lkup_type = fm_list->fltr_info.lkup_type;
1484	vsi_list_id = fm_list->fltr_info.fwd_id.vsi_list_id;
1485	status = ice_update_vsi_list_rule(hw, &vsi_handle, 1, vsi_list_id, true,
1486					  ice_aqc_opc_update_sw_rules,
1487					  lkup_type);
1488	if (status)
1489		return status;
1490
1491	fm_list->vsi_count--;
1492	clear_bit(vsi_handle, fm_list->vsi_list_info->vsi_map);
1493
1494	if (fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) {
1495		struct ice_fltr_info tmp_fltr_info = fm_list->fltr_info;
1496		struct ice_vsi_list_map_info *vsi_list_info =
1497			fm_list->vsi_list_info;
1498		u16 rem_vsi_handle;
1499
1500		rem_vsi_handle = find_first_bit(vsi_list_info->vsi_map,
1501						ICE_MAX_VSI);
1502		if (!ice_is_vsi_valid(hw, rem_vsi_handle))
1503			return ICE_ERR_OUT_OF_RANGE;
1504
1505		/* Make sure VSI list is empty before removing it below */
1506		status = ice_update_vsi_list_rule(hw, &rem_vsi_handle, 1,
1507						  vsi_list_id, true,
1508						  ice_aqc_opc_update_sw_rules,
1509						  lkup_type);
1510		if (status)
1511			return status;
1512
1513		tmp_fltr_info.fltr_act = ICE_FWD_TO_VSI;
1514		tmp_fltr_info.fwd_id.hw_vsi_id =
1515			ice_get_hw_vsi_num(hw, rem_vsi_handle);
1516		tmp_fltr_info.vsi_handle = rem_vsi_handle;
1517		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr_info);
1518		if (status) {
1519			ice_debug(hw, ICE_DBG_SW,
1520				  "Failed to update pkt fwd rule to FWD_TO_VSI on HW VSI %d, error %d\n",
1521				  tmp_fltr_info.fwd_id.hw_vsi_id, status);
1522			return status;
1523		}
1524
1525		fm_list->fltr_info = tmp_fltr_info;
1526	}
1527
1528	if ((fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) ||
1529	    (fm_list->vsi_count == 0 && lkup_type == ICE_SW_LKUP_VLAN)) {
1530		struct ice_vsi_list_map_info *vsi_list_info =
1531			fm_list->vsi_list_info;
1532
1533		/* Remove the VSI list since it is no longer used */
1534		status = ice_remove_vsi_list_rule(hw, vsi_list_id, lkup_type);
1535		if (status) {
1536			ice_debug(hw, ICE_DBG_SW,
1537				  "Failed to remove VSI list %d, error %d\n",
1538				  vsi_list_id, status);
1539			return status;
1540		}
1541
1542		list_del(&vsi_list_info->list_entry);
1543		devm_kfree(ice_hw_to_dev(hw), vsi_list_info);
1544		fm_list->vsi_list_info = NULL;
1545	}
1546
1547	return status;
1548}
1549
1550/**
1551 * ice_remove_rule_internal - Remove a filter rule of a given type
1552 * @hw: pointer to the hardware structure
1553 * @recp_id: recipe ID for which the rule needs to removed
1554 * @f_entry: rule entry containing filter information
1555 */
1556static enum ice_status
1557ice_remove_rule_internal(struct ice_hw *hw, u8 recp_id,
1558			 struct ice_fltr_list_entry *f_entry)
1559{
1560	struct ice_switch_info *sw = hw->switch_info;
1561	struct ice_fltr_mgmt_list_entry *list_elem;
1562	struct mutex *rule_lock; /* Lock to protect filter rule list */
1563	enum ice_status status = 0;
1564	bool remove_rule = false;
1565	u16 vsi_handle;
1566
1567	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1568		return ICE_ERR_PARAM;
1569	f_entry->fltr_info.fwd_id.hw_vsi_id =
1570		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1571
1572	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1573	mutex_lock(rule_lock);
1574	list_elem = ice_find_rule_entry(hw, recp_id, &f_entry->fltr_info);
1575	if (!list_elem) {
1576		status = ICE_ERR_DOES_NOT_EXIST;
1577		goto exit;
1578	}
1579
1580	if (list_elem->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST) {
1581		remove_rule = true;
1582	} else if (!list_elem->vsi_list_info) {
1583		status = ICE_ERR_DOES_NOT_EXIST;
1584		goto exit;
1585	} else if (list_elem->vsi_list_info->ref_cnt > 1) {
1586		/* a ref_cnt > 1 indicates that the vsi_list is being
1587		 * shared by multiple rules. Decrement the ref_cnt and
1588		 * remove this rule, but do not modify the list, as it
1589		 * is in-use by other rules.
1590		 */
1591		list_elem->vsi_list_info->ref_cnt--;
1592		remove_rule = true;
1593	} else {
1594		/* a ref_cnt of 1 indicates the vsi_list is only used
1595		 * by one rule. However, the original removal request is only
1596		 * for a single VSI. Update the vsi_list first, and only
1597		 * remove the rule if there are no further VSIs in this list.
1598		 */
1599		vsi_handle = f_entry->fltr_info.vsi_handle;
1600		status = ice_rem_update_vsi_list(hw, vsi_handle, list_elem);
1601		if (status)
1602			goto exit;
1603		/* if VSI count goes to zero after updating the VSI list */
1604		if (list_elem->vsi_count == 0)
1605			remove_rule = true;
1606	}
1607
1608	if (remove_rule) {
1609		/* Remove the lookup rule */
1610		struct ice_aqc_sw_rules_elem *s_rule;
1611
1612		s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1613				      ICE_SW_RULE_RX_TX_NO_HDR_SIZE,
1614				      GFP_KERNEL);
1615		if (!s_rule) {
1616			status = ICE_ERR_NO_MEMORY;
1617			goto exit;
1618		}
1619
1620		ice_fill_sw_rule(hw, &list_elem->fltr_info, s_rule,
1621				 ice_aqc_opc_remove_sw_rules);
1622
1623		status = ice_aq_sw_rules(hw, s_rule,
1624					 ICE_SW_RULE_RX_TX_NO_HDR_SIZE, 1,
1625					 ice_aqc_opc_remove_sw_rules, NULL);
1626
1627		/* Remove a book keeping from the list */
1628		devm_kfree(ice_hw_to_dev(hw), s_rule);
1629
1630		if (status)
1631			goto exit;
1632
1633		list_del(&list_elem->list_entry);
1634		devm_kfree(ice_hw_to_dev(hw), list_elem);
1635	}
1636exit:
1637	mutex_unlock(rule_lock);
1638	return status;
1639}
1640
1641/**
1642 * ice_add_mac - Add a MAC address based filter rule
1643 * @hw: pointer to the hardware structure
1644 * @m_list: list of MAC addresses and forwarding information
1645 *
1646 * IMPORTANT: When the ucast_shared flag is set to false and m_list has
1647 * multiple unicast addresses, the function assumes that all the
1648 * addresses are unique in a given add_mac call. It doesn't
1649 * check for duplicates in this case, removing duplicates from a given
1650 * list should be taken care of in the caller of this function.
1651 */
1652enum ice_status
1653ice_add_mac(struct ice_hw *hw, struct list_head *m_list)
1654{
1655	struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1656	struct ice_fltr_list_entry *m_list_itr;
1657	struct list_head *rule_head;
1658	u16 elem_sent, total_elem_left;
1659	struct ice_switch_info *sw;
1660	struct mutex *rule_lock; /* Lock to protect filter rule list */
1661	enum ice_status status = 0;
1662	u16 num_unicast = 0;
1663	u16 s_rule_size;
1664
1665	if (!m_list || !hw)
1666		return ICE_ERR_PARAM;
1667
1668	s_rule = NULL;
1669	sw = hw->switch_info;
1670	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1671	list_for_each_entry(m_list_itr, m_list, list_entry) {
1672		u8 *add = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1673		u16 vsi_handle;
1674		u16 hw_vsi_id;
1675
1676		m_list_itr->fltr_info.flag = ICE_FLTR_TX;
1677		vsi_handle = m_list_itr->fltr_info.vsi_handle;
1678		if (!ice_is_vsi_valid(hw, vsi_handle))
1679			return ICE_ERR_PARAM;
1680		hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
1681		m_list_itr->fltr_info.fwd_id.hw_vsi_id = hw_vsi_id;
1682		/* update the src in case it is VSI num */
1683		if (m_list_itr->fltr_info.src_id != ICE_SRC_ID_VSI)
1684			return ICE_ERR_PARAM;
1685		m_list_itr->fltr_info.src = hw_vsi_id;
1686		if (m_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_MAC ||
1687		    is_zero_ether_addr(add))
1688			return ICE_ERR_PARAM;
1689		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
1690			/* Don't overwrite the unicast address */
1691			mutex_lock(rule_lock);
1692			if (ice_find_rule_entry(hw, ICE_SW_LKUP_MAC,
1693						&m_list_itr->fltr_info)) {
1694				mutex_unlock(rule_lock);
1695				return ICE_ERR_ALREADY_EXISTS;
1696			}
1697			mutex_unlock(rule_lock);
1698			num_unicast++;
1699		} else if (is_multicast_ether_addr(add) ||
1700			   (is_unicast_ether_addr(add) && hw->ucast_shared)) {
1701			m_list_itr->status =
1702				ice_add_rule_internal(hw, ICE_SW_LKUP_MAC,
1703						      m_list_itr);
1704			if (m_list_itr->status)
1705				return m_list_itr->status;
 
1706		}
1707	}
1708
1709	mutex_lock(rule_lock);
1710	/* Exit if no suitable entries were found for adding bulk switch rule */
1711	if (!num_unicast) {
1712		status = 0;
1713		goto ice_add_mac_exit;
1714	}
1715
1716	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1717
1718	/* Allocate switch rule buffer for the bulk update for unicast */
1719	s_rule_size = ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
1720	s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1721			      GFP_KERNEL);
1722	if (!s_rule) {
1723		status = ICE_ERR_NO_MEMORY;
1724		goto ice_add_mac_exit;
1725	}
1726
1727	r_iter = s_rule;
1728	list_for_each_entry(m_list_itr, m_list, list_entry) {
1729		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1730		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
1731
1732		if (is_unicast_ether_addr(mac_addr)) {
1733			ice_fill_sw_rule(hw, &m_list_itr->fltr_info, r_iter,
1734					 ice_aqc_opc_add_sw_rules);
1735			r_iter = (struct ice_aqc_sw_rules_elem *)
1736				((u8 *)r_iter + s_rule_size);
1737		}
1738	}
1739
1740	/* Call AQ bulk switch rule update for all unicast addresses */
1741	r_iter = s_rule;
1742	/* Call AQ switch rule in AQ_MAX chunk */
1743	for (total_elem_left = num_unicast; total_elem_left > 0;
1744	     total_elem_left -= elem_sent) {
1745		struct ice_aqc_sw_rules_elem *entry = r_iter;
1746
1747		elem_sent = min(total_elem_left,
1748				(u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1749		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1750					 elem_sent, ice_aqc_opc_add_sw_rules,
1751					 NULL);
1752		if (status)
1753			goto ice_add_mac_exit;
1754		r_iter = (struct ice_aqc_sw_rules_elem *)
1755			((u8 *)r_iter + (elem_sent * s_rule_size));
1756	}
1757
1758	/* Fill up rule ID based on the value returned from FW */
1759	r_iter = s_rule;
1760	list_for_each_entry(m_list_itr, m_list, list_entry) {
1761		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1762		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
 
1763		struct ice_fltr_mgmt_list_entry *fm_entry;
1764
1765		if (is_unicast_ether_addr(mac_addr)) {
1766			f_info->fltr_rule_id =
1767				le16_to_cpu(r_iter->pdata.lkup_tx_rx.index);
1768			f_info->fltr_act = ICE_FWD_TO_VSI;
1769			/* Create an entry to track this MAC address */
1770			fm_entry = devm_kzalloc(ice_hw_to_dev(hw),
1771						sizeof(*fm_entry), GFP_KERNEL);
1772			if (!fm_entry) {
1773				status = ICE_ERR_NO_MEMORY;
1774				goto ice_add_mac_exit;
1775			}
1776			fm_entry->fltr_info = *f_info;
1777			fm_entry->vsi_count = 1;
1778			/* The book keeping entries will get removed when
1779			 * base driver calls remove filter AQ command
1780			 */
 
 
 
1781
1782			list_add(&fm_entry->list_entry, rule_head);
1783			r_iter = (struct ice_aqc_sw_rules_elem *)
1784				((u8 *)r_iter + s_rule_size);
1785		}
1786	}
1787
1788ice_add_mac_exit:
1789	mutex_unlock(rule_lock);
1790	if (s_rule)
1791		devm_kfree(ice_hw_to_dev(hw), s_rule);
1792	return status;
1793}
1794
1795/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1796 * ice_add_vlan_internal - Add one VLAN based filter rule
1797 * @hw: pointer to the hardware structure
1798 * @f_entry: filter entry containing one VLAN information
1799 */
1800static enum ice_status
1801ice_add_vlan_internal(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1802{
1803	struct ice_switch_info *sw = hw->switch_info;
1804	struct ice_fltr_mgmt_list_entry *v_list_itr;
1805	struct ice_fltr_info *new_fltr, *cur_fltr;
1806	enum ice_sw_lkup_type lkup_type;
1807	u16 vsi_list_id = 0, vsi_handle;
1808	struct mutex *rule_lock; /* Lock to protect filter rule list */
1809	enum ice_status status = 0;
1810
1811	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1812		return ICE_ERR_PARAM;
1813
1814	f_entry->fltr_info.fwd_id.hw_vsi_id =
1815		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1816	new_fltr = &f_entry->fltr_info;
1817
1818	/* VLAN ID should only be 12 bits */
1819	if (new_fltr->l_data.vlan.vlan_id > ICE_MAX_VLAN_ID)
1820		return ICE_ERR_PARAM;
1821
1822	if (new_fltr->src_id != ICE_SRC_ID_VSI)
1823		return ICE_ERR_PARAM;
1824
1825	new_fltr->src = new_fltr->fwd_id.hw_vsi_id;
1826	lkup_type = new_fltr->lkup_type;
1827	vsi_handle = new_fltr->vsi_handle;
1828	rule_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
1829	mutex_lock(rule_lock);
1830	v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN, new_fltr);
1831	if (!v_list_itr) {
1832		struct ice_vsi_list_map_info *map_info = NULL;
 
 
1833
1834		if (new_fltr->fltr_act == ICE_FWD_TO_VSI) {
1835			/* All VLAN pruning rules use a VSI list. Check if
1836			 * there is already a VSI list containing VSI that we
1837			 * want to add. If found, use the same vsi_list_id for
1838			 * this new VLAN rule or else create a new list.
1839			 */
1840			map_info = ice_find_vsi_list_entry(hw, ICE_SW_LKUP_VLAN,
1841							   vsi_handle,
1842							   &vsi_list_id);
1843			if (!map_info) {
1844				status = ice_create_vsi_list_rule(hw,
1845								  &vsi_handle,
1846								  1,
1847								  &vsi_list_id,
1848								  lkup_type);
1849				if (status)
1850					goto exit;
1851			}
1852			/* Convert the action to forwarding to a VSI list. */
1853			new_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1854			new_fltr->fwd_id.vsi_list_id = vsi_list_id;
1855		}
1856
1857		status = ice_create_pkt_fwd_rule(hw, f_entry);
1858		if (!status) {
1859			v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN,
1860							 new_fltr);
1861			if (!v_list_itr) {
1862				status = ICE_ERR_DOES_NOT_EXIST;
1863				goto exit;
1864			}
1865			/* reuse VSI list for new rule and increment ref_cnt */
1866			if (map_info) {
1867				v_list_itr->vsi_list_info = map_info;
1868				map_info->ref_cnt++;
1869			} else {
1870				v_list_itr->vsi_list_info =
1871					ice_create_vsi_list_map(hw, &vsi_handle,
1872								1, vsi_list_id);
1873			}
1874		}
1875	} else if (v_list_itr->vsi_list_info->ref_cnt == 1) {
1876		/* Update existing VSI list to add new VSI ID only if it used
1877		 * by one VLAN rule.
1878		 */
1879		cur_fltr = &v_list_itr->fltr_info;
1880		status = ice_add_update_vsi_list(hw, v_list_itr, cur_fltr,
1881						 new_fltr);
1882	} else {
1883		/* If VLAN rule exists and VSI list being used by this rule is
1884		 * referenced by more than 1 VLAN rule. Then create a new VSI
1885		 * list appending previous VSI with new VSI and update existing
1886		 * VLAN rule to point to new VSI list ID
1887		 */
1888		struct ice_fltr_info tmp_fltr;
1889		u16 vsi_handle_arr[2];
1890		u16 cur_handle;
1891
1892		/* Current implementation only supports reusing VSI list with
1893		 * one VSI count. We should never hit below condition
1894		 */
1895		if (v_list_itr->vsi_count > 1 &&
1896		    v_list_itr->vsi_list_info->ref_cnt > 1) {
1897			ice_debug(hw, ICE_DBG_SW,
1898				  "Invalid configuration: Optimization to reuse VSI list with more than one VSI is not being done yet\n");
1899			status = ICE_ERR_CFG;
1900			goto exit;
1901		}
1902
1903		cur_handle =
1904			find_first_bit(v_list_itr->vsi_list_info->vsi_map,
1905				       ICE_MAX_VSI);
1906
1907		/* A rule already exists with the new VSI being added */
1908		if (cur_handle == vsi_handle) {
1909			status = ICE_ERR_ALREADY_EXISTS;
1910			goto exit;
1911		}
1912
1913		vsi_handle_arr[0] = cur_handle;
1914		vsi_handle_arr[1] = vsi_handle;
1915		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1916						  &vsi_list_id, lkup_type);
1917		if (status)
1918			goto exit;
1919
1920		tmp_fltr = v_list_itr->fltr_info;
1921		tmp_fltr.fltr_rule_id = v_list_itr->fltr_info.fltr_rule_id;
1922		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1923		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1924		/* Update the previous switch rule to a new VSI list which
1925		 * includes current VSI that is requested
1926		 */
1927		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
1928		if (status)
1929			goto exit;
1930
1931		/* before overriding VSI list map info. decrement ref_cnt of
1932		 * previous VSI list
1933		 */
1934		v_list_itr->vsi_list_info->ref_cnt--;
1935
1936		/* now update to newly created list */
1937		v_list_itr->fltr_info.fwd_id.vsi_list_id = vsi_list_id;
1938		v_list_itr->vsi_list_info =
1939			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1940						vsi_list_id);
1941		v_list_itr->vsi_count++;
1942	}
1943
1944exit:
1945	mutex_unlock(rule_lock);
1946	return status;
1947}
1948
1949/**
1950 * ice_add_vlan - Add VLAN based filter rule
1951 * @hw: pointer to the hardware structure
1952 * @v_list: list of VLAN entries and forwarding information
1953 */
1954enum ice_status
1955ice_add_vlan(struct ice_hw *hw, struct list_head *v_list)
1956{
1957	struct ice_fltr_list_entry *v_list_itr;
1958
1959	if (!v_list || !hw)
1960		return ICE_ERR_PARAM;
1961
1962	list_for_each_entry(v_list_itr, v_list, list_entry) {
 
 
1963		if (v_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_VLAN)
1964			return ICE_ERR_PARAM;
1965		v_list_itr->fltr_info.flag = ICE_FLTR_TX;
1966		v_list_itr->status = ice_add_vlan_internal(hw, v_list_itr);
1967		if (v_list_itr->status)
1968			return v_list_itr->status;
 
 
 
1969	}
1970	return 0;
1971}
1972
1973/**
1974 * ice_add_eth_mac - Add ethertype and MAC based filter rule
1975 * @hw: pointer to the hardware structure
1976 * @em_list: list of ether type MAC filter, MAC is optional
1977 *
1978 * This function requires the caller to populate the entries in
1979 * the filter list with the necessary fields (including flags to
1980 * indicate Tx or Rx rules).
1981 */
1982enum ice_status
1983ice_add_eth_mac(struct ice_hw *hw, struct list_head *em_list)
 
1984{
1985	struct ice_fltr_list_entry *em_list_itr;
 
 
1986
1987	if (!em_list || !hw)
1988		return ICE_ERR_PARAM;
 
 
1989
1990	list_for_each_entry(em_list_itr, em_list, list_entry) {
1991		enum ice_sw_lkup_type l_type =
1992			em_list_itr->fltr_info.lkup_type;
 
 
 
1993
1994		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
1995		    l_type != ICE_SW_LKUP_ETHERTYPE)
1996			return ICE_ERR_PARAM;
 
 
 
1997
1998		em_list_itr->status = ice_add_rule_internal(hw, l_type,
1999							    em_list_itr);
2000		if (em_list_itr->status)
2001			return em_list_itr->status;
2002	}
2003	return 0;
2004}
2005
2006/**
2007 * ice_remove_eth_mac - Remove an ethertype (or MAC) based filter rule
2008 * @hw: pointer to the hardware structure
2009 * @em_list: list of ethertype or ethertype MAC entries
 
 
2010 */
2011enum ice_status
2012ice_remove_eth_mac(struct ice_hw *hw, struct list_head *em_list)
 
2013{
2014	struct ice_fltr_list_entry *em_list_itr, *tmp;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2015
2016	if (!em_list || !hw)
2017		return ICE_ERR_PARAM;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2018
2019	list_for_each_entry_safe(em_list_itr, tmp, em_list, list_entry) {
2020		enum ice_sw_lkup_type l_type =
2021			em_list_itr->fltr_info.lkup_type;
2022
2023		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
2024		    l_type != ICE_SW_LKUP_ETHERTYPE)
2025			return ICE_ERR_PARAM;
 
 
2026
2027		em_list_itr->status = ice_remove_rule_internal(hw, l_type,
2028							       em_list_itr);
2029		if (em_list_itr->status)
2030			return em_list_itr->status;
 
 
2031	}
2032	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2033}
2034
2035/**
2036 * ice_rem_sw_rule_info
2037 * @hw: pointer to the hardware structure
2038 * @rule_head: pointer to the switch list structure that we want to delete
 
 
 
 
 
 
 
 
2039 */
2040static void
2041ice_rem_sw_rule_info(struct ice_hw *hw, struct list_head *rule_head)
2042{
2043	if (!list_empty(rule_head)) {
2044		struct ice_fltr_mgmt_list_entry *entry;
2045		struct ice_fltr_mgmt_list_entry *tmp;
2046
2047		list_for_each_entry_safe(entry, tmp, rule_head, list_entry) {
2048			list_del(&entry->list_entry);
2049			devm_kfree(ice_hw_to_dev(hw), entry);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2050		}
2051	}
 
 
 
 
2052}
2053
2054/**
2055 * ice_cfg_dflt_vsi - change state of VSI to set/clear default
 
2056 * @hw: pointer to the hardware structure
2057 * @vsi_handle: VSI handle to set as default
2058 * @set: true to add the above mentioned switch rule, false to remove it
2059 * @direction: ICE_FLTR_RX or ICE_FLTR_TX
2060 *
2061 * add filter rule to set/unset given VSI as default VSI for the switch
2062 * (represented by swid)
2063 */
2064enum ice_status
2065ice_cfg_dflt_vsi(struct ice_hw *hw, u16 vsi_handle, bool set, u8 direction)
2066{
2067	struct ice_aqc_sw_rules_elem *s_rule;
2068	struct ice_fltr_info f_info;
2069	enum ice_adminq_opc opcode;
2070	enum ice_status status;
2071	u16 s_rule_size;
2072	u16 hw_vsi_id;
2073
2074	if (!ice_is_vsi_valid(hw, vsi_handle))
2075		return ICE_ERR_PARAM;
2076	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2077
2078	s_rule_size = set ? ICE_SW_RULE_RX_TX_ETH_HDR_SIZE :
2079			    ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
2080	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
2081	if (!s_rule)
2082		return ICE_ERR_NO_MEMORY;
2083
2084	memset(&f_info, 0, sizeof(f_info));
2085
2086	f_info.lkup_type = ICE_SW_LKUP_DFLT;
2087	f_info.flag = direction;
2088	f_info.fltr_act = ICE_FWD_TO_VSI;
2089	f_info.fwd_id.hw_vsi_id = hw_vsi_id;
2090
2091	if (f_info.flag & ICE_FLTR_RX) {
2092		f_info.src = hw->port_info->lport;
2093		f_info.src_id = ICE_SRC_ID_LPORT;
2094		if (!set)
2095			f_info.fltr_rule_id =
2096				hw->port_info->dflt_rx_vsi_rule_id;
2097	} else if (f_info.flag & ICE_FLTR_TX) {
2098		f_info.src_id = ICE_SRC_ID_VSI;
2099		f_info.src = hw_vsi_id;
2100		if (!set)
2101			f_info.fltr_rule_id =
2102				hw->port_info->dflt_tx_vsi_rule_id;
2103	}
2104
2105	if (set)
2106		opcode = ice_aqc_opc_add_sw_rules;
2107	else
2108		opcode = ice_aqc_opc_remove_sw_rules;
2109
2110	ice_fill_sw_rule(hw, &f_info, s_rule, opcode);
2111
2112	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opcode, NULL);
2113	if (status || !(f_info.flag & ICE_FLTR_TX_RX))
2114		goto out;
2115	if (set) {
2116		u16 index = le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
2117
2118		if (f_info.flag & ICE_FLTR_TX) {
2119			hw->port_info->dflt_tx_vsi_num = hw_vsi_id;
2120			hw->port_info->dflt_tx_vsi_rule_id = index;
2121		} else if (f_info.flag & ICE_FLTR_RX) {
2122			hw->port_info->dflt_rx_vsi_num = hw_vsi_id;
2123			hw->port_info->dflt_rx_vsi_rule_id = index;
2124		}
2125	} else {
2126		if (f_info.flag & ICE_FLTR_TX) {
2127			hw->port_info->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
2128			hw->port_info->dflt_tx_vsi_rule_id = ICE_INVAL_ACT;
2129		} else if (f_info.flag & ICE_FLTR_RX) {
2130			hw->port_info->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
2131			hw->port_info->dflt_rx_vsi_rule_id = ICE_INVAL_ACT;
2132		}
2133	}
2134
2135out:
2136	devm_kfree(ice_hw_to_dev(hw), s_rule);
2137	return status;
2138}
2139
2140/**
2141 * ice_find_ucast_rule_entry - Search for a unicast MAC filter rule entry
2142 * @hw: pointer to the hardware structure
2143 * @recp_id: lookup type for which the specified rule needs to be searched
2144 * @f_info: rule information
2145 *
2146 * Helper function to search for a unicast rule entry - this is to be used
2147 * to remove unicast MAC filter that is not shared with other VSIs on the
2148 * PF switch.
2149 *
2150 * Returns pointer to entry storing the rule if found
2151 */
2152static struct ice_fltr_mgmt_list_entry *
2153ice_find_ucast_rule_entry(struct ice_hw *hw, u8 recp_id,
2154			  struct ice_fltr_info *f_info)
2155{
2156	struct ice_switch_info *sw = hw->switch_info;
2157	struct ice_fltr_mgmt_list_entry *list_itr;
2158	struct list_head *list_head;
2159
2160	list_head = &sw->recp_list[recp_id].filt_rules;
2161	list_for_each_entry(list_itr, list_head, list_entry) {
2162		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
2163			    sizeof(f_info->l_data)) &&
2164		    f_info->fwd_id.hw_vsi_id ==
2165		    list_itr->fltr_info.fwd_id.hw_vsi_id &&
2166		    f_info->flag == list_itr->fltr_info.flag)
2167			return list_itr;
2168	}
2169	return NULL;
2170}
2171
2172/**
2173 * ice_remove_mac - remove a MAC address based filter rule
2174 * @hw: pointer to the hardware structure
2175 * @m_list: list of MAC addresses and forwarding information
2176 *
2177 * This function removes either a MAC filter rule or a specific VSI from a
2178 * VSI list for a multicast MAC address.
2179 *
2180 * Returns ICE_ERR_DOES_NOT_EXIST if a given entry was not added by
2181 * ice_add_mac. Caller should be aware that this call will only work if all
2182 * the entries passed into m_list were added previously. It will not attempt to
2183 * do a partial remove of entries that were found.
2184 */
2185enum ice_status
2186ice_remove_mac(struct ice_hw *hw, struct list_head *m_list)
2187{
2188	struct ice_fltr_list_entry *list_itr, *tmp;
2189	struct mutex *rule_lock; /* Lock to protect filter rule list */
2190
2191	if (!m_list)
2192		return ICE_ERR_PARAM;
2193
2194	rule_lock = &hw->switch_info->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
2195	list_for_each_entry_safe(list_itr, tmp, m_list, list_entry) {
2196		enum ice_sw_lkup_type l_type = list_itr->fltr_info.lkup_type;
2197		u8 *add = &list_itr->fltr_info.l_data.mac.mac_addr[0];
2198		u16 vsi_handle;
2199
2200		if (l_type != ICE_SW_LKUP_MAC)
2201			return ICE_ERR_PARAM;
2202
2203		vsi_handle = list_itr->fltr_info.vsi_handle;
2204		if (!ice_is_vsi_valid(hw, vsi_handle))
2205			return ICE_ERR_PARAM;
2206
2207		list_itr->fltr_info.fwd_id.hw_vsi_id =
2208					ice_get_hw_vsi_num(hw, vsi_handle);
2209		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
2210			/* Don't remove the unicast address that belongs to
2211			 * another VSI on the switch, since it is not being
2212			 * shared...
2213			 */
2214			mutex_lock(rule_lock);
2215			if (!ice_find_ucast_rule_entry(hw, ICE_SW_LKUP_MAC,
2216						       &list_itr->fltr_info)) {
2217				mutex_unlock(rule_lock);
2218				return ICE_ERR_DOES_NOT_EXIST;
2219			}
2220			mutex_unlock(rule_lock);
2221		}
2222		list_itr->status = ice_remove_rule_internal(hw,
2223							    ICE_SW_LKUP_MAC,
2224							    list_itr);
2225		if (list_itr->status)
2226			return list_itr->status;
2227	}
2228	return 0;
2229}
2230
2231/**
2232 * ice_remove_vlan - Remove VLAN based filter rule
2233 * @hw: pointer to the hardware structure
2234 * @v_list: list of VLAN entries and forwarding information
2235 */
2236enum ice_status
2237ice_remove_vlan(struct ice_hw *hw, struct list_head *v_list)
2238{
2239	struct ice_fltr_list_entry *v_list_itr, *tmp;
 
2240
2241	if (!v_list || !hw)
2242		return ICE_ERR_PARAM;
2243
2244	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2245		enum ice_sw_lkup_type l_type = v_list_itr->fltr_info.lkup_type;
2246
2247		if (l_type != ICE_SW_LKUP_VLAN)
2248			return ICE_ERR_PARAM;
2249		v_list_itr->status = ice_remove_rule_internal(hw,
2250							      ICE_SW_LKUP_VLAN,
2251							      v_list_itr);
2252		if (v_list_itr->status)
2253			return v_list_itr->status;
2254	}
2255	return 0;
2256}
2257
2258/**
2259 * ice_vsi_uses_fltr - Determine if given VSI uses specified filter
2260 * @fm_entry: filter entry to inspect
2261 * @vsi_handle: VSI handle to compare with filter info
2262 */
2263static bool
2264ice_vsi_uses_fltr(struct ice_fltr_mgmt_list_entry *fm_entry, u16 vsi_handle)
2265{
2266	return ((fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI &&
2267		 fm_entry->fltr_info.vsi_handle == vsi_handle) ||
2268		(fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI_LIST &&
2269		 (test_bit(vsi_handle, fm_entry->vsi_list_info->vsi_map))));
2270}
2271
2272/**
2273 * ice_add_entry_to_vsi_fltr_list - Add copy of fltr_list_entry to remove list
2274 * @hw: pointer to the hardware structure
2275 * @vsi_handle: VSI handle to remove filters from
2276 * @vsi_list_head: pointer to the list to add entry to
2277 * @fi: pointer to fltr_info of filter entry to copy & add
2278 *
2279 * Helper function, used when creating a list of filters to remove from
2280 * a specific VSI. The entry added to vsi_list_head is a COPY of the
2281 * original filter entry, with the exception of fltr_info.fltr_act and
2282 * fltr_info.fwd_id fields. These are set such that later logic can
2283 * extract which VSI to remove the fltr from, and pass on that information.
2284 */
2285static enum ice_status
2286ice_add_entry_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2287			       struct list_head *vsi_list_head,
2288			       struct ice_fltr_info *fi)
2289{
2290	struct ice_fltr_list_entry *tmp;
2291
2292	/* this memory is freed up in the caller function
2293	 * once filters for this VSI are removed
2294	 */
2295	tmp = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*tmp), GFP_KERNEL);
2296	if (!tmp)
2297		return ICE_ERR_NO_MEMORY;
2298
2299	tmp->fltr_info = *fi;
2300
2301	/* Overwrite these fields to indicate which VSI to remove filter from,
2302	 * so find and remove logic can extract the information from the
2303	 * list entries. Note that original entries will still have proper
2304	 * values.
2305	 */
2306	tmp->fltr_info.fltr_act = ICE_FWD_TO_VSI;
2307	tmp->fltr_info.vsi_handle = vsi_handle;
2308	tmp->fltr_info.fwd_id.hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2309
2310	list_add(&tmp->list_entry, vsi_list_head);
2311
2312	return 0;
2313}
2314
2315/**
2316 * ice_add_to_vsi_fltr_list - Add VSI filters to the list
2317 * @hw: pointer to the hardware structure
2318 * @vsi_handle: VSI handle to remove filters from
2319 * @lkup_list_head: pointer to the list that has certain lookup type filters
2320 * @vsi_list_head: pointer to the list pertaining to VSI with vsi_handle
2321 *
2322 * Locates all filters in lkup_list_head that are used by the given VSI,
2323 * and adds COPIES of those entries to vsi_list_head (intended to be used
2324 * to remove the listed filters).
2325 * Note that this means all entries in vsi_list_head must be explicitly
2326 * deallocated by the caller when done with list.
2327 */
2328static enum ice_status
2329ice_add_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2330			 struct list_head *lkup_list_head,
2331			 struct list_head *vsi_list_head)
2332{
2333	struct ice_fltr_mgmt_list_entry *fm_entry;
2334	enum ice_status status = 0;
2335
2336	/* check to make sure VSI ID is valid and within boundary */
2337	if (!ice_is_vsi_valid(hw, vsi_handle))
 
2338		return ICE_ERR_PARAM;
2339
2340	list_for_each_entry(fm_entry, lkup_list_head, list_entry) {
2341		struct ice_fltr_info *fi;
2342
2343		fi = &fm_entry->fltr_info;
2344		if (!fi || !ice_vsi_uses_fltr(fm_entry, vsi_handle))
2345			continue;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2346
2347		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2348							vsi_list_head, fi);
2349		if (status)
2350			return status;
2351	}
2352	return status;
2353}
2354
2355/**
2356 * ice_determine_promisc_mask
2357 * @fi: filter info to parse
2358 *
2359 * Helper function to determine which ICE_PROMISC_ mask corresponds
2360 * to given filter into.
2361 */
2362static u8 ice_determine_promisc_mask(struct ice_fltr_info *fi)
2363{
2364	u16 vid = fi->l_data.mac_vlan.vlan_id;
2365	u8 *macaddr = fi->l_data.mac.mac_addr;
2366	bool is_tx_fltr = false;
2367	u8 promisc_mask = 0;
2368
2369	if (fi->flag == ICE_FLTR_TX)
2370		is_tx_fltr = true;
2371
2372	if (is_broadcast_ether_addr(macaddr))
2373		promisc_mask |= is_tx_fltr ?
2374			ICE_PROMISC_BCAST_TX : ICE_PROMISC_BCAST_RX;
2375	else if (is_multicast_ether_addr(macaddr))
2376		promisc_mask |= is_tx_fltr ?
2377			ICE_PROMISC_MCAST_TX : ICE_PROMISC_MCAST_RX;
2378	else if (is_unicast_ether_addr(macaddr))
2379		promisc_mask |= is_tx_fltr ?
2380			ICE_PROMISC_UCAST_TX : ICE_PROMISC_UCAST_RX;
2381	if (vid)
2382		promisc_mask |= is_tx_fltr ?
2383			ICE_PROMISC_VLAN_TX : ICE_PROMISC_VLAN_RX;
2384
2385	return promisc_mask;
2386}
2387
2388/**
2389 * ice_remove_promisc - Remove promisc based filter rules
2390 * @hw: pointer to the hardware structure
2391 * @recp_id: recipe ID for which the rule needs to removed
2392 * @v_list: list of promisc entries
2393 */
2394static enum ice_status
2395ice_remove_promisc(struct ice_hw *hw, u8 recp_id,
2396		   struct list_head *v_list)
2397{
2398	struct ice_fltr_list_entry *v_list_itr, *tmp;
2399
2400	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2401		v_list_itr->status =
2402			ice_remove_rule_internal(hw, recp_id, v_list_itr);
2403		if (v_list_itr->status)
2404			return v_list_itr->status;
2405	}
2406	return 0;
2407}
2408
2409/**
2410 * ice_clear_vsi_promisc - clear specified promiscuous mode(s) for given VSI
2411 * @hw: pointer to the hardware structure
2412 * @vsi_handle: VSI handle to clear mode
2413 * @promisc_mask: mask of promiscuous config bits to clear
2414 * @vid: VLAN ID to clear VLAN promiscuous
2415 */
2416enum ice_status
2417ice_clear_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2418		      u16 vid)
2419{
2420	struct ice_switch_info *sw = hw->switch_info;
2421	struct ice_fltr_list_entry *fm_entry, *tmp;
2422	struct list_head remove_list_head;
2423	struct ice_fltr_mgmt_list_entry *itr;
2424	struct list_head *rule_head;
2425	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2426	enum ice_status status = 0;
2427	u8 recipe_id;
2428
2429	if (!ice_is_vsi_valid(hw, vsi_handle))
2430		return ICE_ERR_PARAM;
2431
2432	if (vid)
2433		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2434	else
2435		recipe_id = ICE_SW_LKUP_PROMISC;
2436
2437	rule_head = &sw->recp_list[recipe_id].filt_rules;
2438	rule_lock = &sw->recp_list[recipe_id].filt_rule_lock;
2439
2440	INIT_LIST_HEAD(&remove_list_head);
2441
2442	mutex_lock(rule_lock);
2443	list_for_each_entry(itr, rule_head, list_entry) {
2444		u8 fltr_promisc_mask = 0;
2445
2446		if (!ice_vsi_uses_fltr(itr, vsi_handle))
2447			continue;
2448
2449		fltr_promisc_mask |=
2450			ice_determine_promisc_mask(&itr->fltr_info);
2451
2452		/* Skip if filter is not completely specified by given mask */
2453		if (fltr_promisc_mask & ~promisc_mask)
2454			continue;
2455
2456		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2457							&remove_list_head,
2458							&itr->fltr_info);
2459		if (status) {
2460			mutex_unlock(rule_lock);
2461			goto free_fltr_list;
2462		}
2463	}
2464	mutex_unlock(rule_lock);
2465
2466	status = ice_remove_promisc(hw, recipe_id, &remove_list_head);
2467
2468free_fltr_list:
2469	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2470		list_del(&fm_entry->list_entry);
2471		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2472	}
2473
2474	return status;
2475}
2476
2477/**
2478 * ice_set_vsi_promisc - set given VSI to given promiscuous mode(s)
2479 * @hw: pointer to the hardware structure
2480 * @vsi_handle: VSI handle to configure
2481 * @promisc_mask: mask of promiscuous config bits
2482 * @vid: VLAN ID to set VLAN promiscuous
2483 */
2484enum ice_status
2485ice_set_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask, u16 vid)
2486{
2487	enum { UCAST_FLTR = 1, MCAST_FLTR, BCAST_FLTR };
2488	struct ice_fltr_list_entry f_list_entry;
2489	struct ice_fltr_info new_fltr;
2490	enum ice_status status = 0;
2491	bool is_tx_fltr;
2492	u16 hw_vsi_id;
2493	int pkt_type;
2494	u8 recipe_id;
2495
2496	if (!ice_is_vsi_valid(hw, vsi_handle))
2497		return ICE_ERR_PARAM;
2498	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2499
2500	memset(&new_fltr, 0, sizeof(new_fltr));
2501
2502	if (promisc_mask & (ICE_PROMISC_VLAN_RX | ICE_PROMISC_VLAN_TX)) {
2503		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC_VLAN;
2504		new_fltr.l_data.mac_vlan.vlan_id = vid;
2505		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2506	} else {
2507		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC;
2508		recipe_id = ICE_SW_LKUP_PROMISC;
2509	}
2510
2511	/* Separate filters must be set for each direction/packet type
2512	 * combination, so we will loop over the mask value, store the
2513	 * individual type, and clear it out in the input mask as it
2514	 * is found.
2515	 */
2516	while (promisc_mask) {
2517		u8 *mac_addr;
2518
2519		pkt_type = 0;
2520		is_tx_fltr = false;
2521
2522		if (promisc_mask & ICE_PROMISC_UCAST_RX) {
2523			promisc_mask &= ~ICE_PROMISC_UCAST_RX;
2524			pkt_type = UCAST_FLTR;
2525		} else if (promisc_mask & ICE_PROMISC_UCAST_TX) {
2526			promisc_mask &= ~ICE_PROMISC_UCAST_TX;
2527			pkt_type = UCAST_FLTR;
2528			is_tx_fltr = true;
2529		} else if (promisc_mask & ICE_PROMISC_MCAST_RX) {
2530			promisc_mask &= ~ICE_PROMISC_MCAST_RX;
2531			pkt_type = MCAST_FLTR;
2532		} else if (promisc_mask & ICE_PROMISC_MCAST_TX) {
2533			promisc_mask &= ~ICE_PROMISC_MCAST_TX;
2534			pkt_type = MCAST_FLTR;
2535			is_tx_fltr = true;
2536		} else if (promisc_mask & ICE_PROMISC_BCAST_RX) {
2537			promisc_mask &= ~ICE_PROMISC_BCAST_RX;
2538			pkt_type = BCAST_FLTR;
2539		} else if (promisc_mask & ICE_PROMISC_BCAST_TX) {
2540			promisc_mask &= ~ICE_PROMISC_BCAST_TX;
2541			pkt_type = BCAST_FLTR;
2542			is_tx_fltr = true;
2543		}
2544
2545		/* Check for VLAN promiscuous flag */
2546		if (promisc_mask & ICE_PROMISC_VLAN_RX) {
2547			promisc_mask &= ~ICE_PROMISC_VLAN_RX;
2548		} else if (promisc_mask & ICE_PROMISC_VLAN_TX) {
2549			promisc_mask &= ~ICE_PROMISC_VLAN_TX;
2550			is_tx_fltr = true;
2551		}
2552
2553		/* Set filter DA based on packet type */
2554		mac_addr = new_fltr.l_data.mac.mac_addr;
2555		if (pkt_type == BCAST_FLTR) {
2556			eth_broadcast_addr(mac_addr);
2557		} else if (pkt_type == MCAST_FLTR ||
2558			   pkt_type == UCAST_FLTR) {
2559			/* Use the dummy ether header DA */
2560			ether_addr_copy(mac_addr, dummy_eth_header);
2561			if (pkt_type == MCAST_FLTR)
2562				mac_addr[0] |= 0x1;	/* Set multicast bit */
2563		}
2564
2565		/* Need to reset this to zero for all iterations */
2566		new_fltr.flag = 0;
2567		if (is_tx_fltr) {
2568			new_fltr.flag |= ICE_FLTR_TX;
2569			new_fltr.src = hw_vsi_id;
2570		} else {
2571			new_fltr.flag |= ICE_FLTR_RX;
2572			new_fltr.src = hw->port_info->lport;
2573		}
2574
2575		new_fltr.fltr_act = ICE_FWD_TO_VSI;
2576		new_fltr.vsi_handle = vsi_handle;
2577		new_fltr.fwd_id.hw_vsi_id = hw_vsi_id;
2578		f_list_entry.fltr_info = new_fltr;
2579
2580		status = ice_add_rule_internal(hw, recipe_id, &f_list_entry);
2581		if (status)
2582			goto set_promisc_exit;
2583	}
2584
2585set_promisc_exit:
2586	return status;
2587}
2588
2589/**
2590 * ice_set_vlan_vsi_promisc
2591 * @hw: pointer to the hardware structure
2592 * @vsi_handle: VSI handle to configure
2593 * @promisc_mask: mask of promiscuous config bits
2594 * @rm_vlan_promisc: Clear VLANs VSI promisc mode
2595 *
2596 * Configure VSI with all associated VLANs to given promiscuous mode(s)
2597 */
2598enum ice_status
2599ice_set_vlan_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2600			 bool rm_vlan_promisc)
2601{
2602	struct ice_switch_info *sw = hw->switch_info;
2603	struct ice_fltr_list_entry *list_itr, *tmp;
2604	struct list_head vsi_list_head;
2605	struct list_head *vlan_head;
2606	struct mutex *vlan_lock; /* Lock to protect filter rule list */
2607	enum ice_status status;
2608	u16 vlan_id;
2609
2610	INIT_LIST_HEAD(&vsi_list_head);
2611	vlan_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
2612	vlan_head = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rules;
2613	mutex_lock(vlan_lock);
2614	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, vlan_head,
2615					  &vsi_list_head);
2616	mutex_unlock(vlan_lock);
2617	if (status)
2618		goto free_fltr_list;
2619
2620	list_for_each_entry(list_itr, &vsi_list_head, list_entry) {
2621		vlan_id = list_itr->fltr_info.l_data.vlan.vlan_id;
2622		if (rm_vlan_promisc)
2623			status = ice_clear_vsi_promisc(hw, vsi_handle,
2624						       promisc_mask, vlan_id);
2625		else
2626			status = ice_set_vsi_promisc(hw, vsi_handle,
2627						     promisc_mask, vlan_id);
2628		if (status)
2629			break;
2630	}
2631
2632free_fltr_list:
2633	list_for_each_entry_safe(list_itr, tmp, &vsi_list_head, list_entry) {
2634		list_del(&list_itr->list_entry);
2635		devm_kfree(ice_hw_to_dev(hw), list_itr);
2636	}
2637	return status;
2638}
2639
2640/**
2641 * ice_remove_vsi_lkup_fltr - Remove lookup type filters for a VSI
2642 * @hw: pointer to the hardware structure
2643 * @vsi_handle: VSI handle to remove filters from
2644 * @lkup: switch rule filter lookup type
2645 */
2646static void
2647ice_remove_vsi_lkup_fltr(struct ice_hw *hw, u16 vsi_handle,
2648			 enum ice_sw_lkup_type lkup)
2649{
2650	struct ice_switch_info *sw = hw->switch_info;
2651	struct ice_fltr_list_entry *fm_entry;
2652	struct list_head remove_list_head;
2653	struct list_head *rule_head;
2654	struct ice_fltr_list_entry *tmp;
2655	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2656	enum ice_status status;
2657
2658	INIT_LIST_HEAD(&remove_list_head);
2659	rule_lock = &sw->recp_list[lkup].filt_rule_lock;
2660	rule_head = &sw->recp_list[lkup].filt_rules;
2661	mutex_lock(rule_lock);
2662	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, rule_head,
2663					  &remove_list_head);
2664	mutex_unlock(rule_lock);
2665	if (status)
2666		return;
2667
2668	switch (lkup) {
2669	case ICE_SW_LKUP_MAC:
2670		ice_remove_mac(hw, &remove_list_head);
 
 
 
 
 
 
 
 
2671		break;
2672	case ICE_SW_LKUP_VLAN:
2673		ice_remove_vlan(hw, &remove_list_head);
2674		break;
2675	case ICE_SW_LKUP_PROMISC:
2676	case ICE_SW_LKUP_PROMISC_VLAN:
2677		ice_remove_promisc(hw, lkup, &remove_list_head);
 
 
 
 
2678		break;
2679	case ICE_SW_LKUP_MAC_VLAN:
2680	case ICE_SW_LKUP_ETHERTYPE:
2681	case ICE_SW_LKUP_ETHERTYPE_MAC:
 
 
2682	case ICE_SW_LKUP_DFLT:
2683	case ICE_SW_LKUP_LAST:
2684	default:
2685		ice_debug(hw, ICE_DBG_SW, "Unsupported lookup type %d\n", lkup);
2686		break;
2687	}
2688
 
 
2689	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2690		list_del(&fm_entry->list_entry);
2691		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2692	}
2693}
2694
2695/**
2696 * ice_remove_vsi_fltr - Remove all filters for a VSI
2697 * @hw: pointer to the hardware structure
2698 * @vsi_handle: VSI handle to remove filters from
2699 */
2700void ice_remove_vsi_fltr(struct ice_hw *hw, u16 vsi_handle)
2701{
2702	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC);
2703	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC_VLAN);
2704	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC);
2705	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_VLAN);
2706	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_DFLT);
2707	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE);
2708	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE_MAC);
2709	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC_VLAN);
2710}
2711
2712/**
2713 * ice_replay_vsi_fltr - Replay filters for requested VSI
2714 * @hw: pointer to the hardware structure
2715 * @vsi_handle: driver VSI handle
2716 * @recp_id: Recipe ID for which rules need to be replayed
2717 * @list_head: list for which filters need to be replayed
2718 *
2719 * Replays the filter of recipe recp_id for a VSI represented via vsi_handle.
2720 * It is required to pass valid VSI handle.
2721 */
2722static enum ice_status
2723ice_replay_vsi_fltr(struct ice_hw *hw, u16 vsi_handle, u8 recp_id,
2724		    struct list_head *list_head)
2725{
2726	struct ice_fltr_mgmt_list_entry *itr;
2727	enum ice_status status = 0;
2728	u16 hw_vsi_id;
2729
2730	if (list_empty(list_head))
2731		return status;
2732	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2733
2734	list_for_each_entry(itr, list_head, list_entry) {
2735		struct ice_fltr_list_entry f_entry;
2736
2737		f_entry.fltr_info = itr->fltr_info;
2738		if (itr->vsi_count < 2 && recp_id != ICE_SW_LKUP_VLAN &&
2739		    itr->fltr_info.vsi_handle == vsi_handle) {
2740			/* update the src in case it is VSI num */
2741			if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2742				f_entry.fltr_info.src = hw_vsi_id;
2743			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2744			if (status)
2745				goto end;
2746			continue;
2747		}
2748		if (!itr->vsi_list_info ||
2749		    !test_bit(vsi_handle, itr->vsi_list_info->vsi_map))
2750			continue;
2751		/* Clearing it so that the logic can add it back */
2752		clear_bit(vsi_handle, itr->vsi_list_info->vsi_map);
2753		f_entry.fltr_info.vsi_handle = vsi_handle;
2754		f_entry.fltr_info.fltr_act = ICE_FWD_TO_VSI;
2755		/* update the src in case it is VSI num */
2756		if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2757			f_entry.fltr_info.src = hw_vsi_id;
2758		if (recp_id == ICE_SW_LKUP_VLAN)
2759			status = ice_add_vlan_internal(hw, &f_entry);
2760		else
2761			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2762		if (status)
2763			goto end;
2764	}
2765end:
2766	return status;
2767}
2768
2769/**
2770 * ice_replay_vsi_all_fltr - replay all filters stored in bookkeeping lists
2771 * @hw: pointer to the hardware structure
2772 * @vsi_handle: driver VSI handle
2773 *
2774 * Replays filters for requested VSI via vsi_handle.
2775 */
2776enum ice_status ice_replay_vsi_all_fltr(struct ice_hw *hw, u16 vsi_handle)
2777{
2778	struct ice_switch_info *sw = hw->switch_info;
2779	enum ice_status status = 0;
2780	u8 i;
2781
2782	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2783		struct list_head *head;
2784
2785		head = &sw->recp_list[i].filt_replay_rules;
2786		status = ice_replay_vsi_fltr(hw, vsi_handle, i, head);
2787		if (status)
2788			return status;
2789	}
2790	return status;
2791}
2792
2793/**
2794 * ice_rm_all_sw_replay_rule_info - deletes filter replay rules
2795 * @hw: pointer to the HW struct
2796 *
2797 * Deletes the filter replay rules.
2798 */
2799void ice_rm_all_sw_replay_rule_info(struct ice_hw *hw)
2800{
2801	struct ice_switch_info *sw = hw->switch_info;
2802	u8 i;
2803
2804	if (!sw)
2805		return;
2806
2807	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2808		if (!list_empty(&sw->recp_list[i].filt_replay_rules)) {
2809			struct list_head *l_head;
2810
2811			l_head = &sw->recp_list[i].filt_replay_rules;
2812			ice_rem_sw_rule_info(hw, l_head);
2813		}
2814	}
2815}