Linux Audio

Check our new training course

Loading...
v5.4
   1// SPDX-License-Identifier: GPL-2.0
   2/* Copyright (c) 2018, Intel Corporation. */
   3
   4#include "ice_switch.h"
   5
   6#define ICE_ETH_DA_OFFSET		0
   7#define ICE_ETH_ETHTYPE_OFFSET		12
   8#define ICE_ETH_VLAN_TCI_OFFSET		14
   9#define ICE_MAX_VLAN_ID			0xFFF
  10
  11/* Dummy ethernet header needed in the ice_aqc_sw_rules_elem
  12 * struct to configure any switch filter rules.
  13 * {DA (6 bytes), SA(6 bytes),
  14 * Ether type (2 bytes for header without VLAN tag) OR
  15 * VLAN tag (4 bytes for header with VLAN tag) }
  16 *
  17 * Word on Hardcoded values
  18 * byte 0 = 0x2: to identify it as locally administered DA MAC
  19 * byte 6 = 0x2: to identify it as locally administered SA MAC
  20 * byte 12 = 0x81 & byte 13 = 0x00:
  21 *	In case of VLAN filter first two bytes defines ether type (0x8100)
  22 *	and remaining two bytes are placeholder for programming a given VLAN ID
  23 *	In case of Ether type filter it is treated as header without VLAN tag
  24 *	and byte 12 and 13 is used to program a given Ether type instead
  25 */
  26#define DUMMY_ETH_HDR_LEN		16
  27static const u8 dummy_eth_header[DUMMY_ETH_HDR_LEN] = { 0x2, 0, 0, 0, 0, 0,
  28							0x2, 0, 0, 0, 0, 0,
  29							0x81, 0, 0, 0};
  30
  31#define ICE_SW_RULE_RX_TX_ETH_HDR_SIZE \
  32	(sizeof(struct ice_aqc_sw_rules_elem) - \
  33	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  34	 sizeof(struct ice_sw_rule_lkup_rx_tx) + DUMMY_ETH_HDR_LEN - 1)
  35#define ICE_SW_RULE_RX_TX_NO_HDR_SIZE \
  36	(sizeof(struct ice_aqc_sw_rules_elem) - \
  37	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  38	 sizeof(struct ice_sw_rule_lkup_rx_tx) - 1)
  39#define ICE_SW_RULE_LG_ACT_SIZE(n) \
  40	(sizeof(struct ice_aqc_sw_rules_elem) - \
  41	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  42	 sizeof(struct ice_sw_rule_lg_act) - \
  43	 sizeof(((struct ice_sw_rule_lg_act *)0)->act) + \
  44	 ((n) * sizeof(((struct ice_sw_rule_lg_act *)0)->act)))
  45#define ICE_SW_RULE_VSI_LIST_SIZE(n) \
  46	(sizeof(struct ice_aqc_sw_rules_elem) - \
  47	 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
  48	 sizeof(struct ice_sw_rule_vsi_list) - \
  49	 sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi) + \
  50	 ((n) * sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi)))
  51
  52/**
  53 * ice_aq_alloc_free_res - command to allocate/free resources
  54 * @hw: pointer to the HW struct
  55 * @num_entries: number of resource entries in buffer
  56 * @buf: Indirect buffer to hold data parameters and response
  57 * @buf_size: size of buffer for indirect commands
  58 * @opc: pass in the command opcode
  59 * @cd: pointer to command details structure or NULL
  60 *
  61 * Helper function to allocate/free resources using the admin queue commands
  62 */
  63static enum ice_status
  64ice_aq_alloc_free_res(struct ice_hw *hw, u16 num_entries,
  65		      struct ice_aqc_alloc_free_res_elem *buf, u16 buf_size,
  66		      enum ice_adminq_opc opc, struct ice_sq_cd *cd)
  67{
  68	struct ice_aqc_alloc_free_res_cmd *cmd;
  69	struct ice_aq_desc desc;
  70
  71	cmd = &desc.params.sw_res_ctrl;
  72
  73	if (!buf)
  74		return ICE_ERR_PARAM;
  75
  76	if (buf_size < (num_entries * sizeof(buf->elem[0])))
  77		return ICE_ERR_PARAM;
  78
  79	ice_fill_dflt_direct_cmd_desc(&desc, opc);
  80
  81	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
  82
  83	cmd->num_entries = cpu_to_le16(num_entries);
  84
  85	return ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
  86}
  87
  88/**
  89 * ice_init_def_sw_recp - initialize the recipe book keeping tables
  90 * @hw: pointer to the HW struct
  91 *
  92 * Allocate memory for the entire recipe table and initialize the structures/
  93 * entries corresponding to basic recipes.
  94 */
  95enum ice_status ice_init_def_sw_recp(struct ice_hw *hw)
  96{
  97	struct ice_sw_recipe *recps;
  98	u8 i;
  99
 100	recps = devm_kcalloc(ice_hw_to_dev(hw), ICE_MAX_NUM_RECIPES,
 101			     sizeof(*recps), GFP_KERNEL);
 102	if (!recps)
 103		return ICE_ERR_NO_MEMORY;
 104
 105	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
 106		recps[i].root_rid = i;
 107		INIT_LIST_HEAD(&recps[i].filt_rules);
 108		INIT_LIST_HEAD(&recps[i].filt_replay_rules);
 109		mutex_init(&recps[i].filt_rule_lock);
 110	}
 111
 112	hw->switch_info->recp_list = recps;
 113
 114	return 0;
 115}
 116
 117/**
 118 * ice_aq_get_sw_cfg - get switch configuration
 119 * @hw: pointer to the hardware structure
 120 * @buf: pointer to the result buffer
 121 * @buf_size: length of the buffer available for response
 122 * @req_desc: pointer to requested descriptor
 123 * @num_elems: pointer to number of elements
 124 * @cd: pointer to command details structure or NULL
 125 *
 126 * Get switch configuration (0x0200) to be placed in 'buff'.
 127 * This admin command returns information such as initial VSI/port number
 128 * and switch ID it belongs to.
 129 *
 130 * NOTE: *req_desc is both an input/output parameter.
 131 * The caller of this function first calls this function with *request_desc set
 132 * to 0. If the response from f/w has *req_desc set to 0, all the switch
 133 * configuration information has been returned; if non-zero (meaning not all
 134 * the information was returned), the caller should call this function again
 135 * with *req_desc set to the previous value returned by f/w to get the
 136 * next block of switch configuration information.
 137 *
 138 * *num_elems is output only parameter. This reflects the number of elements
 139 * in response buffer. The caller of this function to use *num_elems while
 140 * parsing the response buffer.
 141 */
 142static enum ice_status
 143ice_aq_get_sw_cfg(struct ice_hw *hw, struct ice_aqc_get_sw_cfg_resp *buf,
 144		  u16 buf_size, u16 *req_desc, u16 *num_elems,
 145		  struct ice_sq_cd *cd)
 146{
 147	struct ice_aqc_get_sw_cfg *cmd;
 148	enum ice_status status;
 149	struct ice_aq_desc desc;
 
 150
 151	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_get_sw_cfg);
 152	cmd = &desc.params.get_sw_conf;
 153	cmd->element = cpu_to_le16(*req_desc);
 154
 155	status = ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
 156	if (!status) {
 157		*req_desc = le16_to_cpu(cmd->element);
 158		*num_elems = le16_to_cpu(cmd->num_elems);
 159	}
 160
 161	return status;
 162}
 163
 164/**
 165 * ice_aq_add_vsi
 166 * @hw: pointer to the HW struct
 167 * @vsi_ctx: pointer to a VSI context struct
 168 * @cd: pointer to command details structure or NULL
 169 *
 170 * Add a VSI context to the hardware (0x0210)
 171 */
 172static enum ice_status
 173ice_aq_add_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 174	       struct ice_sq_cd *cd)
 175{
 176	struct ice_aqc_add_update_free_vsi_resp *res;
 177	struct ice_aqc_add_get_update_free_vsi *cmd;
 178	struct ice_aq_desc desc;
 179	enum ice_status status;
 180
 181	cmd = &desc.params.vsi_cmd;
 182	res = &desc.params.add_update_free_vsi_res;
 183
 184	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_add_vsi);
 185
 186	if (!vsi_ctx->alloc_from_pool)
 187		cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num |
 188					   ICE_AQ_VSI_IS_VALID);
 189	cmd->vf_id = vsi_ctx->vf_num;
 190
 191	cmd->vsi_flags = cpu_to_le16(vsi_ctx->flags);
 192
 193	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 194
 195	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 196				 sizeof(vsi_ctx->info), cd);
 197
 198	if (!status) {
 199		vsi_ctx->vsi_num = le16_to_cpu(res->vsi_num) & ICE_AQ_VSI_NUM_M;
 200		vsi_ctx->vsis_allocd = le16_to_cpu(res->vsi_used);
 201		vsi_ctx->vsis_unallocated = le16_to_cpu(res->vsi_free);
 202	}
 203
 204	return status;
 205}
 206
 207/**
 208 * ice_aq_free_vsi
 209 * @hw: pointer to the HW struct
 210 * @vsi_ctx: pointer to a VSI context struct
 211 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 212 * @cd: pointer to command details structure or NULL
 213 *
 214 * Free VSI context info from hardware (0x0213)
 215 */
 216static enum ice_status
 217ice_aq_free_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 218		bool keep_vsi_alloc, struct ice_sq_cd *cd)
 219{
 220	struct ice_aqc_add_update_free_vsi_resp *resp;
 221	struct ice_aqc_add_get_update_free_vsi *cmd;
 222	struct ice_aq_desc desc;
 223	enum ice_status status;
 224
 225	cmd = &desc.params.vsi_cmd;
 226	resp = &desc.params.add_update_free_vsi_res;
 227
 228	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_free_vsi);
 229
 230	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 231	if (keep_vsi_alloc)
 232		cmd->cmd_flags = cpu_to_le16(ICE_AQ_VSI_KEEP_ALLOC);
 233
 234	status = ice_aq_send_cmd(hw, &desc, NULL, 0, cd);
 235	if (!status) {
 236		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 237		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 238	}
 239
 240	return status;
 241}
 242
 243/**
 244 * ice_aq_update_vsi
 245 * @hw: pointer to the HW struct
 246 * @vsi_ctx: pointer to a VSI context struct
 247 * @cd: pointer to command details structure or NULL
 248 *
 249 * Update VSI context in the hardware (0x0211)
 250 */
 251static enum ice_status
 252ice_aq_update_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 253		  struct ice_sq_cd *cd)
 254{
 255	struct ice_aqc_add_update_free_vsi_resp *resp;
 256	struct ice_aqc_add_get_update_free_vsi *cmd;
 257	struct ice_aq_desc desc;
 258	enum ice_status status;
 259
 260	cmd = &desc.params.vsi_cmd;
 261	resp = &desc.params.add_update_free_vsi_res;
 262
 263	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_update_vsi);
 264
 265	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 266
 267	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 268
 269	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 270				 sizeof(vsi_ctx->info), cd);
 271
 272	if (!status) {
 273		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 274		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 275	}
 276
 277	return status;
 278}
 279
 280/**
 281 * ice_is_vsi_valid - check whether the VSI is valid or not
 282 * @hw: pointer to the HW struct
 283 * @vsi_handle: VSI handle
 284 *
 285 * check whether the VSI is valid or not
 286 */
 287bool ice_is_vsi_valid(struct ice_hw *hw, u16 vsi_handle)
 288{
 289	return vsi_handle < ICE_MAX_VSI && hw->vsi_ctx[vsi_handle];
 290}
 291
 292/**
 293 * ice_get_hw_vsi_num - return the HW VSI number
 294 * @hw: pointer to the HW struct
 295 * @vsi_handle: VSI handle
 296 *
 297 * return the HW VSI number
 298 * Caution: call this function only if VSI is valid (ice_is_vsi_valid)
 299 */
 300u16 ice_get_hw_vsi_num(struct ice_hw *hw, u16 vsi_handle)
 301{
 302	return hw->vsi_ctx[vsi_handle]->vsi_num;
 303}
 304
 305/**
 306 * ice_get_vsi_ctx - return the VSI context entry for a given VSI handle
 307 * @hw: pointer to the HW struct
 308 * @vsi_handle: VSI handle
 309 *
 310 * return the VSI context entry for a given VSI handle
 311 */
 312struct ice_vsi_ctx *ice_get_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 313{
 314	return (vsi_handle >= ICE_MAX_VSI) ? NULL : hw->vsi_ctx[vsi_handle];
 315}
 316
 317/**
 318 * ice_save_vsi_ctx - save the VSI context for a given VSI handle
 319 * @hw: pointer to the HW struct
 320 * @vsi_handle: VSI handle
 321 * @vsi: VSI context pointer
 322 *
 323 * save the VSI context entry for a given VSI handle
 324 */
 325static void
 326ice_save_vsi_ctx(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi)
 327{
 328	hw->vsi_ctx[vsi_handle] = vsi;
 329}
 330
 331/**
 332 * ice_clear_vsi_q_ctx - clear VSI queue contexts for all TCs
 333 * @hw: pointer to the HW struct
 334 * @vsi_handle: VSI handle
 335 */
 336static void ice_clear_vsi_q_ctx(struct ice_hw *hw, u16 vsi_handle)
 337{
 338	struct ice_vsi_ctx *vsi;
 339	u8 i;
 340
 341	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 342	if (!vsi)
 343		return;
 344	ice_for_each_traffic_class(i) {
 345		if (vsi->lan_q_ctx[i]) {
 346			devm_kfree(ice_hw_to_dev(hw), vsi->lan_q_ctx[i]);
 347			vsi->lan_q_ctx[i] = NULL;
 348		}
 349	}
 350}
 351
 352/**
 353 * ice_clear_vsi_ctx - clear the VSI context entry
 354 * @hw: pointer to the HW struct
 355 * @vsi_handle: VSI handle
 356 *
 357 * clear the VSI context entry
 358 */
 359static void ice_clear_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 360{
 361	struct ice_vsi_ctx *vsi;
 362
 363	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 364	if (vsi) {
 365		ice_clear_vsi_q_ctx(hw, vsi_handle);
 366		devm_kfree(ice_hw_to_dev(hw), vsi);
 367		hw->vsi_ctx[vsi_handle] = NULL;
 368	}
 369}
 370
 371/**
 372 * ice_clear_all_vsi_ctx - clear all the VSI context entries
 373 * @hw: pointer to the HW struct
 374 */
 375void ice_clear_all_vsi_ctx(struct ice_hw *hw)
 376{
 377	u16 i;
 378
 379	for (i = 0; i < ICE_MAX_VSI; i++)
 380		ice_clear_vsi_ctx(hw, i);
 381}
 382
 383/**
 384 * ice_add_vsi - add VSI context to the hardware and VSI handle list
 385 * @hw: pointer to the HW struct
 386 * @vsi_handle: unique VSI handle provided by drivers
 387 * @vsi_ctx: pointer to a VSI context struct
 388 * @cd: pointer to command details structure or NULL
 389 *
 390 * Add a VSI context to the hardware also add it into the VSI handle list.
 391 * If this function gets called after reset for existing VSIs then update
 392 * with the new HW VSI number in the corresponding VSI handle list entry.
 393 */
 394enum ice_status
 395ice_add_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 396	    struct ice_sq_cd *cd)
 397{
 398	struct ice_vsi_ctx *tmp_vsi_ctx;
 399	enum ice_status status;
 400
 401	if (vsi_handle >= ICE_MAX_VSI)
 402		return ICE_ERR_PARAM;
 403	status = ice_aq_add_vsi(hw, vsi_ctx, cd);
 404	if (status)
 405		return status;
 406	tmp_vsi_ctx = ice_get_vsi_ctx(hw, vsi_handle);
 407	if (!tmp_vsi_ctx) {
 408		/* Create a new VSI context */
 409		tmp_vsi_ctx = devm_kzalloc(ice_hw_to_dev(hw),
 410					   sizeof(*tmp_vsi_ctx), GFP_KERNEL);
 411		if (!tmp_vsi_ctx) {
 412			ice_aq_free_vsi(hw, vsi_ctx, false, cd);
 413			return ICE_ERR_NO_MEMORY;
 414		}
 415		*tmp_vsi_ctx = *vsi_ctx;
 416		ice_save_vsi_ctx(hw, vsi_handle, tmp_vsi_ctx);
 417	} else {
 418		/* update with new HW VSI num */
 419		if (tmp_vsi_ctx->vsi_num != vsi_ctx->vsi_num)
 420			tmp_vsi_ctx->vsi_num = vsi_ctx->vsi_num;
 421	}
 422
 423	return 0;
 424}
 425
 426/**
 427 * ice_free_vsi- free VSI context from hardware and VSI handle list
 428 * @hw: pointer to the HW struct
 429 * @vsi_handle: unique VSI handle
 430 * @vsi_ctx: pointer to a VSI context struct
 431 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 432 * @cd: pointer to command details structure or NULL
 433 *
 434 * Free VSI context info from hardware as well as from VSI handle list
 435 */
 436enum ice_status
 437ice_free_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 438	     bool keep_vsi_alloc, struct ice_sq_cd *cd)
 439{
 440	enum ice_status status;
 441
 442	if (!ice_is_vsi_valid(hw, vsi_handle))
 443		return ICE_ERR_PARAM;
 444	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 445	status = ice_aq_free_vsi(hw, vsi_ctx, keep_vsi_alloc, cd);
 446	if (!status)
 447		ice_clear_vsi_ctx(hw, vsi_handle);
 448	return status;
 449}
 450
 451/**
 452 * ice_update_vsi
 453 * @hw: pointer to the HW struct
 454 * @vsi_handle: unique VSI handle
 455 * @vsi_ctx: pointer to a VSI context struct
 456 * @cd: pointer to command details structure or NULL
 457 *
 458 * Update VSI context in the hardware
 459 */
 460enum ice_status
 461ice_update_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 462	       struct ice_sq_cd *cd)
 463{
 464	if (!ice_is_vsi_valid(hw, vsi_handle))
 465		return ICE_ERR_PARAM;
 466	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 467	return ice_aq_update_vsi(hw, vsi_ctx, cd);
 468}
 469
 470/**
 471 * ice_aq_alloc_free_vsi_list
 472 * @hw: pointer to the HW struct
 473 * @vsi_list_id: VSI list ID returned or used for lookup
 474 * @lkup_type: switch rule filter lookup type
 475 * @opc: switch rules population command type - pass in the command opcode
 476 *
 477 * allocates or free a VSI list resource
 478 */
 479static enum ice_status
 480ice_aq_alloc_free_vsi_list(struct ice_hw *hw, u16 *vsi_list_id,
 481			   enum ice_sw_lkup_type lkup_type,
 482			   enum ice_adminq_opc opc)
 483{
 484	struct ice_aqc_alloc_free_res_elem *sw_buf;
 485	struct ice_aqc_res_elem *vsi_ele;
 486	enum ice_status status;
 487	u16 buf_len;
 488
 489	buf_len = sizeof(*sw_buf);
 490	sw_buf = devm_kzalloc(ice_hw_to_dev(hw), buf_len, GFP_KERNEL);
 491	if (!sw_buf)
 492		return ICE_ERR_NO_MEMORY;
 493	sw_buf->num_elems = cpu_to_le16(1);
 494
 495	if (lkup_type == ICE_SW_LKUP_MAC ||
 496	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 497	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 498	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 499	    lkup_type == ICE_SW_LKUP_PROMISC ||
 500	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN) {
 501		sw_buf->res_type = cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_REP);
 502	} else if (lkup_type == ICE_SW_LKUP_VLAN) {
 503		sw_buf->res_type =
 504			cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_PRUNE);
 505	} else {
 506		status = ICE_ERR_PARAM;
 507		goto ice_aq_alloc_free_vsi_list_exit;
 508	}
 509
 510	if (opc == ice_aqc_opc_free_res)
 511		sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
 512
 513	status = ice_aq_alloc_free_res(hw, 1, sw_buf, buf_len, opc, NULL);
 514	if (status)
 515		goto ice_aq_alloc_free_vsi_list_exit;
 516
 517	if (opc == ice_aqc_opc_alloc_res) {
 518		vsi_ele = &sw_buf->elem[0];
 519		*vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
 520	}
 521
 522ice_aq_alloc_free_vsi_list_exit:
 523	devm_kfree(ice_hw_to_dev(hw), sw_buf);
 524	return status;
 525}
 526
 527/**
 528 * ice_aq_sw_rules - add/update/remove switch rules
 529 * @hw: pointer to the HW struct
 530 * @rule_list: pointer to switch rule population list
 531 * @rule_list_sz: total size of the rule list in bytes
 532 * @num_rules: number of switch rules in the rule_list
 533 * @opc: switch rules population command type - pass in the command opcode
 534 * @cd: pointer to command details structure or NULL
 535 *
 536 * Add(0x02a0)/Update(0x02a1)/Remove(0x02a2) switch rules commands to firmware
 537 */
 538static enum ice_status
 539ice_aq_sw_rules(struct ice_hw *hw, void *rule_list, u16 rule_list_sz,
 540		u8 num_rules, enum ice_adminq_opc opc, struct ice_sq_cd *cd)
 541{
 542	struct ice_aq_desc desc;
 
 543
 544	if (opc != ice_aqc_opc_add_sw_rules &&
 545	    opc != ice_aqc_opc_update_sw_rules &&
 546	    opc != ice_aqc_opc_remove_sw_rules)
 547		return ICE_ERR_PARAM;
 548
 549	ice_fill_dflt_direct_cmd_desc(&desc, opc);
 550
 551	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 552	desc.params.sw_rules.num_rules_fltr_entry_index =
 553		cpu_to_le16(num_rules);
 554	return ice_aq_send_cmd(hw, &desc, rule_list, rule_list_sz, cd);
 
 
 
 
 
 555}
 556
 557/* ice_init_port_info - Initialize port_info with switch configuration data
 558 * @pi: pointer to port_info
 559 * @vsi_port_num: VSI number or port number
 560 * @type: Type of switch element (port or VSI)
 561 * @swid: switch ID of the switch the element is attached to
 562 * @pf_vf_num: PF or VF number
 563 * @is_vf: true if the element is a VF, false otherwise
 564 */
 565static void
 566ice_init_port_info(struct ice_port_info *pi, u16 vsi_port_num, u8 type,
 567		   u16 swid, u16 pf_vf_num, bool is_vf)
 568{
 569	switch (type) {
 570	case ICE_AQC_GET_SW_CONF_RESP_PHYS_PORT:
 571		pi->lport = (u8)(vsi_port_num & ICE_LPORT_MASK);
 572		pi->sw_id = swid;
 573		pi->pf_vf_num = pf_vf_num;
 574		pi->is_vf = is_vf;
 575		pi->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
 576		pi->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
 577		break;
 578	default:
 579		ice_debug(pi->hw, ICE_DBG_SW,
 580			  "incorrect VSI/port type received\n");
 581		break;
 582	}
 583}
 584
 585/* ice_get_initial_sw_cfg - Get initial port and default VSI data
 586 * @hw: pointer to the hardware structure
 587 */
 588enum ice_status ice_get_initial_sw_cfg(struct ice_hw *hw)
 589{
 590	struct ice_aqc_get_sw_cfg_resp *rbuf;
 591	enum ice_status status;
 592	u16 req_desc = 0;
 593	u16 num_elems;
 594	u16 i;
 595
 596	rbuf = devm_kzalloc(ice_hw_to_dev(hw), ICE_SW_CFG_MAX_BUF_LEN,
 597			    GFP_KERNEL);
 598
 599	if (!rbuf)
 600		return ICE_ERR_NO_MEMORY;
 601
 602	/* Multiple calls to ice_aq_get_sw_cfg may be required
 603	 * to get all the switch configuration information. The need
 604	 * for additional calls is indicated by ice_aq_get_sw_cfg
 605	 * writing a non-zero value in req_desc
 606	 */
 607	do {
 
 
 608		status = ice_aq_get_sw_cfg(hw, rbuf, ICE_SW_CFG_MAX_BUF_LEN,
 609					   &req_desc, &num_elems, NULL);
 610
 611		if (status)
 612			break;
 613
 614		for (i = 0; i < num_elems; i++) {
 615			struct ice_aqc_get_sw_cfg_resp_elem *ele;
 616			u16 pf_vf_num, swid, vsi_port_num;
 617			bool is_vf = false;
 618			u8 type;
 619
 620			ele = rbuf[i].elements;
 621			vsi_port_num = le16_to_cpu(ele->vsi_port_num) &
 622				ICE_AQC_GET_SW_CONF_RESP_VSI_PORT_NUM_M;
 623
 624			pf_vf_num = le16_to_cpu(ele->pf_vf_num) &
 625				ICE_AQC_GET_SW_CONF_RESP_FUNC_NUM_M;
 626
 627			swid = le16_to_cpu(ele->swid);
 628
 629			if (le16_to_cpu(ele->pf_vf_num) &
 630			    ICE_AQC_GET_SW_CONF_RESP_IS_VF)
 631				is_vf = true;
 632
 633			type = le16_to_cpu(ele->vsi_port_num) >>
 634				ICE_AQC_GET_SW_CONF_RESP_TYPE_S;
 635
 636			if (type == ICE_AQC_GET_SW_CONF_RESP_VSI) {
 637				/* FW VSI is not needed. Just continue. */
 638				continue;
 639			}
 640
 641			ice_init_port_info(hw->port_info, vsi_port_num,
 642					   type, swid, pf_vf_num, is_vf);
 643		}
 644	} while (req_desc && !status);
 645
 646	devm_kfree(ice_hw_to_dev(hw), (void *)rbuf);
 647	return status;
 648}
 649
 650/**
 651 * ice_fill_sw_info - Helper function to populate lb_en and lan_en
 652 * @hw: pointer to the hardware structure
 653 * @fi: filter info structure to fill/update
 654 *
 655 * This helper function populates the lb_en and lan_en elements of the provided
 656 * ice_fltr_info struct using the switch's type and characteristics of the
 657 * switch rule being configured.
 658 */
 659static void ice_fill_sw_info(struct ice_hw *hw, struct ice_fltr_info *fi)
 660{
 661	fi->lb_en = false;
 662	fi->lan_en = false;
 663	if ((fi->flag & ICE_FLTR_TX) &&
 664	    (fi->fltr_act == ICE_FWD_TO_VSI ||
 665	     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
 666	     fi->fltr_act == ICE_FWD_TO_Q ||
 667	     fi->fltr_act == ICE_FWD_TO_QGRP)) {
 668		/* Setting LB for prune actions will result in replicated
 669		 * packets to the internal switch that will be dropped.
 670		 */
 671		if (fi->lkup_type != ICE_SW_LKUP_VLAN)
 672			fi->lb_en = true;
 673
 674		/* Set lan_en to TRUE if
 675		 * 1. The switch is a VEB AND
 676		 * 2
 677		 * 2.1 The lookup is a directional lookup like ethertype,
 678		 * promiscuous, ethertype-MAC, promiscuous-VLAN
 679		 * and default-port OR
 680		 * 2.2 The lookup is VLAN, OR
 681		 * 2.3 The lookup is MAC with mcast or bcast addr for MAC, OR
 682		 * 2.4 The lookup is MAC_VLAN with mcast or bcast addr for MAC.
 683		 *
 684		 * OR
 685		 *
 686		 * The switch is a VEPA.
 687		 *
 688		 * In all other cases, the LAN enable has to be set to false.
 689		 */
 690		if (hw->evb_veb) {
 691			if (fi->lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 692			    fi->lkup_type == ICE_SW_LKUP_PROMISC ||
 693			    fi->lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 694			    fi->lkup_type == ICE_SW_LKUP_PROMISC_VLAN ||
 695			    fi->lkup_type == ICE_SW_LKUP_DFLT ||
 696			    fi->lkup_type == ICE_SW_LKUP_VLAN ||
 697			    (fi->lkup_type == ICE_SW_LKUP_MAC &&
 698			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)) ||
 699			    (fi->lkup_type == ICE_SW_LKUP_MAC_VLAN &&
 700			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)))
 701				fi->lan_en = true;
 702		} else {
 703			fi->lan_en = true;
 704		}
 705	}
 706}
 707
 708/**
 709 * ice_fill_sw_rule - Helper function to fill switch rule structure
 710 * @hw: pointer to the hardware structure
 711 * @f_info: entry containing packet forwarding information
 712 * @s_rule: switch rule structure to be filled in based on mac_entry
 713 * @opc: switch rules population command type - pass in the command opcode
 714 */
 715static void
 716ice_fill_sw_rule(struct ice_hw *hw, struct ice_fltr_info *f_info,
 717		 struct ice_aqc_sw_rules_elem *s_rule, enum ice_adminq_opc opc)
 718{
 719	u16 vlan_id = ICE_MAX_VLAN_ID + 1;
 720	void *daddr = NULL;
 721	u16 eth_hdr_sz;
 722	u8 *eth_hdr;
 723	u32 act = 0;
 724	__be16 *off;
 725	u8 q_rgn;
 726
 727	if (opc == ice_aqc_opc_remove_sw_rules) {
 728		s_rule->pdata.lkup_tx_rx.act = 0;
 729		s_rule->pdata.lkup_tx_rx.index =
 730			cpu_to_le16(f_info->fltr_rule_id);
 731		s_rule->pdata.lkup_tx_rx.hdr_len = 0;
 732		return;
 733	}
 734
 735	eth_hdr_sz = sizeof(dummy_eth_header);
 736	eth_hdr = s_rule->pdata.lkup_tx_rx.hdr;
 737
 738	/* initialize the ether header with a dummy header */
 739	memcpy(eth_hdr, dummy_eth_header, eth_hdr_sz);
 740	ice_fill_sw_info(hw, f_info);
 741
 742	switch (f_info->fltr_act) {
 743	case ICE_FWD_TO_VSI:
 744		act |= (f_info->fwd_id.hw_vsi_id << ICE_SINGLE_ACT_VSI_ID_S) &
 745			ICE_SINGLE_ACT_VSI_ID_M;
 746		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 747			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 748				ICE_SINGLE_ACT_VALID_BIT;
 749		break;
 750	case ICE_FWD_TO_VSI_LIST:
 751		act |= ICE_SINGLE_ACT_VSI_LIST;
 752		act |= (f_info->fwd_id.vsi_list_id <<
 753			ICE_SINGLE_ACT_VSI_LIST_ID_S) &
 754			ICE_SINGLE_ACT_VSI_LIST_ID_M;
 755		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 756			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 757				ICE_SINGLE_ACT_VALID_BIT;
 758		break;
 759	case ICE_FWD_TO_Q:
 760		act |= ICE_SINGLE_ACT_TO_Q;
 761		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 762			ICE_SINGLE_ACT_Q_INDEX_M;
 763		break;
 764	case ICE_DROP_PACKET:
 765		act |= ICE_SINGLE_ACT_VSI_FORWARDING | ICE_SINGLE_ACT_DROP |
 766			ICE_SINGLE_ACT_VALID_BIT;
 767		break;
 768	case ICE_FWD_TO_QGRP:
 769		q_rgn = f_info->qgrp_size > 0 ?
 770			(u8)ilog2(f_info->qgrp_size) : 0;
 771		act |= ICE_SINGLE_ACT_TO_Q;
 772		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 773			ICE_SINGLE_ACT_Q_INDEX_M;
 774		act |= (q_rgn << ICE_SINGLE_ACT_Q_REGION_S) &
 775			ICE_SINGLE_ACT_Q_REGION_M;
 776		break;
 777	default:
 778		return;
 779	}
 780
 781	if (f_info->lb_en)
 782		act |= ICE_SINGLE_ACT_LB_ENABLE;
 783	if (f_info->lan_en)
 784		act |= ICE_SINGLE_ACT_LAN_ENABLE;
 785
 786	switch (f_info->lkup_type) {
 787	case ICE_SW_LKUP_MAC:
 788		daddr = f_info->l_data.mac.mac_addr;
 789		break;
 790	case ICE_SW_LKUP_VLAN:
 791		vlan_id = f_info->l_data.vlan.vlan_id;
 792		if (f_info->fltr_act == ICE_FWD_TO_VSI ||
 793		    f_info->fltr_act == ICE_FWD_TO_VSI_LIST) {
 794			act |= ICE_SINGLE_ACT_PRUNE;
 795			act |= ICE_SINGLE_ACT_EGRESS | ICE_SINGLE_ACT_INGRESS;
 796		}
 797		break;
 798	case ICE_SW_LKUP_ETHERTYPE_MAC:
 799		daddr = f_info->l_data.ethertype_mac.mac_addr;
 800		/* fall-through */
 801	case ICE_SW_LKUP_ETHERTYPE:
 802		off = (__force __be16 *)(eth_hdr + ICE_ETH_ETHTYPE_OFFSET);
 803		*off = cpu_to_be16(f_info->l_data.ethertype_mac.ethertype);
 804		break;
 805	case ICE_SW_LKUP_MAC_VLAN:
 806		daddr = f_info->l_data.mac_vlan.mac_addr;
 807		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 808		break;
 809	case ICE_SW_LKUP_PROMISC_VLAN:
 810		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 811		/* fall-through */
 812	case ICE_SW_LKUP_PROMISC:
 813		daddr = f_info->l_data.mac_vlan.mac_addr;
 814		break;
 815	default:
 816		break;
 817	}
 818
 819	s_rule->type = (f_info->flag & ICE_FLTR_RX) ?
 820		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_RX) :
 821		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_TX);
 822
 823	/* Recipe set depending on lookup type */
 824	s_rule->pdata.lkup_tx_rx.recipe_id = cpu_to_le16(f_info->lkup_type);
 825	s_rule->pdata.lkup_tx_rx.src = cpu_to_le16(f_info->src);
 826	s_rule->pdata.lkup_tx_rx.act = cpu_to_le32(act);
 827
 828	if (daddr)
 829		ether_addr_copy(eth_hdr + ICE_ETH_DA_OFFSET, daddr);
 830
 831	if (!(vlan_id > ICE_MAX_VLAN_ID)) {
 832		off = (__force __be16 *)(eth_hdr + ICE_ETH_VLAN_TCI_OFFSET);
 833		*off = cpu_to_be16(vlan_id);
 834	}
 835
 836	/* Create the switch rule with the final dummy Ethernet header */
 837	if (opc != ice_aqc_opc_update_sw_rules)
 838		s_rule->pdata.lkup_tx_rx.hdr_len = cpu_to_le16(eth_hdr_sz);
 839}
 840
 841/**
 842 * ice_add_marker_act
 843 * @hw: pointer to the hardware structure
 844 * @m_ent: the management entry for which sw marker needs to be added
 845 * @sw_marker: sw marker to tag the Rx descriptor with
 846 * @l_id: large action resource ID
 847 *
 848 * Create a large action to hold software marker and update the switch rule
 849 * entry pointed by m_ent with newly created large action
 850 */
 851static enum ice_status
 852ice_add_marker_act(struct ice_hw *hw, struct ice_fltr_mgmt_list_entry *m_ent,
 853		   u16 sw_marker, u16 l_id)
 854{
 855	struct ice_aqc_sw_rules_elem *lg_act, *rx_tx;
 856	/* For software marker we need 3 large actions
 857	 * 1. FWD action: FWD TO VSI or VSI LIST
 858	 * 2. GENERIC VALUE action to hold the profile ID
 859	 * 3. GENERIC VALUE action to hold the software marker ID
 860	 */
 861	const u16 num_lg_acts = 3;
 862	enum ice_status status;
 863	u16 lg_act_size;
 864	u16 rules_size;
 865	u32 act;
 866	u16 id;
 867
 868	if (m_ent->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
 869		return ICE_ERR_PARAM;
 870
 871	/* Create two back-to-back switch rules and submit them to the HW using
 872	 * one memory buffer:
 873	 *    1. Large Action
 874	 *    2. Look up Tx Rx
 875	 */
 876	lg_act_size = (u16)ICE_SW_RULE_LG_ACT_SIZE(num_lg_acts);
 877	rules_size = lg_act_size + ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
 878	lg_act = devm_kzalloc(ice_hw_to_dev(hw), rules_size, GFP_KERNEL);
 879	if (!lg_act)
 880		return ICE_ERR_NO_MEMORY;
 881
 882	rx_tx = (struct ice_aqc_sw_rules_elem *)((u8 *)lg_act + lg_act_size);
 883
 884	/* Fill in the first switch rule i.e. large action */
 885	lg_act->type = cpu_to_le16(ICE_AQC_SW_RULES_T_LG_ACT);
 886	lg_act->pdata.lg_act.index = cpu_to_le16(l_id);
 887	lg_act->pdata.lg_act.size = cpu_to_le16(num_lg_acts);
 888
 889	/* First action VSI forwarding or VSI list forwarding depending on how
 890	 * many VSIs
 891	 */
 892	id = (m_ent->vsi_count > 1) ? m_ent->fltr_info.fwd_id.vsi_list_id :
 893		m_ent->fltr_info.fwd_id.hw_vsi_id;
 894
 895	act = ICE_LG_ACT_VSI_FORWARDING | ICE_LG_ACT_VALID_BIT;
 896	act |= (id << ICE_LG_ACT_VSI_LIST_ID_S) &
 897		ICE_LG_ACT_VSI_LIST_ID_M;
 898	if (m_ent->vsi_count > 1)
 899		act |= ICE_LG_ACT_VSI_LIST;
 900	lg_act->pdata.lg_act.act[0] = cpu_to_le32(act);
 901
 902	/* Second action descriptor type */
 903	act = ICE_LG_ACT_GENERIC;
 904
 905	act |= (1 << ICE_LG_ACT_GENERIC_VALUE_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 906	lg_act->pdata.lg_act.act[1] = cpu_to_le32(act);
 907
 908	act = (ICE_LG_ACT_GENERIC_OFF_RX_DESC_PROF_IDX <<
 909	       ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_OFFSET_M;
 910
 911	/* Third action Marker value */
 912	act |= ICE_LG_ACT_GENERIC;
 913	act |= (sw_marker << ICE_LG_ACT_GENERIC_VALUE_S) &
 914		ICE_LG_ACT_GENERIC_VALUE_M;
 915
 916	lg_act->pdata.lg_act.act[2] = cpu_to_le32(act);
 917
 918	/* call the fill switch rule to fill the lookup Tx Rx structure */
 919	ice_fill_sw_rule(hw, &m_ent->fltr_info, rx_tx,
 920			 ice_aqc_opc_update_sw_rules);
 921
 922	/* Update the action to point to the large action ID */
 923	rx_tx->pdata.lkup_tx_rx.act =
 924		cpu_to_le32(ICE_SINGLE_ACT_PTR |
 925			    ((l_id << ICE_SINGLE_ACT_PTR_VAL_S) &
 926			     ICE_SINGLE_ACT_PTR_VAL_M));
 927
 928	/* Use the filter rule ID of the previously created rule with single
 929	 * act. Once the update happens, hardware will treat this as large
 930	 * action
 931	 */
 932	rx_tx->pdata.lkup_tx_rx.index =
 933		cpu_to_le16(m_ent->fltr_info.fltr_rule_id);
 934
 935	status = ice_aq_sw_rules(hw, lg_act, rules_size, 2,
 936				 ice_aqc_opc_update_sw_rules, NULL);
 937	if (!status) {
 938		m_ent->lg_act_idx = l_id;
 939		m_ent->sw_marker_id = sw_marker;
 940	}
 941
 942	devm_kfree(ice_hw_to_dev(hw), lg_act);
 943	return status;
 944}
 945
 946/**
 947 * ice_create_vsi_list_map
 948 * @hw: pointer to the hardware structure
 949 * @vsi_handle_arr: array of VSI handles to set in the VSI mapping
 950 * @num_vsi: number of VSI handles in the array
 951 * @vsi_list_id: VSI list ID generated as part of allocate resource
 952 *
 953 * Helper function to create a new entry of VSI list ID to VSI mapping
 954 * using the given VSI list ID
 955 */
 956static struct ice_vsi_list_map_info *
 957ice_create_vsi_list_map(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 958			u16 vsi_list_id)
 959{
 960	struct ice_switch_info *sw = hw->switch_info;
 961	struct ice_vsi_list_map_info *v_map;
 962	int i;
 963
 964	v_map = devm_kcalloc(ice_hw_to_dev(hw), 1, sizeof(*v_map), GFP_KERNEL);
 965	if (!v_map)
 966		return NULL;
 967
 968	v_map->vsi_list_id = vsi_list_id;
 969	v_map->ref_cnt = 1;
 970	for (i = 0; i < num_vsi; i++)
 971		set_bit(vsi_handle_arr[i], v_map->vsi_map);
 972
 973	list_add(&v_map->list_entry, &sw->vsi_list_map_head);
 974	return v_map;
 975}
 976
 977/**
 978 * ice_update_vsi_list_rule
 979 * @hw: pointer to the hardware structure
 980 * @vsi_handle_arr: array of VSI handles to form a VSI list
 981 * @num_vsi: number of VSI handles in the array
 982 * @vsi_list_id: VSI list ID generated as part of allocate resource
 983 * @remove: Boolean value to indicate if this is a remove action
 984 * @opc: switch rules population command type - pass in the command opcode
 985 * @lkup_type: lookup type of the filter
 986 *
 987 * Call AQ command to add a new switch rule or update existing switch rule
 988 * using the given VSI list ID
 989 */
 990static enum ice_status
 991ice_update_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 992			 u16 vsi_list_id, bool remove, enum ice_adminq_opc opc,
 993			 enum ice_sw_lkup_type lkup_type)
 994{
 995	struct ice_aqc_sw_rules_elem *s_rule;
 996	enum ice_status status;
 997	u16 s_rule_size;
 998	u16 type;
 999	int i;
1000
1001	if (!num_vsi)
1002		return ICE_ERR_PARAM;
1003
1004	if (lkup_type == ICE_SW_LKUP_MAC ||
1005	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
1006	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
1007	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
1008	    lkup_type == ICE_SW_LKUP_PROMISC ||
1009	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN)
1010		type = remove ? ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR :
1011				ICE_AQC_SW_RULES_T_VSI_LIST_SET;
1012	else if (lkup_type == ICE_SW_LKUP_VLAN)
1013		type = remove ? ICE_AQC_SW_RULES_T_PRUNE_LIST_CLEAR :
1014				ICE_AQC_SW_RULES_T_PRUNE_LIST_SET;
1015	else
1016		return ICE_ERR_PARAM;
1017
1018	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(num_vsi);
1019	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1020	if (!s_rule)
1021		return ICE_ERR_NO_MEMORY;
1022	for (i = 0; i < num_vsi; i++) {
1023		if (!ice_is_vsi_valid(hw, vsi_handle_arr[i])) {
1024			status = ICE_ERR_PARAM;
1025			goto exit;
1026		}
1027		/* AQ call requires hw_vsi_id(s) */
1028		s_rule->pdata.vsi_list.vsi[i] =
1029			cpu_to_le16(ice_get_hw_vsi_num(hw, vsi_handle_arr[i]));
1030	}
1031
1032	s_rule->type = cpu_to_le16(type);
1033	s_rule->pdata.vsi_list.number_vsi = cpu_to_le16(num_vsi);
1034	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1035
1036	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opc, NULL);
1037
1038exit:
1039	devm_kfree(ice_hw_to_dev(hw), s_rule);
1040	return status;
1041}
1042
1043/**
1044 * ice_create_vsi_list_rule - Creates and populates a VSI list rule
1045 * @hw: pointer to the HW struct
1046 * @vsi_handle_arr: array of VSI handles to form a VSI list
1047 * @num_vsi: number of VSI handles in the array
1048 * @vsi_list_id: stores the ID of the VSI list to be created
1049 * @lkup_type: switch rule filter's lookup type
1050 */
1051static enum ice_status
1052ice_create_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
1053			 u16 *vsi_list_id, enum ice_sw_lkup_type lkup_type)
1054{
1055	enum ice_status status;
1056
1057	status = ice_aq_alloc_free_vsi_list(hw, vsi_list_id, lkup_type,
1058					    ice_aqc_opc_alloc_res);
1059	if (status)
1060		return status;
1061
1062	/* Update the newly created VSI list to include the specified VSIs */
1063	return ice_update_vsi_list_rule(hw, vsi_handle_arr, num_vsi,
1064					*vsi_list_id, false,
1065					ice_aqc_opc_add_sw_rules, lkup_type);
1066}
1067
1068/**
1069 * ice_create_pkt_fwd_rule
1070 * @hw: pointer to the hardware structure
1071 * @f_entry: entry containing packet forwarding information
1072 *
1073 * Create switch rule with given filter information and add an entry
1074 * to the corresponding filter management list to track this switch rule
1075 * and VSI mapping
1076 */
1077static enum ice_status
1078ice_create_pkt_fwd_rule(struct ice_hw *hw,
1079			struct ice_fltr_list_entry *f_entry)
1080{
1081	struct ice_fltr_mgmt_list_entry *fm_entry;
1082	struct ice_aqc_sw_rules_elem *s_rule;
1083	enum ice_sw_lkup_type l_type;
1084	struct ice_sw_recipe *recp;
1085	enum ice_status status;
1086
1087	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1088			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1089	if (!s_rule)
1090		return ICE_ERR_NO_MEMORY;
1091	fm_entry = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*fm_entry),
1092				GFP_KERNEL);
1093	if (!fm_entry) {
1094		status = ICE_ERR_NO_MEMORY;
1095		goto ice_create_pkt_fwd_rule_exit;
1096	}
1097
1098	fm_entry->fltr_info = f_entry->fltr_info;
1099
1100	/* Initialize all the fields for the management entry */
1101	fm_entry->vsi_count = 1;
1102	fm_entry->lg_act_idx = ICE_INVAL_LG_ACT_INDEX;
1103	fm_entry->sw_marker_id = ICE_INVAL_SW_MARKER_ID;
1104	fm_entry->counter_index = ICE_INVAL_COUNTER_ID;
1105
1106	ice_fill_sw_rule(hw, &fm_entry->fltr_info, s_rule,
1107			 ice_aqc_opc_add_sw_rules);
1108
1109	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1110				 ice_aqc_opc_add_sw_rules, NULL);
1111	if (status) {
1112		devm_kfree(ice_hw_to_dev(hw), fm_entry);
1113		goto ice_create_pkt_fwd_rule_exit;
1114	}
1115
1116	f_entry->fltr_info.fltr_rule_id =
1117		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1118	fm_entry->fltr_info.fltr_rule_id =
1119		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1120
1121	/* The book keeping entries will get removed when base driver
1122	 * calls remove filter AQ command
1123	 */
1124	l_type = fm_entry->fltr_info.lkup_type;
1125	recp = &hw->switch_info->recp_list[l_type];
1126	list_add(&fm_entry->list_entry, &recp->filt_rules);
1127
1128ice_create_pkt_fwd_rule_exit:
1129	devm_kfree(ice_hw_to_dev(hw), s_rule);
1130	return status;
1131}
1132
1133/**
1134 * ice_update_pkt_fwd_rule
1135 * @hw: pointer to the hardware structure
1136 * @f_info: filter information for switch rule
1137 *
1138 * Call AQ command to update a previously created switch rule with a
1139 * VSI list ID
1140 */
1141static enum ice_status
1142ice_update_pkt_fwd_rule(struct ice_hw *hw, struct ice_fltr_info *f_info)
1143{
1144	struct ice_aqc_sw_rules_elem *s_rule;
1145	enum ice_status status;
1146
1147	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1148			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1149	if (!s_rule)
1150		return ICE_ERR_NO_MEMORY;
1151
1152	ice_fill_sw_rule(hw, f_info, s_rule, ice_aqc_opc_update_sw_rules);
1153
1154	s_rule->pdata.lkup_tx_rx.index = cpu_to_le16(f_info->fltr_rule_id);
1155
1156	/* Update switch rule with new rule set to forward VSI list */
1157	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1158				 ice_aqc_opc_update_sw_rules, NULL);
1159
1160	devm_kfree(ice_hw_to_dev(hw), s_rule);
1161	return status;
1162}
1163
1164/**
1165 * ice_update_sw_rule_bridge_mode
1166 * @hw: pointer to the HW struct
1167 *
1168 * Updates unicast switch filter rules based on VEB/VEPA mode
1169 */
1170enum ice_status ice_update_sw_rule_bridge_mode(struct ice_hw *hw)
1171{
1172	struct ice_switch_info *sw = hw->switch_info;
1173	struct ice_fltr_mgmt_list_entry *fm_entry;
1174	enum ice_status status = 0;
1175	struct list_head *rule_head;
1176	struct mutex *rule_lock; /* Lock to protect filter rule list */
1177
1178	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1179	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1180
1181	mutex_lock(rule_lock);
1182	list_for_each_entry(fm_entry, rule_head, list_entry) {
1183		struct ice_fltr_info *fi = &fm_entry->fltr_info;
1184		u8 *addr = fi->l_data.mac.mac_addr;
1185
1186		/* Update unicast Tx rules to reflect the selected
1187		 * VEB/VEPA mode
1188		 */
1189		if ((fi->flag & ICE_FLTR_TX) && is_unicast_ether_addr(addr) &&
1190		    (fi->fltr_act == ICE_FWD_TO_VSI ||
1191		     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
1192		     fi->fltr_act == ICE_FWD_TO_Q ||
1193		     fi->fltr_act == ICE_FWD_TO_QGRP)) {
1194			status = ice_update_pkt_fwd_rule(hw, fi);
1195			if (status)
1196				break;
1197		}
1198	}
1199
1200	mutex_unlock(rule_lock);
1201
1202	return status;
1203}
1204
1205/**
1206 * ice_add_update_vsi_list
1207 * @hw: pointer to the hardware structure
1208 * @m_entry: pointer to current filter management list entry
1209 * @cur_fltr: filter information from the book keeping entry
1210 * @new_fltr: filter information with the new VSI to be added
1211 *
1212 * Call AQ command to add or update previously created VSI list with new VSI.
1213 *
1214 * Helper function to do book keeping associated with adding filter information
1215 * The algorithm to do the book keeping is described below :
1216 * When a VSI needs to subscribe to a given filter (MAC/VLAN/Ethtype etc.)
1217 *	if only one VSI has been added till now
1218 *		Allocate a new VSI list and add two VSIs
1219 *		to this list using switch rule command
1220 *		Update the previously created switch rule with the
1221 *		newly created VSI list ID
1222 *	if a VSI list was previously created
1223 *		Add the new VSI to the previously created VSI list set
1224 *		using the update switch rule command
1225 */
1226static enum ice_status
1227ice_add_update_vsi_list(struct ice_hw *hw,
1228			struct ice_fltr_mgmt_list_entry *m_entry,
1229			struct ice_fltr_info *cur_fltr,
1230			struct ice_fltr_info *new_fltr)
1231{
1232	enum ice_status status = 0;
1233	u16 vsi_list_id = 0;
1234
1235	if ((cur_fltr->fltr_act == ICE_FWD_TO_Q ||
1236	     cur_fltr->fltr_act == ICE_FWD_TO_QGRP))
1237		return ICE_ERR_NOT_IMPL;
1238
1239	if ((new_fltr->fltr_act == ICE_FWD_TO_Q ||
1240	     new_fltr->fltr_act == ICE_FWD_TO_QGRP) &&
1241	    (cur_fltr->fltr_act == ICE_FWD_TO_VSI ||
1242	     cur_fltr->fltr_act == ICE_FWD_TO_VSI_LIST))
1243		return ICE_ERR_NOT_IMPL;
1244
1245	if (m_entry->vsi_count < 2 && !m_entry->vsi_list_info) {
1246		/* Only one entry existed in the mapping and it was not already
1247		 * a part of a VSI list. So, create a VSI list with the old and
1248		 * new VSIs.
1249		 */
1250		struct ice_fltr_info tmp_fltr;
1251		u16 vsi_handle_arr[2];
1252
1253		/* A rule already exists with the new VSI being added */
1254		if (cur_fltr->fwd_id.hw_vsi_id == new_fltr->fwd_id.hw_vsi_id)
1255			return ICE_ERR_ALREADY_EXISTS;
1256
1257		vsi_handle_arr[0] = cur_fltr->vsi_handle;
1258		vsi_handle_arr[1] = new_fltr->vsi_handle;
1259		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1260						  &vsi_list_id,
1261						  new_fltr->lkup_type);
1262		if (status)
1263			return status;
1264
1265		tmp_fltr = *new_fltr;
1266		tmp_fltr.fltr_rule_id = cur_fltr->fltr_rule_id;
1267		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1268		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1269		/* Update the previous switch rule of "MAC forward to VSI" to
1270		 * "MAC fwd to VSI list"
1271		 */
1272		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
1273		if (status)
1274			return status;
1275
1276		cur_fltr->fwd_id.vsi_list_id = vsi_list_id;
1277		cur_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1278		m_entry->vsi_list_info =
1279			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1280						vsi_list_id);
1281
1282		/* If this entry was large action then the large action needs
1283		 * to be updated to point to FWD to VSI list
1284		 */
1285		if (m_entry->sw_marker_id != ICE_INVAL_SW_MARKER_ID)
1286			status =
1287			    ice_add_marker_act(hw, m_entry,
1288					       m_entry->sw_marker_id,
1289					       m_entry->lg_act_idx);
1290	} else {
1291		u16 vsi_handle = new_fltr->vsi_handle;
1292		enum ice_adminq_opc opcode;
1293
1294		if (!m_entry->vsi_list_info)
1295			return ICE_ERR_CFG;
1296
1297		/* A rule already exists with the new VSI being added */
1298		if (test_bit(vsi_handle, m_entry->vsi_list_info->vsi_map))
1299			return 0;
1300
1301		/* Update the previously created VSI list set with
1302		 * the new VSI ID passed in
1303		 */
1304		vsi_list_id = cur_fltr->fwd_id.vsi_list_id;
1305		opcode = ice_aqc_opc_update_sw_rules;
1306
1307		status = ice_update_vsi_list_rule(hw, &vsi_handle, 1,
1308						  vsi_list_id, false, opcode,
1309						  new_fltr->lkup_type);
1310		/* update VSI list mapping info with new VSI ID */
1311		if (!status)
1312			set_bit(vsi_handle, m_entry->vsi_list_info->vsi_map);
1313	}
1314	if (!status)
1315		m_entry->vsi_count++;
1316	return status;
1317}
1318
1319/**
1320 * ice_find_rule_entry - Search a rule entry
1321 * @hw: pointer to the hardware structure
1322 * @recp_id: lookup type for which the specified rule needs to be searched
1323 * @f_info: rule information
1324 *
1325 * Helper function to search for a given rule entry
1326 * Returns pointer to entry storing the rule if found
1327 */
1328static struct ice_fltr_mgmt_list_entry *
1329ice_find_rule_entry(struct ice_hw *hw, u8 recp_id, struct ice_fltr_info *f_info)
1330{
1331	struct ice_fltr_mgmt_list_entry *list_itr, *ret = NULL;
1332	struct ice_switch_info *sw = hw->switch_info;
1333	struct list_head *list_head;
1334
1335	list_head = &sw->recp_list[recp_id].filt_rules;
1336	list_for_each_entry(list_itr, list_head, list_entry) {
1337		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
1338			    sizeof(f_info->l_data)) &&
1339		    f_info->flag == list_itr->fltr_info.flag) {
1340			ret = list_itr;
1341			break;
1342		}
1343	}
1344	return ret;
1345}
1346
1347/**
1348 * ice_find_vsi_list_entry - Search VSI list map with VSI count 1
1349 * @hw: pointer to the hardware structure
1350 * @recp_id: lookup type for which VSI lists needs to be searched
1351 * @vsi_handle: VSI handle to be found in VSI list
1352 * @vsi_list_id: VSI list ID found containing vsi_handle
1353 *
1354 * Helper function to search a VSI list with single entry containing given VSI
1355 * handle element. This can be extended further to search VSI list with more
1356 * than 1 vsi_count. Returns pointer to VSI list entry if found.
1357 */
1358static struct ice_vsi_list_map_info *
1359ice_find_vsi_list_entry(struct ice_hw *hw, u8 recp_id, u16 vsi_handle,
1360			u16 *vsi_list_id)
1361{
1362	struct ice_vsi_list_map_info *map_info = NULL;
1363	struct ice_switch_info *sw = hw->switch_info;
1364	struct ice_fltr_mgmt_list_entry *list_itr;
1365	struct list_head *list_head;
1366
1367	list_head = &sw->recp_list[recp_id].filt_rules;
1368	list_for_each_entry(list_itr, list_head, list_entry) {
1369		if (list_itr->vsi_count == 1 && list_itr->vsi_list_info) {
1370			map_info = list_itr->vsi_list_info;
1371			if (test_bit(vsi_handle, map_info->vsi_map)) {
1372				*vsi_list_id = map_info->vsi_list_id;
1373				return map_info;
1374			}
1375		}
1376	}
1377	return NULL;
1378}
1379
1380/**
1381 * ice_add_rule_internal - add rule for a given lookup type
1382 * @hw: pointer to the hardware structure
1383 * @recp_id: lookup type (recipe ID) for which rule has to be added
1384 * @f_entry: structure containing MAC forwarding information
1385 *
1386 * Adds or updates the rule lists for a given recipe
1387 */
1388static enum ice_status
1389ice_add_rule_internal(struct ice_hw *hw, u8 recp_id,
1390		      struct ice_fltr_list_entry *f_entry)
1391{
1392	struct ice_switch_info *sw = hw->switch_info;
1393	struct ice_fltr_info *new_fltr, *cur_fltr;
1394	struct ice_fltr_mgmt_list_entry *m_entry;
1395	struct mutex *rule_lock; /* Lock to protect filter rule list */
1396	enum ice_status status = 0;
1397
1398	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1399		return ICE_ERR_PARAM;
1400	f_entry->fltr_info.fwd_id.hw_vsi_id =
1401		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1402
1403	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1404
1405	mutex_lock(rule_lock);
1406	new_fltr = &f_entry->fltr_info;
1407	if (new_fltr->flag & ICE_FLTR_RX)
1408		new_fltr->src = hw->port_info->lport;
1409	else if (new_fltr->flag & ICE_FLTR_TX)
1410		new_fltr->src = f_entry->fltr_info.fwd_id.hw_vsi_id;
1411
1412	m_entry = ice_find_rule_entry(hw, recp_id, new_fltr);
1413	if (!m_entry) {
1414		mutex_unlock(rule_lock);
1415		return ice_create_pkt_fwd_rule(hw, f_entry);
1416	}
1417
1418	cur_fltr = &m_entry->fltr_info;
1419	status = ice_add_update_vsi_list(hw, m_entry, cur_fltr, new_fltr);
1420	mutex_unlock(rule_lock);
1421
1422	return status;
1423}
1424
1425/**
1426 * ice_remove_vsi_list_rule
1427 * @hw: pointer to the hardware structure
1428 * @vsi_list_id: VSI list ID generated as part of allocate resource
1429 * @lkup_type: switch rule filter lookup type
1430 *
1431 * The VSI list should be emptied before this function is called to remove the
1432 * VSI list.
1433 */
1434static enum ice_status
1435ice_remove_vsi_list_rule(struct ice_hw *hw, u16 vsi_list_id,
1436			 enum ice_sw_lkup_type lkup_type)
1437{
1438	struct ice_aqc_sw_rules_elem *s_rule;
1439	enum ice_status status;
1440	u16 s_rule_size;
1441
1442	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(0);
1443	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1444	if (!s_rule)
1445		return ICE_ERR_NO_MEMORY;
1446
1447	s_rule->type = cpu_to_le16(ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR);
1448	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1449
1450	/* Free the vsi_list resource that we allocated. It is assumed that the
1451	 * list is empty at this point.
1452	 */
1453	status = ice_aq_alloc_free_vsi_list(hw, &vsi_list_id, lkup_type,
1454					    ice_aqc_opc_free_res);
1455
1456	devm_kfree(ice_hw_to_dev(hw), s_rule);
1457	return status;
1458}
1459
1460/**
1461 * ice_rem_update_vsi_list
1462 * @hw: pointer to the hardware structure
1463 * @vsi_handle: VSI handle of the VSI to remove
1464 * @fm_list: filter management entry for which the VSI list management needs to
1465 *           be done
1466 */
1467static enum ice_status
1468ice_rem_update_vsi_list(struct ice_hw *hw, u16 vsi_handle,
1469			struct ice_fltr_mgmt_list_entry *fm_list)
1470{
1471	enum ice_sw_lkup_type lkup_type;
1472	enum ice_status status = 0;
1473	u16 vsi_list_id;
1474
1475	if (fm_list->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST ||
1476	    fm_list->vsi_count == 0)
1477		return ICE_ERR_PARAM;
1478
1479	/* A rule with the VSI being removed does not exist */
1480	if (!test_bit(vsi_handle, fm_list->vsi_list_info->vsi_map))
1481		return ICE_ERR_DOES_NOT_EXIST;
1482
1483	lkup_type = fm_list->fltr_info.lkup_type;
1484	vsi_list_id = fm_list->fltr_info.fwd_id.vsi_list_id;
1485	status = ice_update_vsi_list_rule(hw, &vsi_handle, 1, vsi_list_id, true,
1486					  ice_aqc_opc_update_sw_rules,
1487					  lkup_type);
1488	if (status)
1489		return status;
1490
1491	fm_list->vsi_count--;
1492	clear_bit(vsi_handle, fm_list->vsi_list_info->vsi_map);
1493
1494	if (fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) {
1495		struct ice_fltr_info tmp_fltr_info = fm_list->fltr_info;
1496		struct ice_vsi_list_map_info *vsi_list_info =
1497			fm_list->vsi_list_info;
1498		u16 rem_vsi_handle;
1499
1500		rem_vsi_handle = find_first_bit(vsi_list_info->vsi_map,
1501						ICE_MAX_VSI);
1502		if (!ice_is_vsi_valid(hw, rem_vsi_handle))
1503			return ICE_ERR_OUT_OF_RANGE;
1504
1505		/* Make sure VSI list is empty before removing it below */
1506		status = ice_update_vsi_list_rule(hw, &rem_vsi_handle, 1,
1507						  vsi_list_id, true,
1508						  ice_aqc_opc_update_sw_rules,
1509						  lkup_type);
1510		if (status)
1511			return status;
1512
1513		tmp_fltr_info.fltr_act = ICE_FWD_TO_VSI;
1514		tmp_fltr_info.fwd_id.hw_vsi_id =
1515			ice_get_hw_vsi_num(hw, rem_vsi_handle);
1516		tmp_fltr_info.vsi_handle = rem_vsi_handle;
1517		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr_info);
1518		if (status) {
1519			ice_debug(hw, ICE_DBG_SW,
1520				  "Failed to update pkt fwd rule to FWD_TO_VSI on HW VSI %d, error %d\n",
1521				  tmp_fltr_info.fwd_id.hw_vsi_id, status);
1522			return status;
1523		}
1524
1525		fm_list->fltr_info = tmp_fltr_info;
1526	}
1527
1528	if ((fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) ||
1529	    (fm_list->vsi_count == 0 && lkup_type == ICE_SW_LKUP_VLAN)) {
1530		struct ice_vsi_list_map_info *vsi_list_info =
1531			fm_list->vsi_list_info;
1532
1533		/* Remove the VSI list since it is no longer used */
1534		status = ice_remove_vsi_list_rule(hw, vsi_list_id, lkup_type);
1535		if (status) {
1536			ice_debug(hw, ICE_DBG_SW,
1537				  "Failed to remove VSI list %d, error %d\n",
1538				  vsi_list_id, status);
1539			return status;
1540		}
1541
1542		list_del(&vsi_list_info->list_entry);
1543		devm_kfree(ice_hw_to_dev(hw), vsi_list_info);
1544		fm_list->vsi_list_info = NULL;
1545	}
1546
1547	return status;
1548}
1549
1550/**
1551 * ice_remove_rule_internal - Remove a filter rule of a given type
1552 * @hw: pointer to the hardware structure
1553 * @recp_id: recipe ID for which the rule needs to removed
1554 * @f_entry: rule entry containing filter information
1555 */
1556static enum ice_status
1557ice_remove_rule_internal(struct ice_hw *hw, u8 recp_id,
1558			 struct ice_fltr_list_entry *f_entry)
1559{
1560	struct ice_switch_info *sw = hw->switch_info;
1561	struct ice_fltr_mgmt_list_entry *list_elem;
1562	struct mutex *rule_lock; /* Lock to protect filter rule list */
1563	enum ice_status status = 0;
1564	bool remove_rule = false;
1565	u16 vsi_handle;
1566
1567	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1568		return ICE_ERR_PARAM;
1569	f_entry->fltr_info.fwd_id.hw_vsi_id =
1570		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1571
1572	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1573	mutex_lock(rule_lock);
1574	list_elem = ice_find_rule_entry(hw, recp_id, &f_entry->fltr_info);
1575	if (!list_elem) {
1576		status = ICE_ERR_DOES_NOT_EXIST;
1577		goto exit;
1578	}
1579
1580	if (list_elem->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST) {
1581		remove_rule = true;
1582	} else if (!list_elem->vsi_list_info) {
1583		status = ICE_ERR_DOES_NOT_EXIST;
1584		goto exit;
1585	} else if (list_elem->vsi_list_info->ref_cnt > 1) {
1586		/* a ref_cnt > 1 indicates that the vsi_list is being
1587		 * shared by multiple rules. Decrement the ref_cnt and
1588		 * remove this rule, but do not modify the list, as it
1589		 * is in-use by other rules.
1590		 */
1591		list_elem->vsi_list_info->ref_cnt--;
1592		remove_rule = true;
1593	} else {
1594		/* a ref_cnt of 1 indicates the vsi_list is only used
1595		 * by one rule. However, the original removal request is only
1596		 * for a single VSI. Update the vsi_list first, and only
1597		 * remove the rule if there are no further VSIs in this list.
1598		 */
1599		vsi_handle = f_entry->fltr_info.vsi_handle;
1600		status = ice_rem_update_vsi_list(hw, vsi_handle, list_elem);
1601		if (status)
1602			goto exit;
1603		/* if VSI count goes to zero after updating the VSI list */
1604		if (list_elem->vsi_count == 0)
1605			remove_rule = true;
1606	}
1607
1608	if (remove_rule) {
1609		/* Remove the lookup rule */
1610		struct ice_aqc_sw_rules_elem *s_rule;
1611
1612		s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1613				      ICE_SW_RULE_RX_TX_NO_HDR_SIZE,
1614				      GFP_KERNEL);
1615		if (!s_rule) {
1616			status = ICE_ERR_NO_MEMORY;
1617			goto exit;
1618		}
1619
1620		ice_fill_sw_rule(hw, &list_elem->fltr_info, s_rule,
1621				 ice_aqc_opc_remove_sw_rules);
1622
1623		status = ice_aq_sw_rules(hw, s_rule,
1624					 ICE_SW_RULE_RX_TX_NO_HDR_SIZE, 1,
1625					 ice_aqc_opc_remove_sw_rules, NULL);
1626
1627		/* Remove a book keeping from the list */
1628		devm_kfree(ice_hw_to_dev(hw), s_rule);
1629
1630		if (status)
1631			goto exit;
1632
1633		list_del(&list_elem->list_entry);
1634		devm_kfree(ice_hw_to_dev(hw), list_elem);
1635	}
1636exit:
1637	mutex_unlock(rule_lock);
1638	return status;
1639}
1640
1641/**
1642 * ice_add_mac - Add a MAC address based filter rule
1643 * @hw: pointer to the hardware structure
1644 * @m_list: list of MAC addresses and forwarding information
1645 *
1646 * IMPORTANT: When the ucast_shared flag is set to false and m_list has
1647 * multiple unicast addresses, the function assumes that all the
1648 * addresses are unique in a given add_mac call. It doesn't
1649 * check for duplicates in this case, removing duplicates from a given
1650 * list should be taken care of in the caller of this function.
1651 */
1652enum ice_status
1653ice_add_mac(struct ice_hw *hw, struct list_head *m_list)
1654{
1655	struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1656	struct ice_fltr_list_entry *m_list_itr;
1657	struct list_head *rule_head;
1658	u16 elem_sent, total_elem_left;
1659	struct ice_switch_info *sw;
1660	struct mutex *rule_lock; /* Lock to protect filter rule list */
1661	enum ice_status status = 0;
1662	u16 num_unicast = 0;
1663	u16 s_rule_size;
1664
1665	if (!m_list || !hw)
1666		return ICE_ERR_PARAM;
1667
1668	s_rule = NULL;
1669	sw = hw->switch_info;
1670	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1671	list_for_each_entry(m_list_itr, m_list, list_entry) {
1672		u8 *add = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1673		u16 vsi_handle;
1674		u16 hw_vsi_id;
1675
1676		m_list_itr->fltr_info.flag = ICE_FLTR_TX;
1677		vsi_handle = m_list_itr->fltr_info.vsi_handle;
1678		if (!ice_is_vsi_valid(hw, vsi_handle))
1679			return ICE_ERR_PARAM;
1680		hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
1681		m_list_itr->fltr_info.fwd_id.hw_vsi_id = hw_vsi_id;
1682		/* update the src in case it is VSI num */
1683		if (m_list_itr->fltr_info.src_id != ICE_SRC_ID_VSI)
1684			return ICE_ERR_PARAM;
1685		m_list_itr->fltr_info.src = hw_vsi_id;
1686		if (m_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_MAC ||
1687		    is_zero_ether_addr(add))
1688			return ICE_ERR_PARAM;
1689		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
1690			/* Don't overwrite the unicast address */
1691			mutex_lock(rule_lock);
1692			if (ice_find_rule_entry(hw, ICE_SW_LKUP_MAC,
1693						&m_list_itr->fltr_info)) {
1694				mutex_unlock(rule_lock);
1695				return ICE_ERR_ALREADY_EXISTS;
1696			}
1697			mutex_unlock(rule_lock);
1698			num_unicast++;
1699		} else if (is_multicast_ether_addr(add) ||
1700			   (is_unicast_ether_addr(add) && hw->ucast_shared)) {
1701			m_list_itr->status =
1702				ice_add_rule_internal(hw, ICE_SW_LKUP_MAC,
1703						      m_list_itr);
1704			if (m_list_itr->status)
1705				return m_list_itr->status;
1706		}
1707	}
1708
1709	mutex_lock(rule_lock);
1710	/* Exit if no suitable entries were found for adding bulk switch rule */
1711	if (!num_unicast) {
1712		status = 0;
1713		goto ice_add_mac_exit;
1714	}
1715
1716	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1717
1718	/* Allocate switch rule buffer for the bulk update for unicast */
1719	s_rule_size = ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
1720	s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1721			      GFP_KERNEL);
1722	if (!s_rule) {
1723		status = ICE_ERR_NO_MEMORY;
1724		goto ice_add_mac_exit;
1725	}
1726
1727	r_iter = s_rule;
1728	list_for_each_entry(m_list_itr, m_list, list_entry) {
1729		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1730		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
1731
1732		if (is_unicast_ether_addr(mac_addr)) {
1733			ice_fill_sw_rule(hw, &m_list_itr->fltr_info, r_iter,
1734					 ice_aqc_opc_add_sw_rules);
1735			r_iter = (struct ice_aqc_sw_rules_elem *)
1736				((u8 *)r_iter + s_rule_size);
1737		}
1738	}
1739
1740	/* Call AQ bulk switch rule update for all unicast addresses */
1741	r_iter = s_rule;
1742	/* Call AQ switch rule in AQ_MAX chunk */
1743	for (total_elem_left = num_unicast; total_elem_left > 0;
1744	     total_elem_left -= elem_sent) {
1745		struct ice_aqc_sw_rules_elem *entry = r_iter;
1746
1747		elem_sent = min(total_elem_left,
1748				(u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1749		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1750					 elem_sent, ice_aqc_opc_add_sw_rules,
1751					 NULL);
1752		if (status)
1753			goto ice_add_mac_exit;
1754		r_iter = (struct ice_aqc_sw_rules_elem *)
1755			((u8 *)r_iter + (elem_sent * s_rule_size));
1756	}
1757
1758	/* Fill up rule ID based on the value returned from FW */
1759	r_iter = s_rule;
1760	list_for_each_entry(m_list_itr, m_list, list_entry) {
1761		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1762		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
1763		struct ice_fltr_mgmt_list_entry *fm_entry;
1764
1765		if (is_unicast_ether_addr(mac_addr)) {
1766			f_info->fltr_rule_id =
1767				le16_to_cpu(r_iter->pdata.lkup_tx_rx.index);
1768			f_info->fltr_act = ICE_FWD_TO_VSI;
1769			/* Create an entry to track this MAC address */
1770			fm_entry = devm_kzalloc(ice_hw_to_dev(hw),
1771						sizeof(*fm_entry), GFP_KERNEL);
1772			if (!fm_entry) {
1773				status = ICE_ERR_NO_MEMORY;
1774				goto ice_add_mac_exit;
1775			}
1776			fm_entry->fltr_info = *f_info;
1777			fm_entry->vsi_count = 1;
1778			/* The book keeping entries will get removed when
1779			 * base driver calls remove filter AQ command
1780			 */
1781
1782			list_add(&fm_entry->list_entry, rule_head);
1783			r_iter = (struct ice_aqc_sw_rules_elem *)
1784				((u8 *)r_iter + s_rule_size);
1785		}
1786	}
1787
1788ice_add_mac_exit:
1789	mutex_unlock(rule_lock);
1790	if (s_rule)
1791		devm_kfree(ice_hw_to_dev(hw), s_rule);
1792	return status;
1793}
1794
1795/**
1796 * ice_add_vlan_internal - Add one VLAN based filter rule
1797 * @hw: pointer to the hardware structure
1798 * @f_entry: filter entry containing one VLAN information
1799 */
1800static enum ice_status
1801ice_add_vlan_internal(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1802{
1803	struct ice_switch_info *sw = hw->switch_info;
1804	struct ice_fltr_mgmt_list_entry *v_list_itr;
1805	struct ice_fltr_info *new_fltr, *cur_fltr;
1806	enum ice_sw_lkup_type lkup_type;
1807	u16 vsi_list_id = 0, vsi_handle;
1808	struct mutex *rule_lock; /* Lock to protect filter rule list */
1809	enum ice_status status = 0;
1810
1811	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1812		return ICE_ERR_PARAM;
1813
1814	f_entry->fltr_info.fwd_id.hw_vsi_id =
1815		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1816	new_fltr = &f_entry->fltr_info;
1817
1818	/* VLAN ID should only be 12 bits */
1819	if (new_fltr->l_data.vlan.vlan_id > ICE_MAX_VLAN_ID)
1820		return ICE_ERR_PARAM;
1821
1822	if (new_fltr->src_id != ICE_SRC_ID_VSI)
1823		return ICE_ERR_PARAM;
1824
1825	new_fltr->src = new_fltr->fwd_id.hw_vsi_id;
1826	lkup_type = new_fltr->lkup_type;
1827	vsi_handle = new_fltr->vsi_handle;
1828	rule_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
1829	mutex_lock(rule_lock);
1830	v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN, new_fltr);
1831	if (!v_list_itr) {
1832		struct ice_vsi_list_map_info *map_info = NULL;
1833
1834		if (new_fltr->fltr_act == ICE_FWD_TO_VSI) {
1835			/* All VLAN pruning rules use a VSI list. Check if
1836			 * there is already a VSI list containing VSI that we
1837			 * want to add. If found, use the same vsi_list_id for
1838			 * this new VLAN rule or else create a new list.
1839			 */
1840			map_info = ice_find_vsi_list_entry(hw, ICE_SW_LKUP_VLAN,
1841							   vsi_handle,
1842							   &vsi_list_id);
1843			if (!map_info) {
1844				status = ice_create_vsi_list_rule(hw,
1845								  &vsi_handle,
1846								  1,
1847								  &vsi_list_id,
1848								  lkup_type);
1849				if (status)
1850					goto exit;
1851			}
1852			/* Convert the action to forwarding to a VSI list. */
1853			new_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1854			new_fltr->fwd_id.vsi_list_id = vsi_list_id;
1855		}
1856
1857		status = ice_create_pkt_fwd_rule(hw, f_entry);
1858		if (!status) {
1859			v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN,
1860							 new_fltr);
1861			if (!v_list_itr) {
1862				status = ICE_ERR_DOES_NOT_EXIST;
1863				goto exit;
1864			}
1865			/* reuse VSI list for new rule and increment ref_cnt */
1866			if (map_info) {
1867				v_list_itr->vsi_list_info = map_info;
1868				map_info->ref_cnt++;
1869			} else {
1870				v_list_itr->vsi_list_info =
1871					ice_create_vsi_list_map(hw, &vsi_handle,
1872								1, vsi_list_id);
1873			}
1874		}
1875	} else if (v_list_itr->vsi_list_info->ref_cnt == 1) {
1876		/* Update existing VSI list to add new VSI ID only if it used
1877		 * by one VLAN rule.
1878		 */
1879		cur_fltr = &v_list_itr->fltr_info;
1880		status = ice_add_update_vsi_list(hw, v_list_itr, cur_fltr,
1881						 new_fltr);
1882	} else {
1883		/* If VLAN rule exists and VSI list being used by this rule is
1884		 * referenced by more than 1 VLAN rule. Then create a new VSI
1885		 * list appending previous VSI with new VSI and update existing
1886		 * VLAN rule to point to new VSI list ID
1887		 */
1888		struct ice_fltr_info tmp_fltr;
1889		u16 vsi_handle_arr[2];
1890		u16 cur_handle;
1891
1892		/* Current implementation only supports reusing VSI list with
1893		 * one VSI count. We should never hit below condition
1894		 */
1895		if (v_list_itr->vsi_count > 1 &&
1896		    v_list_itr->vsi_list_info->ref_cnt > 1) {
1897			ice_debug(hw, ICE_DBG_SW,
1898				  "Invalid configuration: Optimization to reuse VSI list with more than one VSI is not being done yet\n");
1899			status = ICE_ERR_CFG;
1900			goto exit;
1901		}
1902
1903		cur_handle =
1904			find_first_bit(v_list_itr->vsi_list_info->vsi_map,
1905				       ICE_MAX_VSI);
1906
1907		/* A rule already exists with the new VSI being added */
1908		if (cur_handle == vsi_handle) {
1909			status = ICE_ERR_ALREADY_EXISTS;
1910			goto exit;
1911		}
1912
1913		vsi_handle_arr[0] = cur_handle;
1914		vsi_handle_arr[1] = vsi_handle;
1915		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1916						  &vsi_list_id, lkup_type);
1917		if (status)
1918			goto exit;
1919
1920		tmp_fltr = v_list_itr->fltr_info;
1921		tmp_fltr.fltr_rule_id = v_list_itr->fltr_info.fltr_rule_id;
1922		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1923		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1924		/* Update the previous switch rule to a new VSI list which
1925		 * includes current VSI that is requested
1926		 */
1927		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
1928		if (status)
1929			goto exit;
1930
1931		/* before overriding VSI list map info. decrement ref_cnt of
1932		 * previous VSI list
1933		 */
1934		v_list_itr->vsi_list_info->ref_cnt--;
1935
1936		/* now update to newly created list */
1937		v_list_itr->fltr_info.fwd_id.vsi_list_id = vsi_list_id;
1938		v_list_itr->vsi_list_info =
1939			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1940						vsi_list_id);
1941		v_list_itr->vsi_count++;
1942	}
1943
1944exit:
1945	mutex_unlock(rule_lock);
1946	return status;
1947}
1948
1949/**
1950 * ice_add_vlan - Add VLAN based filter rule
1951 * @hw: pointer to the hardware structure
1952 * @v_list: list of VLAN entries and forwarding information
1953 */
1954enum ice_status
1955ice_add_vlan(struct ice_hw *hw, struct list_head *v_list)
1956{
1957	struct ice_fltr_list_entry *v_list_itr;
1958
1959	if (!v_list || !hw)
1960		return ICE_ERR_PARAM;
1961
1962	list_for_each_entry(v_list_itr, v_list, list_entry) {
1963		if (v_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_VLAN)
1964			return ICE_ERR_PARAM;
1965		v_list_itr->fltr_info.flag = ICE_FLTR_TX;
1966		v_list_itr->status = ice_add_vlan_internal(hw, v_list_itr);
1967		if (v_list_itr->status)
1968			return v_list_itr->status;
1969	}
1970	return 0;
1971}
1972
1973/**
1974 * ice_add_eth_mac - Add ethertype and MAC based filter rule
1975 * @hw: pointer to the hardware structure
1976 * @em_list: list of ether type MAC filter, MAC is optional
1977 *
1978 * This function requires the caller to populate the entries in
1979 * the filter list with the necessary fields (including flags to
1980 * indicate Tx or Rx rules).
1981 */
1982enum ice_status
1983ice_add_eth_mac(struct ice_hw *hw, struct list_head *em_list)
1984{
1985	struct ice_fltr_list_entry *em_list_itr;
1986
1987	if (!em_list || !hw)
1988		return ICE_ERR_PARAM;
1989
1990	list_for_each_entry(em_list_itr, em_list, list_entry) {
1991		enum ice_sw_lkup_type l_type =
1992			em_list_itr->fltr_info.lkup_type;
1993
1994		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
1995		    l_type != ICE_SW_LKUP_ETHERTYPE)
1996			return ICE_ERR_PARAM;
1997
1998		em_list_itr->status = ice_add_rule_internal(hw, l_type,
1999							    em_list_itr);
2000		if (em_list_itr->status)
2001			return em_list_itr->status;
2002	}
2003	return 0;
2004}
2005
2006/**
2007 * ice_remove_eth_mac - Remove an ethertype (or MAC) based filter rule
2008 * @hw: pointer to the hardware structure
2009 * @em_list: list of ethertype or ethertype MAC entries
2010 */
2011enum ice_status
2012ice_remove_eth_mac(struct ice_hw *hw, struct list_head *em_list)
2013{
2014	struct ice_fltr_list_entry *em_list_itr, *tmp;
2015
2016	if (!em_list || !hw)
2017		return ICE_ERR_PARAM;
2018
2019	list_for_each_entry_safe(em_list_itr, tmp, em_list, list_entry) {
2020		enum ice_sw_lkup_type l_type =
2021			em_list_itr->fltr_info.lkup_type;
2022
2023		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
2024		    l_type != ICE_SW_LKUP_ETHERTYPE)
2025			return ICE_ERR_PARAM;
2026
2027		em_list_itr->status = ice_remove_rule_internal(hw, l_type,
2028							       em_list_itr);
2029		if (em_list_itr->status)
2030			return em_list_itr->status;
2031	}
2032	return 0;
2033}
2034
2035/**
2036 * ice_rem_sw_rule_info
2037 * @hw: pointer to the hardware structure
2038 * @rule_head: pointer to the switch list structure that we want to delete
2039 */
2040static void
2041ice_rem_sw_rule_info(struct ice_hw *hw, struct list_head *rule_head)
2042{
2043	if (!list_empty(rule_head)) {
2044		struct ice_fltr_mgmt_list_entry *entry;
2045		struct ice_fltr_mgmt_list_entry *tmp;
2046
2047		list_for_each_entry_safe(entry, tmp, rule_head, list_entry) {
2048			list_del(&entry->list_entry);
2049			devm_kfree(ice_hw_to_dev(hw), entry);
2050		}
2051	}
2052}
2053
2054/**
2055 * ice_cfg_dflt_vsi - change state of VSI to set/clear default
2056 * @hw: pointer to the hardware structure
2057 * @vsi_handle: VSI handle to set as default
2058 * @set: true to add the above mentioned switch rule, false to remove it
2059 * @direction: ICE_FLTR_RX or ICE_FLTR_TX
2060 *
2061 * add filter rule to set/unset given VSI as default VSI for the switch
2062 * (represented by swid)
2063 */
2064enum ice_status
2065ice_cfg_dflt_vsi(struct ice_hw *hw, u16 vsi_handle, bool set, u8 direction)
2066{
2067	struct ice_aqc_sw_rules_elem *s_rule;
2068	struct ice_fltr_info f_info;
2069	enum ice_adminq_opc opcode;
2070	enum ice_status status;
2071	u16 s_rule_size;
2072	u16 hw_vsi_id;
2073
2074	if (!ice_is_vsi_valid(hw, vsi_handle))
2075		return ICE_ERR_PARAM;
2076	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2077
2078	s_rule_size = set ? ICE_SW_RULE_RX_TX_ETH_HDR_SIZE :
2079			    ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
 
2080	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
2081	if (!s_rule)
2082		return ICE_ERR_NO_MEMORY;
2083
2084	memset(&f_info, 0, sizeof(f_info));
2085
2086	f_info.lkup_type = ICE_SW_LKUP_DFLT;
2087	f_info.flag = direction;
2088	f_info.fltr_act = ICE_FWD_TO_VSI;
2089	f_info.fwd_id.hw_vsi_id = hw_vsi_id;
2090
2091	if (f_info.flag & ICE_FLTR_RX) {
2092		f_info.src = hw->port_info->lport;
2093		f_info.src_id = ICE_SRC_ID_LPORT;
2094		if (!set)
2095			f_info.fltr_rule_id =
2096				hw->port_info->dflt_rx_vsi_rule_id;
2097	} else if (f_info.flag & ICE_FLTR_TX) {
2098		f_info.src_id = ICE_SRC_ID_VSI;
2099		f_info.src = hw_vsi_id;
2100		if (!set)
2101			f_info.fltr_rule_id =
2102				hw->port_info->dflt_tx_vsi_rule_id;
2103	}
2104
2105	if (set)
2106		opcode = ice_aqc_opc_add_sw_rules;
2107	else
2108		opcode = ice_aqc_opc_remove_sw_rules;
2109
2110	ice_fill_sw_rule(hw, &f_info, s_rule, opcode);
2111
2112	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opcode, NULL);
2113	if (status || !(f_info.flag & ICE_FLTR_TX_RX))
2114		goto out;
2115	if (set) {
2116		u16 index = le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
2117
2118		if (f_info.flag & ICE_FLTR_TX) {
2119			hw->port_info->dflt_tx_vsi_num = hw_vsi_id;
2120			hw->port_info->dflt_tx_vsi_rule_id = index;
2121		} else if (f_info.flag & ICE_FLTR_RX) {
2122			hw->port_info->dflt_rx_vsi_num = hw_vsi_id;
2123			hw->port_info->dflt_rx_vsi_rule_id = index;
2124		}
2125	} else {
2126		if (f_info.flag & ICE_FLTR_TX) {
2127			hw->port_info->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
2128			hw->port_info->dflt_tx_vsi_rule_id = ICE_INVAL_ACT;
2129		} else if (f_info.flag & ICE_FLTR_RX) {
2130			hw->port_info->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
2131			hw->port_info->dflt_rx_vsi_rule_id = ICE_INVAL_ACT;
2132		}
2133	}
2134
2135out:
2136	devm_kfree(ice_hw_to_dev(hw), s_rule);
2137	return status;
2138}
2139
2140/**
2141 * ice_find_ucast_rule_entry - Search for a unicast MAC filter rule entry
2142 * @hw: pointer to the hardware structure
2143 * @recp_id: lookup type for which the specified rule needs to be searched
2144 * @f_info: rule information
2145 *
2146 * Helper function to search for a unicast rule entry - this is to be used
2147 * to remove unicast MAC filter that is not shared with other VSIs on the
2148 * PF switch.
2149 *
2150 * Returns pointer to entry storing the rule if found
2151 */
2152static struct ice_fltr_mgmt_list_entry *
2153ice_find_ucast_rule_entry(struct ice_hw *hw, u8 recp_id,
2154			  struct ice_fltr_info *f_info)
2155{
2156	struct ice_switch_info *sw = hw->switch_info;
2157	struct ice_fltr_mgmt_list_entry *list_itr;
2158	struct list_head *list_head;
2159
2160	list_head = &sw->recp_list[recp_id].filt_rules;
2161	list_for_each_entry(list_itr, list_head, list_entry) {
2162		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
2163			    sizeof(f_info->l_data)) &&
2164		    f_info->fwd_id.hw_vsi_id ==
2165		    list_itr->fltr_info.fwd_id.hw_vsi_id &&
2166		    f_info->flag == list_itr->fltr_info.flag)
2167			return list_itr;
2168	}
2169	return NULL;
2170}
2171
2172/**
2173 * ice_remove_mac - remove a MAC address based filter rule
2174 * @hw: pointer to the hardware structure
2175 * @m_list: list of MAC addresses and forwarding information
2176 *
2177 * This function removes either a MAC filter rule or a specific VSI from a
2178 * VSI list for a multicast MAC address.
2179 *
2180 * Returns ICE_ERR_DOES_NOT_EXIST if a given entry was not added by
2181 * ice_add_mac. Caller should be aware that this call will only work if all
2182 * the entries passed into m_list were added previously. It will not attempt to
2183 * do a partial remove of entries that were found.
2184 */
2185enum ice_status
2186ice_remove_mac(struct ice_hw *hw, struct list_head *m_list)
2187{
2188	struct ice_fltr_list_entry *list_itr, *tmp;
2189	struct mutex *rule_lock; /* Lock to protect filter rule list */
2190
2191	if (!m_list)
2192		return ICE_ERR_PARAM;
2193
2194	rule_lock = &hw->switch_info->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
2195	list_for_each_entry_safe(list_itr, tmp, m_list, list_entry) {
2196		enum ice_sw_lkup_type l_type = list_itr->fltr_info.lkup_type;
2197		u8 *add = &list_itr->fltr_info.l_data.mac.mac_addr[0];
2198		u16 vsi_handle;
2199
2200		if (l_type != ICE_SW_LKUP_MAC)
2201			return ICE_ERR_PARAM;
2202
2203		vsi_handle = list_itr->fltr_info.vsi_handle;
2204		if (!ice_is_vsi_valid(hw, vsi_handle))
2205			return ICE_ERR_PARAM;
2206
2207		list_itr->fltr_info.fwd_id.hw_vsi_id =
2208					ice_get_hw_vsi_num(hw, vsi_handle);
2209		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
2210			/* Don't remove the unicast address that belongs to
2211			 * another VSI on the switch, since it is not being
2212			 * shared...
2213			 */
2214			mutex_lock(rule_lock);
2215			if (!ice_find_ucast_rule_entry(hw, ICE_SW_LKUP_MAC,
2216						       &list_itr->fltr_info)) {
2217				mutex_unlock(rule_lock);
2218				return ICE_ERR_DOES_NOT_EXIST;
2219			}
2220			mutex_unlock(rule_lock);
2221		}
2222		list_itr->status = ice_remove_rule_internal(hw,
2223							    ICE_SW_LKUP_MAC,
2224							    list_itr);
2225		if (list_itr->status)
2226			return list_itr->status;
2227	}
2228	return 0;
2229}
2230
2231/**
2232 * ice_remove_vlan - Remove VLAN based filter rule
2233 * @hw: pointer to the hardware structure
2234 * @v_list: list of VLAN entries and forwarding information
2235 */
2236enum ice_status
2237ice_remove_vlan(struct ice_hw *hw, struct list_head *v_list)
2238{
2239	struct ice_fltr_list_entry *v_list_itr, *tmp;
2240
2241	if (!v_list || !hw)
2242		return ICE_ERR_PARAM;
2243
2244	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2245		enum ice_sw_lkup_type l_type = v_list_itr->fltr_info.lkup_type;
2246
2247		if (l_type != ICE_SW_LKUP_VLAN)
2248			return ICE_ERR_PARAM;
2249		v_list_itr->status = ice_remove_rule_internal(hw,
2250							      ICE_SW_LKUP_VLAN,
2251							      v_list_itr);
2252		if (v_list_itr->status)
2253			return v_list_itr->status;
2254	}
2255	return 0;
2256}
2257
2258/**
2259 * ice_vsi_uses_fltr - Determine if given VSI uses specified filter
2260 * @fm_entry: filter entry to inspect
2261 * @vsi_handle: VSI handle to compare with filter info
2262 */
2263static bool
2264ice_vsi_uses_fltr(struct ice_fltr_mgmt_list_entry *fm_entry, u16 vsi_handle)
2265{
2266	return ((fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI &&
2267		 fm_entry->fltr_info.vsi_handle == vsi_handle) ||
2268		(fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI_LIST &&
2269		 (test_bit(vsi_handle, fm_entry->vsi_list_info->vsi_map))));
2270}
2271
2272/**
2273 * ice_add_entry_to_vsi_fltr_list - Add copy of fltr_list_entry to remove list
2274 * @hw: pointer to the hardware structure
2275 * @vsi_handle: VSI handle to remove filters from
2276 * @vsi_list_head: pointer to the list to add entry to
2277 * @fi: pointer to fltr_info of filter entry to copy & add
2278 *
2279 * Helper function, used when creating a list of filters to remove from
2280 * a specific VSI. The entry added to vsi_list_head is a COPY of the
2281 * original filter entry, with the exception of fltr_info.fltr_act and
2282 * fltr_info.fwd_id fields. These are set such that later logic can
2283 * extract which VSI to remove the fltr from, and pass on that information.
2284 */
2285static enum ice_status
2286ice_add_entry_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2287			       struct list_head *vsi_list_head,
2288			       struct ice_fltr_info *fi)
2289{
2290	struct ice_fltr_list_entry *tmp;
2291
2292	/* this memory is freed up in the caller function
2293	 * once filters for this VSI are removed
2294	 */
2295	tmp = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*tmp), GFP_KERNEL);
2296	if (!tmp)
2297		return ICE_ERR_NO_MEMORY;
2298
2299	tmp->fltr_info = *fi;
2300
2301	/* Overwrite these fields to indicate which VSI to remove filter from,
2302	 * so find and remove logic can extract the information from the
2303	 * list entries. Note that original entries will still have proper
2304	 * values.
2305	 */
2306	tmp->fltr_info.fltr_act = ICE_FWD_TO_VSI;
2307	tmp->fltr_info.vsi_handle = vsi_handle;
2308	tmp->fltr_info.fwd_id.hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2309
2310	list_add(&tmp->list_entry, vsi_list_head);
2311
2312	return 0;
2313}
2314
2315/**
2316 * ice_add_to_vsi_fltr_list - Add VSI filters to the list
2317 * @hw: pointer to the hardware structure
2318 * @vsi_handle: VSI handle to remove filters from
2319 * @lkup_list_head: pointer to the list that has certain lookup type filters
2320 * @vsi_list_head: pointer to the list pertaining to VSI with vsi_handle
2321 *
2322 * Locates all filters in lkup_list_head that are used by the given VSI,
2323 * and adds COPIES of those entries to vsi_list_head (intended to be used
2324 * to remove the listed filters).
2325 * Note that this means all entries in vsi_list_head must be explicitly
2326 * deallocated by the caller when done with list.
2327 */
2328static enum ice_status
2329ice_add_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2330			 struct list_head *lkup_list_head,
2331			 struct list_head *vsi_list_head)
2332{
2333	struct ice_fltr_mgmt_list_entry *fm_entry;
2334	enum ice_status status = 0;
2335
2336	/* check to make sure VSI ID is valid and within boundary */
2337	if (!ice_is_vsi_valid(hw, vsi_handle))
2338		return ICE_ERR_PARAM;
2339
2340	list_for_each_entry(fm_entry, lkup_list_head, list_entry) {
2341		struct ice_fltr_info *fi;
2342
2343		fi = &fm_entry->fltr_info;
2344		if (!fi || !ice_vsi_uses_fltr(fm_entry, vsi_handle))
2345			continue;
2346
2347		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2348							vsi_list_head, fi);
2349		if (status)
2350			return status;
2351	}
2352	return status;
2353}
2354
2355/**
2356 * ice_determine_promisc_mask
2357 * @fi: filter info to parse
2358 *
2359 * Helper function to determine which ICE_PROMISC_ mask corresponds
2360 * to given filter into.
2361 */
2362static u8 ice_determine_promisc_mask(struct ice_fltr_info *fi)
2363{
2364	u16 vid = fi->l_data.mac_vlan.vlan_id;
2365	u8 *macaddr = fi->l_data.mac.mac_addr;
2366	bool is_tx_fltr = false;
2367	u8 promisc_mask = 0;
2368
2369	if (fi->flag == ICE_FLTR_TX)
2370		is_tx_fltr = true;
2371
2372	if (is_broadcast_ether_addr(macaddr))
2373		promisc_mask |= is_tx_fltr ?
2374			ICE_PROMISC_BCAST_TX : ICE_PROMISC_BCAST_RX;
2375	else if (is_multicast_ether_addr(macaddr))
2376		promisc_mask |= is_tx_fltr ?
2377			ICE_PROMISC_MCAST_TX : ICE_PROMISC_MCAST_RX;
2378	else if (is_unicast_ether_addr(macaddr))
2379		promisc_mask |= is_tx_fltr ?
2380			ICE_PROMISC_UCAST_TX : ICE_PROMISC_UCAST_RX;
2381	if (vid)
2382		promisc_mask |= is_tx_fltr ?
2383			ICE_PROMISC_VLAN_TX : ICE_PROMISC_VLAN_RX;
2384
2385	return promisc_mask;
2386}
2387
2388/**
2389 * ice_remove_promisc - Remove promisc based filter rules
2390 * @hw: pointer to the hardware structure
2391 * @recp_id: recipe ID for which the rule needs to removed
2392 * @v_list: list of promisc entries
2393 */
2394static enum ice_status
2395ice_remove_promisc(struct ice_hw *hw, u8 recp_id,
2396		   struct list_head *v_list)
2397{
2398	struct ice_fltr_list_entry *v_list_itr, *tmp;
2399
2400	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2401		v_list_itr->status =
2402			ice_remove_rule_internal(hw, recp_id, v_list_itr);
2403		if (v_list_itr->status)
2404			return v_list_itr->status;
2405	}
2406	return 0;
2407}
2408
2409/**
2410 * ice_clear_vsi_promisc - clear specified promiscuous mode(s) for given VSI
2411 * @hw: pointer to the hardware structure
2412 * @vsi_handle: VSI handle to clear mode
2413 * @promisc_mask: mask of promiscuous config bits to clear
2414 * @vid: VLAN ID to clear VLAN promiscuous
2415 */
2416enum ice_status
2417ice_clear_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2418		      u16 vid)
2419{
2420	struct ice_switch_info *sw = hw->switch_info;
2421	struct ice_fltr_list_entry *fm_entry, *tmp;
2422	struct list_head remove_list_head;
2423	struct ice_fltr_mgmt_list_entry *itr;
2424	struct list_head *rule_head;
2425	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2426	enum ice_status status = 0;
2427	u8 recipe_id;
2428
2429	if (!ice_is_vsi_valid(hw, vsi_handle))
2430		return ICE_ERR_PARAM;
2431
2432	if (vid)
2433		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2434	else
2435		recipe_id = ICE_SW_LKUP_PROMISC;
2436
2437	rule_head = &sw->recp_list[recipe_id].filt_rules;
2438	rule_lock = &sw->recp_list[recipe_id].filt_rule_lock;
2439
2440	INIT_LIST_HEAD(&remove_list_head);
2441
2442	mutex_lock(rule_lock);
2443	list_for_each_entry(itr, rule_head, list_entry) {
 
2444		u8 fltr_promisc_mask = 0;
2445
2446		if (!ice_vsi_uses_fltr(itr, vsi_handle))
2447			continue;
 
 
 
 
 
2448
2449		fltr_promisc_mask |=
2450			ice_determine_promisc_mask(&itr->fltr_info);
2451
2452		/* Skip if filter is not completely specified by given mask */
2453		if (fltr_promisc_mask & ~promisc_mask)
2454			continue;
2455
2456		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2457							&remove_list_head,
2458							&itr->fltr_info);
2459		if (status) {
2460			mutex_unlock(rule_lock);
2461			goto free_fltr_list;
2462		}
2463	}
2464	mutex_unlock(rule_lock);
2465
2466	status = ice_remove_promisc(hw, recipe_id, &remove_list_head);
2467
2468free_fltr_list:
2469	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2470		list_del(&fm_entry->list_entry);
2471		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2472	}
2473
2474	return status;
2475}
2476
2477/**
2478 * ice_set_vsi_promisc - set given VSI to given promiscuous mode(s)
2479 * @hw: pointer to the hardware structure
2480 * @vsi_handle: VSI handle to configure
2481 * @promisc_mask: mask of promiscuous config bits
2482 * @vid: VLAN ID to set VLAN promiscuous
2483 */
2484enum ice_status
2485ice_set_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask, u16 vid)
2486{
2487	enum { UCAST_FLTR = 1, MCAST_FLTR, BCAST_FLTR };
2488	struct ice_fltr_list_entry f_list_entry;
2489	struct ice_fltr_info new_fltr;
2490	enum ice_status status = 0;
2491	bool is_tx_fltr;
2492	u16 hw_vsi_id;
2493	int pkt_type;
2494	u8 recipe_id;
2495
2496	if (!ice_is_vsi_valid(hw, vsi_handle))
2497		return ICE_ERR_PARAM;
2498	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2499
2500	memset(&new_fltr, 0, sizeof(new_fltr));
2501
2502	if (promisc_mask & (ICE_PROMISC_VLAN_RX | ICE_PROMISC_VLAN_TX)) {
2503		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC_VLAN;
2504		new_fltr.l_data.mac_vlan.vlan_id = vid;
2505		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2506	} else {
2507		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC;
2508		recipe_id = ICE_SW_LKUP_PROMISC;
2509	}
2510
2511	/* Separate filters must be set for each direction/packet type
2512	 * combination, so we will loop over the mask value, store the
2513	 * individual type, and clear it out in the input mask as it
2514	 * is found.
2515	 */
2516	while (promisc_mask) {
2517		u8 *mac_addr;
2518
2519		pkt_type = 0;
2520		is_tx_fltr = false;
2521
2522		if (promisc_mask & ICE_PROMISC_UCAST_RX) {
2523			promisc_mask &= ~ICE_PROMISC_UCAST_RX;
2524			pkt_type = UCAST_FLTR;
2525		} else if (promisc_mask & ICE_PROMISC_UCAST_TX) {
2526			promisc_mask &= ~ICE_PROMISC_UCAST_TX;
2527			pkt_type = UCAST_FLTR;
2528			is_tx_fltr = true;
2529		} else if (promisc_mask & ICE_PROMISC_MCAST_RX) {
2530			promisc_mask &= ~ICE_PROMISC_MCAST_RX;
2531			pkt_type = MCAST_FLTR;
2532		} else if (promisc_mask & ICE_PROMISC_MCAST_TX) {
2533			promisc_mask &= ~ICE_PROMISC_MCAST_TX;
2534			pkt_type = MCAST_FLTR;
2535			is_tx_fltr = true;
2536		} else if (promisc_mask & ICE_PROMISC_BCAST_RX) {
2537			promisc_mask &= ~ICE_PROMISC_BCAST_RX;
2538			pkt_type = BCAST_FLTR;
2539		} else if (promisc_mask & ICE_PROMISC_BCAST_TX) {
2540			promisc_mask &= ~ICE_PROMISC_BCAST_TX;
2541			pkt_type = BCAST_FLTR;
2542			is_tx_fltr = true;
2543		}
2544
2545		/* Check for VLAN promiscuous flag */
2546		if (promisc_mask & ICE_PROMISC_VLAN_RX) {
2547			promisc_mask &= ~ICE_PROMISC_VLAN_RX;
2548		} else if (promisc_mask & ICE_PROMISC_VLAN_TX) {
2549			promisc_mask &= ~ICE_PROMISC_VLAN_TX;
2550			is_tx_fltr = true;
2551		}
2552
2553		/* Set filter DA based on packet type */
2554		mac_addr = new_fltr.l_data.mac.mac_addr;
2555		if (pkt_type == BCAST_FLTR) {
2556			eth_broadcast_addr(mac_addr);
2557		} else if (pkt_type == MCAST_FLTR ||
2558			   pkt_type == UCAST_FLTR) {
2559			/* Use the dummy ether header DA */
2560			ether_addr_copy(mac_addr, dummy_eth_header);
2561			if (pkt_type == MCAST_FLTR)
2562				mac_addr[0] |= 0x1;	/* Set multicast bit */
2563		}
2564
2565		/* Need to reset this to zero for all iterations */
2566		new_fltr.flag = 0;
2567		if (is_tx_fltr) {
2568			new_fltr.flag |= ICE_FLTR_TX;
2569			new_fltr.src = hw_vsi_id;
2570		} else {
2571			new_fltr.flag |= ICE_FLTR_RX;
2572			new_fltr.src = hw->port_info->lport;
2573		}
2574
2575		new_fltr.fltr_act = ICE_FWD_TO_VSI;
2576		new_fltr.vsi_handle = vsi_handle;
2577		new_fltr.fwd_id.hw_vsi_id = hw_vsi_id;
2578		f_list_entry.fltr_info = new_fltr;
2579
2580		status = ice_add_rule_internal(hw, recipe_id, &f_list_entry);
2581		if (status)
2582			goto set_promisc_exit;
2583	}
2584
2585set_promisc_exit:
2586	return status;
2587}
2588
2589/**
2590 * ice_set_vlan_vsi_promisc
2591 * @hw: pointer to the hardware structure
2592 * @vsi_handle: VSI handle to configure
2593 * @promisc_mask: mask of promiscuous config bits
2594 * @rm_vlan_promisc: Clear VLANs VSI promisc mode
2595 *
2596 * Configure VSI with all associated VLANs to given promiscuous mode(s)
2597 */
2598enum ice_status
2599ice_set_vlan_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2600			 bool rm_vlan_promisc)
2601{
2602	struct ice_switch_info *sw = hw->switch_info;
2603	struct ice_fltr_list_entry *list_itr, *tmp;
2604	struct list_head vsi_list_head;
2605	struct list_head *vlan_head;
2606	struct mutex *vlan_lock; /* Lock to protect filter rule list */
2607	enum ice_status status;
2608	u16 vlan_id;
2609
2610	INIT_LIST_HEAD(&vsi_list_head);
2611	vlan_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
2612	vlan_head = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rules;
2613	mutex_lock(vlan_lock);
2614	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, vlan_head,
2615					  &vsi_list_head);
2616	mutex_unlock(vlan_lock);
2617	if (status)
2618		goto free_fltr_list;
2619
2620	list_for_each_entry(list_itr, &vsi_list_head, list_entry) {
2621		vlan_id = list_itr->fltr_info.l_data.vlan.vlan_id;
2622		if (rm_vlan_promisc)
2623			status = ice_clear_vsi_promisc(hw, vsi_handle,
2624						       promisc_mask, vlan_id);
2625		else
2626			status = ice_set_vsi_promisc(hw, vsi_handle,
2627						     promisc_mask, vlan_id);
2628		if (status)
2629			break;
2630	}
2631
2632free_fltr_list:
2633	list_for_each_entry_safe(list_itr, tmp, &vsi_list_head, list_entry) {
2634		list_del(&list_itr->list_entry);
2635		devm_kfree(ice_hw_to_dev(hw), list_itr);
2636	}
2637	return status;
2638}
2639
2640/**
2641 * ice_remove_vsi_lkup_fltr - Remove lookup type filters for a VSI
2642 * @hw: pointer to the hardware structure
2643 * @vsi_handle: VSI handle to remove filters from
2644 * @lkup: switch rule filter lookup type
2645 */
2646static void
2647ice_remove_vsi_lkup_fltr(struct ice_hw *hw, u16 vsi_handle,
2648			 enum ice_sw_lkup_type lkup)
2649{
2650	struct ice_switch_info *sw = hw->switch_info;
2651	struct ice_fltr_list_entry *fm_entry;
2652	struct list_head remove_list_head;
2653	struct list_head *rule_head;
2654	struct ice_fltr_list_entry *tmp;
2655	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2656	enum ice_status status;
2657
2658	INIT_LIST_HEAD(&remove_list_head);
2659	rule_lock = &sw->recp_list[lkup].filt_rule_lock;
2660	rule_head = &sw->recp_list[lkup].filt_rules;
2661	mutex_lock(rule_lock);
2662	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, rule_head,
2663					  &remove_list_head);
2664	mutex_unlock(rule_lock);
2665	if (status)
2666		return;
2667
2668	switch (lkup) {
2669	case ICE_SW_LKUP_MAC:
2670		ice_remove_mac(hw, &remove_list_head);
2671		break;
2672	case ICE_SW_LKUP_VLAN:
2673		ice_remove_vlan(hw, &remove_list_head);
2674		break;
2675	case ICE_SW_LKUP_PROMISC:
2676	case ICE_SW_LKUP_PROMISC_VLAN:
2677		ice_remove_promisc(hw, lkup, &remove_list_head);
2678		break;
2679	case ICE_SW_LKUP_MAC_VLAN:
2680	case ICE_SW_LKUP_ETHERTYPE:
2681	case ICE_SW_LKUP_ETHERTYPE_MAC:
2682	case ICE_SW_LKUP_DFLT:
2683	case ICE_SW_LKUP_LAST:
2684	default:
2685		ice_debug(hw, ICE_DBG_SW, "Unsupported lookup type %d\n", lkup);
2686		break;
2687	}
2688
2689	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2690		list_del(&fm_entry->list_entry);
2691		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2692	}
2693}
2694
2695/**
2696 * ice_remove_vsi_fltr - Remove all filters for a VSI
2697 * @hw: pointer to the hardware structure
2698 * @vsi_handle: VSI handle to remove filters from
2699 */
2700void ice_remove_vsi_fltr(struct ice_hw *hw, u16 vsi_handle)
2701{
2702	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC);
2703	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC_VLAN);
2704	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC);
2705	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_VLAN);
2706	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_DFLT);
2707	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE);
2708	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE_MAC);
2709	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC_VLAN);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2710}
2711
2712/**
2713 * ice_replay_vsi_fltr - Replay filters for requested VSI
2714 * @hw: pointer to the hardware structure
2715 * @vsi_handle: driver VSI handle
2716 * @recp_id: Recipe ID for which rules need to be replayed
2717 * @list_head: list for which filters need to be replayed
2718 *
2719 * Replays the filter of recipe recp_id for a VSI represented via vsi_handle.
2720 * It is required to pass valid VSI handle.
2721 */
2722static enum ice_status
2723ice_replay_vsi_fltr(struct ice_hw *hw, u16 vsi_handle, u8 recp_id,
2724		    struct list_head *list_head)
2725{
2726	struct ice_fltr_mgmt_list_entry *itr;
2727	enum ice_status status = 0;
2728	u16 hw_vsi_id;
2729
2730	if (list_empty(list_head))
2731		return status;
2732	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2733
2734	list_for_each_entry(itr, list_head, list_entry) {
2735		struct ice_fltr_list_entry f_entry;
2736
2737		f_entry.fltr_info = itr->fltr_info;
2738		if (itr->vsi_count < 2 && recp_id != ICE_SW_LKUP_VLAN &&
2739		    itr->fltr_info.vsi_handle == vsi_handle) {
2740			/* update the src in case it is VSI num */
2741			if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2742				f_entry.fltr_info.src = hw_vsi_id;
2743			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2744			if (status)
2745				goto end;
2746			continue;
2747		}
2748		if (!itr->vsi_list_info ||
2749		    !test_bit(vsi_handle, itr->vsi_list_info->vsi_map))
2750			continue;
2751		/* Clearing it so that the logic can add it back */
2752		clear_bit(vsi_handle, itr->vsi_list_info->vsi_map);
2753		f_entry.fltr_info.vsi_handle = vsi_handle;
2754		f_entry.fltr_info.fltr_act = ICE_FWD_TO_VSI;
2755		/* update the src in case it is VSI num */
2756		if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2757			f_entry.fltr_info.src = hw_vsi_id;
2758		if (recp_id == ICE_SW_LKUP_VLAN)
2759			status = ice_add_vlan_internal(hw, &f_entry);
2760		else
2761			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2762		if (status)
2763			goto end;
2764	}
2765end:
2766	return status;
2767}
2768
2769/**
2770 * ice_replay_vsi_all_fltr - replay all filters stored in bookkeeping lists
2771 * @hw: pointer to the hardware structure
2772 * @vsi_handle: driver VSI handle
2773 *
2774 * Replays filters for requested VSI via vsi_handle.
2775 */
2776enum ice_status ice_replay_vsi_all_fltr(struct ice_hw *hw, u16 vsi_handle)
2777{
2778	struct ice_switch_info *sw = hw->switch_info;
2779	enum ice_status status = 0;
2780	u8 i;
2781
2782	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2783		struct list_head *head;
2784
2785		head = &sw->recp_list[i].filt_replay_rules;
2786		status = ice_replay_vsi_fltr(hw, vsi_handle, i, head);
2787		if (status)
2788			return status;
2789	}
2790	return status;
2791}
2792
2793/**
2794 * ice_rm_all_sw_replay_rule_info - deletes filter replay rules
2795 * @hw: pointer to the HW struct
2796 *
2797 * Deletes the filter replay rules.
2798 */
2799void ice_rm_all_sw_replay_rule_info(struct ice_hw *hw)
2800{
2801	struct ice_switch_info *sw = hw->switch_info;
2802	u8 i;
2803
2804	if (!sw)
2805		return;
2806
2807	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2808		if (!list_empty(&sw->recp_list[i].filt_replay_rules)) {
2809			struct list_head *l_head;
2810
2811			l_head = &sw->recp_list[i].filt_replay_rules;
2812			ice_rem_sw_rule_info(hw, l_head);
2813		}
2814	}
2815}
v5.9
   1// SPDX-License-Identifier: GPL-2.0
   2/* Copyright (c) 2018, Intel Corporation. */
   3
   4#include "ice_switch.h"
   5
   6#define ICE_ETH_DA_OFFSET		0
   7#define ICE_ETH_ETHTYPE_OFFSET		12
   8#define ICE_ETH_VLAN_TCI_OFFSET		14
   9#define ICE_MAX_VLAN_ID			0xFFF
  10
  11/* Dummy ethernet header needed in the ice_aqc_sw_rules_elem
  12 * struct to configure any switch filter rules.
  13 * {DA (6 bytes), SA(6 bytes),
  14 * Ether type (2 bytes for header without VLAN tag) OR
  15 * VLAN tag (4 bytes for header with VLAN tag) }
  16 *
  17 * Word on Hardcoded values
  18 * byte 0 = 0x2: to identify it as locally administered DA MAC
  19 * byte 6 = 0x2: to identify it as locally administered SA MAC
  20 * byte 12 = 0x81 & byte 13 = 0x00:
  21 *	In case of VLAN filter first two bytes defines ether type (0x8100)
  22 *	and remaining two bytes are placeholder for programming a given VLAN ID
  23 *	In case of Ether type filter it is treated as header without VLAN tag
  24 *	and byte 12 and 13 is used to program a given Ether type instead
  25 */
  26#define DUMMY_ETH_HDR_LEN		16
  27static const u8 dummy_eth_header[DUMMY_ETH_HDR_LEN] = { 0x2, 0, 0, 0, 0, 0,
  28							0x2, 0, 0, 0, 0, 0,
  29							0x81, 0, 0, 0};
  30
  31#define ICE_SW_RULE_RX_TX_ETH_HDR_SIZE \
  32	(offsetof(struct ice_aqc_sw_rules_elem, pdata.lkup_tx_rx.hdr) + \
  33	 (DUMMY_ETH_HDR_LEN * \
  34	  sizeof(((struct ice_sw_rule_lkup_rx_tx *)0)->hdr[0])))
  35#define ICE_SW_RULE_RX_TX_NO_HDR_SIZE \
  36	(offsetof(struct ice_aqc_sw_rules_elem, pdata.lkup_tx_rx.hdr))
 
 
  37#define ICE_SW_RULE_LG_ACT_SIZE(n) \
  38	(offsetof(struct ice_aqc_sw_rules_elem, pdata.lg_act.act) + \
  39	 ((n) * sizeof(((struct ice_sw_rule_lg_act *)0)->act[0])))
 
 
 
  40#define ICE_SW_RULE_VSI_LIST_SIZE(n) \
  41	(offsetof(struct ice_aqc_sw_rules_elem, pdata.vsi_list.vsi) + \
  42	 ((n) * sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi[0])))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  43
  44/**
  45 * ice_init_def_sw_recp - initialize the recipe book keeping tables
  46 * @hw: pointer to the HW struct
  47 *
  48 * Allocate memory for the entire recipe table and initialize the structures/
  49 * entries corresponding to basic recipes.
  50 */
  51enum ice_status ice_init_def_sw_recp(struct ice_hw *hw)
  52{
  53	struct ice_sw_recipe *recps;
  54	u8 i;
  55
  56	recps = devm_kcalloc(ice_hw_to_dev(hw), ICE_MAX_NUM_RECIPES,
  57			     sizeof(*recps), GFP_KERNEL);
  58	if (!recps)
  59		return ICE_ERR_NO_MEMORY;
  60
  61	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
  62		recps[i].root_rid = i;
  63		INIT_LIST_HEAD(&recps[i].filt_rules);
  64		INIT_LIST_HEAD(&recps[i].filt_replay_rules);
  65		mutex_init(&recps[i].filt_rule_lock);
  66	}
  67
  68	hw->switch_info->recp_list = recps;
  69
  70	return 0;
  71}
  72
  73/**
  74 * ice_aq_get_sw_cfg - get switch configuration
  75 * @hw: pointer to the hardware structure
  76 * @buf: pointer to the result buffer
  77 * @buf_size: length of the buffer available for response
  78 * @req_desc: pointer to requested descriptor
  79 * @num_elems: pointer to number of elements
  80 * @cd: pointer to command details structure or NULL
  81 *
  82 * Get switch configuration (0x0200) to be placed in buf.
  83 * This admin command returns information such as initial VSI/port number
  84 * and switch ID it belongs to.
  85 *
  86 * NOTE: *req_desc is both an input/output parameter.
  87 * The caller of this function first calls this function with *request_desc set
  88 * to 0. If the response from f/w has *req_desc set to 0, all the switch
  89 * configuration information has been returned; if non-zero (meaning not all
  90 * the information was returned), the caller should call this function again
  91 * with *req_desc set to the previous value returned by f/w to get the
  92 * next block of switch configuration information.
  93 *
  94 * *num_elems is output only parameter. This reflects the number of elements
  95 * in response buffer. The caller of this function to use *num_elems while
  96 * parsing the response buffer.
  97 */
  98static enum ice_status
  99ice_aq_get_sw_cfg(struct ice_hw *hw, struct ice_aqc_get_sw_cfg_resp_elem *buf,
 100		  u16 buf_size, u16 *req_desc, u16 *num_elems,
 101		  struct ice_sq_cd *cd)
 102{
 103	struct ice_aqc_get_sw_cfg *cmd;
 
 104	struct ice_aq_desc desc;
 105	enum ice_status status;
 106
 107	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_get_sw_cfg);
 108	cmd = &desc.params.get_sw_conf;
 109	cmd->element = cpu_to_le16(*req_desc);
 110
 111	status = ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
 112	if (!status) {
 113		*req_desc = le16_to_cpu(cmd->element);
 114		*num_elems = le16_to_cpu(cmd->num_elems);
 115	}
 116
 117	return status;
 118}
 119
 120/**
 121 * ice_aq_add_vsi
 122 * @hw: pointer to the HW struct
 123 * @vsi_ctx: pointer to a VSI context struct
 124 * @cd: pointer to command details structure or NULL
 125 *
 126 * Add a VSI context to the hardware (0x0210)
 127 */
 128static enum ice_status
 129ice_aq_add_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 130	       struct ice_sq_cd *cd)
 131{
 132	struct ice_aqc_add_update_free_vsi_resp *res;
 133	struct ice_aqc_add_get_update_free_vsi *cmd;
 134	struct ice_aq_desc desc;
 135	enum ice_status status;
 136
 137	cmd = &desc.params.vsi_cmd;
 138	res = &desc.params.add_update_free_vsi_res;
 139
 140	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_add_vsi);
 141
 142	if (!vsi_ctx->alloc_from_pool)
 143		cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num |
 144					   ICE_AQ_VSI_IS_VALID);
 145	cmd->vf_id = vsi_ctx->vf_num;
 146
 147	cmd->vsi_flags = cpu_to_le16(vsi_ctx->flags);
 148
 149	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 150
 151	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 152				 sizeof(vsi_ctx->info), cd);
 153
 154	if (!status) {
 155		vsi_ctx->vsi_num = le16_to_cpu(res->vsi_num) & ICE_AQ_VSI_NUM_M;
 156		vsi_ctx->vsis_allocd = le16_to_cpu(res->vsi_used);
 157		vsi_ctx->vsis_unallocated = le16_to_cpu(res->vsi_free);
 158	}
 159
 160	return status;
 161}
 162
 163/**
 164 * ice_aq_free_vsi
 165 * @hw: pointer to the HW struct
 166 * @vsi_ctx: pointer to a VSI context struct
 167 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 168 * @cd: pointer to command details structure or NULL
 169 *
 170 * Free VSI context info from hardware (0x0213)
 171 */
 172static enum ice_status
 173ice_aq_free_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 174		bool keep_vsi_alloc, struct ice_sq_cd *cd)
 175{
 176	struct ice_aqc_add_update_free_vsi_resp *resp;
 177	struct ice_aqc_add_get_update_free_vsi *cmd;
 178	struct ice_aq_desc desc;
 179	enum ice_status status;
 180
 181	cmd = &desc.params.vsi_cmd;
 182	resp = &desc.params.add_update_free_vsi_res;
 183
 184	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_free_vsi);
 185
 186	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 187	if (keep_vsi_alloc)
 188		cmd->cmd_flags = cpu_to_le16(ICE_AQ_VSI_KEEP_ALLOC);
 189
 190	status = ice_aq_send_cmd(hw, &desc, NULL, 0, cd);
 191	if (!status) {
 192		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 193		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 194	}
 195
 196	return status;
 197}
 198
 199/**
 200 * ice_aq_update_vsi
 201 * @hw: pointer to the HW struct
 202 * @vsi_ctx: pointer to a VSI context struct
 203 * @cd: pointer to command details structure or NULL
 204 *
 205 * Update VSI context in the hardware (0x0211)
 206 */
 207static enum ice_status
 208ice_aq_update_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
 209		  struct ice_sq_cd *cd)
 210{
 211	struct ice_aqc_add_update_free_vsi_resp *resp;
 212	struct ice_aqc_add_get_update_free_vsi *cmd;
 213	struct ice_aq_desc desc;
 214	enum ice_status status;
 215
 216	cmd = &desc.params.vsi_cmd;
 217	resp = &desc.params.add_update_free_vsi_res;
 218
 219	ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_update_vsi);
 220
 221	cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
 222
 223	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 224
 225	status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
 226				 sizeof(vsi_ctx->info), cd);
 227
 228	if (!status) {
 229		vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
 230		vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
 231	}
 232
 233	return status;
 234}
 235
 236/**
 237 * ice_is_vsi_valid - check whether the VSI is valid or not
 238 * @hw: pointer to the HW struct
 239 * @vsi_handle: VSI handle
 240 *
 241 * check whether the VSI is valid or not
 242 */
 243bool ice_is_vsi_valid(struct ice_hw *hw, u16 vsi_handle)
 244{
 245	return vsi_handle < ICE_MAX_VSI && hw->vsi_ctx[vsi_handle];
 246}
 247
 248/**
 249 * ice_get_hw_vsi_num - return the HW VSI number
 250 * @hw: pointer to the HW struct
 251 * @vsi_handle: VSI handle
 252 *
 253 * return the HW VSI number
 254 * Caution: call this function only if VSI is valid (ice_is_vsi_valid)
 255 */
 256u16 ice_get_hw_vsi_num(struct ice_hw *hw, u16 vsi_handle)
 257{
 258	return hw->vsi_ctx[vsi_handle]->vsi_num;
 259}
 260
 261/**
 262 * ice_get_vsi_ctx - return the VSI context entry for a given VSI handle
 263 * @hw: pointer to the HW struct
 264 * @vsi_handle: VSI handle
 265 *
 266 * return the VSI context entry for a given VSI handle
 267 */
 268struct ice_vsi_ctx *ice_get_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 269{
 270	return (vsi_handle >= ICE_MAX_VSI) ? NULL : hw->vsi_ctx[vsi_handle];
 271}
 272
 273/**
 274 * ice_save_vsi_ctx - save the VSI context for a given VSI handle
 275 * @hw: pointer to the HW struct
 276 * @vsi_handle: VSI handle
 277 * @vsi: VSI context pointer
 278 *
 279 * save the VSI context entry for a given VSI handle
 280 */
 281static void
 282ice_save_vsi_ctx(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi)
 283{
 284	hw->vsi_ctx[vsi_handle] = vsi;
 285}
 286
 287/**
 288 * ice_clear_vsi_q_ctx - clear VSI queue contexts for all TCs
 289 * @hw: pointer to the HW struct
 290 * @vsi_handle: VSI handle
 291 */
 292static void ice_clear_vsi_q_ctx(struct ice_hw *hw, u16 vsi_handle)
 293{
 294	struct ice_vsi_ctx *vsi;
 295	u8 i;
 296
 297	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 298	if (!vsi)
 299		return;
 300	ice_for_each_traffic_class(i) {
 301		if (vsi->lan_q_ctx[i]) {
 302			devm_kfree(ice_hw_to_dev(hw), vsi->lan_q_ctx[i]);
 303			vsi->lan_q_ctx[i] = NULL;
 304		}
 305	}
 306}
 307
 308/**
 309 * ice_clear_vsi_ctx - clear the VSI context entry
 310 * @hw: pointer to the HW struct
 311 * @vsi_handle: VSI handle
 312 *
 313 * clear the VSI context entry
 314 */
 315static void ice_clear_vsi_ctx(struct ice_hw *hw, u16 vsi_handle)
 316{
 317	struct ice_vsi_ctx *vsi;
 318
 319	vsi = ice_get_vsi_ctx(hw, vsi_handle);
 320	if (vsi) {
 321		ice_clear_vsi_q_ctx(hw, vsi_handle);
 322		devm_kfree(ice_hw_to_dev(hw), vsi);
 323		hw->vsi_ctx[vsi_handle] = NULL;
 324	}
 325}
 326
 327/**
 328 * ice_clear_all_vsi_ctx - clear all the VSI context entries
 329 * @hw: pointer to the HW struct
 330 */
 331void ice_clear_all_vsi_ctx(struct ice_hw *hw)
 332{
 333	u16 i;
 334
 335	for (i = 0; i < ICE_MAX_VSI; i++)
 336		ice_clear_vsi_ctx(hw, i);
 337}
 338
 339/**
 340 * ice_add_vsi - add VSI context to the hardware and VSI handle list
 341 * @hw: pointer to the HW struct
 342 * @vsi_handle: unique VSI handle provided by drivers
 343 * @vsi_ctx: pointer to a VSI context struct
 344 * @cd: pointer to command details structure or NULL
 345 *
 346 * Add a VSI context to the hardware also add it into the VSI handle list.
 347 * If this function gets called after reset for existing VSIs then update
 348 * with the new HW VSI number in the corresponding VSI handle list entry.
 349 */
 350enum ice_status
 351ice_add_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 352	    struct ice_sq_cd *cd)
 353{
 354	struct ice_vsi_ctx *tmp_vsi_ctx;
 355	enum ice_status status;
 356
 357	if (vsi_handle >= ICE_MAX_VSI)
 358		return ICE_ERR_PARAM;
 359	status = ice_aq_add_vsi(hw, vsi_ctx, cd);
 360	if (status)
 361		return status;
 362	tmp_vsi_ctx = ice_get_vsi_ctx(hw, vsi_handle);
 363	if (!tmp_vsi_ctx) {
 364		/* Create a new VSI context */
 365		tmp_vsi_ctx = devm_kzalloc(ice_hw_to_dev(hw),
 366					   sizeof(*tmp_vsi_ctx), GFP_KERNEL);
 367		if (!tmp_vsi_ctx) {
 368			ice_aq_free_vsi(hw, vsi_ctx, false, cd);
 369			return ICE_ERR_NO_MEMORY;
 370		}
 371		*tmp_vsi_ctx = *vsi_ctx;
 372		ice_save_vsi_ctx(hw, vsi_handle, tmp_vsi_ctx);
 373	} else {
 374		/* update with new HW VSI num */
 375		tmp_vsi_ctx->vsi_num = vsi_ctx->vsi_num;
 
 376	}
 377
 378	return 0;
 379}
 380
 381/**
 382 * ice_free_vsi- free VSI context from hardware and VSI handle list
 383 * @hw: pointer to the HW struct
 384 * @vsi_handle: unique VSI handle
 385 * @vsi_ctx: pointer to a VSI context struct
 386 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
 387 * @cd: pointer to command details structure or NULL
 388 *
 389 * Free VSI context info from hardware as well as from VSI handle list
 390 */
 391enum ice_status
 392ice_free_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 393	     bool keep_vsi_alloc, struct ice_sq_cd *cd)
 394{
 395	enum ice_status status;
 396
 397	if (!ice_is_vsi_valid(hw, vsi_handle))
 398		return ICE_ERR_PARAM;
 399	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 400	status = ice_aq_free_vsi(hw, vsi_ctx, keep_vsi_alloc, cd);
 401	if (!status)
 402		ice_clear_vsi_ctx(hw, vsi_handle);
 403	return status;
 404}
 405
 406/**
 407 * ice_update_vsi
 408 * @hw: pointer to the HW struct
 409 * @vsi_handle: unique VSI handle
 410 * @vsi_ctx: pointer to a VSI context struct
 411 * @cd: pointer to command details structure or NULL
 412 *
 413 * Update VSI context in the hardware
 414 */
 415enum ice_status
 416ice_update_vsi(struct ice_hw *hw, u16 vsi_handle, struct ice_vsi_ctx *vsi_ctx,
 417	       struct ice_sq_cd *cd)
 418{
 419	if (!ice_is_vsi_valid(hw, vsi_handle))
 420		return ICE_ERR_PARAM;
 421	vsi_ctx->vsi_num = ice_get_hw_vsi_num(hw, vsi_handle);
 422	return ice_aq_update_vsi(hw, vsi_ctx, cd);
 423}
 424
 425/**
 426 * ice_aq_alloc_free_vsi_list
 427 * @hw: pointer to the HW struct
 428 * @vsi_list_id: VSI list ID returned or used for lookup
 429 * @lkup_type: switch rule filter lookup type
 430 * @opc: switch rules population command type - pass in the command opcode
 431 *
 432 * allocates or free a VSI list resource
 433 */
 434static enum ice_status
 435ice_aq_alloc_free_vsi_list(struct ice_hw *hw, u16 *vsi_list_id,
 436			   enum ice_sw_lkup_type lkup_type,
 437			   enum ice_adminq_opc opc)
 438{
 439	struct ice_aqc_alloc_free_res_elem *sw_buf;
 440	struct ice_aqc_res_elem *vsi_ele;
 441	enum ice_status status;
 442	u16 buf_len;
 443
 444	buf_len = struct_size(sw_buf, elem, 1);
 445	sw_buf = devm_kzalloc(ice_hw_to_dev(hw), buf_len, GFP_KERNEL);
 446	if (!sw_buf)
 447		return ICE_ERR_NO_MEMORY;
 448	sw_buf->num_elems = cpu_to_le16(1);
 449
 450	if (lkup_type == ICE_SW_LKUP_MAC ||
 451	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 452	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 453	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 454	    lkup_type == ICE_SW_LKUP_PROMISC ||
 455	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN) {
 456		sw_buf->res_type = cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_REP);
 457	} else if (lkup_type == ICE_SW_LKUP_VLAN) {
 458		sw_buf->res_type =
 459			cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_PRUNE);
 460	} else {
 461		status = ICE_ERR_PARAM;
 462		goto ice_aq_alloc_free_vsi_list_exit;
 463	}
 464
 465	if (opc == ice_aqc_opc_free_res)
 466		sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
 467
 468	status = ice_aq_alloc_free_res(hw, 1, sw_buf, buf_len, opc, NULL);
 469	if (status)
 470		goto ice_aq_alloc_free_vsi_list_exit;
 471
 472	if (opc == ice_aqc_opc_alloc_res) {
 473		vsi_ele = &sw_buf->elem[0];
 474		*vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
 475	}
 476
 477ice_aq_alloc_free_vsi_list_exit:
 478	devm_kfree(ice_hw_to_dev(hw), sw_buf);
 479	return status;
 480}
 481
 482/**
 483 * ice_aq_sw_rules - add/update/remove switch rules
 484 * @hw: pointer to the HW struct
 485 * @rule_list: pointer to switch rule population list
 486 * @rule_list_sz: total size of the rule list in bytes
 487 * @num_rules: number of switch rules in the rule_list
 488 * @opc: switch rules population command type - pass in the command opcode
 489 * @cd: pointer to command details structure or NULL
 490 *
 491 * Add(0x02a0)/Update(0x02a1)/Remove(0x02a2) switch rules commands to firmware
 492 */
 493static enum ice_status
 494ice_aq_sw_rules(struct ice_hw *hw, void *rule_list, u16 rule_list_sz,
 495		u8 num_rules, enum ice_adminq_opc opc, struct ice_sq_cd *cd)
 496{
 497	struct ice_aq_desc desc;
 498	enum ice_status status;
 499
 500	if (opc != ice_aqc_opc_add_sw_rules &&
 501	    opc != ice_aqc_opc_update_sw_rules &&
 502	    opc != ice_aqc_opc_remove_sw_rules)
 503		return ICE_ERR_PARAM;
 504
 505	ice_fill_dflt_direct_cmd_desc(&desc, opc);
 506
 507	desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
 508	desc.params.sw_rules.num_rules_fltr_entry_index =
 509		cpu_to_le16(num_rules);
 510	status = ice_aq_send_cmd(hw, &desc, rule_list, rule_list_sz, cd);
 511	if (opc != ice_aqc_opc_add_sw_rules &&
 512	    hw->adminq.sq_last_status == ICE_AQ_RC_ENOENT)
 513		status = ICE_ERR_DOES_NOT_EXIST;
 514
 515	return status;
 516}
 517
 518/* ice_init_port_info - Initialize port_info with switch configuration data
 519 * @pi: pointer to port_info
 520 * @vsi_port_num: VSI number or port number
 521 * @type: Type of switch element (port or VSI)
 522 * @swid: switch ID of the switch the element is attached to
 523 * @pf_vf_num: PF or VF number
 524 * @is_vf: true if the element is a VF, false otherwise
 525 */
 526static void
 527ice_init_port_info(struct ice_port_info *pi, u16 vsi_port_num, u8 type,
 528		   u16 swid, u16 pf_vf_num, bool is_vf)
 529{
 530	switch (type) {
 531	case ICE_AQC_GET_SW_CONF_RESP_PHYS_PORT:
 532		pi->lport = (u8)(vsi_port_num & ICE_LPORT_MASK);
 533		pi->sw_id = swid;
 534		pi->pf_vf_num = pf_vf_num;
 535		pi->is_vf = is_vf;
 536		pi->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
 537		pi->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
 538		break;
 539	default:
 540		ice_debug(pi->hw, ICE_DBG_SW,
 541			  "incorrect VSI/port type received\n");
 542		break;
 543	}
 544}
 545
 546/* ice_get_initial_sw_cfg - Get initial port and default VSI data
 547 * @hw: pointer to the hardware structure
 548 */
 549enum ice_status ice_get_initial_sw_cfg(struct ice_hw *hw)
 550{
 551	struct ice_aqc_get_sw_cfg_resp_elem *rbuf;
 552	enum ice_status status;
 553	u16 req_desc = 0;
 554	u16 num_elems;
 555	u16 i;
 556
 557	rbuf = devm_kzalloc(ice_hw_to_dev(hw), ICE_SW_CFG_MAX_BUF_LEN,
 558			    GFP_KERNEL);
 559
 560	if (!rbuf)
 561		return ICE_ERR_NO_MEMORY;
 562
 563	/* Multiple calls to ice_aq_get_sw_cfg may be required
 564	 * to get all the switch configuration information. The need
 565	 * for additional calls is indicated by ice_aq_get_sw_cfg
 566	 * writing a non-zero value in req_desc
 567	 */
 568	do {
 569		struct ice_aqc_get_sw_cfg_resp_elem *ele;
 570
 571		status = ice_aq_get_sw_cfg(hw, rbuf, ICE_SW_CFG_MAX_BUF_LEN,
 572					   &req_desc, &num_elems, NULL);
 573
 574		if (status)
 575			break;
 576
 577		for (i = 0, ele = rbuf; i < num_elems; i++, ele++) {
 
 578			u16 pf_vf_num, swid, vsi_port_num;
 579			bool is_vf = false;
 580			u8 res_type;
 581
 
 582			vsi_port_num = le16_to_cpu(ele->vsi_port_num) &
 583				ICE_AQC_GET_SW_CONF_RESP_VSI_PORT_NUM_M;
 584
 585			pf_vf_num = le16_to_cpu(ele->pf_vf_num) &
 586				ICE_AQC_GET_SW_CONF_RESP_FUNC_NUM_M;
 587
 588			swid = le16_to_cpu(ele->swid);
 589
 590			if (le16_to_cpu(ele->pf_vf_num) &
 591			    ICE_AQC_GET_SW_CONF_RESP_IS_VF)
 592				is_vf = true;
 593
 594			res_type = (u8)(le16_to_cpu(ele->vsi_port_num) >>
 595					ICE_AQC_GET_SW_CONF_RESP_TYPE_S);
 596
 597			if (res_type == ICE_AQC_GET_SW_CONF_RESP_VSI) {
 598				/* FW VSI is not needed. Just continue. */
 599				continue;
 600			}
 601
 602			ice_init_port_info(hw->port_info, vsi_port_num,
 603					   res_type, swid, pf_vf_num, is_vf);
 604		}
 605	} while (req_desc && !status);
 606
 607	devm_kfree(ice_hw_to_dev(hw), (void *)rbuf);
 608	return status;
 609}
 610
 611/**
 612 * ice_fill_sw_info - Helper function to populate lb_en and lan_en
 613 * @hw: pointer to the hardware structure
 614 * @fi: filter info structure to fill/update
 615 *
 616 * This helper function populates the lb_en and lan_en elements of the provided
 617 * ice_fltr_info struct using the switch's type and characteristics of the
 618 * switch rule being configured.
 619 */
 620static void ice_fill_sw_info(struct ice_hw *hw, struct ice_fltr_info *fi)
 621{
 622	fi->lb_en = false;
 623	fi->lan_en = false;
 624	if ((fi->flag & ICE_FLTR_TX) &&
 625	    (fi->fltr_act == ICE_FWD_TO_VSI ||
 626	     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
 627	     fi->fltr_act == ICE_FWD_TO_Q ||
 628	     fi->fltr_act == ICE_FWD_TO_QGRP)) {
 629		/* Setting LB for prune actions will result in replicated
 630		 * packets to the internal switch that will be dropped.
 631		 */
 632		if (fi->lkup_type != ICE_SW_LKUP_VLAN)
 633			fi->lb_en = true;
 634
 635		/* Set lan_en to TRUE if
 636		 * 1. The switch is a VEB AND
 637		 * 2
 638		 * 2.1 The lookup is a directional lookup like ethertype,
 639		 * promiscuous, ethertype-MAC, promiscuous-VLAN
 640		 * and default-port OR
 641		 * 2.2 The lookup is VLAN, OR
 642		 * 2.3 The lookup is MAC with mcast or bcast addr for MAC, OR
 643		 * 2.4 The lookup is MAC_VLAN with mcast or bcast addr for MAC.
 644		 *
 645		 * OR
 646		 *
 647		 * The switch is a VEPA.
 648		 *
 649		 * In all other cases, the LAN enable has to be set to false.
 650		 */
 651		if (hw->evb_veb) {
 652			if (fi->lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 653			    fi->lkup_type == ICE_SW_LKUP_PROMISC ||
 654			    fi->lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 655			    fi->lkup_type == ICE_SW_LKUP_PROMISC_VLAN ||
 656			    fi->lkup_type == ICE_SW_LKUP_DFLT ||
 657			    fi->lkup_type == ICE_SW_LKUP_VLAN ||
 658			    (fi->lkup_type == ICE_SW_LKUP_MAC &&
 659			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)) ||
 660			    (fi->lkup_type == ICE_SW_LKUP_MAC_VLAN &&
 661			     !is_unicast_ether_addr(fi->l_data.mac.mac_addr)))
 662				fi->lan_en = true;
 663		} else {
 664			fi->lan_en = true;
 665		}
 666	}
 667}
 668
 669/**
 670 * ice_fill_sw_rule - Helper function to fill switch rule structure
 671 * @hw: pointer to the hardware structure
 672 * @f_info: entry containing packet forwarding information
 673 * @s_rule: switch rule structure to be filled in based on mac_entry
 674 * @opc: switch rules population command type - pass in the command opcode
 675 */
 676static void
 677ice_fill_sw_rule(struct ice_hw *hw, struct ice_fltr_info *f_info,
 678		 struct ice_aqc_sw_rules_elem *s_rule, enum ice_adminq_opc opc)
 679{
 680	u16 vlan_id = ICE_MAX_VLAN_ID + 1;
 681	void *daddr = NULL;
 682	u16 eth_hdr_sz;
 683	u8 *eth_hdr;
 684	u32 act = 0;
 685	__be16 *off;
 686	u8 q_rgn;
 687
 688	if (opc == ice_aqc_opc_remove_sw_rules) {
 689		s_rule->pdata.lkup_tx_rx.act = 0;
 690		s_rule->pdata.lkup_tx_rx.index =
 691			cpu_to_le16(f_info->fltr_rule_id);
 692		s_rule->pdata.lkup_tx_rx.hdr_len = 0;
 693		return;
 694	}
 695
 696	eth_hdr_sz = sizeof(dummy_eth_header);
 697	eth_hdr = s_rule->pdata.lkup_tx_rx.hdr;
 698
 699	/* initialize the ether header with a dummy header */
 700	memcpy(eth_hdr, dummy_eth_header, eth_hdr_sz);
 701	ice_fill_sw_info(hw, f_info);
 702
 703	switch (f_info->fltr_act) {
 704	case ICE_FWD_TO_VSI:
 705		act |= (f_info->fwd_id.hw_vsi_id << ICE_SINGLE_ACT_VSI_ID_S) &
 706			ICE_SINGLE_ACT_VSI_ID_M;
 707		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 708			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 709				ICE_SINGLE_ACT_VALID_BIT;
 710		break;
 711	case ICE_FWD_TO_VSI_LIST:
 712		act |= ICE_SINGLE_ACT_VSI_LIST;
 713		act |= (f_info->fwd_id.vsi_list_id <<
 714			ICE_SINGLE_ACT_VSI_LIST_ID_S) &
 715			ICE_SINGLE_ACT_VSI_LIST_ID_M;
 716		if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
 717			act |= ICE_SINGLE_ACT_VSI_FORWARDING |
 718				ICE_SINGLE_ACT_VALID_BIT;
 719		break;
 720	case ICE_FWD_TO_Q:
 721		act |= ICE_SINGLE_ACT_TO_Q;
 722		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 723			ICE_SINGLE_ACT_Q_INDEX_M;
 724		break;
 725	case ICE_DROP_PACKET:
 726		act |= ICE_SINGLE_ACT_VSI_FORWARDING | ICE_SINGLE_ACT_DROP |
 727			ICE_SINGLE_ACT_VALID_BIT;
 728		break;
 729	case ICE_FWD_TO_QGRP:
 730		q_rgn = f_info->qgrp_size > 0 ?
 731			(u8)ilog2(f_info->qgrp_size) : 0;
 732		act |= ICE_SINGLE_ACT_TO_Q;
 733		act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
 734			ICE_SINGLE_ACT_Q_INDEX_M;
 735		act |= (q_rgn << ICE_SINGLE_ACT_Q_REGION_S) &
 736			ICE_SINGLE_ACT_Q_REGION_M;
 737		break;
 738	default:
 739		return;
 740	}
 741
 742	if (f_info->lb_en)
 743		act |= ICE_SINGLE_ACT_LB_ENABLE;
 744	if (f_info->lan_en)
 745		act |= ICE_SINGLE_ACT_LAN_ENABLE;
 746
 747	switch (f_info->lkup_type) {
 748	case ICE_SW_LKUP_MAC:
 749		daddr = f_info->l_data.mac.mac_addr;
 750		break;
 751	case ICE_SW_LKUP_VLAN:
 752		vlan_id = f_info->l_data.vlan.vlan_id;
 753		if (f_info->fltr_act == ICE_FWD_TO_VSI ||
 754		    f_info->fltr_act == ICE_FWD_TO_VSI_LIST) {
 755			act |= ICE_SINGLE_ACT_PRUNE;
 756			act |= ICE_SINGLE_ACT_EGRESS | ICE_SINGLE_ACT_INGRESS;
 757		}
 758		break;
 759	case ICE_SW_LKUP_ETHERTYPE_MAC:
 760		daddr = f_info->l_data.ethertype_mac.mac_addr;
 761		fallthrough;
 762	case ICE_SW_LKUP_ETHERTYPE:
 763		off = (__force __be16 *)(eth_hdr + ICE_ETH_ETHTYPE_OFFSET);
 764		*off = cpu_to_be16(f_info->l_data.ethertype_mac.ethertype);
 765		break;
 766	case ICE_SW_LKUP_MAC_VLAN:
 767		daddr = f_info->l_data.mac_vlan.mac_addr;
 768		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 769		break;
 770	case ICE_SW_LKUP_PROMISC_VLAN:
 771		vlan_id = f_info->l_data.mac_vlan.vlan_id;
 772		fallthrough;
 773	case ICE_SW_LKUP_PROMISC:
 774		daddr = f_info->l_data.mac_vlan.mac_addr;
 775		break;
 776	default:
 777		break;
 778	}
 779
 780	s_rule->type = (f_info->flag & ICE_FLTR_RX) ?
 781		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_RX) :
 782		cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_TX);
 783
 784	/* Recipe set depending on lookup type */
 785	s_rule->pdata.lkup_tx_rx.recipe_id = cpu_to_le16(f_info->lkup_type);
 786	s_rule->pdata.lkup_tx_rx.src = cpu_to_le16(f_info->src);
 787	s_rule->pdata.lkup_tx_rx.act = cpu_to_le32(act);
 788
 789	if (daddr)
 790		ether_addr_copy(eth_hdr + ICE_ETH_DA_OFFSET, daddr);
 791
 792	if (!(vlan_id > ICE_MAX_VLAN_ID)) {
 793		off = (__force __be16 *)(eth_hdr + ICE_ETH_VLAN_TCI_OFFSET);
 794		*off = cpu_to_be16(vlan_id);
 795	}
 796
 797	/* Create the switch rule with the final dummy Ethernet header */
 798	if (opc != ice_aqc_opc_update_sw_rules)
 799		s_rule->pdata.lkup_tx_rx.hdr_len = cpu_to_le16(eth_hdr_sz);
 800}
 801
 802/**
 803 * ice_add_marker_act
 804 * @hw: pointer to the hardware structure
 805 * @m_ent: the management entry for which sw marker needs to be added
 806 * @sw_marker: sw marker to tag the Rx descriptor with
 807 * @l_id: large action resource ID
 808 *
 809 * Create a large action to hold software marker and update the switch rule
 810 * entry pointed by m_ent with newly created large action
 811 */
 812static enum ice_status
 813ice_add_marker_act(struct ice_hw *hw, struct ice_fltr_mgmt_list_entry *m_ent,
 814		   u16 sw_marker, u16 l_id)
 815{
 816	struct ice_aqc_sw_rules_elem *lg_act, *rx_tx;
 817	/* For software marker we need 3 large actions
 818	 * 1. FWD action: FWD TO VSI or VSI LIST
 819	 * 2. GENERIC VALUE action to hold the profile ID
 820	 * 3. GENERIC VALUE action to hold the software marker ID
 821	 */
 822	const u16 num_lg_acts = 3;
 823	enum ice_status status;
 824	u16 lg_act_size;
 825	u16 rules_size;
 826	u32 act;
 827	u16 id;
 828
 829	if (m_ent->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
 830		return ICE_ERR_PARAM;
 831
 832	/* Create two back-to-back switch rules and submit them to the HW using
 833	 * one memory buffer:
 834	 *    1. Large Action
 835	 *    2. Look up Tx Rx
 836	 */
 837	lg_act_size = (u16)ICE_SW_RULE_LG_ACT_SIZE(num_lg_acts);
 838	rules_size = lg_act_size + ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
 839	lg_act = devm_kzalloc(ice_hw_to_dev(hw), rules_size, GFP_KERNEL);
 840	if (!lg_act)
 841		return ICE_ERR_NO_MEMORY;
 842
 843	rx_tx = (struct ice_aqc_sw_rules_elem *)((u8 *)lg_act + lg_act_size);
 844
 845	/* Fill in the first switch rule i.e. large action */
 846	lg_act->type = cpu_to_le16(ICE_AQC_SW_RULES_T_LG_ACT);
 847	lg_act->pdata.lg_act.index = cpu_to_le16(l_id);
 848	lg_act->pdata.lg_act.size = cpu_to_le16(num_lg_acts);
 849
 850	/* First action VSI forwarding or VSI list forwarding depending on how
 851	 * many VSIs
 852	 */
 853	id = (m_ent->vsi_count > 1) ? m_ent->fltr_info.fwd_id.vsi_list_id :
 854		m_ent->fltr_info.fwd_id.hw_vsi_id;
 855
 856	act = ICE_LG_ACT_VSI_FORWARDING | ICE_LG_ACT_VALID_BIT;
 857	act |= (id << ICE_LG_ACT_VSI_LIST_ID_S) & ICE_LG_ACT_VSI_LIST_ID_M;
 
 858	if (m_ent->vsi_count > 1)
 859		act |= ICE_LG_ACT_VSI_LIST;
 860	lg_act->pdata.lg_act.act[0] = cpu_to_le32(act);
 861
 862	/* Second action descriptor type */
 863	act = ICE_LG_ACT_GENERIC;
 864
 865	act |= (1 << ICE_LG_ACT_GENERIC_VALUE_S) & ICE_LG_ACT_GENERIC_VALUE_M;
 866	lg_act->pdata.lg_act.act[1] = cpu_to_le32(act);
 867
 868	act = (ICE_LG_ACT_GENERIC_OFF_RX_DESC_PROF_IDX <<
 869	       ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_OFFSET_M;
 870
 871	/* Third action Marker value */
 872	act |= ICE_LG_ACT_GENERIC;
 873	act |= (sw_marker << ICE_LG_ACT_GENERIC_VALUE_S) &
 874		ICE_LG_ACT_GENERIC_VALUE_M;
 875
 876	lg_act->pdata.lg_act.act[2] = cpu_to_le32(act);
 877
 878	/* call the fill switch rule to fill the lookup Tx Rx structure */
 879	ice_fill_sw_rule(hw, &m_ent->fltr_info, rx_tx,
 880			 ice_aqc_opc_update_sw_rules);
 881
 882	/* Update the action to point to the large action ID */
 883	rx_tx->pdata.lkup_tx_rx.act =
 884		cpu_to_le32(ICE_SINGLE_ACT_PTR |
 885			    ((l_id << ICE_SINGLE_ACT_PTR_VAL_S) &
 886			     ICE_SINGLE_ACT_PTR_VAL_M));
 887
 888	/* Use the filter rule ID of the previously created rule with single
 889	 * act. Once the update happens, hardware will treat this as large
 890	 * action
 891	 */
 892	rx_tx->pdata.lkup_tx_rx.index =
 893		cpu_to_le16(m_ent->fltr_info.fltr_rule_id);
 894
 895	status = ice_aq_sw_rules(hw, lg_act, rules_size, 2,
 896				 ice_aqc_opc_update_sw_rules, NULL);
 897	if (!status) {
 898		m_ent->lg_act_idx = l_id;
 899		m_ent->sw_marker_id = sw_marker;
 900	}
 901
 902	devm_kfree(ice_hw_to_dev(hw), lg_act);
 903	return status;
 904}
 905
 906/**
 907 * ice_create_vsi_list_map
 908 * @hw: pointer to the hardware structure
 909 * @vsi_handle_arr: array of VSI handles to set in the VSI mapping
 910 * @num_vsi: number of VSI handles in the array
 911 * @vsi_list_id: VSI list ID generated as part of allocate resource
 912 *
 913 * Helper function to create a new entry of VSI list ID to VSI mapping
 914 * using the given VSI list ID
 915 */
 916static struct ice_vsi_list_map_info *
 917ice_create_vsi_list_map(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 918			u16 vsi_list_id)
 919{
 920	struct ice_switch_info *sw = hw->switch_info;
 921	struct ice_vsi_list_map_info *v_map;
 922	int i;
 923
 924	v_map = devm_kcalloc(ice_hw_to_dev(hw), 1, sizeof(*v_map), GFP_KERNEL);
 925	if (!v_map)
 926		return NULL;
 927
 928	v_map->vsi_list_id = vsi_list_id;
 929	v_map->ref_cnt = 1;
 930	for (i = 0; i < num_vsi; i++)
 931		set_bit(vsi_handle_arr[i], v_map->vsi_map);
 932
 933	list_add(&v_map->list_entry, &sw->vsi_list_map_head);
 934	return v_map;
 935}
 936
 937/**
 938 * ice_update_vsi_list_rule
 939 * @hw: pointer to the hardware structure
 940 * @vsi_handle_arr: array of VSI handles to form a VSI list
 941 * @num_vsi: number of VSI handles in the array
 942 * @vsi_list_id: VSI list ID generated as part of allocate resource
 943 * @remove: Boolean value to indicate if this is a remove action
 944 * @opc: switch rules population command type - pass in the command opcode
 945 * @lkup_type: lookup type of the filter
 946 *
 947 * Call AQ command to add a new switch rule or update existing switch rule
 948 * using the given VSI list ID
 949 */
 950static enum ice_status
 951ice_update_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
 952			 u16 vsi_list_id, bool remove, enum ice_adminq_opc opc,
 953			 enum ice_sw_lkup_type lkup_type)
 954{
 955	struct ice_aqc_sw_rules_elem *s_rule;
 956	enum ice_status status;
 957	u16 s_rule_size;
 958	u16 rule_type;
 959	int i;
 960
 961	if (!num_vsi)
 962		return ICE_ERR_PARAM;
 963
 964	if (lkup_type == ICE_SW_LKUP_MAC ||
 965	    lkup_type == ICE_SW_LKUP_MAC_VLAN ||
 966	    lkup_type == ICE_SW_LKUP_ETHERTYPE ||
 967	    lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
 968	    lkup_type == ICE_SW_LKUP_PROMISC ||
 969	    lkup_type == ICE_SW_LKUP_PROMISC_VLAN)
 970		rule_type = remove ? ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR :
 971			ICE_AQC_SW_RULES_T_VSI_LIST_SET;
 972	else if (lkup_type == ICE_SW_LKUP_VLAN)
 973		rule_type = remove ? ICE_AQC_SW_RULES_T_PRUNE_LIST_CLEAR :
 974			ICE_AQC_SW_RULES_T_PRUNE_LIST_SET;
 975	else
 976		return ICE_ERR_PARAM;
 977
 978	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(num_vsi);
 979	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
 980	if (!s_rule)
 981		return ICE_ERR_NO_MEMORY;
 982	for (i = 0; i < num_vsi; i++) {
 983		if (!ice_is_vsi_valid(hw, vsi_handle_arr[i])) {
 984			status = ICE_ERR_PARAM;
 985			goto exit;
 986		}
 987		/* AQ call requires hw_vsi_id(s) */
 988		s_rule->pdata.vsi_list.vsi[i] =
 989			cpu_to_le16(ice_get_hw_vsi_num(hw, vsi_handle_arr[i]));
 990	}
 991
 992	s_rule->type = cpu_to_le16(rule_type);
 993	s_rule->pdata.vsi_list.number_vsi = cpu_to_le16(num_vsi);
 994	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
 995
 996	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opc, NULL);
 997
 998exit:
 999	devm_kfree(ice_hw_to_dev(hw), s_rule);
1000	return status;
1001}
1002
1003/**
1004 * ice_create_vsi_list_rule - Creates and populates a VSI list rule
1005 * @hw: pointer to the HW struct
1006 * @vsi_handle_arr: array of VSI handles to form a VSI list
1007 * @num_vsi: number of VSI handles in the array
1008 * @vsi_list_id: stores the ID of the VSI list to be created
1009 * @lkup_type: switch rule filter's lookup type
1010 */
1011static enum ice_status
1012ice_create_vsi_list_rule(struct ice_hw *hw, u16 *vsi_handle_arr, u16 num_vsi,
1013			 u16 *vsi_list_id, enum ice_sw_lkup_type lkup_type)
1014{
1015	enum ice_status status;
1016
1017	status = ice_aq_alloc_free_vsi_list(hw, vsi_list_id, lkup_type,
1018					    ice_aqc_opc_alloc_res);
1019	if (status)
1020		return status;
1021
1022	/* Update the newly created VSI list to include the specified VSIs */
1023	return ice_update_vsi_list_rule(hw, vsi_handle_arr, num_vsi,
1024					*vsi_list_id, false,
1025					ice_aqc_opc_add_sw_rules, lkup_type);
1026}
1027
1028/**
1029 * ice_create_pkt_fwd_rule
1030 * @hw: pointer to the hardware structure
1031 * @f_entry: entry containing packet forwarding information
1032 *
1033 * Create switch rule with given filter information and add an entry
1034 * to the corresponding filter management list to track this switch rule
1035 * and VSI mapping
1036 */
1037static enum ice_status
1038ice_create_pkt_fwd_rule(struct ice_hw *hw,
1039			struct ice_fltr_list_entry *f_entry)
1040{
1041	struct ice_fltr_mgmt_list_entry *fm_entry;
1042	struct ice_aqc_sw_rules_elem *s_rule;
1043	enum ice_sw_lkup_type l_type;
1044	struct ice_sw_recipe *recp;
1045	enum ice_status status;
1046
1047	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1048			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1049	if (!s_rule)
1050		return ICE_ERR_NO_MEMORY;
1051	fm_entry = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*fm_entry),
1052				GFP_KERNEL);
1053	if (!fm_entry) {
1054		status = ICE_ERR_NO_MEMORY;
1055		goto ice_create_pkt_fwd_rule_exit;
1056	}
1057
1058	fm_entry->fltr_info = f_entry->fltr_info;
1059
1060	/* Initialize all the fields for the management entry */
1061	fm_entry->vsi_count = 1;
1062	fm_entry->lg_act_idx = ICE_INVAL_LG_ACT_INDEX;
1063	fm_entry->sw_marker_id = ICE_INVAL_SW_MARKER_ID;
1064	fm_entry->counter_index = ICE_INVAL_COUNTER_ID;
1065
1066	ice_fill_sw_rule(hw, &fm_entry->fltr_info, s_rule,
1067			 ice_aqc_opc_add_sw_rules);
1068
1069	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1070				 ice_aqc_opc_add_sw_rules, NULL);
1071	if (status) {
1072		devm_kfree(ice_hw_to_dev(hw), fm_entry);
1073		goto ice_create_pkt_fwd_rule_exit;
1074	}
1075
1076	f_entry->fltr_info.fltr_rule_id =
1077		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1078	fm_entry->fltr_info.fltr_rule_id =
1079		le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1080
1081	/* The book keeping entries will get removed when base driver
1082	 * calls remove filter AQ command
1083	 */
1084	l_type = fm_entry->fltr_info.lkup_type;
1085	recp = &hw->switch_info->recp_list[l_type];
1086	list_add(&fm_entry->list_entry, &recp->filt_rules);
1087
1088ice_create_pkt_fwd_rule_exit:
1089	devm_kfree(ice_hw_to_dev(hw), s_rule);
1090	return status;
1091}
1092
1093/**
1094 * ice_update_pkt_fwd_rule
1095 * @hw: pointer to the hardware structure
1096 * @f_info: filter information for switch rule
1097 *
1098 * Call AQ command to update a previously created switch rule with a
1099 * VSI list ID
1100 */
1101static enum ice_status
1102ice_update_pkt_fwd_rule(struct ice_hw *hw, struct ice_fltr_info *f_info)
1103{
1104	struct ice_aqc_sw_rules_elem *s_rule;
1105	enum ice_status status;
1106
1107	s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1108			      ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
1109	if (!s_rule)
1110		return ICE_ERR_NO_MEMORY;
1111
1112	ice_fill_sw_rule(hw, f_info, s_rule, ice_aqc_opc_update_sw_rules);
1113
1114	s_rule->pdata.lkup_tx_rx.index = cpu_to_le16(f_info->fltr_rule_id);
1115
1116	/* Update switch rule with new rule set to forward VSI list */
1117	status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1118				 ice_aqc_opc_update_sw_rules, NULL);
1119
1120	devm_kfree(ice_hw_to_dev(hw), s_rule);
1121	return status;
1122}
1123
1124/**
1125 * ice_update_sw_rule_bridge_mode
1126 * @hw: pointer to the HW struct
1127 *
1128 * Updates unicast switch filter rules based on VEB/VEPA mode
1129 */
1130enum ice_status ice_update_sw_rule_bridge_mode(struct ice_hw *hw)
1131{
1132	struct ice_switch_info *sw = hw->switch_info;
1133	struct ice_fltr_mgmt_list_entry *fm_entry;
1134	enum ice_status status = 0;
1135	struct list_head *rule_head;
1136	struct mutex *rule_lock; /* Lock to protect filter rule list */
1137
1138	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1139	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1140
1141	mutex_lock(rule_lock);
1142	list_for_each_entry(fm_entry, rule_head, list_entry) {
1143		struct ice_fltr_info *fi = &fm_entry->fltr_info;
1144		u8 *addr = fi->l_data.mac.mac_addr;
1145
1146		/* Update unicast Tx rules to reflect the selected
1147		 * VEB/VEPA mode
1148		 */
1149		if ((fi->flag & ICE_FLTR_TX) && is_unicast_ether_addr(addr) &&
1150		    (fi->fltr_act == ICE_FWD_TO_VSI ||
1151		     fi->fltr_act == ICE_FWD_TO_VSI_LIST ||
1152		     fi->fltr_act == ICE_FWD_TO_Q ||
1153		     fi->fltr_act == ICE_FWD_TO_QGRP)) {
1154			status = ice_update_pkt_fwd_rule(hw, fi);
1155			if (status)
1156				break;
1157		}
1158	}
1159
1160	mutex_unlock(rule_lock);
1161
1162	return status;
1163}
1164
1165/**
1166 * ice_add_update_vsi_list
1167 * @hw: pointer to the hardware structure
1168 * @m_entry: pointer to current filter management list entry
1169 * @cur_fltr: filter information from the book keeping entry
1170 * @new_fltr: filter information with the new VSI to be added
1171 *
1172 * Call AQ command to add or update previously created VSI list with new VSI.
1173 *
1174 * Helper function to do book keeping associated with adding filter information
1175 * The algorithm to do the book keeping is described below :
1176 * When a VSI needs to subscribe to a given filter (MAC/VLAN/Ethtype etc.)
1177 *	if only one VSI has been added till now
1178 *		Allocate a new VSI list and add two VSIs
1179 *		to this list using switch rule command
1180 *		Update the previously created switch rule with the
1181 *		newly created VSI list ID
1182 *	if a VSI list was previously created
1183 *		Add the new VSI to the previously created VSI list set
1184 *		using the update switch rule command
1185 */
1186static enum ice_status
1187ice_add_update_vsi_list(struct ice_hw *hw,
1188			struct ice_fltr_mgmt_list_entry *m_entry,
1189			struct ice_fltr_info *cur_fltr,
1190			struct ice_fltr_info *new_fltr)
1191{
1192	enum ice_status status = 0;
1193	u16 vsi_list_id = 0;
1194
1195	if ((cur_fltr->fltr_act == ICE_FWD_TO_Q ||
1196	     cur_fltr->fltr_act == ICE_FWD_TO_QGRP))
1197		return ICE_ERR_NOT_IMPL;
1198
1199	if ((new_fltr->fltr_act == ICE_FWD_TO_Q ||
1200	     new_fltr->fltr_act == ICE_FWD_TO_QGRP) &&
1201	    (cur_fltr->fltr_act == ICE_FWD_TO_VSI ||
1202	     cur_fltr->fltr_act == ICE_FWD_TO_VSI_LIST))
1203		return ICE_ERR_NOT_IMPL;
1204
1205	if (m_entry->vsi_count < 2 && !m_entry->vsi_list_info) {
1206		/* Only one entry existed in the mapping and it was not already
1207		 * a part of a VSI list. So, create a VSI list with the old and
1208		 * new VSIs.
1209		 */
1210		struct ice_fltr_info tmp_fltr;
1211		u16 vsi_handle_arr[2];
1212
1213		/* A rule already exists with the new VSI being added */
1214		if (cur_fltr->fwd_id.hw_vsi_id == new_fltr->fwd_id.hw_vsi_id)
1215			return ICE_ERR_ALREADY_EXISTS;
1216
1217		vsi_handle_arr[0] = cur_fltr->vsi_handle;
1218		vsi_handle_arr[1] = new_fltr->vsi_handle;
1219		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1220						  &vsi_list_id,
1221						  new_fltr->lkup_type);
1222		if (status)
1223			return status;
1224
1225		tmp_fltr = *new_fltr;
1226		tmp_fltr.fltr_rule_id = cur_fltr->fltr_rule_id;
1227		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1228		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1229		/* Update the previous switch rule of "MAC forward to VSI" to
1230		 * "MAC fwd to VSI list"
1231		 */
1232		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
1233		if (status)
1234			return status;
1235
1236		cur_fltr->fwd_id.vsi_list_id = vsi_list_id;
1237		cur_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1238		m_entry->vsi_list_info =
1239			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1240						vsi_list_id);
1241
1242		/* If this entry was large action then the large action needs
1243		 * to be updated to point to FWD to VSI list
1244		 */
1245		if (m_entry->sw_marker_id != ICE_INVAL_SW_MARKER_ID)
1246			status =
1247			    ice_add_marker_act(hw, m_entry,
1248					       m_entry->sw_marker_id,
1249					       m_entry->lg_act_idx);
1250	} else {
1251		u16 vsi_handle = new_fltr->vsi_handle;
1252		enum ice_adminq_opc opcode;
1253
1254		if (!m_entry->vsi_list_info)
1255			return ICE_ERR_CFG;
1256
1257		/* A rule already exists with the new VSI being added */
1258		if (test_bit(vsi_handle, m_entry->vsi_list_info->vsi_map))
1259			return 0;
1260
1261		/* Update the previously created VSI list set with
1262		 * the new VSI ID passed in
1263		 */
1264		vsi_list_id = cur_fltr->fwd_id.vsi_list_id;
1265		opcode = ice_aqc_opc_update_sw_rules;
1266
1267		status = ice_update_vsi_list_rule(hw, &vsi_handle, 1,
1268						  vsi_list_id, false, opcode,
1269						  new_fltr->lkup_type);
1270		/* update VSI list mapping info with new VSI ID */
1271		if (!status)
1272			set_bit(vsi_handle, m_entry->vsi_list_info->vsi_map);
1273	}
1274	if (!status)
1275		m_entry->vsi_count++;
1276	return status;
1277}
1278
1279/**
1280 * ice_find_rule_entry - Search a rule entry
1281 * @hw: pointer to the hardware structure
1282 * @recp_id: lookup type for which the specified rule needs to be searched
1283 * @f_info: rule information
1284 *
1285 * Helper function to search for a given rule entry
1286 * Returns pointer to entry storing the rule if found
1287 */
1288static struct ice_fltr_mgmt_list_entry *
1289ice_find_rule_entry(struct ice_hw *hw, u8 recp_id, struct ice_fltr_info *f_info)
1290{
1291	struct ice_fltr_mgmt_list_entry *list_itr, *ret = NULL;
1292	struct ice_switch_info *sw = hw->switch_info;
1293	struct list_head *list_head;
1294
1295	list_head = &sw->recp_list[recp_id].filt_rules;
1296	list_for_each_entry(list_itr, list_head, list_entry) {
1297		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
1298			    sizeof(f_info->l_data)) &&
1299		    f_info->flag == list_itr->fltr_info.flag) {
1300			ret = list_itr;
1301			break;
1302		}
1303	}
1304	return ret;
1305}
1306
1307/**
1308 * ice_find_vsi_list_entry - Search VSI list map with VSI count 1
1309 * @hw: pointer to the hardware structure
1310 * @recp_id: lookup type for which VSI lists needs to be searched
1311 * @vsi_handle: VSI handle to be found in VSI list
1312 * @vsi_list_id: VSI list ID found containing vsi_handle
1313 *
1314 * Helper function to search a VSI list with single entry containing given VSI
1315 * handle element. This can be extended further to search VSI list with more
1316 * than 1 vsi_count. Returns pointer to VSI list entry if found.
1317 */
1318static struct ice_vsi_list_map_info *
1319ice_find_vsi_list_entry(struct ice_hw *hw, u8 recp_id, u16 vsi_handle,
1320			u16 *vsi_list_id)
1321{
1322	struct ice_vsi_list_map_info *map_info = NULL;
1323	struct ice_switch_info *sw = hw->switch_info;
1324	struct ice_fltr_mgmt_list_entry *list_itr;
1325	struct list_head *list_head;
1326
1327	list_head = &sw->recp_list[recp_id].filt_rules;
1328	list_for_each_entry(list_itr, list_head, list_entry) {
1329		if (list_itr->vsi_count == 1 && list_itr->vsi_list_info) {
1330			map_info = list_itr->vsi_list_info;
1331			if (test_bit(vsi_handle, map_info->vsi_map)) {
1332				*vsi_list_id = map_info->vsi_list_id;
1333				return map_info;
1334			}
1335		}
1336	}
1337	return NULL;
1338}
1339
1340/**
1341 * ice_add_rule_internal - add rule for a given lookup type
1342 * @hw: pointer to the hardware structure
1343 * @recp_id: lookup type (recipe ID) for which rule has to be added
1344 * @f_entry: structure containing MAC forwarding information
1345 *
1346 * Adds or updates the rule lists for a given recipe
1347 */
1348static enum ice_status
1349ice_add_rule_internal(struct ice_hw *hw, u8 recp_id,
1350		      struct ice_fltr_list_entry *f_entry)
1351{
1352	struct ice_switch_info *sw = hw->switch_info;
1353	struct ice_fltr_info *new_fltr, *cur_fltr;
1354	struct ice_fltr_mgmt_list_entry *m_entry;
1355	struct mutex *rule_lock; /* Lock to protect filter rule list */
1356	enum ice_status status = 0;
1357
1358	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1359		return ICE_ERR_PARAM;
1360	f_entry->fltr_info.fwd_id.hw_vsi_id =
1361		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1362
1363	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1364
1365	mutex_lock(rule_lock);
1366	new_fltr = &f_entry->fltr_info;
1367	if (new_fltr->flag & ICE_FLTR_RX)
1368		new_fltr->src = hw->port_info->lport;
1369	else if (new_fltr->flag & ICE_FLTR_TX)
1370		new_fltr->src = f_entry->fltr_info.fwd_id.hw_vsi_id;
1371
1372	m_entry = ice_find_rule_entry(hw, recp_id, new_fltr);
1373	if (!m_entry) {
1374		mutex_unlock(rule_lock);
1375		return ice_create_pkt_fwd_rule(hw, f_entry);
1376	}
1377
1378	cur_fltr = &m_entry->fltr_info;
1379	status = ice_add_update_vsi_list(hw, m_entry, cur_fltr, new_fltr);
1380	mutex_unlock(rule_lock);
1381
1382	return status;
1383}
1384
1385/**
1386 * ice_remove_vsi_list_rule
1387 * @hw: pointer to the hardware structure
1388 * @vsi_list_id: VSI list ID generated as part of allocate resource
1389 * @lkup_type: switch rule filter lookup type
1390 *
1391 * The VSI list should be emptied before this function is called to remove the
1392 * VSI list.
1393 */
1394static enum ice_status
1395ice_remove_vsi_list_rule(struct ice_hw *hw, u16 vsi_list_id,
1396			 enum ice_sw_lkup_type lkup_type)
1397{
1398	struct ice_aqc_sw_rules_elem *s_rule;
1399	enum ice_status status;
1400	u16 s_rule_size;
1401
1402	s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(0);
1403	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1404	if (!s_rule)
1405		return ICE_ERR_NO_MEMORY;
1406
1407	s_rule->type = cpu_to_le16(ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR);
1408	s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1409
1410	/* Free the vsi_list resource that we allocated. It is assumed that the
1411	 * list is empty at this point.
1412	 */
1413	status = ice_aq_alloc_free_vsi_list(hw, &vsi_list_id, lkup_type,
1414					    ice_aqc_opc_free_res);
1415
1416	devm_kfree(ice_hw_to_dev(hw), s_rule);
1417	return status;
1418}
1419
1420/**
1421 * ice_rem_update_vsi_list
1422 * @hw: pointer to the hardware structure
1423 * @vsi_handle: VSI handle of the VSI to remove
1424 * @fm_list: filter management entry for which the VSI list management needs to
1425 *           be done
1426 */
1427static enum ice_status
1428ice_rem_update_vsi_list(struct ice_hw *hw, u16 vsi_handle,
1429			struct ice_fltr_mgmt_list_entry *fm_list)
1430{
1431	enum ice_sw_lkup_type lkup_type;
1432	enum ice_status status = 0;
1433	u16 vsi_list_id;
1434
1435	if (fm_list->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST ||
1436	    fm_list->vsi_count == 0)
1437		return ICE_ERR_PARAM;
1438
1439	/* A rule with the VSI being removed does not exist */
1440	if (!test_bit(vsi_handle, fm_list->vsi_list_info->vsi_map))
1441		return ICE_ERR_DOES_NOT_EXIST;
1442
1443	lkup_type = fm_list->fltr_info.lkup_type;
1444	vsi_list_id = fm_list->fltr_info.fwd_id.vsi_list_id;
1445	status = ice_update_vsi_list_rule(hw, &vsi_handle, 1, vsi_list_id, true,
1446					  ice_aqc_opc_update_sw_rules,
1447					  lkup_type);
1448	if (status)
1449		return status;
1450
1451	fm_list->vsi_count--;
1452	clear_bit(vsi_handle, fm_list->vsi_list_info->vsi_map);
1453
1454	if (fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) {
1455		struct ice_fltr_info tmp_fltr_info = fm_list->fltr_info;
1456		struct ice_vsi_list_map_info *vsi_list_info =
1457			fm_list->vsi_list_info;
1458		u16 rem_vsi_handle;
1459
1460		rem_vsi_handle = find_first_bit(vsi_list_info->vsi_map,
1461						ICE_MAX_VSI);
1462		if (!ice_is_vsi_valid(hw, rem_vsi_handle))
1463			return ICE_ERR_OUT_OF_RANGE;
1464
1465		/* Make sure VSI list is empty before removing it below */
1466		status = ice_update_vsi_list_rule(hw, &rem_vsi_handle, 1,
1467						  vsi_list_id, true,
1468						  ice_aqc_opc_update_sw_rules,
1469						  lkup_type);
1470		if (status)
1471			return status;
1472
1473		tmp_fltr_info.fltr_act = ICE_FWD_TO_VSI;
1474		tmp_fltr_info.fwd_id.hw_vsi_id =
1475			ice_get_hw_vsi_num(hw, rem_vsi_handle);
1476		tmp_fltr_info.vsi_handle = rem_vsi_handle;
1477		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr_info);
1478		if (status) {
1479			ice_debug(hw, ICE_DBG_SW,
1480				  "Failed to update pkt fwd rule to FWD_TO_VSI on HW VSI %d, error %d\n",
1481				  tmp_fltr_info.fwd_id.hw_vsi_id, status);
1482			return status;
1483		}
1484
1485		fm_list->fltr_info = tmp_fltr_info;
1486	}
1487
1488	if ((fm_list->vsi_count == 1 && lkup_type != ICE_SW_LKUP_VLAN) ||
1489	    (fm_list->vsi_count == 0 && lkup_type == ICE_SW_LKUP_VLAN)) {
1490		struct ice_vsi_list_map_info *vsi_list_info =
1491			fm_list->vsi_list_info;
1492
1493		/* Remove the VSI list since it is no longer used */
1494		status = ice_remove_vsi_list_rule(hw, vsi_list_id, lkup_type);
1495		if (status) {
1496			ice_debug(hw, ICE_DBG_SW,
1497				  "Failed to remove VSI list %d, error %d\n",
1498				  vsi_list_id, status);
1499			return status;
1500		}
1501
1502		list_del(&vsi_list_info->list_entry);
1503		devm_kfree(ice_hw_to_dev(hw), vsi_list_info);
1504		fm_list->vsi_list_info = NULL;
1505	}
1506
1507	return status;
1508}
1509
1510/**
1511 * ice_remove_rule_internal - Remove a filter rule of a given type
1512 * @hw: pointer to the hardware structure
1513 * @recp_id: recipe ID for which the rule needs to removed
1514 * @f_entry: rule entry containing filter information
1515 */
1516static enum ice_status
1517ice_remove_rule_internal(struct ice_hw *hw, u8 recp_id,
1518			 struct ice_fltr_list_entry *f_entry)
1519{
1520	struct ice_switch_info *sw = hw->switch_info;
1521	struct ice_fltr_mgmt_list_entry *list_elem;
1522	struct mutex *rule_lock; /* Lock to protect filter rule list */
1523	enum ice_status status = 0;
1524	bool remove_rule = false;
1525	u16 vsi_handle;
1526
1527	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1528		return ICE_ERR_PARAM;
1529	f_entry->fltr_info.fwd_id.hw_vsi_id =
1530		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1531
1532	rule_lock = &sw->recp_list[recp_id].filt_rule_lock;
1533	mutex_lock(rule_lock);
1534	list_elem = ice_find_rule_entry(hw, recp_id, &f_entry->fltr_info);
1535	if (!list_elem) {
1536		status = ICE_ERR_DOES_NOT_EXIST;
1537		goto exit;
1538	}
1539
1540	if (list_elem->fltr_info.fltr_act != ICE_FWD_TO_VSI_LIST) {
1541		remove_rule = true;
1542	} else if (!list_elem->vsi_list_info) {
1543		status = ICE_ERR_DOES_NOT_EXIST;
1544		goto exit;
1545	} else if (list_elem->vsi_list_info->ref_cnt > 1) {
1546		/* a ref_cnt > 1 indicates that the vsi_list is being
1547		 * shared by multiple rules. Decrement the ref_cnt and
1548		 * remove this rule, but do not modify the list, as it
1549		 * is in-use by other rules.
1550		 */
1551		list_elem->vsi_list_info->ref_cnt--;
1552		remove_rule = true;
1553	} else {
1554		/* a ref_cnt of 1 indicates the vsi_list is only used
1555		 * by one rule. However, the original removal request is only
1556		 * for a single VSI. Update the vsi_list first, and only
1557		 * remove the rule if there are no further VSIs in this list.
1558		 */
1559		vsi_handle = f_entry->fltr_info.vsi_handle;
1560		status = ice_rem_update_vsi_list(hw, vsi_handle, list_elem);
1561		if (status)
1562			goto exit;
1563		/* if VSI count goes to zero after updating the VSI list */
1564		if (list_elem->vsi_count == 0)
1565			remove_rule = true;
1566	}
1567
1568	if (remove_rule) {
1569		/* Remove the lookup rule */
1570		struct ice_aqc_sw_rules_elem *s_rule;
1571
1572		s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1573				      ICE_SW_RULE_RX_TX_NO_HDR_SIZE,
1574				      GFP_KERNEL);
1575		if (!s_rule) {
1576			status = ICE_ERR_NO_MEMORY;
1577			goto exit;
1578		}
1579
1580		ice_fill_sw_rule(hw, &list_elem->fltr_info, s_rule,
1581				 ice_aqc_opc_remove_sw_rules);
1582
1583		status = ice_aq_sw_rules(hw, s_rule,
1584					 ICE_SW_RULE_RX_TX_NO_HDR_SIZE, 1,
1585					 ice_aqc_opc_remove_sw_rules, NULL);
1586
1587		/* Remove a book keeping from the list */
1588		devm_kfree(ice_hw_to_dev(hw), s_rule);
1589
1590		if (status)
1591			goto exit;
1592
1593		list_del(&list_elem->list_entry);
1594		devm_kfree(ice_hw_to_dev(hw), list_elem);
1595	}
1596exit:
1597	mutex_unlock(rule_lock);
1598	return status;
1599}
1600
1601/**
1602 * ice_add_mac - Add a MAC address based filter rule
1603 * @hw: pointer to the hardware structure
1604 * @m_list: list of MAC addresses and forwarding information
1605 *
1606 * IMPORTANT: When the ucast_shared flag is set to false and m_list has
1607 * multiple unicast addresses, the function assumes that all the
1608 * addresses are unique in a given add_mac call. It doesn't
1609 * check for duplicates in this case, removing duplicates from a given
1610 * list should be taken care of in the caller of this function.
1611 */
1612enum ice_status ice_add_mac(struct ice_hw *hw, struct list_head *m_list)
 
1613{
1614	struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1615	struct ice_fltr_list_entry *m_list_itr;
1616	struct list_head *rule_head;
1617	u16 total_elem_left, s_rule_size;
1618	struct ice_switch_info *sw;
1619	struct mutex *rule_lock; /* Lock to protect filter rule list */
1620	enum ice_status status = 0;
1621	u16 num_unicast = 0;
1622	u8 elem_sent;
1623
1624	if (!m_list || !hw)
1625		return ICE_ERR_PARAM;
1626
1627	s_rule = NULL;
1628	sw = hw->switch_info;
1629	rule_lock = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
1630	list_for_each_entry(m_list_itr, m_list, list_entry) {
1631		u8 *add = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1632		u16 vsi_handle;
1633		u16 hw_vsi_id;
1634
1635		m_list_itr->fltr_info.flag = ICE_FLTR_TX;
1636		vsi_handle = m_list_itr->fltr_info.vsi_handle;
1637		if (!ice_is_vsi_valid(hw, vsi_handle))
1638			return ICE_ERR_PARAM;
1639		hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
1640		m_list_itr->fltr_info.fwd_id.hw_vsi_id = hw_vsi_id;
1641		/* update the src in case it is VSI num */
1642		if (m_list_itr->fltr_info.src_id != ICE_SRC_ID_VSI)
1643			return ICE_ERR_PARAM;
1644		m_list_itr->fltr_info.src = hw_vsi_id;
1645		if (m_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_MAC ||
1646		    is_zero_ether_addr(add))
1647			return ICE_ERR_PARAM;
1648		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
1649			/* Don't overwrite the unicast address */
1650			mutex_lock(rule_lock);
1651			if (ice_find_rule_entry(hw, ICE_SW_LKUP_MAC,
1652						&m_list_itr->fltr_info)) {
1653				mutex_unlock(rule_lock);
1654				return ICE_ERR_ALREADY_EXISTS;
1655			}
1656			mutex_unlock(rule_lock);
1657			num_unicast++;
1658		} else if (is_multicast_ether_addr(add) ||
1659			   (is_unicast_ether_addr(add) && hw->ucast_shared)) {
1660			m_list_itr->status =
1661				ice_add_rule_internal(hw, ICE_SW_LKUP_MAC,
1662						      m_list_itr);
1663			if (m_list_itr->status)
1664				return m_list_itr->status;
1665		}
1666	}
1667
1668	mutex_lock(rule_lock);
1669	/* Exit if no suitable entries were found for adding bulk switch rule */
1670	if (!num_unicast) {
1671		status = 0;
1672		goto ice_add_mac_exit;
1673	}
1674
1675	rule_head = &sw->recp_list[ICE_SW_LKUP_MAC].filt_rules;
1676
1677	/* Allocate switch rule buffer for the bulk update for unicast */
1678	s_rule_size = ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
1679	s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1680			      GFP_KERNEL);
1681	if (!s_rule) {
1682		status = ICE_ERR_NO_MEMORY;
1683		goto ice_add_mac_exit;
1684	}
1685
1686	r_iter = s_rule;
1687	list_for_each_entry(m_list_itr, m_list, list_entry) {
1688		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1689		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
1690
1691		if (is_unicast_ether_addr(mac_addr)) {
1692			ice_fill_sw_rule(hw, &m_list_itr->fltr_info, r_iter,
1693					 ice_aqc_opc_add_sw_rules);
1694			r_iter = (struct ice_aqc_sw_rules_elem *)
1695				((u8 *)r_iter + s_rule_size);
1696		}
1697	}
1698
1699	/* Call AQ bulk switch rule update for all unicast addresses */
1700	r_iter = s_rule;
1701	/* Call AQ switch rule in AQ_MAX chunk */
1702	for (total_elem_left = num_unicast; total_elem_left > 0;
1703	     total_elem_left -= elem_sent) {
1704		struct ice_aqc_sw_rules_elem *entry = r_iter;
1705
1706		elem_sent = min_t(u8, total_elem_left,
1707				  (ICE_AQ_MAX_BUF_LEN / s_rule_size));
1708		status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1709					 elem_sent, ice_aqc_opc_add_sw_rules,
1710					 NULL);
1711		if (status)
1712			goto ice_add_mac_exit;
1713		r_iter = (struct ice_aqc_sw_rules_elem *)
1714			((u8 *)r_iter + (elem_sent * s_rule_size));
1715	}
1716
1717	/* Fill up rule ID based on the value returned from FW */
1718	r_iter = s_rule;
1719	list_for_each_entry(m_list_itr, m_list, list_entry) {
1720		struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1721		u8 *mac_addr = &f_info->l_data.mac.mac_addr[0];
1722		struct ice_fltr_mgmt_list_entry *fm_entry;
1723
1724		if (is_unicast_ether_addr(mac_addr)) {
1725			f_info->fltr_rule_id =
1726				le16_to_cpu(r_iter->pdata.lkup_tx_rx.index);
1727			f_info->fltr_act = ICE_FWD_TO_VSI;
1728			/* Create an entry to track this MAC address */
1729			fm_entry = devm_kzalloc(ice_hw_to_dev(hw),
1730						sizeof(*fm_entry), GFP_KERNEL);
1731			if (!fm_entry) {
1732				status = ICE_ERR_NO_MEMORY;
1733				goto ice_add_mac_exit;
1734			}
1735			fm_entry->fltr_info = *f_info;
1736			fm_entry->vsi_count = 1;
1737			/* The book keeping entries will get removed when
1738			 * base driver calls remove filter AQ command
1739			 */
1740
1741			list_add(&fm_entry->list_entry, rule_head);
1742			r_iter = (struct ice_aqc_sw_rules_elem *)
1743				((u8 *)r_iter + s_rule_size);
1744		}
1745	}
1746
1747ice_add_mac_exit:
1748	mutex_unlock(rule_lock);
1749	if (s_rule)
1750		devm_kfree(ice_hw_to_dev(hw), s_rule);
1751	return status;
1752}
1753
1754/**
1755 * ice_add_vlan_internal - Add one VLAN based filter rule
1756 * @hw: pointer to the hardware structure
1757 * @f_entry: filter entry containing one VLAN information
1758 */
1759static enum ice_status
1760ice_add_vlan_internal(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1761{
1762	struct ice_switch_info *sw = hw->switch_info;
1763	struct ice_fltr_mgmt_list_entry *v_list_itr;
1764	struct ice_fltr_info *new_fltr, *cur_fltr;
1765	enum ice_sw_lkup_type lkup_type;
1766	u16 vsi_list_id = 0, vsi_handle;
1767	struct mutex *rule_lock; /* Lock to protect filter rule list */
1768	enum ice_status status = 0;
1769
1770	if (!ice_is_vsi_valid(hw, f_entry->fltr_info.vsi_handle))
1771		return ICE_ERR_PARAM;
1772
1773	f_entry->fltr_info.fwd_id.hw_vsi_id =
1774		ice_get_hw_vsi_num(hw, f_entry->fltr_info.vsi_handle);
1775	new_fltr = &f_entry->fltr_info;
1776
1777	/* VLAN ID should only be 12 bits */
1778	if (new_fltr->l_data.vlan.vlan_id > ICE_MAX_VLAN_ID)
1779		return ICE_ERR_PARAM;
1780
1781	if (new_fltr->src_id != ICE_SRC_ID_VSI)
1782		return ICE_ERR_PARAM;
1783
1784	new_fltr->src = new_fltr->fwd_id.hw_vsi_id;
1785	lkup_type = new_fltr->lkup_type;
1786	vsi_handle = new_fltr->vsi_handle;
1787	rule_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
1788	mutex_lock(rule_lock);
1789	v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN, new_fltr);
1790	if (!v_list_itr) {
1791		struct ice_vsi_list_map_info *map_info = NULL;
1792
1793		if (new_fltr->fltr_act == ICE_FWD_TO_VSI) {
1794			/* All VLAN pruning rules use a VSI list. Check if
1795			 * there is already a VSI list containing VSI that we
1796			 * want to add. If found, use the same vsi_list_id for
1797			 * this new VLAN rule or else create a new list.
1798			 */
1799			map_info = ice_find_vsi_list_entry(hw, ICE_SW_LKUP_VLAN,
1800							   vsi_handle,
1801							   &vsi_list_id);
1802			if (!map_info) {
1803				status = ice_create_vsi_list_rule(hw,
1804								  &vsi_handle,
1805								  1,
1806								  &vsi_list_id,
1807								  lkup_type);
1808				if (status)
1809					goto exit;
1810			}
1811			/* Convert the action to forwarding to a VSI list. */
1812			new_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1813			new_fltr->fwd_id.vsi_list_id = vsi_list_id;
1814		}
1815
1816		status = ice_create_pkt_fwd_rule(hw, f_entry);
1817		if (!status) {
1818			v_list_itr = ice_find_rule_entry(hw, ICE_SW_LKUP_VLAN,
1819							 new_fltr);
1820			if (!v_list_itr) {
1821				status = ICE_ERR_DOES_NOT_EXIST;
1822				goto exit;
1823			}
1824			/* reuse VSI list for new rule and increment ref_cnt */
1825			if (map_info) {
1826				v_list_itr->vsi_list_info = map_info;
1827				map_info->ref_cnt++;
1828			} else {
1829				v_list_itr->vsi_list_info =
1830					ice_create_vsi_list_map(hw, &vsi_handle,
1831								1, vsi_list_id);
1832			}
1833		}
1834	} else if (v_list_itr->vsi_list_info->ref_cnt == 1) {
1835		/* Update existing VSI list to add new VSI ID only if it used
1836		 * by one VLAN rule.
1837		 */
1838		cur_fltr = &v_list_itr->fltr_info;
1839		status = ice_add_update_vsi_list(hw, v_list_itr, cur_fltr,
1840						 new_fltr);
1841	} else {
1842		/* If VLAN rule exists and VSI list being used by this rule is
1843		 * referenced by more than 1 VLAN rule. Then create a new VSI
1844		 * list appending previous VSI with new VSI and update existing
1845		 * VLAN rule to point to new VSI list ID
1846		 */
1847		struct ice_fltr_info tmp_fltr;
1848		u16 vsi_handle_arr[2];
1849		u16 cur_handle;
1850
1851		/* Current implementation only supports reusing VSI list with
1852		 * one VSI count. We should never hit below condition
1853		 */
1854		if (v_list_itr->vsi_count > 1 &&
1855		    v_list_itr->vsi_list_info->ref_cnt > 1) {
1856			ice_debug(hw, ICE_DBG_SW,
1857				  "Invalid configuration: Optimization to reuse VSI list with more than one VSI is not being done yet\n");
1858			status = ICE_ERR_CFG;
1859			goto exit;
1860		}
1861
1862		cur_handle =
1863			find_first_bit(v_list_itr->vsi_list_info->vsi_map,
1864				       ICE_MAX_VSI);
1865
1866		/* A rule already exists with the new VSI being added */
1867		if (cur_handle == vsi_handle) {
1868			status = ICE_ERR_ALREADY_EXISTS;
1869			goto exit;
1870		}
1871
1872		vsi_handle_arr[0] = cur_handle;
1873		vsi_handle_arr[1] = vsi_handle;
1874		status = ice_create_vsi_list_rule(hw, &vsi_handle_arr[0], 2,
1875						  &vsi_list_id, lkup_type);
1876		if (status)
1877			goto exit;
1878
1879		tmp_fltr = v_list_itr->fltr_info;
1880		tmp_fltr.fltr_rule_id = v_list_itr->fltr_info.fltr_rule_id;
1881		tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
1882		tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
1883		/* Update the previous switch rule to a new VSI list which
1884		 * includes current VSI that is requested
1885		 */
1886		status = ice_update_pkt_fwd_rule(hw, &tmp_fltr);
1887		if (status)
1888			goto exit;
1889
1890		/* before overriding VSI list map info. decrement ref_cnt of
1891		 * previous VSI list
1892		 */
1893		v_list_itr->vsi_list_info->ref_cnt--;
1894
1895		/* now update to newly created list */
1896		v_list_itr->fltr_info.fwd_id.vsi_list_id = vsi_list_id;
1897		v_list_itr->vsi_list_info =
1898			ice_create_vsi_list_map(hw, &vsi_handle_arr[0], 2,
1899						vsi_list_id);
1900		v_list_itr->vsi_count++;
1901	}
1902
1903exit:
1904	mutex_unlock(rule_lock);
1905	return status;
1906}
1907
1908/**
1909 * ice_add_vlan - Add VLAN based filter rule
1910 * @hw: pointer to the hardware structure
1911 * @v_list: list of VLAN entries and forwarding information
1912 */
1913enum ice_status ice_add_vlan(struct ice_hw *hw, struct list_head *v_list)
 
1914{
1915	struct ice_fltr_list_entry *v_list_itr;
1916
1917	if (!v_list || !hw)
1918		return ICE_ERR_PARAM;
1919
1920	list_for_each_entry(v_list_itr, v_list, list_entry) {
1921		if (v_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_VLAN)
1922			return ICE_ERR_PARAM;
1923		v_list_itr->fltr_info.flag = ICE_FLTR_TX;
1924		v_list_itr->status = ice_add_vlan_internal(hw, v_list_itr);
1925		if (v_list_itr->status)
1926			return v_list_itr->status;
1927	}
1928	return 0;
1929}
1930
1931/**
1932 * ice_add_eth_mac - Add ethertype and MAC based filter rule
1933 * @hw: pointer to the hardware structure
1934 * @em_list: list of ether type MAC filter, MAC is optional
1935 *
1936 * This function requires the caller to populate the entries in
1937 * the filter list with the necessary fields (including flags to
1938 * indicate Tx or Rx rules).
1939 */
1940enum ice_status
1941ice_add_eth_mac(struct ice_hw *hw, struct list_head *em_list)
1942{
1943	struct ice_fltr_list_entry *em_list_itr;
1944
1945	if (!em_list || !hw)
1946		return ICE_ERR_PARAM;
1947
1948	list_for_each_entry(em_list_itr, em_list, list_entry) {
1949		enum ice_sw_lkup_type l_type =
1950			em_list_itr->fltr_info.lkup_type;
1951
1952		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
1953		    l_type != ICE_SW_LKUP_ETHERTYPE)
1954			return ICE_ERR_PARAM;
1955
1956		em_list_itr->status = ice_add_rule_internal(hw, l_type,
1957							    em_list_itr);
1958		if (em_list_itr->status)
1959			return em_list_itr->status;
1960	}
1961	return 0;
1962}
1963
1964/**
1965 * ice_remove_eth_mac - Remove an ethertype (or MAC) based filter rule
1966 * @hw: pointer to the hardware structure
1967 * @em_list: list of ethertype or ethertype MAC entries
1968 */
1969enum ice_status
1970ice_remove_eth_mac(struct ice_hw *hw, struct list_head *em_list)
1971{
1972	struct ice_fltr_list_entry *em_list_itr, *tmp;
1973
1974	if (!em_list || !hw)
1975		return ICE_ERR_PARAM;
1976
1977	list_for_each_entry_safe(em_list_itr, tmp, em_list, list_entry) {
1978		enum ice_sw_lkup_type l_type =
1979			em_list_itr->fltr_info.lkup_type;
1980
1981		if (l_type != ICE_SW_LKUP_ETHERTYPE_MAC &&
1982		    l_type != ICE_SW_LKUP_ETHERTYPE)
1983			return ICE_ERR_PARAM;
1984
1985		em_list_itr->status = ice_remove_rule_internal(hw, l_type,
1986							       em_list_itr);
1987		if (em_list_itr->status)
1988			return em_list_itr->status;
1989	}
1990	return 0;
1991}
1992
1993/**
1994 * ice_rem_sw_rule_info
1995 * @hw: pointer to the hardware structure
1996 * @rule_head: pointer to the switch list structure that we want to delete
1997 */
1998static void
1999ice_rem_sw_rule_info(struct ice_hw *hw, struct list_head *rule_head)
2000{
2001	if (!list_empty(rule_head)) {
2002		struct ice_fltr_mgmt_list_entry *entry;
2003		struct ice_fltr_mgmt_list_entry *tmp;
2004
2005		list_for_each_entry_safe(entry, tmp, rule_head, list_entry) {
2006			list_del(&entry->list_entry);
2007			devm_kfree(ice_hw_to_dev(hw), entry);
2008		}
2009	}
2010}
2011
2012/**
2013 * ice_cfg_dflt_vsi - change state of VSI to set/clear default
2014 * @hw: pointer to the hardware structure
2015 * @vsi_handle: VSI handle to set as default
2016 * @set: true to add the above mentioned switch rule, false to remove it
2017 * @direction: ICE_FLTR_RX or ICE_FLTR_TX
2018 *
2019 * add filter rule to set/unset given VSI as default VSI for the switch
2020 * (represented by swid)
2021 */
2022enum ice_status
2023ice_cfg_dflt_vsi(struct ice_hw *hw, u16 vsi_handle, bool set, u8 direction)
2024{
2025	struct ice_aqc_sw_rules_elem *s_rule;
2026	struct ice_fltr_info f_info;
2027	enum ice_adminq_opc opcode;
2028	enum ice_status status;
2029	u16 s_rule_size;
2030	u16 hw_vsi_id;
2031
2032	if (!ice_is_vsi_valid(hw, vsi_handle))
2033		return ICE_ERR_PARAM;
2034	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2035
2036	s_rule_size = set ? ICE_SW_RULE_RX_TX_ETH_HDR_SIZE :
2037		ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
2038
2039	s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
2040	if (!s_rule)
2041		return ICE_ERR_NO_MEMORY;
2042
2043	memset(&f_info, 0, sizeof(f_info));
2044
2045	f_info.lkup_type = ICE_SW_LKUP_DFLT;
2046	f_info.flag = direction;
2047	f_info.fltr_act = ICE_FWD_TO_VSI;
2048	f_info.fwd_id.hw_vsi_id = hw_vsi_id;
2049
2050	if (f_info.flag & ICE_FLTR_RX) {
2051		f_info.src = hw->port_info->lport;
2052		f_info.src_id = ICE_SRC_ID_LPORT;
2053		if (!set)
2054			f_info.fltr_rule_id =
2055				hw->port_info->dflt_rx_vsi_rule_id;
2056	} else if (f_info.flag & ICE_FLTR_TX) {
2057		f_info.src_id = ICE_SRC_ID_VSI;
2058		f_info.src = hw_vsi_id;
2059		if (!set)
2060			f_info.fltr_rule_id =
2061				hw->port_info->dflt_tx_vsi_rule_id;
2062	}
2063
2064	if (set)
2065		opcode = ice_aqc_opc_add_sw_rules;
2066	else
2067		opcode = ice_aqc_opc_remove_sw_rules;
2068
2069	ice_fill_sw_rule(hw, &f_info, s_rule, opcode);
2070
2071	status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opcode, NULL);
2072	if (status || !(f_info.flag & ICE_FLTR_TX_RX))
2073		goto out;
2074	if (set) {
2075		u16 index = le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
2076
2077		if (f_info.flag & ICE_FLTR_TX) {
2078			hw->port_info->dflt_tx_vsi_num = hw_vsi_id;
2079			hw->port_info->dflt_tx_vsi_rule_id = index;
2080		} else if (f_info.flag & ICE_FLTR_RX) {
2081			hw->port_info->dflt_rx_vsi_num = hw_vsi_id;
2082			hw->port_info->dflt_rx_vsi_rule_id = index;
2083		}
2084	} else {
2085		if (f_info.flag & ICE_FLTR_TX) {
2086			hw->port_info->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
2087			hw->port_info->dflt_tx_vsi_rule_id = ICE_INVAL_ACT;
2088		} else if (f_info.flag & ICE_FLTR_RX) {
2089			hw->port_info->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
2090			hw->port_info->dflt_rx_vsi_rule_id = ICE_INVAL_ACT;
2091		}
2092	}
2093
2094out:
2095	devm_kfree(ice_hw_to_dev(hw), s_rule);
2096	return status;
2097}
2098
2099/**
2100 * ice_find_ucast_rule_entry - Search for a unicast MAC filter rule entry
2101 * @hw: pointer to the hardware structure
2102 * @recp_id: lookup type for which the specified rule needs to be searched
2103 * @f_info: rule information
2104 *
2105 * Helper function to search for a unicast rule entry - this is to be used
2106 * to remove unicast MAC filter that is not shared with other VSIs on the
2107 * PF switch.
2108 *
2109 * Returns pointer to entry storing the rule if found
2110 */
2111static struct ice_fltr_mgmt_list_entry *
2112ice_find_ucast_rule_entry(struct ice_hw *hw, u8 recp_id,
2113			  struct ice_fltr_info *f_info)
2114{
2115	struct ice_switch_info *sw = hw->switch_info;
2116	struct ice_fltr_mgmt_list_entry *list_itr;
2117	struct list_head *list_head;
2118
2119	list_head = &sw->recp_list[recp_id].filt_rules;
2120	list_for_each_entry(list_itr, list_head, list_entry) {
2121		if (!memcmp(&f_info->l_data, &list_itr->fltr_info.l_data,
2122			    sizeof(f_info->l_data)) &&
2123		    f_info->fwd_id.hw_vsi_id ==
2124		    list_itr->fltr_info.fwd_id.hw_vsi_id &&
2125		    f_info->flag == list_itr->fltr_info.flag)
2126			return list_itr;
2127	}
2128	return NULL;
2129}
2130
2131/**
2132 * ice_remove_mac - remove a MAC address based filter rule
2133 * @hw: pointer to the hardware structure
2134 * @m_list: list of MAC addresses and forwarding information
2135 *
2136 * This function removes either a MAC filter rule or a specific VSI from a
2137 * VSI list for a multicast MAC address.
2138 *
2139 * Returns ICE_ERR_DOES_NOT_EXIST if a given entry was not added by
2140 * ice_add_mac. Caller should be aware that this call will only work if all
2141 * the entries passed into m_list were added previously. It will not attempt to
2142 * do a partial remove of entries that were found.
2143 */
2144enum ice_status ice_remove_mac(struct ice_hw *hw, struct list_head *m_list)
 
2145{
2146	struct ice_fltr_list_entry *list_itr, *tmp;
2147	struct mutex *rule_lock; /* Lock to protect filter rule list */
2148
2149	if (!m_list)
2150		return ICE_ERR_PARAM;
2151
2152	rule_lock = &hw->switch_info->recp_list[ICE_SW_LKUP_MAC].filt_rule_lock;
2153	list_for_each_entry_safe(list_itr, tmp, m_list, list_entry) {
2154		enum ice_sw_lkup_type l_type = list_itr->fltr_info.lkup_type;
2155		u8 *add = &list_itr->fltr_info.l_data.mac.mac_addr[0];
2156		u16 vsi_handle;
2157
2158		if (l_type != ICE_SW_LKUP_MAC)
2159			return ICE_ERR_PARAM;
2160
2161		vsi_handle = list_itr->fltr_info.vsi_handle;
2162		if (!ice_is_vsi_valid(hw, vsi_handle))
2163			return ICE_ERR_PARAM;
2164
2165		list_itr->fltr_info.fwd_id.hw_vsi_id =
2166					ice_get_hw_vsi_num(hw, vsi_handle);
2167		if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
2168			/* Don't remove the unicast address that belongs to
2169			 * another VSI on the switch, since it is not being
2170			 * shared...
2171			 */
2172			mutex_lock(rule_lock);
2173			if (!ice_find_ucast_rule_entry(hw, ICE_SW_LKUP_MAC,
2174						       &list_itr->fltr_info)) {
2175				mutex_unlock(rule_lock);
2176				return ICE_ERR_DOES_NOT_EXIST;
2177			}
2178			mutex_unlock(rule_lock);
2179		}
2180		list_itr->status = ice_remove_rule_internal(hw,
2181							    ICE_SW_LKUP_MAC,
2182							    list_itr);
2183		if (list_itr->status)
2184			return list_itr->status;
2185	}
2186	return 0;
2187}
2188
2189/**
2190 * ice_remove_vlan - Remove VLAN based filter rule
2191 * @hw: pointer to the hardware structure
2192 * @v_list: list of VLAN entries and forwarding information
2193 */
2194enum ice_status
2195ice_remove_vlan(struct ice_hw *hw, struct list_head *v_list)
2196{
2197	struct ice_fltr_list_entry *v_list_itr, *tmp;
2198
2199	if (!v_list || !hw)
2200		return ICE_ERR_PARAM;
2201
2202	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2203		enum ice_sw_lkup_type l_type = v_list_itr->fltr_info.lkup_type;
2204
2205		if (l_type != ICE_SW_LKUP_VLAN)
2206			return ICE_ERR_PARAM;
2207		v_list_itr->status = ice_remove_rule_internal(hw,
2208							      ICE_SW_LKUP_VLAN,
2209							      v_list_itr);
2210		if (v_list_itr->status)
2211			return v_list_itr->status;
2212	}
2213	return 0;
2214}
2215
2216/**
2217 * ice_vsi_uses_fltr - Determine if given VSI uses specified filter
2218 * @fm_entry: filter entry to inspect
2219 * @vsi_handle: VSI handle to compare with filter info
2220 */
2221static bool
2222ice_vsi_uses_fltr(struct ice_fltr_mgmt_list_entry *fm_entry, u16 vsi_handle)
2223{
2224	return ((fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI &&
2225		 fm_entry->fltr_info.vsi_handle == vsi_handle) ||
2226		(fm_entry->fltr_info.fltr_act == ICE_FWD_TO_VSI_LIST &&
2227		 (test_bit(vsi_handle, fm_entry->vsi_list_info->vsi_map))));
2228}
2229
2230/**
2231 * ice_add_entry_to_vsi_fltr_list - Add copy of fltr_list_entry to remove list
2232 * @hw: pointer to the hardware structure
2233 * @vsi_handle: VSI handle to remove filters from
2234 * @vsi_list_head: pointer to the list to add entry to
2235 * @fi: pointer to fltr_info of filter entry to copy & add
2236 *
2237 * Helper function, used when creating a list of filters to remove from
2238 * a specific VSI. The entry added to vsi_list_head is a COPY of the
2239 * original filter entry, with the exception of fltr_info.fltr_act and
2240 * fltr_info.fwd_id fields. These are set such that later logic can
2241 * extract which VSI to remove the fltr from, and pass on that information.
2242 */
2243static enum ice_status
2244ice_add_entry_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2245			       struct list_head *vsi_list_head,
2246			       struct ice_fltr_info *fi)
2247{
2248	struct ice_fltr_list_entry *tmp;
2249
2250	/* this memory is freed up in the caller function
2251	 * once filters for this VSI are removed
2252	 */
2253	tmp = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*tmp), GFP_KERNEL);
2254	if (!tmp)
2255		return ICE_ERR_NO_MEMORY;
2256
2257	tmp->fltr_info = *fi;
2258
2259	/* Overwrite these fields to indicate which VSI to remove filter from,
2260	 * so find and remove logic can extract the information from the
2261	 * list entries. Note that original entries will still have proper
2262	 * values.
2263	 */
2264	tmp->fltr_info.fltr_act = ICE_FWD_TO_VSI;
2265	tmp->fltr_info.vsi_handle = vsi_handle;
2266	tmp->fltr_info.fwd_id.hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2267
2268	list_add(&tmp->list_entry, vsi_list_head);
2269
2270	return 0;
2271}
2272
2273/**
2274 * ice_add_to_vsi_fltr_list - Add VSI filters to the list
2275 * @hw: pointer to the hardware structure
2276 * @vsi_handle: VSI handle to remove filters from
2277 * @lkup_list_head: pointer to the list that has certain lookup type filters
2278 * @vsi_list_head: pointer to the list pertaining to VSI with vsi_handle
2279 *
2280 * Locates all filters in lkup_list_head that are used by the given VSI,
2281 * and adds COPIES of those entries to vsi_list_head (intended to be used
2282 * to remove the listed filters).
2283 * Note that this means all entries in vsi_list_head must be explicitly
2284 * deallocated by the caller when done with list.
2285 */
2286static enum ice_status
2287ice_add_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_handle,
2288			 struct list_head *lkup_list_head,
2289			 struct list_head *vsi_list_head)
2290{
2291	struct ice_fltr_mgmt_list_entry *fm_entry;
2292	enum ice_status status = 0;
2293
2294	/* check to make sure VSI ID is valid and within boundary */
2295	if (!ice_is_vsi_valid(hw, vsi_handle))
2296		return ICE_ERR_PARAM;
2297
2298	list_for_each_entry(fm_entry, lkup_list_head, list_entry) {
2299		struct ice_fltr_info *fi;
2300
2301		fi = &fm_entry->fltr_info;
2302		if (!fi || !ice_vsi_uses_fltr(fm_entry, vsi_handle))
2303			continue;
2304
2305		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2306							vsi_list_head, fi);
2307		if (status)
2308			return status;
2309	}
2310	return status;
2311}
2312
2313/**
2314 * ice_determine_promisc_mask
2315 * @fi: filter info to parse
2316 *
2317 * Helper function to determine which ICE_PROMISC_ mask corresponds
2318 * to given filter into.
2319 */
2320static u8 ice_determine_promisc_mask(struct ice_fltr_info *fi)
2321{
2322	u16 vid = fi->l_data.mac_vlan.vlan_id;
2323	u8 *macaddr = fi->l_data.mac.mac_addr;
2324	bool is_tx_fltr = false;
2325	u8 promisc_mask = 0;
2326
2327	if (fi->flag == ICE_FLTR_TX)
2328		is_tx_fltr = true;
2329
2330	if (is_broadcast_ether_addr(macaddr))
2331		promisc_mask |= is_tx_fltr ?
2332			ICE_PROMISC_BCAST_TX : ICE_PROMISC_BCAST_RX;
2333	else if (is_multicast_ether_addr(macaddr))
2334		promisc_mask |= is_tx_fltr ?
2335			ICE_PROMISC_MCAST_TX : ICE_PROMISC_MCAST_RX;
2336	else if (is_unicast_ether_addr(macaddr))
2337		promisc_mask |= is_tx_fltr ?
2338			ICE_PROMISC_UCAST_TX : ICE_PROMISC_UCAST_RX;
2339	if (vid)
2340		promisc_mask |= is_tx_fltr ?
2341			ICE_PROMISC_VLAN_TX : ICE_PROMISC_VLAN_RX;
2342
2343	return promisc_mask;
2344}
2345
2346/**
2347 * ice_remove_promisc - Remove promisc based filter rules
2348 * @hw: pointer to the hardware structure
2349 * @recp_id: recipe ID for which the rule needs to removed
2350 * @v_list: list of promisc entries
2351 */
2352static enum ice_status
2353ice_remove_promisc(struct ice_hw *hw, u8 recp_id,
2354		   struct list_head *v_list)
2355{
2356	struct ice_fltr_list_entry *v_list_itr, *tmp;
2357
2358	list_for_each_entry_safe(v_list_itr, tmp, v_list, list_entry) {
2359		v_list_itr->status =
2360			ice_remove_rule_internal(hw, recp_id, v_list_itr);
2361		if (v_list_itr->status)
2362			return v_list_itr->status;
2363	}
2364	return 0;
2365}
2366
2367/**
2368 * ice_clear_vsi_promisc - clear specified promiscuous mode(s) for given VSI
2369 * @hw: pointer to the hardware structure
2370 * @vsi_handle: VSI handle to clear mode
2371 * @promisc_mask: mask of promiscuous config bits to clear
2372 * @vid: VLAN ID to clear VLAN promiscuous
2373 */
2374enum ice_status
2375ice_clear_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2376		      u16 vid)
2377{
2378	struct ice_switch_info *sw = hw->switch_info;
2379	struct ice_fltr_list_entry *fm_entry, *tmp;
2380	struct list_head remove_list_head;
2381	struct ice_fltr_mgmt_list_entry *itr;
2382	struct list_head *rule_head;
2383	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2384	enum ice_status status = 0;
2385	u8 recipe_id;
2386
2387	if (!ice_is_vsi_valid(hw, vsi_handle))
2388		return ICE_ERR_PARAM;
2389
2390	if (promisc_mask & (ICE_PROMISC_VLAN_RX | ICE_PROMISC_VLAN_TX))
2391		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2392	else
2393		recipe_id = ICE_SW_LKUP_PROMISC;
2394
2395	rule_head = &sw->recp_list[recipe_id].filt_rules;
2396	rule_lock = &sw->recp_list[recipe_id].filt_rule_lock;
2397
2398	INIT_LIST_HEAD(&remove_list_head);
2399
2400	mutex_lock(rule_lock);
2401	list_for_each_entry(itr, rule_head, list_entry) {
2402		struct ice_fltr_info *fltr_info;
2403		u8 fltr_promisc_mask = 0;
2404
2405		if (!ice_vsi_uses_fltr(itr, vsi_handle))
2406			continue;
2407		fltr_info = &itr->fltr_info;
2408
2409		if (recipe_id == ICE_SW_LKUP_PROMISC_VLAN &&
2410		    vid != fltr_info->l_data.mac_vlan.vlan_id)
2411			continue;
2412
2413		fltr_promisc_mask |= ice_determine_promisc_mask(fltr_info);
 
2414
2415		/* Skip if filter is not completely specified by given mask */
2416		if (fltr_promisc_mask & ~promisc_mask)
2417			continue;
2418
2419		status = ice_add_entry_to_vsi_fltr_list(hw, vsi_handle,
2420							&remove_list_head,
2421							fltr_info);
2422		if (status) {
2423			mutex_unlock(rule_lock);
2424			goto free_fltr_list;
2425		}
2426	}
2427	mutex_unlock(rule_lock);
2428
2429	status = ice_remove_promisc(hw, recipe_id, &remove_list_head);
2430
2431free_fltr_list:
2432	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2433		list_del(&fm_entry->list_entry);
2434		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2435	}
2436
2437	return status;
2438}
2439
2440/**
2441 * ice_set_vsi_promisc - set given VSI to given promiscuous mode(s)
2442 * @hw: pointer to the hardware structure
2443 * @vsi_handle: VSI handle to configure
2444 * @promisc_mask: mask of promiscuous config bits
2445 * @vid: VLAN ID to set VLAN promiscuous
2446 */
2447enum ice_status
2448ice_set_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask, u16 vid)
2449{
2450	enum { UCAST_FLTR = 1, MCAST_FLTR, BCAST_FLTR };
2451	struct ice_fltr_list_entry f_list_entry;
2452	struct ice_fltr_info new_fltr;
2453	enum ice_status status = 0;
2454	bool is_tx_fltr;
2455	u16 hw_vsi_id;
2456	int pkt_type;
2457	u8 recipe_id;
2458
2459	if (!ice_is_vsi_valid(hw, vsi_handle))
2460		return ICE_ERR_PARAM;
2461	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2462
2463	memset(&new_fltr, 0, sizeof(new_fltr));
2464
2465	if (promisc_mask & (ICE_PROMISC_VLAN_RX | ICE_PROMISC_VLAN_TX)) {
2466		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC_VLAN;
2467		new_fltr.l_data.mac_vlan.vlan_id = vid;
2468		recipe_id = ICE_SW_LKUP_PROMISC_VLAN;
2469	} else {
2470		new_fltr.lkup_type = ICE_SW_LKUP_PROMISC;
2471		recipe_id = ICE_SW_LKUP_PROMISC;
2472	}
2473
2474	/* Separate filters must be set for each direction/packet type
2475	 * combination, so we will loop over the mask value, store the
2476	 * individual type, and clear it out in the input mask as it
2477	 * is found.
2478	 */
2479	while (promisc_mask) {
2480		u8 *mac_addr;
2481
2482		pkt_type = 0;
2483		is_tx_fltr = false;
2484
2485		if (promisc_mask & ICE_PROMISC_UCAST_RX) {
2486			promisc_mask &= ~ICE_PROMISC_UCAST_RX;
2487			pkt_type = UCAST_FLTR;
2488		} else if (promisc_mask & ICE_PROMISC_UCAST_TX) {
2489			promisc_mask &= ~ICE_PROMISC_UCAST_TX;
2490			pkt_type = UCAST_FLTR;
2491			is_tx_fltr = true;
2492		} else if (promisc_mask & ICE_PROMISC_MCAST_RX) {
2493			promisc_mask &= ~ICE_PROMISC_MCAST_RX;
2494			pkt_type = MCAST_FLTR;
2495		} else if (promisc_mask & ICE_PROMISC_MCAST_TX) {
2496			promisc_mask &= ~ICE_PROMISC_MCAST_TX;
2497			pkt_type = MCAST_FLTR;
2498			is_tx_fltr = true;
2499		} else if (promisc_mask & ICE_PROMISC_BCAST_RX) {
2500			promisc_mask &= ~ICE_PROMISC_BCAST_RX;
2501			pkt_type = BCAST_FLTR;
2502		} else if (promisc_mask & ICE_PROMISC_BCAST_TX) {
2503			promisc_mask &= ~ICE_PROMISC_BCAST_TX;
2504			pkt_type = BCAST_FLTR;
2505			is_tx_fltr = true;
2506		}
2507
2508		/* Check for VLAN promiscuous flag */
2509		if (promisc_mask & ICE_PROMISC_VLAN_RX) {
2510			promisc_mask &= ~ICE_PROMISC_VLAN_RX;
2511		} else if (promisc_mask & ICE_PROMISC_VLAN_TX) {
2512			promisc_mask &= ~ICE_PROMISC_VLAN_TX;
2513			is_tx_fltr = true;
2514		}
2515
2516		/* Set filter DA based on packet type */
2517		mac_addr = new_fltr.l_data.mac.mac_addr;
2518		if (pkt_type == BCAST_FLTR) {
2519			eth_broadcast_addr(mac_addr);
2520		} else if (pkt_type == MCAST_FLTR ||
2521			   pkt_type == UCAST_FLTR) {
2522			/* Use the dummy ether header DA */
2523			ether_addr_copy(mac_addr, dummy_eth_header);
2524			if (pkt_type == MCAST_FLTR)
2525				mac_addr[0] |= 0x1;	/* Set multicast bit */
2526		}
2527
2528		/* Need to reset this to zero for all iterations */
2529		new_fltr.flag = 0;
2530		if (is_tx_fltr) {
2531			new_fltr.flag |= ICE_FLTR_TX;
2532			new_fltr.src = hw_vsi_id;
2533		} else {
2534			new_fltr.flag |= ICE_FLTR_RX;
2535			new_fltr.src = hw->port_info->lport;
2536		}
2537
2538		new_fltr.fltr_act = ICE_FWD_TO_VSI;
2539		new_fltr.vsi_handle = vsi_handle;
2540		new_fltr.fwd_id.hw_vsi_id = hw_vsi_id;
2541		f_list_entry.fltr_info = new_fltr;
2542
2543		status = ice_add_rule_internal(hw, recipe_id, &f_list_entry);
2544		if (status)
2545			goto set_promisc_exit;
2546	}
2547
2548set_promisc_exit:
2549	return status;
2550}
2551
2552/**
2553 * ice_set_vlan_vsi_promisc
2554 * @hw: pointer to the hardware structure
2555 * @vsi_handle: VSI handle to configure
2556 * @promisc_mask: mask of promiscuous config bits
2557 * @rm_vlan_promisc: Clear VLANs VSI promisc mode
2558 *
2559 * Configure VSI with all associated VLANs to given promiscuous mode(s)
2560 */
2561enum ice_status
2562ice_set_vlan_vsi_promisc(struct ice_hw *hw, u16 vsi_handle, u8 promisc_mask,
2563			 bool rm_vlan_promisc)
2564{
2565	struct ice_switch_info *sw = hw->switch_info;
2566	struct ice_fltr_list_entry *list_itr, *tmp;
2567	struct list_head vsi_list_head;
2568	struct list_head *vlan_head;
2569	struct mutex *vlan_lock; /* Lock to protect filter rule list */
2570	enum ice_status status;
2571	u16 vlan_id;
2572
2573	INIT_LIST_HEAD(&vsi_list_head);
2574	vlan_lock = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rule_lock;
2575	vlan_head = &sw->recp_list[ICE_SW_LKUP_VLAN].filt_rules;
2576	mutex_lock(vlan_lock);
2577	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, vlan_head,
2578					  &vsi_list_head);
2579	mutex_unlock(vlan_lock);
2580	if (status)
2581		goto free_fltr_list;
2582
2583	list_for_each_entry(list_itr, &vsi_list_head, list_entry) {
2584		vlan_id = list_itr->fltr_info.l_data.vlan.vlan_id;
2585		if (rm_vlan_promisc)
2586			status = ice_clear_vsi_promisc(hw, vsi_handle,
2587						       promisc_mask, vlan_id);
2588		else
2589			status = ice_set_vsi_promisc(hw, vsi_handle,
2590						     promisc_mask, vlan_id);
2591		if (status)
2592			break;
2593	}
2594
2595free_fltr_list:
2596	list_for_each_entry_safe(list_itr, tmp, &vsi_list_head, list_entry) {
2597		list_del(&list_itr->list_entry);
2598		devm_kfree(ice_hw_to_dev(hw), list_itr);
2599	}
2600	return status;
2601}
2602
2603/**
2604 * ice_remove_vsi_lkup_fltr - Remove lookup type filters for a VSI
2605 * @hw: pointer to the hardware structure
2606 * @vsi_handle: VSI handle to remove filters from
2607 * @lkup: switch rule filter lookup type
2608 */
2609static void
2610ice_remove_vsi_lkup_fltr(struct ice_hw *hw, u16 vsi_handle,
2611			 enum ice_sw_lkup_type lkup)
2612{
2613	struct ice_switch_info *sw = hw->switch_info;
2614	struct ice_fltr_list_entry *fm_entry;
2615	struct list_head remove_list_head;
2616	struct list_head *rule_head;
2617	struct ice_fltr_list_entry *tmp;
2618	struct mutex *rule_lock;	/* Lock to protect filter rule list */
2619	enum ice_status status;
2620
2621	INIT_LIST_HEAD(&remove_list_head);
2622	rule_lock = &sw->recp_list[lkup].filt_rule_lock;
2623	rule_head = &sw->recp_list[lkup].filt_rules;
2624	mutex_lock(rule_lock);
2625	status = ice_add_to_vsi_fltr_list(hw, vsi_handle, rule_head,
2626					  &remove_list_head);
2627	mutex_unlock(rule_lock);
2628	if (status)
2629		return;
2630
2631	switch (lkup) {
2632	case ICE_SW_LKUP_MAC:
2633		ice_remove_mac(hw, &remove_list_head);
2634		break;
2635	case ICE_SW_LKUP_VLAN:
2636		ice_remove_vlan(hw, &remove_list_head);
2637		break;
2638	case ICE_SW_LKUP_PROMISC:
2639	case ICE_SW_LKUP_PROMISC_VLAN:
2640		ice_remove_promisc(hw, lkup, &remove_list_head);
2641		break;
2642	case ICE_SW_LKUP_MAC_VLAN:
2643	case ICE_SW_LKUP_ETHERTYPE:
2644	case ICE_SW_LKUP_ETHERTYPE_MAC:
2645	case ICE_SW_LKUP_DFLT:
2646	case ICE_SW_LKUP_LAST:
2647	default:
2648		ice_debug(hw, ICE_DBG_SW, "Unsupported lookup type %d\n", lkup);
2649		break;
2650	}
2651
2652	list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
2653		list_del(&fm_entry->list_entry);
2654		devm_kfree(ice_hw_to_dev(hw), fm_entry);
2655	}
2656}
2657
2658/**
2659 * ice_remove_vsi_fltr - Remove all filters for a VSI
2660 * @hw: pointer to the hardware structure
2661 * @vsi_handle: VSI handle to remove filters from
2662 */
2663void ice_remove_vsi_fltr(struct ice_hw *hw, u16 vsi_handle)
2664{
2665	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC);
2666	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_MAC_VLAN);
2667	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC);
2668	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_VLAN);
2669	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_DFLT);
2670	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE);
2671	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_ETHERTYPE_MAC);
2672	ice_remove_vsi_lkup_fltr(hw, vsi_handle, ICE_SW_LKUP_PROMISC_VLAN);
2673}
2674
2675/**
2676 * ice_alloc_res_cntr - allocating resource counter
2677 * @hw: pointer to the hardware structure
2678 * @type: type of resource
2679 * @alloc_shared: if set it is shared else dedicated
2680 * @num_items: number of entries requested for FD resource type
2681 * @counter_id: counter index returned by AQ call
2682 */
2683enum ice_status
2684ice_alloc_res_cntr(struct ice_hw *hw, u8 type, u8 alloc_shared, u16 num_items,
2685		   u16 *counter_id)
2686{
2687	struct ice_aqc_alloc_free_res_elem *buf;
2688	enum ice_status status;
2689	u16 buf_len;
2690
2691	/* Allocate resource */
2692	buf_len = struct_size(buf, elem, 1);
2693	buf = kzalloc(buf_len, GFP_KERNEL);
2694	if (!buf)
2695		return ICE_ERR_NO_MEMORY;
2696
2697	buf->num_elems = cpu_to_le16(num_items);
2698	buf->res_type = cpu_to_le16(((type << ICE_AQC_RES_TYPE_S) &
2699				      ICE_AQC_RES_TYPE_M) | alloc_shared);
2700
2701	status = ice_aq_alloc_free_res(hw, 1, buf, buf_len,
2702				       ice_aqc_opc_alloc_res, NULL);
2703	if (status)
2704		goto exit;
2705
2706	*counter_id = le16_to_cpu(buf->elem[0].e.sw_resp);
2707
2708exit:
2709	kfree(buf);
2710	return status;
2711}
2712
2713/**
2714 * ice_free_res_cntr - free resource counter
2715 * @hw: pointer to the hardware structure
2716 * @type: type of resource
2717 * @alloc_shared: if set it is shared else dedicated
2718 * @num_items: number of entries to be freed for FD resource type
2719 * @counter_id: counter ID resource which needs to be freed
2720 */
2721enum ice_status
2722ice_free_res_cntr(struct ice_hw *hw, u8 type, u8 alloc_shared, u16 num_items,
2723		  u16 counter_id)
2724{
2725	struct ice_aqc_alloc_free_res_elem *buf;
2726	enum ice_status status;
2727	u16 buf_len;
2728
2729	/* Free resource */
2730	buf_len = struct_size(buf, elem, 1);
2731	buf = kzalloc(buf_len, GFP_KERNEL);
2732	if (!buf)
2733		return ICE_ERR_NO_MEMORY;
2734
2735	buf->num_elems = cpu_to_le16(num_items);
2736	buf->res_type = cpu_to_le16(((type << ICE_AQC_RES_TYPE_S) &
2737				      ICE_AQC_RES_TYPE_M) | alloc_shared);
2738	buf->elem[0].e.sw_resp = cpu_to_le16(counter_id);
2739
2740	status = ice_aq_alloc_free_res(hw, 1, buf, buf_len,
2741				       ice_aqc_opc_free_res, NULL);
2742	if (status)
2743		ice_debug(hw, ICE_DBG_SW,
2744			  "counter resource could not be freed\n");
2745
2746	kfree(buf);
2747	return status;
2748}
2749
2750/**
2751 * ice_replay_vsi_fltr - Replay filters for requested VSI
2752 * @hw: pointer to the hardware structure
2753 * @vsi_handle: driver VSI handle
2754 * @recp_id: Recipe ID for which rules need to be replayed
2755 * @list_head: list for which filters need to be replayed
2756 *
2757 * Replays the filter of recipe recp_id for a VSI represented via vsi_handle.
2758 * It is required to pass valid VSI handle.
2759 */
2760static enum ice_status
2761ice_replay_vsi_fltr(struct ice_hw *hw, u16 vsi_handle, u8 recp_id,
2762		    struct list_head *list_head)
2763{
2764	struct ice_fltr_mgmt_list_entry *itr;
2765	enum ice_status status = 0;
2766	u16 hw_vsi_id;
2767
2768	if (list_empty(list_head))
2769		return status;
2770	hw_vsi_id = ice_get_hw_vsi_num(hw, vsi_handle);
2771
2772	list_for_each_entry(itr, list_head, list_entry) {
2773		struct ice_fltr_list_entry f_entry;
2774
2775		f_entry.fltr_info = itr->fltr_info;
2776		if (itr->vsi_count < 2 && recp_id != ICE_SW_LKUP_VLAN &&
2777		    itr->fltr_info.vsi_handle == vsi_handle) {
2778			/* update the src in case it is VSI num */
2779			if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2780				f_entry.fltr_info.src = hw_vsi_id;
2781			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2782			if (status)
2783				goto end;
2784			continue;
2785		}
2786		if (!itr->vsi_list_info ||
2787		    !test_bit(vsi_handle, itr->vsi_list_info->vsi_map))
2788			continue;
2789		/* Clearing it so that the logic can add it back */
2790		clear_bit(vsi_handle, itr->vsi_list_info->vsi_map);
2791		f_entry.fltr_info.vsi_handle = vsi_handle;
2792		f_entry.fltr_info.fltr_act = ICE_FWD_TO_VSI;
2793		/* update the src in case it is VSI num */
2794		if (f_entry.fltr_info.src_id == ICE_SRC_ID_VSI)
2795			f_entry.fltr_info.src = hw_vsi_id;
2796		if (recp_id == ICE_SW_LKUP_VLAN)
2797			status = ice_add_vlan_internal(hw, &f_entry);
2798		else
2799			status = ice_add_rule_internal(hw, recp_id, &f_entry);
2800		if (status)
2801			goto end;
2802	}
2803end:
2804	return status;
2805}
2806
2807/**
2808 * ice_replay_vsi_all_fltr - replay all filters stored in bookkeeping lists
2809 * @hw: pointer to the hardware structure
2810 * @vsi_handle: driver VSI handle
2811 *
2812 * Replays filters for requested VSI via vsi_handle.
2813 */
2814enum ice_status ice_replay_vsi_all_fltr(struct ice_hw *hw, u16 vsi_handle)
2815{
2816	struct ice_switch_info *sw = hw->switch_info;
2817	enum ice_status status = 0;
2818	u8 i;
2819
2820	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2821		struct list_head *head;
2822
2823		head = &sw->recp_list[i].filt_replay_rules;
2824		status = ice_replay_vsi_fltr(hw, vsi_handle, i, head);
2825		if (status)
2826			return status;
2827	}
2828	return status;
2829}
2830
2831/**
2832 * ice_rm_all_sw_replay_rule_info - deletes filter replay rules
2833 * @hw: pointer to the HW struct
2834 *
2835 * Deletes the filter replay rules.
2836 */
2837void ice_rm_all_sw_replay_rule_info(struct ice_hw *hw)
2838{
2839	struct ice_switch_info *sw = hw->switch_info;
2840	u8 i;
2841
2842	if (!sw)
2843		return;
2844
2845	for (i = 0; i < ICE_SW_LKUP_LAST; i++) {
2846		if (!list_empty(&sw->recp_list[i].filt_replay_rules)) {
2847			struct list_head *l_head;
2848
2849			l_head = &sw->recp_list[i].filt_replay_rules;
2850			ice_rem_sw_rule_info(hw, l_head);
2851		}
2852	}
2853}