Loading...
Note: File does not exist in v3.1.
1// SPDX-License-Identifier: GPL-2.0
2/* Copyright (c) 2018, Intel Corporation. */
3
4#include "ice_switch.h"
5
6#define ICE_ETH_DA_OFFSET 0
7#define ICE_ETH_ETHTYPE_OFFSET 12
8#define ICE_ETH_VLAN_TCI_OFFSET 14
9#define ICE_MAX_VLAN_ID 0xFFF
10
11/* Dummy ethernet header needed in the ice_aqc_sw_rules_elem
12 * struct to configure any switch filter rules.
13 * {DA (6 bytes), SA(6 bytes),
14 * Ether type (2 bytes for header without VLAN tag) OR
15 * VLAN tag (4 bytes for header with VLAN tag) }
16 *
17 * Word on Hardcoded values
18 * byte 0 = 0x2: to identify it as locally administered DA MAC
19 * byte 6 = 0x2: to identify it as locally administered SA MAC
20 * byte 12 = 0x81 & byte 13 = 0x00:
21 * In case of VLAN filter first two bytes defines ether type (0x8100)
22 * and remaining two bytes are placeholder for programming a given VLAN id
23 * In case of Ether type filter it is treated as header without VLAN tag
24 * and byte 12 and 13 is used to program a given Ether type instead
25 */
26#define DUMMY_ETH_HDR_LEN 16
27static const u8 dummy_eth_header[DUMMY_ETH_HDR_LEN] = { 0x2, 0, 0, 0, 0, 0,
28 0x2, 0, 0, 0, 0, 0,
29 0x81, 0, 0, 0};
30
31#define ICE_SW_RULE_RX_TX_ETH_HDR_SIZE \
32 (sizeof(struct ice_aqc_sw_rules_elem) - \
33 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
34 sizeof(struct ice_sw_rule_lkup_rx_tx) + DUMMY_ETH_HDR_LEN - 1)
35#define ICE_SW_RULE_RX_TX_NO_HDR_SIZE \
36 (sizeof(struct ice_aqc_sw_rules_elem) - \
37 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
38 sizeof(struct ice_sw_rule_lkup_rx_tx) - 1)
39#define ICE_SW_RULE_LG_ACT_SIZE(n) \
40 (sizeof(struct ice_aqc_sw_rules_elem) - \
41 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
42 sizeof(struct ice_sw_rule_lg_act) - \
43 sizeof(((struct ice_sw_rule_lg_act *)0)->act) + \
44 ((n) * sizeof(((struct ice_sw_rule_lg_act *)0)->act)))
45#define ICE_SW_RULE_VSI_LIST_SIZE(n) \
46 (sizeof(struct ice_aqc_sw_rules_elem) - \
47 sizeof(((struct ice_aqc_sw_rules_elem *)0)->pdata) + \
48 sizeof(struct ice_sw_rule_vsi_list) - \
49 sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi) + \
50 ((n) * sizeof(((struct ice_sw_rule_vsi_list *)0)->vsi)))
51
52/**
53 * ice_aq_alloc_free_res - command to allocate/free resources
54 * @hw: pointer to the hw struct
55 * @num_entries: number of resource entries in buffer
56 * @buf: Indirect buffer to hold data parameters and response
57 * @buf_size: size of buffer for indirect commands
58 * @opc: pass in the command opcode
59 * @cd: pointer to command details structure or NULL
60 *
61 * Helper function to allocate/free resources using the admin queue commands
62 */
63static enum ice_status
64ice_aq_alloc_free_res(struct ice_hw *hw, u16 num_entries,
65 struct ice_aqc_alloc_free_res_elem *buf, u16 buf_size,
66 enum ice_adminq_opc opc, struct ice_sq_cd *cd)
67{
68 struct ice_aqc_alloc_free_res_cmd *cmd;
69 struct ice_aq_desc desc;
70
71 cmd = &desc.params.sw_res_ctrl;
72
73 if (!buf)
74 return ICE_ERR_PARAM;
75
76 if (buf_size < (num_entries * sizeof(buf->elem[0])))
77 return ICE_ERR_PARAM;
78
79 ice_fill_dflt_direct_cmd_desc(&desc, opc);
80
81 desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
82
83 cmd->num_entries = cpu_to_le16(num_entries);
84
85 return ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
86}
87
88/**
89 * ice_aq_get_sw_cfg - get switch configuration
90 * @hw: pointer to the hardware structure
91 * @buf: pointer to the result buffer
92 * @buf_size: length of the buffer available for response
93 * @req_desc: pointer to requested descriptor
94 * @num_elems: pointer to number of elements
95 * @cd: pointer to command details structure or NULL
96 *
97 * Get switch configuration (0x0200) to be placed in 'buff'.
98 * This admin command returns information such as initial VSI/port number
99 * and switch ID it belongs to.
100 *
101 * NOTE: *req_desc is both an input/output parameter.
102 * The caller of this function first calls this function with *request_desc set
103 * to 0. If the response from f/w has *req_desc set to 0, all the switch
104 * configuration information has been returned; if non-zero (meaning not all
105 * the information was returned), the caller should call this function again
106 * with *req_desc set to the previous value returned by f/w to get the
107 * next block of switch configuration information.
108 *
109 * *num_elems is output only parameter. This reflects the number of elements
110 * in response buffer. The caller of this function to use *num_elems while
111 * parsing the response buffer.
112 */
113static enum ice_status
114ice_aq_get_sw_cfg(struct ice_hw *hw, struct ice_aqc_get_sw_cfg_resp *buf,
115 u16 buf_size, u16 *req_desc, u16 *num_elems,
116 struct ice_sq_cd *cd)
117{
118 struct ice_aqc_get_sw_cfg *cmd;
119 enum ice_status status;
120 struct ice_aq_desc desc;
121
122 ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_get_sw_cfg);
123 cmd = &desc.params.get_sw_conf;
124 cmd->element = cpu_to_le16(*req_desc);
125
126 status = ice_aq_send_cmd(hw, &desc, buf, buf_size, cd);
127 if (!status) {
128 *req_desc = le16_to_cpu(cmd->element);
129 *num_elems = le16_to_cpu(cmd->num_elems);
130 }
131
132 return status;
133}
134
135/**
136 * ice_aq_add_vsi
137 * @hw: pointer to the hw struct
138 * @vsi_ctx: pointer to a VSI context struct
139 * @cd: pointer to command details structure or NULL
140 *
141 * Add a VSI context to the hardware (0x0210)
142 */
143enum ice_status
144ice_aq_add_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
145 struct ice_sq_cd *cd)
146{
147 struct ice_aqc_add_update_free_vsi_resp *res;
148 struct ice_aqc_add_get_update_free_vsi *cmd;
149 enum ice_status status;
150 struct ice_aq_desc desc;
151
152 cmd = &desc.params.vsi_cmd;
153 res = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
154
155 ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_add_vsi);
156
157 if (!vsi_ctx->alloc_from_pool)
158 cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num |
159 ICE_AQ_VSI_IS_VALID);
160
161 cmd->vsi_flags = cpu_to_le16(vsi_ctx->flags);
162
163 desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
164
165 status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
166 sizeof(vsi_ctx->info), cd);
167
168 if (!status) {
169 vsi_ctx->vsi_num = le16_to_cpu(res->vsi_num) & ICE_AQ_VSI_NUM_M;
170 vsi_ctx->vsis_allocd = le16_to_cpu(res->vsi_used);
171 vsi_ctx->vsis_unallocated = le16_to_cpu(res->vsi_free);
172 }
173
174 return status;
175}
176
177/**
178 * ice_aq_update_vsi
179 * @hw: pointer to the hw struct
180 * @vsi_ctx: pointer to a VSI context struct
181 * @cd: pointer to command details structure or NULL
182 *
183 * Update VSI context in the hardware (0x0211)
184 */
185enum ice_status
186ice_aq_update_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
187 struct ice_sq_cd *cd)
188{
189 struct ice_aqc_add_update_free_vsi_resp *resp;
190 struct ice_aqc_add_get_update_free_vsi *cmd;
191 struct ice_aq_desc desc;
192 enum ice_status status;
193
194 cmd = &desc.params.vsi_cmd;
195 resp = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
196
197 ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_update_vsi);
198
199 cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
200
201 desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
202
203 status = ice_aq_send_cmd(hw, &desc, &vsi_ctx->info,
204 sizeof(vsi_ctx->info), cd);
205
206 if (!status) {
207 vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
208 vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
209 }
210
211 return status;
212}
213
214/**
215 * ice_aq_free_vsi
216 * @hw: pointer to the hw struct
217 * @vsi_ctx: pointer to a VSI context struct
218 * @keep_vsi_alloc: keep VSI allocation as part of this PF's resources
219 * @cd: pointer to command details structure or NULL
220 *
221 * Get VSI context info from hardware (0x0213)
222 */
223enum ice_status
224ice_aq_free_vsi(struct ice_hw *hw, struct ice_vsi_ctx *vsi_ctx,
225 bool keep_vsi_alloc, struct ice_sq_cd *cd)
226{
227 struct ice_aqc_add_update_free_vsi_resp *resp;
228 struct ice_aqc_add_get_update_free_vsi *cmd;
229 struct ice_aq_desc desc;
230 enum ice_status status;
231
232 cmd = &desc.params.vsi_cmd;
233 resp = (struct ice_aqc_add_update_free_vsi_resp *)&desc.params.raw;
234
235 ice_fill_dflt_direct_cmd_desc(&desc, ice_aqc_opc_free_vsi);
236
237 cmd->vsi_num = cpu_to_le16(vsi_ctx->vsi_num | ICE_AQ_VSI_IS_VALID);
238 if (keep_vsi_alloc)
239 cmd->cmd_flags = cpu_to_le16(ICE_AQ_VSI_KEEP_ALLOC);
240
241 status = ice_aq_send_cmd(hw, &desc, NULL, 0, cd);
242 if (!status) {
243 vsi_ctx->vsis_allocd = le16_to_cpu(resp->vsi_used);
244 vsi_ctx->vsis_unallocated = le16_to_cpu(resp->vsi_free);
245 }
246
247 return status;
248}
249
250/**
251 * ice_aq_alloc_free_vsi_list
252 * @hw: pointer to the hw struct
253 * @vsi_list_id: VSI list id returned or used for lookup
254 * @lkup_type: switch rule filter lookup type
255 * @opc: switch rules population command type - pass in the command opcode
256 *
257 * allocates or free a VSI list resource
258 */
259static enum ice_status
260ice_aq_alloc_free_vsi_list(struct ice_hw *hw, u16 *vsi_list_id,
261 enum ice_sw_lkup_type lkup_type,
262 enum ice_adminq_opc opc)
263{
264 struct ice_aqc_alloc_free_res_elem *sw_buf;
265 struct ice_aqc_res_elem *vsi_ele;
266 enum ice_status status;
267 u16 buf_len;
268
269 buf_len = sizeof(*sw_buf);
270 sw_buf = devm_kzalloc(ice_hw_to_dev(hw), buf_len, GFP_KERNEL);
271 if (!sw_buf)
272 return ICE_ERR_NO_MEMORY;
273 sw_buf->num_elems = cpu_to_le16(1);
274
275 if (lkup_type == ICE_SW_LKUP_MAC ||
276 lkup_type == ICE_SW_LKUP_MAC_VLAN ||
277 lkup_type == ICE_SW_LKUP_ETHERTYPE ||
278 lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
279 lkup_type == ICE_SW_LKUP_PROMISC ||
280 lkup_type == ICE_SW_LKUP_PROMISC_VLAN) {
281 sw_buf->res_type = cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_REP);
282 } else if (lkup_type == ICE_SW_LKUP_VLAN) {
283 sw_buf->res_type =
284 cpu_to_le16(ICE_AQC_RES_TYPE_VSI_LIST_PRUNE);
285 } else {
286 status = ICE_ERR_PARAM;
287 goto ice_aq_alloc_free_vsi_list_exit;
288 }
289
290 if (opc == ice_aqc_opc_free_res)
291 sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
292
293 status = ice_aq_alloc_free_res(hw, 1, sw_buf, buf_len, opc, NULL);
294 if (status)
295 goto ice_aq_alloc_free_vsi_list_exit;
296
297 if (opc == ice_aqc_opc_alloc_res) {
298 vsi_ele = &sw_buf->elem[0];
299 *vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
300 }
301
302ice_aq_alloc_free_vsi_list_exit:
303 devm_kfree(ice_hw_to_dev(hw), sw_buf);
304 return status;
305}
306
307/**
308 * ice_aq_sw_rules - add/update/remove switch rules
309 * @hw: pointer to the hw struct
310 * @rule_list: pointer to switch rule population list
311 * @rule_list_sz: total size of the rule list in bytes
312 * @num_rules: number of switch rules in the rule_list
313 * @opc: switch rules population command type - pass in the command opcode
314 * @cd: pointer to command details structure or NULL
315 *
316 * Add(0x02a0)/Update(0x02a1)/Remove(0x02a2) switch rules commands to firmware
317 */
318static enum ice_status
319ice_aq_sw_rules(struct ice_hw *hw, void *rule_list, u16 rule_list_sz,
320 u8 num_rules, enum ice_adminq_opc opc, struct ice_sq_cd *cd)
321{
322 struct ice_aq_desc desc;
323
324 if (opc != ice_aqc_opc_add_sw_rules &&
325 opc != ice_aqc_opc_update_sw_rules &&
326 opc != ice_aqc_opc_remove_sw_rules)
327 return ICE_ERR_PARAM;
328
329 ice_fill_dflt_direct_cmd_desc(&desc, opc);
330
331 desc.flags |= cpu_to_le16(ICE_AQ_FLAG_RD);
332 desc.params.sw_rules.num_rules_fltr_entry_index =
333 cpu_to_le16(num_rules);
334 return ice_aq_send_cmd(hw, &desc, rule_list, rule_list_sz, cd);
335}
336
337/* ice_init_port_info - Initialize port_info with switch configuration data
338 * @pi: pointer to port_info
339 * @vsi_port_num: VSI number or port number
340 * @type: Type of switch element (port or VSI)
341 * @swid: switch ID of the switch the element is attached to
342 * @pf_vf_num: PF or VF number
343 * @is_vf: true if the element is a VF, false otherwise
344 */
345static void
346ice_init_port_info(struct ice_port_info *pi, u16 vsi_port_num, u8 type,
347 u16 swid, u16 pf_vf_num, bool is_vf)
348{
349 switch (type) {
350 case ICE_AQC_GET_SW_CONF_RESP_PHYS_PORT:
351 pi->lport = (u8)(vsi_port_num & ICE_LPORT_MASK);
352 pi->sw_id = swid;
353 pi->pf_vf_num = pf_vf_num;
354 pi->is_vf = is_vf;
355 pi->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
356 pi->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
357 break;
358 default:
359 ice_debug(pi->hw, ICE_DBG_SW,
360 "incorrect VSI/port type received\n");
361 break;
362 }
363}
364
365/* ice_get_initial_sw_cfg - Get initial port and default VSI data
366 * @hw: pointer to the hardware structure
367 */
368enum ice_status ice_get_initial_sw_cfg(struct ice_hw *hw)
369{
370 struct ice_aqc_get_sw_cfg_resp *rbuf;
371 enum ice_status status;
372 u16 req_desc = 0;
373 u16 num_elems;
374 u16 i;
375
376 rbuf = devm_kzalloc(ice_hw_to_dev(hw), ICE_SW_CFG_MAX_BUF_LEN,
377 GFP_KERNEL);
378
379 if (!rbuf)
380 return ICE_ERR_NO_MEMORY;
381
382 /* Multiple calls to ice_aq_get_sw_cfg may be required
383 * to get all the switch configuration information. The need
384 * for additional calls is indicated by ice_aq_get_sw_cfg
385 * writing a non-zero value in req_desc
386 */
387 do {
388 status = ice_aq_get_sw_cfg(hw, rbuf, ICE_SW_CFG_MAX_BUF_LEN,
389 &req_desc, &num_elems, NULL);
390
391 if (status)
392 break;
393
394 for (i = 0; i < num_elems; i++) {
395 struct ice_aqc_get_sw_cfg_resp_elem *ele;
396 u16 pf_vf_num, swid, vsi_port_num;
397 bool is_vf = false;
398 u8 type;
399
400 ele = rbuf[i].elements;
401 vsi_port_num = le16_to_cpu(ele->vsi_port_num) &
402 ICE_AQC_GET_SW_CONF_RESP_VSI_PORT_NUM_M;
403
404 pf_vf_num = le16_to_cpu(ele->pf_vf_num) &
405 ICE_AQC_GET_SW_CONF_RESP_FUNC_NUM_M;
406
407 swid = le16_to_cpu(ele->swid);
408
409 if (le16_to_cpu(ele->pf_vf_num) &
410 ICE_AQC_GET_SW_CONF_RESP_IS_VF)
411 is_vf = true;
412
413 type = le16_to_cpu(ele->vsi_port_num) >>
414 ICE_AQC_GET_SW_CONF_RESP_TYPE_S;
415
416 if (type == ICE_AQC_GET_SW_CONF_RESP_VSI) {
417 /* FW VSI is not needed. Just continue. */
418 continue;
419 }
420
421 ice_init_port_info(hw->port_info, vsi_port_num,
422 type, swid, pf_vf_num, is_vf);
423 }
424 } while (req_desc && !status);
425
426 devm_kfree(ice_hw_to_dev(hw), (void *)rbuf);
427 return status;
428}
429
430/**
431 * ice_fill_sw_info - Helper function to populate lb_en and lan_en
432 * @hw: pointer to the hardware structure
433 * @f_info: filter info structure to fill/update
434 *
435 * This helper function populates the lb_en and lan_en elements of the provided
436 * ice_fltr_info struct using the switch's type and characteristics of the
437 * switch rule being configured.
438 */
439static void ice_fill_sw_info(struct ice_hw *hw, struct ice_fltr_info *f_info)
440{
441 f_info->lb_en = false;
442 f_info->lan_en = false;
443 if ((f_info->flag & ICE_FLTR_TX) &&
444 (f_info->fltr_act == ICE_FWD_TO_VSI ||
445 f_info->fltr_act == ICE_FWD_TO_VSI_LIST ||
446 f_info->fltr_act == ICE_FWD_TO_Q ||
447 f_info->fltr_act == ICE_FWD_TO_QGRP)) {
448 f_info->lb_en = true;
449 if (!(hw->evb_veb && f_info->lkup_type == ICE_SW_LKUP_MAC &&
450 is_unicast_ether_addr(f_info->l_data.mac.mac_addr)))
451 f_info->lan_en = true;
452 }
453}
454
455/**
456 * ice_fill_sw_rule - Helper function to fill switch rule structure
457 * @hw: pointer to the hardware structure
458 * @f_info: entry containing packet forwarding information
459 * @s_rule: switch rule structure to be filled in based on mac_entry
460 * @opc: switch rules population command type - pass in the command opcode
461 */
462static void
463ice_fill_sw_rule(struct ice_hw *hw, struct ice_fltr_info *f_info,
464 struct ice_aqc_sw_rules_elem *s_rule, enum ice_adminq_opc opc)
465{
466 u16 vlan_id = ICE_MAX_VLAN_ID + 1;
467 u8 eth_hdr[DUMMY_ETH_HDR_LEN];
468 void *daddr = NULL;
469 u32 act = 0;
470 __be16 *off;
471
472 if (opc == ice_aqc_opc_remove_sw_rules) {
473 s_rule->pdata.lkup_tx_rx.act = 0;
474 s_rule->pdata.lkup_tx_rx.index =
475 cpu_to_le16(f_info->fltr_rule_id);
476 s_rule->pdata.lkup_tx_rx.hdr_len = 0;
477 return;
478 }
479
480 /* initialize the ether header with a dummy header */
481 memcpy(eth_hdr, dummy_eth_header, sizeof(dummy_eth_header));
482 ice_fill_sw_info(hw, f_info);
483
484 switch (f_info->fltr_act) {
485 case ICE_FWD_TO_VSI:
486 act |= (f_info->fwd_id.vsi_id << ICE_SINGLE_ACT_VSI_ID_S) &
487 ICE_SINGLE_ACT_VSI_ID_M;
488 if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
489 act |= ICE_SINGLE_ACT_VSI_FORWARDING |
490 ICE_SINGLE_ACT_VALID_BIT;
491 break;
492 case ICE_FWD_TO_VSI_LIST:
493 act |= ICE_SINGLE_ACT_VSI_LIST;
494 act |= (f_info->fwd_id.vsi_list_id <<
495 ICE_SINGLE_ACT_VSI_LIST_ID_S) &
496 ICE_SINGLE_ACT_VSI_LIST_ID_M;
497 if (f_info->lkup_type != ICE_SW_LKUP_VLAN)
498 act |= ICE_SINGLE_ACT_VSI_FORWARDING |
499 ICE_SINGLE_ACT_VALID_BIT;
500 break;
501 case ICE_FWD_TO_Q:
502 act |= ICE_SINGLE_ACT_TO_Q;
503 act |= (f_info->fwd_id.q_id << ICE_SINGLE_ACT_Q_INDEX_S) &
504 ICE_SINGLE_ACT_Q_INDEX_M;
505 break;
506 case ICE_FWD_TO_QGRP:
507 act |= ICE_SINGLE_ACT_TO_Q;
508 act |= (f_info->qgrp_size << ICE_SINGLE_ACT_Q_REGION_S) &
509 ICE_SINGLE_ACT_Q_REGION_M;
510 break;
511 case ICE_DROP_PACKET:
512 act |= ICE_SINGLE_ACT_VSI_FORWARDING | ICE_SINGLE_ACT_DROP;
513 break;
514 default:
515 return;
516 }
517
518 if (f_info->lb_en)
519 act |= ICE_SINGLE_ACT_LB_ENABLE;
520 if (f_info->lan_en)
521 act |= ICE_SINGLE_ACT_LAN_ENABLE;
522
523 switch (f_info->lkup_type) {
524 case ICE_SW_LKUP_MAC:
525 daddr = f_info->l_data.mac.mac_addr;
526 break;
527 case ICE_SW_LKUP_VLAN:
528 vlan_id = f_info->l_data.vlan.vlan_id;
529 if (f_info->fltr_act == ICE_FWD_TO_VSI ||
530 f_info->fltr_act == ICE_FWD_TO_VSI_LIST) {
531 act |= ICE_SINGLE_ACT_PRUNE;
532 act |= ICE_SINGLE_ACT_EGRESS | ICE_SINGLE_ACT_INGRESS;
533 }
534 break;
535 case ICE_SW_LKUP_ETHERTYPE_MAC:
536 daddr = f_info->l_data.ethertype_mac.mac_addr;
537 /* fall-through */
538 case ICE_SW_LKUP_ETHERTYPE:
539 off = (__be16 *)ð_hdr[ICE_ETH_ETHTYPE_OFFSET];
540 *off = cpu_to_be16(f_info->l_data.ethertype_mac.ethertype);
541 break;
542 case ICE_SW_LKUP_MAC_VLAN:
543 daddr = f_info->l_data.mac_vlan.mac_addr;
544 vlan_id = f_info->l_data.mac_vlan.vlan_id;
545 break;
546 case ICE_SW_LKUP_PROMISC_VLAN:
547 vlan_id = f_info->l_data.mac_vlan.vlan_id;
548 /* fall-through */
549 case ICE_SW_LKUP_PROMISC:
550 daddr = f_info->l_data.mac_vlan.mac_addr;
551 break;
552 default:
553 break;
554 }
555
556 s_rule->type = (f_info->flag & ICE_FLTR_RX) ?
557 cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_RX) :
558 cpu_to_le16(ICE_AQC_SW_RULES_T_LKUP_TX);
559
560 /* Recipe set depending on lookup type */
561 s_rule->pdata.lkup_tx_rx.recipe_id = cpu_to_le16(f_info->lkup_type);
562 s_rule->pdata.lkup_tx_rx.src = cpu_to_le16(f_info->src);
563 s_rule->pdata.lkup_tx_rx.act = cpu_to_le32(act);
564
565 if (daddr)
566 ether_addr_copy(ð_hdr[ICE_ETH_DA_OFFSET], daddr);
567
568 if (!(vlan_id > ICE_MAX_VLAN_ID)) {
569 off = (__be16 *)ð_hdr[ICE_ETH_VLAN_TCI_OFFSET];
570 *off = cpu_to_be16(vlan_id);
571 }
572
573 /* Create the switch rule with the final dummy Ethernet header */
574 if (opc != ice_aqc_opc_update_sw_rules)
575 s_rule->pdata.lkup_tx_rx.hdr_len = cpu_to_le16(sizeof(eth_hdr));
576
577 memcpy(s_rule->pdata.lkup_tx_rx.hdr, eth_hdr, sizeof(eth_hdr));
578}
579
580/**
581 * ice_add_marker_act
582 * @hw: pointer to the hardware structure
583 * @m_ent: the management entry for which sw marker needs to be added
584 * @sw_marker: sw marker to tag the Rx descriptor with
585 * @l_id: large action resource id
586 *
587 * Create a large action to hold software marker and update the switch rule
588 * entry pointed by m_ent with newly created large action
589 */
590static enum ice_status
591ice_add_marker_act(struct ice_hw *hw, struct ice_fltr_mgmt_list_entry *m_ent,
592 u16 sw_marker, u16 l_id)
593{
594 struct ice_aqc_sw_rules_elem *lg_act, *rx_tx;
595 /* For software marker we need 3 large actions
596 * 1. FWD action: FWD TO VSI or VSI LIST
597 * 2. GENERIC VALUE action to hold the profile id
598 * 3. GENERIC VALUE action to hold the software marker id
599 */
600 const u16 num_lg_acts = 3;
601 enum ice_status status;
602 u16 lg_act_size;
603 u16 rules_size;
604 u16 vsi_info;
605 u32 act;
606
607 if (m_ent->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
608 return ICE_ERR_PARAM;
609
610 /* Create two back-to-back switch rules and submit them to the HW using
611 * one memory buffer:
612 * 1. Large Action
613 * 2. Look up tx rx
614 */
615 lg_act_size = (u16)ICE_SW_RULE_LG_ACT_SIZE(num_lg_acts);
616 rules_size = lg_act_size + ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
617 lg_act = devm_kzalloc(ice_hw_to_dev(hw), rules_size, GFP_KERNEL);
618 if (!lg_act)
619 return ICE_ERR_NO_MEMORY;
620
621 rx_tx = (struct ice_aqc_sw_rules_elem *)((u8 *)lg_act + lg_act_size);
622
623 /* Fill in the first switch rule i.e. large action */
624 lg_act->type = cpu_to_le16(ICE_AQC_SW_RULES_T_LG_ACT);
625 lg_act->pdata.lg_act.index = cpu_to_le16(l_id);
626 lg_act->pdata.lg_act.size = cpu_to_le16(num_lg_acts);
627
628 /* First action VSI forwarding or VSI list forwarding depending on how
629 * many VSIs
630 */
631 vsi_info = (m_ent->vsi_count > 1) ?
632 m_ent->fltr_info.fwd_id.vsi_list_id :
633 m_ent->fltr_info.fwd_id.vsi_id;
634
635 act = ICE_LG_ACT_VSI_FORWARDING | ICE_LG_ACT_VALID_BIT;
636 act |= (vsi_info << ICE_LG_ACT_VSI_LIST_ID_S) &
637 ICE_LG_ACT_VSI_LIST_ID_M;
638 if (m_ent->vsi_count > 1)
639 act |= ICE_LG_ACT_VSI_LIST;
640 lg_act->pdata.lg_act.act[0] = cpu_to_le32(act);
641
642 /* Second action descriptor type */
643 act = ICE_LG_ACT_GENERIC;
644
645 act |= (1 << ICE_LG_ACT_GENERIC_VALUE_S) & ICE_LG_ACT_GENERIC_VALUE_M;
646 lg_act->pdata.lg_act.act[1] = cpu_to_le32(act);
647
648 act = (7 << ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_VALUE_M;
649
650 /* Third action Marker value */
651 act |= ICE_LG_ACT_GENERIC;
652 act |= (sw_marker << ICE_LG_ACT_GENERIC_VALUE_S) &
653 ICE_LG_ACT_GENERIC_VALUE_M;
654
655 act |= (0 << ICE_LG_ACT_GENERIC_OFFSET_S) & ICE_LG_ACT_GENERIC_VALUE_M;
656 lg_act->pdata.lg_act.act[2] = cpu_to_le32(act);
657
658 /* call the fill switch rule to fill the lookup tx rx structure */
659 ice_fill_sw_rule(hw, &m_ent->fltr_info, rx_tx,
660 ice_aqc_opc_update_sw_rules);
661
662 /* Update the action to point to the large action id */
663 rx_tx->pdata.lkup_tx_rx.act =
664 cpu_to_le32(ICE_SINGLE_ACT_PTR |
665 ((l_id << ICE_SINGLE_ACT_PTR_VAL_S) &
666 ICE_SINGLE_ACT_PTR_VAL_M));
667
668 /* Use the filter rule id of the previously created rule with single
669 * act. Once the update happens, hardware will treat this as large
670 * action
671 */
672 rx_tx->pdata.lkup_tx_rx.index =
673 cpu_to_le16(m_ent->fltr_info.fltr_rule_id);
674
675 status = ice_aq_sw_rules(hw, lg_act, rules_size, 2,
676 ice_aqc_opc_update_sw_rules, NULL);
677 if (!status) {
678 m_ent->lg_act_idx = l_id;
679 m_ent->sw_marker_id = sw_marker;
680 }
681
682 devm_kfree(ice_hw_to_dev(hw), lg_act);
683 return status;
684}
685
686/**
687 * ice_create_vsi_list_map
688 * @hw: pointer to the hardware structure
689 * @vsi_array: array of VSIs to form a VSI list
690 * @num_vsi: num VSI in the array
691 * @vsi_list_id: VSI list id generated as part of allocate resource
692 *
693 * Helper function to create a new entry of VSI list id to VSI mapping
694 * using the given VSI list id
695 */
696static struct ice_vsi_list_map_info *
697ice_create_vsi_list_map(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
698 u16 vsi_list_id)
699{
700 struct ice_switch_info *sw = hw->switch_info;
701 struct ice_vsi_list_map_info *v_map;
702 int i;
703
704 v_map = devm_kcalloc(ice_hw_to_dev(hw), 1, sizeof(*v_map), GFP_KERNEL);
705 if (!v_map)
706 return NULL;
707
708 v_map->vsi_list_id = vsi_list_id;
709
710 for (i = 0; i < num_vsi; i++)
711 set_bit(vsi_array[i], v_map->vsi_map);
712
713 list_add(&v_map->list_entry, &sw->vsi_list_map_head);
714 return v_map;
715}
716
717/**
718 * ice_update_vsi_list_rule
719 * @hw: pointer to the hardware structure
720 * @vsi_array: array of VSIs to form a VSI list
721 * @num_vsi: num VSI in the array
722 * @vsi_list_id: VSI list id generated as part of allocate resource
723 * @remove: Boolean value to indicate if this is a remove action
724 * @opc: switch rules population command type - pass in the command opcode
725 * @lkup_type: lookup type of the filter
726 *
727 * Call AQ command to add a new switch rule or update existing switch rule
728 * using the given VSI list id
729 */
730static enum ice_status
731ice_update_vsi_list_rule(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
732 u16 vsi_list_id, bool remove, enum ice_adminq_opc opc,
733 enum ice_sw_lkup_type lkup_type)
734{
735 struct ice_aqc_sw_rules_elem *s_rule;
736 enum ice_status status;
737 u16 s_rule_size;
738 u16 type;
739 int i;
740
741 if (!num_vsi)
742 return ICE_ERR_PARAM;
743
744 if (lkup_type == ICE_SW_LKUP_MAC ||
745 lkup_type == ICE_SW_LKUP_MAC_VLAN ||
746 lkup_type == ICE_SW_LKUP_ETHERTYPE ||
747 lkup_type == ICE_SW_LKUP_ETHERTYPE_MAC ||
748 lkup_type == ICE_SW_LKUP_PROMISC ||
749 lkup_type == ICE_SW_LKUP_PROMISC_VLAN)
750 type = remove ? ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR :
751 ICE_AQC_SW_RULES_T_VSI_LIST_SET;
752 else if (lkup_type == ICE_SW_LKUP_VLAN)
753 type = remove ? ICE_AQC_SW_RULES_T_PRUNE_LIST_CLEAR :
754 ICE_AQC_SW_RULES_T_PRUNE_LIST_SET;
755 else
756 return ICE_ERR_PARAM;
757
758 s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(num_vsi);
759 s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
760 if (!s_rule)
761 return ICE_ERR_NO_MEMORY;
762
763 for (i = 0; i < num_vsi; i++)
764 s_rule->pdata.vsi_list.vsi[i] = cpu_to_le16(vsi_array[i]);
765
766 s_rule->type = cpu_to_le16(type);
767 s_rule->pdata.vsi_list.number_vsi = cpu_to_le16(num_vsi);
768 s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
769
770 status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opc, NULL);
771
772 devm_kfree(ice_hw_to_dev(hw), s_rule);
773 return status;
774}
775
776/**
777 * ice_create_vsi_list_rule - Creates and populates a VSI list rule
778 * @hw: pointer to the hw struct
779 * @vsi_array: array of VSIs to form a VSI list
780 * @num_vsi: number of VSIs in the array
781 * @vsi_list_id: stores the ID of the VSI list to be created
782 * @lkup_type: switch rule filter's lookup type
783 */
784static enum ice_status
785ice_create_vsi_list_rule(struct ice_hw *hw, u16 *vsi_array, u16 num_vsi,
786 u16 *vsi_list_id, enum ice_sw_lkup_type lkup_type)
787{
788 enum ice_status status;
789 int i;
790
791 for (i = 0; i < num_vsi; i++)
792 if (vsi_array[i] >= ICE_MAX_VSI)
793 return ICE_ERR_OUT_OF_RANGE;
794
795 status = ice_aq_alloc_free_vsi_list(hw, vsi_list_id, lkup_type,
796 ice_aqc_opc_alloc_res);
797 if (status)
798 return status;
799
800 /* Update the newly created VSI list to include the specified VSIs */
801 return ice_update_vsi_list_rule(hw, vsi_array, num_vsi, *vsi_list_id,
802 false, ice_aqc_opc_add_sw_rules,
803 lkup_type);
804}
805
806/**
807 * ice_create_pkt_fwd_rule
808 * @hw: pointer to the hardware structure
809 * @f_entry: entry containing packet forwarding information
810 *
811 * Create switch rule with given filter information and add an entry
812 * to the corresponding filter management list to track this switch rule
813 * and VSI mapping
814 */
815static enum ice_status
816ice_create_pkt_fwd_rule(struct ice_hw *hw,
817 struct ice_fltr_list_entry *f_entry)
818{
819 struct ice_switch_info *sw = hw->switch_info;
820 struct ice_fltr_mgmt_list_entry *fm_entry;
821 struct ice_aqc_sw_rules_elem *s_rule;
822 enum ice_sw_lkup_type l_type;
823 enum ice_status status;
824
825 s_rule = devm_kzalloc(ice_hw_to_dev(hw),
826 ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
827 if (!s_rule)
828 return ICE_ERR_NO_MEMORY;
829 fm_entry = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*fm_entry),
830 GFP_KERNEL);
831 if (!fm_entry) {
832 status = ICE_ERR_NO_MEMORY;
833 goto ice_create_pkt_fwd_rule_exit;
834 }
835
836 fm_entry->fltr_info = f_entry->fltr_info;
837
838 /* Initialize all the fields for the management entry */
839 fm_entry->vsi_count = 1;
840 fm_entry->lg_act_idx = ICE_INVAL_LG_ACT_INDEX;
841 fm_entry->sw_marker_id = ICE_INVAL_SW_MARKER_ID;
842 fm_entry->counter_index = ICE_INVAL_COUNTER_ID;
843
844 ice_fill_sw_rule(hw, &fm_entry->fltr_info, s_rule,
845 ice_aqc_opc_add_sw_rules);
846
847 status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
848 ice_aqc_opc_add_sw_rules, NULL);
849 if (status) {
850 devm_kfree(ice_hw_to_dev(hw), fm_entry);
851 goto ice_create_pkt_fwd_rule_exit;
852 }
853
854 f_entry->fltr_info.fltr_rule_id =
855 le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
856 fm_entry->fltr_info.fltr_rule_id =
857 le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
858
859 /* The book keeping entries will get removed when base driver
860 * calls remove filter AQ command
861 */
862 l_type = fm_entry->fltr_info.lkup_type;
863 if (l_type == ICE_SW_LKUP_MAC) {
864 mutex_lock(&sw->mac_list_lock);
865 list_add(&fm_entry->list_entry, &sw->mac_list_head);
866 mutex_unlock(&sw->mac_list_lock);
867 } else if (l_type == ICE_SW_LKUP_VLAN) {
868 mutex_lock(&sw->vlan_list_lock);
869 list_add(&fm_entry->list_entry, &sw->vlan_list_head);
870 mutex_unlock(&sw->vlan_list_lock);
871 } else if (l_type == ICE_SW_LKUP_ETHERTYPE ||
872 l_type == ICE_SW_LKUP_ETHERTYPE_MAC) {
873 mutex_lock(&sw->eth_m_list_lock);
874 list_add(&fm_entry->list_entry, &sw->eth_m_list_head);
875 mutex_unlock(&sw->eth_m_list_lock);
876 } else if (l_type == ICE_SW_LKUP_PROMISC ||
877 l_type == ICE_SW_LKUP_PROMISC_VLAN) {
878 mutex_lock(&sw->promisc_list_lock);
879 list_add(&fm_entry->list_entry, &sw->promisc_list_head);
880 mutex_unlock(&sw->promisc_list_lock);
881 } else if (fm_entry->fltr_info.lkup_type == ICE_SW_LKUP_MAC_VLAN) {
882 mutex_lock(&sw->mac_vlan_list_lock);
883 list_add(&fm_entry->list_entry, &sw->mac_vlan_list_head);
884 mutex_unlock(&sw->mac_vlan_list_lock);
885 } else {
886 status = ICE_ERR_NOT_IMPL;
887 }
888ice_create_pkt_fwd_rule_exit:
889 devm_kfree(ice_hw_to_dev(hw), s_rule);
890 return status;
891}
892
893/**
894 * ice_update_pkt_fwd_rule
895 * @hw: pointer to the hardware structure
896 * @rule_id: rule of previously created switch rule to update
897 * @vsi_list_id: VSI list id to be updated with
898 * @f_info: ice_fltr_info to pull other information for switch rule
899 *
900 * Call AQ command to update a previously created switch rule with a
901 * VSI list id
902 */
903static enum ice_status
904ice_update_pkt_fwd_rule(struct ice_hw *hw, u16 rule_id, u16 vsi_list_id,
905 struct ice_fltr_info f_info)
906{
907 struct ice_aqc_sw_rules_elem *s_rule;
908 struct ice_fltr_info tmp_fltr;
909 enum ice_status status;
910
911 s_rule = devm_kzalloc(ice_hw_to_dev(hw),
912 ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, GFP_KERNEL);
913 if (!s_rule)
914 return ICE_ERR_NO_MEMORY;
915
916 tmp_fltr = f_info;
917 tmp_fltr.fltr_act = ICE_FWD_TO_VSI_LIST;
918 tmp_fltr.fwd_id.vsi_list_id = vsi_list_id;
919
920 ice_fill_sw_rule(hw, &tmp_fltr, s_rule,
921 ice_aqc_opc_update_sw_rules);
922
923 s_rule->pdata.lkup_tx_rx.index = cpu_to_le16(rule_id);
924
925 /* Update switch rule with new rule set to forward VSI list */
926 status = ice_aq_sw_rules(hw, s_rule, ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
927 ice_aqc_opc_update_sw_rules, NULL);
928
929 devm_kfree(ice_hw_to_dev(hw), s_rule);
930 return status;
931}
932
933/**
934 * ice_handle_vsi_list_mgmt
935 * @hw: pointer to the hardware structure
936 * @m_entry: pointer to current filter management list entry
937 * @cur_fltr: filter information from the book keeping entry
938 * @new_fltr: filter information with the new VSI to be added
939 *
940 * Call AQ command to add or update previously created VSI list with new VSI.
941 *
942 * Helper function to do book keeping associated with adding filter information
943 * The algorithm to do the booking keeping is described below :
944 * When a VSI needs to subscribe to a given filter( MAC/VLAN/Ethtype etc.)
945 * if only one VSI has been added till now
946 * Allocate a new VSI list and add two VSIs
947 * to this list using switch rule command
948 * Update the previously created switch rule with the
949 * newly created VSI list id
950 * if a VSI list was previously created
951 * Add the new VSI to the previously created VSI list set
952 * using the update switch rule command
953 */
954static enum ice_status
955ice_handle_vsi_list_mgmt(struct ice_hw *hw,
956 struct ice_fltr_mgmt_list_entry *m_entry,
957 struct ice_fltr_info *cur_fltr,
958 struct ice_fltr_info *new_fltr)
959{
960 enum ice_status status = 0;
961 u16 vsi_list_id = 0;
962
963 if ((cur_fltr->fltr_act == ICE_FWD_TO_Q ||
964 cur_fltr->fltr_act == ICE_FWD_TO_QGRP))
965 return ICE_ERR_NOT_IMPL;
966
967 if ((new_fltr->fltr_act == ICE_FWD_TO_Q ||
968 new_fltr->fltr_act == ICE_FWD_TO_QGRP) &&
969 (cur_fltr->fltr_act == ICE_FWD_TO_VSI ||
970 cur_fltr->fltr_act == ICE_FWD_TO_VSI_LIST))
971 return ICE_ERR_NOT_IMPL;
972
973 if (m_entry->vsi_count < 2 && !m_entry->vsi_list_info) {
974 /* Only one entry existed in the mapping and it was not already
975 * a part of a VSI list. So, create a VSI list with the old and
976 * new VSIs.
977 */
978 u16 vsi_id_arr[2];
979 u16 fltr_rule;
980
981 /* A rule already exists with the new VSI being added */
982 if (cur_fltr->fwd_id.vsi_id == new_fltr->fwd_id.vsi_id)
983 return ICE_ERR_ALREADY_EXISTS;
984
985 vsi_id_arr[0] = cur_fltr->fwd_id.vsi_id;
986 vsi_id_arr[1] = new_fltr->fwd_id.vsi_id;
987 status = ice_create_vsi_list_rule(hw, &vsi_id_arr[0], 2,
988 &vsi_list_id,
989 new_fltr->lkup_type);
990 if (status)
991 return status;
992
993 fltr_rule = cur_fltr->fltr_rule_id;
994 /* Update the previous switch rule of "MAC forward to VSI" to
995 * "MAC fwd to VSI list"
996 */
997 status = ice_update_pkt_fwd_rule(hw, fltr_rule, vsi_list_id,
998 *new_fltr);
999 if (status)
1000 return status;
1001
1002 cur_fltr->fwd_id.vsi_list_id = vsi_list_id;
1003 cur_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1004 m_entry->vsi_list_info =
1005 ice_create_vsi_list_map(hw, &vsi_id_arr[0], 2,
1006 vsi_list_id);
1007
1008 /* If this entry was large action then the large action needs
1009 * to be updated to point to FWD to VSI list
1010 */
1011 if (m_entry->sw_marker_id != ICE_INVAL_SW_MARKER_ID)
1012 status =
1013 ice_add_marker_act(hw, m_entry,
1014 m_entry->sw_marker_id,
1015 m_entry->lg_act_idx);
1016 } else {
1017 u16 vsi_id = new_fltr->fwd_id.vsi_id;
1018 enum ice_adminq_opc opcode;
1019
1020 /* A rule already exists with the new VSI being added */
1021 if (test_bit(vsi_id, m_entry->vsi_list_info->vsi_map))
1022 return 0;
1023
1024 /* Update the previously created VSI list set with
1025 * the new VSI id passed in
1026 */
1027 vsi_list_id = cur_fltr->fwd_id.vsi_list_id;
1028 opcode = ice_aqc_opc_update_sw_rules;
1029
1030 status = ice_update_vsi_list_rule(hw, &vsi_id, 1, vsi_list_id,
1031 false, opcode,
1032 new_fltr->lkup_type);
1033 /* update VSI list mapping info with new VSI id */
1034 if (!status)
1035 set_bit(vsi_id, m_entry->vsi_list_info->vsi_map);
1036 }
1037 if (!status)
1038 m_entry->vsi_count++;
1039 return status;
1040}
1041
1042/**
1043 * ice_find_mac_entry
1044 * @hw: pointer to the hardware structure
1045 * @mac_addr: MAC address to search for
1046 *
1047 * Helper function to search for a MAC entry using a given MAC address
1048 * Returns pointer to the entry if found.
1049 */
1050static struct ice_fltr_mgmt_list_entry *
1051ice_find_mac_entry(struct ice_hw *hw, u8 *mac_addr)
1052{
1053 struct ice_fltr_mgmt_list_entry *m_list_itr, *mac_ret = NULL;
1054 struct ice_switch_info *sw = hw->switch_info;
1055
1056 mutex_lock(&sw->mac_list_lock);
1057 list_for_each_entry(m_list_itr, &sw->mac_list_head, list_entry) {
1058 u8 *buf = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1059
1060 if (ether_addr_equal(buf, mac_addr)) {
1061 mac_ret = m_list_itr;
1062 break;
1063 }
1064 }
1065 mutex_unlock(&sw->mac_list_lock);
1066 return mac_ret;
1067}
1068
1069/**
1070 * ice_add_shared_mac - Add one MAC shared filter rule
1071 * @hw: pointer to the hardware structure
1072 * @f_entry: structure containing MAC forwarding information
1073 *
1074 * Adds or updates the book keeping list for the MAC addresses
1075 */
1076static enum ice_status
1077ice_add_shared_mac(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1078{
1079 struct ice_fltr_info *new_fltr, *cur_fltr;
1080 struct ice_fltr_mgmt_list_entry *m_entry;
1081
1082 new_fltr = &f_entry->fltr_info;
1083
1084 m_entry = ice_find_mac_entry(hw, &new_fltr->l_data.mac.mac_addr[0]);
1085 if (!m_entry)
1086 return ice_create_pkt_fwd_rule(hw, f_entry);
1087
1088 cur_fltr = &m_entry->fltr_info;
1089
1090 return ice_handle_vsi_list_mgmt(hw, m_entry, cur_fltr, new_fltr);
1091}
1092
1093/**
1094 * ice_add_mac - Add a MAC address based filter rule
1095 * @hw: pointer to the hardware structure
1096 * @m_list: list of MAC addresses and forwarding information
1097 *
1098 * IMPORTANT: When the ucast_shared flag is set to false and m_list has
1099 * multiple unicast addresses, the function assumes that all the
1100 * addresses are unique in a given add_mac call. It doesn't
1101 * check for duplicates in this case, removing duplicates from a given
1102 * list should be taken care of in the caller of this function.
1103 */
1104enum ice_status
1105ice_add_mac(struct ice_hw *hw, struct list_head *m_list)
1106{
1107 struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1108 struct ice_fltr_list_entry *m_list_itr;
1109 u16 elem_sent, total_elem_left;
1110 enum ice_status status = 0;
1111 u16 num_unicast = 0;
1112 u16 s_rule_size;
1113
1114 if (!m_list || !hw)
1115 return ICE_ERR_PARAM;
1116
1117 list_for_each_entry(m_list_itr, m_list, list_entry) {
1118 u8 *add = &m_list_itr->fltr_info.l_data.mac.mac_addr[0];
1119
1120 if (m_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_MAC)
1121 return ICE_ERR_PARAM;
1122 if (is_zero_ether_addr(add))
1123 return ICE_ERR_PARAM;
1124 if (is_unicast_ether_addr(add) && !hw->ucast_shared) {
1125 /* Don't overwrite the unicast address */
1126 if (ice_find_mac_entry(hw, add))
1127 return ICE_ERR_ALREADY_EXISTS;
1128 num_unicast++;
1129 } else if (is_multicast_ether_addr(add) ||
1130 (is_unicast_ether_addr(add) && hw->ucast_shared)) {
1131 status = ice_add_shared_mac(hw, m_list_itr);
1132 if (status) {
1133 m_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1134 return status;
1135 }
1136 m_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
1137 }
1138 }
1139
1140 /* Exit if no suitable entries were found for adding bulk switch rule */
1141 if (!num_unicast)
1142 return 0;
1143
1144 /* Allocate switch rule buffer for the bulk update for unicast */
1145 s_rule_size = ICE_SW_RULE_RX_TX_ETH_HDR_SIZE;
1146 s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1147 GFP_KERNEL);
1148 if (!s_rule)
1149 return ICE_ERR_NO_MEMORY;
1150
1151 r_iter = s_rule;
1152 list_for_each_entry(m_list_itr, m_list, list_entry) {
1153 struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1154 u8 *addr = &f_info->l_data.mac.mac_addr[0];
1155
1156 if (is_unicast_ether_addr(addr)) {
1157 ice_fill_sw_rule(hw, &m_list_itr->fltr_info,
1158 r_iter, ice_aqc_opc_add_sw_rules);
1159 r_iter = (struct ice_aqc_sw_rules_elem *)
1160 ((u8 *)r_iter + s_rule_size);
1161 }
1162 }
1163
1164 /* Call AQ bulk switch rule update for all unicast addresses */
1165 r_iter = s_rule;
1166 /* Call AQ switch rule in AQ_MAX chunk */
1167 for (total_elem_left = num_unicast; total_elem_left > 0;
1168 total_elem_left -= elem_sent) {
1169 struct ice_aqc_sw_rules_elem *entry = r_iter;
1170
1171 elem_sent = min(total_elem_left,
1172 (u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1173 status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1174 elem_sent, ice_aqc_opc_add_sw_rules,
1175 NULL);
1176 if (status)
1177 goto ice_add_mac_exit;
1178 r_iter = (struct ice_aqc_sw_rules_elem *)
1179 ((u8 *)r_iter + (elem_sent * s_rule_size));
1180 }
1181
1182 /* Fill up rule id based on the value returned from FW */
1183 r_iter = s_rule;
1184 list_for_each_entry(m_list_itr, m_list, list_entry) {
1185 struct ice_fltr_info *f_info = &m_list_itr->fltr_info;
1186 u8 *addr = &f_info->l_data.mac.mac_addr[0];
1187 struct ice_switch_info *sw = hw->switch_info;
1188 struct ice_fltr_mgmt_list_entry *fm_entry;
1189
1190 if (is_unicast_ether_addr(addr)) {
1191 f_info->fltr_rule_id =
1192 le16_to_cpu(r_iter->pdata.lkup_tx_rx.index);
1193 f_info->fltr_act = ICE_FWD_TO_VSI;
1194 /* Create an entry to track this MAC address */
1195 fm_entry = devm_kzalloc(ice_hw_to_dev(hw),
1196 sizeof(*fm_entry), GFP_KERNEL);
1197 if (!fm_entry) {
1198 status = ICE_ERR_NO_MEMORY;
1199 goto ice_add_mac_exit;
1200 }
1201 fm_entry->fltr_info = *f_info;
1202 fm_entry->vsi_count = 1;
1203 /* The book keeping entries will get removed when
1204 * base driver calls remove filter AQ command
1205 */
1206 mutex_lock(&sw->mac_list_lock);
1207 list_add(&fm_entry->list_entry, &sw->mac_list_head);
1208 mutex_unlock(&sw->mac_list_lock);
1209
1210 r_iter = (struct ice_aqc_sw_rules_elem *)
1211 ((u8 *)r_iter + s_rule_size);
1212 }
1213 }
1214
1215ice_add_mac_exit:
1216 devm_kfree(ice_hw_to_dev(hw), s_rule);
1217 return status;
1218}
1219
1220/**
1221 * ice_find_vlan_entry
1222 * @hw: pointer to the hardware structure
1223 * @vlan_id: VLAN id to search for
1224 *
1225 * Helper function to search for a VLAN entry using a given VLAN id
1226 * Returns pointer to the entry if found.
1227 */
1228static struct ice_fltr_mgmt_list_entry *
1229ice_find_vlan_entry(struct ice_hw *hw, u16 vlan_id)
1230{
1231 struct ice_fltr_mgmt_list_entry *vlan_list_itr, *vlan_ret = NULL;
1232 struct ice_switch_info *sw = hw->switch_info;
1233
1234 mutex_lock(&sw->vlan_list_lock);
1235 list_for_each_entry(vlan_list_itr, &sw->vlan_list_head, list_entry)
1236 if (vlan_list_itr->fltr_info.l_data.vlan.vlan_id == vlan_id) {
1237 vlan_ret = vlan_list_itr;
1238 break;
1239 }
1240
1241 mutex_unlock(&sw->vlan_list_lock);
1242 return vlan_ret;
1243}
1244
1245/**
1246 * ice_add_vlan_internal - Add one VLAN based filter rule
1247 * @hw: pointer to the hardware structure
1248 * @f_entry: filter entry containing one VLAN information
1249 */
1250static enum ice_status
1251ice_add_vlan_internal(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1252{
1253 struct ice_fltr_info *new_fltr, *cur_fltr;
1254 struct ice_fltr_mgmt_list_entry *v_list_itr;
1255 u16 vlan_id;
1256
1257 new_fltr = &f_entry->fltr_info;
1258 /* VLAN id should only be 12 bits */
1259 if (new_fltr->l_data.vlan.vlan_id > ICE_MAX_VLAN_ID)
1260 return ICE_ERR_PARAM;
1261
1262 vlan_id = new_fltr->l_data.vlan.vlan_id;
1263 v_list_itr = ice_find_vlan_entry(hw, vlan_id);
1264 if (!v_list_itr) {
1265 u16 vsi_id = ICE_VSI_INVAL_ID;
1266 enum ice_status status;
1267 u16 vsi_list_id = 0;
1268
1269 if (new_fltr->fltr_act == ICE_FWD_TO_VSI) {
1270 enum ice_sw_lkup_type lkup_type = new_fltr->lkup_type;
1271
1272 /* All VLAN pruning rules use a VSI list.
1273 * Convert the action to forwarding to a VSI list.
1274 */
1275 vsi_id = new_fltr->fwd_id.vsi_id;
1276 status = ice_create_vsi_list_rule(hw, &vsi_id, 1,
1277 &vsi_list_id,
1278 lkup_type);
1279 if (status)
1280 return status;
1281 new_fltr->fltr_act = ICE_FWD_TO_VSI_LIST;
1282 new_fltr->fwd_id.vsi_list_id = vsi_list_id;
1283 }
1284
1285 status = ice_create_pkt_fwd_rule(hw, f_entry);
1286 if (!status && vsi_id != ICE_VSI_INVAL_ID) {
1287 v_list_itr = ice_find_vlan_entry(hw, vlan_id);
1288 if (!v_list_itr)
1289 return ICE_ERR_DOES_NOT_EXIST;
1290 v_list_itr->vsi_list_info =
1291 ice_create_vsi_list_map(hw, &vsi_id, 1,
1292 vsi_list_id);
1293 }
1294
1295 return status;
1296 }
1297
1298 cur_fltr = &v_list_itr->fltr_info;
1299 return ice_handle_vsi_list_mgmt(hw, v_list_itr, cur_fltr, new_fltr);
1300}
1301
1302/**
1303 * ice_add_vlan - Add VLAN based filter rule
1304 * @hw: pointer to the hardware structure
1305 * @v_list: list of VLAN entries and forwarding information
1306 */
1307enum ice_status
1308ice_add_vlan(struct ice_hw *hw, struct list_head *v_list)
1309{
1310 struct ice_fltr_list_entry *v_list_itr;
1311
1312 if (!v_list || !hw)
1313 return ICE_ERR_PARAM;
1314
1315 list_for_each_entry(v_list_itr, v_list, list_entry) {
1316 enum ice_status status;
1317
1318 if (v_list_itr->fltr_info.lkup_type != ICE_SW_LKUP_VLAN)
1319 return ICE_ERR_PARAM;
1320
1321 status = ice_add_vlan_internal(hw, v_list_itr);
1322 if (status) {
1323 v_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1324 return status;
1325 }
1326 v_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
1327 }
1328 return 0;
1329}
1330
1331/**
1332 * ice_remove_vsi_list_rule
1333 * @hw: pointer to the hardware structure
1334 * @vsi_list_id: VSI list id generated as part of allocate resource
1335 * @lkup_type: switch rule filter lookup type
1336 */
1337static enum ice_status
1338ice_remove_vsi_list_rule(struct ice_hw *hw, u16 vsi_list_id,
1339 enum ice_sw_lkup_type lkup_type)
1340{
1341 struct ice_aqc_sw_rules_elem *s_rule;
1342 enum ice_status status;
1343 u16 s_rule_size;
1344
1345 s_rule_size = (u16)ICE_SW_RULE_VSI_LIST_SIZE(0);
1346 s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1347 if (!s_rule)
1348 return ICE_ERR_NO_MEMORY;
1349
1350 s_rule->type = cpu_to_le16(ICE_AQC_SW_RULES_T_VSI_LIST_CLEAR);
1351 s_rule->pdata.vsi_list.index = cpu_to_le16(vsi_list_id);
1352 /* FW expects number of VSIs in vsi_list resource to be 0 for clear
1353 * command. Since memory is zero'ed out during initialization, it's not
1354 * necessary to explicitly initialize the variable to 0.
1355 */
1356
1357 status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1,
1358 ice_aqc_opc_remove_sw_rules, NULL);
1359 if (!status)
1360 /* Free the vsi_list resource that we allocated */
1361 status = ice_aq_alloc_free_vsi_list(hw, &vsi_list_id, lkup_type,
1362 ice_aqc_opc_free_res);
1363
1364 devm_kfree(ice_hw_to_dev(hw), s_rule);
1365 return status;
1366}
1367
1368/**
1369 * ice_handle_rem_vsi_list_mgmt
1370 * @hw: pointer to the hardware structure
1371 * @vsi_id: ID of the VSI to remove
1372 * @fm_list_itr: filter management entry for which the VSI list management
1373 * needs to be done
1374 */
1375static enum ice_status
1376ice_handle_rem_vsi_list_mgmt(struct ice_hw *hw, u16 vsi_id,
1377 struct ice_fltr_mgmt_list_entry *fm_list_itr)
1378{
1379 struct ice_switch_info *sw = hw->switch_info;
1380 enum ice_status status = 0;
1381 enum ice_sw_lkup_type lkup_type;
1382 bool is_last_elem = true;
1383 bool conv_list = false;
1384 bool del_list = false;
1385 u16 vsi_list_id;
1386
1387 lkup_type = fm_list_itr->fltr_info.lkup_type;
1388 vsi_list_id = fm_list_itr->fltr_info.fwd_id.vsi_list_id;
1389
1390 if (fm_list_itr->vsi_count > 1) {
1391 status = ice_update_vsi_list_rule(hw, &vsi_id, 1, vsi_list_id,
1392 true,
1393 ice_aqc_opc_update_sw_rules,
1394 lkup_type);
1395 if (status)
1396 return status;
1397 fm_list_itr->vsi_count--;
1398 is_last_elem = false;
1399 clear_bit(vsi_id, fm_list_itr->vsi_list_info->vsi_map);
1400 }
1401
1402 /* For non-VLAN rules that forward packets to a VSI list, convert them
1403 * to forwarding packets to a VSI if there is only one VSI left in the
1404 * list. Unused lists are then removed.
1405 * VLAN rules need to use VSI lists even with only one VSI.
1406 */
1407 if (fm_list_itr->fltr_info.fltr_act == ICE_FWD_TO_VSI_LIST) {
1408 if (lkup_type == ICE_SW_LKUP_VLAN) {
1409 del_list = is_last_elem;
1410 } else if (fm_list_itr->vsi_count == 1) {
1411 conv_list = true;
1412 del_list = true;
1413 }
1414 }
1415
1416 if (del_list) {
1417 /* Remove the VSI list since it is no longer used */
1418 struct ice_vsi_list_map_info *vsi_list_info =
1419 fm_list_itr->vsi_list_info;
1420
1421 status = ice_remove_vsi_list_rule(hw, vsi_list_id, lkup_type);
1422 if (status)
1423 return status;
1424
1425 if (conv_list) {
1426 u16 rem_vsi_id;
1427
1428 rem_vsi_id = find_first_bit(vsi_list_info->vsi_map,
1429 ICE_MAX_VSI);
1430
1431 /* Error out when the expected last element is not in
1432 * the VSI list map
1433 */
1434 if (rem_vsi_id == ICE_MAX_VSI)
1435 return ICE_ERR_OUT_OF_RANGE;
1436
1437 /* Change the list entry action from VSI_LIST to VSI */
1438 fm_list_itr->fltr_info.fltr_act = ICE_FWD_TO_VSI;
1439 fm_list_itr->fltr_info.fwd_id.vsi_id = rem_vsi_id;
1440 }
1441
1442 list_del(&vsi_list_info->list_entry);
1443 devm_kfree(ice_hw_to_dev(hw), vsi_list_info);
1444 fm_list_itr->vsi_list_info = NULL;
1445 }
1446
1447 if (conv_list) {
1448 /* Convert the rule's forward action to forwarding packets to
1449 * a VSI
1450 */
1451 struct ice_aqc_sw_rules_elem *s_rule;
1452
1453 s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1454 ICE_SW_RULE_RX_TX_ETH_HDR_SIZE,
1455 GFP_KERNEL);
1456 if (!s_rule)
1457 return ICE_ERR_NO_MEMORY;
1458
1459 ice_fill_sw_rule(hw, &fm_list_itr->fltr_info, s_rule,
1460 ice_aqc_opc_update_sw_rules);
1461
1462 s_rule->pdata.lkup_tx_rx.index =
1463 cpu_to_le16(fm_list_itr->fltr_info.fltr_rule_id);
1464
1465 status = ice_aq_sw_rules(hw, s_rule,
1466 ICE_SW_RULE_RX_TX_ETH_HDR_SIZE, 1,
1467 ice_aqc_opc_update_sw_rules, NULL);
1468 devm_kfree(ice_hw_to_dev(hw), s_rule);
1469 if (status)
1470 return status;
1471 }
1472
1473 if (is_last_elem) {
1474 /* Remove the lookup rule */
1475 struct ice_aqc_sw_rules_elem *s_rule;
1476
1477 s_rule = devm_kzalloc(ice_hw_to_dev(hw),
1478 ICE_SW_RULE_RX_TX_NO_HDR_SIZE,
1479 GFP_KERNEL);
1480 if (!s_rule)
1481 return ICE_ERR_NO_MEMORY;
1482
1483 ice_fill_sw_rule(hw, &fm_list_itr->fltr_info, s_rule,
1484 ice_aqc_opc_remove_sw_rules);
1485
1486 status = ice_aq_sw_rules(hw, s_rule,
1487 ICE_SW_RULE_RX_TX_NO_HDR_SIZE, 1,
1488 ice_aqc_opc_remove_sw_rules, NULL);
1489 if (status)
1490 return status;
1491
1492 /* Remove a book keeping entry from the MAC address list */
1493 mutex_lock(&sw->mac_list_lock);
1494 list_del(&fm_list_itr->list_entry);
1495 mutex_unlock(&sw->mac_list_lock);
1496 devm_kfree(ice_hw_to_dev(hw), fm_list_itr);
1497 devm_kfree(ice_hw_to_dev(hw), s_rule);
1498 }
1499 return status;
1500}
1501
1502/**
1503 * ice_remove_mac_entry
1504 * @hw: pointer to the hardware structure
1505 * @f_entry: structure containing MAC forwarding information
1506 */
1507static enum ice_status
1508ice_remove_mac_entry(struct ice_hw *hw, struct ice_fltr_list_entry *f_entry)
1509{
1510 struct ice_fltr_mgmt_list_entry *m_entry;
1511 u16 vsi_id;
1512 u8 *add;
1513
1514 add = &f_entry->fltr_info.l_data.mac.mac_addr[0];
1515
1516 m_entry = ice_find_mac_entry(hw, add);
1517 if (!m_entry)
1518 return ICE_ERR_PARAM;
1519
1520 vsi_id = f_entry->fltr_info.fwd_id.vsi_id;
1521 return ice_handle_rem_vsi_list_mgmt(hw, vsi_id, m_entry);
1522}
1523
1524/**
1525 * ice_remove_mac - remove a MAC address based filter rule
1526 * @hw: pointer to the hardware structure
1527 * @m_list: list of MAC addresses and forwarding information
1528 *
1529 * This function removes either a MAC filter rule or a specific VSI from a
1530 * VSI list for a multicast MAC address.
1531 *
1532 * Returns ICE_ERR_DOES_NOT_EXIST if a given entry was not added by
1533 * ice_add_mac. Caller should be aware that this call will only work if all
1534 * the entries passed into m_list were added previously. It will not attempt to
1535 * do a partial remove of entries that were found.
1536 */
1537enum ice_status
1538ice_remove_mac(struct ice_hw *hw, struct list_head *m_list)
1539{
1540 struct ice_aqc_sw_rules_elem *s_rule, *r_iter;
1541 u8 s_rule_size = ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
1542 struct ice_switch_info *sw = hw->switch_info;
1543 struct ice_fltr_mgmt_list_entry *m_entry;
1544 struct ice_fltr_list_entry *m_list_itr;
1545 u16 elem_sent, total_elem_left;
1546 enum ice_status status = 0;
1547 u16 num_unicast = 0;
1548
1549 if (!m_list)
1550 return ICE_ERR_PARAM;
1551
1552 list_for_each_entry(m_list_itr, m_list, list_entry) {
1553 u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1554
1555 if (is_unicast_ether_addr(addr) && !hw->ucast_shared)
1556 num_unicast++;
1557 else if (is_multicast_ether_addr(addr) ||
1558 (is_unicast_ether_addr(addr) && hw->ucast_shared))
1559 ice_remove_mac_entry(hw, m_list_itr);
1560 }
1561
1562 /* Exit if no unicast addresses found. Multicast switch rules
1563 * were added individually
1564 */
1565 if (!num_unicast)
1566 return 0;
1567
1568 /* Allocate switch rule buffer for the bulk update for unicast */
1569 s_rule = devm_kcalloc(ice_hw_to_dev(hw), num_unicast, s_rule_size,
1570 GFP_KERNEL);
1571 if (!s_rule)
1572 return ICE_ERR_NO_MEMORY;
1573
1574 r_iter = s_rule;
1575 list_for_each_entry(m_list_itr, m_list, list_entry) {
1576 u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1577
1578 if (is_unicast_ether_addr(addr)) {
1579 m_entry = ice_find_mac_entry(hw, addr);
1580 if (!m_entry) {
1581 status = ICE_ERR_DOES_NOT_EXIST;
1582 goto ice_remove_mac_exit;
1583 }
1584
1585 ice_fill_sw_rule(hw, &m_entry->fltr_info,
1586 r_iter, ice_aqc_opc_remove_sw_rules);
1587 r_iter = (struct ice_aqc_sw_rules_elem *)
1588 ((u8 *)r_iter + s_rule_size);
1589 }
1590 }
1591
1592 /* Call AQ bulk switch rule update for all unicast addresses */
1593 r_iter = s_rule;
1594 /* Call AQ switch rule in AQ_MAX chunk */
1595 for (total_elem_left = num_unicast; total_elem_left > 0;
1596 total_elem_left -= elem_sent) {
1597 struct ice_aqc_sw_rules_elem *entry = r_iter;
1598
1599 elem_sent = min(total_elem_left,
1600 (u16)(ICE_AQ_MAX_BUF_LEN / s_rule_size));
1601 status = ice_aq_sw_rules(hw, entry, elem_sent * s_rule_size,
1602 elem_sent, ice_aqc_opc_remove_sw_rules,
1603 NULL);
1604 if (status)
1605 break;
1606 r_iter = (struct ice_aqc_sw_rules_elem *)
1607 ((u8 *)r_iter + s_rule_size);
1608 }
1609
1610 list_for_each_entry(m_list_itr, m_list, list_entry) {
1611 u8 *addr = m_list_itr->fltr_info.l_data.mac.mac_addr;
1612
1613 if (is_unicast_ether_addr(addr)) {
1614 m_entry = ice_find_mac_entry(hw, addr);
1615 if (!m_entry)
1616 return ICE_ERR_OUT_OF_RANGE;
1617 mutex_lock(&sw->mac_list_lock);
1618 list_del(&m_entry->list_entry);
1619 mutex_unlock(&sw->mac_list_lock);
1620 devm_kfree(ice_hw_to_dev(hw), m_entry);
1621 }
1622 }
1623
1624ice_remove_mac_exit:
1625 devm_kfree(ice_hw_to_dev(hw), s_rule);
1626 return status;
1627}
1628
1629/**
1630 * ice_cfg_dflt_vsi - add filter rule to set/unset given VSI as default
1631 * VSI for the switch (represented by swid)
1632 * @hw: pointer to the hardware structure
1633 * @vsi_id: number of VSI to set as default
1634 * @set: true to add the above mentioned switch rule, false to remove it
1635 * @direction: ICE_FLTR_RX or ICE_FLTR_TX
1636 */
1637enum ice_status
1638ice_cfg_dflt_vsi(struct ice_hw *hw, u16 vsi_id, bool set, u8 direction)
1639{
1640 struct ice_aqc_sw_rules_elem *s_rule;
1641 struct ice_fltr_info f_info;
1642 enum ice_adminq_opc opcode;
1643 enum ice_status status;
1644 u16 s_rule_size;
1645
1646 s_rule_size = set ? ICE_SW_RULE_RX_TX_ETH_HDR_SIZE :
1647 ICE_SW_RULE_RX_TX_NO_HDR_SIZE;
1648 s_rule = devm_kzalloc(ice_hw_to_dev(hw), s_rule_size, GFP_KERNEL);
1649 if (!s_rule)
1650 return ICE_ERR_NO_MEMORY;
1651
1652 memset(&f_info, 0, sizeof(f_info));
1653
1654 f_info.lkup_type = ICE_SW_LKUP_DFLT;
1655 f_info.flag = direction;
1656 f_info.fltr_act = ICE_FWD_TO_VSI;
1657 f_info.fwd_id.vsi_id = vsi_id;
1658
1659 if (f_info.flag & ICE_FLTR_RX) {
1660 f_info.src = hw->port_info->lport;
1661 if (!set)
1662 f_info.fltr_rule_id =
1663 hw->port_info->dflt_rx_vsi_rule_id;
1664 } else if (f_info.flag & ICE_FLTR_TX) {
1665 f_info.src = vsi_id;
1666 if (!set)
1667 f_info.fltr_rule_id =
1668 hw->port_info->dflt_tx_vsi_rule_id;
1669 }
1670
1671 if (set)
1672 opcode = ice_aqc_opc_add_sw_rules;
1673 else
1674 opcode = ice_aqc_opc_remove_sw_rules;
1675
1676 ice_fill_sw_rule(hw, &f_info, s_rule, opcode);
1677
1678 status = ice_aq_sw_rules(hw, s_rule, s_rule_size, 1, opcode, NULL);
1679 if (status || !(f_info.flag & ICE_FLTR_TX_RX))
1680 goto out;
1681 if (set) {
1682 u16 index = le16_to_cpu(s_rule->pdata.lkup_tx_rx.index);
1683
1684 if (f_info.flag & ICE_FLTR_TX) {
1685 hw->port_info->dflt_tx_vsi_num = vsi_id;
1686 hw->port_info->dflt_tx_vsi_rule_id = index;
1687 } else if (f_info.flag & ICE_FLTR_RX) {
1688 hw->port_info->dflt_rx_vsi_num = vsi_id;
1689 hw->port_info->dflt_rx_vsi_rule_id = index;
1690 }
1691 } else {
1692 if (f_info.flag & ICE_FLTR_TX) {
1693 hw->port_info->dflt_tx_vsi_num = ICE_DFLT_VSI_INVAL;
1694 hw->port_info->dflt_tx_vsi_rule_id = ICE_INVAL_ACT;
1695 } else if (f_info.flag & ICE_FLTR_RX) {
1696 hw->port_info->dflt_rx_vsi_num = ICE_DFLT_VSI_INVAL;
1697 hw->port_info->dflt_rx_vsi_rule_id = ICE_INVAL_ACT;
1698 }
1699 }
1700
1701out:
1702 devm_kfree(ice_hw_to_dev(hw), s_rule);
1703 return status;
1704}
1705
1706/**
1707 * ice_remove_vlan_internal - Remove one VLAN based filter rule
1708 * @hw: pointer to the hardware structure
1709 * @f_entry: filter entry containing one VLAN information
1710 */
1711static enum ice_status
1712ice_remove_vlan_internal(struct ice_hw *hw,
1713 struct ice_fltr_list_entry *f_entry)
1714{
1715 struct ice_fltr_info *new_fltr;
1716 struct ice_fltr_mgmt_list_entry *v_list_elem;
1717 u16 vsi_id;
1718
1719 new_fltr = &f_entry->fltr_info;
1720
1721 v_list_elem = ice_find_vlan_entry(hw, new_fltr->l_data.vlan.vlan_id);
1722 if (!v_list_elem)
1723 return ICE_ERR_PARAM;
1724
1725 vsi_id = f_entry->fltr_info.fwd_id.vsi_id;
1726 return ice_handle_rem_vsi_list_mgmt(hw, vsi_id, v_list_elem);
1727}
1728
1729/**
1730 * ice_remove_vlan - Remove VLAN based filter rule
1731 * @hw: pointer to the hardware structure
1732 * @v_list: list of VLAN entries and forwarding information
1733 */
1734enum ice_status
1735ice_remove_vlan(struct ice_hw *hw, struct list_head *v_list)
1736{
1737 struct ice_fltr_list_entry *v_list_itr;
1738 enum ice_status status = 0;
1739
1740 if (!v_list || !hw)
1741 return ICE_ERR_PARAM;
1742
1743 list_for_each_entry(v_list_itr, v_list, list_entry) {
1744 status = ice_remove_vlan_internal(hw, v_list_itr);
1745 if (status) {
1746 v_list_itr->status = ICE_FLTR_STATUS_FW_FAIL;
1747 return status;
1748 }
1749 v_list_itr->status = ICE_FLTR_STATUS_FW_SUCCESS;
1750 }
1751 return status;
1752}
1753
1754/**
1755 * ice_add_to_vsi_fltr_list - Add VSI filters to the list
1756 * @hw: pointer to the hardware structure
1757 * @vsi_id: ID of VSI to remove filters from
1758 * @lkup_list_head: pointer to the list that has certain lookup type filters
1759 * @vsi_list_head: pointer to the list pertaining to VSI with vsi_id
1760 */
1761static enum ice_status
1762ice_add_to_vsi_fltr_list(struct ice_hw *hw, u16 vsi_id,
1763 struct list_head *lkup_list_head,
1764 struct list_head *vsi_list_head)
1765{
1766 struct ice_fltr_mgmt_list_entry *fm_entry;
1767
1768 /* check to make sure VSI id is valid and within boundary */
1769 if (vsi_id >=
1770 (sizeof(fm_entry->vsi_list_info->vsi_map) * BITS_PER_BYTE - 1))
1771 return ICE_ERR_PARAM;
1772
1773 list_for_each_entry(fm_entry, lkup_list_head, list_entry) {
1774 struct ice_fltr_info *fi;
1775
1776 fi = &fm_entry->fltr_info;
1777 if ((fi->fltr_act == ICE_FWD_TO_VSI &&
1778 fi->fwd_id.vsi_id == vsi_id) ||
1779 (fi->fltr_act == ICE_FWD_TO_VSI_LIST &&
1780 (test_bit(vsi_id, fm_entry->vsi_list_info->vsi_map)))) {
1781 struct ice_fltr_list_entry *tmp;
1782
1783 /* this memory is freed up in the caller function
1784 * ice_remove_vsi_lkup_fltr() once filters for
1785 * this VSI are removed
1786 */
1787 tmp = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*tmp),
1788 GFP_KERNEL);
1789 if (!tmp)
1790 return ICE_ERR_NO_MEMORY;
1791
1792 memcpy(&tmp->fltr_info, fi, sizeof(*fi));
1793
1794 /* Expected below fields to be set to ICE_FWD_TO_VSI and
1795 * the particular VSI id since we are only removing this
1796 * one VSI
1797 */
1798 if (fi->fltr_act == ICE_FWD_TO_VSI_LIST) {
1799 tmp->fltr_info.fltr_act = ICE_FWD_TO_VSI;
1800 tmp->fltr_info.fwd_id.vsi_id = vsi_id;
1801 }
1802
1803 list_add(&tmp->list_entry, vsi_list_head);
1804 }
1805 }
1806 return 0;
1807}
1808
1809/**
1810 * ice_remove_vsi_lkup_fltr - Remove lookup type filters for a VSI
1811 * @hw: pointer to the hardware structure
1812 * @vsi_id: ID of VSI to remove filters from
1813 * @lkup: switch rule filter lookup type
1814 */
1815static void
1816ice_remove_vsi_lkup_fltr(struct ice_hw *hw, u16 vsi_id,
1817 enum ice_sw_lkup_type lkup)
1818{
1819 struct ice_switch_info *sw = hw->switch_info;
1820 struct ice_fltr_list_entry *fm_entry;
1821 struct list_head remove_list_head;
1822 struct ice_fltr_list_entry *tmp;
1823 enum ice_status status;
1824
1825 INIT_LIST_HEAD(&remove_list_head);
1826 switch (lkup) {
1827 case ICE_SW_LKUP_MAC:
1828 mutex_lock(&sw->mac_list_lock);
1829 status = ice_add_to_vsi_fltr_list(hw, vsi_id,
1830 &sw->mac_list_head,
1831 &remove_list_head);
1832 mutex_unlock(&sw->mac_list_lock);
1833 if (!status) {
1834 ice_remove_mac(hw, &remove_list_head);
1835 goto free_fltr_list;
1836 }
1837 break;
1838 case ICE_SW_LKUP_VLAN:
1839 mutex_lock(&sw->vlan_list_lock);
1840 status = ice_add_to_vsi_fltr_list(hw, vsi_id,
1841 &sw->vlan_list_head,
1842 &remove_list_head);
1843 mutex_unlock(&sw->vlan_list_lock);
1844 if (!status) {
1845 ice_remove_vlan(hw, &remove_list_head);
1846 goto free_fltr_list;
1847 }
1848 break;
1849 case ICE_SW_LKUP_MAC_VLAN:
1850 case ICE_SW_LKUP_ETHERTYPE:
1851 case ICE_SW_LKUP_ETHERTYPE_MAC:
1852 case ICE_SW_LKUP_PROMISC:
1853 case ICE_SW_LKUP_PROMISC_VLAN:
1854 case ICE_SW_LKUP_DFLT:
1855 ice_debug(hw, ICE_DBG_SW,
1856 "Remove filters for this lookup type hasn't been implemented yet\n");
1857 break;
1858 }
1859
1860 return;
1861free_fltr_list:
1862 list_for_each_entry_safe(fm_entry, tmp, &remove_list_head, list_entry) {
1863 list_del(&fm_entry->list_entry);
1864 devm_kfree(ice_hw_to_dev(hw), fm_entry);
1865 }
1866}
1867
1868/**
1869 * ice_remove_vsi_fltr - Remove all filters for a VSI
1870 * @hw: pointer to the hardware structure
1871 * @vsi_id: ID of VSI to remove filters from
1872 */
1873void ice_remove_vsi_fltr(struct ice_hw *hw, u16 vsi_id)
1874{
1875 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_MAC);
1876 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_MAC_VLAN);
1877 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_PROMISC);
1878 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_VLAN);
1879 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_DFLT);
1880 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_ETHERTYPE);
1881 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_ETHERTYPE_MAC);
1882 ice_remove_vsi_lkup_fltr(hw, vsi_id, ICE_SW_LKUP_PROMISC_VLAN);
1883}