Loading...
1// SPDX-License-Identifier: GPL-2.0
2/* Copyright (C) 2021-2023, Intel Corporation. */
3
4#include "ice.h"
5#include "ice_base.h"
6#include "ice_lib.h"
7#include "ice_flow.h"
8#include "ice_vf_lib_private.h"
9
10#define to_fltr_conf_from_desc(p) \
11 container_of(p, struct virtchnl_fdir_fltr_conf, input)
12
13#define GTPU_TEID_OFFSET 4
14#define GTPU_EH_QFI_OFFSET 1
15#define GTPU_EH_QFI_MASK 0x3F
16#define PFCP_S_OFFSET 0
17#define PFCP_S_MASK 0x1
18#define PFCP_PORT_NR 8805
19
20#define FDIR_INSET_FLAG_ESP_S 0
21#define FDIR_INSET_FLAG_ESP_M BIT_ULL(FDIR_INSET_FLAG_ESP_S)
22#define FDIR_INSET_FLAG_ESP_UDP BIT_ULL(FDIR_INSET_FLAG_ESP_S)
23#define FDIR_INSET_FLAG_ESP_IPSEC (0ULL << FDIR_INSET_FLAG_ESP_S)
24
25enum ice_fdir_tunnel_type {
26 ICE_FDIR_TUNNEL_TYPE_NONE = 0,
27 ICE_FDIR_TUNNEL_TYPE_GTPU,
28 ICE_FDIR_TUNNEL_TYPE_GTPU_EH,
29 ICE_FDIR_TUNNEL_TYPE_ECPRI,
30 ICE_FDIR_TUNNEL_TYPE_GTPU_INNER,
31 ICE_FDIR_TUNNEL_TYPE_GTPU_EH_INNER,
32 ICE_FDIR_TUNNEL_TYPE_GRE,
33 ICE_FDIR_TUNNEL_TYPE_GTPOGRE,
34 ICE_FDIR_TUNNEL_TYPE_GTPOGRE_INNER,
35 ICE_FDIR_TUNNEL_TYPE_GRE_INNER,
36 ICE_FDIR_TUNNEL_TYPE_L2TPV2,
37 ICE_FDIR_TUNNEL_TYPE_L2TPV2_INNER,
38};
39
40struct virtchnl_fdir_fltr_conf {
41 struct ice_fdir_fltr input;
42 enum ice_fdir_tunnel_type ttype;
43 u64 inset_flag;
44 u32 flow_id;
45
46 struct ice_parser_profile *prof;
47 bool parser_ena;
48 u8 *pkt_buf;
49 u8 pkt_len;
50};
51
52struct virtchnl_fdir_inset_map {
53 enum virtchnl_proto_hdr_field field;
54 enum ice_flow_field fld;
55 u64 flag;
56 u64 mask;
57};
58
59static const struct virtchnl_fdir_inset_map fdir_inset_map[] = {
60 {VIRTCHNL_PROTO_HDR_ETH_ETHERTYPE, ICE_FLOW_FIELD_IDX_ETH_TYPE, 0, 0},
61 {VIRTCHNL_PROTO_HDR_IPV4_SRC, ICE_FLOW_FIELD_IDX_IPV4_SA, 0, 0},
62 {VIRTCHNL_PROTO_HDR_IPV4_DST, ICE_FLOW_FIELD_IDX_IPV4_DA, 0, 0},
63 {VIRTCHNL_PROTO_HDR_IPV4_DSCP, ICE_FLOW_FIELD_IDX_IPV4_DSCP, 0, 0},
64 {VIRTCHNL_PROTO_HDR_IPV4_TTL, ICE_FLOW_FIELD_IDX_IPV4_TTL, 0, 0},
65 {VIRTCHNL_PROTO_HDR_IPV4_PROT, ICE_FLOW_FIELD_IDX_IPV4_PROT, 0, 0},
66 {VIRTCHNL_PROTO_HDR_IPV6_SRC, ICE_FLOW_FIELD_IDX_IPV6_SA, 0, 0},
67 {VIRTCHNL_PROTO_HDR_IPV6_DST, ICE_FLOW_FIELD_IDX_IPV6_DA, 0, 0},
68 {VIRTCHNL_PROTO_HDR_IPV6_TC, ICE_FLOW_FIELD_IDX_IPV6_DSCP, 0, 0},
69 {VIRTCHNL_PROTO_HDR_IPV6_HOP_LIMIT, ICE_FLOW_FIELD_IDX_IPV6_TTL, 0, 0},
70 {VIRTCHNL_PROTO_HDR_IPV6_PROT, ICE_FLOW_FIELD_IDX_IPV6_PROT, 0, 0},
71 {VIRTCHNL_PROTO_HDR_UDP_SRC_PORT, ICE_FLOW_FIELD_IDX_UDP_SRC_PORT, 0, 0},
72 {VIRTCHNL_PROTO_HDR_UDP_DST_PORT, ICE_FLOW_FIELD_IDX_UDP_DST_PORT, 0, 0},
73 {VIRTCHNL_PROTO_HDR_TCP_SRC_PORT, ICE_FLOW_FIELD_IDX_TCP_SRC_PORT, 0, 0},
74 {VIRTCHNL_PROTO_HDR_TCP_DST_PORT, ICE_FLOW_FIELD_IDX_TCP_DST_PORT, 0, 0},
75 {VIRTCHNL_PROTO_HDR_SCTP_SRC_PORT, ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT, 0, 0},
76 {VIRTCHNL_PROTO_HDR_SCTP_DST_PORT, ICE_FLOW_FIELD_IDX_SCTP_DST_PORT, 0, 0},
77 {VIRTCHNL_PROTO_HDR_GTPU_IP_TEID, ICE_FLOW_FIELD_IDX_GTPU_IP_TEID, 0, 0},
78 {VIRTCHNL_PROTO_HDR_GTPU_EH_QFI, ICE_FLOW_FIELD_IDX_GTPU_EH_QFI, 0, 0},
79 {VIRTCHNL_PROTO_HDR_ESP_SPI, ICE_FLOW_FIELD_IDX_ESP_SPI,
80 FDIR_INSET_FLAG_ESP_IPSEC, FDIR_INSET_FLAG_ESP_M},
81 {VIRTCHNL_PROTO_HDR_ESP_SPI, ICE_FLOW_FIELD_IDX_NAT_T_ESP_SPI,
82 FDIR_INSET_FLAG_ESP_UDP, FDIR_INSET_FLAG_ESP_M},
83 {VIRTCHNL_PROTO_HDR_AH_SPI, ICE_FLOW_FIELD_IDX_AH_SPI, 0, 0},
84 {VIRTCHNL_PROTO_HDR_L2TPV3_SESS_ID, ICE_FLOW_FIELD_IDX_L2TPV3_SESS_ID, 0, 0},
85 {VIRTCHNL_PROTO_HDR_PFCP_S_FIELD, ICE_FLOW_FIELD_IDX_UDP_DST_PORT, 0, 0},
86};
87
88/**
89 * ice_vc_fdir_param_check
90 * @vf: pointer to the VF structure
91 * @vsi_id: VF relative VSI ID
92 *
93 * Check for the valid VSI ID, PF's state and VF's state
94 *
95 * Return: 0 on success, and -EINVAL on error.
96 */
97static int
98ice_vc_fdir_param_check(struct ice_vf *vf, u16 vsi_id)
99{
100 struct ice_pf *pf = vf->pf;
101
102 if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
103 return -EINVAL;
104
105 if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states))
106 return -EINVAL;
107
108 if (!(vf->driver_caps & VIRTCHNL_VF_OFFLOAD_FDIR_PF))
109 return -EINVAL;
110
111 if (!ice_vc_isvalid_vsi_id(vf, vsi_id))
112 return -EINVAL;
113
114 if (!ice_get_vf_vsi(vf))
115 return -EINVAL;
116
117 return 0;
118}
119
120/**
121 * ice_vf_start_ctrl_vsi
122 * @vf: pointer to the VF structure
123 *
124 * Allocate ctrl_vsi for the first time and open the ctrl_vsi port for VF
125 *
126 * Return: 0 on success, and other on error.
127 */
128static int ice_vf_start_ctrl_vsi(struct ice_vf *vf)
129{
130 struct ice_pf *pf = vf->pf;
131 struct ice_vsi *ctrl_vsi;
132 struct device *dev;
133 int err;
134
135 dev = ice_pf_to_dev(pf);
136 if (vf->ctrl_vsi_idx != ICE_NO_VSI)
137 return -EEXIST;
138
139 ctrl_vsi = ice_vf_ctrl_vsi_setup(vf);
140 if (!ctrl_vsi) {
141 dev_dbg(dev, "Could not setup control VSI for VF %d\n",
142 vf->vf_id);
143 return -ENOMEM;
144 }
145
146 err = ice_vsi_open_ctrl(ctrl_vsi);
147 if (err) {
148 dev_dbg(dev, "Could not open control VSI for VF %d\n",
149 vf->vf_id);
150 goto err_vsi_open;
151 }
152
153 return 0;
154
155err_vsi_open:
156 ice_vsi_release(ctrl_vsi);
157 if (vf->ctrl_vsi_idx != ICE_NO_VSI) {
158 pf->vsi[vf->ctrl_vsi_idx] = NULL;
159 vf->ctrl_vsi_idx = ICE_NO_VSI;
160 }
161 return err;
162}
163
164/**
165 * ice_vc_fdir_alloc_prof - allocate profile for this filter flow type
166 * @vf: pointer to the VF structure
167 * @flow: filter flow type
168 *
169 * Return: 0 on success, and other on error.
170 */
171static int
172ice_vc_fdir_alloc_prof(struct ice_vf *vf, enum ice_fltr_ptype flow)
173{
174 struct ice_vf_fdir *fdir = &vf->fdir;
175
176 if (!fdir->fdir_prof) {
177 fdir->fdir_prof = devm_kcalloc(ice_pf_to_dev(vf->pf),
178 ICE_FLTR_PTYPE_MAX,
179 sizeof(*fdir->fdir_prof),
180 GFP_KERNEL);
181 if (!fdir->fdir_prof)
182 return -ENOMEM;
183 }
184
185 if (!fdir->fdir_prof[flow]) {
186 fdir->fdir_prof[flow] = devm_kzalloc(ice_pf_to_dev(vf->pf),
187 sizeof(**fdir->fdir_prof),
188 GFP_KERNEL);
189 if (!fdir->fdir_prof[flow])
190 return -ENOMEM;
191 }
192
193 return 0;
194}
195
196/**
197 * ice_vc_fdir_free_prof - free profile for this filter flow type
198 * @vf: pointer to the VF structure
199 * @flow: filter flow type
200 */
201static void
202ice_vc_fdir_free_prof(struct ice_vf *vf, enum ice_fltr_ptype flow)
203{
204 struct ice_vf_fdir *fdir = &vf->fdir;
205
206 if (!fdir->fdir_prof)
207 return;
208
209 if (!fdir->fdir_prof[flow])
210 return;
211
212 devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof[flow]);
213 fdir->fdir_prof[flow] = NULL;
214}
215
216/**
217 * ice_vc_fdir_free_prof_all - free all the profile for this VF
218 * @vf: pointer to the VF structure
219 */
220static void ice_vc_fdir_free_prof_all(struct ice_vf *vf)
221{
222 struct ice_vf_fdir *fdir = &vf->fdir;
223 enum ice_fltr_ptype flow;
224
225 if (!fdir->fdir_prof)
226 return;
227
228 for (flow = ICE_FLTR_PTYPE_NONF_NONE; flow < ICE_FLTR_PTYPE_MAX; flow++)
229 ice_vc_fdir_free_prof(vf, flow);
230
231 devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof);
232 fdir->fdir_prof = NULL;
233}
234
235/**
236 * ice_vc_fdir_parse_flow_fld
237 * @proto_hdr: virtual channel protocol filter header
238 * @conf: FDIR configuration for each filter
239 * @fld: field type array
240 * @fld_cnt: field counter
241 *
242 * Parse the virtual channel filter header and store them into field type array
243 *
244 * Return: 0 on success, and other on error.
245 */
246static int
247ice_vc_fdir_parse_flow_fld(struct virtchnl_proto_hdr *proto_hdr,
248 struct virtchnl_fdir_fltr_conf *conf,
249 enum ice_flow_field *fld, int *fld_cnt)
250{
251 struct virtchnl_proto_hdr hdr;
252 u32 i;
253
254 memcpy(&hdr, proto_hdr, sizeof(hdr));
255
256 for (i = 0; (i < ARRAY_SIZE(fdir_inset_map)) &&
257 VIRTCHNL_GET_PROTO_HDR_FIELD(&hdr); i++)
258 if (VIRTCHNL_TEST_PROTO_HDR(&hdr, fdir_inset_map[i].field)) {
259 if (fdir_inset_map[i].mask &&
260 ((fdir_inset_map[i].mask & conf->inset_flag) !=
261 fdir_inset_map[i].flag))
262 continue;
263
264 fld[*fld_cnt] = fdir_inset_map[i].fld;
265 *fld_cnt += 1;
266 if (*fld_cnt >= ICE_FLOW_FIELD_IDX_MAX)
267 return -EINVAL;
268 VIRTCHNL_DEL_PROTO_HDR_FIELD(&hdr,
269 fdir_inset_map[i].field);
270 }
271
272 return 0;
273}
274
275/**
276 * ice_vc_fdir_set_flow_fld
277 * @vf: pointer to the VF structure
278 * @fltr: virtual channel add cmd buffer
279 * @conf: FDIR configuration for each filter
280 * @seg: array of one or more packet segments that describe the flow
281 *
282 * Parse the virtual channel add msg buffer's field vector and store them into
283 * flow's packet segment field
284 *
285 * Return: 0 on success, and other on error.
286 */
287static int
288ice_vc_fdir_set_flow_fld(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
289 struct virtchnl_fdir_fltr_conf *conf,
290 struct ice_flow_seg_info *seg)
291{
292 struct virtchnl_fdir_rule *rule = &fltr->rule_cfg;
293 enum ice_flow_field fld[ICE_FLOW_FIELD_IDX_MAX];
294 struct device *dev = ice_pf_to_dev(vf->pf);
295 struct virtchnl_proto_hdrs *proto;
296 int fld_cnt = 0;
297 int i;
298
299 proto = &rule->proto_hdrs;
300 for (i = 0; i < proto->count; i++) {
301 struct virtchnl_proto_hdr *hdr = &proto->proto_hdr[i];
302 int ret;
303
304 ret = ice_vc_fdir_parse_flow_fld(hdr, conf, fld, &fld_cnt);
305 if (ret)
306 return ret;
307 }
308
309 if (fld_cnt == 0) {
310 dev_dbg(dev, "Empty input set for VF %d\n", vf->vf_id);
311 return -EINVAL;
312 }
313
314 for (i = 0; i < fld_cnt; i++)
315 ice_flow_set_fld(seg, fld[i],
316 ICE_FLOW_FLD_OFF_INVAL,
317 ICE_FLOW_FLD_OFF_INVAL,
318 ICE_FLOW_FLD_OFF_INVAL, false);
319
320 return 0;
321}
322
323/**
324 * ice_vc_fdir_set_flow_hdr - config the flow's packet segment header
325 * @vf: pointer to the VF structure
326 * @conf: FDIR configuration for each filter
327 * @seg: array of one or more packet segments that describe the flow
328 *
329 * Return: 0 on success, and other on error.
330 */
331static int
332ice_vc_fdir_set_flow_hdr(struct ice_vf *vf,
333 struct virtchnl_fdir_fltr_conf *conf,
334 struct ice_flow_seg_info *seg)
335{
336 enum ice_fltr_ptype flow = conf->input.flow_type;
337 enum ice_fdir_tunnel_type ttype = conf->ttype;
338 struct device *dev = ice_pf_to_dev(vf->pf);
339
340 switch (flow) {
341 case ICE_FLTR_PTYPE_NON_IP_L2:
342 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ETH_NON_IP);
343 break;
344 case ICE_FLTR_PTYPE_NONF_IPV4_L2TPV3:
345 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
346 ICE_FLOW_SEG_HDR_IPV4 |
347 ICE_FLOW_SEG_HDR_IPV_OTHER);
348 break;
349 case ICE_FLTR_PTYPE_NONF_IPV4_ESP:
350 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
351 ICE_FLOW_SEG_HDR_IPV4 |
352 ICE_FLOW_SEG_HDR_IPV_OTHER);
353 break;
354 case ICE_FLTR_PTYPE_NONF_IPV4_AH:
355 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
356 ICE_FLOW_SEG_HDR_IPV4 |
357 ICE_FLOW_SEG_HDR_IPV_OTHER);
358 break;
359 case ICE_FLTR_PTYPE_NONF_IPV4_NAT_T_ESP:
360 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
361 ICE_FLOW_SEG_HDR_IPV4 |
362 ICE_FLOW_SEG_HDR_IPV_OTHER);
363 break;
364 case ICE_FLTR_PTYPE_NONF_IPV4_PFCP_NODE:
365 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
366 ICE_FLOW_SEG_HDR_IPV4 |
367 ICE_FLOW_SEG_HDR_IPV_OTHER);
368 break;
369 case ICE_FLTR_PTYPE_NONF_IPV4_PFCP_SESSION:
370 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
371 ICE_FLOW_SEG_HDR_IPV4 |
372 ICE_FLOW_SEG_HDR_IPV_OTHER);
373 break;
374 case ICE_FLTR_PTYPE_NONF_IPV4_OTHER:
375 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4 |
376 ICE_FLOW_SEG_HDR_IPV_OTHER);
377 break;
378 case ICE_FLTR_PTYPE_NONF_IPV4_TCP:
379 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
380 ICE_FLOW_SEG_HDR_IPV4 |
381 ICE_FLOW_SEG_HDR_IPV_OTHER);
382 break;
383 case ICE_FLTR_PTYPE_NONF_IPV4_UDP:
384 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
385 ICE_FLOW_SEG_HDR_IPV4 |
386 ICE_FLOW_SEG_HDR_IPV_OTHER);
387 break;
388 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_UDP:
389 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_TCP:
390 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_ICMP:
391 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_OTHER:
392 if (ttype == ICE_FDIR_TUNNEL_TYPE_GTPU) {
393 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_IP |
394 ICE_FLOW_SEG_HDR_IPV4 |
395 ICE_FLOW_SEG_HDR_IPV_OTHER);
396 } else if (ttype == ICE_FDIR_TUNNEL_TYPE_GTPU_EH) {
397 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_EH |
398 ICE_FLOW_SEG_HDR_GTPU_IP |
399 ICE_FLOW_SEG_HDR_IPV4 |
400 ICE_FLOW_SEG_HDR_IPV_OTHER);
401 } else {
402 dev_dbg(dev, "Invalid tunnel type 0x%x for VF %d\n",
403 flow, vf->vf_id);
404 return -EINVAL;
405 }
406 break;
407 case ICE_FLTR_PTYPE_NONF_IPV4_SCTP:
408 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
409 ICE_FLOW_SEG_HDR_IPV4 |
410 ICE_FLOW_SEG_HDR_IPV_OTHER);
411 break;
412 case ICE_FLTR_PTYPE_NONF_IPV6_L2TPV3:
413 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
414 ICE_FLOW_SEG_HDR_IPV6 |
415 ICE_FLOW_SEG_HDR_IPV_OTHER);
416 break;
417 case ICE_FLTR_PTYPE_NONF_IPV6_ESP:
418 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
419 ICE_FLOW_SEG_HDR_IPV6 |
420 ICE_FLOW_SEG_HDR_IPV_OTHER);
421 break;
422 case ICE_FLTR_PTYPE_NONF_IPV6_AH:
423 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
424 ICE_FLOW_SEG_HDR_IPV6 |
425 ICE_FLOW_SEG_HDR_IPV_OTHER);
426 break;
427 case ICE_FLTR_PTYPE_NONF_IPV6_NAT_T_ESP:
428 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
429 ICE_FLOW_SEG_HDR_IPV6 |
430 ICE_FLOW_SEG_HDR_IPV_OTHER);
431 break;
432 case ICE_FLTR_PTYPE_NONF_IPV6_PFCP_NODE:
433 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
434 ICE_FLOW_SEG_HDR_IPV6 |
435 ICE_FLOW_SEG_HDR_IPV_OTHER);
436 break;
437 case ICE_FLTR_PTYPE_NONF_IPV6_PFCP_SESSION:
438 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
439 ICE_FLOW_SEG_HDR_IPV6 |
440 ICE_FLOW_SEG_HDR_IPV_OTHER);
441 break;
442 case ICE_FLTR_PTYPE_NONF_IPV6_OTHER:
443 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6 |
444 ICE_FLOW_SEG_HDR_IPV_OTHER);
445 break;
446 case ICE_FLTR_PTYPE_NONF_IPV6_TCP:
447 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
448 ICE_FLOW_SEG_HDR_IPV6 |
449 ICE_FLOW_SEG_HDR_IPV_OTHER);
450 break;
451 case ICE_FLTR_PTYPE_NONF_IPV6_UDP:
452 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
453 ICE_FLOW_SEG_HDR_IPV6 |
454 ICE_FLOW_SEG_HDR_IPV_OTHER);
455 break;
456 case ICE_FLTR_PTYPE_NONF_IPV6_SCTP:
457 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
458 ICE_FLOW_SEG_HDR_IPV6 |
459 ICE_FLOW_SEG_HDR_IPV_OTHER);
460 break;
461 default:
462 dev_dbg(dev, "Invalid flow type 0x%x for VF %d failed\n",
463 flow, vf->vf_id);
464 return -EINVAL;
465 }
466
467 return 0;
468}
469
470/**
471 * ice_vc_fdir_rem_prof - remove profile for this filter flow type
472 * @vf: pointer to the VF structure
473 * @flow: filter flow type
474 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
475 */
476static void
477ice_vc_fdir_rem_prof(struct ice_vf *vf, enum ice_fltr_ptype flow, int tun)
478{
479 struct ice_vf_fdir *fdir = &vf->fdir;
480 struct ice_fd_hw_prof *vf_prof;
481 struct ice_pf *pf = vf->pf;
482 struct ice_vsi *vf_vsi;
483 struct device *dev;
484 struct ice_hw *hw;
485 u64 prof_id;
486 int i;
487
488 dev = ice_pf_to_dev(pf);
489 hw = &pf->hw;
490 if (!fdir->fdir_prof || !fdir->fdir_prof[flow])
491 return;
492
493 vf_prof = fdir->fdir_prof[flow];
494 prof_id = vf_prof->prof_id[tun];
495
496 vf_vsi = ice_get_vf_vsi(vf);
497 if (!vf_vsi) {
498 dev_dbg(dev, "NULL vf %d vsi pointer\n", vf->vf_id);
499 return;
500 }
501
502 if (!fdir->prof_entry_cnt[flow][tun])
503 return;
504
505 for (i = 0; i < fdir->prof_entry_cnt[flow][tun]; i++)
506 if (vf_prof->entry_h[i][tun]) {
507 u16 vsi_num = ice_get_hw_vsi_num(hw, vf_prof->vsi_h[i]);
508
509 ice_rem_prof_id_flow(hw, ICE_BLK_FD, vsi_num, prof_id);
510 ice_flow_rem_entry(hw, ICE_BLK_FD,
511 vf_prof->entry_h[i][tun]);
512 vf_prof->entry_h[i][tun] = 0;
513 }
514
515 ice_flow_rem_prof(hw, ICE_BLK_FD, prof_id);
516 devm_kfree(dev, vf_prof->fdir_seg[tun]);
517 vf_prof->fdir_seg[tun] = NULL;
518
519 for (i = 0; i < vf_prof->cnt; i++)
520 vf_prof->vsi_h[i] = 0;
521
522 fdir->prof_entry_cnt[flow][tun] = 0;
523}
524
525/**
526 * ice_vc_fdir_rem_prof_all - remove profile for this VF
527 * @vf: pointer to the VF structure
528 */
529static void ice_vc_fdir_rem_prof_all(struct ice_vf *vf)
530{
531 enum ice_fltr_ptype flow;
532
533 for (flow = ICE_FLTR_PTYPE_NONF_NONE;
534 flow < ICE_FLTR_PTYPE_MAX; flow++) {
535 ice_vc_fdir_rem_prof(vf, flow, 0);
536 ice_vc_fdir_rem_prof(vf, flow, 1);
537 }
538}
539
540/**
541 * ice_vc_fdir_reset_cnt_all - reset all FDIR counters for this VF FDIR
542 * @fdir: pointer to the VF FDIR structure
543 */
544static void ice_vc_fdir_reset_cnt_all(struct ice_vf_fdir *fdir)
545{
546 enum ice_fltr_ptype flow;
547
548 for (flow = ICE_FLTR_PTYPE_NONF_NONE;
549 flow < ICE_FLTR_PTYPE_MAX; flow++) {
550 fdir->fdir_fltr_cnt[flow][0] = 0;
551 fdir->fdir_fltr_cnt[flow][1] = 0;
552 }
553
554 fdir->fdir_fltr_cnt_total = 0;
555}
556
557/**
558 * ice_vc_fdir_has_prof_conflict
559 * @vf: pointer to the VF structure
560 * @conf: FDIR configuration for each filter
561 *
562 * Check if @conf has conflicting profile with existing profiles
563 *
564 * Return: true on success, and false on error.
565 */
566static bool
567ice_vc_fdir_has_prof_conflict(struct ice_vf *vf,
568 struct virtchnl_fdir_fltr_conf *conf)
569{
570 struct ice_fdir_fltr *desc;
571
572 list_for_each_entry(desc, &vf->fdir.fdir_rule_list, fltr_node) {
573 struct virtchnl_fdir_fltr_conf *existing_conf;
574 enum ice_fltr_ptype flow_type_a, flow_type_b;
575 struct ice_fdir_fltr *a, *b;
576
577 existing_conf = to_fltr_conf_from_desc(desc);
578 a = &existing_conf->input;
579 b = &conf->input;
580 flow_type_a = a->flow_type;
581 flow_type_b = b->flow_type;
582
583 /* No need to compare two rules with different tunnel types or
584 * with the same protocol type.
585 */
586 if (existing_conf->ttype != conf->ttype ||
587 flow_type_a == flow_type_b)
588 continue;
589
590 switch (flow_type_a) {
591 case ICE_FLTR_PTYPE_NONF_IPV4_UDP:
592 case ICE_FLTR_PTYPE_NONF_IPV4_TCP:
593 case ICE_FLTR_PTYPE_NONF_IPV4_SCTP:
594 if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_OTHER)
595 return true;
596 break;
597 case ICE_FLTR_PTYPE_NONF_IPV4_OTHER:
598 if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_UDP ||
599 flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_TCP ||
600 flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_SCTP)
601 return true;
602 break;
603 case ICE_FLTR_PTYPE_NONF_IPV6_UDP:
604 case ICE_FLTR_PTYPE_NONF_IPV6_TCP:
605 case ICE_FLTR_PTYPE_NONF_IPV6_SCTP:
606 if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_OTHER)
607 return true;
608 break;
609 case ICE_FLTR_PTYPE_NONF_IPV6_OTHER:
610 if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_UDP ||
611 flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_TCP ||
612 flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_SCTP)
613 return true;
614 break;
615 default:
616 break;
617 }
618 }
619
620 return false;
621}
622
623/**
624 * ice_vc_fdir_write_flow_prof
625 * @vf: pointer to the VF structure
626 * @flow: filter flow type
627 * @seg: array of one or more packet segments that describe the flow
628 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
629 *
630 * Write the flow's profile config and packet segment into the hardware
631 *
632 * Return: 0 on success, and other on error.
633 */
634static int
635ice_vc_fdir_write_flow_prof(struct ice_vf *vf, enum ice_fltr_ptype flow,
636 struct ice_flow_seg_info *seg, int tun)
637{
638 struct ice_vf_fdir *fdir = &vf->fdir;
639 struct ice_vsi *vf_vsi, *ctrl_vsi;
640 struct ice_flow_seg_info *old_seg;
641 struct ice_flow_prof *prof = NULL;
642 struct ice_fd_hw_prof *vf_prof;
643 struct device *dev;
644 struct ice_pf *pf;
645 struct ice_hw *hw;
646 u64 entry1_h = 0;
647 u64 entry2_h = 0;
648 int ret;
649
650 pf = vf->pf;
651 dev = ice_pf_to_dev(pf);
652 hw = &pf->hw;
653 vf_vsi = ice_get_vf_vsi(vf);
654 if (!vf_vsi)
655 return -EINVAL;
656
657 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
658 if (!ctrl_vsi)
659 return -EINVAL;
660
661 vf_prof = fdir->fdir_prof[flow];
662 old_seg = vf_prof->fdir_seg[tun];
663 if (old_seg) {
664 if (!memcmp(old_seg, seg, sizeof(*seg))) {
665 dev_dbg(dev, "Duplicated profile for VF %d!\n",
666 vf->vf_id);
667 return -EEXIST;
668 }
669
670 if (fdir->fdir_fltr_cnt[flow][tun]) {
671 ret = -EINVAL;
672 dev_dbg(dev, "Input set conflicts for VF %d\n",
673 vf->vf_id);
674 goto err_exit;
675 }
676
677 /* remove previously allocated profile */
678 ice_vc_fdir_rem_prof(vf, flow, tun);
679 }
680
681 ret = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, seg,
682 tun + 1, false, &prof);
683 if (ret) {
684 dev_dbg(dev, "Could not add VSI flow 0x%x for VF %d\n",
685 flow, vf->vf_id);
686 goto err_exit;
687 }
688
689 ret = ice_flow_add_entry(hw, ICE_BLK_FD, prof->id, vf_vsi->idx,
690 vf_vsi->idx, ICE_FLOW_PRIO_NORMAL,
691 seg, &entry1_h);
692 if (ret) {
693 dev_dbg(dev, "Could not add flow 0x%x VSI entry for VF %d\n",
694 flow, vf->vf_id);
695 goto err_prof;
696 }
697
698 ret = ice_flow_add_entry(hw, ICE_BLK_FD, prof->id, vf_vsi->idx,
699 ctrl_vsi->idx, ICE_FLOW_PRIO_NORMAL,
700 seg, &entry2_h);
701 if (ret) {
702 dev_dbg(dev,
703 "Could not add flow 0x%x Ctrl VSI entry for VF %d\n",
704 flow, vf->vf_id);
705 goto err_entry_1;
706 }
707
708 vf_prof->fdir_seg[tun] = seg;
709 vf_prof->cnt = 0;
710 fdir->prof_entry_cnt[flow][tun] = 0;
711
712 vf_prof->entry_h[vf_prof->cnt][tun] = entry1_h;
713 vf_prof->vsi_h[vf_prof->cnt] = vf_vsi->idx;
714 vf_prof->cnt++;
715 fdir->prof_entry_cnt[flow][tun]++;
716
717 vf_prof->entry_h[vf_prof->cnt][tun] = entry2_h;
718 vf_prof->vsi_h[vf_prof->cnt] = ctrl_vsi->idx;
719 vf_prof->cnt++;
720 fdir->prof_entry_cnt[flow][tun]++;
721
722 vf_prof->prof_id[tun] = prof->id;
723
724 return 0;
725
726err_entry_1:
727 ice_rem_prof_id_flow(hw, ICE_BLK_FD,
728 ice_get_hw_vsi_num(hw, vf_vsi->idx), prof->id);
729 ice_flow_rem_entry(hw, ICE_BLK_FD, entry1_h);
730err_prof:
731 ice_flow_rem_prof(hw, ICE_BLK_FD, prof->id);
732err_exit:
733 return ret;
734}
735
736/**
737 * ice_vc_fdir_config_input_set
738 * @vf: pointer to the VF structure
739 * @fltr: virtual channel add cmd buffer
740 * @conf: FDIR configuration for each filter
741 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
742 *
743 * Config the input set type and value for virtual channel add msg buffer
744 *
745 * Return: 0 on success, and other on error.
746 */
747static int
748ice_vc_fdir_config_input_set(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
749 struct virtchnl_fdir_fltr_conf *conf, int tun)
750{
751 struct ice_fdir_fltr *input = &conf->input;
752 struct device *dev = ice_pf_to_dev(vf->pf);
753 struct ice_flow_seg_info *seg;
754 enum ice_fltr_ptype flow;
755 int ret;
756
757 ret = ice_vc_fdir_has_prof_conflict(vf, conf);
758 if (ret) {
759 dev_dbg(dev, "Found flow profile conflict for VF %d\n",
760 vf->vf_id);
761 return ret;
762 }
763
764 flow = input->flow_type;
765 ret = ice_vc_fdir_alloc_prof(vf, flow);
766 if (ret) {
767 dev_dbg(dev, "Alloc flow prof for VF %d failed\n", vf->vf_id);
768 return ret;
769 }
770
771 seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
772 if (!seg)
773 return -ENOMEM;
774
775 ret = ice_vc_fdir_set_flow_fld(vf, fltr, conf, seg);
776 if (ret) {
777 dev_dbg(dev, "Set flow field for VF %d failed\n", vf->vf_id);
778 goto err_exit;
779 }
780
781 ret = ice_vc_fdir_set_flow_hdr(vf, conf, seg);
782 if (ret) {
783 dev_dbg(dev, "Set flow hdr for VF %d failed\n", vf->vf_id);
784 goto err_exit;
785 }
786
787 ret = ice_vc_fdir_write_flow_prof(vf, flow, seg, tun);
788 if (ret == -EEXIST) {
789 devm_kfree(dev, seg);
790 } else if (ret) {
791 dev_dbg(dev, "Write flow profile for VF %d failed\n",
792 vf->vf_id);
793 goto err_exit;
794 }
795
796 return 0;
797
798err_exit:
799 devm_kfree(dev, seg);
800 return ret;
801}
802
803/**
804 * ice_vc_fdir_is_raw_flow - check if FDIR flow is raw (binary)
805 * @proto: virtchnl protocol headers
806 *
807 * Check if the FDIR rule is raw flow (protocol agnostic flow) or not. Note
808 * that common FDIR rule must have non-zero proto->count. Thus, we choose the
809 * tunnel_level and count of proto as the indicators. If both tunnel_level and
810 * count of proto are zero, this FDIR rule will be regarded as raw flow.
811 *
812 * Returns: true if headers describe raw flow, false otherwise.
813 */
814static bool
815ice_vc_fdir_is_raw_flow(struct virtchnl_proto_hdrs *proto)
816{
817 return (proto->tunnel_level == 0 && proto->count == 0);
818}
819
820/**
821 * ice_vc_fdir_parse_raw - parse a virtchnl raw FDIR rule
822 * @vf: pointer to the VF info
823 * @proto: virtchnl protocol headers
824 * @conf: FDIR configuration for each filter
825 *
826 * Parse the virtual channel filter's raw flow and store it in @conf
827 *
828 * Return: 0 on success or negative errno on failure.
829 */
830static int
831ice_vc_fdir_parse_raw(struct ice_vf *vf,
832 struct virtchnl_proto_hdrs *proto,
833 struct virtchnl_fdir_fltr_conf *conf)
834{
835 u8 *pkt_buf, *msk_buf __free(kfree);
836 struct ice_parser_result rslt;
837 struct ice_pf *pf = vf->pf;
838 struct ice_parser *psr;
839 int status = -ENOMEM;
840 struct ice_hw *hw;
841 u16 udp_port = 0;
842
843 pkt_buf = kzalloc(proto->raw.pkt_len, GFP_KERNEL);
844 msk_buf = kzalloc(proto->raw.pkt_len, GFP_KERNEL);
845 if (!pkt_buf || !msk_buf)
846 goto err_mem_alloc;
847
848 memcpy(pkt_buf, proto->raw.spec, proto->raw.pkt_len);
849 memcpy(msk_buf, proto->raw.mask, proto->raw.pkt_len);
850
851 hw = &pf->hw;
852
853 /* Get raw profile info via Parser Lib */
854 psr = ice_parser_create(hw);
855 if (IS_ERR(psr)) {
856 status = PTR_ERR(psr);
857 goto err_mem_alloc;
858 }
859
860 ice_parser_dvm_set(psr, ice_is_dvm_ena(hw));
861
862 if (ice_get_open_tunnel_port(hw, &udp_port, TNL_VXLAN))
863 ice_parser_vxlan_tunnel_set(psr, udp_port, true);
864
865 status = ice_parser_run(psr, pkt_buf, proto->raw.pkt_len, &rslt);
866 if (status)
867 goto err_parser_destroy;
868
869 if (hw->debug_mask & ICE_DBG_PARSER)
870 ice_parser_result_dump(hw, &rslt);
871
872 conf->prof = kzalloc(sizeof(*conf->prof), GFP_KERNEL);
873 if (!conf->prof) {
874 status = -ENOMEM;
875 goto err_parser_destroy;
876 }
877
878 status = ice_parser_profile_init(&rslt, pkt_buf, msk_buf,
879 proto->raw.pkt_len, ICE_BLK_FD,
880 conf->prof);
881 if (status)
882 goto err_parser_profile_init;
883
884 if (hw->debug_mask & ICE_DBG_PARSER)
885 ice_parser_profile_dump(hw, conf->prof);
886
887 /* Store raw flow info into @conf */
888 conf->pkt_len = proto->raw.pkt_len;
889 conf->pkt_buf = pkt_buf;
890 conf->parser_ena = true;
891
892 ice_parser_destroy(psr);
893 return 0;
894
895err_parser_profile_init:
896 kfree(conf->prof);
897err_parser_destroy:
898 ice_parser_destroy(psr);
899err_mem_alloc:
900 kfree(pkt_buf);
901 return status;
902}
903
904/**
905 * ice_vc_fdir_parse_pattern
906 * @vf: pointer to the VF info
907 * @fltr: virtual channel add cmd buffer
908 * @conf: FDIR configuration for each filter
909 *
910 * Parse the virtual channel filter's pattern and store them into conf
911 *
912 * Return: 0 on success, and other on error.
913 */
914static int
915ice_vc_fdir_parse_pattern(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
916 struct virtchnl_fdir_fltr_conf *conf)
917{
918 struct virtchnl_proto_hdrs *proto = &fltr->rule_cfg.proto_hdrs;
919 enum virtchnl_proto_hdr_type l3 = VIRTCHNL_PROTO_HDR_NONE;
920 enum virtchnl_proto_hdr_type l4 = VIRTCHNL_PROTO_HDR_NONE;
921 struct device *dev = ice_pf_to_dev(vf->pf);
922 struct ice_fdir_fltr *input = &conf->input;
923 int i;
924
925 if (proto->count > VIRTCHNL_MAX_NUM_PROTO_HDRS) {
926 dev_dbg(dev, "Invalid protocol count:0x%x for VF %d\n",
927 proto->count, vf->vf_id);
928 return -EINVAL;
929 }
930
931 /* For raw FDIR filters created by the parser */
932 if (ice_vc_fdir_is_raw_flow(proto))
933 return ice_vc_fdir_parse_raw(vf, proto, conf);
934
935 for (i = 0; i < proto->count; i++) {
936 struct virtchnl_proto_hdr *hdr = &proto->proto_hdr[i];
937 struct ip_esp_hdr *esph;
938 struct ip_auth_hdr *ah;
939 struct sctphdr *sctph;
940 struct ipv6hdr *ip6h;
941 struct udphdr *udph;
942 struct tcphdr *tcph;
943 struct ethhdr *eth;
944 struct iphdr *iph;
945 u8 s_field;
946 u8 *rawh;
947
948 switch (hdr->type) {
949 case VIRTCHNL_PROTO_HDR_ETH:
950 eth = (struct ethhdr *)hdr->buffer;
951 input->flow_type = ICE_FLTR_PTYPE_NON_IP_L2;
952
953 if (hdr->field_selector)
954 input->ext_data.ether_type = eth->h_proto;
955 break;
956 case VIRTCHNL_PROTO_HDR_IPV4:
957 iph = (struct iphdr *)hdr->buffer;
958 l3 = VIRTCHNL_PROTO_HDR_IPV4;
959 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_OTHER;
960
961 if (hdr->field_selector) {
962 input->ip.v4.src_ip = iph->saddr;
963 input->ip.v4.dst_ip = iph->daddr;
964 input->ip.v4.tos = iph->tos;
965 input->ip.v4.proto = iph->protocol;
966 }
967 break;
968 case VIRTCHNL_PROTO_HDR_IPV6:
969 ip6h = (struct ipv6hdr *)hdr->buffer;
970 l3 = VIRTCHNL_PROTO_HDR_IPV6;
971 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_OTHER;
972
973 if (hdr->field_selector) {
974 memcpy(input->ip.v6.src_ip,
975 ip6h->saddr.in6_u.u6_addr8,
976 sizeof(ip6h->saddr));
977 memcpy(input->ip.v6.dst_ip,
978 ip6h->daddr.in6_u.u6_addr8,
979 sizeof(ip6h->daddr));
980 input->ip.v6.tc = ((u8)(ip6h->priority) << 4) |
981 (ip6h->flow_lbl[0] >> 4);
982 input->ip.v6.proto = ip6h->nexthdr;
983 }
984 break;
985 case VIRTCHNL_PROTO_HDR_TCP:
986 tcph = (struct tcphdr *)hdr->buffer;
987 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
988 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_TCP;
989 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
990 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_TCP;
991
992 if (hdr->field_selector) {
993 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
994 input->ip.v4.src_port = tcph->source;
995 input->ip.v4.dst_port = tcph->dest;
996 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
997 input->ip.v6.src_port = tcph->source;
998 input->ip.v6.dst_port = tcph->dest;
999 }
1000 }
1001 break;
1002 case VIRTCHNL_PROTO_HDR_UDP:
1003 udph = (struct udphdr *)hdr->buffer;
1004 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1005 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_UDP;
1006 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1007 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_UDP;
1008
1009 if (hdr->field_selector) {
1010 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
1011 input->ip.v4.src_port = udph->source;
1012 input->ip.v4.dst_port = udph->dest;
1013 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
1014 input->ip.v6.src_port = udph->source;
1015 input->ip.v6.dst_port = udph->dest;
1016 }
1017 }
1018 break;
1019 case VIRTCHNL_PROTO_HDR_SCTP:
1020 sctph = (struct sctphdr *)hdr->buffer;
1021 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1022 input->flow_type =
1023 ICE_FLTR_PTYPE_NONF_IPV4_SCTP;
1024 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1025 input->flow_type =
1026 ICE_FLTR_PTYPE_NONF_IPV6_SCTP;
1027
1028 if (hdr->field_selector) {
1029 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
1030 input->ip.v4.src_port = sctph->source;
1031 input->ip.v4.dst_port = sctph->dest;
1032 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
1033 input->ip.v6.src_port = sctph->source;
1034 input->ip.v6.dst_port = sctph->dest;
1035 }
1036 }
1037 break;
1038 case VIRTCHNL_PROTO_HDR_L2TPV3:
1039 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1040 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_L2TPV3;
1041 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1042 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_L2TPV3;
1043
1044 if (hdr->field_selector)
1045 input->l2tpv3_data.session_id = *((__be32 *)hdr->buffer);
1046 break;
1047 case VIRTCHNL_PROTO_HDR_ESP:
1048 esph = (struct ip_esp_hdr *)hdr->buffer;
1049 if (l3 == VIRTCHNL_PROTO_HDR_IPV4 &&
1050 l4 == VIRTCHNL_PROTO_HDR_UDP)
1051 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_NAT_T_ESP;
1052 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 &&
1053 l4 == VIRTCHNL_PROTO_HDR_UDP)
1054 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_NAT_T_ESP;
1055 else if (l3 == VIRTCHNL_PROTO_HDR_IPV4 &&
1056 l4 == VIRTCHNL_PROTO_HDR_NONE)
1057 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_ESP;
1058 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 &&
1059 l4 == VIRTCHNL_PROTO_HDR_NONE)
1060 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_ESP;
1061
1062 if (l4 == VIRTCHNL_PROTO_HDR_UDP)
1063 conf->inset_flag |= FDIR_INSET_FLAG_ESP_UDP;
1064 else
1065 conf->inset_flag |= FDIR_INSET_FLAG_ESP_IPSEC;
1066
1067 if (hdr->field_selector) {
1068 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1069 input->ip.v4.sec_parm_idx = esph->spi;
1070 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1071 input->ip.v6.sec_parm_idx = esph->spi;
1072 }
1073 break;
1074 case VIRTCHNL_PROTO_HDR_AH:
1075 ah = (struct ip_auth_hdr *)hdr->buffer;
1076 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1077 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_AH;
1078 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1079 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_AH;
1080
1081 if (hdr->field_selector) {
1082 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1083 input->ip.v4.sec_parm_idx = ah->spi;
1084 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1085 input->ip.v6.sec_parm_idx = ah->spi;
1086 }
1087 break;
1088 case VIRTCHNL_PROTO_HDR_PFCP:
1089 rawh = (u8 *)hdr->buffer;
1090 s_field = (rawh[0] >> PFCP_S_OFFSET) & PFCP_S_MASK;
1091 if (l3 == VIRTCHNL_PROTO_HDR_IPV4 && s_field == 0)
1092 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_PFCP_NODE;
1093 else if (l3 == VIRTCHNL_PROTO_HDR_IPV4 && s_field == 1)
1094 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_PFCP_SESSION;
1095 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 && s_field == 0)
1096 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_PFCP_NODE;
1097 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 && s_field == 1)
1098 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_PFCP_SESSION;
1099
1100 if (hdr->field_selector) {
1101 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
1102 input->ip.v4.dst_port = cpu_to_be16(PFCP_PORT_NR);
1103 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
1104 input->ip.v6.dst_port = cpu_to_be16(PFCP_PORT_NR);
1105 }
1106 break;
1107 case VIRTCHNL_PROTO_HDR_GTPU_IP:
1108 rawh = (u8 *)hdr->buffer;
1109 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_OTHER;
1110
1111 if (hdr->field_selector)
1112 input->gtpu_data.teid = *(__be32 *)(&rawh[GTPU_TEID_OFFSET]);
1113 conf->ttype = ICE_FDIR_TUNNEL_TYPE_GTPU;
1114 break;
1115 case VIRTCHNL_PROTO_HDR_GTPU_EH:
1116 rawh = (u8 *)hdr->buffer;
1117
1118 if (hdr->field_selector)
1119 input->gtpu_data.qfi = rawh[GTPU_EH_QFI_OFFSET] & GTPU_EH_QFI_MASK;
1120 conf->ttype = ICE_FDIR_TUNNEL_TYPE_GTPU_EH;
1121 break;
1122 default:
1123 dev_dbg(dev, "Invalid header type 0x:%x for VF %d\n",
1124 hdr->type, vf->vf_id);
1125 return -EINVAL;
1126 }
1127 }
1128
1129 return 0;
1130}
1131
1132/**
1133 * ice_vc_fdir_parse_action
1134 * @vf: pointer to the VF info
1135 * @fltr: virtual channel add cmd buffer
1136 * @conf: FDIR configuration for each filter
1137 *
1138 * Parse the virtual channel filter's action and store them into conf
1139 *
1140 * Return: 0 on success, and other on error.
1141 */
1142static int
1143ice_vc_fdir_parse_action(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
1144 struct virtchnl_fdir_fltr_conf *conf)
1145{
1146 struct virtchnl_filter_action_set *as = &fltr->rule_cfg.action_set;
1147 struct device *dev = ice_pf_to_dev(vf->pf);
1148 struct ice_fdir_fltr *input = &conf->input;
1149 u32 dest_num = 0;
1150 u32 mark_num = 0;
1151 int i;
1152
1153 if (as->count > VIRTCHNL_MAX_NUM_ACTIONS) {
1154 dev_dbg(dev, "Invalid action numbers:0x%x for VF %d\n",
1155 as->count, vf->vf_id);
1156 return -EINVAL;
1157 }
1158
1159 for (i = 0; i < as->count; i++) {
1160 struct virtchnl_filter_action *action = &as->actions[i];
1161
1162 switch (action->type) {
1163 case VIRTCHNL_ACTION_PASSTHRU:
1164 dest_num++;
1165 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_OTHER;
1166 break;
1167 case VIRTCHNL_ACTION_DROP:
1168 dest_num++;
1169 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DROP_PKT;
1170 break;
1171 case VIRTCHNL_ACTION_QUEUE:
1172 dest_num++;
1173 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_QINDEX;
1174 input->q_index = action->act_conf.queue.index;
1175 break;
1176 case VIRTCHNL_ACTION_Q_REGION:
1177 dest_num++;
1178 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_QGROUP;
1179 input->q_index = action->act_conf.queue.index;
1180 input->q_region = action->act_conf.queue.region;
1181 break;
1182 case VIRTCHNL_ACTION_MARK:
1183 mark_num++;
1184 input->fltr_id = action->act_conf.mark_id;
1185 input->fdid_prio = ICE_FXD_FLTR_QW1_FDID_PRI_THREE;
1186 break;
1187 default:
1188 dev_dbg(dev, "Invalid action type:0x%x for VF %d\n",
1189 action->type, vf->vf_id);
1190 return -EINVAL;
1191 }
1192 }
1193
1194 if (dest_num == 0 || dest_num >= 2) {
1195 dev_dbg(dev, "Invalid destination action for VF %d\n",
1196 vf->vf_id);
1197 return -EINVAL;
1198 }
1199
1200 if (mark_num >= 2) {
1201 dev_dbg(dev, "Too many mark actions for VF %d\n", vf->vf_id);
1202 return -EINVAL;
1203 }
1204
1205 return 0;
1206}
1207
1208/**
1209 * ice_vc_validate_fdir_fltr - validate the virtual channel filter
1210 * @vf: pointer to the VF info
1211 * @fltr: virtual channel add cmd buffer
1212 * @conf: FDIR configuration for each filter
1213 *
1214 * Return: 0 on success, and other on error.
1215 */
1216static int
1217ice_vc_validate_fdir_fltr(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
1218 struct virtchnl_fdir_fltr_conf *conf)
1219{
1220 struct virtchnl_proto_hdrs *proto = &fltr->rule_cfg.proto_hdrs;
1221 int ret;
1222
1223 /* For raw FDIR filters created by the parser */
1224 if (!ice_vc_fdir_is_raw_flow(proto))
1225 if (!ice_vc_validate_pattern(vf, proto))
1226 return -EINVAL;
1227
1228 ret = ice_vc_fdir_parse_pattern(vf, fltr, conf);
1229 if (ret)
1230 return ret;
1231
1232 return ice_vc_fdir_parse_action(vf, fltr, conf);
1233}
1234
1235/**
1236 * ice_vc_fdir_comp_rules - compare if two filter rules have the same value
1237 * @conf_a: FDIR configuration for filter a
1238 * @conf_b: FDIR configuration for filter b
1239 *
1240 * Return: 0 on success, and other on error.
1241 */
1242static bool
1243ice_vc_fdir_comp_rules(struct virtchnl_fdir_fltr_conf *conf_a,
1244 struct virtchnl_fdir_fltr_conf *conf_b)
1245{
1246 struct ice_fdir_fltr *a = &conf_a->input;
1247 struct ice_fdir_fltr *b = &conf_b->input;
1248
1249 if (conf_a->ttype != conf_b->ttype)
1250 return false;
1251 if (a->flow_type != b->flow_type)
1252 return false;
1253 if (memcmp(&a->ip, &b->ip, sizeof(a->ip)))
1254 return false;
1255 if (memcmp(&a->mask, &b->mask, sizeof(a->mask)))
1256 return false;
1257 if (memcmp(&a->gtpu_data, &b->gtpu_data, sizeof(a->gtpu_data)))
1258 return false;
1259 if (memcmp(&a->gtpu_mask, &b->gtpu_mask, sizeof(a->gtpu_mask)))
1260 return false;
1261 if (memcmp(&a->l2tpv3_data, &b->l2tpv3_data, sizeof(a->l2tpv3_data)))
1262 return false;
1263 if (memcmp(&a->l2tpv3_mask, &b->l2tpv3_mask, sizeof(a->l2tpv3_mask)))
1264 return false;
1265 if (memcmp(&a->ext_data, &b->ext_data, sizeof(a->ext_data)))
1266 return false;
1267 if (memcmp(&a->ext_mask, &b->ext_mask, sizeof(a->ext_mask)))
1268 return false;
1269
1270 return true;
1271}
1272
1273/**
1274 * ice_vc_fdir_is_dup_fltr
1275 * @vf: pointer to the VF info
1276 * @conf: FDIR configuration for each filter
1277 *
1278 * Check if there is duplicated rule with same conf value
1279 *
1280 * Return: 0 true success, and false on error.
1281 */
1282static bool
1283ice_vc_fdir_is_dup_fltr(struct ice_vf *vf, struct virtchnl_fdir_fltr_conf *conf)
1284{
1285 struct ice_fdir_fltr *desc;
1286 bool ret;
1287
1288 list_for_each_entry(desc, &vf->fdir.fdir_rule_list, fltr_node) {
1289 struct virtchnl_fdir_fltr_conf *node =
1290 to_fltr_conf_from_desc(desc);
1291
1292 ret = ice_vc_fdir_comp_rules(node, conf);
1293 if (ret)
1294 return true;
1295 }
1296
1297 return false;
1298}
1299
1300/**
1301 * ice_vc_fdir_insert_entry
1302 * @vf: pointer to the VF info
1303 * @conf: FDIR configuration for each filter
1304 * @id: pointer to ID value allocated by driver
1305 *
1306 * Insert FDIR conf entry into list and allocate ID for this filter
1307 *
1308 * Return: 0 true success, and other on error.
1309 */
1310static int
1311ice_vc_fdir_insert_entry(struct ice_vf *vf,
1312 struct virtchnl_fdir_fltr_conf *conf, u32 *id)
1313{
1314 struct ice_fdir_fltr *input = &conf->input;
1315 int i;
1316
1317 /* alloc ID corresponding with conf */
1318 i = idr_alloc(&vf->fdir.fdir_rule_idr, conf, 0,
1319 ICE_FDIR_MAX_FLTRS, GFP_KERNEL);
1320 if (i < 0)
1321 return -EINVAL;
1322 *id = i;
1323
1324 list_add(&input->fltr_node, &vf->fdir.fdir_rule_list);
1325 return 0;
1326}
1327
1328/**
1329 * ice_vc_fdir_remove_entry - remove FDIR conf entry by ID value
1330 * @vf: pointer to the VF info
1331 * @conf: FDIR configuration for each filter
1332 * @id: filter rule's ID
1333 */
1334static void
1335ice_vc_fdir_remove_entry(struct ice_vf *vf,
1336 struct virtchnl_fdir_fltr_conf *conf, u32 id)
1337{
1338 struct ice_fdir_fltr *input = &conf->input;
1339
1340 idr_remove(&vf->fdir.fdir_rule_idr, id);
1341 list_del(&input->fltr_node);
1342}
1343
1344/**
1345 * ice_vc_fdir_lookup_entry - lookup FDIR conf entry by ID value
1346 * @vf: pointer to the VF info
1347 * @id: filter rule's ID
1348 *
1349 * Return: NULL on error, and other on success.
1350 */
1351static struct virtchnl_fdir_fltr_conf *
1352ice_vc_fdir_lookup_entry(struct ice_vf *vf, u32 id)
1353{
1354 return idr_find(&vf->fdir.fdir_rule_idr, id);
1355}
1356
1357/**
1358 * ice_vc_fdir_flush_entry - remove all FDIR conf entry
1359 * @vf: pointer to the VF info
1360 */
1361static void ice_vc_fdir_flush_entry(struct ice_vf *vf)
1362{
1363 struct virtchnl_fdir_fltr_conf *conf;
1364 struct ice_fdir_fltr *desc, *temp;
1365
1366 list_for_each_entry_safe(desc, temp,
1367 &vf->fdir.fdir_rule_list, fltr_node) {
1368 conf = to_fltr_conf_from_desc(desc);
1369 list_del(&desc->fltr_node);
1370 devm_kfree(ice_pf_to_dev(vf->pf), conf);
1371 }
1372}
1373
1374/**
1375 * ice_vc_fdir_write_fltr - write filter rule into hardware
1376 * @vf: pointer to the VF info
1377 * @conf: FDIR configuration for each filter
1378 * @add: true implies add rule, false implies del rules
1379 * @is_tun: false implies non-tunnel type filter, true implies tunnel filter
1380 *
1381 * Return: 0 on success, and other on error.
1382 */
1383static int ice_vc_fdir_write_fltr(struct ice_vf *vf,
1384 struct virtchnl_fdir_fltr_conf *conf,
1385 bool add, bool is_tun)
1386{
1387 struct ice_fdir_fltr *input = &conf->input;
1388 struct ice_vsi *vsi, *ctrl_vsi;
1389 struct ice_fltr_desc desc;
1390 struct device *dev;
1391 struct ice_pf *pf;
1392 struct ice_hw *hw;
1393 int ret;
1394 u8 *pkt;
1395
1396 pf = vf->pf;
1397 dev = ice_pf_to_dev(pf);
1398 hw = &pf->hw;
1399 vsi = ice_get_vf_vsi(vf);
1400 if (!vsi) {
1401 dev_dbg(dev, "Invalid vsi for VF %d\n", vf->vf_id);
1402 return -EINVAL;
1403 }
1404
1405 input->dest_vsi = vsi->idx;
1406 input->comp_report = ICE_FXD_FLTR_QW0_COMP_REPORT_SW;
1407
1408 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
1409 if (!ctrl_vsi) {
1410 dev_dbg(dev, "Invalid ctrl_vsi for VF %d\n", vf->vf_id);
1411 return -EINVAL;
1412 }
1413
1414 pkt = devm_kzalloc(dev, ICE_FDIR_MAX_RAW_PKT_SIZE, GFP_KERNEL);
1415 if (!pkt)
1416 return -ENOMEM;
1417
1418 ice_fdir_get_prgm_desc(hw, input, &desc, add);
1419 if (conf->parser_ena) {
1420 memcpy(pkt, conf->pkt_buf, conf->pkt_len);
1421 } else {
1422 ret = ice_fdir_get_gen_prgm_pkt(hw, input, pkt, false, is_tun);
1423 if (ret) {
1424 dev_dbg(dev, "Gen training pkt for VF %d ptype %d failed\n",
1425 vf->vf_id, input->flow_type);
1426 goto err_free_pkt;
1427 }
1428 }
1429
1430 ret = ice_prgm_fdir_fltr(ctrl_vsi, &desc, pkt);
1431 if (ret)
1432 goto err_free_pkt;
1433
1434 return 0;
1435
1436err_free_pkt:
1437 devm_kfree(dev, pkt);
1438 return ret;
1439}
1440
1441/**
1442 * ice_vf_fdir_timer - FDIR program waiting timer interrupt handler
1443 * @t: pointer to timer_list
1444 */
1445static void ice_vf_fdir_timer(struct timer_list *t)
1446{
1447 struct ice_vf_fdir_ctx *ctx_irq = from_timer(ctx_irq, t, rx_tmr);
1448 struct ice_vf_fdir_ctx *ctx_done;
1449 struct ice_vf_fdir *fdir;
1450 unsigned long flags;
1451 struct ice_vf *vf;
1452 struct ice_pf *pf;
1453
1454 fdir = container_of(ctx_irq, struct ice_vf_fdir, ctx_irq);
1455 vf = container_of(fdir, struct ice_vf, fdir);
1456 ctx_done = &fdir->ctx_done;
1457 pf = vf->pf;
1458 spin_lock_irqsave(&fdir->ctx_lock, flags);
1459 if (!(ctx_irq->flags & ICE_VF_FDIR_CTX_VALID)) {
1460 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1461 WARN_ON_ONCE(1);
1462 return;
1463 }
1464
1465 ctx_irq->flags &= ~ICE_VF_FDIR_CTX_VALID;
1466
1467 ctx_done->flags |= ICE_VF_FDIR_CTX_VALID;
1468 ctx_done->conf = ctx_irq->conf;
1469 ctx_done->stat = ICE_FDIR_CTX_TIMEOUT;
1470 ctx_done->v_opcode = ctx_irq->v_opcode;
1471 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1472
1473 set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
1474 ice_service_task_schedule(pf);
1475}
1476
1477/**
1478 * ice_vc_fdir_irq_handler - ctrl_vsi Rx queue interrupt handler
1479 * @ctrl_vsi: pointer to a VF's CTRL VSI
1480 * @rx_desc: pointer to FDIR Rx queue descriptor
1481 */
1482void
1483ice_vc_fdir_irq_handler(struct ice_vsi *ctrl_vsi,
1484 union ice_32b_rx_flex_desc *rx_desc)
1485{
1486 struct ice_pf *pf = ctrl_vsi->back;
1487 struct ice_vf *vf = ctrl_vsi->vf;
1488 struct ice_vf_fdir_ctx *ctx_done;
1489 struct ice_vf_fdir_ctx *ctx_irq;
1490 struct ice_vf_fdir *fdir;
1491 unsigned long flags;
1492 struct device *dev;
1493 int ret;
1494
1495 if (WARN_ON(!vf))
1496 return;
1497
1498 fdir = &vf->fdir;
1499 ctx_done = &fdir->ctx_done;
1500 ctx_irq = &fdir->ctx_irq;
1501 dev = ice_pf_to_dev(pf);
1502 spin_lock_irqsave(&fdir->ctx_lock, flags);
1503 if (!(ctx_irq->flags & ICE_VF_FDIR_CTX_VALID)) {
1504 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1505 WARN_ON_ONCE(1);
1506 return;
1507 }
1508
1509 ctx_irq->flags &= ~ICE_VF_FDIR_CTX_VALID;
1510
1511 ctx_done->flags |= ICE_VF_FDIR_CTX_VALID;
1512 ctx_done->conf = ctx_irq->conf;
1513 ctx_done->stat = ICE_FDIR_CTX_IRQ;
1514 ctx_done->v_opcode = ctx_irq->v_opcode;
1515 memcpy(&ctx_done->rx_desc, rx_desc, sizeof(*rx_desc));
1516 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1517
1518 ret = del_timer(&ctx_irq->rx_tmr);
1519 if (!ret)
1520 dev_err(dev, "VF %d: Unexpected inactive timer!\n", vf->vf_id);
1521
1522 set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
1523 ice_service_task_schedule(pf);
1524}
1525
1526/**
1527 * ice_vf_fdir_dump_info - dump FDIR information for diagnosis
1528 * @vf: pointer to the VF info
1529 */
1530static void ice_vf_fdir_dump_info(struct ice_vf *vf)
1531{
1532 u32 fd_size, fd_cnt, fd_size_g, fd_cnt_g, fd_size_b, fd_cnt_b;
1533 struct ice_vsi *vf_vsi;
1534 struct device *dev;
1535 struct ice_pf *pf;
1536 struct ice_hw *hw;
1537 u16 vsi_num;
1538
1539 pf = vf->pf;
1540 hw = &pf->hw;
1541 dev = ice_pf_to_dev(pf);
1542 vf_vsi = ice_get_vf_vsi(vf);
1543 if (!vf_vsi) {
1544 dev_dbg(dev, "VF %d: invalid VSI pointer\n", vf->vf_id);
1545 return;
1546 }
1547
1548 vsi_num = ice_get_hw_vsi_num(hw, vf_vsi->idx);
1549
1550 fd_size = rd32(hw, VSIQF_FD_SIZE(vsi_num));
1551 fd_cnt = rd32(hw, VSIQF_FD_CNT(vsi_num));
1552 switch (hw->mac_type) {
1553 case ICE_MAC_E830:
1554 fd_size_g = FIELD_GET(E830_VSIQF_FD_CNT_FD_GCNT_M, fd_size);
1555 fd_size_b = FIELD_GET(E830_VSIQF_FD_CNT_FD_BCNT_M, fd_size);
1556 fd_cnt_g = FIELD_GET(E830_VSIQF_FD_CNT_FD_GCNT_M, fd_cnt);
1557 fd_cnt_b = FIELD_GET(E830_VSIQF_FD_CNT_FD_BCNT_M, fd_cnt);
1558 break;
1559 case ICE_MAC_E810:
1560 default:
1561 fd_size_g = FIELD_GET(E800_VSIQF_FD_CNT_FD_GCNT_M, fd_size);
1562 fd_size_b = FIELD_GET(E800_VSIQF_FD_CNT_FD_BCNT_M, fd_size);
1563 fd_cnt_g = FIELD_GET(E800_VSIQF_FD_CNT_FD_GCNT_M, fd_cnt);
1564 fd_cnt_b = FIELD_GET(E800_VSIQF_FD_CNT_FD_BCNT_M, fd_cnt);
1565 }
1566
1567 dev_dbg(dev, "VF %d: Size in the FD table: guaranteed:0x%x, best effort:0x%x\n",
1568 vf->vf_id, fd_size_g, fd_size_b);
1569 dev_dbg(dev, "VF %d: Filter counter in the FD table: guaranteed:0x%x, best effort:0x%x\n",
1570 vf->vf_id, fd_cnt_g, fd_cnt_b);
1571}
1572
1573/**
1574 * ice_vf_verify_rx_desc - verify received FDIR programming status descriptor
1575 * @vf: pointer to the VF info
1576 * @ctx: FDIR context info for post processing
1577 * @status: virtchnl FDIR program status
1578 *
1579 * Return: 0 on success, and other on error.
1580 */
1581static int
1582ice_vf_verify_rx_desc(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1583 enum virtchnl_fdir_prgm_status *status)
1584{
1585 struct device *dev = ice_pf_to_dev(vf->pf);
1586 u32 stat_err, error, prog_id;
1587 int ret;
1588
1589 stat_err = le16_to_cpu(ctx->rx_desc.wb.status_error0);
1590 if (FIELD_GET(ICE_FXD_FLTR_WB_QW1_DD_M, stat_err) !=
1591 ICE_FXD_FLTR_WB_QW1_DD_YES) {
1592 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1593 dev_err(dev, "VF %d: Desc Done not set\n", vf->vf_id);
1594 ret = -EINVAL;
1595 goto err_exit;
1596 }
1597
1598 prog_id = FIELD_GET(ICE_FXD_FLTR_WB_QW1_PROG_ID_M, stat_err);
1599 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_ADD &&
1600 ctx->v_opcode != VIRTCHNL_OP_ADD_FDIR_FILTER) {
1601 dev_err(dev, "VF %d: Desc show add, but ctx not",
1602 vf->vf_id);
1603 *status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
1604 ret = -EINVAL;
1605 goto err_exit;
1606 }
1607
1608 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_DEL &&
1609 ctx->v_opcode != VIRTCHNL_OP_DEL_FDIR_FILTER) {
1610 dev_err(dev, "VF %d: Desc show del, but ctx not",
1611 vf->vf_id);
1612 *status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
1613 ret = -EINVAL;
1614 goto err_exit;
1615 }
1616
1617 error = FIELD_GET(ICE_FXD_FLTR_WB_QW1_FAIL_M, stat_err);
1618 if (error == ICE_FXD_FLTR_WB_QW1_FAIL_YES) {
1619 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_ADD) {
1620 dev_err(dev, "VF %d, Failed to add FDIR rule due to no space in the table",
1621 vf->vf_id);
1622 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1623 } else {
1624 dev_err(dev, "VF %d, Failed to remove FDIR rule, attempt to remove non-existent entry",
1625 vf->vf_id);
1626 *status = VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST;
1627 }
1628 ret = -EINVAL;
1629 goto err_exit;
1630 }
1631
1632 error = FIELD_GET(ICE_FXD_FLTR_WB_QW1_FAIL_PROF_M, stat_err);
1633 if (error == ICE_FXD_FLTR_WB_QW1_FAIL_PROF_YES) {
1634 dev_err(dev, "VF %d: Profile matching error", vf->vf_id);
1635 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1636 ret = -EINVAL;
1637 goto err_exit;
1638 }
1639
1640 *status = VIRTCHNL_FDIR_SUCCESS;
1641
1642 return 0;
1643
1644err_exit:
1645 ice_vf_fdir_dump_info(vf);
1646 return ret;
1647}
1648
1649static int ice_fdir_is_tunnel(enum ice_fdir_tunnel_type ttype)
1650{
1651 return (ttype == ICE_FDIR_TUNNEL_TYPE_GRE_INNER ||
1652 ttype == ICE_FDIR_TUNNEL_TYPE_GTPU_INNER ||
1653 ttype == ICE_FDIR_TUNNEL_TYPE_GTPU_EH_INNER ||
1654 ttype == ICE_FDIR_TUNNEL_TYPE_GTPOGRE_INNER ||
1655 ttype == ICE_FDIR_TUNNEL_TYPE_ECPRI ||
1656 ttype == ICE_FDIR_TUNNEL_TYPE_L2TPV2_INNER);
1657}
1658
1659/**
1660 * ice_vc_add_fdir_fltr_post
1661 * @vf: pointer to the VF structure
1662 * @ctx: FDIR context info for post processing
1663 * @status: virtchnl FDIR program status
1664 * @success: true implies success, false implies failure
1665 *
1666 * Post process for flow director add command. If success, then do post process
1667 * and send back success msg by virtchnl. Otherwise, do context reversion and
1668 * send back failure msg by virtchnl.
1669 *
1670 * Return: 0 on success, and other on error.
1671 */
1672static int
1673ice_vc_add_fdir_fltr_post(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1674 enum virtchnl_fdir_prgm_status status,
1675 bool success)
1676{
1677 struct virtchnl_fdir_fltr_conf *conf = ctx->conf;
1678 struct device *dev = ice_pf_to_dev(vf->pf);
1679 enum virtchnl_status_code v_ret;
1680 struct virtchnl_fdir_add *resp;
1681 int ret, len, is_tun;
1682
1683 v_ret = VIRTCHNL_STATUS_SUCCESS;
1684 len = sizeof(*resp);
1685 resp = kzalloc(len, GFP_KERNEL);
1686 if (!resp) {
1687 len = 0;
1688 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1689 dev_dbg(dev, "VF %d: Alloc resp buf fail", vf->vf_id);
1690 goto err_exit;
1691 }
1692
1693 if (!success)
1694 goto err_exit;
1695
1696 is_tun = 0;
1697 resp->status = status;
1698 resp->flow_id = conf->flow_id;
1699 vf->fdir.fdir_fltr_cnt[conf->input.flow_type][is_tun]++;
1700 vf->fdir.fdir_fltr_cnt_total++;
1701
1702 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1703 (u8 *)resp, len);
1704 kfree(resp);
1705
1706 dev_dbg(dev, "VF %d: flow_id:0x%X, FDIR %s success!\n",
1707 vf->vf_id, conf->flow_id,
1708 (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER) ?
1709 "add" : "del");
1710 return ret;
1711
1712err_exit:
1713 if (resp)
1714 resp->status = status;
1715 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
1716 devm_kfree(dev, conf);
1717
1718 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1719 (u8 *)resp, len);
1720 kfree(resp);
1721 return ret;
1722}
1723
1724/**
1725 * ice_vc_del_fdir_fltr_post
1726 * @vf: pointer to the VF structure
1727 * @ctx: FDIR context info for post processing
1728 * @status: virtchnl FDIR program status
1729 * @success: true implies success, false implies failure
1730 *
1731 * Post process for flow director del command. If success, then do post process
1732 * and send back success msg by virtchnl. Otherwise, do context reversion and
1733 * send back failure msg by virtchnl.
1734 *
1735 * Return: 0 on success, and other on error.
1736 */
1737static int
1738ice_vc_del_fdir_fltr_post(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1739 enum virtchnl_fdir_prgm_status status,
1740 bool success)
1741{
1742 struct virtchnl_fdir_fltr_conf *conf = ctx->conf;
1743 struct device *dev = ice_pf_to_dev(vf->pf);
1744 enum virtchnl_status_code v_ret;
1745 struct virtchnl_fdir_del *resp;
1746 int ret, len, is_tun;
1747
1748 v_ret = VIRTCHNL_STATUS_SUCCESS;
1749 len = sizeof(*resp);
1750 resp = kzalloc(len, GFP_KERNEL);
1751 if (!resp) {
1752 len = 0;
1753 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1754 dev_dbg(dev, "VF %d: Alloc resp buf fail", vf->vf_id);
1755 goto err_exit;
1756 }
1757
1758 if (!success)
1759 goto err_exit;
1760
1761 is_tun = 0;
1762 resp->status = status;
1763 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
1764 vf->fdir.fdir_fltr_cnt[conf->input.flow_type][is_tun]--;
1765 vf->fdir.fdir_fltr_cnt_total--;
1766
1767 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1768 (u8 *)resp, len);
1769 kfree(resp);
1770
1771 dev_dbg(dev, "VF %d: flow_id:0x%X, FDIR %s success!\n",
1772 vf->vf_id, conf->flow_id,
1773 (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER) ?
1774 "add" : "del");
1775 devm_kfree(dev, conf);
1776 return ret;
1777
1778err_exit:
1779 if (resp)
1780 resp->status = status;
1781 if (success)
1782 devm_kfree(dev, conf);
1783
1784 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1785 (u8 *)resp, len);
1786 kfree(resp);
1787 return ret;
1788}
1789
1790/**
1791 * ice_flush_fdir_ctx
1792 * @pf: pointer to the PF structure
1793 *
1794 * Flush all the pending event on ctx_done list and process them.
1795 */
1796void ice_flush_fdir_ctx(struct ice_pf *pf)
1797{
1798 struct ice_vf *vf;
1799 unsigned int bkt;
1800
1801 if (!test_and_clear_bit(ICE_FD_VF_FLUSH_CTX, pf->state))
1802 return;
1803
1804 mutex_lock(&pf->vfs.table_lock);
1805 ice_for_each_vf(pf, bkt, vf) {
1806 struct device *dev = ice_pf_to_dev(pf);
1807 enum virtchnl_fdir_prgm_status status;
1808 struct ice_vf_fdir_ctx *ctx;
1809 unsigned long flags;
1810 int ret;
1811
1812 if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states))
1813 continue;
1814
1815 if (vf->ctrl_vsi_idx == ICE_NO_VSI)
1816 continue;
1817
1818 ctx = &vf->fdir.ctx_done;
1819 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1820 if (!(ctx->flags & ICE_VF_FDIR_CTX_VALID)) {
1821 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1822 continue;
1823 }
1824 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1825
1826 WARN_ON(ctx->stat == ICE_FDIR_CTX_READY);
1827 if (ctx->stat == ICE_FDIR_CTX_TIMEOUT) {
1828 status = VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT;
1829 dev_err(dev, "VF %d: ctrl_vsi irq timeout\n",
1830 vf->vf_id);
1831 goto err_exit;
1832 }
1833
1834 ret = ice_vf_verify_rx_desc(vf, ctx, &status);
1835 if (ret)
1836 goto err_exit;
1837
1838 if (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER)
1839 ice_vc_add_fdir_fltr_post(vf, ctx, status, true);
1840 else if (ctx->v_opcode == VIRTCHNL_OP_DEL_FDIR_FILTER)
1841 ice_vc_del_fdir_fltr_post(vf, ctx, status, true);
1842 else
1843 dev_err(dev, "VF %d: Unsupported opcode\n", vf->vf_id);
1844
1845 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1846 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1847 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1848 continue;
1849err_exit:
1850 if (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER)
1851 ice_vc_add_fdir_fltr_post(vf, ctx, status, false);
1852 else if (ctx->v_opcode == VIRTCHNL_OP_DEL_FDIR_FILTER)
1853 ice_vc_del_fdir_fltr_post(vf, ctx, status, false);
1854 else
1855 dev_err(dev, "VF %d: Unsupported opcode\n", vf->vf_id);
1856
1857 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1858 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1859 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1860 }
1861 mutex_unlock(&pf->vfs.table_lock);
1862}
1863
1864/**
1865 * ice_vc_fdir_set_irq_ctx - set FDIR context info for later IRQ handler
1866 * @vf: pointer to the VF structure
1867 * @conf: FDIR configuration for each filter
1868 * @v_opcode: virtual channel operation code
1869 *
1870 * Return: 0 on success, and other on error.
1871 */
1872static int
1873ice_vc_fdir_set_irq_ctx(struct ice_vf *vf, struct virtchnl_fdir_fltr_conf *conf,
1874 enum virtchnl_ops v_opcode)
1875{
1876 struct device *dev = ice_pf_to_dev(vf->pf);
1877 struct ice_vf_fdir_ctx *ctx;
1878 unsigned long flags;
1879
1880 ctx = &vf->fdir.ctx_irq;
1881 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1882 if ((vf->fdir.ctx_irq.flags & ICE_VF_FDIR_CTX_VALID) ||
1883 (vf->fdir.ctx_done.flags & ICE_VF_FDIR_CTX_VALID)) {
1884 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1885 dev_dbg(dev, "VF %d: Last request is still in progress\n",
1886 vf->vf_id);
1887 return -EBUSY;
1888 }
1889 ctx->flags |= ICE_VF_FDIR_CTX_VALID;
1890 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1891
1892 ctx->conf = conf;
1893 ctx->v_opcode = v_opcode;
1894 ctx->stat = ICE_FDIR_CTX_READY;
1895 timer_setup(&ctx->rx_tmr, ice_vf_fdir_timer, 0);
1896
1897 mod_timer(&ctx->rx_tmr, round_jiffies(msecs_to_jiffies(10) + jiffies));
1898
1899 return 0;
1900}
1901
1902/**
1903 * ice_vc_fdir_clear_irq_ctx - clear FDIR context info for IRQ handler
1904 * @vf: pointer to the VF structure
1905 *
1906 * Return: 0 on success, and other on error.
1907 */
1908static void ice_vc_fdir_clear_irq_ctx(struct ice_vf *vf)
1909{
1910 struct ice_vf_fdir_ctx *ctx = &vf->fdir.ctx_irq;
1911 unsigned long flags;
1912
1913 del_timer(&ctx->rx_tmr);
1914 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1915 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1916 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1917}
1918
1919/**
1920 * ice_vc_parser_fv_check_diff - check two parsed FDIR profile fv context
1921 * @fv_a: struct of parsed FDIR profile field vector
1922 * @fv_b: struct of parsed FDIR profile field vector
1923 *
1924 * Check if the two parsed FDIR profile field vector context are different,
1925 * including proto_id, offset and mask.
1926 *
1927 * Return: true on different, false on otherwise.
1928 */
1929static bool ice_vc_parser_fv_check_diff(struct ice_parser_fv *fv_a,
1930 struct ice_parser_fv *fv_b)
1931{
1932 return (fv_a->proto_id != fv_b->proto_id ||
1933 fv_a->offset != fv_b->offset ||
1934 fv_a->msk != fv_b->msk);
1935}
1936
1937/**
1938 * ice_vc_parser_fv_save - save parsed FDIR profile fv context
1939 * @fv: struct of parsed FDIR profile field vector
1940 * @fv_src: parsed FDIR profile field vector context to save
1941 *
1942 * Save the parsed FDIR profile field vector context, including proto_id,
1943 * offset and mask.
1944 *
1945 * Return: Void.
1946 */
1947static void ice_vc_parser_fv_save(struct ice_parser_fv *fv,
1948 struct ice_parser_fv *fv_src)
1949{
1950 fv->proto_id = fv_src->proto_id;
1951 fv->offset = fv_src->offset;
1952 fv->msk = fv_src->msk;
1953 fv->spec = 0;
1954}
1955
1956/**
1957 * ice_vc_add_fdir_raw - add a raw FDIR filter for VF
1958 * @vf: pointer to the VF info
1959 * @conf: FDIR configuration for each filter
1960 * @v_ret: the final VIRTCHNL code
1961 * @stat: pointer to the VIRTCHNL_OP_ADD_FDIR_FILTER
1962 * @len: length of the stat
1963 *
1964 * Return: 0 on success or negative errno on failure.
1965 */
1966static int
1967ice_vc_add_fdir_raw(struct ice_vf *vf,
1968 struct virtchnl_fdir_fltr_conf *conf,
1969 enum virtchnl_status_code *v_ret,
1970 struct virtchnl_fdir_add *stat, int len)
1971{
1972 struct ice_vsi *vf_vsi, *ctrl_vsi;
1973 struct ice_fdir_prof_info *pi;
1974 struct ice_pf *pf = vf->pf;
1975 int ret, ptg, id, i;
1976 struct device *dev;
1977 struct ice_hw *hw;
1978 bool fv_found;
1979
1980 dev = ice_pf_to_dev(pf);
1981 hw = &pf->hw;
1982 *v_ret = VIRTCHNL_STATUS_ERR_PARAM;
1983 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1984
1985 id = find_first_bit(conf->prof->ptypes, ICE_FLOW_PTYPE_MAX);
1986 ptg = hw->blk[ICE_BLK_FD].xlt1.t[id];
1987
1988 vf_vsi = ice_get_vf_vsi(vf);
1989 if (!vf_vsi) {
1990 dev_err(dev, "Can not get FDIR vf_vsi for VF %d\n", vf->vf_id);
1991 return -ENODEV;
1992 }
1993
1994 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
1995 if (!ctrl_vsi) {
1996 dev_err(dev, "Can not get FDIR ctrl_vsi for VF %d\n",
1997 vf->vf_id);
1998 return -ENODEV;
1999 }
2000
2001 fv_found = false;
2002
2003 /* Check if profile info already exists, then update the counter */
2004 pi = &vf->fdir_prof_info[ptg];
2005 if (pi->fdir_active_cnt != 0) {
2006 for (i = 0; i < ICE_MAX_FV_WORDS; i++)
2007 if (ice_vc_parser_fv_check_diff(&pi->prof.fv[i],
2008 &conf->prof->fv[i]))
2009 break;
2010 if (i == ICE_MAX_FV_WORDS) {
2011 fv_found = true;
2012 pi->fdir_active_cnt++;
2013 }
2014 }
2015
2016 /* HW profile setting is only required for the first time */
2017 if (!fv_found) {
2018 ret = ice_flow_set_parser_prof(hw, vf_vsi->idx,
2019 ctrl_vsi->idx, conf->prof,
2020 ICE_BLK_FD);
2021
2022 if (ret) {
2023 *v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
2024 dev_dbg(dev, "VF %d: insert hw prof failed\n",
2025 vf->vf_id);
2026 return ret;
2027 }
2028 }
2029
2030 ret = ice_vc_fdir_insert_entry(vf, conf, &conf->flow_id);
2031 if (ret) {
2032 *v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
2033 dev_dbg(dev, "VF %d: insert FDIR list failed\n",
2034 vf->vf_id);
2035 return ret;
2036 }
2037
2038 ret = ice_vc_fdir_set_irq_ctx(vf, conf,
2039 VIRTCHNL_OP_ADD_FDIR_FILTER);
2040 if (ret) {
2041 dev_dbg(dev, "VF %d: set FDIR context failed\n",
2042 vf->vf_id);
2043 goto err_rem_entry;
2044 }
2045
2046 ret = ice_vc_fdir_write_fltr(vf, conf, true, false);
2047 if (ret) {
2048 dev_err(dev, "VF %d: adding FDIR raw flow rule failed, ret:%d\n",
2049 vf->vf_id, ret);
2050 goto err_clr_irq;
2051 }
2052
2053 /* Save parsed profile fv info of the FDIR rule for the first time */
2054 if (!fv_found) {
2055 for (i = 0; i < conf->prof->fv_num; i++)
2056 ice_vc_parser_fv_save(&pi->prof.fv[i],
2057 &conf->prof->fv[i]);
2058 pi->prof.fv_num = conf->prof->fv_num;
2059 pi->fdir_active_cnt = 1;
2060 }
2061
2062 return 0;
2063
2064err_clr_irq:
2065 ice_vc_fdir_clear_irq_ctx(vf);
2066err_rem_entry:
2067 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
2068 return ret;
2069}
2070
2071/**
2072 * ice_vc_add_fdir_fltr - add a FDIR filter for VF by the msg buffer
2073 * @vf: pointer to the VF info
2074 * @msg: pointer to the msg buffer
2075 *
2076 * Return: 0 on success, and other on error.
2077 */
2078int ice_vc_add_fdir_fltr(struct ice_vf *vf, u8 *msg)
2079{
2080 struct virtchnl_fdir_add *fltr = (struct virtchnl_fdir_add *)msg;
2081 struct virtchnl_fdir_add *stat = NULL;
2082 struct virtchnl_fdir_fltr_conf *conf;
2083 enum virtchnl_status_code v_ret;
2084 struct ice_vsi *vf_vsi;
2085 struct device *dev;
2086 struct ice_pf *pf;
2087 int is_tun = 0;
2088 int len = 0;
2089 int ret;
2090
2091 pf = vf->pf;
2092 dev = ice_pf_to_dev(pf);
2093 vf_vsi = ice_get_vf_vsi(vf);
2094
2095#define ICE_VF_MAX_FDIR_FILTERS 128
2096 if (!ice_fdir_num_avail_fltr(&pf->hw, vf_vsi) ||
2097 vf->fdir.fdir_fltr_cnt_total >= ICE_VF_MAX_FDIR_FILTERS) {
2098 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2099 dev_err(dev, "Max number of FDIR filters for VF %d is reached\n",
2100 vf->vf_id);
2101 goto err_exit;
2102 }
2103
2104 ret = ice_vc_fdir_param_check(vf, fltr->vsi_id);
2105 if (ret) {
2106 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2107 dev_dbg(dev, "Parameter check for VF %d failed\n", vf->vf_id);
2108 goto err_exit;
2109 }
2110
2111 ret = ice_vf_start_ctrl_vsi(vf);
2112 if (ret && (ret != -EEXIST)) {
2113 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2114 dev_err(dev, "Init FDIR for VF %d failed, ret:%d\n",
2115 vf->vf_id, ret);
2116 goto err_exit;
2117 }
2118
2119 stat = kzalloc(sizeof(*stat), GFP_KERNEL);
2120 if (!stat) {
2121 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
2122 dev_dbg(dev, "Alloc stat for VF %d failed\n", vf->vf_id);
2123 goto err_exit;
2124 }
2125
2126 conf = devm_kzalloc(dev, sizeof(*conf), GFP_KERNEL);
2127 if (!conf) {
2128 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
2129 dev_dbg(dev, "Alloc conf for VF %d failed\n", vf->vf_id);
2130 goto err_exit;
2131 }
2132
2133 len = sizeof(*stat);
2134 ret = ice_vc_validate_fdir_fltr(vf, fltr, conf);
2135 if (ret) {
2136 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2137 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
2138 dev_dbg(dev, "Invalid FDIR filter from VF %d\n", vf->vf_id);
2139 goto err_free_conf;
2140 }
2141
2142 if (fltr->validate_only) {
2143 v_ret = VIRTCHNL_STATUS_SUCCESS;
2144 stat->status = VIRTCHNL_FDIR_SUCCESS;
2145 devm_kfree(dev, conf);
2146 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_ADD_FDIR_FILTER,
2147 v_ret, (u8 *)stat, len);
2148 goto exit;
2149 }
2150
2151 /* For raw FDIR filters created by the parser */
2152 if (conf->parser_ena) {
2153 ret = ice_vc_add_fdir_raw(vf, conf, &v_ret, stat, len);
2154 if (ret)
2155 goto err_free_conf;
2156 goto exit;
2157 }
2158
2159 is_tun = ice_fdir_is_tunnel(conf->ttype);
2160 ret = ice_vc_fdir_config_input_set(vf, fltr, conf, is_tun);
2161 if (ret) {
2162 v_ret = VIRTCHNL_STATUS_SUCCESS;
2163 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_CONFLICT;
2164 dev_err(dev, "VF %d: FDIR input set configure failed, ret:%d\n",
2165 vf->vf_id, ret);
2166 goto err_free_conf;
2167 }
2168
2169 ret = ice_vc_fdir_is_dup_fltr(vf, conf);
2170 if (ret) {
2171 v_ret = VIRTCHNL_STATUS_SUCCESS;
2172 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_EXIST;
2173 dev_dbg(dev, "VF %d: duplicated FDIR rule detected\n",
2174 vf->vf_id);
2175 goto err_free_conf;
2176 }
2177
2178 ret = ice_vc_fdir_insert_entry(vf, conf, &conf->flow_id);
2179 if (ret) {
2180 v_ret = VIRTCHNL_STATUS_SUCCESS;
2181 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2182 dev_dbg(dev, "VF %d: insert FDIR list failed\n", vf->vf_id);
2183 goto err_free_conf;
2184 }
2185
2186 ret = ice_vc_fdir_set_irq_ctx(vf, conf, VIRTCHNL_OP_ADD_FDIR_FILTER);
2187 if (ret) {
2188 v_ret = VIRTCHNL_STATUS_SUCCESS;
2189 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2190 dev_dbg(dev, "VF %d: set FDIR context failed\n", vf->vf_id);
2191 goto err_rem_entry;
2192 }
2193
2194 ret = ice_vc_fdir_write_fltr(vf, conf, true, is_tun);
2195 if (ret) {
2196 v_ret = VIRTCHNL_STATUS_SUCCESS;
2197 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2198 dev_err(dev, "VF %d: writing FDIR rule failed, ret:%d\n",
2199 vf->vf_id, ret);
2200 goto err_clr_irq;
2201 }
2202
2203exit:
2204 kfree(stat);
2205 return ret;
2206
2207err_clr_irq:
2208 ice_vc_fdir_clear_irq_ctx(vf);
2209err_rem_entry:
2210 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
2211err_free_conf:
2212 devm_kfree(dev, conf);
2213err_exit:
2214 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_ADD_FDIR_FILTER, v_ret,
2215 (u8 *)stat, len);
2216 kfree(stat);
2217 return ret;
2218}
2219
2220/**
2221 * ice_vc_del_fdir_raw - delete a raw FDIR filter for VF
2222 * @vf: pointer to the VF info
2223 * @conf: FDIR configuration for each filter
2224 * @v_ret: the final VIRTCHNL code
2225 * @stat: pointer to the VIRTCHNL_OP_DEL_FDIR_FILTER
2226 * @len: length of the stat
2227 *
2228 * Return: 0 on success or negative errno on failure.
2229 */
2230static int
2231ice_vc_del_fdir_raw(struct ice_vf *vf,
2232 struct virtchnl_fdir_fltr_conf *conf,
2233 enum virtchnl_status_code *v_ret,
2234 struct virtchnl_fdir_del *stat, int len)
2235{
2236 struct ice_vsi *vf_vsi, *ctrl_vsi;
2237 enum ice_block blk = ICE_BLK_FD;
2238 struct ice_fdir_prof_info *pi;
2239 struct ice_pf *pf = vf->pf;
2240 struct device *dev;
2241 struct ice_hw *hw;
2242 unsigned long id;
2243 u16 vsi_num;
2244 int ptg;
2245 int ret;
2246
2247 dev = ice_pf_to_dev(pf);
2248 hw = &pf->hw;
2249 *v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2250 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2251
2252 id = find_first_bit(conf->prof->ptypes, ICE_FLOW_PTYPE_MAX);
2253 ptg = hw->blk[ICE_BLK_FD].xlt1.t[id];
2254
2255 ret = ice_vc_fdir_write_fltr(vf, conf, false, false);
2256 if (ret) {
2257 dev_err(dev, "VF %u: deleting FDIR raw flow rule failed: %d\n",
2258 vf->vf_id, ret);
2259 return ret;
2260 }
2261
2262 vf_vsi = ice_get_vf_vsi(vf);
2263 if (!vf_vsi) {
2264 dev_err(dev, "Can not get FDIR vf_vsi for VF %u\n", vf->vf_id);
2265 return -ENODEV;
2266 }
2267
2268 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
2269 if (!ctrl_vsi) {
2270 dev_err(dev, "Can not get FDIR ctrl_vsi for VF %u\n",
2271 vf->vf_id);
2272 return -ENODEV;
2273 }
2274
2275 pi = &vf->fdir_prof_info[ptg];
2276 if (pi->fdir_active_cnt != 0) {
2277 pi->fdir_active_cnt--;
2278 /* Remove the profile id flow if no active FDIR rule left */
2279 if (!pi->fdir_active_cnt) {
2280 vsi_num = ice_get_hw_vsi_num(hw, ctrl_vsi->idx);
2281 ice_rem_prof_id_flow(hw, blk, vsi_num, id);
2282
2283 vsi_num = ice_get_hw_vsi_num(hw, vf_vsi->idx);
2284 ice_rem_prof_id_flow(hw, blk, vsi_num, id);
2285 }
2286 }
2287
2288 conf->parser_ena = false;
2289 return 0;
2290}
2291
2292/**
2293 * ice_vc_del_fdir_fltr - delete a FDIR filter for VF by the msg buffer
2294 * @vf: pointer to the VF info
2295 * @msg: pointer to the msg buffer
2296 *
2297 * Return: 0 on success, and other on error.
2298 */
2299int ice_vc_del_fdir_fltr(struct ice_vf *vf, u8 *msg)
2300{
2301 struct virtchnl_fdir_del *fltr = (struct virtchnl_fdir_del *)msg;
2302 struct virtchnl_fdir_del *stat = NULL;
2303 struct virtchnl_fdir_fltr_conf *conf;
2304 struct ice_vf_fdir *fdir = &vf->fdir;
2305 enum virtchnl_status_code v_ret;
2306 struct ice_fdir_fltr *input;
2307 enum ice_fltr_ptype flow;
2308 struct device *dev;
2309 struct ice_pf *pf;
2310 int is_tun = 0;
2311 int len = 0;
2312 int ret;
2313
2314 pf = vf->pf;
2315 dev = ice_pf_to_dev(pf);
2316 ret = ice_vc_fdir_param_check(vf, fltr->vsi_id);
2317 if (ret) {
2318 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
2319 dev_dbg(dev, "Parameter check for VF %d failed\n", vf->vf_id);
2320 goto err_exit;
2321 }
2322
2323 stat = kzalloc(sizeof(*stat), GFP_KERNEL);
2324 if (!stat) {
2325 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
2326 dev_dbg(dev, "Alloc stat for VF %d failed\n", vf->vf_id);
2327 goto err_exit;
2328 }
2329
2330 len = sizeof(*stat);
2331
2332 conf = ice_vc_fdir_lookup_entry(vf, fltr->flow_id);
2333 if (!conf) {
2334 v_ret = VIRTCHNL_STATUS_SUCCESS;
2335 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST;
2336 dev_dbg(dev, "VF %d: FDIR invalid flow_id:0x%X\n",
2337 vf->vf_id, fltr->flow_id);
2338 goto err_exit;
2339 }
2340
2341 /* Just return failure when ctrl_vsi idx is invalid */
2342 if (vf->ctrl_vsi_idx == ICE_NO_VSI) {
2343 v_ret = VIRTCHNL_STATUS_SUCCESS;
2344 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2345 dev_err(dev, "Invalid FDIR ctrl_vsi for VF %d\n", vf->vf_id);
2346 goto err_exit;
2347 }
2348
2349 ret = ice_vc_fdir_set_irq_ctx(vf, conf, VIRTCHNL_OP_DEL_FDIR_FILTER);
2350 if (ret) {
2351 v_ret = VIRTCHNL_STATUS_SUCCESS;
2352 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2353 dev_dbg(dev, "VF %d: set FDIR context failed\n", vf->vf_id);
2354 goto err_exit;
2355 }
2356
2357 /* For raw FDIR filters created by the parser */
2358 if (conf->parser_ena) {
2359 ret = ice_vc_del_fdir_raw(vf, conf, &v_ret, stat, len);
2360 if (ret)
2361 goto err_del_tmr;
2362 goto exit;
2363 }
2364
2365 is_tun = ice_fdir_is_tunnel(conf->ttype);
2366 ret = ice_vc_fdir_write_fltr(vf, conf, false, is_tun);
2367 if (ret) {
2368 v_ret = VIRTCHNL_STATUS_SUCCESS;
2369 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
2370 dev_err(dev, "VF %d: writing FDIR rule failed, ret:%d\n",
2371 vf->vf_id, ret);
2372 goto err_del_tmr;
2373 }
2374
2375 /* Remove unused profiles to avoid unexpected behaviors */
2376 input = &conf->input;
2377 flow = input->flow_type;
2378 if (fdir->fdir_fltr_cnt[flow][is_tun] == 1)
2379 ice_vc_fdir_rem_prof(vf, flow, is_tun);
2380
2381exit:
2382 kfree(stat);
2383
2384 return ret;
2385
2386err_del_tmr:
2387 ice_vc_fdir_clear_irq_ctx(vf);
2388err_exit:
2389 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_DEL_FDIR_FILTER, v_ret,
2390 (u8 *)stat, len);
2391 kfree(stat);
2392 return ret;
2393}
2394
2395/**
2396 * ice_vf_fdir_init - init FDIR resource for VF
2397 * @vf: pointer to the VF info
2398 */
2399void ice_vf_fdir_init(struct ice_vf *vf)
2400{
2401 struct ice_vf_fdir *fdir = &vf->fdir;
2402
2403 idr_init(&fdir->fdir_rule_idr);
2404 INIT_LIST_HEAD(&fdir->fdir_rule_list);
2405
2406 spin_lock_init(&fdir->ctx_lock);
2407 fdir->ctx_irq.flags = 0;
2408 fdir->ctx_done.flags = 0;
2409 ice_vc_fdir_reset_cnt_all(fdir);
2410}
2411
2412/**
2413 * ice_vf_fdir_exit - destroy FDIR resource for VF
2414 * @vf: pointer to the VF info
2415 */
2416void ice_vf_fdir_exit(struct ice_vf *vf)
2417{
2418 ice_vc_fdir_flush_entry(vf);
2419 idr_destroy(&vf->fdir.fdir_rule_idr);
2420 ice_vc_fdir_rem_prof_all(vf);
2421 ice_vc_fdir_free_prof_all(vf);
2422}
1// SPDX-License-Identifier: GPL-2.0
2/* Copyright (C) 2021, Intel Corporation. */
3
4#include "ice.h"
5#include "ice_base.h"
6#include "ice_lib.h"
7#include "ice_flow.h"
8#include "ice_vf_lib_private.h"
9
10#define to_fltr_conf_from_desc(p) \
11 container_of(p, struct virtchnl_fdir_fltr_conf, input)
12
13#define ICE_FLOW_PROF_TYPE_S 0
14#define ICE_FLOW_PROF_TYPE_M (0xFFFFFFFFULL << ICE_FLOW_PROF_TYPE_S)
15#define ICE_FLOW_PROF_VSI_S 32
16#define ICE_FLOW_PROF_VSI_M (0xFFFFFFFFULL << ICE_FLOW_PROF_VSI_S)
17
18/* Flow profile ID format:
19 * [0:31] - flow type, flow + tun_offs
20 * [32:63] - VSI index
21 */
22#define ICE_FLOW_PROF_FD(vsi, flow, tun_offs) \
23 ((u64)(((((flow) + (tun_offs)) & ICE_FLOW_PROF_TYPE_M)) | \
24 (((u64)(vsi) << ICE_FLOW_PROF_VSI_S) & ICE_FLOW_PROF_VSI_M)))
25
26#define GTPU_TEID_OFFSET 4
27#define GTPU_EH_QFI_OFFSET 1
28#define GTPU_EH_QFI_MASK 0x3F
29#define PFCP_S_OFFSET 0
30#define PFCP_S_MASK 0x1
31#define PFCP_PORT_NR 8805
32
33#define FDIR_INSET_FLAG_ESP_S 0
34#define FDIR_INSET_FLAG_ESP_M BIT_ULL(FDIR_INSET_FLAG_ESP_S)
35#define FDIR_INSET_FLAG_ESP_UDP BIT_ULL(FDIR_INSET_FLAG_ESP_S)
36#define FDIR_INSET_FLAG_ESP_IPSEC (0ULL << FDIR_INSET_FLAG_ESP_S)
37
38enum ice_fdir_tunnel_type {
39 ICE_FDIR_TUNNEL_TYPE_NONE = 0,
40 ICE_FDIR_TUNNEL_TYPE_GTPU,
41 ICE_FDIR_TUNNEL_TYPE_GTPU_EH,
42};
43
44struct virtchnl_fdir_fltr_conf {
45 struct ice_fdir_fltr input;
46 enum ice_fdir_tunnel_type ttype;
47 u64 inset_flag;
48 u32 flow_id;
49};
50
51struct virtchnl_fdir_inset_map {
52 enum virtchnl_proto_hdr_field field;
53 enum ice_flow_field fld;
54 u64 flag;
55 u64 mask;
56};
57
58static const struct virtchnl_fdir_inset_map fdir_inset_map[] = {
59 {VIRTCHNL_PROTO_HDR_ETH_ETHERTYPE, ICE_FLOW_FIELD_IDX_ETH_TYPE, 0, 0},
60 {VIRTCHNL_PROTO_HDR_IPV4_SRC, ICE_FLOW_FIELD_IDX_IPV4_SA, 0, 0},
61 {VIRTCHNL_PROTO_HDR_IPV4_DST, ICE_FLOW_FIELD_IDX_IPV4_DA, 0, 0},
62 {VIRTCHNL_PROTO_HDR_IPV4_DSCP, ICE_FLOW_FIELD_IDX_IPV4_DSCP, 0, 0},
63 {VIRTCHNL_PROTO_HDR_IPV4_TTL, ICE_FLOW_FIELD_IDX_IPV4_TTL, 0, 0},
64 {VIRTCHNL_PROTO_HDR_IPV4_PROT, ICE_FLOW_FIELD_IDX_IPV4_PROT, 0, 0},
65 {VIRTCHNL_PROTO_HDR_IPV6_SRC, ICE_FLOW_FIELD_IDX_IPV6_SA, 0, 0},
66 {VIRTCHNL_PROTO_HDR_IPV6_DST, ICE_FLOW_FIELD_IDX_IPV6_DA, 0, 0},
67 {VIRTCHNL_PROTO_HDR_IPV6_TC, ICE_FLOW_FIELD_IDX_IPV6_DSCP, 0, 0},
68 {VIRTCHNL_PROTO_HDR_IPV6_HOP_LIMIT, ICE_FLOW_FIELD_IDX_IPV6_TTL, 0, 0},
69 {VIRTCHNL_PROTO_HDR_IPV6_PROT, ICE_FLOW_FIELD_IDX_IPV6_PROT, 0, 0},
70 {VIRTCHNL_PROTO_HDR_UDP_SRC_PORT, ICE_FLOW_FIELD_IDX_UDP_SRC_PORT, 0, 0},
71 {VIRTCHNL_PROTO_HDR_UDP_DST_PORT, ICE_FLOW_FIELD_IDX_UDP_DST_PORT, 0, 0},
72 {VIRTCHNL_PROTO_HDR_TCP_SRC_PORT, ICE_FLOW_FIELD_IDX_TCP_SRC_PORT, 0, 0},
73 {VIRTCHNL_PROTO_HDR_TCP_DST_PORT, ICE_FLOW_FIELD_IDX_TCP_DST_PORT, 0, 0},
74 {VIRTCHNL_PROTO_HDR_SCTP_SRC_PORT, ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT, 0, 0},
75 {VIRTCHNL_PROTO_HDR_SCTP_DST_PORT, ICE_FLOW_FIELD_IDX_SCTP_DST_PORT, 0, 0},
76 {VIRTCHNL_PROTO_HDR_GTPU_IP_TEID, ICE_FLOW_FIELD_IDX_GTPU_IP_TEID, 0, 0},
77 {VIRTCHNL_PROTO_HDR_GTPU_EH_QFI, ICE_FLOW_FIELD_IDX_GTPU_EH_QFI, 0, 0},
78 {VIRTCHNL_PROTO_HDR_ESP_SPI, ICE_FLOW_FIELD_IDX_ESP_SPI,
79 FDIR_INSET_FLAG_ESP_IPSEC, FDIR_INSET_FLAG_ESP_M},
80 {VIRTCHNL_PROTO_HDR_ESP_SPI, ICE_FLOW_FIELD_IDX_NAT_T_ESP_SPI,
81 FDIR_INSET_FLAG_ESP_UDP, FDIR_INSET_FLAG_ESP_M},
82 {VIRTCHNL_PROTO_HDR_AH_SPI, ICE_FLOW_FIELD_IDX_AH_SPI, 0, 0},
83 {VIRTCHNL_PROTO_HDR_L2TPV3_SESS_ID, ICE_FLOW_FIELD_IDX_L2TPV3_SESS_ID, 0, 0},
84 {VIRTCHNL_PROTO_HDR_PFCP_S_FIELD, ICE_FLOW_FIELD_IDX_UDP_DST_PORT, 0, 0},
85};
86
87/**
88 * ice_vc_fdir_param_check
89 * @vf: pointer to the VF structure
90 * @vsi_id: VF relative VSI ID
91 *
92 * Check for the valid VSI ID, PF's state and VF's state
93 *
94 * Return: 0 on success, and -EINVAL on error.
95 */
96static int
97ice_vc_fdir_param_check(struct ice_vf *vf, u16 vsi_id)
98{
99 struct ice_pf *pf = vf->pf;
100
101 if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
102 return -EINVAL;
103
104 if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states))
105 return -EINVAL;
106
107 if (!(vf->driver_caps & VIRTCHNL_VF_OFFLOAD_FDIR_PF))
108 return -EINVAL;
109
110 if (vsi_id != vf->lan_vsi_num)
111 return -EINVAL;
112
113 if (!ice_vc_isvalid_vsi_id(vf, vsi_id))
114 return -EINVAL;
115
116 if (!pf->vsi[vf->lan_vsi_idx])
117 return -EINVAL;
118
119 return 0;
120}
121
122/**
123 * ice_vf_start_ctrl_vsi
124 * @vf: pointer to the VF structure
125 *
126 * Allocate ctrl_vsi for the first time and open the ctrl_vsi port for VF
127 *
128 * Return: 0 on success, and other on error.
129 */
130static int ice_vf_start_ctrl_vsi(struct ice_vf *vf)
131{
132 struct ice_pf *pf = vf->pf;
133 struct ice_vsi *ctrl_vsi;
134 struct device *dev;
135 int err;
136
137 dev = ice_pf_to_dev(pf);
138 if (vf->ctrl_vsi_idx != ICE_NO_VSI)
139 return -EEXIST;
140
141 ctrl_vsi = ice_vf_ctrl_vsi_setup(vf);
142 if (!ctrl_vsi) {
143 dev_dbg(dev, "Could not setup control VSI for VF %d\n",
144 vf->vf_id);
145 return -ENOMEM;
146 }
147
148 err = ice_vsi_open_ctrl(ctrl_vsi);
149 if (err) {
150 dev_dbg(dev, "Could not open control VSI for VF %d\n",
151 vf->vf_id);
152 goto err_vsi_open;
153 }
154
155 return 0;
156
157err_vsi_open:
158 ice_vsi_release(ctrl_vsi);
159 if (vf->ctrl_vsi_idx != ICE_NO_VSI) {
160 pf->vsi[vf->ctrl_vsi_idx] = NULL;
161 vf->ctrl_vsi_idx = ICE_NO_VSI;
162 }
163 return err;
164}
165
166/**
167 * ice_vc_fdir_alloc_prof - allocate profile for this filter flow type
168 * @vf: pointer to the VF structure
169 * @flow: filter flow type
170 *
171 * Return: 0 on success, and other on error.
172 */
173static int
174ice_vc_fdir_alloc_prof(struct ice_vf *vf, enum ice_fltr_ptype flow)
175{
176 struct ice_vf_fdir *fdir = &vf->fdir;
177
178 if (!fdir->fdir_prof) {
179 fdir->fdir_prof = devm_kcalloc(ice_pf_to_dev(vf->pf),
180 ICE_FLTR_PTYPE_MAX,
181 sizeof(*fdir->fdir_prof),
182 GFP_KERNEL);
183 if (!fdir->fdir_prof)
184 return -ENOMEM;
185 }
186
187 if (!fdir->fdir_prof[flow]) {
188 fdir->fdir_prof[flow] = devm_kzalloc(ice_pf_to_dev(vf->pf),
189 sizeof(**fdir->fdir_prof),
190 GFP_KERNEL);
191 if (!fdir->fdir_prof[flow])
192 return -ENOMEM;
193 }
194
195 return 0;
196}
197
198/**
199 * ice_vc_fdir_free_prof - free profile for this filter flow type
200 * @vf: pointer to the VF structure
201 * @flow: filter flow type
202 */
203static void
204ice_vc_fdir_free_prof(struct ice_vf *vf, enum ice_fltr_ptype flow)
205{
206 struct ice_vf_fdir *fdir = &vf->fdir;
207
208 if (!fdir->fdir_prof)
209 return;
210
211 if (!fdir->fdir_prof[flow])
212 return;
213
214 devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof[flow]);
215 fdir->fdir_prof[flow] = NULL;
216}
217
218/**
219 * ice_vc_fdir_free_prof_all - free all the profile for this VF
220 * @vf: pointer to the VF structure
221 */
222static void ice_vc_fdir_free_prof_all(struct ice_vf *vf)
223{
224 struct ice_vf_fdir *fdir = &vf->fdir;
225 enum ice_fltr_ptype flow;
226
227 if (!fdir->fdir_prof)
228 return;
229
230 for (flow = ICE_FLTR_PTYPE_NONF_NONE; flow < ICE_FLTR_PTYPE_MAX; flow++)
231 ice_vc_fdir_free_prof(vf, flow);
232
233 devm_kfree(ice_pf_to_dev(vf->pf), fdir->fdir_prof);
234 fdir->fdir_prof = NULL;
235}
236
237/**
238 * ice_vc_fdir_parse_flow_fld
239 * @proto_hdr: virtual channel protocol filter header
240 * @conf: FDIR configuration for each filter
241 * @fld: field type array
242 * @fld_cnt: field counter
243 *
244 * Parse the virtual channel filter header and store them into field type array
245 *
246 * Return: 0 on success, and other on error.
247 */
248static int
249ice_vc_fdir_parse_flow_fld(struct virtchnl_proto_hdr *proto_hdr,
250 struct virtchnl_fdir_fltr_conf *conf,
251 enum ice_flow_field *fld, int *fld_cnt)
252{
253 struct virtchnl_proto_hdr hdr;
254 u32 i;
255
256 memcpy(&hdr, proto_hdr, sizeof(hdr));
257
258 for (i = 0; (i < ARRAY_SIZE(fdir_inset_map)) &&
259 VIRTCHNL_GET_PROTO_HDR_FIELD(&hdr); i++)
260 if (VIRTCHNL_TEST_PROTO_HDR(&hdr, fdir_inset_map[i].field)) {
261 if (fdir_inset_map[i].mask &&
262 ((fdir_inset_map[i].mask & conf->inset_flag) !=
263 fdir_inset_map[i].flag))
264 continue;
265
266 fld[*fld_cnt] = fdir_inset_map[i].fld;
267 *fld_cnt += 1;
268 if (*fld_cnt >= ICE_FLOW_FIELD_IDX_MAX)
269 return -EINVAL;
270 VIRTCHNL_DEL_PROTO_HDR_FIELD(&hdr,
271 fdir_inset_map[i].field);
272 }
273
274 return 0;
275}
276
277/**
278 * ice_vc_fdir_set_flow_fld
279 * @vf: pointer to the VF structure
280 * @fltr: virtual channel add cmd buffer
281 * @conf: FDIR configuration for each filter
282 * @seg: array of one or more packet segments that describe the flow
283 *
284 * Parse the virtual channel add msg buffer's field vector and store them into
285 * flow's packet segment field
286 *
287 * Return: 0 on success, and other on error.
288 */
289static int
290ice_vc_fdir_set_flow_fld(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
291 struct virtchnl_fdir_fltr_conf *conf,
292 struct ice_flow_seg_info *seg)
293{
294 struct virtchnl_fdir_rule *rule = &fltr->rule_cfg;
295 enum ice_flow_field fld[ICE_FLOW_FIELD_IDX_MAX];
296 struct device *dev = ice_pf_to_dev(vf->pf);
297 struct virtchnl_proto_hdrs *proto;
298 int fld_cnt = 0;
299 int i;
300
301 proto = &rule->proto_hdrs;
302 for (i = 0; i < proto->count; i++) {
303 struct virtchnl_proto_hdr *hdr = &proto->proto_hdr[i];
304 int ret;
305
306 ret = ice_vc_fdir_parse_flow_fld(hdr, conf, fld, &fld_cnt);
307 if (ret)
308 return ret;
309 }
310
311 if (fld_cnt == 0) {
312 dev_dbg(dev, "Empty input set for VF %d\n", vf->vf_id);
313 return -EINVAL;
314 }
315
316 for (i = 0; i < fld_cnt; i++)
317 ice_flow_set_fld(seg, fld[i],
318 ICE_FLOW_FLD_OFF_INVAL,
319 ICE_FLOW_FLD_OFF_INVAL,
320 ICE_FLOW_FLD_OFF_INVAL, false);
321
322 return 0;
323}
324
325/**
326 * ice_vc_fdir_set_flow_hdr - config the flow's packet segment header
327 * @vf: pointer to the VF structure
328 * @conf: FDIR configuration for each filter
329 * @seg: array of one or more packet segments that describe the flow
330 *
331 * Return: 0 on success, and other on error.
332 */
333static int
334ice_vc_fdir_set_flow_hdr(struct ice_vf *vf,
335 struct virtchnl_fdir_fltr_conf *conf,
336 struct ice_flow_seg_info *seg)
337{
338 enum ice_fltr_ptype flow = conf->input.flow_type;
339 enum ice_fdir_tunnel_type ttype = conf->ttype;
340 struct device *dev = ice_pf_to_dev(vf->pf);
341
342 switch (flow) {
343 case ICE_FLTR_PTYPE_NON_IP_L2:
344 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ETH_NON_IP);
345 break;
346 case ICE_FLTR_PTYPE_NONF_IPV4_L2TPV3:
347 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
348 ICE_FLOW_SEG_HDR_IPV4 |
349 ICE_FLOW_SEG_HDR_IPV_OTHER);
350 break;
351 case ICE_FLTR_PTYPE_NONF_IPV4_ESP:
352 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
353 ICE_FLOW_SEG_HDR_IPV4 |
354 ICE_FLOW_SEG_HDR_IPV_OTHER);
355 break;
356 case ICE_FLTR_PTYPE_NONF_IPV4_AH:
357 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
358 ICE_FLOW_SEG_HDR_IPV4 |
359 ICE_FLOW_SEG_HDR_IPV_OTHER);
360 break;
361 case ICE_FLTR_PTYPE_NONF_IPV4_NAT_T_ESP:
362 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
363 ICE_FLOW_SEG_HDR_IPV4 |
364 ICE_FLOW_SEG_HDR_IPV_OTHER);
365 break;
366 case ICE_FLTR_PTYPE_NONF_IPV4_PFCP_NODE:
367 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
368 ICE_FLOW_SEG_HDR_IPV4 |
369 ICE_FLOW_SEG_HDR_IPV_OTHER);
370 break;
371 case ICE_FLTR_PTYPE_NONF_IPV4_PFCP_SESSION:
372 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
373 ICE_FLOW_SEG_HDR_IPV4 |
374 ICE_FLOW_SEG_HDR_IPV_OTHER);
375 break;
376 case ICE_FLTR_PTYPE_NONF_IPV4_OTHER:
377 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4 |
378 ICE_FLOW_SEG_HDR_IPV_OTHER);
379 break;
380 case ICE_FLTR_PTYPE_NONF_IPV4_TCP:
381 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
382 ICE_FLOW_SEG_HDR_IPV4 |
383 ICE_FLOW_SEG_HDR_IPV_OTHER);
384 break;
385 case ICE_FLTR_PTYPE_NONF_IPV4_UDP:
386 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
387 ICE_FLOW_SEG_HDR_IPV4 |
388 ICE_FLOW_SEG_HDR_IPV_OTHER);
389 break;
390 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_UDP:
391 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_TCP:
392 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_ICMP:
393 case ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_OTHER:
394 if (ttype == ICE_FDIR_TUNNEL_TYPE_GTPU) {
395 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_IP |
396 ICE_FLOW_SEG_HDR_IPV4 |
397 ICE_FLOW_SEG_HDR_IPV_OTHER);
398 } else if (ttype == ICE_FDIR_TUNNEL_TYPE_GTPU_EH) {
399 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_EH |
400 ICE_FLOW_SEG_HDR_GTPU_IP |
401 ICE_FLOW_SEG_HDR_IPV4 |
402 ICE_FLOW_SEG_HDR_IPV_OTHER);
403 } else {
404 dev_dbg(dev, "Invalid tunnel type 0x%x for VF %d\n",
405 flow, vf->vf_id);
406 return -EINVAL;
407 }
408 break;
409 case ICE_FLTR_PTYPE_NONF_IPV4_SCTP:
410 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
411 ICE_FLOW_SEG_HDR_IPV4 |
412 ICE_FLOW_SEG_HDR_IPV_OTHER);
413 break;
414 case ICE_FLTR_PTYPE_NONF_IPV6_L2TPV3:
415 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
416 ICE_FLOW_SEG_HDR_IPV6 |
417 ICE_FLOW_SEG_HDR_IPV_OTHER);
418 break;
419 case ICE_FLTR_PTYPE_NONF_IPV6_ESP:
420 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
421 ICE_FLOW_SEG_HDR_IPV6 |
422 ICE_FLOW_SEG_HDR_IPV_OTHER);
423 break;
424 case ICE_FLTR_PTYPE_NONF_IPV6_AH:
425 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
426 ICE_FLOW_SEG_HDR_IPV6 |
427 ICE_FLOW_SEG_HDR_IPV_OTHER);
428 break;
429 case ICE_FLTR_PTYPE_NONF_IPV6_NAT_T_ESP:
430 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
431 ICE_FLOW_SEG_HDR_IPV6 |
432 ICE_FLOW_SEG_HDR_IPV_OTHER);
433 break;
434 case ICE_FLTR_PTYPE_NONF_IPV6_PFCP_NODE:
435 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
436 ICE_FLOW_SEG_HDR_IPV6 |
437 ICE_FLOW_SEG_HDR_IPV_OTHER);
438 break;
439 case ICE_FLTR_PTYPE_NONF_IPV6_PFCP_SESSION:
440 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
441 ICE_FLOW_SEG_HDR_IPV6 |
442 ICE_FLOW_SEG_HDR_IPV_OTHER);
443 break;
444 case ICE_FLTR_PTYPE_NONF_IPV6_OTHER:
445 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6 |
446 ICE_FLOW_SEG_HDR_IPV_OTHER);
447 break;
448 case ICE_FLTR_PTYPE_NONF_IPV6_TCP:
449 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
450 ICE_FLOW_SEG_HDR_IPV6 |
451 ICE_FLOW_SEG_HDR_IPV_OTHER);
452 break;
453 case ICE_FLTR_PTYPE_NONF_IPV6_UDP:
454 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
455 ICE_FLOW_SEG_HDR_IPV6 |
456 ICE_FLOW_SEG_HDR_IPV_OTHER);
457 break;
458 case ICE_FLTR_PTYPE_NONF_IPV6_SCTP:
459 ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
460 ICE_FLOW_SEG_HDR_IPV6 |
461 ICE_FLOW_SEG_HDR_IPV_OTHER);
462 break;
463 default:
464 dev_dbg(dev, "Invalid flow type 0x%x for VF %d failed\n",
465 flow, vf->vf_id);
466 return -EINVAL;
467 }
468
469 return 0;
470}
471
472/**
473 * ice_vc_fdir_rem_prof - remove profile for this filter flow type
474 * @vf: pointer to the VF structure
475 * @flow: filter flow type
476 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
477 */
478static void
479ice_vc_fdir_rem_prof(struct ice_vf *vf, enum ice_fltr_ptype flow, int tun)
480{
481 struct ice_vf_fdir *fdir = &vf->fdir;
482 struct ice_fd_hw_prof *vf_prof;
483 struct ice_pf *pf = vf->pf;
484 struct ice_vsi *vf_vsi;
485 struct device *dev;
486 struct ice_hw *hw;
487 u64 prof_id;
488 int i;
489
490 dev = ice_pf_to_dev(pf);
491 hw = &pf->hw;
492 if (!fdir->fdir_prof || !fdir->fdir_prof[flow])
493 return;
494
495 vf_prof = fdir->fdir_prof[flow];
496
497 vf_vsi = pf->vsi[vf->lan_vsi_idx];
498 if (!vf_vsi) {
499 dev_dbg(dev, "NULL vf %d vsi pointer\n", vf->vf_id);
500 return;
501 }
502
503 if (!fdir->prof_entry_cnt[flow][tun])
504 return;
505
506 prof_id = ICE_FLOW_PROF_FD(vf_vsi->vsi_num,
507 flow, tun ? ICE_FLTR_PTYPE_MAX : 0);
508
509 for (i = 0; i < fdir->prof_entry_cnt[flow][tun]; i++)
510 if (vf_prof->entry_h[i][tun]) {
511 u16 vsi_num = ice_get_hw_vsi_num(hw, vf_prof->vsi_h[i]);
512
513 ice_rem_prof_id_flow(hw, ICE_BLK_FD, vsi_num, prof_id);
514 ice_flow_rem_entry(hw, ICE_BLK_FD,
515 vf_prof->entry_h[i][tun]);
516 vf_prof->entry_h[i][tun] = 0;
517 }
518
519 ice_flow_rem_prof(hw, ICE_BLK_FD, prof_id);
520 devm_kfree(dev, vf_prof->fdir_seg[tun]);
521 vf_prof->fdir_seg[tun] = NULL;
522
523 for (i = 0; i < vf_prof->cnt; i++)
524 vf_prof->vsi_h[i] = 0;
525
526 fdir->prof_entry_cnt[flow][tun] = 0;
527}
528
529/**
530 * ice_vc_fdir_rem_prof_all - remove profile for this VF
531 * @vf: pointer to the VF structure
532 */
533static void ice_vc_fdir_rem_prof_all(struct ice_vf *vf)
534{
535 enum ice_fltr_ptype flow;
536
537 for (flow = ICE_FLTR_PTYPE_NONF_NONE;
538 flow < ICE_FLTR_PTYPE_MAX; flow++) {
539 ice_vc_fdir_rem_prof(vf, flow, 0);
540 ice_vc_fdir_rem_prof(vf, flow, 1);
541 }
542}
543
544/**
545 * ice_vc_fdir_write_flow_prof
546 * @vf: pointer to the VF structure
547 * @flow: filter flow type
548 * @seg: array of one or more packet segments that describe the flow
549 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
550 *
551 * Write the flow's profile config and packet segment into the hardware
552 *
553 * Return: 0 on success, and other on error.
554 */
555static int
556ice_vc_fdir_write_flow_prof(struct ice_vf *vf, enum ice_fltr_ptype flow,
557 struct ice_flow_seg_info *seg, int tun)
558{
559 struct ice_vf_fdir *fdir = &vf->fdir;
560 struct ice_vsi *vf_vsi, *ctrl_vsi;
561 struct ice_flow_seg_info *old_seg;
562 struct ice_flow_prof *prof = NULL;
563 struct ice_fd_hw_prof *vf_prof;
564 struct device *dev;
565 struct ice_pf *pf;
566 struct ice_hw *hw;
567 u64 entry1_h = 0;
568 u64 entry2_h = 0;
569 u64 prof_id;
570 int ret;
571
572 pf = vf->pf;
573 dev = ice_pf_to_dev(pf);
574 hw = &pf->hw;
575 vf_vsi = pf->vsi[vf->lan_vsi_idx];
576 if (!vf_vsi)
577 return -EINVAL;
578
579 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
580 if (!ctrl_vsi)
581 return -EINVAL;
582
583 vf_prof = fdir->fdir_prof[flow];
584 old_seg = vf_prof->fdir_seg[tun];
585 if (old_seg) {
586 if (!memcmp(old_seg, seg, sizeof(*seg))) {
587 dev_dbg(dev, "Duplicated profile for VF %d!\n",
588 vf->vf_id);
589 return -EEXIST;
590 }
591
592 if (fdir->fdir_fltr_cnt[flow][tun]) {
593 ret = -EINVAL;
594 dev_dbg(dev, "Input set conflicts for VF %d\n",
595 vf->vf_id);
596 goto err_exit;
597 }
598
599 /* remove previously allocated profile */
600 ice_vc_fdir_rem_prof(vf, flow, tun);
601 }
602
603 prof_id = ICE_FLOW_PROF_FD(vf_vsi->vsi_num, flow,
604 tun ? ICE_FLTR_PTYPE_MAX : 0);
605
606 ret = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, prof_id, seg,
607 tun + 1, &prof);
608 if (ret) {
609 dev_dbg(dev, "Could not add VSI flow 0x%x for VF %d\n",
610 flow, vf->vf_id);
611 goto err_exit;
612 }
613
614 ret = ice_flow_add_entry(hw, ICE_BLK_FD, prof_id, vf_vsi->idx,
615 vf_vsi->idx, ICE_FLOW_PRIO_NORMAL,
616 seg, &entry1_h);
617 if (ret) {
618 dev_dbg(dev, "Could not add flow 0x%x VSI entry for VF %d\n",
619 flow, vf->vf_id);
620 goto err_prof;
621 }
622
623 ret = ice_flow_add_entry(hw, ICE_BLK_FD, prof_id, vf_vsi->idx,
624 ctrl_vsi->idx, ICE_FLOW_PRIO_NORMAL,
625 seg, &entry2_h);
626 if (ret) {
627 dev_dbg(dev,
628 "Could not add flow 0x%x Ctrl VSI entry for VF %d\n",
629 flow, vf->vf_id);
630 goto err_entry_1;
631 }
632
633 vf_prof->fdir_seg[tun] = seg;
634 vf_prof->cnt = 0;
635 fdir->prof_entry_cnt[flow][tun] = 0;
636
637 vf_prof->entry_h[vf_prof->cnt][tun] = entry1_h;
638 vf_prof->vsi_h[vf_prof->cnt] = vf_vsi->idx;
639 vf_prof->cnt++;
640 fdir->prof_entry_cnt[flow][tun]++;
641
642 vf_prof->entry_h[vf_prof->cnt][tun] = entry2_h;
643 vf_prof->vsi_h[vf_prof->cnt] = ctrl_vsi->idx;
644 vf_prof->cnt++;
645 fdir->prof_entry_cnt[flow][tun]++;
646
647 return 0;
648
649err_entry_1:
650 ice_rem_prof_id_flow(hw, ICE_BLK_FD,
651 ice_get_hw_vsi_num(hw, vf_vsi->idx), prof_id);
652 ice_flow_rem_entry(hw, ICE_BLK_FD, entry1_h);
653err_prof:
654 ice_flow_rem_prof(hw, ICE_BLK_FD, prof_id);
655err_exit:
656 return ret;
657}
658
659/**
660 * ice_vc_fdir_config_input_set
661 * @vf: pointer to the VF structure
662 * @fltr: virtual channel add cmd buffer
663 * @conf: FDIR configuration for each filter
664 * @tun: 0 implies non-tunnel type filter, 1 implies tunnel type filter
665 *
666 * Config the input set type and value for virtual channel add msg buffer
667 *
668 * Return: 0 on success, and other on error.
669 */
670static int
671ice_vc_fdir_config_input_set(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
672 struct virtchnl_fdir_fltr_conf *conf, int tun)
673{
674 struct ice_fdir_fltr *input = &conf->input;
675 struct device *dev = ice_pf_to_dev(vf->pf);
676 struct ice_flow_seg_info *seg;
677 enum ice_fltr_ptype flow;
678 int ret;
679
680 flow = input->flow_type;
681 ret = ice_vc_fdir_alloc_prof(vf, flow);
682 if (ret) {
683 dev_dbg(dev, "Alloc flow prof for VF %d failed\n", vf->vf_id);
684 return ret;
685 }
686
687 seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
688 if (!seg)
689 return -ENOMEM;
690
691 ret = ice_vc_fdir_set_flow_fld(vf, fltr, conf, seg);
692 if (ret) {
693 dev_dbg(dev, "Set flow field for VF %d failed\n", vf->vf_id);
694 goto err_exit;
695 }
696
697 ret = ice_vc_fdir_set_flow_hdr(vf, conf, seg);
698 if (ret) {
699 dev_dbg(dev, "Set flow hdr for VF %d failed\n", vf->vf_id);
700 goto err_exit;
701 }
702
703 ret = ice_vc_fdir_write_flow_prof(vf, flow, seg, tun);
704 if (ret == -EEXIST) {
705 devm_kfree(dev, seg);
706 } else if (ret) {
707 dev_dbg(dev, "Write flow profile for VF %d failed\n",
708 vf->vf_id);
709 goto err_exit;
710 }
711
712 return 0;
713
714err_exit:
715 devm_kfree(dev, seg);
716 return ret;
717}
718
719/**
720 * ice_vc_fdir_parse_pattern
721 * @vf: pointer to the VF info
722 * @fltr: virtual channel add cmd buffer
723 * @conf: FDIR configuration for each filter
724 *
725 * Parse the virtual channel filter's pattern and store them into conf
726 *
727 * Return: 0 on success, and other on error.
728 */
729static int
730ice_vc_fdir_parse_pattern(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
731 struct virtchnl_fdir_fltr_conf *conf)
732{
733 struct virtchnl_proto_hdrs *proto = &fltr->rule_cfg.proto_hdrs;
734 enum virtchnl_proto_hdr_type l3 = VIRTCHNL_PROTO_HDR_NONE;
735 enum virtchnl_proto_hdr_type l4 = VIRTCHNL_PROTO_HDR_NONE;
736 struct device *dev = ice_pf_to_dev(vf->pf);
737 struct ice_fdir_fltr *input = &conf->input;
738 int i;
739
740 if (proto->count > VIRTCHNL_MAX_NUM_PROTO_HDRS) {
741 dev_dbg(dev, "Invalid protocol count:0x%x for VF %d\n",
742 proto->count, vf->vf_id);
743 return -EINVAL;
744 }
745
746 for (i = 0; i < proto->count; i++) {
747 struct virtchnl_proto_hdr *hdr = &proto->proto_hdr[i];
748 struct ip_esp_hdr *esph;
749 struct ip_auth_hdr *ah;
750 struct sctphdr *sctph;
751 struct ipv6hdr *ip6h;
752 struct udphdr *udph;
753 struct tcphdr *tcph;
754 struct ethhdr *eth;
755 struct iphdr *iph;
756 u8 s_field;
757 u8 *rawh;
758
759 switch (hdr->type) {
760 case VIRTCHNL_PROTO_HDR_ETH:
761 eth = (struct ethhdr *)hdr->buffer;
762 input->flow_type = ICE_FLTR_PTYPE_NON_IP_L2;
763
764 if (hdr->field_selector)
765 input->ext_data.ether_type = eth->h_proto;
766 break;
767 case VIRTCHNL_PROTO_HDR_IPV4:
768 iph = (struct iphdr *)hdr->buffer;
769 l3 = VIRTCHNL_PROTO_HDR_IPV4;
770 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_OTHER;
771
772 if (hdr->field_selector) {
773 input->ip.v4.src_ip = iph->saddr;
774 input->ip.v4.dst_ip = iph->daddr;
775 input->ip.v4.tos = iph->tos;
776 input->ip.v4.proto = iph->protocol;
777 }
778 break;
779 case VIRTCHNL_PROTO_HDR_IPV6:
780 ip6h = (struct ipv6hdr *)hdr->buffer;
781 l3 = VIRTCHNL_PROTO_HDR_IPV6;
782 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_OTHER;
783
784 if (hdr->field_selector) {
785 memcpy(input->ip.v6.src_ip,
786 ip6h->saddr.in6_u.u6_addr8,
787 sizeof(ip6h->saddr));
788 memcpy(input->ip.v6.dst_ip,
789 ip6h->daddr.in6_u.u6_addr8,
790 sizeof(ip6h->daddr));
791 input->ip.v6.tc = ((u8)(ip6h->priority) << 4) |
792 (ip6h->flow_lbl[0] >> 4);
793 input->ip.v6.proto = ip6h->nexthdr;
794 }
795 break;
796 case VIRTCHNL_PROTO_HDR_TCP:
797 tcph = (struct tcphdr *)hdr->buffer;
798 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
799 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_TCP;
800 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
801 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_TCP;
802
803 if (hdr->field_selector) {
804 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
805 input->ip.v4.src_port = tcph->source;
806 input->ip.v4.dst_port = tcph->dest;
807 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
808 input->ip.v6.src_port = tcph->source;
809 input->ip.v6.dst_port = tcph->dest;
810 }
811 }
812 break;
813 case VIRTCHNL_PROTO_HDR_UDP:
814 udph = (struct udphdr *)hdr->buffer;
815 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
816 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_UDP;
817 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
818 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_UDP;
819
820 if (hdr->field_selector) {
821 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
822 input->ip.v4.src_port = udph->source;
823 input->ip.v4.dst_port = udph->dest;
824 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
825 input->ip.v6.src_port = udph->source;
826 input->ip.v6.dst_port = udph->dest;
827 }
828 }
829 break;
830 case VIRTCHNL_PROTO_HDR_SCTP:
831 sctph = (struct sctphdr *)hdr->buffer;
832 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
833 input->flow_type =
834 ICE_FLTR_PTYPE_NONF_IPV4_SCTP;
835 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
836 input->flow_type =
837 ICE_FLTR_PTYPE_NONF_IPV6_SCTP;
838
839 if (hdr->field_selector) {
840 if (l3 == VIRTCHNL_PROTO_HDR_IPV4) {
841 input->ip.v4.src_port = sctph->source;
842 input->ip.v4.dst_port = sctph->dest;
843 } else if (l3 == VIRTCHNL_PROTO_HDR_IPV6) {
844 input->ip.v6.src_port = sctph->source;
845 input->ip.v6.dst_port = sctph->dest;
846 }
847 }
848 break;
849 case VIRTCHNL_PROTO_HDR_L2TPV3:
850 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
851 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_L2TPV3;
852 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
853 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_L2TPV3;
854
855 if (hdr->field_selector)
856 input->l2tpv3_data.session_id = *((__be32 *)hdr->buffer);
857 break;
858 case VIRTCHNL_PROTO_HDR_ESP:
859 esph = (struct ip_esp_hdr *)hdr->buffer;
860 if (l3 == VIRTCHNL_PROTO_HDR_IPV4 &&
861 l4 == VIRTCHNL_PROTO_HDR_UDP)
862 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_NAT_T_ESP;
863 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 &&
864 l4 == VIRTCHNL_PROTO_HDR_UDP)
865 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_NAT_T_ESP;
866 else if (l3 == VIRTCHNL_PROTO_HDR_IPV4 &&
867 l4 == VIRTCHNL_PROTO_HDR_NONE)
868 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_ESP;
869 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 &&
870 l4 == VIRTCHNL_PROTO_HDR_NONE)
871 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_ESP;
872
873 if (l4 == VIRTCHNL_PROTO_HDR_UDP)
874 conf->inset_flag |= FDIR_INSET_FLAG_ESP_UDP;
875 else
876 conf->inset_flag |= FDIR_INSET_FLAG_ESP_IPSEC;
877
878 if (hdr->field_selector) {
879 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
880 input->ip.v4.sec_parm_idx = esph->spi;
881 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
882 input->ip.v6.sec_parm_idx = esph->spi;
883 }
884 break;
885 case VIRTCHNL_PROTO_HDR_AH:
886 ah = (struct ip_auth_hdr *)hdr->buffer;
887 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
888 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_AH;
889 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
890 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_AH;
891
892 if (hdr->field_selector) {
893 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
894 input->ip.v4.sec_parm_idx = ah->spi;
895 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
896 input->ip.v6.sec_parm_idx = ah->spi;
897 }
898 break;
899 case VIRTCHNL_PROTO_HDR_PFCP:
900 rawh = (u8 *)hdr->buffer;
901 s_field = (rawh[0] >> PFCP_S_OFFSET) & PFCP_S_MASK;
902 if (l3 == VIRTCHNL_PROTO_HDR_IPV4 && s_field == 0)
903 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_PFCP_NODE;
904 else if (l3 == VIRTCHNL_PROTO_HDR_IPV4 && s_field == 1)
905 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_PFCP_SESSION;
906 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 && s_field == 0)
907 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_PFCP_NODE;
908 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6 && s_field == 1)
909 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV6_PFCP_SESSION;
910
911 if (hdr->field_selector) {
912 if (l3 == VIRTCHNL_PROTO_HDR_IPV4)
913 input->ip.v4.dst_port = cpu_to_be16(PFCP_PORT_NR);
914 else if (l3 == VIRTCHNL_PROTO_HDR_IPV6)
915 input->ip.v6.dst_port = cpu_to_be16(PFCP_PORT_NR);
916 }
917 break;
918 case VIRTCHNL_PROTO_HDR_GTPU_IP:
919 rawh = (u8 *)hdr->buffer;
920 input->flow_type = ICE_FLTR_PTYPE_NONF_IPV4_GTPU_IPV4_OTHER;
921
922 if (hdr->field_selector)
923 input->gtpu_data.teid = *(__be32 *)(&rawh[GTPU_TEID_OFFSET]);
924 conf->ttype = ICE_FDIR_TUNNEL_TYPE_GTPU;
925 break;
926 case VIRTCHNL_PROTO_HDR_GTPU_EH:
927 rawh = (u8 *)hdr->buffer;
928
929 if (hdr->field_selector)
930 input->gtpu_data.qfi = rawh[GTPU_EH_QFI_OFFSET] & GTPU_EH_QFI_MASK;
931 conf->ttype = ICE_FDIR_TUNNEL_TYPE_GTPU_EH;
932 break;
933 default:
934 dev_dbg(dev, "Invalid header type 0x:%x for VF %d\n",
935 hdr->type, vf->vf_id);
936 return -EINVAL;
937 }
938 }
939
940 return 0;
941}
942
943/**
944 * ice_vc_fdir_parse_action
945 * @vf: pointer to the VF info
946 * @fltr: virtual channel add cmd buffer
947 * @conf: FDIR configuration for each filter
948 *
949 * Parse the virtual channel filter's action and store them into conf
950 *
951 * Return: 0 on success, and other on error.
952 */
953static int
954ice_vc_fdir_parse_action(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
955 struct virtchnl_fdir_fltr_conf *conf)
956{
957 struct virtchnl_filter_action_set *as = &fltr->rule_cfg.action_set;
958 struct device *dev = ice_pf_to_dev(vf->pf);
959 struct ice_fdir_fltr *input = &conf->input;
960 u32 dest_num = 0;
961 u32 mark_num = 0;
962 int i;
963
964 if (as->count > VIRTCHNL_MAX_NUM_ACTIONS) {
965 dev_dbg(dev, "Invalid action numbers:0x%x for VF %d\n",
966 as->count, vf->vf_id);
967 return -EINVAL;
968 }
969
970 for (i = 0; i < as->count; i++) {
971 struct virtchnl_filter_action *action = &as->actions[i];
972
973 switch (action->type) {
974 case VIRTCHNL_ACTION_PASSTHRU:
975 dest_num++;
976 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_OTHER;
977 break;
978 case VIRTCHNL_ACTION_DROP:
979 dest_num++;
980 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DROP_PKT;
981 break;
982 case VIRTCHNL_ACTION_QUEUE:
983 dest_num++;
984 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_QINDEX;
985 input->q_index = action->act_conf.queue.index;
986 break;
987 case VIRTCHNL_ACTION_Q_REGION:
988 dest_num++;
989 input->dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_QGROUP;
990 input->q_index = action->act_conf.queue.index;
991 input->q_region = action->act_conf.queue.region;
992 break;
993 case VIRTCHNL_ACTION_MARK:
994 mark_num++;
995 input->fltr_id = action->act_conf.mark_id;
996 input->fdid_prio = ICE_FXD_FLTR_QW1_FDID_PRI_THREE;
997 break;
998 default:
999 dev_dbg(dev, "Invalid action type:0x%x for VF %d\n",
1000 action->type, vf->vf_id);
1001 return -EINVAL;
1002 }
1003 }
1004
1005 if (dest_num == 0 || dest_num >= 2) {
1006 dev_dbg(dev, "Invalid destination action for VF %d\n",
1007 vf->vf_id);
1008 return -EINVAL;
1009 }
1010
1011 if (mark_num >= 2) {
1012 dev_dbg(dev, "Too many mark actions for VF %d\n", vf->vf_id);
1013 return -EINVAL;
1014 }
1015
1016 return 0;
1017}
1018
1019/**
1020 * ice_vc_validate_fdir_fltr - validate the virtual channel filter
1021 * @vf: pointer to the VF info
1022 * @fltr: virtual channel add cmd buffer
1023 * @conf: FDIR configuration for each filter
1024 *
1025 * Return: 0 on success, and other on error.
1026 */
1027static int
1028ice_vc_validate_fdir_fltr(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
1029 struct virtchnl_fdir_fltr_conf *conf)
1030{
1031 struct virtchnl_proto_hdrs *proto = &fltr->rule_cfg.proto_hdrs;
1032 int ret;
1033
1034 if (!ice_vc_validate_pattern(vf, proto))
1035 return -EINVAL;
1036
1037 ret = ice_vc_fdir_parse_pattern(vf, fltr, conf);
1038 if (ret)
1039 return ret;
1040
1041 return ice_vc_fdir_parse_action(vf, fltr, conf);
1042}
1043
1044/**
1045 * ice_vc_fdir_comp_rules - compare if two filter rules have the same value
1046 * @conf_a: FDIR configuration for filter a
1047 * @conf_b: FDIR configuration for filter b
1048 *
1049 * Return: 0 on success, and other on error.
1050 */
1051static bool
1052ice_vc_fdir_comp_rules(struct virtchnl_fdir_fltr_conf *conf_a,
1053 struct virtchnl_fdir_fltr_conf *conf_b)
1054{
1055 struct ice_fdir_fltr *a = &conf_a->input;
1056 struct ice_fdir_fltr *b = &conf_b->input;
1057
1058 if (conf_a->ttype != conf_b->ttype)
1059 return false;
1060 if (a->flow_type != b->flow_type)
1061 return false;
1062 if (memcmp(&a->ip, &b->ip, sizeof(a->ip)))
1063 return false;
1064 if (memcmp(&a->mask, &b->mask, sizeof(a->mask)))
1065 return false;
1066 if (memcmp(&a->gtpu_data, &b->gtpu_data, sizeof(a->gtpu_data)))
1067 return false;
1068 if (memcmp(&a->gtpu_mask, &b->gtpu_mask, sizeof(a->gtpu_mask)))
1069 return false;
1070 if (memcmp(&a->l2tpv3_data, &b->l2tpv3_data, sizeof(a->l2tpv3_data)))
1071 return false;
1072 if (memcmp(&a->l2tpv3_mask, &b->l2tpv3_mask, sizeof(a->l2tpv3_mask)))
1073 return false;
1074 if (memcmp(&a->ext_data, &b->ext_data, sizeof(a->ext_data)))
1075 return false;
1076 if (memcmp(&a->ext_mask, &b->ext_mask, sizeof(a->ext_mask)))
1077 return false;
1078
1079 return true;
1080}
1081
1082/**
1083 * ice_vc_fdir_is_dup_fltr
1084 * @vf: pointer to the VF info
1085 * @conf: FDIR configuration for each filter
1086 *
1087 * Check if there is duplicated rule with same conf value
1088 *
1089 * Return: 0 true success, and false on error.
1090 */
1091static bool
1092ice_vc_fdir_is_dup_fltr(struct ice_vf *vf, struct virtchnl_fdir_fltr_conf *conf)
1093{
1094 struct ice_fdir_fltr *desc;
1095 bool ret;
1096
1097 list_for_each_entry(desc, &vf->fdir.fdir_rule_list, fltr_node) {
1098 struct virtchnl_fdir_fltr_conf *node =
1099 to_fltr_conf_from_desc(desc);
1100
1101 ret = ice_vc_fdir_comp_rules(node, conf);
1102 if (ret)
1103 return true;
1104 }
1105
1106 return false;
1107}
1108
1109/**
1110 * ice_vc_fdir_insert_entry
1111 * @vf: pointer to the VF info
1112 * @conf: FDIR configuration for each filter
1113 * @id: pointer to ID value allocated by driver
1114 *
1115 * Insert FDIR conf entry into list and allocate ID for this filter
1116 *
1117 * Return: 0 true success, and other on error.
1118 */
1119static int
1120ice_vc_fdir_insert_entry(struct ice_vf *vf,
1121 struct virtchnl_fdir_fltr_conf *conf, u32 *id)
1122{
1123 struct ice_fdir_fltr *input = &conf->input;
1124 int i;
1125
1126 /* alloc ID corresponding with conf */
1127 i = idr_alloc(&vf->fdir.fdir_rule_idr, conf, 0,
1128 ICE_FDIR_MAX_FLTRS, GFP_KERNEL);
1129 if (i < 0)
1130 return -EINVAL;
1131 *id = i;
1132
1133 list_add(&input->fltr_node, &vf->fdir.fdir_rule_list);
1134 return 0;
1135}
1136
1137/**
1138 * ice_vc_fdir_remove_entry - remove FDIR conf entry by ID value
1139 * @vf: pointer to the VF info
1140 * @conf: FDIR configuration for each filter
1141 * @id: filter rule's ID
1142 */
1143static void
1144ice_vc_fdir_remove_entry(struct ice_vf *vf,
1145 struct virtchnl_fdir_fltr_conf *conf, u32 id)
1146{
1147 struct ice_fdir_fltr *input = &conf->input;
1148
1149 idr_remove(&vf->fdir.fdir_rule_idr, id);
1150 list_del(&input->fltr_node);
1151}
1152
1153/**
1154 * ice_vc_fdir_lookup_entry - lookup FDIR conf entry by ID value
1155 * @vf: pointer to the VF info
1156 * @id: filter rule's ID
1157 *
1158 * Return: NULL on error, and other on success.
1159 */
1160static struct virtchnl_fdir_fltr_conf *
1161ice_vc_fdir_lookup_entry(struct ice_vf *vf, u32 id)
1162{
1163 return idr_find(&vf->fdir.fdir_rule_idr, id);
1164}
1165
1166/**
1167 * ice_vc_fdir_flush_entry - remove all FDIR conf entry
1168 * @vf: pointer to the VF info
1169 */
1170static void ice_vc_fdir_flush_entry(struct ice_vf *vf)
1171{
1172 struct virtchnl_fdir_fltr_conf *conf;
1173 struct ice_fdir_fltr *desc, *temp;
1174
1175 list_for_each_entry_safe(desc, temp,
1176 &vf->fdir.fdir_rule_list, fltr_node) {
1177 conf = to_fltr_conf_from_desc(desc);
1178 list_del(&desc->fltr_node);
1179 devm_kfree(ice_pf_to_dev(vf->pf), conf);
1180 }
1181}
1182
1183/**
1184 * ice_vc_fdir_write_fltr - write filter rule into hardware
1185 * @vf: pointer to the VF info
1186 * @conf: FDIR configuration for each filter
1187 * @add: true implies add rule, false implies del rules
1188 * @is_tun: false implies non-tunnel type filter, true implies tunnel filter
1189 *
1190 * Return: 0 on success, and other on error.
1191 */
1192static int ice_vc_fdir_write_fltr(struct ice_vf *vf,
1193 struct virtchnl_fdir_fltr_conf *conf,
1194 bool add, bool is_tun)
1195{
1196 struct ice_fdir_fltr *input = &conf->input;
1197 struct ice_vsi *vsi, *ctrl_vsi;
1198 struct ice_fltr_desc desc;
1199 struct device *dev;
1200 struct ice_pf *pf;
1201 struct ice_hw *hw;
1202 int ret;
1203 u8 *pkt;
1204
1205 pf = vf->pf;
1206 dev = ice_pf_to_dev(pf);
1207 hw = &pf->hw;
1208 vsi = pf->vsi[vf->lan_vsi_idx];
1209 if (!vsi) {
1210 dev_dbg(dev, "Invalid vsi for VF %d\n", vf->vf_id);
1211 return -EINVAL;
1212 }
1213
1214 input->dest_vsi = vsi->idx;
1215 input->comp_report = ICE_FXD_FLTR_QW0_COMP_REPORT_SW;
1216
1217 ctrl_vsi = pf->vsi[vf->ctrl_vsi_idx];
1218 if (!ctrl_vsi) {
1219 dev_dbg(dev, "Invalid ctrl_vsi for VF %d\n", vf->vf_id);
1220 return -EINVAL;
1221 }
1222
1223 pkt = devm_kzalloc(dev, ICE_FDIR_MAX_RAW_PKT_SIZE, GFP_KERNEL);
1224 if (!pkt)
1225 return -ENOMEM;
1226
1227 ice_fdir_get_prgm_desc(hw, input, &desc, add);
1228 ret = ice_fdir_get_gen_prgm_pkt(hw, input, pkt, false, is_tun);
1229 if (ret) {
1230 dev_dbg(dev, "Gen training pkt for VF %d ptype %d failed\n",
1231 vf->vf_id, input->flow_type);
1232 goto err_free_pkt;
1233 }
1234
1235 ret = ice_prgm_fdir_fltr(ctrl_vsi, &desc, pkt);
1236 if (ret)
1237 goto err_free_pkt;
1238
1239 return 0;
1240
1241err_free_pkt:
1242 devm_kfree(dev, pkt);
1243 return ret;
1244}
1245
1246/**
1247 * ice_vf_fdir_timer - FDIR program waiting timer interrupt handler
1248 * @t: pointer to timer_list
1249 */
1250static void ice_vf_fdir_timer(struct timer_list *t)
1251{
1252 struct ice_vf_fdir_ctx *ctx_irq = from_timer(ctx_irq, t, rx_tmr);
1253 struct ice_vf_fdir_ctx *ctx_done;
1254 struct ice_vf_fdir *fdir;
1255 unsigned long flags;
1256 struct ice_vf *vf;
1257 struct ice_pf *pf;
1258
1259 fdir = container_of(ctx_irq, struct ice_vf_fdir, ctx_irq);
1260 vf = container_of(fdir, struct ice_vf, fdir);
1261 ctx_done = &fdir->ctx_done;
1262 pf = vf->pf;
1263 spin_lock_irqsave(&fdir->ctx_lock, flags);
1264 if (!(ctx_irq->flags & ICE_VF_FDIR_CTX_VALID)) {
1265 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1266 WARN_ON_ONCE(1);
1267 return;
1268 }
1269
1270 ctx_irq->flags &= ~ICE_VF_FDIR_CTX_VALID;
1271
1272 ctx_done->flags |= ICE_VF_FDIR_CTX_VALID;
1273 ctx_done->conf = ctx_irq->conf;
1274 ctx_done->stat = ICE_FDIR_CTX_TIMEOUT;
1275 ctx_done->v_opcode = ctx_irq->v_opcode;
1276 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1277
1278 set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
1279 ice_service_task_schedule(pf);
1280}
1281
1282/**
1283 * ice_vc_fdir_irq_handler - ctrl_vsi Rx queue interrupt handler
1284 * @ctrl_vsi: pointer to a VF's CTRL VSI
1285 * @rx_desc: pointer to FDIR Rx queue descriptor
1286 */
1287void
1288ice_vc_fdir_irq_handler(struct ice_vsi *ctrl_vsi,
1289 union ice_32b_rx_flex_desc *rx_desc)
1290{
1291 struct ice_pf *pf = ctrl_vsi->back;
1292 struct ice_vf *vf = ctrl_vsi->vf;
1293 struct ice_vf_fdir_ctx *ctx_done;
1294 struct ice_vf_fdir_ctx *ctx_irq;
1295 struct ice_vf_fdir *fdir;
1296 unsigned long flags;
1297 struct device *dev;
1298 int ret;
1299
1300 if (WARN_ON(!vf))
1301 return;
1302
1303 fdir = &vf->fdir;
1304 ctx_done = &fdir->ctx_done;
1305 ctx_irq = &fdir->ctx_irq;
1306 dev = ice_pf_to_dev(pf);
1307 spin_lock_irqsave(&fdir->ctx_lock, flags);
1308 if (!(ctx_irq->flags & ICE_VF_FDIR_CTX_VALID)) {
1309 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1310 WARN_ON_ONCE(1);
1311 return;
1312 }
1313
1314 ctx_irq->flags &= ~ICE_VF_FDIR_CTX_VALID;
1315
1316 ctx_done->flags |= ICE_VF_FDIR_CTX_VALID;
1317 ctx_done->conf = ctx_irq->conf;
1318 ctx_done->stat = ICE_FDIR_CTX_IRQ;
1319 ctx_done->v_opcode = ctx_irq->v_opcode;
1320 memcpy(&ctx_done->rx_desc, rx_desc, sizeof(*rx_desc));
1321 spin_unlock_irqrestore(&fdir->ctx_lock, flags);
1322
1323 ret = del_timer(&ctx_irq->rx_tmr);
1324 if (!ret)
1325 dev_err(dev, "VF %d: Unexpected inactive timer!\n", vf->vf_id);
1326
1327 set_bit(ICE_FD_VF_FLUSH_CTX, pf->state);
1328 ice_service_task_schedule(pf);
1329}
1330
1331/**
1332 * ice_vf_fdir_dump_info - dump FDIR information for diagnosis
1333 * @vf: pointer to the VF info
1334 */
1335static void ice_vf_fdir_dump_info(struct ice_vf *vf)
1336{
1337 struct ice_vsi *vf_vsi;
1338 u32 fd_size, fd_cnt;
1339 struct device *dev;
1340 struct ice_pf *pf;
1341 struct ice_hw *hw;
1342 u16 vsi_num;
1343
1344 pf = vf->pf;
1345 hw = &pf->hw;
1346 dev = ice_pf_to_dev(pf);
1347 vf_vsi = ice_get_vf_vsi(vf);
1348 if (!vf_vsi) {
1349 dev_dbg(dev, "VF %d: invalid VSI pointer\n", vf->vf_id);
1350 return;
1351 }
1352
1353 vsi_num = ice_get_hw_vsi_num(hw, vf_vsi->idx);
1354
1355 fd_size = rd32(hw, VSIQF_FD_SIZE(vsi_num));
1356 fd_cnt = rd32(hw, VSIQF_FD_CNT(vsi_num));
1357 dev_dbg(dev, "VF %d: space allocated: guar:0x%x, be:0x%x, space consumed: guar:0x%x, be:0x%x\n",
1358 vf->vf_id,
1359 (fd_size & VSIQF_FD_CNT_FD_GCNT_M) >> VSIQF_FD_CNT_FD_GCNT_S,
1360 (fd_size & VSIQF_FD_CNT_FD_BCNT_M) >> VSIQF_FD_CNT_FD_BCNT_S,
1361 (fd_cnt & VSIQF_FD_CNT_FD_GCNT_M) >> VSIQF_FD_CNT_FD_GCNT_S,
1362 (fd_cnt & VSIQF_FD_CNT_FD_BCNT_M) >> VSIQF_FD_CNT_FD_BCNT_S);
1363}
1364
1365/**
1366 * ice_vf_verify_rx_desc - verify received FDIR programming status descriptor
1367 * @vf: pointer to the VF info
1368 * @ctx: FDIR context info for post processing
1369 * @status: virtchnl FDIR program status
1370 *
1371 * Return: 0 on success, and other on error.
1372 */
1373static int
1374ice_vf_verify_rx_desc(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1375 enum virtchnl_fdir_prgm_status *status)
1376{
1377 struct device *dev = ice_pf_to_dev(vf->pf);
1378 u32 stat_err, error, prog_id;
1379 int ret;
1380
1381 stat_err = le16_to_cpu(ctx->rx_desc.wb.status_error0);
1382 if (((stat_err & ICE_FXD_FLTR_WB_QW1_DD_M) >>
1383 ICE_FXD_FLTR_WB_QW1_DD_S) != ICE_FXD_FLTR_WB_QW1_DD_YES) {
1384 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1385 dev_err(dev, "VF %d: Desc Done not set\n", vf->vf_id);
1386 ret = -EINVAL;
1387 goto err_exit;
1388 }
1389
1390 prog_id = (stat_err & ICE_FXD_FLTR_WB_QW1_PROG_ID_M) >>
1391 ICE_FXD_FLTR_WB_QW1_PROG_ID_S;
1392 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_ADD &&
1393 ctx->v_opcode != VIRTCHNL_OP_ADD_FDIR_FILTER) {
1394 dev_err(dev, "VF %d: Desc show add, but ctx not",
1395 vf->vf_id);
1396 *status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
1397 ret = -EINVAL;
1398 goto err_exit;
1399 }
1400
1401 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_DEL &&
1402 ctx->v_opcode != VIRTCHNL_OP_DEL_FDIR_FILTER) {
1403 dev_err(dev, "VF %d: Desc show del, but ctx not",
1404 vf->vf_id);
1405 *status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
1406 ret = -EINVAL;
1407 goto err_exit;
1408 }
1409
1410 error = (stat_err & ICE_FXD_FLTR_WB_QW1_FAIL_M) >>
1411 ICE_FXD_FLTR_WB_QW1_FAIL_S;
1412 if (error == ICE_FXD_FLTR_WB_QW1_FAIL_YES) {
1413 if (prog_id == ICE_FXD_FLTR_WB_QW1_PROG_ADD) {
1414 dev_err(dev, "VF %d, Failed to add FDIR rule due to no space in the table",
1415 vf->vf_id);
1416 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1417 } else {
1418 dev_err(dev, "VF %d, Failed to remove FDIR rule, attempt to remove non-existent entry",
1419 vf->vf_id);
1420 *status = VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST;
1421 }
1422 ret = -EINVAL;
1423 goto err_exit;
1424 }
1425
1426 error = (stat_err & ICE_FXD_FLTR_WB_QW1_FAIL_PROF_M) >>
1427 ICE_FXD_FLTR_WB_QW1_FAIL_PROF_S;
1428 if (error == ICE_FXD_FLTR_WB_QW1_FAIL_PROF_YES) {
1429 dev_err(dev, "VF %d: Profile matching error", vf->vf_id);
1430 *status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1431 ret = -EINVAL;
1432 goto err_exit;
1433 }
1434
1435 *status = VIRTCHNL_FDIR_SUCCESS;
1436
1437 return 0;
1438
1439err_exit:
1440 ice_vf_fdir_dump_info(vf);
1441 return ret;
1442}
1443
1444/**
1445 * ice_vc_add_fdir_fltr_post
1446 * @vf: pointer to the VF structure
1447 * @ctx: FDIR context info for post processing
1448 * @status: virtchnl FDIR program status
1449 * @success: true implies success, false implies failure
1450 *
1451 * Post process for flow director add command. If success, then do post process
1452 * and send back success msg by virtchnl. Otherwise, do context reversion and
1453 * send back failure msg by virtchnl.
1454 *
1455 * Return: 0 on success, and other on error.
1456 */
1457static int
1458ice_vc_add_fdir_fltr_post(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1459 enum virtchnl_fdir_prgm_status status,
1460 bool success)
1461{
1462 struct virtchnl_fdir_fltr_conf *conf = ctx->conf;
1463 struct device *dev = ice_pf_to_dev(vf->pf);
1464 enum virtchnl_status_code v_ret;
1465 struct virtchnl_fdir_add *resp;
1466 int ret, len, is_tun;
1467
1468 v_ret = VIRTCHNL_STATUS_SUCCESS;
1469 len = sizeof(*resp);
1470 resp = kzalloc(len, GFP_KERNEL);
1471 if (!resp) {
1472 len = 0;
1473 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1474 dev_dbg(dev, "VF %d: Alloc resp buf fail", vf->vf_id);
1475 goto err_exit;
1476 }
1477
1478 if (!success)
1479 goto err_exit;
1480
1481 is_tun = 0;
1482 resp->status = status;
1483 resp->flow_id = conf->flow_id;
1484 vf->fdir.fdir_fltr_cnt[conf->input.flow_type][is_tun]++;
1485
1486 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1487 (u8 *)resp, len);
1488 kfree(resp);
1489
1490 dev_dbg(dev, "VF %d: flow_id:0x%X, FDIR %s success!\n",
1491 vf->vf_id, conf->flow_id,
1492 (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER) ?
1493 "add" : "del");
1494 return ret;
1495
1496err_exit:
1497 if (resp)
1498 resp->status = status;
1499 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
1500 devm_kfree(dev, conf);
1501
1502 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1503 (u8 *)resp, len);
1504 kfree(resp);
1505 return ret;
1506}
1507
1508/**
1509 * ice_vc_del_fdir_fltr_post
1510 * @vf: pointer to the VF structure
1511 * @ctx: FDIR context info for post processing
1512 * @status: virtchnl FDIR program status
1513 * @success: true implies success, false implies failure
1514 *
1515 * Post process for flow director del command. If success, then do post process
1516 * and send back success msg by virtchnl. Otherwise, do context reversion and
1517 * send back failure msg by virtchnl.
1518 *
1519 * Return: 0 on success, and other on error.
1520 */
1521static int
1522ice_vc_del_fdir_fltr_post(struct ice_vf *vf, struct ice_vf_fdir_ctx *ctx,
1523 enum virtchnl_fdir_prgm_status status,
1524 bool success)
1525{
1526 struct virtchnl_fdir_fltr_conf *conf = ctx->conf;
1527 struct device *dev = ice_pf_to_dev(vf->pf);
1528 enum virtchnl_status_code v_ret;
1529 struct virtchnl_fdir_del *resp;
1530 int ret, len, is_tun;
1531
1532 v_ret = VIRTCHNL_STATUS_SUCCESS;
1533 len = sizeof(*resp);
1534 resp = kzalloc(len, GFP_KERNEL);
1535 if (!resp) {
1536 len = 0;
1537 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1538 dev_dbg(dev, "VF %d: Alloc resp buf fail", vf->vf_id);
1539 goto err_exit;
1540 }
1541
1542 if (!success)
1543 goto err_exit;
1544
1545 is_tun = 0;
1546 resp->status = status;
1547 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
1548 vf->fdir.fdir_fltr_cnt[conf->input.flow_type][is_tun]--;
1549
1550 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1551 (u8 *)resp, len);
1552 kfree(resp);
1553
1554 dev_dbg(dev, "VF %d: flow_id:0x%X, FDIR %s success!\n",
1555 vf->vf_id, conf->flow_id,
1556 (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER) ?
1557 "add" : "del");
1558 devm_kfree(dev, conf);
1559 return ret;
1560
1561err_exit:
1562 if (resp)
1563 resp->status = status;
1564 if (success)
1565 devm_kfree(dev, conf);
1566
1567 ret = ice_vc_send_msg_to_vf(vf, ctx->v_opcode, v_ret,
1568 (u8 *)resp, len);
1569 kfree(resp);
1570 return ret;
1571}
1572
1573/**
1574 * ice_flush_fdir_ctx
1575 * @pf: pointer to the PF structure
1576 *
1577 * Flush all the pending event on ctx_done list and process them.
1578 */
1579void ice_flush_fdir_ctx(struct ice_pf *pf)
1580{
1581 struct ice_vf *vf;
1582 unsigned int bkt;
1583
1584 if (!test_and_clear_bit(ICE_FD_VF_FLUSH_CTX, pf->state))
1585 return;
1586
1587 mutex_lock(&pf->vfs.table_lock);
1588 ice_for_each_vf(pf, bkt, vf) {
1589 struct device *dev = ice_pf_to_dev(pf);
1590 enum virtchnl_fdir_prgm_status status;
1591 struct ice_vf_fdir_ctx *ctx;
1592 unsigned long flags;
1593 int ret;
1594
1595 if (!test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states))
1596 continue;
1597
1598 if (vf->ctrl_vsi_idx == ICE_NO_VSI)
1599 continue;
1600
1601 ctx = &vf->fdir.ctx_done;
1602 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1603 if (!(ctx->flags & ICE_VF_FDIR_CTX_VALID)) {
1604 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1605 continue;
1606 }
1607 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1608
1609 WARN_ON(ctx->stat == ICE_FDIR_CTX_READY);
1610 if (ctx->stat == ICE_FDIR_CTX_TIMEOUT) {
1611 status = VIRTCHNL_FDIR_FAILURE_RULE_TIMEOUT;
1612 dev_err(dev, "VF %d: ctrl_vsi irq timeout\n",
1613 vf->vf_id);
1614 goto err_exit;
1615 }
1616
1617 ret = ice_vf_verify_rx_desc(vf, ctx, &status);
1618 if (ret)
1619 goto err_exit;
1620
1621 if (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER)
1622 ice_vc_add_fdir_fltr_post(vf, ctx, status, true);
1623 else if (ctx->v_opcode == VIRTCHNL_OP_DEL_FDIR_FILTER)
1624 ice_vc_del_fdir_fltr_post(vf, ctx, status, true);
1625 else
1626 dev_err(dev, "VF %d: Unsupported opcode\n", vf->vf_id);
1627
1628 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1629 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1630 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1631 continue;
1632err_exit:
1633 if (ctx->v_opcode == VIRTCHNL_OP_ADD_FDIR_FILTER)
1634 ice_vc_add_fdir_fltr_post(vf, ctx, status, false);
1635 else if (ctx->v_opcode == VIRTCHNL_OP_DEL_FDIR_FILTER)
1636 ice_vc_del_fdir_fltr_post(vf, ctx, status, false);
1637 else
1638 dev_err(dev, "VF %d: Unsupported opcode\n", vf->vf_id);
1639
1640 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1641 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1642 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1643 }
1644 mutex_unlock(&pf->vfs.table_lock);
1645}
1646
1647/**
1648 * ice_vc_fdir_set_irq_ctx - set FDIR context info for later IRQ handler
1649 * @vf: pointer to the VF structure
1650 * @conf: FDIR configuration for each filter
1651 * @v_opcode: virtual channel operation code
1652 *
1653 * Return: 0 on success, and other on error.
1654 */
1655static int
1656ice_vc_fdir_set_irq_ctx(struct ice_vf *vf, struct virtchnl_fdir_fltr_conf *conf,
1657 enum virtchnl_ops v_opcode)
1658{
1659 struct device *dev = ice_pf_to_dev(vf->pf);
1660 struct ice_vf_fdir_ctx *ctx;
1661 unsigned long flags;
1662
1663 ctx = &vf->fdir.ctx_irq;
1664 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1665 if ((vf->fdir.ctx_irq.flags & ICE_VF_FDIR_CTX_VALID) ||
1666 (vf->fdir.ctx_done.flags & ICE_VF_FDIR_CTX_VALID)) {
1667 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1668 dev_dbg(dev, "VF %d: Last request is still in progress\n",
1669 vf->vf_id);
1670 return -EBUSY;
1671 }
1672 ctx->flags |= ICE_VF_FDIR_CTX_VALID;
1673 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1674
1675 ctx->conf = conf;
1676 ctx->v_opcode = v_opcode;
1677 ctx->stat = ICE_FDIR_CTX_READY;
1678 timer_setup(&ctx->rx_tmr, ice_vf_fdir_timer, 0);
1679
1680 mod_timer(&ctx->rx_tmr, round_jiffies(msecs_to_jiffies(10) + jiffies));
1681
1682 return 0;
1683}
1684
1685/**
1686 * ice_vc_fdir_clear_irq_ctx - clear FDIR context info for IRQ handler
1687 * @vf: pointer to the VF structure
1688 *
1689 * Return: 0 on success, and other on error.
1690 */
1691static void ice_vc_fdir_clear_irq_ctx(struct ice_vf *vf)
1692{
1693 struct ice_vf_fdir_ctx *ctx = &vf->fdir.ctx_irq;
1694 unsigned long flags;
1695
1696 del_timer(&ctx->rx_tmr);
1697 spin_lock_irqsave(&vf->fdir.ctx_lock, flags);
1698 ctx->flags &= ~ICE_VF_FDIR_CTX_VALID;
1699 spin_unlock_irqrestore(&vf->fdir.ctx_lock, flags);
1700}
1701
1702/**
1703 * ice_vc_add_fdir_fltr - add a FDIR filter for VF by the msg buffer
1704 * @vf: pointer to the VF info
1705 * @msg: pointer to the msg buffer
1706 *
1707 * Return: 0 on success, and other on error.
1708 */
1709int ice_vc_add_fdir_fltr(struct ice_vf *vf, u8 *msg)
1710{
1711 struct virtchnl_fdir_add *fltr = (struct virtchnl_fdir_add *)msg;
1712 struct virtchnl_fdir_add *stat = NULL;
1713 struct virtchnl_fdir_fltr_conf *conf;
1714 enum virtchnl_status_code v_ret;
1715 struct device *dev;
1716 struct ice_pf *pf;
1717 int is_tun = 0;
1718 int len = 0;
1719 int ret;
1720
1721 pf = vf->pf;
1722 dev = ice_pf_to_dev(pf);
1723 ret = ice_vc_fdir_param_check(vf, fltr->vsi_id);
1724 if (ret) {
1725 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
1726 dev_dbg(dev, "Parameter check for VF %d failed\n", vf->vf_id);
1727 goto err_exit;
1728 }
1729
1730 ret = ice_vf_start_ctrl_vsi(vf);
1731 if (ret && (ret != -EEXIST)) {
1732 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
1733 dev_err(dev, "Init FDIR for VF %d failed, ret:%d\n",
1734 vf->vf_id, ret);
1735 goto err_exit;
1736 }
1737
1738 stat = kzalloc(sizeof(*stat), GFP_KERNEL);
1739 if (!stat) {
1740 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1741 dev_dbg(dev, "Alloc stat for VF %d failed\n", vf->vf_id);
1742 goto err_exit;
1743 }
1744
1745 conf = devm_kzalloc(dev, sizeof(*conf), GFP_KERNEL);
1746 if (!conf) {
1747 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1748 dev_dbg(dev, "Alloc conf for VF %d failed\n", vf->vf_id);
1749 goto err_exit;
1750 }
1751
1752 len = sizeof(*stat);
1753 ret = ice_vc_validate_fdir_fltr(vf, fltr, conf);
1754 if (ret) {
1755 v_ret = VIRTCHNL_STATUS_SUCCESS;
1756 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_INVALID;
1757 dev_dbg(dev, "Invalid FDIR filter from VF %d\n", vf->vf_id);
1758 goto err_free_conf;
1759 }
1760
1761 if (fltr->validate_only) {
1762 v_ret = VIRTCHNL_STATUS_SUCCESS;
1763 stat->status = VIRTCHNL_FDIR_SUCCESS;
1764 devm_kfree(dev, conf);
1765 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_ADD_FDIR_FILTER,
1766 v_ret, (u8 *)stat, len);
1767 goto exit;
1768 }
1769
1770 ret = ice_vc_fdir_config_input_set(vf, fltr, conf, is_tun);
1771 if (ret) {
1772 v_ret = VIRTCHNL_STATUS_SUCCESS;
1773 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_CONFLICT;
1774 dev_err(dev, "VF %d: FDIR input set configure failed, ret:%d\n",
1775 vf->vf_id, ret);
1776 goto err_free_conf;
1777 }
1778
1779 ret = ice_vc_fdir_is_dup_fltr(vf, conf);
1780 if (ret) {
1781 v_ret = VIRTCHNL_STATUS_SUCCESS;
1782 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_EXIST;
1783 dev_dbg(dev, "VF %d: duplicated FDIR rule detected\n",
1784 vf->vf_id);
1785 goto err_free_conf;
1786 }
1787
1788 ret = ice_vc_fdir_insert_entry(vf, conf, &conf->flow_id);
1789 if (ret) {
1790 v_ret = VIRTCHNL_STATUS_SUCCESS;
1791 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1792 dev_dbg(dev, "VF %d: insert FDIR list failed\n", vf->vf_id);
1793 goto err_free_conf;
1794 }
1795
1796 ret = ice_vc_fdir_set_irq_ctx(vf, conf, VIRTCHNL_OP_ADD_FDIR_FILTER);
1797 if (ret) {
1798 v_ret = VIRTCHNL_STATUS_SUCCESS;
1799 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1800 dev_dbg(dev, "VF %d: set FDIR context failed\n", vf->vf_id);
1801 goto err_free_conf;
1802 }
1803
1804 ret = ice_vc_fdir_write_fltr(vf, conf, true, is_tun);
1805 if (ret) {
1806 v_ret = VIRTCHNL_STATUS_SUCCESS;
1807 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1808 dev_err(dev, "VF %d: writing FDIR rule failed, ret:%d\n",
1809 vf->vf_id, ret);
1810 goto err_rem_entry;
1811 }
1812
1813exit:
1814 kfree(stat);
1815 return ret;
1816
1817err_rem_entry:
1818 ice_vc_fdir_clear_irq_ctx(vf);
1819 ice_vc_fdir_remove_entry(vf, conf, conf->flow_id);
1820err_free_conf:
1821 devm_kfree(dev, conf);
1822err_exit:
1823 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_ADD_FDIR_FILTER, v_ret,
1824 (u8 *)stat, len);
1825 kfree(stat);
1826 return ret;
1827}
1828
1829/**
1830 * ice_vc_del_fdir_fltr - delete a FDIR filter for VF by the msg buffer
1831 * @vf: pointer to the VF info
1832 * @msg: pointer to the msg buffer
1833 *
1834 * Return: 0 on success, and other on error.
1835 */
1836int ice_vc_del_fdir_fltr(struct ice_vf *vf, u8 *msg)
1837{
1838 struct virtchnl_fdir_del *fltr = (struct virtchnl_fdir_del *)msg;
1839 struct virtchnl_fdir_del *stat = NULL;
1840 struct virtchnl_fdir_fltr_conf *conf;
1841 enum virtchnl_status_code v_ret;
1842 struct device *dev;
1843 struct ice_pf *pf;
1844 int is_tun = 0;
1845 int len = 0;
1846 int ret;
1847
1848 pf = vf->pf;
1849 dev = ice_pf_to_dev(pf);
1850 ret = ice_vc_fdir_param_check(vf, fltr->vsi_id);
1851 if (ret) {
1852 v_ret = VIRTCHNL_STATUS_ERR_PARAM;
1853 dev_dbg(dev, "Parameter check for VF %d failed\n", vf->vf_id);
1854 goto err_exit;
1855 }
1856
1857 stat = kzalloc(sizeof(*stat), GFP_KERNEL);
1858 if (!stat) {
1859 v_ret = VIRTCHNL_STATUS_ERR_NO_MEMORY;
1860 dev_dbg(dev, "Alloc stat for VF %d failed\n", vf->vf_id);
1861 goto err_exit;
1862 }
1863
1864 len = sizeof(*stat);
1865
1866 conf = ice_vc_fdir_lookup_entry(vf, fltr->flow_id);
1867 if (!conf) {
1868 v_ret = VIRTCHNL_STATUS_SUCCESS;
1869 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NONEXIST;
1870 dev_dbg(dev, "VF %d: FDIR invalid flow_id:0x%X\n",
1871 vf->vf_id, fltr->flow_id);
1872 goto err_exit;
1873 }
1874
1875 /* Just return failure when ctrl_vsi idx is invalid */
1876 if (vf->ctrl_vsi_idx == ICE_NO_VSI) {
1877 v_ret = VIRTCHNL_STATUS_SUCCESS;
1878 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1879 dev_err(dev, "Invalid FDIR ctrl_vsi for VF %d\n", vf->vf_id);
1880 goto err_exit;
1881 }
1882
1883 ret = ice_vc_fdir_set_irq_ctx(vf, conf, VIRTCHNL_OP_DEL_FDIR_FILTER);
1884 if (ret) {
1885 v_ret = VIRTCHNL_STATUS_SUCCESS;
1886 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1887 dev_dbg(dev, "VF %d: set FDIR context failed\n", vf->vf_id);
1888 goto err_exit;
1889 }
1890
1891 ret = ice_vc_fdir_write_fltr(vf, conf, false, is_tun);
1892 if (ret) {
1893 v_ret = VIRTCHNL_STATUS_SUCCESS;
1894 stat->status = VIRTCHNL_FDIR_FAILURE_RULE_NORESOURCE;
1895 dev_err(dev, "VF %d: writing FDIR rule failed, ret:%d\n",
1896 vf->vf_id, ret);
1897 goto err_del_tmr;
1898 }
1899
1900 kfree(stat);
1901
1902 return ret;
1903
1904err_del_tmr:
1905 ice_vc_fdir_clear_irq_ctx(vf);
1906err_exit:
1907 ret = ice_vc_send_msg_to_vf(vf, VIRTCHNL_OP_DEL_FDIR_FILTER, v_ret,
1908 (u8 *)stat, len);
1909 kfree(stat);
1910 return ret;
1911}
1912
1913/**
1914 * ice_vf_fdir_init - init FDIR resource for VF
1915 * @vf: pointer to the VF info
1916 */
1917void ice_vf_fdir_init(struct ice_vf *vf)
1918{
1919 struct ice_vf_fdir *fdir = &vf->fdir;
1920
1921 idr_init(&fdir->fdir_rule_idr);
1922 INIT_LIST_HEAD(&fdir->fdir_rule_list);
1923
1924 spin_lock_init(&fdir->ctx_lock);
1925 fdir->ctx_irq.flags = 0;
1926 fdir->ctx_done.flags = 0;
1927}
1928
1929/**
1930 * ice_vf_fdir_exit - destroy FDIR resource for VF
1931 * @vf: pointer to the VF info
1932 */
1933void ice_vf_fdir_exit(struct ice_vf *vf)
1934{
1935 ice_vc_fdir_flush_entry(vf);
1936 idr_destroy(&vf->fdir.fdir_rule_idr);
1937 ice_vc_fdir_rem_prof_all(vf);
1938 ice_vc_fdir_free_prof_all(vf);
1939}