Loading...
1/* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */
2/* QLogic qed NIC Driver
3 * Copyright (c) 2015-2017 QLogic Corporation
4 * Copyright (c) 2019-2020 Marvell International Ltd.
5 */
6
7#ifndef _QED_SRIOV_H
8#define _QED_SRIOV_H
9#include <linux/types.h>
10#include "qed_vf.h"
11
12#define QED_ETH_VF_NUM_MAC_FILTERS 1
13#define QED_ETH_VF_NUM_VLAN_FILTERS 2
14#define QED_VF_ARRAY_LENGTH (3)
15
16#ifdef CONFIG_QED_SRIOV
17#define IS_VF(cdev) ((cdev)->b_is_vf)
18#define IS_PF(cdev) (!((cdev)->b_is_vf))
19#define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->cdev->p_iov_info))
20#else
21#define IS_VF(cdev) (0)
22#define IS_PF(cdev) (1)
23#define IS_PF_SRIOV(p_hwfn) (0)
24#endif
25#define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info))
26
27#define QED_MAX_VF_CHAINS_PER_PF 16
28
29#define QED_ETH_MAX_VF_NUM_VLAN_FILTERS \
30 (MAX_NUM_VFS * QED_ETH_VF_NUM_VLAN_FILTERS)
31
32enum qed_iov_vport_update_flag {
33 QED_IOV_VP_UPDATE_ACTIVATE,
34 QED_IOV_VP_UPDATE_VLAN_STRIP,
35 QED_IOV_VP_UPDATE_TX_SWITCH,
36 QED_IOV_VP_UPDATE_MCAST,
37 QED_IOV_VP_UPDATE_ACCEPT_PARAM,
38 QED_IOV_VP_UPDATE_RSS,
39 QED_IOV_VP_UPDATE_ACCEPT_ANY_VLAN,
40 QED_IOV_VP_UPDATE_SGE_TPA,
41 QED_IOV_VP_UPDATE_MAX,
42};
43
44struct qed_public_vf_info {
45 /* These copies will later be reflected in the bulletin board,
46 * but this copy should be newer.
47 */
48 u8 forced_mac[ETH_ALEN];
49 u16 forced_vlan;
50 u8 mac[ETH_ALEN];
51
52 /* IFLA_VF_LINK_STATE_<X> */
53 int link_state;
54
55 /* Currently configured Tx rate in MB/sec. 0 if unconfigured */
56 int tx_rate;
57
58 /* Trusted VFs can configure promiscuous mode.
59 * Also store shadow promisc configuration if needed.
60 */
61 bool is_trusted_configured;
62 bool is_trusted_request;
63 u8 rx_accept_mode;
64 u8 tx_accept_mode;
65 bool accept_any_vlan;
66};
67
68struct qed_iov_vf_init_params {
69 u16 rel_vf_id;
70
71 /* Number of requested Queues; Currently, don't support different
72 * number of Rx/Tx queues.
73 */
74
75 u16 num_queues;
76
77 /* Allow the client to choose which qzones to use for Rx/Tx,
78 * and which queue_base to use for Tx queues on a per-queue basis.
79 * Notice values should be relative to the PF resources.
80 */
81 u16 req_rx_queue[QED_MAX_VF_CHAINS_PER_PF];
82 u16 req_tx_queue[QED_MAX_VF_CHAINS_PER_PF];
83};
84
85/* This struct is part of qed_dev and contains data relevant to all hwfns;
86 * Initialized only if SR-IOV cpabability is exposed in PCIe config space.
87 */
88struct qed_hw_sriov_info {
89 int pos; /* capability position */
90 int nres; /* number of resources */
91 u32 cap; /* SR-IOV Capabilities */
92 u16 ctrl; /* SR-IOV Control */
93 u16 total_vfs; /* total VFs associated with the PF */
94 u16 num_vfs; /* number of vfs that have been started */
95 u16 initial_vfs; /* initial VFs associated with the PF */
96 u16 nr_virtfn; /* number of VFs available */
97 u16 offset; /* first VF Routing ID offset */
98 u16 stride; /* following VF stride */
99 u16 vf_device_id; /* VF device id */
100 u32 pgsz; /* page size for BAR alignment */
101 u8 link; /* Function Dependency Link */
102
103 u32 first_vf_in_pf;
104};
105
106/* This mailbox is maintained per VF in its PF contains all information
107 * required for sending / receiving a message.
108 */
109struct qed_iov_vf_mbx {
110 union vfpf_tlvs *req_virt;
111 dma_addr_t req_phys;
112 union pfvf_tlvs *reply_virt;
113 dma_addr_t reply_phys;
114
115 /* Address in VF where a pending message is located */
116 dma_addr_t pending_req;
117
118 /* Message from VF awaits handling */
119 bool b_pending_msg;
120
121 u8 *offset;
122
123 /* saved VF request header */
124 struct vfpf_first_tlv first_tlv;
125};
126
127#define QED_IOV_LEGACY_QID_RX (0)
128#define QED_IOV_LEGACY_QID_TX (1)
129#define QED_IOV_QID_INVALID (0xFE)
130
131struct qed_vf_queue_cid {
132 bool b_is_tx;
133 struct qed_queue_cid *p_cid;
134};
135
136/* Describes a qzone associated with the VF */
137struct qed_vf_queue {
138 u16 fw_rx_qid;
139 u16 fw_tx_qid;
140
141 struct qed_vf_queue_cid cids[MAX_QUEUES_PER_QZONE];
142};
143
144enum vf_state {
145 VF_FREE = 0, /* VF ready to be acquired holds no resc */
146 VF_ACQUIRED, /* VF, acquired, but not initialized */
147 VF_ENABLED, /* VF, Enabled */
148 VF_RESET, /* VF, FLR'd, pending cleanup */
149 VF_STOPPED /* VF, Stopped */
150};
151
152struct qed_vf_vlan_shadow {
153 bool used;
154 u16 vid;
155};
156
157struct qed_vf_shadow_config {
158 /* Shadow copy of all guest vlans */
159 struct qed_vf_vlan_shadow vlans[QED_ETH_VF_NUM_VLAN_FILTERS + 1];
160
161 /* Shadow copy of all configured MACs; Empty if forcing MACs */
162 u8 macs[QED_ETH_VF_NUM_MAC_FILTERS][ETH_ALEN];
163 u8 inner_vlan_removal;
164};
165
166/* PFs maintain an array of this structure, per VF */
167struct qed_vf_info {
168 struct qed_iov_vf_mbx vf_mbx;
169 enum vf_state state;
170 bool b_init;
171 bool b_malicious;
172 u8 to_disable;
173
174 struct qed_bulletin bulletin;
175 dma_addr_t vf_bulletin;
176
177 /* PF saves a copy of the last VF acquire message */
178 struct vfpf_acquire_tlv acquire;
179
180 u32 concrete_fid;
181 u16 opaque_fid;
182 u16 mtu;
183
184 u8 vport_id;
185 u8 relative_vf_id;
186 u8 abs_vf_id;
187#define QED_VF_ABS_ID(p_hwfn, p_vf) (QED_PATH_ID(p_hwfn) ? \
188 (p_vf)->abs_vf_id + MAX_NUM_VFS_BB : \
189 (p_vf)->abs_vf_id)
190
191 u8 vport_instance;
192 u8 num_rxqs;
193 u8 num_txqs;
194
195 u16 rx_coal;
196 u16 tx_coal;
197
198 u8 num_sbs;
199
200 u8 num_mac_filters;
201 u8 num_vlan_filters;
202
203 struct qed_vf_queue vf_queues[QED_MAX_VF_CHAINS_PER_PF];
204 u16 igu_sbs[QED_MAX_VF_CHAINS_PER_PF];
205 u8 num_active_rxqs;
206 struct qed_public_vf_info p_vf_info;
207 bool spoof_chk;
208 bool req_spoofchk_val;
209
210 /* Stores the configuration requested by VF */
211 struct qed_vf_shadow_config shadow_config;
212
213 /* A bitfield using bulletin's valid-map bits, used to indicate
214 * which of the bulletin board features have been configured.
215 */
216 u64 configured_features;
217#define QED_IOV_CONFIGURED_FEATURES_MASK ((1 << MAC_ADDR_FORCED) | \
218 (1 << VLAN_ADDR_FORCED))
219};
220
221/* This structure is part of qed_hwfn and used only for PFs that have sriov
222 * capability enabled.
223 */
224struct qed_pf_iov {
225 struct qed_vf_info vfs_array[MAX_NUM_VFS];
226 u64 pending_flr[QED_VF_ARRAY_LENGTH];
227
228 /* Allocate message address continuosuly and split to each VF */
229 void *mbx_msg_virt_addr;
230 dma_addr_t mbx_msg_phys_addr;
231 u32 mbx_msg_size;
232 void *mbx_reply_virt_addr;
233 dma_addr_t mbx_reply_phys_addr;
234 u32 mbx_reply_size;
235 void *p_bulletins;
236 dma_addr_t bulletins_phys;
237 u32 bulletins_size;
238};
239
240enum qed_iov_wq_flag {
241 QED_IOV_WQ_MSG_FLAG,
242 QED_IOV_WQ_SET_UNICAST_FILTER_FLAG,
243 QED_IOV_WQ_BULLETIN_UPDATE_FLAG,
244 QED_IOV_WQ_STOP_WQ_FLAG,
245 QED_IOV_WQ_FLR_FLAG,
246 QED_IOV_WQ_TRUST_FLAG,
247 QED_IOV_WQ_VF_FORCE_LINK_QUERY_FLAG,
248};
249
250extern const struct qed_iov_hv_ops qed_iov_ops_pass;
251
252#ifdef CONFIG_QED_SRIOV
253/**
254 * qed_iov_is_valid_vfid(): Check if given VF ID @vfid is valid
255 * w.r.t. @b_enabled_only value
256 * if b_enabled_only = true - only enabled
257 * VF id is valid.
258 * else any VF id less than max_vfs is valid.
259 *
260 * @p_hwfn: HW device data.
261 * @rel_vf_id: Relative VF ID.
262 * @b_enabled_only: consider only enabled VF.
263 * @b_non_malicious: true iff we want to validate vf isn't malicious.
264 *
265 * Return: bool - true for valid VF ID
266 */
267bool qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
268 int rel_vf_id,
269 bool b_enabled_only, bool b_non_malicious);
270
271/**
272 * qed_iov_get_next_active_vf(): Given a VF index, return index of
273 * next [including that] active VF.
274 *
275 * @p_hwfn: HW device data.
276 * @rel_vf_id: VF ID.
277 *
278 * Return: MAX_NUM_VFS in case no further active VFs, otherwise index.
279 */
280u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn, u16 rel_vf_id);
281
282void qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn,
283 int vfid, u16 vxlan_port, u16 geneve_port);
284
285/**
286 * qed_iov_hw_info(): Read sriov related information and allocated resources
287 * reads from configuration space, shmem, etc.
288 *
289 * @p_hwfn: HW device data.
290 *
291 * Return: Int.
292 */
293int qed_iov_hw_info(struct qed_hwfn *p_hwfn);
294
295/**
296 * qed_add_tlv(): place a given tlv on the tlv buffer at next offset
297 *
298 * @p_hwfn: HW device data.
299 * @offset: offset.
300 * @type: Type
301 * @length: Length.
302 *
303 * Return: pointer to the newly placed tlv
304 */
305void *qed_add_tlv(struct qed_hwfn *p_hwfn, u8 **offset, u16 type, u16 length);
306
307/**
308 * qed_dp_tlv_list(): list the types and lengths of the tlvs on the buffer
309 *
310 * @p_hwfn: HW device data.
311 * @tlvs_list: Tlvs_list.
312 *
313 * Return: Void.
314 */
315void qed_dp_tlv_list(struct qed_hwfn *p_hwfn, void *tlvs_list);
316
317/**
318 * qed_sriov_vfpf_malicious(): Handle malicious VF/PF.
319 *
320 * @p_hwfn: HW device data.
321 * @p_data: Pointer to data.
322 *
323 * Return: Void.
324 */
325void qed_sriov_vfpf_malicious(struct qed_hwfn *p_hwfn,
326 struct fw_err_data *p_data);
327
328/**
329 * qed_sriov_eqe_event(): Callback for SRIOV events.
330 *
331 * @p_hwfn: HW device data.
332 * @opcode: Opcode.
333 * @echo: Echo.
334 * @data: data
335 * @fw_return_code: FW return code.
336 *
337 * Return: Int.
338 */
339int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn, u8 opcode, __le16 echo,
340 union event_ring_data *data, u8 fw_return_code);
341
342/**
343 * qed_iov_alloc(): allocate sriov related resources
344 *
345 * @p_hwfn: HW device data.
346 *
347 * Return: Int.
348 */
349int qed_iov_alloc(struct qed_hwfn *p_hwfn);
350
351/**
352 * qed_iov_setup(): setup sriov related resources
353 *
354 * @p_hwfn: HW device data.
355 *
356 * Return: Void.
357 */
358void qed_iov_setup(struct qed_hwfn *p_hwfn);
359
360/**
361 * qed_iov_free(): free sriov related resources
362 *
363 * @p_hwfn: HW device data.
364 *
365 * Return: Void.
366 */
367void qed_iov_free(struct qed_hwfn *p_hwfn);
368
369/**
370 * qed_iov_free_hw_info(): free sriov related memory that was
371 * allocated during hw_prepare
372 *
373 * @cdev: Qed dev pointer.
374 *
375 * Return: Void.
376 */
377void qed_iov_free_hw_info(struct qed_dev *cdev);
378
379/**
380 * qed_iov_mark_vf_flr(): Mark structs of vfs that have been FLR-ed.
381 *
382 * @p_hwfn: HW device data.
383 * @disabled_vfs: bitmask of all VFs on path that were FLRed
384 *
385 * Return: true iff one of the PF's vfs got FLRed. false otherwise.
386 */
387bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn, u32 *disabled_vfs);
388
389/**
390 * qed_iov_search_list_tlvs(): Search extended TLVs in request/reply buffer.
391 *
392 * @p_hwfn: HW device data.
393 * @p_tlvs_list: Pointer to tlvs list
394 * @req_type: Type of TLV
395 *
396 * Return: pointer to tlv type if found, otherwise returns NULL.
397 */
398void *qed_iov_search_list_tlvs(struct qed_hwfn *p_hwfn,
399 void *p_tlvs_list, u16 req_type);
400
401void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first);
402int qed_iov_wq_start(struct qed_dev *cdev);
403
404void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag);
405void qed_vf_start_iov_wq(struct qed_dev *cdev);
406int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled);
407void qed_inform_vf_link_state(struct qed_hwfn *hwfn);
408#else
409static inline bool
410qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
411 int rel_vf_id, bool b_enabled_only, bool b_non_malicious)
412{
413 return false;
414}
415
416static inline u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn,
417 u16 rel_vf_id)
418{
419 return MAX_NUM_VFS;
420}
421
422static inline void
423qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn, int vfid,
424 u16 vxlan_port, u16 geneve_port)
425{
426}
427
428static inline int qed_iov_hw_info(struct qed_hwfn *p_hwfn)
429{
430 return 0;
431}
432
433static inline int qed_iov_alloc(struct qed_hwfn *p_hwfn)
434{
435 return 0;
436}
437
438static inline void qed_iov_setup(struct qed_hwfn *p_hwfn)
439{
440}
441
442static inline void qed_iov_free(struct qed_hwfn *p_hwfn)
443{
444}
445
446static inline void qed_iov_free_hw_info(struct qed_dev *cdev)
447{
448}
449
450static inline bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn,
451 u32 *disabled_vfs)
452{
453 return false;
454}
455
456static inline void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first)
457{
458}
459
460static inline int qed_iov_wq_start(struct qed_dev *cdev)
461{
462 return 0;
463}
464
465static inline void qed_schedule_iov(struct qed_hwfn *hwfn,
466 enum qed_iov_wq_flag flag)
467{
468}
469
470static inline void qed_vf_start_iov_wq(struct qed_dev *cdev)
471{
472}
473
474static inline int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled)
475{
476 return 0;
477}
478
479static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn)
480{
481}
482
483static inline void qed_sriov_vfpf_malicious(struct qed_hwfn *p_hwfn,
484 struct fw_err_data *p_data)
485{
486}
487
488static inline int qed_sriov_eqe_event(struct qed_hwfn *p_hwfn, u8 opcode,
489 __le16 echo, union event_ring_data *data,
490 u8 fw_return_code)
491{
492 return 0;
493}
494#endif
495
496#define qed_for_each_vf(_p_hwfn, _i) \
497 for (_i = qed_iov_get_next_active_vf(_p_hwfn, 0); \
498 _i < MAX_NUM_VFS; \
499 _i = qed_iov_get_next_active_vf(_p_hwfn, _i + 1))
500
501#endif
1/* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */
2/* QLogic qed NIC Driver
3 * Copyright (c) 2015-2017 QLogic Corporation
4 * Copyright (c) 2019-2020 Marvell International Ltd.
5 */
6
7#ifndef _QED_SRIOV_H
8#define _QED_SRIOV_H
9#include <linux/types.h>
10#include "qed_vf.h"
11
12#define QED_ETH_VF_NUM_MAC_FILTERS 1
13#define QED_ETH_VF_NUM_VLAN_FILTERS 2
14#define QED_VF_ARRAY_LENGTH (3)
15
16#ifdef CONFIG_QED_SRIOV
17#define IS_VF(cdev) ((cdev)->b_is_vf)
18#define IS_PF(cdev) (!((cdev)->b_is_vf))
19#define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->cdev->p_iov_info))
20#else
21#define IS_VF(cdev) (0)
22#define IS_PF(cdev) (1)
23#define IS_PF_SRIOV(p_hwfn) (0)
24#endif
25#define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info))
26
27#define QED_MAX_VF_CHAINS_PER_PF 16
28
29#define QED_ETH_MAX_VF_NUM_VLAN_FILTERS \
30 (MAX_NUM_VFS * QED_ETH_VF_NUM_VLAN_FILTERS)
31
32enum qed_iov_vport_update_flag {
33 QED_IOV_VP_UPDATE_ACTIVATE,
34 QED_IOV_VP_UPDATE_VLAN_STRIP,
35 QED_IOV_VP_UPDATE_TX_SWITCH,
36 QED_IOV_VP_UPDATE_MCAST,
37 QED_IOV_VP_UPDATE_ACCEPT_PARAM,
38 QED_IOV_VP_UPDATE_RSS,
39 QED_IOV_VP_UPDATE_ACCEPT_ANY_VLAN,
40 QED_IOV_VP_UPDATE_SGE_TPA,
41 QED_IOV_VP_UPDATE_MAX,
42};
43
44struct qed_public_vf_info {
45 /* These copies will later be reflected in the bulletin board,
46 * but this copy should be newer.
47 */
48 u8 forced_mac[ETH_ALEN];
49 u16 forced_vlan;
50 u8 mac[ETH_ALEN];
51
52 /* IFLA_VF_LINK_STATE_<X> */
53 int link_state;
54
55 /* Currently configured Tx rate in MB/sec. 0 if unconfigured */
56 int tx_rate;
57
58 /* Trusted VFs can configure promiscuous mode.
59 * Also store shadow promisc configuration if needed.
60 */
61 bool is_trusted_configured;
62 bool is_trusted_request;
63 u8 rx_accept_mode;
64 u8 tx_accept_mode;
65};
66
67struct qed_iov_vf_init_params {
68 u16 rel_vf_id;
69
70 /* Number of requested Queues; Currently, don't support different
71 * number of Rx/Tx queues.
72 */
73
74 u16 num_queues;
75
76 /* Allow the client to choose which qzones to use for Rx/Tx,
77 * and which queue_base to use for Tx queues on a per-queue basis.
78 * Notice values should be relative to the PF resources.
79 */
80 u16 req_rx_queue[QED_MAX_VF_CHAINS_PER_PF];
81 u16 req_tx_queue[QED_MAX_VF_CHAINS_PER_PF];
82};
83
84/* This struct is part of qed_dev and contains data relevant to all hwfns;
85 * Initialized only if SR-IOV cpabability is exposed in PCIe config space.
86 */
87struct qed_hw_sriov_info {
88 int pos; /* capability position */
89 int nres; /* number of resources */
90 u32 cap; /* SR-IOV Capabilities */
91 u16 ctrl; /* SR-IOV Control */
92 u16 total_vfs; /* total VFs associated with the PF */
93 u16 num_vfs; /* number of vfs that have been started */
94 u16 initial_vfs; /* initial VFs associated with the PF */
95 u16 nr_virtfn; /* number of VFs available */
96 u16 offset; /* first VF Routing ID offset */
97 u16 stride; /* following VF stride */
98 u16 vf_device_id; /* VF device id */
99 u32 pgsz; /* page size for BAR alignment */
100 u8 link; /* Function Dependency Link */
101
102 u32 first_vf_in_pf;
103};
104
105/* This mailbox is maintained per VF in its PF contains all information
106 * required for sending / receiving a message.
107 */
108struct qed_iov_vf_mbx {
109 union vfpf_tlvs *req_virt;
110 dma_addr_t req_phys;
111 union pfvf_tlvs *reply_virt;
112 dma_addr_t reply_phys;
113
114 /* Address in VF where a pending message is located */
115 dma_addr_t pending_req;
116
117 /* Message from VF awaits handling */
118 bool b_pending_msg;
119
120 u8 *offset;
121
122 /* saved VF request header */
123 struct vfpf_first_tlv first_tlv;
124};
125
126#define QED_IOV_LEGACY_QID_RX (0)
127#define QED_IOV_LEGACY_QID_TX (1)
128#define QED_IOV_QID_INVALID (0xFE)
129
130struct qed_vf_queue_cid {
131 bool b_is_tx;
132 struct qed_queue_cid *p_cid;
133};
134
135/* Describes a qzone associated with the VF */
136struct qed_vf_queue {
137 u16 fw_rx_qid;
138 u16 fw_tx_qid;
139
140 struct qed_vf_queue_cid cids[MAX_QUEUES_PER_QZONE];
141};
142
143enum vf_state {
144 VF_FREE = 0, /* VF ready to be acquired holds no resc */
145 VF_ACQUIRED, /* VF, acquired, but not initalized */
146 VF_ENABLED, /* VF, Enabled */
147 VF_RESET, /* VF, FLR'd, pending cleanup */
148 VF_STOPPED /* VF, Stopped */
149};
150
151struct qed_vf_vlan_shadow {
152 bool used;
153 u16 vid;
154};
155
156struct qed_vf_shadow_config {
157 /* Shadow copy of all guest vlans */
158 struct qed_vf_vlan_shadow vlans[QED_ETH_VF_NUM_VLAN_FILTERS + 1];
159
160 /* Shadow copy of all configured MACs; Empty if forcing MACs */
161 u8 macs[QED_ETH_VF_NUM_MAC_FILTERS][ETH_ALEN];
162 u8 inner_vlan_removal;
163};
164
165/* PFs maintain an array of this structure, per VF */
166struct qed_vf_info {
167 struct qed_iov_vf_mbx vf_mbx;
168 enum vf_state state;
169 bool b_init;
170 bool b_malicious;
171 u8 to_disable;
172
173 struct qed_bulletin bulletin;
174 dma_addr_t vf_bulletin;
175
176 /* PF saves a copy of the last VF acquire message */
177 struct vfpf_acquire_tlv acquire;
178
179 u32 concrete_fid;
180 u16 opaque_fid;
181 u16 mtu;
182
183 u8 vport_id;
184 u8 relative_vf_id;
185 u8 abs_vf_id;
186#define QED_VF_ABS_ID(p_hwfn, p_vf) (QED_PATH_ID(p_hwfn) ? \
187 (p_vf)->abs_vf_id + MAX_NUM_VFS_BB : \
188 (p_vf)->abs_vf_id)
189
190 u8 vport_instance;
191 u8 num_rxqs;
192 u8 num_txqs;
193
194 u16 rx_coal;
195 u16 tx_coal;
196
197 u8 num_sbs;
198
199 u8 num_mac_filters;
200 u8 num_vlan_filters;
201
202 struct qed_vf_queue vf_queues[QED_MAX_VF_CHAINS_PER_PF];
203 u16 igu_sbs[QED_MAX_VF_CHAINS_PER_PF];
204 u8 num_active_rxqs;
205 struct qed_public_vf_info p_vf_info;
206 bool spoof_chk;
207 bool req_spoofchk_val;
208
209 /* Stores the configuration requested by VF */
210 struct qed_vf_shadow_config shadow_config;
211
212 /* A bitfield using bulletin's valid-map bits, used to indicate
213 * which of the bulletin board features have been configured.
214 */
215 u64 configured_features;
216#define QED_IOV_CONFIGURED_FEATURES_MASK ((1 << MAC_ADDR_FORCED) | \
217 (1 << VLAN_ADDR_FORCED))
218};
219
220/* This structure is part of qed_hwfn and used only for PFs that have sriov
221 * capability enabled.
222 */
223struct qed_pf_iov {
224 struct qed_vf_info vfs_array[MAX_NUM_VFS];
225 u64 pending_flr[QED_VF_ARRAY_LENGTH];
226
227 /* Allocate message address continuosuly and split to each VF */
228 void *mbx_msg_virt_addr;
229 dma_addr_t mbx_msg_phys_addr;
230 u32 mbx_msg_size;
231 void *mbx_reply_virt_addr;
232 dma_addr_t mbx_reply_phys_addr;
233 u32 mbx_reply_size;
234 void *p_bulletins;
235 dma_addr_t bulletins_phys;
236 u32 bulletins_size;
237};
238
239enum qed_iov_wq_flag {
240 QED_IOV_WQ_MSG_FLAG,
241 QED_IOV_WQ_SET_UNICAST_FILTER_FLAG,
242 QED_IOV_WQ_BULLETIN_UPDATE_FLAG,
243 QED_IOV_WQ_STOP_WQ_FLAG,
244 QED_IOV_WQ_FLR_FLAG,
245 QED_IOV_WQ_TRUST_FLAG,
246 QED_IOV_WQ_VF_FORCE_LINK_QUERY_FLAG,
247};
248
249extern const struct qed_iov_hv_ops qed_iov_ops_pass;
250
251#ifdef CONFIG_QED_SRIOV
252/**
253 * @brief Check if given VF ID @vfid is valid
254 * w.r.t. @b_enabled_only value
255 * if b_enabled_only = true - only enabled VF id is valid
256 * else any VF id less than max_vfs is valid
257 *
258 * @param p_hwfn
259 * @param rel_vf_id - Relative VF ID
260 * @param b_enabled_only - consider only enabled VF
261 * @param b_non_malicious - true iff we want to validate vf isn't malicious.
262 *
263 * @return bool - true for valid VF ID
264 */
265bool qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
266 int rel_vf_id,
267 bool b_enabled_only, bool b_non_malicious);
268
269/**
270 * @brief - Given a VF index, return index of next [including that] active VF.
271 *
272 * @param p_hwfn
273 * @param rel_vf_id
274 *
275 * @return MAX_NUM_VFS in case no further active VFs, otherwise index.
276 */
277u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn, u16 rel_vf_id);
278
279void qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn,
280 int vfid, u16 vxlan_port, u16 geneve_port);
281
282/**
283 * @brief Read sriov related information and allocated resources
284 * reads from configuration space, shmem, etc.
285 *
286 * @param p_hwfn
287 *
288 * @return int
289 */
290int qed_iov_hw_info(struct qed_hwfn *p_hwfn);
291
292/**
293 * @brief qed_add_tlv - place a given tlv on the tlv buffer at next offset
294 *
295 * @param p_hwfn
296 * @param p_iov
297 * @param type
298 * @param length
299 *
300 * @return pointer to the newly placed tlv
301 */
302void *qed_add_tlv(struct qed_hwfn *p_hwfn, u8 **offset, u16 type, u16 length);
303
304/**
305 * @brief list the types and lengths of the tlvs on the buffer
306 *
307 * @param p_hwfn
308 * @param tlvs_list
309 */
310void qed_dp_tlv_list(struct qed_hwfn *p_hwfn, void *tlvs_list);
311
312/**
313 * @brief qed_iov_alloc - allocate sriov related resources
314 *
315 * @param p_hwfn
316 *
317 * @return int
318 */
319int qed_iov_alloc(struct qed_hwfn *p_hwfn);
320
321/**
322 * @brief qed_iov_setup - setup sriov related resources
323 *
324 * @param p_hwfn
325 */
326void qed_iov_setup(struct qed_hwfn *p_hwfn);
327
328/**
329 * @brief qed_iov_free - free sriov related resources
330 *
331 * @param p_hwfn
332 */
333void qed_iov_free(struct qed_hwfn *p_hwfn);
334
335/**
336 * @brief free sriov related memory that was allocated during hw_prepare
337 *
338 * @param cdev
339 */
340void qed_iov_free_hw_info(struct qed_dev *cdev);
341
342/**
343 * @brief Mark structs of vfs that have been FLR-ed.
344 *
345 * @param p_hwfn
346 * @param disabled_vfs - bitmask of all VFs on path that were FLRed
347 *
348 * @return true iff one of the PF's vfs got FLRed. false otherwise.
349 */
350bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn, u32 *disabled_vfs);
351
352/**
353 * @brief Search extended TLVs in request/reply buffer.
354 *
355 * @param p_hwfn
356 * @param p_tlvs_list - Pointer to tlvs list
357 * @param req_type - Type of TLV
358 *
359 * @return pointer to tlv type if found, otherwise returns NULL.
360 */
361void *qed_iov_search_list_tlvs(struct qed_hwfn *p_hwfn,
362 void *p_tlvs_list, u16 req_type);
363
364void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first);
365int qed_iov_wq_start(struct qed_dev *cdev);
366
367void qed_schedule_iov(struct qed_hwfn *hwfn, enum qed_iov_wq_flag flag);
368void qed_vf_start_iov_wq(struct qed_dev *cdev);
369int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled);
370void qed_inform_vf_link_state(struct qed_hwfn *hwfn);
371#else
372static inline bool
373qed_iov_is_valid_vfid(struct qed_hwfn *p_hwfn,
374 int rel_vf_id, bool b_enabled_only, bool b_non_malicious)
375{
376 return false;
377}
378
379static inline u16 qed_iov_get_next_active_vf(struct qed_hwfn *p_hwfn,
380 u16 rel_vf_id)
381{
382 return MAX_NUM_VFS;
383}
384
385static inline void
386qed_iov_bulletin_set_udp_ports(struct qed_hwfn *p_hwfn, int vfid,
387 u16 vxlan_port, u16 geneve_port)
388{
389}
390
391static inline int qed_iov_hw_info(struct qed_hwfn *p_hwfn)
392{
393 return 0;
394}
395
396static inline int qed_iov_alloc(struct qed_hwfn *p_hwfn)
397{
398 return 0;
399}
400
401static inline void qed_iov_setup(struct qed_hwfn *p_hwfn)
402{
403}
404
405static inline void qed_iov_free(struct qed_hwfn *p_hwfn)
406{
407}
408
409static inline void qed_iov_free_hw_info(struct qed_dev *cdev)
410{
411}
412
413static inline bool qed_iov_mark_vf_flr(struct qed_hwfn *p_hwfn,
414 u32 *disabled_vfs)
415{
416 return false;
417}
418
419static inline void qed_iov_wq_stop(struct qed_dev *cdev, bool schedule_first)
420{
421}
422
423static inline int qed_iov_wq_start(struct qed_dev *cdev)
424{
425 return 0;
426}
427
428static inline void qed_schedule_iov(struct qed_hwfn *hwfn,
429 enum qed_iov_wq_flag flag)
430{
431}
432
433static inline void qed_vf_start_iov_wq(struct qed_dev *cdev)
434{
435}
436
437static inline int qed_sriov_disable(struct qed_dev *cdev, bool pci_enabled)
438{
439 return 0;
440}
441
442static inline void qed_inform_vf_link_state(struct qed_hwfn *hwfn)
443{
444}
445#endif
446
447#define qed_for_each_vf(_p_hwfn, _i) \
448 for (_i = qed_iov_get_next_active_vf(_p_hwfn, 0); \
449 _i < MAX_NUM_VFS; \
450 _i = qed_iov_get_next_active_vf(_p_hwfn, _i + 1))
451
452#endif