Linux Audio

Check our new training course

Loading...
  1/* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */
  2/*
  3 * Copyright(c) 2018 Intel Corporation.
  4 *
  5 */
  6#ifndef HFI1_TID_RDMA_H
  7#define HFI1_TID_RDMA_H
  8
  9#include <linux/circ_buf.h>
 10#include "common.h"
 11
 12/* Add a convenience helper */
 13#define CIRC_ADD(val, add, size) (((val) + (add)) & ((size) - 1))
 14#define CIRC_NEXT(val, size) CIRC_ADD(val, 1, size)
 15#define CIRC_PREV(val, size) CIRC_ADD(val, -1, size)
 16
 17#define TID_RDMA_MIN_SEGMENT_SIZE       BIT(18)   /* 256 KiB (for now) */
 18#define TID_RDMA_MAX_SEGMENT_SIZE       BIT(18)   /* 256 KiB (for now) */
 19#define TID_RDMA_MAX_PAGES              (BIT(18) >> PAGE_SHIFT)
 20#define TID_RDMA_SEGMENT_SHIFT		18
 21
 22/*
 23 * Bit definitions for priv->s_flags.
 24 * These bit flags overload the bit flags defined for the QP's s_flags.
 25 * Due to the fact that these bit fields are used only for the QP priv
 26 * s_flags, there are no collisions.
 27 *
 28 * HFI1_S_TID_WAIT_INTERLCK - QP is waiting for requester interlock
 29 * HFI1_R_TID_WAIT_INTERLCK - QP is waiting for responder interlock
 30 */
 31#define HFI1_S_TID_BUSY_SET       BIT(0)
 32/* BIT(1) reserved for RVT_S_BUSY. */
 33#define HFI1_R_TID_RSC_TIMER      BIT(2)
 34/* BIT(3) reserved for RVT_S_RESP_PENDING. */
 35/* BIT(4) reserved for RVT_S_ACK_PENDING. */
 36#define HFI1_S_TID_WAIT_INTERLCK  BIT(5)
 37#define HFI1_R_TID_WAIT_INTERLCK  BIT(6)
 38/* BIT(7) - BIT(15) reserved for RVT_S_WAIT_*. */
 39/* BIT(16) reserved for RVT_S_SEND_ONE */
 40#define HFI1_S_TID_RETRY_TIMER    BIT(17)
 41/* BIT(18) reserved for RVT_S_ECN. */
 42#define HFI1_R_TID_SW_PSN         BIT(19)
 43/* BIT(26) reserved for HFI1_S_WAIT_HALT */
 44/* BIT(27) reserved for HFI1_S_WAIT_TID_RESP */
 45/* BIT(28) reserved for HFI1_S_WAIT_TID_SPACE */
 46
 47/*
 48 * Unlike regular IB RDMA VERBS, which do not require an entry
 49 * in the s_ack_queue, TID RDMA WRITE requests do because they
 50 * generate responses.
 51 * Therefore, the s_ack_queue needs to be extended by a certain
 52 * amount. The key point is that the queue needs to be extended
 53 * without letting the "user" know so they user doesn't end up
 54 * using these extra entries.
 55 */
 56#define HFI1_TID_RDMA_WRITE_CNT 8
 57
 58struct tid_rdma_params {
 59	struct rcu_head rcu_head;
 60	u32 qp;
 61	u32 max_len;
 62	u16 jkey;
 63	u8 max_read;
 64	u8 max_write;
 65	u8 timeout;
 66	u8 urg;
 67	u8 version;
 68};
 69
 70struct tid_rdma_qp_params {
 71	struct work_struct trigger_work;
 72	struct tid_rdma_params local;
 73	struct tid_rdma_params __rcu *remote;
 74};
 75
 76/* Track state for each hardware flow */
 77struct tid_flow_state {
 78	u32 generation;
 79	u32 psn;
 80	u8 index;
 81	u8 last_index;
 82};
 83
 84enum tid_rdma_req_state {
 85	TID_REQUEST_INACTIVE = 0,
 86	TID_REQUEST_INIT,
 87	TID_REQUEST_INIT_RESEND,
 88	TID_REQUEST_ACTIVE,
 89	TID_REQUEST_RESEND,
 90	TID_REQUEST_RESEND_ACTIVE,
 91	TID_REQUEST_QUEUED,
 92	TID_REQUEST_SYNC,
 93	TID_REQUEST_RNR_NAK,
 94	TID_REQUEST_COMPLETE,
 95};
 96
 97struct tid_rdma_request {
 98	struct rvt_qp *qp;
 99	struct hfi1_ctxtdata *rcd;
100	union {
101		struct rvt_swqe *swqe;
102		struct rvt_ack_entry *ack;
103	} e;
104
105	struct tid_rdma_flow *flows;	/* array of tid flows */
106	struct rvt_sge_state ss; /* SGE state for TID RDMA requests */
107	u16 n_flows;		/* size of the flow buffer window */
108	u16 setup_head;		/* flow index we are setting up */
109	u16 clear_tail;		/* flow index we are clearing */
110	u16 flow_idx;		/* flow index most recently set up */
111	u16 acked_tail;
112
113	u32 seg_len;
114	u32 total_len;
115	u32 r_ack_psn;          /* next expected ack PSN */
116	u32 r_flow_psn;         /* IB PSN of next segment start */
117	u32 r_last_acked;       /* IB PSN of last ACK'ed packet */
118	u32 s_next_psn;		/* IB PSN of next segment start for read */
119
120	u32 total_segs;		/* segments required to complete a request */
121	u32 cur_seg;		/* index of current segment */
122	u32 comp_seg;           /* index of last completed segment */
123	u32 ack_seg;            /* index of last ack'ed segment */
124	u32 alloc_seg;          /* index of next segment to be allocated */
125	u32 isge;		/* index of "current" sge */
126	u32 ack_pending;        /* num acks pending for this request */
127
128	enum tid_rdma_req_state state;
129};
130
131/*
132 * When header suppression is used, PSNs associated with a "flow" are
133 * relevant (and not the PSNs maintained by verbs). Track per-flow
134 * PSNs here for a TID RDMA segment.
135 *
136 */
137struct flow_state {
138	u32 flags;
139	u32 resp_ib_psn;     /* The IB PSN of the response for this flow */
140	u32 generation;      /* generation of flow */
141	u32 spsn;            /* starting PSN in TID space */
142	u32 lpsn;            /* last PSN in TID space */
143	u32 r_next_psn;      /* next PSN to be received (in TID space) */
144
145	/* For tid rdma read */
146	u32 ib_spsn;         /* starting PSN in Verbs space */
147	u32 ib_lpsn;         /* last PSn in Verbs space */
148};
149
150struct tid_rdma_pageset {
151	dma_addr_t addr : 48; /* Only needed for the first page */
152	u8 idx: 8;
153	u8 count : 7;
154	u8 mapped: 1;
155};
156
157/**
158 * kern_tid_node - used for managing TID's in TID groups
159 *
160 * @grp_idx: rcd relative index to tid_group
161 * @map: grp->map captured prior to programming this TID group in HW
162 * @cnt: Only @cnt of available group entries are actually programmed
163 */
164struct kern_tid_node {
165	struct tid_group *grp;
166	u8 map;
167	u8 cnt;
168};
169
170/* Overall info for a TID RDMA segment */
171struct tid_rdma_flow {
172	/*
173	 * While a TID RDMA segment is being transferred, it uses a QP number
174	 * from the "KDETH section of QP numbers" (which is different from the
175	 * QP number that originated the request). Bits 11-15 of these QP
176	 * numbers identify the "TID flow" for the segment.
177	 */
178	struct flow_state flow_state;
179	struct tid_rdma_request *req;
180	u32 tid_qpn;
181	u32 tid_offset;
182	u32 length;
183	u32 sent;
184	u8 tnode_cnt;
185	u8 tidcnt;
186	u8 tid_idx;
187	u8 idx;
188	u8 npagesets;
189	u8 npkts;
190	u8 pkt;
191	u8 resync_npkts;
192	struct kern_tid_node tnode[TID_RDMA_MAX_PAGES];
193	struct tid_rdma_pageset pagesets[TID_RDMA_MAX_PAGES];
194	u32 tid_entry[TID_RDMA_MAX_PAGES];
195};
196
197enum tid_rnr_nak_state {
198	TID_RNR_NAK_INIT = 0,
199	TID_RNR_NAK_SEND,
200	TID_RNR_NAK_SENT,
201};
202
203bool tid_rdma_conn_req(struct rvt_qp *qp, u64 *data);
204bool tid_rdma_conn_reply(struct rvt_qp *qp, u64 data);
205bool tid_rdma_conn_resp(struct rvt_qp *qp, u64 *data);
206void tid_rdma_conn_error(struct rvt_qp *qp);
207void tid_rdma_opfn_init(struct rvt_qp *qp, struct tid_rdma_params *p);
208
209int hfi1_kern_exp_rcv_init(struct hfi1_ctxtdata *rcd, int reinit);
210int hfi1_kern_exp_rcv_setup(struct tid_rdma_request *req,
211			    struct rvt_sge_state *ss, bool *last);
212int hfi1_kern_exp_rcv_clear(struct tid_rdma_request *req);
213void hfi1_kern_exp_rcv_clear_all(struct tid_rdma_request *req);
214void __trdma_clean_swqe(struct rvt_qp *qp, struct rvt_swqe *wqe);
215
216/**
217 * trdma_clean_swqe - clean flows for swqe if large send queue
218 * @qp: the qp
219 * @wqe: the send wqe
220 */
221static inline void trdma_clean_swqe(struct rvt_qp *qp, struct rvt_swqe *wqe)
222{
223	if (!wqe->priv)
224		return;
225	__trdma_clean_swqe(qp, wqe);
226}
227
228void hfi1_kern_read_tid_flow_free(struct rvt_qp *qp);
229
230int hfi1_qp_priv_init(struct rvt_dev_info *rdi, struct rvt_qp *qp,
231		      struct ib_qp_init_attr *init_attr);
232void hfi1_qp_priv_tid_free(struct rvt_dev_info *rdi, struct rvt_qp *qp);
233
234void hfi1_tid_rdma_flush_wait(struct rvt_qp *qp);
235
236int hfi1_kern_setup_hw_flow(struct hfi1_ctxtdata *rcd, struct rvt_qp *qp);
237void hfi1_kern_clear_hw_flow(struct hfi1_ctxtdata *rcd, struct rvt_qp *qp);
238void hfi1_kern_init_ctxt_generations(struct hfi1_ctxtdata *rcd);
239
240struct cntr_entry;
241u64 hfi1_access_sw_tid_wait(const struct cntr_entry *entry,
242			    void *context, int vl, int mode, u64 data);
243
244u32 hfi1_build_tid_rdma_read_packet(struct rvt_swqe *wqe,
245				    struct ib_other_headers *ohdr,
246				    u32 *bth1, u32 *bth2, u32 *len);
247u32 hfi1_build_tid_rdma_read_req(struct rvt_qp *qp, struct rvt_swqe *wqe,
248				 struct ib_other_headers *ohdr, u32 *bth1,
249				 u32 *bth2, u32 *len);
250void hfi1_rc_rcv_tid_rdma_read_req(struct hfi1_packet *packet);
251u32 hfi1_build_tid_rdma_read_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
252				  struct ib_other_headers *ohdr, u32 *bth0,
253				  u32 *bth1, u32 *bth2, u32 *len, bool *last);
254void hfi1_rc_rcv_tid_rdma_read_resp(struct hfi1_packet *packet);
255bool hfi1_handle_kdeth_eflags(struct hfi1_ctxtdata *rcd,
256			      struct hfi1_pportdata *ppd,
257			      struct hfi1_packet *packet);
258void hfi1_tid_rdma_restart_req(struct rvt_qp *qp, struct rvt_swqe *wqe,
259			       u32 *bth2);
260void hfi1_qp_kern_exp_rcv_clear_all(struct rvt_qp *qp);
261bool hfi1_tid_rdma_wqe_interlock(struct rvt_qp *qp, struct rvt_swqe *wqe);
262
263void setup_tid_rdma_wqe(struct rvt_qp *qp, struct rvt_swqe *wqe);
264static inline void hfi1_setup_tid_rdma_wqe(struct rvt_qp *qp,
265					   struct rvt_swqe *wqe)
266{
267	if (wqe->priv &&
268	    (wqe->wr.opcode == IB_WR_RDMA_READ ||
269	     wqe->wr.opcode == IB_WR_RDMA_WRITE) &&
270	    wqe->length >= TID_RDMA_MIN_SEGMENT_SIZE)
271		setup_tid_rdma_wqe(qp, wqe);
272}
273
274u32 hfi1_build_tid_rdma_write_req(struct rvt_qp *qp, struct rvt_swqe *wqe,
275				  struct ib_other_headers *ohdr,
276				  u32 *bth1, u32 *bth2, u32 *len);
277
278void hfi1_rc_rcv_tid_rdma_write_req(struct hfi1_packet *packet);
279
280u32 hfi1_build_tid_rdma_write_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
281				   struct ib_other_headers *ohdr, u32 *bth1,
282				   u32 bth2, u32 *len,
283				   struct rvt_sge_state **ss);
284
285void hfi1_del_tid_reap_timer(struct rvt_qp *qp);
286
287void hfi1_rc_rcv_tid_rdma_write_resp(struct hfi1_packet *packet);
288
289bool hfi1_build_tid_rdma_packet(struct rvt_swqe *wqe,
290				struct ib_other_headers *ohdr,
291				u32 *bth1, u32 *bth2, u32 *len);
292
293void hfi1_rc_rcv_tid_rdma_write_data(struct hfi1_packet *packet);
294
295u32 hfi1_build_tid_rdma_write_ack(struct rvt_qp *qp, struct rvt_ack_entry *e,
296				  struct ib_other_headers *ohdr, u16 iflow,
297				  u32 *bth1, u32 *bth2);
298
299void hfi1_rc_rcv_tid_rdma_ack(struct hfi1_packet *packet);
300
301void hfi1_add_tid_retry_timer(struct rvt_qp *qp);
302void hfi1_del_tid_retry_timer(struct rvt_qp *qp);
303
304u32 hfi1_build_tid_rdma_resync(struct rvt_qp *qp, struct rvt_swqe *wqe,
305			       struct ib_other_headers *ohdr, u32 *bth1,
306			       u32 *bth2, u16 fidx);
307
308void hfi1_rc_rcv_tid_rdma_resync(struct hfi1_packet *packet);
309
310struct hfi1_pkt_state;
311int hfi1_make_tid_rdma_pkt(struct rvt_qp *qp, struct hfi1_pkt_state *ps);
312
313void _hfi1_do_tid_send(struct work_struct *work);
314
315bool hfi1_schedule_tid_send(struct rvt_qp *qp);
316
317bool hfi1_tid_rdma_ack_interlock(struct rvt_qp *qp, struct rvt_ack_entry *e);
318
319#endif /* HFI1_TID_RDMA_H */