Linux Audio

Check our new training course

Loading...
v6.8
  1/* SPDX-License-Identifier: GPL-2.0-or-later */
  2/*
  3 * RDMA Transport Layer
  4 *
  5 * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved.
  6 * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved.
  7 * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
  8 */
  9
 10#ifndef RTRS_PRI_H
 11#define RTRS_PRI_H
 12
 13#include <linux/uuid.h>
 14#include <rdma/rdma_cm.h>
 15#include <rdma/ib_verbs.h>
 16#include <rdma/ib.h>
 17
 18#include "rtrs.h"
 19
 20#define RTRS_PROTO_VER_MAJOR 2
 21#define RTRS_PROTO_VER_MINOR 0
 22
 23#define RTRS_PROTO_VER_STRING __stringify(RTRS_PROTO_VER_MAJOR) "." \
 24			       __stringify(RTRS_PROTO_VER_MINOR)
 25
 26/*
 27 * Max IB immediate data size is 2^28 (MAX_IMM_PAYL_BITS)
 28 * and the minimum chunk size is 4096 (2^12).
 29 * So the maximum sess_queue_depth is 65535 (2^16 - 1) in theory
 30 * since queue_depth in rtrs_msg_conn_rsp is defined as le16.
 31 * Therefore the pratical max value of sess_queue_depth is
 32 * somewhere between 1 and 65535 and it depends on the system.
 33 */
 34#define MAX_SESS_QUEUE_DEPTH 65535
 35
 36enum rtrs_imm_const {
 37	MAX_IMM_TYPE_BITS = 4,
 38	MAX_IMM_TYPE_MASK = ((1 << MAX_IMM_TYPE_BITS) - 1),
 39	MAX_IMM_PAYL_BITS = 28,
 40	MAX_IMM_PAYL_MASK = ((1 << MAX_IMM_PAYL_BITS) - 1),
 41};
 42
 43enum rtrs_imm_type {
 44	RTRS_IO_REQ_IMM       = 0, /* client to server */
 45	RTRS_IO_RSP_IMM       = 1, /* server to client */
 46	RTRS_IO_RSP_W_INV_IMM = 2, /* server to client */
 47
 48	RTRS_HB_MSG_IMM = 8, /* HB: HeartBeat */
 49	RTRS_HB_ACK_IMM = 9,
 50
 51	RTRS_LAST_IMM,
 52};
 53
 54enum {
 55	SERVICE_CON_QUEUE_DEPTH = 512,
 56
 57	MAX_PATHS_NUM = 128,
 58
 59	MIN_CHUNK_SIZE = 8192,
 60
 61	RTRS_HB_INTERVAL_MS = 5000,
 62	RTRS_HB_MISSED_MAX = 5,
 63
 64	RTRS_MAGIC = 0x1BBD,
 65	RTRS_PROTO_VER = (RTRS_PROTO_VER_MAJOR << 8) | RTRS_PROTO_VER_MINOR,
 66};
 67
 68struct rtrs_ib_dev;
 69
 70struct rtrs_rdma_dev_pd_ops {
 71	int (*init)(struct rtrs_ib_dev *dev);
 72};
 73
 74struct rtrs_rdma_dev_pd {
 75	struct mutex		mutex;
 76	struct list_head	list;
 77	enum ib_pd_flags	pd_flags;
 78	const struct rtrs_rdma_dev_pd_ops *ops;
 79};
 80
 81struct rtrs_ib_dev {
 82	struct ib_device	 *ib_dev;
 83	struct ib_pd		 *ib_pd;
 84	struct kref		 ref;
 85	struct list_head	 entry;
 86	struct rtrs_rdma_dev_pd *pool;
 87};
 88
 89struct rtrs_con {
 90	struct rtrs_path	*path;
 91	struct ib_qp		*qp;
 92	struct ib_cq		*cq;
 93	struct rdma_cm_id	*cm_id;
 94	unsigned int		cid;
 95	int                     nr_cqe;
 96	atomic_t		wr_cnt;
 97	atomic_t		sq_wr_avail;
 98};
 99
100struct rtrs_path {
101	struct list_head	entry;
102	struct sockaddr_storage dst_addr;
103	struct sockaddr_storage src_addr;
104	char			sessname[NAME_MAX];
105	uuid_t			uuid;
106	struct rtrs_con	**con;
107	unsigned int		con_num;
108	unsigned int		irq_con_num;
109	unsigned int		recon_cnt;
110	unsigned int		signal_interval;
111	struct rtrs_ib_dev	*dev;
112	int			dev_ref;
113	struct ib_cqe		*hb_cqe;
114	void			(*hb_err_handler)(struct rtrs_con *con);
115	struct workqueue_struct *hb_wq;
116	struct delayed_work	hb_dwork;
117	unsigned int		hb_interval_ms;
118	unsigned int		hb_missed_cnt;
119	unsigned int		hb_missed_max;
120	ktime_t			hb_last_sent;
121	ktime_t			hb_cur_latency;
122};
123
124/* rtrs information unit */
125struct rtrs_iu {
126	struct ib_cqe           cqe;
127	dma_addr_t              dma_addr;
128	void                    *buf;
129	size_t                  size;
130	enum dma_data_direction direction;
131};
132
133/**
134 * enum rtrs_msg_types - RTRS message types, see also rtrs/README
135 * @RTRS_MSG_INFO_REQ:		Client additional info request to the server
136 * @RTRS_MSG_INFO_RSP:		Server additional info response to the client
137 * @RTRS_MSG_WRITE:		Client writes data per RDMA to server
138 * @RTRS_MSG_READ:		Client requests data transfer from server
139 * @RTRS_MSG_RKEY_RSP:		Server refreshed rkey for rbuf
140 */
141enum rtrs_msg_types {
142	RTRS_MSG_INFO_REQ,
143	RTRS_MSG_INFO_RSP,
144	RTRS_MSG_WRITE,
145	RTRS_MSG_READ,
146	RTRS_MSG_RKEY_RSP,
147};
148
149/**
150 * enum rtrs_msg_flags - RTRS message flags.
151 * @RTRS_NEED_INVAL:	Send invalidation in response.
152 * @RTRS_MSG_NEW_RKEY_F: Send refreshed rkey in response.
153 */
154enum rtrs_msg_flags {
155	RTRS_MSG_NEED_INVAL_F = 1 << 0,
156	RTRS_MSG_NEW_RKEY_F = 1 << 1,
157};
158
159/**
160 * struct rtrs_sg_desc - RDMA-Buffer entry description
161 * @addr:	Address of RDMA destination buffer
162 * @key:	Authorization rkey to write to the buffer
163 * @len:	Size of the buffer
164 */
165struct rtrs_sg_desc {
166	__le64			addr;
167	__le32			key;
168	__le32			len;
169};
170
171/**
172 * struct rtrs_msg_conn_req - Client connection request to the server
173 * @magic:	   RTRS magic
174 * @version:	   RTRS protocol version
175 * @cid:	   Current connection id
176 * @cid_num:	   Number of connections per session
177 * @recon_cnt:	   Reconnections counter
178 * @sess_uuid:	   UUID of a session (path)
179 * @paths_uuid:	   UUID of a group of sessions (paths)
180 *
181 * NOTE: max size 56 bytes, see man rdma_connect().
182 */
183struct rtrs_msg_conn_req {
184	/* Is set to 0 by cma.c in case of AF_IB, do not touch that.
185	 * see https://www.spinics.net/lists/linux-rdma/msg22397.html
186	 */
187	u8		__cma_version;
188	/* On sender side that should be set to 0, or cma_save_ip_info()
189	 * extract garbage and will fail.
190	 */
191	u8		__ip_version;
192	__le16		magic;
193	__le16		version;
194	__le16		cid;
195	__le16		cid_num;
196	__le16		recon_cnt;
197	uuid_t		sess_uuid;
198	uuid_t		paths_uuid;
199	u8		first_conn : 1;
200	u8		reserved_bits : 7;
201	u8		reserved[11];
202};
203
204/**
205 * struct rtrs_msg_conn_rsp - Server connection response to the client
206 * @magic:	   RTRS magic
207 * @version:	   RTRS protocol version
208 * @errno:	   If rdma_accept() then 0, if rdma_reject() indicates error
209 * @queue_depth:   max inflight messages (queue-depth) in this session
210 * @max_io_size:   max io size server supports
211 * @max_hdr_size:  max msg header size server supports
212 *
213 * NOTE: size is 56 bytes, max possible is 136 bytes, see man rdma_accept().
214 */
215struct rtrs_msg_conn_rsp {
216	__le16		magic;
217	__le16		version;
218	__le16		errno;
219	__le16		queue_depth;
220	__le32		max_io_size;
221	__le32		max_hdr_size;
222	__le32		flags;
223	u8		reserved[36];
224};
225
226/**
227 * struct rtrs_msg_info_req
228 * @type:		@RTRS_MSG_INFO_REQ
229 * @pathname:		Path name chosen by client
230 */
231struct rtrs_msg_info_req {
232	__le16		type;
233	u8		pathname[NAME_MAX];
234	u8		reserved[15];
235};
236
237/**
238 * struct rtrs_msg_info_rsp
239 * @type:		@RTRS_MSG_INFO_RSP
240 * @sg_cnt:		Number of @desc entries
241 * @desc:		RDMA buffers where the client can write to server
242 */
243struct rtrs_msg_info_rsp {
244	__le16		type;
245	__le16          sg_cnt;
246	u8              reserved[4];
247	struct rtrs_sg_desc desc[];
248};
249
250/**
251 * struct rtrs_msg_rkey_rsp
252 * @type:		@RTRS_MSG_RKEY_RSP
253 * @buf_id:		RDMA buf_id of the new rkey
254 * @rkey:		new remote key for RDMA buffers id from server
255 */
256struct rtrs_msg_rkey_rsp {
257	__le16		type;
258	__le16          buf_id;
259	__le32		rkey;
260};
261
262/**
263 * struct rtrs_msg_rdma_read - RDMA data transfer request from client
264 * @type:		always @RTRS_MSG_READ
265 * @usr_len:		length of user payload
266 * @sg_cnt:		number of @desc entries
267 * @desc:		RDMA buffers where the server can write the result to
268 */
269struct rtrs_msg_rdma_read {
270	__le16			type;
271	__le16			usr_len;
272	__le16			flags;
273	__le16			sg_cnt;
274	struct rtrs_sg_desc    desc[];
275};
276
277/**
278 * struct_msg_rdma_write - Message transferred to server with RDMA-Write
279 * @type:		always @RTRS_MSG_WRITE
280 * @usr_len:		length of user payload
281 */
282struct rtrs_msg_rdma_write {
283	__le16			type;
284	__le16			usr_len;
285};
286
287/**
288 * struct_msg_rdma_hdr - header for read or write request
289 * @type:		@RTRS_MSG_WRITE | @RTRS_MSG_READ
290 */
291struct rtrs_msg_rdma_hdr {
292	__le16			type;
293};
294
295/* rtrs.c */
296
297struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t,
298			      struct ib_device *dev, enum dma_data_direction,
299			      void (*done)(struct ib_cq *cq, struct ib_wc *wc));
300void rtrs_iu_free(struct rtrs_iu *iu, struct ib_device *dev, u32 queue_num);
301int rtrs_iu_post_recv(struct rtrs_con *con, struct rtrs_iu *iu);
302int rtrs_iu_post_send(struct rtrs_con *con, struct rtrs_iu *iu, size_t size,
303		      struct ib_send_wr *head);
304int rtrs_iu_post_rdma_write_imm(struct rtrs_con *con, struct rtrs_iu *iu,
305				struct ib_sge *sge, unsigned int num_sge,
306				u32 rkey, u64 rdma_addr, u32 imm_data,
307				enum ib_send_flags flags,
308				struct ib_send_wr *head,
309				struct ib_send_wr *tail);
310
311int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe);
312
313int rtrs_cq_qp_create(struct rtrs_path *path, struct rtrs_con *con,
314		      u32 max_send_sge, int cq_vector, int nr_cqe,
315		      u32 max_send_wr, u32 max_recv_wr,
316		      enum ib_poll_context poll_ctx);
317void rtrs_cq_qp_destroy(struct rtrs_con *con);
318
319void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe,
320		  unsigned int interval_ms, unsigned int missed_max,
321		  void (*err_handler)(struct rtrs_con *con),
322		  struct workqueue_struct *wq);
323void rtrs_start_hb(struct rtrs_path *path);
324void rtrs_stop_hb(struct rtrs_path *path);
325void rtrs_send_hb_ack(struct rtrs_path *path);
326
327void rtrs_rdma_dev_pd_init(enum ib_pd_flags pd_flags,
328			   struct rtrs_rdma_dev_pd *pool);
329void rtrs_rdma_dev_pd_deinit(struct rtrs_rdma_dev_pd *pool);
330
331struct rtrs_ib_dev *rtrs_ib_dev_find_or_add(struct ib_device *ib_dev,
332					    struct rtrs_rdma_dev_pd *pool);
333int rtrs_ib_dev_put(struct rtrs_ib_dev *dev);
334
335static inline u32 rtrs_to_imm(u32 type, u32 payload)
336{
337	BUILD_BUG_ON(MAX_IMM_PAYL_BITS + MAX_IMM_TYPE_BITS != 32);
338	BUILD_BUG_ON(RTRS_LAST_IMM > (1<<MAX_IMM_TYPE_BITS));
339	return ((type & MAX_IMM_TYPE_MASK) << MAX_IMM_PAYL_BITS) |
340		(payload & MAX_IMM_PAYL_MASK);
341}
342
343static inline void rtrs_from_imm(u32 imm, u32 *type, u32 *payload)
344{
345	*payload = imm & MAX_IMM_PAYL_MASK;
346	*type = imm >> MAX_IMM_PAYL_BITS;
347}
348
349static inline u32 rtrs_to_io_req_imm(u32 addr)
350{
351	return rtrs_to_imm(RTRS_IO_REQ_IMM, addr);
352}
353
354static inline u32 rtrs_to_io_rsp_imm(u32 msg_id, int errno, bool w_inval)
355{
356	enum rtrs_imm_type type;
357	u32 payload;
358
359	/* 9 bits for errno, 19 bits for msg_id */
360	payload = (abs(errno) & 0x1ff) << 19 | (msg_id & 0x7ffff);
361	type = w_inval ? RTRS_IO_RSP_W_INV_IMM : RTRS_IO_RSP_IMM;
362
363	return rtrs_to_imm(type, payload);
364}
365
366static inline void rtrs_from_io_rsp_imm(u32 payload, u32 *msg_id, int *errno)
367{
368	/* 9 bits for errno, 19 bits for msg_id */
369	*msg_id = payload & 0x7ffff;
370	*errno = -(int)((payload >> 19) & 0x1ff);
371}
372
373#define STAT_STORE_FUNC(type, set_value, reset)				\
374static ssize_t set_value##_store(struct kobject *kobj,			\
375			     struct kobj_attribute *attr,		\
376			     const char *buf, size_t count)		\
377{									\
378	int ret = -EINVAL;						\
379	type *stats = container_of(kobj, type, kobj_stats);		\
380									\
381	if (sysfs_streq(buf, "1"))					\
382		ret = reset(stats, true);			\
383	else if (sysfs_streq(buf, "0"))					\
384		ret = reset(stats, false);			\
385	if (ret)							\
386		return ret;						\
387									\
388	return count;							\
389}
390
391#define STAT_SHOW_FUNC(type, get_value, print)				\
392static ssize_t get_value##_show(struct kobject *kobj,			\
393			   struct kobj_attribute *attr,			\
394			   char *page)					\
395{									\
396	type *stats = container_of(kobj, type, kobj_stats);		\
397									\
398	return print(stats, page);			\
399}
400
401#define STAT_ATTR(type, stat, print, reset)				\
402STAT_STORE_FUNC(type, stat, reset)					\
403STAT_SHOW_FUNC(type, stat, print)					\
404static struct kobj_attribute stat##_attr = __ATTR_RW(stat)
405
406#endif /* RTRS_PRI_H */
v6.9.4
  1/* SPDX-License-Identifier: GPL-2.0-or-later */
  2/*
  3 * RDMA Transport Layer
  4 *
  5 * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved.
  6 * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved.
  7 * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
  8 */
  9
 10#ifndef RTRS_PRI_H
 11#define RTRS_PRI_H
 12
 13#include <linux/uuid.h>
 14#include <rdma/rdma_cm.h>
 15#include <rdma/ib_verbs.h>
 16#include <rdma/ib.h>
 17
 18#include "rtrs.h"
 19
 20#define RTRS_PROTO_VER_MAJOR 2
 21#define RTRS_PROTO_VER_MINOR 0
 22
 23#define RTRS_PROTO_VER_STRING __stringify(RTRS_PROTO_VER_MAJOR) "." \
 24			       __stringify(RTRS_PROTO_VER_MINOR)
 25
 26/*
 27 * Max IB immediate data size is 2^28 (MAX_IMM_PAYL_BITS)
 28 * and the minimum chunk size is 4096 (2^12).
 29 * So the maximum sess_queue_depth is 65535 (2^16 - 1) in theory
 30 * since queue_depth in rtrs_msg_conn_rsp is defined as le16.
 31 * Therefore the pratical max value of sess_queue_depth is
 32 * somewhere between 1 and 65535 and it depends on the system.
 33 */
 34#define MAX_SESS_QUEUE_DEPTH 65535
 35
 36enum rtrs_imm_const {
 37	MAX_IMM_TYPE_BITS = 4,
 38	MAX_IMM_TYPE_MASK = ((1 << MAX_IMM_TYPE_BITS) - 1),
 39	MAX_IMM_PAYL_BITS = 28,
 40	MAX_IMM_PAYL_MASK = ((1 << MAX_IMM_PAYL_BITS) - 1),
 41};
 42
 43enum rtrs_imm_type {
 44	RTRS_IO_REQ_IMM       = 0, /* client to server */
 45	RTRS_IO_RSP_IMM       = 1, /* server to client */
 46	RTRS_IO_RSP_W_INV_IMM = 2, /* server to client */
 47
 48	RTRS_HB_MSG_IMM = 8, /* HB: HeartBeat */
 49	RTRS_HB_ACK_IMM = 9,
 50
 51	RTRS_LAST_IMM,
 52};
 53
 54enum {
 55	SERVICE_CON_QUEUE_DEPTH = 512,
 56
 57	MAX_PATHS_NUM = 128,
 58
 59	MIN_CHUNK_SIZE = 8192,
 60
 61	RTRS_HB_INTERVAL_MS = 5000,
 62	RTRS_HB_MISSED_MAX = 5,
 63
 64	RTRS_MAGIC = 0x1BBD,
 65	RTRS_PROTO_VER = (RTRS_PROTO_VER_MAJOR << 8) | RTRS_PROTO_VER_MINOR,
 66};
 67
 68struct rtrs_ib_dev;
 69
 70struct rtrs_rdma_dev_pd_ops {
 71	int (*init)(struct rtrs_ib_dev *dev);
 72};
 73
 74struct rtrs_rdma_dev_pd {
 75	struct mutex		mutex;
 76	struct list_head	list;
 77	enum ib_pd_flags	pd_flags;
 78	const struct rtrs_rdma_dev_pd_ops *ops;
 79};
 80
 81struct rtrs_ib_dev {
 82	struct ib_device	 *ib_dev;
 83	struct ib_pd		 *ib_pd;
 84	struct kref		 ref;
 85	struct list_head	 entry;
 86	struct rtrs_rdma_dev_pd *pool;
 87};
 88
 89struct rtrs_con {
 90	struct rtrs_path	*path;
 91	struct ib_qp		*qp;
 92	struct ib_cq		*cq;
 93	struct rdma_cm_id	*cm_id;
 94	unsigned int		cid;
 95	int                     nr_cqe;
 96	atomic_t		wr_cnt;
 97	atomic_t		sq_wr_avail;
 98};
 99
100struct rtrs_path {
101	struct list_head	entry;
102	struct sockaddr_storage dst_addr;
103	struct sockaddr_storage src_addr;
104	char			sessname[NAME_MAX];
105	uuid_t			uuid;
106	struct rtrs_con	**con;
107	unsigned int		con_num;
108	unsigned int		irq_con_num;
109	unsigned int		recon_cnt;
110	unsigned int		signal_interval;
111	struct rtrs_ib_dev	*dev;
112	int			dev_ref;
113	struct ib_cqe		*hb_cqe;
114	void			(*hb_err_handler)(struct rtrs_con *con);
115	struct workqueue_struct *hb_wq;
116	struct delayed_work	hb_dwork;
117	unsigned int		hb_interval_ms;
118	unsigned int		hb_missed_cnt;
119	unsigned int		hb_missed_max;
120	ktime_t			hb_last_sent;
121	ktime_t			hb_cur_latency;
122};
123
124/* rtrs information unit */
125struct rtrs_iu {
126	struct ib_cqe           cqe;
127	dma_addr_t              dma_addr;
128	void                    *buf;
129	size_t                  size;
130	enum dma_data_direction direction;
131};
132
133/**
134 * enum rtrs_msg_types - RTRS message types, see also rtrs/README
135 * @RTRS_MSG_INFO_REQ:		Client additional info request to the server
136 * @RTRS_MSG_INFO_RSP:		Server additional info response to the client
137 * @RTRS_MSG_WRITE:		Client writes data per RDMA to server
138 * @RTRS_MSG_READ:		Client requests data transfer from server
139 * @RTRS_MSG_RKEY_RSP:		Server refreshed rkey for rbuf
140 */
141enum rtrs_msg_types {
142	RTRS_MSG_INFO_REQ,
143	RTRS_MSG_INFO_RSP,
144	RTRS_MSG_WRITE,
145	RTRS_MSG_READ,
146	RTRS_MSG_RKEY_RSP,
147};
148
149/**
150 * enum rtrs_msg_flags - RTRS message flags.
151 * @RTRS_NEED_INVAL:	Send invalidation in response.
152 * @RTRS_MSG_NEW_RKEY_F: Send refreshed rkey in response.
153 */
154enum rtrs_msg_flags {
155	RTRS_MSG_NEED_INVAL_F = 1 << 0,
156	RTRS_MSG_NEW_RKEY_F = 1 << 1,
157};
158
159/**
160 * struct rtrs_sg_desc - RDMA-Buffer entry description
161 * @addr:	Address of RDMA destination buffer
162 * @key:	Authorization rkey to write to the buffer
163 * @len:	Size of the buffer
164 */
165struct rtrs_sg_desc {
166	__le64			addr;
167	__le32			key;
168	__le32			len;
169};
170
171/**
172 * struct rtrs_msg_conn_req - Client connection request to the server
173 * @magic:	   RTRS magic
174 * @version:	   RTRS protocol version
175 * @cid:	   Current connection id
176 * @cid_num:	   Number of connections per session
177 * @recon_cnt:	   Reconnections counter
178 * @sess_uuid:	   UUID of a session (path)
179 * @paths_uuid:	   UUID of a group of sessions (paths)
180 *
181 * NOTE: max size 56 bytes, see man rdma_connect().
182 */
183struct rtrs_msg_conn_req {
184	/* Is set to 0 by cma.c in case of AF_IB, do not touch that.
185	 * see https://www.spinics.net/lists/linux-rdma/msg22397.html
186	 */
187	u8		__cma_version;
188	/* On sender side that should be set to 0, or cma_save_ip_info()
189	 * extract garbage and will fail.
190	 */
191	u8		__ip_version;
192	__le16		magic;
193	__le16		version;
194	__le16		cid;
195	__le16		cid_num;
196	__le16		recon_cnt;
197	uuid_t		sess_uuid;
198	uuid_t		paths_uuid;
199	u8		first_conn : 1;
200	u8		reserved_bits : 7;
201	u8		reserved[11];
202};
203
204/**
205 * struct rtrs_msg_conn_rsp - Server connection response to the client
206 * @magic:	   RTRS magic
207 * @version:	   RTRS protocol version
208 * @errno:	   If rdma_accept() then 0, if rdma_reject() indicates error
209 * @queue_depth:   max inflight messages (queue-depth) in this session
210 * @max_io_size:   max io size server supports
211 * @max_hdr_size:  max msg header size server supports
212 *
213 * NOTE: size is 56 bytes, max possible is 136 bytes, see man rdma_accept().
214 */
215struct rtrs_msg_conn_rsp {
216	__le16		magic;
217	__le16		version;
218	__le16		errno;
219	__le16		queue_depth;
220	__le32		max_io_size;
221	__le32		max_hdr_size;
222	__le32		flags;
223	u8		reserved[36];
224};
225
226/**
227 * struct rtrs_msg_info_req
228 * @type:		@RTRS_MSG_INFO_REQ
229 * @pathname:		Path name chosen by client
230 */
231struct rtrs_msg_info_req {
232	__le16		type;
233	u8		pathname[NAME_MAX];
234	u8		reserved[15];
235};
236
237/**
238 * struct rtrs_msg_info_rsp
239 * @type:		@RTRS_MSG_INFO_RSP
240 * @sg_cnt:		Number of @desc entries
241 * @desc:		RDMA buffers where the client can write to server
242 */
243struct rtrs_msg_info_rsp {
244	__le16		type;
245	__le16          sg_cnt;
246	u8              reserved[4];
247	struct rtrs_sg_desc desc[];
248};
249
250/**
251 * struct rtrs_msg_rkey_rsp
252 * @type:		@RTRS_MSG_RKEY_RSP
253 * @buf_id:		RDMA buf_id of the new rkey
254 * @rkey:		new remote key for RDMA buffers id from server
255 */
256struct rtrs_msg_rkey_rsp {
257	__le16		type;
258	__le16          buf_id;
259	__le32		rkey;
260};
261
262/**
263 * struct rtrs_msg_rdma_read - RDMA data transfer request from client
264 * @type:		always @RTRS_MSG_READ
265 * @usr_len:		length of user payload
266 * @sg_cnt:		number of @desc entries
267 * @desc:		RDMA buffers where the server can write the result to
268 */
269struct rtrs_msg_rdma_read {
270	__le16			type;
271	__le16			usr_len;
272	__le16			flags;
273	__le16			sg_cnt;
274	struct rtrs_sg_desc    desc[];
275};
276
277/**
278 * struct_msg_rdma_write - Message transferred to server with RDMA-Write
279 * @type:		always @RTRS_MSG_WRITE
280 * @usr_len:		length of user payload
281 */
282struct rtrs_msg_rdma_write {
283	__le16			type;
284	__le16			usr_len;
285};
286
287/**
288 * struct_msg_rdma_hdr - header for read or write request
289 * @type:		@RTRS_MSG_WRITE | @RTRS_MSG_READ
290 */
291struct rtrs_msg_rdma_hdr {
292	__le16			type;
293};
294
295/* rtrs.c */
296
297struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t,
298			      struct ib_device *dev, enum dma_data_direction,
299			      void (*done)(struct ib_cq *cq, struct ib_wc *wc));
300void rtrs_iu_free(struct rtrs_iu *iu, struct ib_device *dev, u32 queue_num);
301int rtrs_iu_post_recv(struct rtrs_con *con, struct rtrs_iu *iu);
302int rtrs_iu_post_send(struct rtrs_con *con, struct rtrs_iu *iu, size_t size,
303		      struct ib_send_wr *head);
304int rtrs_iu_post_rdma_write_imm(struct rtrs_con *con, struct rtrs_iu *iu,
305				struct ib_sge *sge, unsigned int num_sge,
306				u32 rkey, u64 rdma_addr, u32 imm_data,
307				enum ib_send_flags flags,
308				struct ib_send_wr *head,
309				struct ib_send_wr *tail);
310
311int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe);
312
313int rtrs_cq_qp_create(struct rtrs_path *path, struct rtrs_con *con,
314		      u32 max_send_sge, int cq_vector, int nr_cqe,
315		      u32 max_send_wr, u32 max_recv_wr,
316		      enum ib_poll_context poll_ctx);
317void rtrs_cq_qp_destroy(struct rtrs_con *con);
318
319void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe,
320		  unsigned int interval_ms, unsigned int missed_max,
321		  void (*err_handler)(struct rtrs_con *con),
322		  struct workqueue_struct *wq);
323void rtrs_start_hb(struct rtrs_path *path);
324void rtrs_stop_hb(struct rtrs_path *path);
325void rtrs_send_hb_ack(struct rtrs_path *path);
326
327void rtrs_rdma_dev_pd_init(enum ib_pd_flags pd_flags,
328			   struct rtrs_rdma_dev_pd *pool);
329void rtrs_rdma_dev_pd_deinit(struct rtrs_rdma_dev_pd *pool);
330
331struct rtrs_ib_dev *rtrs_ib_dev_find_or_add(struct ib_device *ib_dev,
332					    struct rtrs_rdma_dev_pd *pool);
333int rtrs_ib_dev_put(struct rtrs_ib_dev *dev);
334
335static inline u32 rtrs_to_imm(u32 type, u32 payload)
336{
337	BUILD_BUG_ON(MAX_IMM_PAYL_BITS + MAX_IMM_TYPE_BITS != 32);
338	BUILD_BUG_ON(RTRS_LAST_IMM > (1<<MAX_IMM_TYPE_BITS));
339	return ((type & MAX_IMM_TYPE_MASK) << MAX_IMM_PAYL_BITS) |
340		(payload & MAX_IMM_PAYL_MASK);
341}
342
343static inline void rtrs_from_imm(u32 imm, u32 *type, u32 *payload)
344{
345	*payload = imm & MAX_IMM_PAYL_MASK;
346	*type = imm >> MAX_IMM_PAYL_BITS;
347}
348
349static inline u32 rtrs_to_io_req_imm(u32 addr)
350{
351	return rtrs_to_imm(RTRS_IO_REQ_IMM, addr);
352}
353
354static inline u32 rtrs_to_io_rsp_imm(u32 msg_id, int errno, bool w_inval)
355{
356	enum rtrs_imm_type type;
357	u32 payload;
358
359	/* 9 bits for errno, 19 bits for msg_id */
360	payload = (abs(errno) & 0x1ff) << 19 | (msg_id & 0x7ffff);
361	type = w_inval ? RTRS_IO_RSP_W_INV_IMM : RTRS_IO_RSP_IMM;
362
363	return rtrs_to_imm(type, payload);
364}
365
366static inline void rtrs_from_io_rsp_imm(u32 payload, u32 *msg_id, int *errno)
367{
368	/* 9 bits for errno, 19 bits for msg_id */
369	*msg_id = payload & 0x7ffff;
370	*errno = -(int)((payload >> 19) & 0x1ff);
371}
372
373#define STAT_STORE_FUNC(type, set_value, reset)				\
374static ssize_t set_value##_store(struct kobject *kobj,			\
375			     struct kobj_attribute *attr,		\
376			     const char *buf, size_t count)		\
377{									\
378	int ret = -EINVAL;						\
379	type *stats = container_of(kobj, type, kobj_stats);		\
380									\
381	if (sysfs_streq(buf, "1"))					\
382		ret = reset(stats, true);			\
383	else if (sysfs_streq(buf, "0"))					\
384		ret = reset(stats, false);			\
385	if (ret)							\
386		return ret;						\
387									\
388	return count;							\
389}
390
391#define STAT_SHOW_FUNC(type, get_value, print)				\
392static ssize_t get_value##_show(struct kobject *kobj,			\
393			   struct kobj_attribute *attr,			\
394			   char *page)					\
395{									\
396	type *stats = container_of(kobj, type, kobj_stats);		\
397									\
398	return print(stats, page);			\
399}
400
401#define STAT_ATTR(type, stat, print, reset)				\
402STAT_STORE_FUNC(type, stat, reset)					\
403STAT_SHOW_FUNC(type, stat, print)					\
404static struct kobj_attribute stat##_attr = __ATTR_RW(stat)
405
406#endif /* RTRS_PRI_H */