Linux Audio

Check our new training course

Loading...
v5.14.15
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Copyright (c) 2008-2009 Patrick McHardy <kaber@trash.net>
  4 *
  5 * Development of this code funded by Astaro AG (http://www.astaro.com/)
  6 */
  7
  8#include <linux/kernel.h>
  9#include <linux/init.h>
 10#include <linux/module.h>
 11#include <linux/netlink.h>
 12#include <linux/netfilter.h>
 13#include <linux/if_arp.h>
 14#include <linux/netfilter/nf_tables.h>
 15#include <net/netfilter/nf_tables_core.h>
 16#include <net/netfilter/nf_tables_offload.h>
 17#include <net/netfilter/nf_tables.h>
 18
 19struct nft_cmp_expr {
 20	struct nft_data		data;
 21	u8			sreg;
 22	u8			len;
 23	enum nft_cmp_ops	op:8;
 24};
 25
 26void nft_cmp_eval(const struct nft_expr *expr,
 27		  struct nft_regs *regs,
 28		  const struct nft_pktinfo *pkt)
 29{
 30	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
 31	int d;
 32
 33	d = memcmp(&regs->data[priv->sreg], &priv->data, priv->len);
 34	switch (priv->op) {
 35	case NFT_CMP_EQ:
 36		if (d != 0)
 37			goto mismatch;
 38		break;
 39	case NFT_CMP_NEQ:
 40		if (d == 0)
 41			goto mismatch;
 42		break;
 43	case NFT_CMP_LT:
 44		if (d == 0)
 45			goto mismatch;
 46		fallthrough;
 47	case NFT_CMP_LTE:
 48		if (d > 0)
 49			goto mismatch;
 50		break;
 51	case NFT_CMP_GT:
 52		if (d == 0)
 53			goto mismatch;
 54		fallthrough;
 55	case NFT_CMP_GTE:
 56		if (d < 0)
 57			goto mismatch;
 58		break;
 59	}
 60	return;
 61
 62mismatch:
 63	regs->verdict.code = NFT_BREAK;
 64}
 65
 66static const struct nla_policy nft_cmp_policy[NFTA_CMP_MAX + 1] = {
 67	[NFTA_CMP_SREG]		= { .type = NLA_U32 },
 68	[NFTA_CMP_OP]		= { .type = NLA_U32 },
 69	[NFTA_CMP_DATA]		= { .type = NLA_NESTED },
 70};
 71
 72static int nft_cmp_init(const struct nft_ctx *ctx, const struct nft_expr *expr,
 73			const struct nlattr * const tb[])
 74{
 75	struct nft_cmp_expr *priv = nft_expr_priv(expr);
 76	struct nft_data_desc desc;
 
 
 
 77	int err;
 78
 79	err = nft_data_init(NULL, &priv->data, sizeof(priv->data), &desc,
 80			    tb[NFTA_CMP_DATA]);
 81	if (err < 0)
 82		return err;
 83
 84	if (desc.type != NFT_DATA_VALUE) {
 85		err = -EINVAL;
 86		nft_data_release(&priv->data, desc.type);
 87		return err;
 88	}
 89
 90	err = nft_parse_register_load(tb[NFTA_CMP_SREG], &priv->sreg, desc.len);
 91	if (err < 0)
 92		return err;
 93
 94	priv->op  = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
 95	priv->len = desc.len;
 96	return 0;
 97}
 98
 99static int nft_cmp_dump(struct sk_buff *skb, const struct nft_expr *expr)
 
100{
101	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
102
103	if (nft_dump_register(skb, NFTA_CMP_SREG, priv->sreg))
104		goto nla_put_failure;
105	if (nla_put_be32(skb, NFTA_CMP_OP, htonl(priv->op)))
106		goto nla_put_failure;
107
108	if (nft_data_dump(skb, NFTA_CMP_DATA, &priv->data,
109			  NFT_DATA_VALUE, priv->len) < 0)
110		goto nla_put_failure;
111	return 0;
112
113nla_put_failure:
114	return -1;
115}
116
117union nft_cmp_offload_data {
118	u16	val16;
119	u32	val32;
120	u64	val64;
121};
122
123static void nft_payload_n2h(union nft_cmp_offload_data *data,
124			    const u8 *val, u32 len)
125{
126	switch (len) {
127	case 2:
128		data->val16 = ntohs(*((u16 *)val));
129		break;
130	case 4:
131		data->val32 = ntohl(*((u32 *)val));
132		break;
133	case 8:
134		data->val64 = be64_to_cpu(*((u64 *)val));
135		break;
136	default:
137		WARN_ON_ONCE(1);
138		break;
139	}
140}
141
142static int __nft_cmp_offload(struct nft_offload_ctx *ctx,
143			     struct nft_flow_rule *flow,
144			     const struct nft_cmp_expr *priv)
145{
146	struct nft_offload_reg *reg = &ctx->regs[priv->sreg];
147	union nft_cmp_offload_data _data, _datamask;
148	u8 *mask = (u8 *)&flow->match.mask;
149	u8 *key = (u8 *)&flow->match.key;
150	u8 *data, *datamask;
151
152	if (priv->op != NFT_CMP_EQ || priv->len > reg->len)
153		return -EOPNOTSUPP;
154
155	if (reg->flags & NFT_OFFLOAD_F_NETWORK2HOST) {
156		nft_payload_n2h(&_data, (u8 *)&priv->data, reg->len);
157		nft_payload_n2h(&_datamask, (u8 *)&reg->mask, reg->len);
158		data = (u8 *)&_data;
159		datamask = (u8 *)&_datamask;
160	} else {
161		data = (u8 *)&priv->data;
162		datamask = (u8 *)&reg->mask;
163	}
164
165	memcpy(key + reg->offset, data, reg->len);
166	memcpy(mask + reg->offset, datamask, reg->len);
167
168	flow->match.dissector.used_keys |= BIT(reg->key);
169	flow->match.dissector.offset[reg->key] = reg->base_offset;
170
171	if (reg->key == FLOW_DISSECTOR_KEY_META &&
172	    reg->offset == offsetof(struct nft_flow_key, meta.ingress_iftype) &&
173	    nft_reg_load16(priv->data.data) != ARPHRD_ETHER)
174		return -EOPNOTSUPP;
175
176	nft_offload_update_dependency(ctx, &priv->data, reg->len);
177
178	return 0;
179}
180
181static int nft_cmp_offload(struct nft_offload_ctx *ctx,
182			   struct nft_flow_rule *flow,
183			   const struct nft_expr *expr)
184{
185	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
186
187	return __nft_cmp_offload(ctx, flow, priv);
188}
189
190static const struct nft_expr_ops nft_cmp_ops = {
191	.type		= &nft_cmp_type,
192	.size		= NFT_EXPR_SIZE(sizeof(struct nft_cmp_expr)),
193	.eval		= nft_cmp_eval,
194	.init		= nft_cmp_init,
195	.dump		= nft_cmp_dump,
 
196	.offload	= nft_cmp_offload,
197};
198
 
 
 
 
 
 
 
 
 
 
 
 
199static int nft_cmp_fast_init(const struct nft_ctx *ctx,
200			     const struct nft_expr *expr,
201			     const struct nlattr * const tb[])
202{
203	struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
204	struct nft_data_desc desc;
205	struct nft_data data;
 
 
 
 
206	int err;
207
208	err = nft_data_init(NULL, &data, sizeof(data), &desc,
209			    tb[NFTA_CMP_DATA]);
210	if (err < 0)
211		return err;
212
213	err = nft_parse_register_load(tb[NFTA_CMP_SREG], &priv->sreg, desc.len);
214	if (err < 0)
215		return err;
216
217	desc.len *= BITS_PER_BYTE;
218
219	priv->mask = nft_cmp_fast_mask(desc.len);
220	priv->data = data.data[0] & priv->mask;
221	priv->len  = desc.len;
222	priv->inv  = ntohl(nla_get_be32(tb[NFTA_CMP_OP])) != NFT_CMP_EQ;
223	return 0;
224}
225
226static int nft_cmp_fast_offload(struct nft_offload_ctx *ctx,
227				struct nft_flow_rule *flow,
228				const struct nft_expr *expr)
229{
230	const struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
231	struct nft_cmp_expr cmp = {
232		.data	= {
233			.data	= {
234				[0] = priv->data,
235			},
236		},
237		.sreg	= priv->sreg,
238		.len	= priv->len / BITS_PER_BYTE,
239		.op	= priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ,
240	};
241
242	return __nft_cmp_offload(ctx, flow, &cmp);
243}
244
245static int nft_cmp_fast_dump(struct sk_buff *skb, const struct nft_expr *expr)
 
246{
247	const struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
248	enum nft_cmp_ops op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ;
249	struct nft_data data;
250
251	if (nft_dump_register(skb, NFTA_CMP_SREG, priv->sreg))
252		goto nla_put_failure;
253	if (nla_put_be32(skb, NFTA_CMP_OP, htonl(op)))
254		goto nla_put_failure;
255
256	data.data[0] = priv->data;
257	if (nft_data_dump(skb, NFTA_CMP_DATA, &data,
258			  NFT_DATA_VALUE, priv->len / BITS_PER_BYTE) < 0)
259		goto nla_put_failure;
260	return 0;
261
262nla_put_failure:
263	return -1;
264}
265
266const struct nft_expr_ops nft_cmp_fast_ops = {
267	.type		= &nft_cmp_type,
268	.size		= NFT_EXPR_SIZE(sizeof(struct nft_cmp_fast_expr)),
269	.eval		= NULL,	/* inlined */
270	.init		= nft_cmp_fast_init,
271	.dump		= nft_cmp_fast_dump,
 
272	.offload	= nft_cmp_fast_offload,
273};
274
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
275static const struct nft_expr_ops *
276nft_cmp_select_ops(const struct nft_ctx *ctx, const struct nlattr * const tb[])
277{
278	struct nft_data_desc desc;
279	struct nft_data data;
 
 
 
 
280	enum nft_cmp_ops op;
 
281	int err;
282
283	if (tb[NFTA_CMP_SREG] == NULL ||
284	    tb[NFTA_CMP_OP] == NULL ||
285	    tb[NFTA_CMP_DATA] == NULL)
286		return ERR_PTR(-EINVAL);
287
288	op = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
289	switch (op) {
290	case NFT_CMP_EQ:
291	case NFT_CMP_NEQ:
292	case NFT_CMP_LT:
293	case NFT_CMP_LTE:
294	case NFT_CMP_GT:
295	case NFT_CMP_GTE:
296		break;
297	default:
298		return ERR_PTR(-EINVAL);
299	}
300
301	err = nft_data_init(NULL, &data, sizeof(data), &desc,
302			    tb[NFTA_CMP_DATA]);
303	if (err < 0)
304		return ERR_PTR(err);
305
306	if (desc.type != NFT_DATA_VALUE)
307		goto err1;
308
309	if (desc.len <= sizeof(u32) && (op == NFT_CMP_EQ || op == NFT_CMP_NEQ))
310		return &nft_cmp_fast_ops;
311
 
 
 
 
 
 
 
 
312	return &nft_cmp_ops;
313err1:
314	nft_data_release(&data, desc.type);
315	return ERR_PTR(-EINVAL);
316}
317
318struct nft_expr_type nft_cmp_type __read_mostly = {
319	.name		= "cmp",
320	.select_ops	= nft_cmp_select_ops,
321	.policy		= nft_cmp_policy,
322	.maxattr	= NFTA_CMP_MAX,
323	.owner		= THIS_MODULE,
324};
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Copyright (c) 2008-2009 Patrick McHardy <kaber@trash.net>
  4 *
  5 * Development of this code funded by Astaro AG (http://www.astaro.com/)
  6 */
  7
  8#include <linux/kernel.h>
  9#include <linux/init.h>
 10#include <linux/module.h>
 11#include <linux/netlink.h>
 12#include <linux/netfilter.h>
 13#include <linux/if_arp.h>
 14#include <linux/netfilter/nf_tables.h>
 15#include <net/netfilter/nf_tables_core.h>
 16#include <net/netfilter/nf_tables_offload.h>
 17#include <net/netfilter/nf_tables.h>
 18
 19struct nft_cmp_expr {
 20	struct nft_data		data;
 21	u8			sreg;
 22	u8			len;
 23	enum nft_cmp_ops	op:8;
 24};
 25
 26void nft_cmp_eval(const struct nft_expr *expr,
 27		  struct nft_regs *regs,
 28		  const struct nft_pktinfo *pkt)
 29{
 30	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
 31	int d;
 32
 33	d = memcmp(&regs->data[priv->sreg], &priv->data, priv->len);
 34	switch (priv->op) {
 35	case NFT_CMP_EQ:
 36		if (d != 0)
 37			goto mismatch;
 38		break;
 39	case NFT_CMP_NEQ:
 40		if (d == 0)
 41			goto mismatch;
 42		break;
 43	case NFT_CMP_LT:
 44		if (d == 0)
 45			goto mismatch;
 46		fallthrough;
 47	case NFT_CMP_LTE:
 48		if (d > 0)
 49			goto mismatch;
 50		break;
 51	case NFT_CMP_GT:
 52		if (d == 0)
 53			goto mismatch;
 54		fallthrough;
 55	case NFT_CMP_GTE:
 56		if (d < 0)
 57			goto mismatch;
 58		break;
 59	}
 60	return;
 61
 62mismatch:
 63	regs->verdict.code = NFT_BREAK;
 64}
 65
 66static const struct nla_policy nft_cmp_policy[NFTA_CMP_MAX + 1] = {
 67	[NFTA_CMP_SREG]		= { .type = NLA_U32 },
 68	[NFTA_CMP_OP]		= { .type = NLA_U32 },
 69	[NFTA_CMP_DATA]		= { .type = NLA_NESTED },
 70};
 71
 72static int nft_cmp_init(const struct nft_ctx *ctx, const struct nft_expr *expr,
 73			const struct nlattr * const tb[])
 74{
 75	struct nft_cmp_expr *priv = nft_expr_priv(expr);
 76	struct nft_data_desc desc = {
 77		.type	= NFT_DATA_VALUE,
 78		.size	= sizeof(priv->data),
 79	};
 80	int err;
 81
 82	err = nft_data_init(NULL, &priv->data, &desc, tb[NFTA_CMP_DATA]);
 
 83	if (err < 0)
 84		return err;
 85
 86	err = nft_parse_register_load(ctx, tb[NFTA_CMP_SREG], &priv->sreg, desc.len);
 
 
 
 
 
 
 87	if (err < 0)
 88		return err;
 89
 90	priv->op  = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
 91	priv->len = desc.len;
 92	return 0;
 93}
 94
 95static int nft_cmp_dump(struct sk_buff *skb,
 96			const struct nft_expr *expr, bool reset)
 97{
 98	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
 99
100	if (nft_dump_register(skb, NFTA_CMP_SREG, priv->sreg))
101		goto nla_put_failure;
102	if (nla_put_be32(skb, NFTA_CMP_OP, htonl(priv->op)))
103		goto nla_put_failure;
104
105	if (nft_data_dump(skb, NFTA_CMP_DATA, &priv->data,
106			  NFT_DATA_VALUE, priv->len) < 0)
107		goto nla_put_failure;
108	return 0;
109
110nla_put_failure:
111	return -1;
112}
113
114union nft_cmp_offload_data {
115	u16	val16;
116	u32	val32;
117	u64	val64;
118};
119
120static void nft_payload_n2h(union nft_cmp_offload_data *data,
121			    const u8 *val, u32 len)
122{
123	switch (len) {
124	case 2:
125		data->val16 = ntohs(*((__be16 *)val));
126		break;
127	case 4:
128		data->val32 = ntohl(*((__be32 *)val));
129		break;
130	case 8:
131		data->val64 = be64_to_cpu(*((__be64 *)val));
132		break;
133	default:
134		WARN_ON_ONCE(1);
135		break;
136	}
137}
138
139static int __nft_cmp_offload(struct nft_offload_ctx *ctx,
140			     struct nft_flow_rule *flow,
141			     const struct nft_cmp_expr *priv)
142{
143	struct nft_offload_reg *reg = &ctx->regs[priv->sreg];
144	union nft_cmp_offload_data _data, _datamask;
145	u8 *mask = (u8 *)&flow->match.mask;
146	u8 *key = (u8 *)&flow->match.key;
147	u8 *data, *datamask;
148
149	if (priv->op != NFT_CMP_EQ || priv->len > reg->len)
150		return -EOPNOTSUPP;
151
152	if (reg->flags & NFT_OFFLOAD_F_NETWORK2HOST) {
153		nft_payload_n2h(&_data, (u8 *)&priv->data, reg->len);
154		nft_payload_n2h(&_datamask, (u8 *)&reg->mask, reg->len);
155		data = (u8 *)&_data;
156		datamask = (u8 *)&_datamask;
157	} else {
158		data = (u8 *)&priv->data;
159		datamask = (u8 *)&reg->mask;
160	}
161
162	memcpy(key + reg->offset, data, reg->len);
163	memcpy(mask + reg->offset, datamask, reg->len);
164
165	flow->match.dissector.used_keys |= BIT_ULL(reg->key);
166	flow->match.dissector.offset[reg->key] = reg->base_offset;
167
168	if (reg->key == FLOW_DISSECTOR_KEY_META &&
169	    reg->offset == offsetof(struct nft_flow_key, meta.ingress_iftype) &&
170	    nft_reg_load16(priv->data.data) != ARPHRD_ETHER)
171		return -EOPNOTSUPP;
172
173	nft_offload_update_dependency(ctx, &priv->data, reg->len);
174
175	return 0;
176}
177
178static int nft_cmp_offload(struct nft_offload_ctx *ctx,
179			   struct nft_flow_rule *flow,
180			   const struct nft_expr *expr)
181{
182	const struct nft_cmp_expr *priv = nft_expr_priv(expr);
183
184	return __nft_cmp_offload(ctx, flow, priv);
185}
186
187static const struct nft_expr_ops nft_cmp_ops = {
188	.type		= &nft_cmp_type,
189	.size		= NFT_EXPR_SIZE(sizeof(struct nft_cmp_expr)),
190	.eval		= nft_cmp_eval,
191	.init		= nft_cmp_init,
192	.dump		= nft_cmp_dump,
193	.reduce		= NFT_REDUCE_READONLY,
194	.offload	= nft_cmp_offload,
195};
196
197/* Calculate the mask for the nft_cmp_fast expression. On big endian the
198 * mask needs to include the *upper* bytes when interpreting that data as
199 * something smaller than the full u32, therefore a cpu_to_le32 is done.
200 */
201static u32 nft_cmp_fast_mask(unsigned int len)
202{
203	__le32 mask = cpu_to_le32(~0U >> (sizeof_field(struct nft_cmp_fast_expr,
204					  data) * BITS_PER_BYTE - len));
205
206	return (__force u32)mask;
207}
208
209static int nft_cmp_fast_init(const struct nft_ctx *ctx,
210			     const struct nft_expr *expr,
211			     const struct nlattr * const tb[])
212{
213	struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
 
214	struct nft_data data;
215	struct nft_data_desc desc = {
216		.type	= NFT_DATA_VALUE,
217		.size	= sizeof(data),
218	};
219	int err;
220
221	err = nft_data_init(NULL, &data, &desc, tb[NFTA_CMP_DATA]);
 
222	if (err < 0)
223		return err;
224
225	err = nft_parse_register_load(ctx, tb[NFTA_CMP_SREG], &priv->sreg, desc.len);
226	if (err < 0)
227		return err;
228
229	desc.len *= BITS_PER_BYTE;
230
231	priv->mask = nft_cmp_fast_mask(desc.len);
232	priv->data = data.data[0] & priv->mask;
233	priv->len  = desc.len;
234	priv->inv  = ntohl(nla_get_be32(tb[NFTA_CMP_OP])) != NFT_CMP_EQ;
235	return 0;
236}
237
238static int nft_cmp_fast_offload(struct nft_offload_ctx *ctx,
239				struct nft_flow_rule *flow,
240				const struct nft_expr *expr)
241{
242	const struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
243	struct nft_cmp_expr cmp = {
244		.data	= {
245			.data	= {
246				[0] = priv->data,
247			},
248		},
249		.sreg	= priv->sreg,
250		.len	= priv->len / BITS_PER_BYTE,
251		.op	= priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ,
252	};
253
254	return __nft_cmp_offload(ctx, flow, &cmp);
255}
256
257static int nft_cmp_fast_dump(struct sk_buff *skb,
258			     const struct nft_expr *expr, bool reset)
259{
260	const struct nft_cmp_fast_expr *priv = nft_expr_priv(expr);
261	enum nft_cmp_ops op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ;
262	struct nft_data data;
263
264	if (nft_dump_register(skb, NFTA_CMP_SREG, priv->sreg))
265		goto nla_put_failure;
266	if (nla_put_be32(skb, NFTA_CMP_OP, htonl(op)))
267		goto nla_put_failure;
268
269	data.data[0] = priv->data;
270	if (nft_data_dump(skb, NFTA_CMP_DATA, &data,
271			  NFT_DATA_VALUE, priv->len / BITS_PER_BYTE) < 0)
272		goto nla_put_failure;
273	return 0;
274
275nla_put_failure:
276	return -1;
277}
278
279const struct nft_expr_ops nft_cmp_fast_ops = {
280	.type		= &nft_cmp_type,
281	.size		= NFT_EXPR_SIZE(sizeof(struct nft_cmp_fast_expr)),
282	.eval		= NULL,	/* inlined */
283	.init		= nft_cmp_fast_init,
284	.dump		= nft_cmp_fast_dump,
285	.reduce		= NFT_REDUCE_READONLY,
286	.offload	= nft_cmp_fast_offload,
287};
288
289static u32 nft_cmp_mask(u32 bitlen)
290{
291	return (__force u32)cpu_to_le32(~0U >> (sizeof(u32) * BITS_PER_BYTE - bitlen));
292}
293
294static void nft_cmp16_fast_mask(struct nft_data *data, unsigned int bitlen)
295{
296	int len = bitlen / BITS_PER_BYTE;
297	int i, words = len / sizeof(u32);
298
299	for (i = 0; i < words; i++) {
300		data->data[i] = 0xffffffff;
301		bitlen -= sizeof(u32) * BITS_PER_BYTE;
302	}
303
304	if (len % sizeof(u32))
305		data->data[i++] = nft_cmp_mask(bitlen);
306
307	for (; i < 4; i++)
308		data->data[i] = 0;
309}
310
311static int nft_cmp16_fast_init(const struct nft_ctx *ctx,
312			       const struct nft_expr *expr,
313			       const struct nlattr * const tb[])
314{
315	struct nft_cmp16_fast_expr *priv = nft_expr_priv(expr);
316	struct nft_data_desc desc = {
317		.type	= NFT_DATA_VALUE,
318		.size	= sizeof(priv->data),
319	};
320	int err;
321
322	err = nft_data_init(NULL, &priv->data, &desc, tb[NFTA_CMP_DATA]);
323	if (err < 0)
324		return err;
325
326	err = nft_parse_register_load(ctx, tb[NFTA_CMP_SREG], &priv->sreg, desc.len);
327	if (err < 0)
328		return err;
329
330	nft_cmp16_fast_mask(&priv->mask, desc.len * BITS_PER_BYTE);
331	priv->inv = ntohl(nla_get_be32(tb[NFTA_CMP_OP])) != NFT_CMP_EQ;
332	priv->len = desc.len;
333
334	return 0;
335}
336
337static int nft_cmp16_fast_offload(struct nft_offload_ctx *ctx,
338				  struct nft_flow_rule *flow,
339				  const struct nft_expr *expr)
340{
341	const struct nft_cmp16_fast_expr *priv = nft_expr_priv(expr);
342	struct nft_cmp_expr cmp = {
343		.data	= priv->data,
344		.sreg	= priv->sreg,
345		.len	= priv->len,
346		.op	= priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ,
347	};
348
349	return __nft_cmp_offload(ctx, flow, &cmp);
350}
351
352static int nft_cmp16_fast_dump(struct sk_buff *skb,
353			       const struct nft_expr *expr, bool reset)
354{
355	const struct nft_cmp16_fast_expr *priv = nft_expr_priv(expr);
356	enum nft_cmp_ops op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ;
357
358	if (nft_dump_register(skb, NFTA_CMP_SREG, priv->sreg))
359		goto nla_put_failure;
360	if (nla_put_be32(skb, NFTA_CMP_OP, htonl(op)))
361		goto nla_put_failure;
362
363	if (nft_data_dump(skb, NFTA_CMP_DATA, &priv->data,
364			  NFT_DATA_VALUE, priv->len) < 0)
365		goto nla_put_failure;
366	return 0;
367
368nla_put_failure:
369	return -1;
370}
371
372
373const struct nft_expr_ops nft_cmp16_fast_ops = {
374	.type		= &nft_cmp_type,
375	.size		= NFT_EXPR_SIZE(sizeof(struct nft_cmp16_fast_expr)),
376	.eval		= NULL,	/* inlined */
377	.init		= nft_cmp16_fast_init,
378	.dump		= nft_cmp16_fast_dump,
379	.reduce		= NFT_REDUCE_READONLY,
380	.offload	= nft_cmp16_fast_offload,
381};
382
383static const struct nft_expr_ops *
384nft_cmp_select_ops(const struct nft_ctx *ctx, const struct nlattr * const tb[])
385{
 
386	struct nft_data data;
387	struct nft_data_desc desc = {
388		.type	= NFT_DATA_VALUE,
389		.size	= sizeof(data),
390	};
391	enum nft_cmp_ops op;
392	u8 sreg;
393	int err;
394
395	if (tb[NFTA_CMP_SREG] == NULL ||
396	    tb[NFTA_CMP_OP] == NULL ||
397	    tb[NFTA_CMP_DATA] == NULL)
398		return ERR_PTR(-EINVAL);
399
400	op = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
401	switch (op) {
402	case NFT_CMP_EQ:
403	case NFT_CMP_NEQ:
404	case NFT_CMP_LT:
405	case NFT_CMP_LTE:
406	case NFT_CMP_GT:
407	case NFT_CMP_GTE:
408		break;
409	default:
410		return ERR_PTR(-EINVAL);
411	}
412
413	err = nft_data_init(NULL, &data, &desc, tb[NFTA_CMP_DATA]);
 
414	if (err < 0)
415		return ERR_PTR(err);
416
417	sreg = ntohl(nla_get_be32(tb[NFTA_CMP_SREG]));
 
 
 
 
418
419	if (op == NFT_CMP_EQ || op == NFT_CMP_NEQ) {
420		if (desc.len <= sizeof(u32))
421			return &nft_cmp_fast_ops;
422		else if (desc.len <= sizeof(data) &&
423			 ((sreg >= NFT_REG_1 && sreg <= NFT_REG_4) ||
424			  (sreg >= NFT_REG32_00 && sreg <= NFT_REG32_12 && sreg % 2 == 0)))
425			return &nft_cmp16_fast_ops;
426	}
427	return &nft_cmp_ops;
 
 
 
428}
429
430struct nft_expr_type nft_cmp_type __read_mostly = {
431	.name		= "cmp",
432	.select_ops	= nft_cmp_select_ops,
433	.policy		= nft_cmp_policy,
434	.maxattr	= NFTA_CMP_MAX,
435	.owner		= THIS_MODULE,
436};