Loading...
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * INET An implementation of the TCP Authentication Option (TCP-AO).
4 * See RFC5925.
5 *
6 * Authors: Dmitry Safonov <dima@arista.com>
7 * Francesco Ruggeri <fruggeri@arista.com>
8 * Salam Noureddine <noureddine@arista.com>
9 */
10#define pr_fmt(fmt) "TCP: " fmt
11
12#include <crypto/hash.h>
13#include <linux/inetdevice.h>
14#include <linux/tcp.h>
15
16#include <net/tcp.h>
17#include <net/ipv6.h>
18#include <net/icmp.h>
19#include <trace/events/tcp.h>
20
21DEFINE_STATIC_KEY_DEFERRED_FALSE(tcp_ao_needed, HZ);
22
23int tcp_ao_calc_traffic_key(struct tcp_ao_key *mkt, u8 *key, void *ctx,
24 unsigned int len, struct tcp_sigpool *hp)
25{
26 struct scatterlist sg;
27 int ret;
28
29 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp->req),
30 mkt->key, mkt->keylen))
31 goto clear_hash;
32
33 ret = crypto_ahash_init(hp->req);
34 if (ret)
35 goto clear_hash;
36
37 sg_init_one(&sg, ctx, len);
38 ahash_request_set_crypt(hp->req, &sg, key, len);
39 crypto_ahash_update(hp->req);
40
41 ret = crypto_ahash_final(hp->req);
42 if (ret)
43 goto clear_hash;
44
45 return 0;
46clear_hash:
47 memset(key, 0, tcp_ao_digest_size(mkt));
48 return 1;
49}
50
51bool tcp_ao_ignore_icmp(const struct sock *sk, int family, int type, int code)
52{
53 bool ignore_icmp = false;
54 struct tcp_ao_info *ao;
55
56 if (!static_branch_unlikely(&tcp_ao_needed.key))
57 return false;
58
59 /* RFC5925, 7.8:
60 * >> A TCP-AO implementation MUST default to ignore incoming ICMPv4
61 * messages of Type 3 (destination unreachable), Codes 2-4 (protocol
62 * unreachable, port unreachable, and fragmentation needed -- ’hard
63 * errors’), and ICMPv6 Type 1 (destination unreachable), Code 1
64 * (administratively prohibited) and Code 4 (port unreachable) intended
65 * for connections in synchronized states (ESTABLISHED, FIN-WAIT-1, FIN-
66 * WAIT-2, CLOSE-WAIT, CLOSING, LAST-ACK, TIME-WAIT) that match MKTs.
67 */
68 if (family == AF_INET) {
69 if (type != ICMP_DEST_UNREACH)
70 return false;
71 if (code < ICMP_PROT_UNREACH || code > ICMP_FRAG_NEEDED)
72 return false;
73 } else {
74 if (type != ICMPV6_DEST_UNREACH)
75 return false;
76 if (code != ICMPV6_ADM_PROHIBITED && code != ICMPV6_PORT_UNREACH)
77 return false;
78 }
79
80 rcu_read_lock();
81 switch (sk->sk_state) {
82 case TCP_TIME_WAIT:
83 ao = rcu_dereference(tcp_twsk(sk)->ao_info);
84 break;
85 case TCP_SYN_SENT:
86 case TCP_SYN_RECV:
87 case TCP_LISTEN:
88 case TCP_NEW_SYN_RECV:
89 /* RFC5925 specifies to ignore ICMPs *only* on connections
90 * in synchronized states.
91 */
92 rcu_read_unlock();
93 return false;
94 default:
95 ao = rcu_dereference(tcp_sk(sk)->ao_info);
96 }
97
98 if (ao && !ao->accept_icmps) {
99 ignore_icmp = true;
100 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAODROPPEDICMPS);
101 atomic64_inc(&ao->counters.dropped_icmp);
102 }
103 rcu_read_unlock();
104
105 return ignore_icmp;
106}
107
108/* Optimized version of tcp_ao_do_lookup(): only for sockets for which
109 * it's known that the keys in ao_info are matching peer's
110 * family/address/VRF/etc.
111 */
112struct tcp_ao_key *tcp_ao_established_key(const struct sock *sk,
113 struct tcp_ao_info *ao,
114 int sndid, int rcvid)
115{
116 struct tcp_ao_key *key;
117
118 hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk)) {
119 if ((sndid >= 0 && key->sndid != sndid) ||
120 (rcvid >= 0 && key->rcvid != rcvid))
121 continue;
122 return key;
123 }
124
125 return NULL;
126}
127
128static int ipv4_prefix_cmp(const struct in_addr *addr1,
129 const struct in_addr *addr2,
130 unsigned int prefixlen)
131{
132 __be32 mask = inet_make_mask(prefixlen);
133 __be32 a1 = addr1->s_addr & mask;
134 __be32 a2 = addr2->s_addr & mask;
135
136 if (a1 == a2)
137 return 0;
138 return memcmp(&a1, &a2, sizeof(a1));
139}
140
141static int __tcp_ao_key_cmp(const struct tcp_ao_key *key, int l3index,
142 const union tcp_ao_addr *addr, u8 prefixlen,
143 int family, int sndid, int rcvid)
144{
145 if (sndid >= 0 && key->sndid != sndid)
146 return (key->sndid > sndid) ? 1 : -1;
147 if (rcvid >= 0 && key->rcvid != rcvid)
148 return (key->rcvid > rcvid) ? 1 : -1;
149 if (l3index >= 0 && (key->keyflags & TCP_AO_KEYF_IFINDEX)) {
150 if (key->l3index != l3index)
151 return (key->l3index > l3index) ? 1 : -1;
152 }
153
154 if (family == AF_UNSPEC)
155 return 0;
156 if (key->family != family)
157 return (key->family > family) ? 1 : -1;
158
159 if (family == AF_INET) {
160 if (ntohl(key->addr.a4.s_addr) == INADDR_ANY)
161 return 0;
162 if (ntohl(addr->a4.s_addr) == INADDR_ANY)
163 return 0;
164 return ipv4_prefix_cmp(&key->addr.a4, &addr->a4, prefixlen);
165#if IS_ENABLED(CONFIG_IPV6)
166 } else {
167 if (ipv6_addr_any(&key->addr.a6) || ipv6_addr_any(&addr->a6))
168 return 0;
169 if (ipv6_prefix_equal(&key->addr.a6, &addr->a6, prefixlen))
170 return 0;
171 return memcmp(&key->addr.a6, &addr->a6, sizeof(addr->a6));
172#endif
173 }
174 return -1;
175}
176
177static int tcp_ao_key_cmp(const struct tcp_ao_key *key, int l3index,
178 const union tcp_ao_addr *addr, u8 prefixlen,
179 int family, int sndid, int rcvid)
180{
181#if IS_ENABLED(CONFIG_IPV6)
182 if (family == AF_INET6 && ipv6_addr_v4mapped(&addr->a6)) {
183 __be32 addr4 = addr->a6.s6_addr32[3];
184
185 return __tcp_ao_key_cmp(key, l3index,
186 (union tcp_ao_addr *)&addr4,
187 prefixlen, AF_INET, sndid, rcvid);
188 }
189#endif
190 return __tcp_ao_key_cmp(key, l3index, addr,
191 prefixlen, family, sndid, rcvid);
192}
193
194static struct tcp_ao_key *__tcp_ao_do_lookup(const struct sock *sk, int l3index,
195 const union tcp_ao_addr *addr, int family, u8 prefix,
196 int sndid, int rcvid)
197{
198 struct tcp_ao_key *key;
199 struct tcp_ao_info *ao;
200
201 if (!static_branch_unlikely(&tcp_ao_needed.key))
202 return NULL;
203
204 ao = rcu_dereference_check(tcp_sk(sk)->ao_info,
205 lockdep_sock_is_held(sk));
206 if (!ao)
207 return NULL;
208
209 hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk)) {
210 u8 prefixlen = min(prefix, key->prefixlen);
211
212 if (!tcp_ao_key_cmp(key, l3index, addr, prefixlen,
213 family, sndid, rcvid))
214 return key;
215 }
216 return NULL;
217}
218
219struct tcp_ao_key *tcp_ao_do_lookup(const struct sock *sk, int l3index,
220 const union tcp_ao_addr *addr,
221 int family, int sndid, int rcvid)
222{
223 return __tcp_ao_do_lookup(sk, l3index, addr, family, U8_MAX, sndid, rcvid);
224}
225
226static struct tcp_ao_info *tcp_ao_alloc_info(gfp_t flags)
227{
228 struct tcp_ao_info *ao;
229
230 ao = kzalloc(sizeof(*ao), flags);
231 if (!ao)
232 return NULL;
233 INIT_HLIST_HEAD(&ao->head);
234 refcount_set(&ao->refcnt, 1);
235
236 return ao;
237}
238
239static void tcp_ao_link_mkt(struct tcp_ao_info *ao, struct tcp_ao_key *mkt)
240{
241 hlist_add_head_rcu(&mkt->node, &ao->head);
242}
243
244static struct tcp_ao_key *tcp_ao_copy_key(struct sock *sk,
245 struct tcp_ao_key *key)
246{
247 struct tcp_ao_key *new_key;
248
249 new_key = sock_kmalloc(sk, tcp_ao_sizeof_key(key),
250 GFP_ATOMIC);
251 if (!new_key)
252 return NULL;
253
254 *new_key = *key;
255 INIT_HLIST_NODE(&new_key->node);
256 tcp_sigpool_get(new_key->tcp_sigpool_id);
257 atomic64_set(&new_key->pkt_good, 0);
258 atomic64_set(&new_key->pkt_bad, 0);
259
260 return new_key;
261}
262
263static void tcp_ao_key_free_rcu(struct rcu_head *head)
264{
265 struct tcp_ao_key *key = container_of(head, struct tcp_ao_key, rcu);
266
267 tcp_sigpool_release(key->tcp_sigpool_id);
268 kfree_sensitive(key);
269}
270
271static void tcp_ao_info_free_rcu(struct rcu_head *head)
272{
273 struct tcp_ao_info *ao = container_of(head, struct tcp_ao_info, rcu);
274 struct tcp_ao_key *key;
275 struct hlist_node *n;
276
277 hlist_for_each_entry_safe(key, n, &ao->head, node) {
278 hlist_del(&key->node);
279 tcp_sigpool_release(key->tcp_sigpool_id);
280 kfree_sensitive(key);
281 }
282 kfree(ao);
283 static_branch_slow_dec_deferred(&tcp_ao_needed);
284}
285
286static void tcp_ao_sk_omem_free(struct sock *sk, struct tcp_ao_info *ao)
287{
288 size_t total_ao_sk_mem = 0;
289 struct tcp_ao_key *key;
290
291 hlist_for_each_entry(key, &ao->head, node)
292 total_ao_sk_mem += tcp_ao_sizeof_key(key);
293 atomic_sub(total_ao_sk_mem, &sk->sk_omem_alloc);
294}
295
296void tcp_ao_destroy_sock(struct sock *sk, bool twsk)
297{
298 struct tcp_ao_info *ao;
299
300 if (twsk) {
301 ao = rcu_dereference_protected(tcp_twsk(sk)->ao_info, 1);
302 rcu_assign_pointer(tcp_twsk(sk)->ao_info, NULL);
303 } else {
304 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info, 1);
305 rcu_assign_pointer(tcp_sk(sk)->ao_info, NULL);
306 }
307
308 if (!ao || !refcount_dec_and_test(&ao->refcnt))
309 return;
310
311 if (!twsk)
312 tcp_ao_sk_omem_free(sk, ao);
313 call_rcu(&ao->rcu, tcp_ao_info_free_rcu);
314}
315
316void tcp_ao_time_wait(struct tcp_timewait_sock *tcptw, struct tcp_sock *tp)
317{
318 struct tcp_ao_info *ao_info = rcu_dereference_protected(tp->ao_info, 1);
319
320 if (ao_info) {
321 struct tcp_ao_key *key;
322 struct hlist_node *n;
323 int omem = 0;
324
325 hlist_for_each_entry_safe(key, n, &ao_info->head, node) {
326 omem += tcp_ao_sizeof_key(key);
327 }
328
329 refcount_inc(&ao_info->refcnt);
330 atomic_sub(omem, &(((struct sock *)tp)->sk_omem_alloc));
331 rcu_assign_pointer(tcptw->ao_info, ao_info);
332 } else {
333 tcptw->ao_info = NULL;
334 }
335}
336
337/* 4 tuple and ISNs are expected in NBO */
338static int tcp_v4_ao_calc_key(struct tcp_ao_key *mkt, u8 *key,
339 __be32 saddr, __be32 daddr,
340 __be16 sport, __be16 dport,
341 __be32 sisn, __be32 disn)
342{
343 /* See RFC5926 3.1.1 */
344 struct kdf_input_block {
345 u8 counter;
346 u8 label[6];
347 struct tcp4_ao_context ctx;
348 __be16 outlen;
349 } __packed * tmp;
350 struct tcp_sigpool hp;
351 int err;
352
353 err = tcp_sigpool_start(mkt->tcp_sigpool_id, &hp);
354 if (err)
355 return err;
356
357 tmp = hp.scratch;
358 tmp->counter = 1;
359 memcpy(tmp->label, "TCP-AO", 6);
360 tmp->ctx.saddr = saddr;
361 tmp->ctx.daddr = daddr;
362 tmp->ctx.sport = sport;
363 tmp->ctx.dport = dport;
364 tmp->ctx.sisn = sisn;
365 tmp->ctx.disn = disn;
366 tmp->outlen = htons(tcp_ao_digest_size(mkt) * 8); /* in bits */
367
368 err = tcp_ao_calc_traffic_key(mkt, key, tmp, sizeof(*tmp), &hp);
369 tcp_sigpool_end(&hp);
370
371 return err;
372}
373
374int tcp_v4_ao_calc_key_sk(struct tcp_ao_key *mkt, u8 *key,
375 const struct sock *sk,
376 __be32 sisn, __be32 disn, bool send)
377{
378 if (send)
379 return tcp_v4_ao_calc_key(mkt, key, sk->sk_rcv_saddr,
380 sk->sk_daddr, htons(sk->sk_num),
381 sk->sk_dport, sisn, disn);
382 else
383 return tcp_v4_ao_calc_key(mkt, key, sk->sk_daddr,
384 sk->sk_rcv_saddr, sk->sk_dport,
385 htons(sk->sk_num), disn, sisn);
386}
387
388static int tcp_ao_calc_key_sk(struct tcp_ao_key *mkt, u8 *key,
389 const struct sock *sk,
390 __be32 sisn, __be32 disn, bool send)
391{
392 if (mkt->family == AF_INET)
393 return tcp_v4_ao_calc_key_sk(mkt, key, sk, sisn, disn, send);
394#if IS_ENABLED(CONFIG_IPV6)
395 else if (mkt->family == AF_INET6)
396 return tcp_v6_ao_calc_key_sk(mkt, key, sk, sisn, disn, send);
397#endif
398 else
399 return -EOPNOTSUPP;
400}
401
402int tcp_v4_ao_calc_key_rsk(struct tcp_ao_key *mkt, u8 *key,
403 struct request_sock *req)
404{
405 struct inet_request_sock *ireq = inet_rsk(req);
406
407 return tcp_v4_ao_calc_key(mkt, key,
408 ireq->ir_loc_addr, ireq->ir_rmt_addr,
409 htons(ireq->ir_num), ireq->ir_rmt_port,
410 htonl(tcp_rsk(req)->snt_isn),
411 htonl(tcp_rsk(req)->rcv_isn));
412}
413
414static int tcp_v4_ao_calc_key_skb(struct tcp_ao_key *mkt, u8 *key,
415 const struct sk_buff *skb,
416 __be32 sisn, __be32 disn)
417{
418 const struct iphdr *iph = ip_hdr(skb);
419 const struct tcphdr *th = tcp_hdr(skb);
420
421 return tcp_v4_ao_calc_key(mkt, key, iph->saddr, iph->daddr,
422 th->source, th->dest, sisn, disn);
423}
424
425static int tcp_ao_calc_key_skb(struct tcp_ao_key *mkt, u8 *key,
426 const struct sk_buff *skb,
427 __be32 sisn, __be32 disn, int family)
428{
429 if (family == AF_INET)
430 return tcp_v4_ao_calc_key_skb(mkt, key, skb, sisn, disn);
431#if IS_ENABLED(CONFIG_IPV6)
432 else if (family == AF_INET6)
433 return tcp_v6_ao_calc_key_skb(mkt, key, skb, sisn, disn);
434#endif
435 return -EAFNOSUPPORT;
436}
437
438static int tcp_v4_ao_hash_pseudoheader(struct tcp_sigpool *hp,
439 __be32 daddr, __be32 saddr,
440 int nbytes)
441{
442 struct tcp4_pseudohdr *bp;
443 struct scatterlist sg;
444
445 bp = hp->scratch;
446 bp->saddr = saddr;
447 bp->daddr = daddr;
448 bp->pad = 0;
449 bp->protocol = IPPROTO_TCP;
450 bp->len = cpu_to_be16(nbytes);
451
452 sg_init_one(&sg, bp, sizeof(*bp));
453 ahash_request_set_crypt(hp->req, &sg, NULL, sizeof(*bp));
454 return crypto_ahash_update(hp->req);
455}
456
457static int tcp_ao_hash_pseudoheader(unsigned short int family,
458 const struct sock *sk,
459 const struct sk_buff *skb,
460 struct tcp_sigpool *hp, int nbytes)
461{
462 const struct tcphdr *th = tcp_hdr(skb);
463
464 /* TODO: Can we rely on checksum being zero to mean outbound pkt? */
465 if (!th->check) {
466 if (family == AF_INET)
467 return tcp_v4_ao_hash_pseudoheader(hp, sk->sk_daddr,
468 sk->sk_rcv_saddr, skb->len);
469#if IS_ENABLED(CONFIG_IPV6)
470 else if (family == AF_INET6)
471 return tcp_v6_ao_hash_pseudoheader(hp, &sk->sk_v6_daddr,
472 &sk->sk_v6_rcv_saddr, skb->len);
473#endif
474 else
475 return -EAFNOSUPPORT;
476 }
477
478 if (family == AF_INET) {
479 const struct iphdr *iph = ip_hdr(skb);
480
481 return tcp_v4_ao_hash_pseudoheader(hp, iph->daddr,
482 iph->saddr, skb->len);
483#if IS_ENABLED(CONFIG_IPV6)
484 } else if (family == AF_INET6) {
485 const struct ipv6hdr *iph = ipv6_hdr(skb);
486
487 return tcp_v6_ao_hash_pseudoheader(hp, &iph->daddr,
488 &iph->saddr, skb->len);
489#endif
490 }
491 return -EAFNOSUPPORT;
492}
493
494u32 tcp_ao_compute_sne(u32 next_sne, u32 next_seq, u32 seq)
495{
496 u32 sne = next_sne;
497
498 if (before(seq, next_seq)) {
499 if (seq > next_seq)
500 sne--;
501 } else {
502 if (seq < next_seq)
503 sne++;
504 }
505
506 return sne;
507}
508
509/* tcp_ao_hash_sne(struct tcp_sigpool *hp)
510 * @hp - used for hashing
511 * @sne - sne value
512 */
513static int tcp_ao_hash_sne(struct tcp_sigpool *hp, u32 sne)
514{
515 struct scatterlist sg;
516 __be32 *bp;
517
518 bp = (__be32 *)hp->scratch;
519 *bp = htonl(sne);
520
521 sg_init_one(&sg, bp, sizeof(*bp));
522 ahash_request_set_crypt(hp->req, &sg, NULL, sizeof(*bp));
523 return crypto_ahash_update(hp->req);
524}
525
526static int tcp_ao_hash_header(struct tcp_sigpool *hp,
527 const struct tcphdr *th,
528 bool exclude_options, u8 *hash,
529 int hash_offset, int hash_len)
530{
531 struct scatterlist sg;
532 u8 *hdr = hp->scratch;
533 int err, len;
534
535 /* We are not allowed to change tcphdr, make a local copy */
536 if (exclude_options) {
537 len = sizeof(*th) + sizeof(struct tcp_ao_hdr) + hash_len;
538 memcpy(hdr, th, sizeof(*th));
539 memcpy(hdr + sizeof(*th),
540 (u8 *)th + hash_offset - sizeof(struct tcp_ao_hdr),
541 sizeof(struct tcp_ao_hdr));
542 memset(hdr + sizeof(*th) + sizeof(struct tcp_ao_hdr),
543 0, hash_len);
544 ((struct tcphdr *)hdr)->check = 0;
545 } else {
546 len = th->doff << 2;
547 memcpy(hdr, th, len);
548 /* zero out tcp-ao hash */
549 ((struct tcphdr *)hdr)->check = 0;
550 memset(hdr + hash_offset, 0, hash_len);
551 }
552
553 sg_init_one(&sg, hdr, len);
554 ahash_request_set_crypt(hp->req, &sg, NULL, len);
555 err = crypto_ahash_update(hp->req);
556 WARN_ON_ONCE(err != 0);
557 return err;
558}
559
560int tcp_ao_hash_hdr(unsigned short int family, char *ao_hash,
561 struct tcp_ao_key *key, const u8 *tkey,
562 const union tcp_ao_addr *daddr,
563 const union tcp_ao_addr *saddr,
564 const struct tcphdr *th, u32 sne)
565{
566 int tkey_len = tcp_ao_digest_size(key);
567 int hash_offset = ao_hash - (char *)th;
568 struct tcp_sigpool hp;
569 void *hash_buf = NULL;
570
571 hash_buf = kmalloc(tkey_len, GFP_ATOMIC);
572 if (!hash_buf)
573 goto clear_hash_noput;
574
575 if (tcp_sigpool_start(key->tcp_sigpool_id, &hp))
576 goto clear_hash_noput;
577
578 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp.req), tkey, tkey_len))
579 goto clear_hash;
580
581 if (crypto_ahash_init(hp.req))
582 goto clear_hash;
583
584 if (tcp_ao_hash_sne(&hp, sne))
585 goto clear_hash;
586 if (family == AF_INET) {
587 if (tcp_v4_ao_hash_pseudoheader(&hp, daddr->a4.s_addr,
588 saddr->a4.s_addr, th->doff * 4))
589 goto clear_hash;
590#if IS_ENABLED(CONFIG_IPV6)
591 } else if (family == AF_INET6) {
592 if (tcp_v6_ao_hash_pseudoheader(&hp, &daddr->a6,
593 &saddr->a6, th->doff * 4))
594 goto clear_hash;
595#endif
596 } else {
597 WARN_ON_ONCE(1);
598 goto clear_hash;
599 }
600 if (tcp_ao_hash_header(&hp, th,
601 !!(key->keyflags & TCP_AO_KEYF_EXCLUDE_OPT),
602 ao_hash, hash_offset, tcp_ao_maclen(key)))
603 goto clear_hash;
604 ahash_request_set_crypt(hp.req, NULL, hash_buf, 0);
605 if (crypto_ahash_final(hp.req))
606 goto clear_hash;
607
608 memcpy(ao_hash, hash_buf, tcp_ao_maclen(key));
609 tcp_sigpool_end(&hp);
610 kfree(hash_buf);
611 return 0;
612
613clear_hash:
614 tcp_sigpool_end(&hp);
615clear_hash_noput:
616 memset(ao_hash, 0, tcp_ao_maclen(key));
617 kfree(hash_buf);
618 return 1;
619}
620
621int tcp_ao_hash_skb(unsigned short int family,
622 char *ao_hash, struct tcp_ao_key *key,
623 const struct sock *sk, const struct sk_buff *skb,
624 const u8 *tkey, int hash_offset, u32 sne)
625{
626 const struct tcphdr *th = tcp_hdr(skb);
627 int tkey_len = tcp_ao_digest_size(key);
628 struct tcp_sigpool hp;
629 void *hash_buf = NULL;
630
631 hash_buf = kmalloc(tkey_len, GFP_ATOMIC);
632 if (!hash_buf)
633 goto clear_hash_noput;
634
635 if (tcp_sigpool_start(key->tcp_sigpool_id, &hp))
636 goto clear_hash_noput;
637
638 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp.req), tkey, tkey_len))
639 goto clear_hash;
640
641 /* For now use sha1 by default. Depends on alg in tcp_ao_key */
642 if (crypto_ahash_init(hp.req))
643 goto clear_hash;
644
645 if (tcp_ao_hash_sne(&hp, sne))
646 goto clear_hash;
647 if (tcp_ao_hash_pseudoheader(family, sk, skb, &hp, skb->len))
648 goto clear_hash;
649 if (tcp_ao_hash_header(&hp, th,
650 !!(key->keyflags & TCP_AO_KEYF_EXCLUDE_OPT),
651 ao_hash, hash_offset, tcp_ao_maclen(key)))
652 goto clear_hash;
653 if (tcp_sigpool_hash_skb_data(&hp, skb, th->doff << 2))
654 goto clear_hash;
655 ahash_request_set_crypt(hp.req, NULL, hash_buf, 0);
656 if (crypto_ahash_final(hp.req))
657 goto clear_hash;
658
659 memcpy(ao_hash, hash_buf, tcp_ao_maclen(key));
660 tcp_sigpool_end(&hp);
661 kfree(hash_buf);
662 return 0;
663
664clear_hash:
665 tcp_sigpool_end(&hp);
666clear_hash_noput:
667 memset(ao_hash, 0, tcp_ao_maclen(key));
668 kfree(hash_buf);
669 return 1;
670}
671
672int tcp_v4_ao_hash_skb(char *ao_hash, struct tcp_ao_key *key,
673 const struct sock *sk, const struct sk_buff *skb,
674 const u8 *tkey, int hash_offset, u32 sne)
675{
676 return tcp_ao_hash_skb(AF_INET, ao_hash, key, sk, skb,
677 tkey, hash_offset, sne);
678}
679
680int tcp_v4_ao_synack_hash(char *ao_hash, struct tcp_ao_key *ao_key,
681 struct request_sock *req, const struct sk_buff *skb,
682 int hash_offset, u32 sne)
683{
684 void *hash_buf = NULL;
685 int err;
686
687 hash_buf = kmalloc(tcp_ao_digest_size(ao_key), GFP_ATOMIC);
688 if (!hash_buf)
689 return -ENOMEM;
690
691 err = tcp_v4_ao_calc_key_rsk(ao_key, hash_buf, req);
692 if (err)
693 goto out;
694
695 err = tcp_ao_hash_skb(AF_INET, ao_hash, ao_key, req_to_sk(req), skb,
696 hash_buf, hash_offset, sne);
697out:
698 kfree(hash_buf);
699 return err;
700}
701
702struct tcp_ao_key *tcp_v4_ao_lookup_rsk(const struct sock *sk,
703 struct request_sock *req,
704 int sndid, int rcvid)
705{
706 struct inet_request_sock *ireq = inet_rsk(req);
707 union tcp_ao_addr *addr = (union tcp_ao_addr *)&ireq->ir_rmt_addr;
708 int l3index;
709
710 l3index = l3mdev_master_ifindex_by_index(sock_net(sk), ireq->ir_iif);
711 return tcp_ao_do_lookup(sk, l3index, addr, AF_INET, sndid, rcvid);
712}
713
714struct tcp_ao_key *tcp_v4_ao_lookup(const struct sock *sk, struct sock *addr_sk,
715 int sndid, int rcvid)
716{
717 int l3index = l3mdev_master_ifindex_by_index(sock_net(sk),
718 addr_sk->sk_bound_dev_if);
719 union tcp_ao_addr *addr = (union tcp_ao_addr *)&addr_sk->sk_daddr;
720
721 return tcp_ao_do_lookup(sk, l3index, addr, AF_INET, sndid, rcvid);
722}
723
724int tcp_ao_prepare_reset(const struct sock *sk, struct sk_buff *skb,
725 const struct tcp_ao_hdr *aoh, int l3index, u32 seq,
726 struct tcp_ao_key **key, char **traffic_key,
727 bool *allocated_traffic_key, u8 *keyid, u32 *sne)
728{
729 const struct tcphdr *th = tcp_hdr(skb);
730 struct tcp_ao_info *ao_info;
731
732 *allocated_traffic_key = false;
733 /* If there's no socket - than initial sisn/disn are unknown.
734 * Drop the segment. RFC5925 (7.7) advises to require graceful
735 * restart [RFC4724]. Alternatively, the RFC5925 advises to
736 * save/restore traffic keys before/after reboot.
737 * Linux TCP-AO support provides TCP_AO_ADD_KEY and TCP_AO_REPAIR
738 * options to restore a socket post-reboot.
739 */
740 if (!sk)
741 return -ENOTCONN;
742
743 if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_NEW_SYN_RECV)) {
744 unsigned int family = READ_ONCE(sk->sk_family);
745 union tcp_ao_addr *addr;
746 __be32 disn, sisn;
747
748 if (sk->sk_state == TCP_NEW_SYN_RECV) {
749 struct request_sock *req = inet_reqsk(sk);
750
751 sisn = htonl(tcp_rsk(req)->rcv_isn);
752 disn = htonl(tcp_rsk(req)->snt_isn);
753 *sne = tcp_ao_compute_sne(0, tcp_rsk(req)->snt_isn, seq);
754 } else {
755 sisn = th->seq;
756 disn = 0;
757 }
758 if (IS_ENABLED(CONFIG_IPV6) && family == AF_INET6)
759 addr = (union tcp_md5_addr *)&ipv6_hdr(skb)->saddr;
760 else
761 addr = (union tcp_md5_addr *)&ip_hdr(skb)->saddr;
762#if IS_ENABLED(CONFIG_IPV6)
763 if (family == AF_INET6 && ipv6_addr_v4mapped(&sk->sk_v6_daddr))
764 family = AF_INET;
765#endif
766
767 sk = sk_const_to_full_sk(sk);
768 ao_info = rcu_dereference(tcp_sk(sk)->ao_info);
769 if (!ao_info)
770 return -ENOENT;
771 *key = tcp_ao_do_lookup(sk, l3index, addr, family,
772 -1, aoh->rnext_keyid);
773 if (!*key)
774 return -ENOENT;
775 *traffic_key = kmalloc(tcp_ao_digest_size(*key), GFP_ATOMIC);
776 if (!*traffic_key)
777 return -ENOMEM;
778 *allocated_traffic_key = true;
779 if (tcp_ao_calc_key_skb(*key, *traffic_key, skb,
780 sisn, disn, family))
781 return -1;
782 *keyid = (*key)->rcvid;
783 } else {
784 struct tcp_ao_key *rnext_key;
785 u32 snd_basis;
786
787 if (sk->sk_state == TCP_TIME_WAIT) {
788 ao_info = rcu_dereference(tcp_twsk(sk)->ao_info);
789 snd_basis = tcp_twsk(sk)->tw_snd_nxt;
790 } else {
791 ao_info = rcu_dereference(tcp_sk(sk)->ao_info);
792 snd_basis = tcp_sk(sk)->snd_una;
793 }
794 if (!ao_info)
795 return -ENOENT;
796
797 *key = tcp_ao_established_key(sk, ao_info, aoh->rnext_keyid, -1);
798 if (!*key)
799 return -ENOENT;
800 *traffic_key = snd_other_key(*key);
801 rnext_key = READ_ONCE(ao_info->rnext_key);
802 *keyid = rnext_key->rcvid;
803 *sne = tcp_ao_compute_sne(READ_ONCE(ao_info->snd_sne),
804 snd_basis, seq);
805 }
806 return 0;
807}
808
809int tcp_ao_transmit_skb(struct sock *sk, struct sk_buff *skb,
810 struct tcp_ao_key *key, struct tcphdr *th,
811 __u8 *hash_location)
812{
813 struct tcp_skb_cb *tcb = TCP_SKB_CB(skb);
814 struct tcp_sock *tp = tcp_sk(sk);
815 struct tcp_ao_info *ao;
816 void *tkey_buf = NULL;
817 u8 *traffic_key;
818 u32 sne;
819
820 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
821 lockdep_sock_is_held(sk));
822 traffic_key = snd_other_key(key);
823 if (unlikely(tcb->tcp_flags & TCPHDR_SYN)) {
824 __be32 disn;
825
826 if (!(tcb->tcp_flags & TCPHDR_ACK)) {
827 disn = 0;
828 tkey_buf = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
829 if (!tkey_buf)
830 return -ENOMEM;
831 traffic_key = tkey_buf;
832 } else {
833 disn = ao->risn;
834 }
835 tp->af_specific->ao_calc_key_sk(key, traffic_key,
836 sk, ao->lisn, disn, true);
837 }
838 sne = tcp_ao_compute_sne(READ_ONCE(ao->snd_sne), READ_ONCE(tp->snd_una),
839 ntohl(th->seq));
840 tp->af_specific->calc_ao_hash(hash_location, key, sk, skb, traffic_key,
841 hash_location - (u8 *)th, sne);
842 kfree(tkey_buf);
843 return 0;
844}
845
846static struct tcp_ao_key *tcp_ao_inbound_lookup(unsigned short int family,
847 const struct sock *sk, const struct sk_buff *skb,
848 int sndid, int rcvid, int l3index)
849{
850 if (family == AF_INET) {
851 const struct iphdr *iph = ip_hdr(skb);
852
853 return tcp_ao_do_lookup(sk, l3index,
854 (union tcp_ao_addr *)&iph->saddr,
855 AF_INET, sndid, rcvid);
856 } else {
857 const struct ipv6hdr *iph = ipv6_hdr(skb);
858
859 return tcp_ao_do_lookup(sk, l3index,
860 (union tcp_ao_addr *)&iph->saddr,
861 AF_INET6, sndid, rcvid);
862 }
863}
864
865void tcp_ao_syncookie(struct sock *sk, const struct sk_buff *skb,
866 struct request_sock *req, unsigned short int family)
867{
868 struct tcp_request_sock *treq = tcp_rsk(req);
869 const struct tcphdr *th = tcp_hdr(skb);
870 const struct tcp_ao_hdr *aoh;
871 struct tcp_ao_key *key;
872 int l3index;
873
874 /* treq->af_specific is used to perform TCP_AO lookup
875 * in tcp_create_openreq_child().
876 */
877#if IS_ENABLED(CONFIG_IPV6)
878 if (family == AF_INET6)
879 treq->af_specific = &tcp_request_sock_ipv6_ops;
880 else
881#endif
882 treq->af_specific = &tcp_request_sock_ipv4_ops;
883
884 treq->used_tcp_ao = false;
885
886 if (tcp_parse_auth_options(th, NULL, &aoh) || !aoh)
887 return;
888
889 l3index = l3mdev_master_ifindex_by_index(sock_net(sk), inet_rsk(req)->ir_iif);
890 key = tcp_ao_inbound_lookup(family, sk, skb, -1, aoh->keyid, l3index);
891 if (!key)
892 /* Key not found, continue without TCP-AO */
893 return;
894
895 treq->ao_rcv_next = aoh->keyid;
896 treq->ao_keyid = aoh->rnext_keyid;
897 treq->used_tcp_ao = true;
898}
899
900static enum skb_drop_reason
901tcp_ao_verify_hash(const struct sock *sk, const struct sk_buff *skb,
902 unsigned short int family, struct tcp_ao_info *info,
903 const struct tcp_ao_hdr *aoh, struct tcp_ao_key *key,
904 u8 *traffic_key, u8 *phash, u32 sne, int l3index)
905{
906 const struct tcphdr *th = tcp_hdr(skb);
907 u8 maclen = tcp_ao_hdr_maclen(aoh);
908 void *hash_buf = NULL;
909
910 if (maclen != tcp_ao_maclen(key)) {
911 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOBAD);
912 atomic64_inc(&info->counters.pkt_bad);
913 atomic64_inc(&key->pkt_bad);
914 trace_tcp_ao_wrong_maclen(sk, skb, aoh->keyid,
915 aoh->rnext_keyid, maclen);
916 return SKB_DROP_REASON_TCP_AOFAILURE;
917 }
918
919 hash_buf = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
920 if (!hash_buf)
921 return SKB_DROP_REASON_NOT_SPECIFIED;
922
923 /* XXX: make it per-AF callback? */
924 tcp_ao_hash_skb(family, hash_buf, key, sk, skb, traffic_key,
925 (phash - (u8 *)th), sne);
926 if (memcmp(phash, hash_buf, maclen)) {
927 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOBAD);
928 atomic64_inc(&info->counters.pkt_bad);
929 atomic64_inc(&key->pkt_bad);
930 trace_tcp_ao_mismatch(sk, skb, aoh->keyid,
931 aoh->rnext_keyid, maclen);
932 kfree(hash_buf);
933 return SKB_DROP_REASON_TCP_AOFAILURE;
934 }
935 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOGOOD);
936 atomic64_inc(&info->counters.pkt_good);
937 atomic64_inc(&key->pkt_good);
938 kfree(hash_buf);
939 return SKB_NOT_DROPPED_YET;
940}
941
942enum skb_drop_reason
943tcp_inbound_ao_hash(struct sock *sk, const struct sk_buff *skb,
944 unsigned short int family, const struct request_sock *req,
945 int l3index, const struct tcp_ao_hdr *aoh)
946{
947 const struct tcphdr *th = tcp_hdr(skb);
948 u8 maclen = tcp_ao_hdr_maclen(aoh);
949 u8 *phash = (u8 *)(aoh + 1); /* hash goes just after the header */
950 struct tcp_ao_info *info;
951 enum skb_drop_reason ret;
952 struct tcp_ao_key *key;
953 __be32 sisn, disn;
954 u8 *traffic_key;
955 int state;
956 u32 sne = 0;
957
958 info = rcu_dereference(tcp_sk(sk)->ao_info);
959 if (!info) {
960 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOKEYNOTFOUND);
961 trace_tcp_ao_key_not_found(sk, skb, aoh->keyid,
962 aoh->rnext_keyid, maclen);
963 return SKB_DROP_REASON_TCP_AOUNEXPECTED;
964 }
965
966 if (unlikely(th->syn)) {
967 sisn = th->seq;
968 disn = 0;
969 }
970
971 state = READ_ONCE(sk->sk_state);
972 /* Fast-path */
973 if (likely((1 << state) & TCP_AO_ESTABLISHED)) {
974 enum skb_drop_reason err;
975 struct tcp_ao_key *current_key;
976
977 /* Check if this socket's rnext_key matches the keyid in the
978 * packet. If not we lookup the key based on the keyid
979 * matching the rcvid in the mkt.
980 */
981 key = READ_ONCE(info->rnext_key);
982 if (key->rcvid != aoh->keyid) {
983 key = tcp_ao_established_key(sk, info, -1, aoh->keyid);
984 if (!key)
985 goto key_not_found;
986 }
987
988 /* Delayed retransmitted SYN */
989 if (unlikely(th->syn && !th->ack))
990 goto verify_hash;
991
992 sne = tcp_ao_compute_sne(info->rcv_sne, tcp_sk(sk)->rcv_nxt,
993 ntohl(th->seq));
994 /* Established socket, traffic key are cached */
995 traffic_key = rcv_other_key(key);
996 err = tcp_ao_verify_hash(sk, skb, family, info, aoh, key,
997 traffic_key, phash, sne, l3index);
998 if (err)
999 return err;
1000 current_key = READ_ONCE(info->current_key);
1001 /* Key rotation: the peer asks us to use new key (RNext) */
1002 if (unlikely(aoh->rnext_keyid != current_key->sndid)) {
1003 trace_tcp_ao_rnext_request(sk, skb, current_key->sndid,
1004 aoh->rnext_keyid,
1005 tcp_ao_hdr_maclen(aoh));
1006 /* If the key is not found we do nothing. */
1007 key = tcp_ao_established_key(sk, info, aoh->rnext_keyid, -1);
1008 if (key)
1009 /* pairs with tcp_ao_del_cmd */
1010 WRITE_ONCE(info->current_key, key);
1011 }
1012 return SKB_NOT_DROPPED_YET;
1013 }
1014
1015 if (unlikely(state == TCP_CLOSE))
1016 return SKB_DROP_REASON_TCP_CLOSE;
1017
1018 /* Lookup key based on peer address and keyid.
1019 * current_key and rnext_key must not be used on tcp listen
1020 * sockets as otherwise:
1021 * - request sockets would race on those key pointers
1022 * - tcp_ao_del_cmd() allows async key removal
1023 */
1024 key = tcp_ao_inbound_lookup(family, sk, skb, -1, aoh->keyid, l3index);
1025 if (!key)
1026 goto key_not_found;
1027
1028 if (th->syn && !th->ack)
1029 goto verify_hash;
1030
1031 if ((1 << state) & (TCPF_LISTEN | TCPF_NEW_SYN_RECV)) {
1032 /* Make the initial syn the likely case here */
1033 if (unlikely(req)) {
1034 sne = tcp_ao_compute_sne(0, tcp_rsk(req)->rcv_isn,
1035 ntohl(th->seq));
1036 sisn = htonl(tcp_rsk(req)->rcv_isn);
1037 disn = htonl(tcp_rsk(req)->snt_isn);
1038 } else if (unlikely(th->ack && !th->syn)) {
1039 /* Possible syncookie packet */
1040 sisn = htonl(ntohl(th->seq) - 1);
1041 disn = htonl(ntohl(th->ack_seq) - 1);
1042 sne = tcp_ao_compute_sne(0, ntohl(sisn),
1043 ntohl(th->seq));
1044 } else if (unlikely(!th->syn)) {
1045 /* no way to figure out initial sisn/disn - drop */
1046 return SKB_DROP_REASON_TCP_FLAGS;
1047 }
1048 } else if ((1 << state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) {
1049 disn = info->lisn;
1050 if (th->syn || th->rst)
1051 sisn = th->seq;
1052 else
1053 sisn = info->risn;
1054 } else {
1055 WARN_ONCE(1, "TCP-AO: Unexpected sk_state %d", state);
1056 return SKB_DROP_REASON_TCP_AOFAILURE;
1057 }
1058verify_hash:
1059 traffic_key = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
1060 if (!traffic_key)
1061 return SKB_DROP_REASON_NOT_SPECIFIED;
1062 tcp_ao_calc_key_skb(key, traffic_key, skb, sisn, disn, family);
1063 ret = tcp_ao_verify_hash(sk, skb, family, info, aoh, key,
1064 traffic_key, phash, sne, l3index);
1065 kfree(traffic_key);
1066 return ret;
1067
1068key_not_found:
1069 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOKEYNOTFOUND);
1070 atomic64_inc(&info->counters.key_not_found);
1071 trace_tcp_ao_key_not_found(sk, skb, aoh->keyid,
1072 aoh->rnext_keyid, maclen);
1073 return SKB_DROP_REASON_TCP_AOKEYNOTFOUND;
1074}
1075
1076static int tcp_ao_cache_traffic_keys(const struct sock *sk,
1077 struct tcp_ao_info *ao,
1078 struct tcp_ao_key *ao_key)
1079{
1080 u8 *traffic_key = snd_other_key(ao_key);
1081 int ret;
1082
1083 ret = tcp_ao_calc_key_sk(ao_key, traffic_key, sk,
1084 ao->lisn, ao->risn, true);
1085 if (ret)
1086 return ret;
1087
1088 traffic_key = rcv_other_key(ao_key);
1089 ret = tcp_ao_calc_key_sk(ao_key, traffic_key, sk,
1090 ao->lisn, ao->risn, false);
1091 return ret;
1092}
1093
1094void tcp_ao_connect_init(struct sock *sk)
1095{
1096 struct tcp_sock *tp = tcp_sk(sk);
1097 struct tcp_ao_info *ao_info;
1098 struct hlist_node *next;
1099 union tcp_ao_addr *addr;
1100 struct tcp_ao_key *key;
1101 int family, l3index;
1102
1103 ao_info = rcu_dereference_protected(tp->ao_info,
1104 lockdep_sock_is_held(sk));
1105 if (!ao_info)
1106 return;
1107
1108 /* Remove all keys that don't match the peer */
1109 family = sk->sk_family;
1110 if (family == AF_INET)
1111 addr = (union tcp_ao_addr *)&sk->sk_daddr;
1112#if IS_ENABLED(CONFIG_IPV6)
1113 else if (family == AF_INET6)
1114 addr = (union tcp_ao_addr *)&sk->sk_v6_daddr;
1115#endif
1116 else
1117 return;
1118 l3index = l3mdev_master_ifindex_by_index(sock_net(sk),
1119 sk->sk_bound_dev_if);
1120
1121 hlist_for_each_entry_safe(key, next, &ao_info->head, node) {
1122 if (!tcp_ao_key_cmp(key, l3index, addr, key->prefixlen, family, -1, -1))
1123 continue;
1124
1125 if (key == ao_info->current_key)
1126 ao_info->current_key = NULL;
1127 if (key == ao_info->rnext_key)
1128 ao_info->rnext_key = NULL;
1129 hlist_del_rcu(&key->node);
1130 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1131 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1132 }
1133
1134 key = tp->af_specific->ao_lookup(sk, sk, -1, -1);
1135 if (key) {
1136 /* if current_key or rnext_key were not provided,
1137 * use the first key matching the peer
1138 */
1139 if (!ao_info->current_key)
1140 ao_info->current_key = key;
1141 if (!ao_info->rnext_key)
1142 ao_info->rnext_key = key;
1143 tp->tcp_header_len += tcp_ao_len_aligned(key);
1144
1145 ao_info->lisn = htonl(tp->write_seq);
1146 ao_info->snd_sne = 0;
1147 } else {
1148 /* Can't happen: tcp_connect() verifies that there's
1149 * at least one tcp-ao key that matches the remote peer.
1150 */
1151 WARN_ON_ONCE(1);
1152 rcu_assign_pointer(tp->ao_info, NULL);
1153 kfree(ao_info);
1154 }
1155}
1156
1157void tcp_ao_established(struct sock *sk)
1158{
1159 struct tcp_ao_info *ao;
1160 struct tcp_ao_key *key;
1161
1162 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
1163 lockdep_sock_is_held(sk));
1164 if (!ao)
1165 return;
1166
1167 hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
1168 tcp_ao_cache_traffic_keys(sk, ao, key);
1169}
1170
1171void tcp_ao_finish_connect(struct sock *sk, struct sk_buff *skb)
1172{
1173 struct tcp_ao_info *ao;
1174 struct tcp_ao_key *key;
1175
1176 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
1177 lockdep_sock_is_held(sk));
1178 if (!ao)
1179 return;
1180
1181 WRITE_ONCE(ao->risn, tcp_hdr(skb)->seq);
1182 ao->rcv_sne = 0;
1183
1184 hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
1185 tcp_ao_cache_traffic_keys(sk, ao, key);
1186}
1187
1188int tcp_ao_copy_all_matching(const struct sock *sk, struct sock *newsk,
1189 struct request_sock *req, struct sk_buff *skb,
1190 int family)
1191{
1192 struct tcp_ao_key *key, *new_key, *first_key;
1193 struct tcp_ao_info *new_ao, *ao;
1194 struct hlist_node *key_head;
1195 int l3index, ret = -ENOMEM;
1196 union tcp_ao_addr *addr;
1197 bool match = false;
1198
1199 ao = rcu_dereference(tcp_sk(sk)->ao_info);
1200 if (!ao)
1201 return 0;
1202
1203 /* New socket without TCP-AO on it */
1204 if (!tcp_rsk_used_ao(req))
1205 return 0;
1206
1207 new_ao = tcp_ao_alloc_info(GFP_ATOMIC);
1208 if (!new_ao)
1209 return -ENOMEM;
1210 new_ao->lisn = htonl(tcp_rsk(req)->snt_isn);
1211 new_ao->risn = htonl(tcp_rsk(req)->rcv_isn);
1212 new_ao->ao_required = ao->ao_required;
1213 new_ao->accept_icmps = ao->accept_icmps;
1214
1215 if (family == AF_INET) {
1216 addr = (union tcp_ao_addr *)&newsk->sk_daddr;
1217#if IS_ENABLED(CONFIG_IPV6)
1218 } else if (family == AF_INET6) {
1219 addr = (union tcp_ao_addr *)&newsk->sk_v6_daddr;
1220#endif
1221 } else {
1222 ret = -EAFNOSUPPORT;
1223 goto free_ao;
1224 }
1225 l3index = l3mdev_master_ifindex_by_index(sock_net(newsk),
1226 newsk->sk_bound_dev_if);
1227
1228 hlist_for_each_entry_rcu(key, &ao->head, node) {
1229 if (tcp_ao_key_cmp(key, l3index, addr, key->prefixlen, family, -1, -1))
1230 continue;
1231
1232 new_key = tcp_ao_copy_key(newsk, key);
1233 if (!new_key)
1234 goto free_and_exit;
1235
1236 tcp_ao_cache_traffic_keys(newsk, new_ao, new_key);
1237 tcp_ao_link_mkt(new_ao, new_key);
1238 match = true;
1239 }
1240
1241 if (!match) {
1242 /* RFC5925 (7.4.1) specifies that the TCP-AO status
1243 * of a connection is determined on the initial SYN.
1244 * At this point the connection was TCP-AO enabled, so
1245 * it can't switch to being unsigned if peer's key
1246 * disappears on the listening socket.
1247 */
1248 ret = -EKEYREJECTED;
1249 goto free_and_exit;
1250 }
1251
1252 if (!static_key_fast_inc_not_disabled(&tcp_ao_needed.key.key)) {
1253 ret = -EUSERS;
1254 goto free_and_exit;
1255 }
1256
1257 key_head = rcu_dereference(hlist_first_rcu(&new_ao->head));
1258 first_key = hlist_entry_safe(key_head, struct tcp_ao_key, node);
1259
1260 key = tcp_ao_established_key(req_to_sk(req), new_ao, tcp_rsk(req)->ao_keyid, -1);
1261 if (key)
1262 new_ao->current_key = key;
1263 else
1264 new_ao->current_key = first_key;
1265
1266 /* set rnext_key */
1267 key = tcp_ao_established_key(req_to_sk(req), new_ao, -1, tcp_rsk(req)->ao_rcv_next);
1268 if (key)
1269 new_ao->rnext_key = key;
1270 else
1271 new_ao->rnext_key = first_key;
1272
1273 sk_gso_disable(newsk);
1274 rcu_assign_pointer(tcp_sk(newsk)->ao_info, new_ao);
1275
1276 return 0;
1277
1278free_and_exit:
1279 hlist_for_each_entry_safe(key, key_head, &new_ao->head, node) {
1280 hlist_del(&key->node);
1281 tcp_sigpool_release(key->tcp_sigpool_id);
1282 atomic_sub(tcp_ao_sizeof_key(key), &newsk->sk_omem_alloc);
1283 kfree_sensitive(key);
1284 }
1285free_ao:
1286 kfree(new_ao);
1287 return ret;
1288}
1289
1290static bool tcp_ao_can_set_current_rnext(struct sock *sk)
1291{
1292 /* There aren't current/rnext keys on TCP_LISTEN sockets */
1293 if (sk->sk_state == TCP_LISTEN)
1294 return false;
1295 return true;
1296}
1297
1298static int tcp_ao_verify_ipv4(struct sock *sk, struct tcp_ao_add *cmd,
1299 union tcp_ao_addr **addr)
1300{
1301 struct sockaddr_in *sin = (struct sockaddr_in *)&cmd->addr;
1302 struct inet_sock *inet = inet_sk(sk);
1303
1304 if (sin->sin_family != AF_INET)
1305 return -EINVAL;
1306
1307 /* Currently matching is not performed on port (or port ranges) */
1308 if (sin->sin_port != 0)
1309 return -EINVAL;
1310
1311 /* Check prefix and trailing 0's in addr */
1312 if (cmd->prefix != 0) {
1313 __be32 mask;
1314
1315 if (ntohl(sin->sin_addr.s_addr) == INADDR_ANY)
1316 return -EINVAL;
1317 if (cmd->prefix > 32)
1318 return -EINVAL;
1319
1320 mask = inet_make_mask(cmd->prefix);
1321 if (sin->sin_addr.s_addr & ~mask)
1322 return -EINVAL;
1323
1324 /* Check that MKT address is consistent with socket */
1325 if (ntohl(inet->inet_daddr) != INADDR_ANY &&
1326 (inet->inet_daddr & mask) != sin->sin_addr.s_addr)
1327 return -EINVAL;
1328 } else {
1329 if (ntohl(sin->sin_addr.s_addr) != INADDR_ANY)
1330 return -EINVAL;
1331 }
1332
1333 *addr = (union tcp_ao_addr *)&sin->sin_addr;
1334 return 0;
1335}
1336
1337static int tcp_ao_parse_crypto(struct tcp_ao_add *cmd, struct tcp_ao_key *key)
1338{
1339 unsigned int syn_tcp_option_space;
1340 bool is_kdf_aes_128_cmac = false;
1341 struct crypto_ahash *tfm;
1342 struct tcp_sigpool hp;
1343 void *tmp_key = NULL;
1344 int err;
1345
1346 /* RFC5926, 3.1.1.2. KDF_AES_128_CMAC */
1347 if (!strcmp("cmac(aes128)", cmd->alg_name)) {
1348 strscpy(cmd->alg_name, "cmac(aes)", sizeof(cmd->alg_name));
1349 is_kdf_aes_128_cmac = (cmd->keylen != 16);
1350 tmp_key = kmalloc(cmd->keylen, GFP_KERNEL);
1351 if (!tmp_key)
1352 return -ENOMEM;
1353 }
1354
1355 key->maclen = cmd->maclen ?: 12; /* 12 is the default in RFC5925 */
1356
1357 /* Check: maclen + tcp-ao header <= (MAX_TCP_OPTION_SPACE - mss
1358 * - tstamp (including sackperm)
1359 * - wscale),
1360 * see tcp_syn_options(), tcp_synack_options(), commit 33ad798c924b.
1361 *
1362 * In order to allow D-SACK with TCP-AO, the header size should be:
1363 * (MAX_TCP_OPTION_SPACE - TCPOLEN_TSTAMP_ALIGNED
1364 * - TCPOLEN_SACK_BASE_ALIGNED
1365 * - 2 * TCPOLEN_SACK_PERBLOCK) = 8 (maclen = 4),
1366 * see tcp_established_options().
1367 *
1368 * RFC5925, 2.2:
1369 * Typical MACs are 96-128 bits (12-16 bytes), but any length
1370 * that fits in the header of the segment being authenticated
1371 * is allowed.
1372 *
1373 * RFC5925, 7.6:
1374 * TCP-AO continues to consume 16 bytes in non-SYN segments,
1375 * leaving a total of 24 bytes for other options, of which
1376 * the timestamp consumes 10. This leaves 14 bytes, of which 10
1377 * are used for a single SACK block. When two SACK blocks are used,
1378 * such as to handle D-SACK, a smaller TCP-AO MAC would be required
1379 * to make room for the additional SACK block (i.e., to leave 18
1380 * bytes for the D-SACK variant of the SACK option) [RFC2883].
1381 * Note that D-SACK is not supportable in TCP MD5 in the presence
1382 * of timestamps, because TCP MD5’s MAC length is fixed and too
1383 * large to leave sufficient option space.
1384 */
1385 syn_tcp_option_space = MAX_TCP_OPTION_SPACE;
1386 syn_tcp_option_space -= TCPOLEN_MSS_ALIGNED;
1387 syn_tcp_option_space -= TCPOLEN_TSTAMP_ALIGNED;
1388 syn_tcp_option_space -= TCPOLEN_WSCALE_ALIGNED;
1389 if (tcp_ao_len_aligned(key) > syn_tcp_option_space) {
1390 err = -EMSGSIZE;
1391 goto err_kfree;
1392 }
1393
1394 key->keylen = cmd->keylen;
1395 memcpy(key->key, cmd->key, cmd->keylen);
1396
1397 err = tcp_sigpool_start(key->tcp_sigpool_id, &hp);
1398 if (err)
1399 goto err_kfree;
1400
1401 tfm = crypto_ahash_reqtfm(hp.req);
1402 if (is_kdf_aes_128_cmac) {
1403 void *scratch = hp.scratch;
1404 struct scatterlist sg;
1405
1406 memcpy(tmp_key, cmd->key, cmd->keylen);
1407 sg_init_one(&sg, tmp_key, cmd->keylen);
1408
1409 /* Using zero-key of 16 bytes as described in RFC5926 */
1410 memset(scratch, 0, 16);
1411 err = crypto_ahash_setkey(tfm, scratch, 16);
1412 if (err)
1413 goto err_pool_end;
1414
1415 err = crypto_ahash_init(hp.req);
1416 if (err)
1417 goto err_pool_end;
1418
1419 ahash_request_set_crypt(hp.req, &sg, key->key, cmd->keylen);
1420 err = crypto_ahash_update(hp.req);
1421 if (err)
1422 goto err_pool_end;
1423
1424 err |= crypto_ahash_final(hp.req);
1425 if (err)
1426 goto err_pool_end;
1427 key->keylen = 16;
1428 }
1429
1430 err = crypto_ahash_setkey(tfm, key->key, key->keylen);
1431 if (err)
1432 goto err_pool_end;
1433
1434 tcp_sigpool_end(&hp);
1435 kfree_sensitive(tmp_key);
1436
1437 if (tcp_ao_maclen(key) > key->digest_size)
1438 return -EINVAL;
1439
1440 return 0;
1441
1442err_pool_end:
1443 tcp_sigpool_end(&hp);
1444err_kfree:
1445 kfree_sensitive(tmp_key);
1446 return err;
1447}
1448
1449#if IS_ENABLED(CONFIG_IPV6)
1450static int tcp_ao_verify_ipv6(struct sock *sk, struct tcp_ao_add *cmd,
1451 union tcp_ao_addr **paddr,
1452 unsigned short int *family)
1453{
1454 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd->addr;
1455 struct in6_addr *addr = &sin6->sin6_addr;
1456 u8 prefix = cmd->prefix;
1457
1458 if (sin6->sin6_family != AF_INET6)
1459 return -EINVAL;
1460
1461 /* Currently matching is not performed on port (or port ranges) */
1462 if (sin6->sin6_port != 0)
1463 return -EINVAL;
1464
1465 /* Check prefix and trailing 0's in addr */
1466 if (cmd->prefix != 0 && ipv6_addr_v4mapped(addr)) {
1467 __be32 addr4 = addr->s6_addr32[3];
1468 __be32 mask;
1469
1470 if (prefix > 32 || ntohl(addr4) == INADDR_ANY)
1471 return -EINVAL;
1472
1473 mask = inet_make_mask(prefix);
1474 if (addr4 & ~mask)
1475 return -EINVAL;
1476
1477 /* Check that MKT address is consistent with socket */
1478 if (!ipv6_addr_any(&sk->sk_v6_daddr)) {
1479 __be32 daddr4 = sk->sk_v6_daddr.s6_addr32[3];
1480
1481 if (!ipv6_addr_v4mapped(&sk->sk_v6_daddr))
1482 return -EINVAL;
1483 if ((daddr4 & mask) != addr4)
1484 return -EINVAL;
1485 }
1486
1487 *paddr = (union tcp_ao_addr *)&addr->s6_addr32[3];
1488 *family = AF_INET;
1489 return 0;
1490 } else if (cmd->prefix != 0) {
1491 struct in6_addr pfx;
1492
1493 if (ipv6_addr_any(addr) || prefix > 128)
1494 return -EINVAL;
1495
1496 ipv6_addr_prefix(&pfx, addr, prefix);
1497 if (ipv6_addr_cmp(&pfx, addr))
1498 return -EINVAL;
1499
1500 /* Check that MKT address is consistent with socket */
1501 if (!ipv6_addr_any(&sk->sk_v6_daddr) &&
1502 !ipv6_prefix_equal(&sk->sk_v6_daddr, addr, prefix))
1503
1504 return -EINVAL;
1505 } else {
1506 if (!ipv6_addr_any(addr))
1507 return -EINVAL;
1508 }
1509
1510 *paddr = (union tcp_ao_addr *)addr;
1511 return 0;
1512}
1513#else
1514static int tcp_ao_verify_ipv6(struct sock *sk, struct tcp_ao_add *cmd,
1515 union tcp_ao_addr **paddr,
1516 unsigned short int *family)
1517{
1518 return -EOPNOTSUPP;
1519}
1520#endif
1521
1522static struct tcp_ao_info *setsockopt_ao_info(struct sock *sk)
1523{
1524 if (sk_fullsock(sk)) {
1525 return rcu_dereference_protected(tcp_sk(sk)->ao_info,
1526 lockdep_sock_is_held(sk));
1527 } else if (sk->sk_state == TCP_TIME_WAIT) {
1528 return rcu_dereference_protected(tcp_twsk(sk)->ao_info,
1529 lockdep_sock_is_held(sk));
1530 }
1531 return ERR_PTR(-ESOCKTNOSUPPORT);
1532}
1533
1534static struct tcp_ao_info *getsockopt_ao_info(struct sock *sk)
1535{
1536 if (sk_fullsock(sk))
1537 return rcu_dereference(tcp_sk(sk)->ao_info);
1538 else if (sk->sk_state == TCP_TIME_WAIT)
1539 return rcu_dereference(tcp_twsk(sk)->ao_info);
1540
1541 return ERR_PTR(-ESOCKTNOSUPPORT);
1542}
1543
1544#define TCP_AO_KEYF_ALL (TCP_AO_KEYF_IFINDEX | TCP_AO_KEYF_EXCLUDE_OPT)
1545#define TCP_AO_GET_KEYF_VALID (TCP_AO_KEYF_IFINDEX)
1546
1547static struct tcp_ao_key *tcp_ao_key_alloc(struct sock *sk,
1548 struct tcp_ao_add *cmd)
1549{
1550 const char *algo = cmd->alg_name;
1551 unsigned int digest_size;
1552 struct crypto_ahash *tfm;
1553 struct tcp_ao_key *key;
1554 struct tcp_sigpool hp;
1555 int err, pool_id;
1556 size_t size;
1557
1558 /* Force null-termination of alg_name */
1559 cmd->alg_name[ARRAY_SIZE(cmd->alg_name) - 1] = '\0';
1560
1561 /* RFC5926, 3.1.1.2. KDF_AES_128_CMAC */
1562 if (!strcmp("cmac(aes128)", algo))
1563 algo = "cmac(aes)";
1564
1565 /* Full TCP header (th->doff << 2) should fit into scratch area,
1566 * see tcp_ao_hash_header().
1567 */
1568 pool_id = tcp_sigpool_alloc_ahash(algo, 60);
1569 if (pool_id < 0)
1570 return ERR_PTR(pool_id);
1571
1572 err = tcp_sigpool_start(pool_id, &hp);
1573 if (err)
1574 goto err_free_pool;
1575
1576 tfm = crypto_ahash_reqtfm(hp.req);
1577 digest_size = crypto_ahash_digestsize(tfm);
1578 tcp_sigpool_end(&hp);
1579
1580 size = sizeof(struct tcp_ao_key) + (digest_size << 1);
1581 key = sock_kmalloc(sk, size, GFP_KERNEL);
1582 if (!key) {
1583 err = -ENOMEM;
1584 goto err_free_pool;
1585 }
1586
1587 key->tcp_sigpool_id = pool_id;
1588 key->digest_size = digest_size;
1589 return key;
1590
1591err_free_pool:
1592 tcp_sigpool_release(pool_id);
1593 return ERR_PTR(err);
1594}
1595
1596static int tcp_ao_add_cmd(struct sock *sk, unsigned short int family,
1597 sockptr_t optval, int optlen)
1598{
1599 struct tcp_ao_info *ao_info;
1600 union tcp_ao_addr *addr;
1601 struct tcp_ao_key *key;
1602 struct tcp_ao_add cmd;
1603 int ret, l3index = 0;
1604 bool first = false;
1605
1606 if (optlen < sizeof(cmd))
1607 return -EINVAL;
1608
1609 ret = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1610 if (ret)
1611 return ret;
1612
1613 if (cmd.keylen > TCP_AO_MAXKEYLEN)
1614 return -EINVAL;
1615
1616 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1617 return -EINVAL;
1618
1619 if (family == AF_INET)
1620 ret = tcp_ao_verify_ipv4(sk, &cmd, &addr);
1621 else
1622 ret = tcp_ao_verify_ipv6(sk, &cmd, &addr, &family);
1623 if (ret)
1624 return ret;
1625
1626 if (cmd.keyflags & ~TCP_AO_KEYF_ALL)
1627 return -EINVAL;
1628
1629 if (cmd.set_current || cmd.set_rnext) {
1630 if (!tcp_ao_can_set_current_rnext(sk))
1631 return -EINVAL;
1632 }
1633
1634 if (cmd.ifindex && !(cmd.keyflags & TCP_AO_KEYF_IFINDEX))
1635 return -EINVAL;
1636
1637 /* For cmd.tcp_ifindex = 0 the key will apply to the default VRF */
1638 if (cmd.keyflags & TCP_AO_KEYF_IFINDEX && cmd.ifindex) {
1639 int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
1640 struct net_device *dev;
1641
1642 rcu_read_lock();
1643 dev = dev_get_by_index_rcu(sock_net(sk), cmd.ifindex);
1644 if (dev && netif_is_l3_master(dev))
1645 l3index = dev->ifindex;
1646 rcu_read_unlock();
1647
1648 if (!dev || !l3index)
1649 return -EINVAL;
1650
1651 if (!bound_dev_if || bound_dev_if != cmd.ifindex) {
1652 /* tcp_ao_established_key() doesn't expect having
1653 * non peer-matching key on an established TCP-AO
1654 * connection.
1655 */
1656 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE)))
1657 return -EINVAL;
1658 }
1659
1660 /* It's still possible to bind after adding keys or even
1661 * re-bind to a different dev (with CAP_NET_RAW).
1662 * So, no reason to return error here, rather try to be
1663 * nice and warn the user.
1664 */
1665 if (bound_dev_if && bound_dev_if != cmd.ifindex)
1666 net_warn_ratelimited("AO key ifindex %d != sk bound ifindex %d\n",
1667 cmd.ifindex, bound_dev_if);
1668 }
1669
1670 /* Don't allow keys for peers that have a matching TCP-MD5 key */
1671 if (cmd.keyflags & TCP_AO_KEYF_IFINDEX) {
1672 /* Non-_exact version of tcp_md5_do_lookup() will
1673 * as well match keys that aren't bound to a specific VRF
1674 * (that will make them match AO key with
1675 * sysctl_tcp_l3dev_accept = 1
1676 */
1677 if (tcp_md5_do_lookup(sk, l3index, addr, family))
1678 return -EKEYREJECTED;
1679 } else {
1680 if (tcp_md5_do_lookup_any_l3index(sk, addr, family))
1681 return -EKEYREJECTED;
1682 }
1683
1684 ao_info = setsockopt_ao_info(sk);
1685 if (IS_ERR(ao_info))
1686 return PTR_ERR(ao_info);
1687
1688 if (!ao_info) {
1689 ao_info = tcp_ao_alloc_info(GFP_KERNEL);
1690 if (!ao_info)
1691 return -ENOMEM;
1692 first = true;
1693 } else {
1694 /* Check that neither RecvID nor SendID match any
1695 * existing key for the peer, RFC5925 3.1:
1696 * > The IDs of MKTs MUST NOT overlap where their
1697 * > TCP connection identifiers overlap.
1698 */
1699 if (__tcp_ao_do_lookup(sk, l3index, addr, family, cmd.prefix, -1, cmd.rcvid))
1700 return -EEXIST;
1701 if (__tcp_ao_do_lookup(sk, l3index, addr, family,
1702 cmd.prefix, cmd.sndid, -1))
1703 return -EEXIST;
1704 }
1705
1706 key = tcp_ao_key_alloc(sk, &cmd);
1707 if (IS_ERR(key)) {
1708 ret = PTR_ERR(key);
1709 goto err_free_ao;
1710 }
1711
1712 INIT_HLIST_NODE(&key->node);
1713 memcpy(&key->addr, addr, (family == AF_INET) ? sizeof(struct in_addr) :
1714 sizeof(struct in6_addr));
1715 key->prefixlen = cmd.prefix;
1716 key->family = family;
1717 key->keyflags = cmd.keyflags;
1718 key->sndid = cmd.sndid;
1719 key->rcvid = cmd.rcvid;
1720 key->l3index = l3index;
1721 atomic64_set(&key->pkt_good, 0);
1722 atomic64_set(&key->pkt_bad, 0);
1723
1724 ret = tcp_ao_parse_crypto(&cmd, key);
1725 if (ret < 0)
1726 goto err_free_sock;
1727
1728 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))) {
1729 tcp_ao_cache_traffic_keys(sk, ao_info, key);
1730 if (first) {
1731 ao_info->current_key = key;
1732 ao_info->rnext_key = key;
1733 }
1734 }
1735
1736 tcp_ao_link_mkt(ao_info, key);
1737 if (first) {
1738 if (!static_branch_inc(&tcp_ao_needed.key)) {
1739 ret = -EUSERS;
1740 goto err_free_sock;
1741 }
1742 sk_gso_disable(sk);
1743 rcu_assign_pointer(tcp_sk(sk)->ao_info, ao_info);
1744 }
1745
1746 if (cmd.set_current)
1747 WRITE_ONCE(ao_info->current_key, key);
1748 if (cmd.set_rnext)
1749 WRITE_ONCE(ao_info->rnext_key, key);
1750 return 0;
1751
1752err_free_sock:
1753 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1754 tcp_sigpool_release(key->tcp_sigpool_id);
1755 kfree_sensitive(key);
1756err_free_ao:
1757 if (first)
1758 kfree(ao_info);
1759 return ret;
1760}
1761
1762static int tcp_ao_delete_key(struct sock *sk, struct tcp_ao_info *ao_info,
1763 bool del_async, struct tcp_ao_key *key,
1764 struct tcp_ao_key *new_current,
1765 struct tcp_ao_key *new_rnext)
1766{
1767 int err;
1768
1769 hlist_del_rcu(&key->node);
1770
1771 /* Support for async delete on listening sockets: as they don't
1772 * need current_key/rnext_key maintaining, we don't need to check
1773 * them and we can just free all resources in RCU fashion.
1774 */
1775 if (del_async) {
1776 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1777 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1778 return 0;
1779 }
1780
1781 /* At this moment another CPU could have looked this key up
1782 * while it was unlinked from the list. Wait for RCU grace period,
1783 * after which the key is off-list and can't be looked up again;
1784 * the rx path [just before RCU came] might have used it and set it
1785 * as current_key (very unlikely).
1786 * Free the key with next RCU grace period (in case it was
1787 * current_key before tcp_ao_current_rnext() might have
1788 * changed it in forced-delete).
1789 */
1790 synchronize_rcu();
1791 if (new_current)
1792 WRITE_ONCE(ao_info->current_key, new_current);
1793 if (new_rnext)
1794 WRITE_ONCE(ao_info->rnext_key, new_rnext);
1795
1796 if (unlikely(READ_ONCE(ao_info->current_key) == key ||
1797 READ_ONCE(ao_info->rnext_key) == key)) {
1798 err = -EBUSY;
1799 goto add_key;
1800 }
1801
1802 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1803 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1804
1805 return 0;
1806add_key:
1807 hlist_add_head_rcu(&key->node, &ao_info->head);
1808 return err;
1809}
1810
1811#define TCP_AO_DEL_KEYF_ALL (TCP_AO_KEYF_IFINDEX)
1812static int tcp_ao_del_cmd(struct sock *sk, unsigned short int family,
1813 sockptr_t optval, int optlen)
1814{
1815 struct tcp_ao_key *key, *new_current = NULL, *new_rnext = NULL;
1816 int err, addr_len, l3index = 0;
1817 struct tcp_ao_info *ao_info;
1818 union tcp_ao_addr *addr;
1819 struct tcp_ao_del cmd;
1820 __u8 prefix;
1821 u16 port;
1822
1823 if (optlen < sizeof(cmd))
1824 return -EINVAL;
1825
1826 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1827 if (err)
1828 return err;
1829
1830 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1831 return -EINVAL;
1832
1833 if (cmd.set_current || cmd.set_rnext) {
1834 if (!tcp_ao_can_set_current_rnext(sk))
1835 return -EINVAL;
1836 }
1837
1838 if (cmd.keyflags & ~TCP_AO_DEL_KEYF_ALL)
1839 return -EINVAL;
1840
1841 /* No sanity check for TCP_AO_KEYF_IFINDEX as if a VRF
1842 * was destroyed, there still should be a way to delete keys,
1843 * that were bound to that l3intf. So, fail late at lookup stage
1844 * if there is no key for that ifindex.
1845 */
1846 if (cmd.ifindex && !(cmd.keyflags & TCP_AO_KEYF_IFINDEX))
1847 return -EINVAL;
1848
1849 ao_info = setsockopt_ao_info(sk);
1850 if (IS_ERR(ao_info))
1851 return PTR_ERR(ao_info);
1852 if (!ao_info)
1853 return -ENOENT;
1854
1855 /* For sockets in TCP_CLOSED it's possible set keys that aren't
1856 * matching the future peer (address/VRF/etc),
1857 * tcp_ao_connect_init() will choose a correct matching MKT
1858 * if there's any.
1859 */
1860 if (cmd.set_current) {
1861 new_current = tcp_ao_established_key(sk, ao_info, cmd.current_key, -1);
1862 if (!new_current)
1863 return -ENOENT;
1864 }
1865 if (cmd.set_rnext) {
1866 new_rnext = tcp_ao_established_key(sk, ao_info, -1, cmd.rnext);
1867 if (!new_rnext)
1868 return -ENOENT;
1869 }
1870 if (cmd.del_async && sk->sk_state != TCP_LISTEN)
1871 return -EINVAL;
1872
1873 if (family == AF_INET) {
1874 struct sockaddr_in *sin = (struct sockaddr_in *)&cmd.addr;
1875
1876 addr = (union tcp_ao_addr *)&sin->sin_addr;
1877 addr_len = sizeof(struct in_addr);
1878 port = ntohs(sin->sin_port);
1879 } else {
1880 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd.addr;
1881 struct in6_addr *addr6 = &sin6->sin6_addr;
1882
1883 if (ipv6_addr_v4mapped(addr6)) {
1884 addr = (union tcp_ao_addr *)&addr6->s6_addr32[3];
1885 addr_len = sizeof(struct in_addr);
1886 family = AF_INET;
1887 } else {
1888 addr = (union tcp_ao_addr *)addr6;
1889 addr_len = sizeof(struct in6_addr);
1890 }
1891 port = ntohs(sin6->sin6_port);
1892 }
1893 prefix = cmd.prefix;
1894
1895 /* Currently matching is not performed on port (or port ranges) */
1896 if (port != 0)
1897 return -EINVAL;
1898
1899 /* We could choose random present key here for current/rnext
1900 * but that's less predictable. Let's be strict and don't
1901 * allow removing a key that's in use. RFC5925 doesn't
1902 * specify how-to coordinate key removal, but says:
1903 * "It is presumed that an MKT affecting a particular
1904 * connection cannot be destroyed during an active connection"
1905 */
1906 hlist_for_each_entry_rcu(key, &ao_info->head, node,
1907 lockdep_sock_is_held(sk)) {
1908 if (cmd.sndid != key->sndid ||
1909 cmd.rcvid != key->rcvid)
1910 continue;
1911
1912 if (family != key->family ||
1913 prefix != key->prefixlen ||
1914 memcmp(addr, &key->addr, addr_len))
1915 continue;
1916
1917 if ((cmd.keyflags & TCP_AO_KEYF_IFINDEX) !=
1918 (key->keyflags & TCP_AO_KEYF_IFINDEX))
1919 continue;
1920
1921 if (key->l3index != l3index)
1922 continue;
1923
1924 if (key == new_current || key == new_rnext)
1925 continue;
1926
1927 return tcp_ao_delete_key(sk, ao_info, cmd.del_async, key,
1928 new_current, new_rnext);
1929 }
1930 return -ENOENT;
1931}
1932
1933/* cmd.ao_required makes a socket TCP-AO only.
1934 * Don't allow any md5 keys for any l3intf on the socket together with it.
1935 * Restricting it early in setsockopt() removes a check for
1936 * ao_info->ao_required on inbound tcp segment fast-path.
1937 */
1938static int tcp_ao_required_verify(struct sock *sk)
1939{
1940#ifdef CONFIG_TCP_MD5SIG
1941 const struct tcp_md5sig_info *md5sig;
1942
1943 if (!static_branch_unlikely(&tcp_md5_needed.key))
1944 return 0;
1945
1946 md5sig = rcu_dereference_check(tcp_sk(sk)->md5sig_info,
1947 lockdep_sock_is_held(sk));
1948 if (!md5sig)
1949 return 0;
1950
1951 if (rcu_dereference_check(hlist_first_rcu(&md5sig->head),
1952 lockdep_sock_is_held(sk)))
1953 return 1;
1954#endif
1955 return 0;
1956}
1957
1958static int tcp_ao_info_cmd(struct sock *sk, unsigned short int family,
1959 sockptr_t optval, int optlen)
1960{
1961 struct tcp_ao_key *new_current = NULL, *new_rnext = NULL;
1962 struct tcp_ao_info *ao_info;
1963 struct tcp_ao_info_opt cmd;
1964 bool first = false;
1965 int err;
1966
1967 if (optlen < sizeof(cmd))
1968 return -EINVAL;
1969
1970 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1971 if (err)
1972 return err;
1973
1974 if (cmd.set_current || cmd.set_rnext) {
1975 if (!tcp_ao_can_set_current_rnext(sk))
1976 return -EINVAL;
1977 }
1978
1979 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1980 return -EINVAL;
1981
1982 ao_info = setsockopt_ao_info(sk);
1983 if (IS_ERR(ao_info))
1984 return PTR_ERR(ao_info);
1985 if (!ao_info) {
1986 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE)))
1987 return -EINVAL;
1988 ao_info = tcp_ao_alloc_info(GFP_KERNEL);
1989 if (!ao_info)
1990 return -ENOMEM;
1991 first = true;
1992 }
1993
1994 if (cmd.ao_required && tcp_ao_required_verify(sk)) {
1995 err = -EKEYREJECTED;
1996 goto out;
1997 }
1998
1999 /* For sockets in TCP_CLOSED it's possible set keys that aren't
2000 * matching the future peer (address/port/VRF/etc),
2001 * tcp_ao_connect_init() will choose a correct matching MKT
2002 * if there's any.
2003 */
2004 if (cmd.set_current) {
2005 new_current = tcp_ao_established_key(sk, ao_info, cmd.current_key, -1);
2006 if (!new_current) {
2007 err = -ENOENT;
2008 goto out;
2009 }
2010 }
2011 if (cmd.set_rnext) {
2012 new_rnext = tcp_ao_established_key(sk, ao_info, -1, cmd.rnext);
2013 if (!new_rnext) {
2014 err = -ENOENT;
2015 goto out;
2016 }
2017 }
2018 if (cmd.set_counters) {
2019 atomic64_set(&ao_info->counters.pkt_good, cmd.pkt_good);
2020 atomic64_set(&ao_info->counters.pkt_bad, cmd.pkt_bad);
2021 atomic64_set(&ao_info->counters.key_not_found, cmd.pkt_key_not_found);
2022 atomic64_set(&ao_info->counters.ao_required, cmd.pkt_ao_required);
2023 atomic64_set(&ao_info->counters.dropped_icmp, cmd.pkt_dropped_icmp);
2024 }
2025
2026 ao_info->ao_required = cmd.ao_required;
2027 ao_info->accept_icmps = cmd.accept_icmps;
2028 if (new_current)
2029 WRITE_ONCE(ao_info->current_key, new_current);
2030 if (new_rnext)
2031 WRITE_ONCE(ao_info->rnext_key, new_rnext);
2032 if (first) {
2033 if (!static_branch_inc(&tcp_ao_needed.key)) {
2034 err = -EUSERS;
2035 goto out;
2036 }
2037 sk_gso_disable(sk);
2038 rcu_assign_pointer(tcp_sk(sk)->ao_info, ao_info);
2039 }
2040 return 0;
2041out:
2042 if (first)
2043 kfree(ao_info);
2044 return err;
2045}
2046
2047int tcp_parse_ao(struct sock *sk, int cmd, unsigned short int family,
2048 sockptr_t optval, int optlen)
2049{
2050 if (WARN_ON_ONCE(family != AF_INET && family != AF_INET6))
2051 return -EAFNOSUPPORT;
2052
2053 switch (cmd) {
2054 case TCP_AO_ADD_KEY:
2055 return tcp_ao_add_cmd(sk, family, optval, optlen);
2056 case TCP_AO_DEL_KEY:
2057 return tcp_ao_del_cmd(sk, family, optval, optlen);
2058 case TCP_AO_INFO:
2059 return tcp_ao_info_cmd(sk, family, optval, optlen);
2060 default:
2061 WARN_ON_ONCE(1);
2062 return -EINVAL;
2063 }
2064}
2065
2066int tcp_v4_parse_ao(struct sock *sk, int cmd, sockptr_t optval, int optlen)
2067{
2068 return tcp_parse_ao(sk, cmd, AF_INET, optval, optlen);
2069}
2070
2071/* tcp_ao_copy_mkts_to_user(ao_info, optval, optlen)
2072 *
2073 * @ao_info: struct tcp_ao_info on the socket that
2074 * socket getsockopt(TCP_AO_GET_KEYS) is executed on
2075 * @optval: pointer to array of tcp_ao_getsockopt structures in user space.
2076 * Must be != NULL.
2077 * @optlen: pointer to size of tcp_ao_getsockopt structure.
2078 * Must be != NULL.
2079 *
2080 * Return value: 0 on success, a negative error number otherwise.
2081 *
2082 * optval points to an array of tcp_ao_getsockopt structures in user space.
2083 * optval[0] is used as both input and output to getsockopt. It determines
2084 * which keys are returned by the kernel.
2085 * optval[0].nkeys is the size of the array in user space. On return it contains
2086 * the number of keys matching the search criteria.
2087 * If tcp_ao_getsockopt::get_all is set, then all keys in the socket are
2088 * returned, otherwise only keys matching <addr, prefix, sndid, rcvid>
2089 * in optval[0] are returned.
2090 * optlen is also used as both input and output. The user provides the size
2091 * of struct tcp_ao_getsockopt in user space, and the kernel returns the size
2092 * of the structure in kernel space.
2093 * The size of struct tcp_ao_getsockopt may differ between user and kernel.
2094 * There are three cases to consider:
2095 * * If usize == ksize, then keys are copied verbatim.
2096 * * If usize < ksize, then the userspace has passed an old struct to a
2097 * newer kernel. The rest of the trailing bytes in optval[0]
2098 * (ksize - usize) are interpreted as 0 by the kernel.
2099 * * If usize > ksize, then the userspace has passed a new struct to an
2100 * older kernel. The trailing bytes unknown to the kernel (usize - ksize)
2101 * are checked to ensure they are zeroed, otherwise -E2BIG is returned.
2102 * On return the kernel fills in min(usize, ksize) in each entry of the array.
2103 * The layout of the fields in the user and kernel structures is expected to
2104 * be the same (including in the 32bit vs 64bit case).
2105 */
2106static int tcp_ao_copy_mkts_to_user(const struct sock *sk,
2107 struct tcp_ao_info *ao_info,
2108 sockptr_t optval, sockptr_t optlen)
2109{
2110 struct tcp_ao_getsockopt opt_in, opt_out;
2111 struct tcp_ao_key *key, *current_key;
2112 bool do_address_matching = true;
2113 union tcp_ao_addr *addr = NULL;
2114 int err, l3index, user_len;
2115 unsigned int max_keys; /* maximum number of keys to copy to user */
2116 size_t out_offset = 0;
2117 size_t bytes_to_write; /* number of bytes to write to user level */
2118 u32 matched_keys; /* keys from ao_info matched so far */
2119 int optlen_out;
2120 __be16 port = 0;
2121
2122 if (copy_from_sockptr(&user_len, optlen, sizeof(int)))
2123 return -EFAULT;
2124
2125 if (user_len <= 0)
2126 return -EINVAL;
2127
2128 memset(&opt_in, 0, sizeof(struct tcp_ao_getsockopt));
2129 err = copy_struct_from_sockptr(&opt_in, sizeof(opt_in),
2130 optval, user_len);
2131 if (err < 0)
2132 return err;
2133
2134 if (opt_in.pkt_good || opt_in.pkt_bad)
2135 return -EINVAL;
2136 if (opt_in.keyflags & ~TCP_AO_GET_KEYF_VALID)
2137 return -EINVAL;
2138 if (opt_in.ifindex && !(opt_in.keyflags & TCP_AO_KEYF_IFINDEX))
2139 return -EINVAL;
2140
2141 if (opt_in.reserved != 0)
2142 return -EINVAL;
2143
2144 max_keys = opt_in.nkeys;
2145 l3index = (opt_in.keyflags & TCP_AO_KEYF_IFINDEX) ? opt_in.ifindex : -1;
2146
2147 if (opt_in.get_all || opt_in.is_current || opt_in.is_rnext) {
2148 if (opt_in.get_all && (opt_in.is_current || opt_in.is_rnext))
2149 return -EINVAL;
2150 do_address_matching = false;
2151 }
2152
2153 switch (opt_in.addr.ss_family) {
2154 case AF_INET: {
2155 struct sockaddr_in *sin;
2156 __be32 mask;
2157
2158 sin = (struct sockaddr_in *)&opt_in.addr;
2159 port = sin->sin_port;
2160 addr = (union tcp_ao_addr *)&sin->sin_addr;
2161
2162 if (opt_in.prefix > 32)
2163 return -EINVAL;
2164
2165 if (ntohl(sin->sin_addr.s_addr) == INADDR_ANY &&
2166 opt_in.prefix != 0)
2167 return -EINVAL;
2168
2169 mask = inet_make_mask(opt_in.prefix);
2170 if (sin->sin_addr.s_addr & ~mask)
2171 return -EINVAL;
2172
2173 break;
2174 }
2175 case AF_INET6: {
2176 struct sockaddr_in6 *sin6;
2177 struct in6_addr *addr6;
2178
2179 sin6 = (struct sockaddr_in6 *)&opt_in.addr;
2180 addr = (union tcp_ao_addr *)&sin6->sin6_addr;
2181 addr6 = &sin6->sin6_addr;
2182 port = sin6->sin6_port;
2183
2184 /* We don't have to change family and @addr here if
2185 * ipv6_addr_v4mapped() like in key adding:
2186 * tcp_ao_key_cmp() does it. Do the sanity checks though.
2187 */
2188 if (opt_in.prefix != 0) {
2189 if (ipv6_addr_v4mapped(addr6)) {
2190 __be32 mask, addr4 = addr6->s6_addr32[3];
2191
2192 if (opt_in.prefix > 32 ||
2193 ntohl(addr4) == INADDR_ANY)
2194 return -EINVAL;
2195 mask = inet_make_mask(opt_in.prefix);
2196 if (addr4 & ~mask)
2197 return -EINVAL;
2198 } else {
2199 struct in6_addr pfx;
2200
2201 if (ipv6_addr_any(addr6) ||
2202 opt_in.prefix > 128)
2203 return -EINVAL;
2204
2205 ipv6_addr_prefix(&pfx, addr6, opt_in.prefix);
2206 if (ipv6_addr_cmp(&pfx, addr6))
2207 return -EINVAL;
2208 }
2209 } else if (!ipv6_addr_any(addr6)) {
2210 return -EINVAL;
2211 }
2212 break;
2213 }
2214 case 0:
2215 if (!do_address_matching)
2216 break;
2217 fallthrough;
2218 default:
2219 return -EAFNOSUPPORT;
2220 }
2221
2222 if (!do_address_matching) {
2223 /* We could just ignore those, but let's do stricter checks */
2224 if (addr || port)
2225 return -EINVAL;
2226 if (opt_in.prefix || opt_in.sndid || opt_in.rcvid)
2227 return -EINVAL;
2228 }
2229
2230 bytes_to_write = min_t(int, user_len, sizeof(struct tcp_ao_getsockopt));
2231 matched_keys = 0;
2232 /* May change in RX, while we're dumping, pre-fetch it */
2233 current_key = READ_ONCE(ao_info->current_key);
2234
2235 hlist_for_each_entry_rcu(key, &ao_info->head, node,
2236 lockdep_sock_is_held(sk)) {
2237 if (opt_in.get_all)
2238 goto match;
2239
2240 if (opt_in.is_current || opt_in.is_rnext) {
2241 if (opt_in.is_current && key == current_key)
2242 goto match;
2243 if (opt_in.is_rnext && key == ao_info->rnext_key)
2244 goto match;
2245 continue;
2246 }
2247
2248 if (tcp_ao_key_cmp(key, l3index, addr, opt_in.prefix,
2249 opt_in.addr.ss_family,
2250 opt_in.sndid, opt_in.rcvid) != 0)
2251 continue;
2252match:
2253 matched_keys++;
2254 if (matched_keys > max_keys)
2255 continue;
2256
2257 memset(&opt_out, 0, sizeof(struct tcp_ao_getsockopt));
2258
2259 if (key->family == AF_INET) {
2260 struct sockaddr_in *sin_out = (struct sockaddr_in *)&opt_out.addr;
2261
2262 sin_out->sin_family = key->family;
2263 sin_out->sin_port = 0;
2264 memcpy(&sin_out->sin_addr, &key->addr, sizeof(struct in_addr));
2265 } else {
2266 struct sockaddr_in6 *sin6_out = (struct sockaddr_in6 *)&opt_out.addr;
2267
2268 sin6_out->sin6_family = key->family;
2269 sin6_out->sin6_port = 0;
2270 memcpy(&sin6_out->sin6_addr, &key->addr, sizeof(struct in6_addr));
2271 }
2272 opt_out.sndid = key->sndid;
2273 opt_out.rcvid = key->rcvid;
2274 opt_out.prefix = key->prefixlen;
2275 opt_out.keyflags = key->keyflags;
2276 opt_out.is_current = (key == current_key);
2277 opt_out.is_rnext = (key == ao_info->rnext_key);
2278 opt_out.nkeys = 0;
2279 opt_out.maclen = key->maclen;
2280 opt_out.keylen = key->keylen;
2281 opt_out.ifindex = key->l3index;
2282 opt_out.pkt_good = atomic64_read(&key->pkt_good);
2283 opt_out.pkt_bad = atomic64_read(&key->pkt_bad);
2284 memcpy(&opt_out.key, key->key, key->keylen);
2285 tcp_sigpool_algo(key->tcp_sigpool_id, opt_out.alg_name, 64);
2286
2287 /* Copy key to user */
2288 if (copy_to_sockptr_offset(optval, out_offset,
2289 &opt_out, bytes_to_write))
2290 return -EFAULT;
2291 out_offset += user_len;
2292 }
2293
2294 optlen_out = (int)sizeof(struct tcp_ao_getsockopt);
2295 if (copy_to_sockptr(optlen, &optlen_out, sizeof(int)))
2296 return -EFAULT;
2297
2298 out_offset = offsetof(struct tcp_ao_getsockopt, nkeys);
2299 if (copy_to_sockptr_offset(optval, out_offset,
2300 &matched_keys, sizeof(u32)))
2301 return -EFAULT;
2302
2303 return 0;
2304}
2305
2306int tcp_ao_get_mkts(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2307{
2308 struct tcp_ao_info *ao_info;
2309
2310 ao_info = setsockopt_ao_info(sk);
2311 if (IS_ERR(ao_info))
2312 return PTR_ERR(ao_info);
2313 if (!ao_info)
2314 return -ENOENT;
2315
2316 return tcp_ao_copy_mkts_to_user(sk, ao_info, optval, optlen);
2317}
2318
2319int tcp_ao_get_sock_info(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2320{
2321 struct tcp_ao_info_opt out, in = {};
2322 struct tcp_ao_key *current_key;
2323 struct tcp_ao_info *ao;
2324 int err, len;
2325
2326 if (copy_from_sockptr(&len, optlen, sizeof(int)))
2327 return -EFAULT;
2328
2329 if (len <= 0)
2330 return -EINVAL;
2331
2332 /* Copying this "in" only to check ::reserved, ::reserved2,
2333 * that may be needed to extend (struct tcp_ao_info_opt) and
2334 * what getsockopt() provides in future.
2335 */
2336 err = copy_struct_from_sockptr(&in, sizeof(in), optval, len);
2337 if (err)
2338 return err;
2339
2340 if (in.reserved != 0 || in.reserved2 != 0)
2341 return -EINVAL;
2342
2343 ao = setsockopt_ao_info(sk);
2344 if (IS_ERR(ao))
2345 return PTR_ERR(ao);
2346 if (!ao)
2347 return -ENOENT;
2348
2349 memset(&out, 0, sizeof(out));
2350 out.ao_required = ao->ao_required;
2351 out.accept_icmps = ao->accept_icmps;
2352 out.pkt_good = atomic64_read(&ao->counters.pkt_good);
2353 out.pkt_bad = atomic64_read(&ao->counters.pkt_bad);
2354 out.pkt_key_not_found = atomic64_read(&ao->counters.key_not_found);
2355 out.pkt_ao_required = atomic64_read(&ao->counters.ao_required);
2356 out.pkt_dropped_icmp = atomic64_read(&ao->counters.dropped_icmp);
2357
2358 current_key = READ_ONCE(ao->current_key);
2359 if (current_key) {
2360 out.set_current = 1;
2361 out.current_key = current_key->sndid;
2362 }
2363 if (ao->rnext_key) {
2364 out.set_rnext = 1;
2365 out.rnext = ao->rnext_key->rcvid;
2366 }
2367
2368 if (copy_to_sockptr(optval, &out, min_t(int, len, sizeof(out))))
2369 return -EFAULT;
2370
2371 return 0;
2372}
2373
2374int tcp_ao_set_repair(struct sock *sk, sockptr_t optval, unsigned int optlen)
2375{
2376 struct tcp_sock *tp = tcp_sk(sk);
2377 struct tcp_ao_repair cmd;
2378 struct tcp_ao_key *key;
2379 struct tcp_ao_info *ao;
2380 int err;
2381
2382 if (optlen < sizeof(cmd))
2383 return -EINVAL;
2384
2385 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
2386 if (err)
2387 return err;
2388
2389 if (!tp->repair)
2390 return -EPERM;
2391
2392 ao = setsockopt_ao_info(sk);
2393 if (IS_ERR(ao))
2394 return PTR_ERR(ao);
2395 if (!ao)
2396 return -ENOENT;
2397
2398 WRITE_ONCE(ao->lisn, cmd.snt_isn);
2399 WRITE_ONCE(ao->risn, cmd.rcv_isn);
2400 WRITE_ONCE(ao->snd_sne, cmd.snd_sne);
2401 WRITE_ONCE(ao->rcv_sne, cmd.rcv_sne);
2402
2403 hlist_for_each_entry_rcu(key, &ao->head, node, lockdep_sock_is_held(sk))
2404 tcp_ao_cache_traffic_keys(sk, ao, key);
2405
2406 return 0;
2407}
2408
2409int tcp_ao_get_repair(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2410{
2411 struct tcp_sock *tp = tcp_sk(sk);
2412 struct tcp_ao_repair opt;
2413 struct tcp_ao_info *ao;
2414 int len;
2415
2416 if (copy_from_sockptr(&len, optlen, sizeof(int)))
2417 return -EFAULT;
2418
2419 if (len <= 0)
2420 return -EINVAL;
2421
2422 if (!tp->repair)
2423 return -EPERM;
2424
2425 rcu_read_lock();
2426 ao = getsockopt_ao_info(sk);
2427 if (IS_ERR_OR_NULL(ao)) {
2428 rcu_read_unlock();
2429 return ao ? PTR_ERR(ao) : -ENOENT;
2430 }
2431
2432 opt.snt_isn = ao->lisn;
2433 opt.rcv_isn = ao->risn;
2434 opt.snd_sne = READ_ONCE(ao->snd_sne);
2435 opt.rcv_sne = READ_ONCE(ao->rcv_sne);
2436 rcu_read_unlock();
2437
2438 if (copy_to_sockptr(optval, &opt, min_t(int, len, sizeof(opt))))
2439 return -EFAULT;
2440 return 0;
2441}
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * INET An implementation of the TCP Authentication Option (TCP-AO).
4 * See RFC5925.
5 *
6 * Authors: Dmitry Safonov <dima@arista.com>
7 * Francesco Ruggeri <fruggeri@arista.com>
8 * Salam Noureddine <noureddine@arista.com>
9 */
10#define pr_fmt(fmt) "TCP: " fmt
11
12#include <crypto/hash.h>
13#include <linux/inetdevice.h>
14#include <linux/tcp.h>
15
16#include <net/tcp.h>
17#include <net/ipv6.h>
18#include <net/icmp.h>
19
20DEFINE_STATIC_KEY_DEFERRED_FALSE(tcp_ao_needed, HZ);
21
22int tcp_ao_calc_traffic_key(struct tcp_ao_key *mkt, u8 *key, void *ctx,
23 unsigned int len, struct tcp_sigpool *hp)
24{
25 struct scatterlist sg;
26 int ret;
27
28 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp->req),
29 mkt->key, mkt->keylen))
30 goto clear_hash;
31
32 ret = crypto_ahash_init(hp->req);
33 if (ret)
34 goto clear_hash;
35
36 sg_init_one(&sg, ctx, len);
37 ahash_request_set_crypt(hp->req, &sg, key, len);
38 crypto_ahash_update(hp->req);
39
40 ret = crypto_ahash_final(hp->req);
41 if (ret)
42 goto clear_hash;
43
44 return 0;
45clear_hash:
46 memset(key, 0, tcp_ao_digest_size(mkt));
47 return 1;
48}
49
50bool tcp_ao_ignore_icmp(const struct sock *sk, int family, int type, int code)
51{
52 bool ignore_icmp = false;
53 struct tcp_ao_info *ao;
54
55 if (!static_branch_unlikely(&tcp_ao_needed.key))
56 return false;
57
58 /* RFC5925, 7.8:
59 * >> A TCP-AO implementation MUST default to ignore incoming ICMPv4
60 * messages of Type 3 (destination unreachable), Codes 2-4 (protocol
61 * unreachable, port unreachable, and fragmentation needed -- ’hard
62 * errors’), and ICMPv6 Type 1 (destination unreachable), Code 1
63 * (administratively prohibited) and Code 4 (port unreachable) intended
64 * for connections in synchronized states (ESTABLISHED, FIN-WAIT-1, FIN-
65 * WAIT-2, CLOSE-WAIT, CLOSING, LAST-ACK, TIME-WAIT) that match MKTs.
66 */
67 if (family == AF_INET) {
68 if (type != ICMP_DEST_UNREACH)
69 return false;
70 if (code < ICMP_PROT_UNREACH || code > ICMP_FRAG_NEEDED)
71 return false;
72 } else {
73 if (type != ICMPV6_DEST_UNREACH)
74 return false;
75 if (code != ICMPV6_ADM_PROHIBITED && code != ICMPV6_PORT_UNREACH)
76 return false;
77 }
78
79 rcu_read_lock();
80 switch (sk->sk_state) {
81 case TCP_TIME_WAIT:
82 ao = rcu_dereference(tcp_twsk(sk)->ao_info);
83 break;
84 case TCP_SYN_SENT:
85 case TCP_SYN_RECV:
86 case TCP_LISTEN:
87 case TCP_NEW_SYN_RECV:
88 /* RFC5925 specifies to ignore ICMPs *only* on connections
89 * in synchronized states.
90 */
91 rcu_read_unlock();
92 return false;
93 default:
94 ao = rcu_dereference(tcp_sk(sk)->ao_info);
95 }
96
97 if (ao && !ao->accept_icmps) {
98 ignore_icmp = true;
99 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAODROPPEDICMPS);
100 atomic64_inc(&ao->counters.dropped_icmp);
101 }
102 rcu_read_unlock();
103
104 return ignore_icmp;
105}
106
107/* Optimized version of tcp_ao_do_lookup(): only for sockets for which
108 * it's known that the keys in ao_info are matching peer's
109 * family/address/VRF/etc.
110 */
111struct tcp_ao_key *tcp_ao_established_key(struct tcp_ao_info *ao,
112 int sndid, int rcvid)
113{
114 struct tcp_ao_key *key;
115
116 hlist_for_each_entry_rcu(key, &ao->head, node) {
117 if ((sndid >= 0 && key->sndid != sndid) ||
118 (rcvid >= 0 && key->rcvid != rcvid))
119 continue;
120 return key;
121 }
122
123 return NULL;
124}
125
126static int ipv4_prefix_cmp(const struct in_addr *addr1,
127 const struct in_addr *addr2,
128 unsigned int prefixlen)
129{
130 __be32 mask = inet_make_mask(prefixlen);
131 __be32 a1 = addr1->s_addr & mask;
132 __be32 a2 = addr2->s_addr & mask;
133
134 if (a1 == a2)
135 return 0;
136 return memcmp(&a1, &a2, sizeof(a1));
137}
138
139static int __tcp_ao_key_cmp(const struct tcp_ao_key *key, int l3index,
140 const union tcp_ao_addr *addr, u8 prefixlen,
141 int family, int sndid, int rcvid)
142{
143 if (sndid >= 0 && key->sndid != sndid)
144 return (key->sndid > sndid) ? 1 : -1;
145 if (rcvid >= 0 && key->rcvid != rcvid)
146 return (key->rcvid > rcvid) ? 1 : -1;
147 if (l3index >= 0 && (key->keyflags & TCP_AO_KEYF_IFINDEX)) {
148 if (key->l3index != l3index)
149 return (key->l3index > l3index) ? 1 : -1;
150 }
151
152 if (family == AF_UNSPEC)
153 return 0;
154 if (key->family != family)
155 return (key->family > family) ? 1 : -1;
156
157 if (family == AF_INET) {
158 if (ntohl(key->addr.a4.s_addr) == INADDR_ANY)
159 return 0;
160 if (ntohl(addr->a4.s_addr) == INADDR_ANY)
161 return 0;
162 return ipv4_prefix_cmp(&key->addr.a4, &addr->a4, prefixlen);
163#if IS_ENABLED(CONFIG_IPV6)
164 } else {
165 if (ipv6_addr_any(&key->addr.a6) || ipv6_addr_any(&addr->a6))
166 return 0;
167 if (ipv6_prefix_equal(&key->addr.a6, &addr->a6, prefixlen))
168 return 0;
169 return memcmp(&key->addr.a6, &addr->a6, sizeof(addr->a6));
170#endif
171 }
172 return -1;
173}
174
175static int tcp_ao_key_cmp(const struct tcp_ao_key *key, int l3index,
176 const union tcp_ao_addr *addr, u8 prefixlen,
177 int family, int sndid, int rcvid)
178{
179#if IS_ENABLED(CONFIG_IPV6)
180 if (family == AF_INET6 && ipv6_addr_v4mapped(&addr->a6)) {
181 __be32 addr4 = addr->a6.s6_addr32[3];
182
183 return __tcp_ao_key_cmp(key, l3index,
184 (union tcp_ao_addr *)&addr4,
185 prefixlen, AF_INET, sndid, rcvid);
186 }
187#endif
188 return __tcp_ao_key_cmp(key, l3index, addr,
189 prefixlen, family, sndid, rcvid);
190}
191
192static struct tcp_ao_key *__tcp_ao_do_lookup(const struct sock *sk, int l3index,
193 const union tcp_ao_addr *addr, int family, u8 prefix,
194 int sndid, int rcvid)
195{
196 struct tcp_ao_key *key;
197 struct tcp_ao_info *ao;
198
199 if (!static_branch_unlikely(&tcp_ao_needed.key))
200 return NULL;
201
202 ao = rcu_dereference_check(tcp_sk(sk)->ao_info,
203 lockdep_sock_is_held(sk));
204 if (!ao)
205 return NULL;
206
207 hlist_for_each_entry_rcu(key, &ao->head, node) {
208 u8 prefixlen = min(prefix, key->prefixlen);
209
210 if (!tcp_ao_key_cmp(key, l3index, addr, prefixlen,
211 family, sndid, rcvid))
212 return key;
213 }
214 return NULL;
215}
216
217struct tcp_ao_key *tcp_ao_do_lookup(const struct sock *sk, int l3index,
218 const union tcp_ao_addr *addr,
219 int family, int sndid, int rcvid)
220{
221 return __tcp_ao_do_lookup(sk, l3index, addr, family, U8_MAX, sndid, rcvid);
222}
223
224static struct tcp_ao_info *tcp_ao_alloc_info(gfp_t flags)
225{
226 struct tcp_ao_info *ao;
227
228 ao = kzalloc(sizeof(*ao), flags);
229 if (!ao)
230 return NULL;
231 INIT_HLIST_HEAD(&ao->head);
232 refcount_set(&ao->refcnt, 1);
233
234 return ao;
235}
236
237static void tcp_ao_link_mkt(struct tcp_ao_info *ao, struct tcp_ao_key *mkt)
238{
239 hlist_add_head_rcu(&mkt->node, &ao->head);
240}
241
242static struct tcp_ao_key *tcp_ao_copy_key(struct sock *sk,
243 struct tcp_ao_key *key)
244{
245 struct tcp_ao_key *new_key;
246
247 new_key = sock_kmalloc(sk, tcp_ao_sizeof_key(key),
248 GFP_ATOMIC);
249 if (!new_key)
250 return NULL;
251
252 *new_key = *key;
253 INIT_HLIST_NODE(&new_key->node);
254 tcp_sigpool_get(new_key->tcp_sigpool_id);
255 atomic64_set(&new_key->pkt_good, 0);
256 atomic64_set(&new_key->pkt_bad, 0);
257
258 return new_key;
259}
260
261static void tcp_ao_key_free_rcu(struct rcu_head *head)
262{
263 struct tcp_ao_key *key = container_of(head, struct tcp_ao_key, rcu);
264
265 tcp_sigpool_release(key->tcp_sigpool_id);
266 kfree_sensitive(key);
267}
268
269void tcp_ao_destroy_sock(struct sock *sk, bool twsk)
270{
271 struct tcp_ao_info *ao;
272 struct tcp_ao_key *key;
273 struct hlist_node *n;
274
275 if (twsk) {
276 ao = rcu_dereference_protected(tcp_twsk(sk)->ao_info, 1);
277 tcp_twsk(sk)->ao_info = NULL;
278 } else {
279 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info, 1);
280 tcp_sk(sk)->ao_info = NULL;
281 }
282
283 if (!ao || !refcount_dec_and_test(&ao->refcnt))
284 return;
285
286 hlist_for_each_entry_safe(key, n, &ao->head, node) {
287 hlist_del_rcu(&key->node);
288 if (!twsk)
289 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
290 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
291 }
292
293 kfree_rcu(ao, rcu);
294 static_branch_slow_dec_deferred(&tcp_ao_needed);
295}
296
297void tcp_ao_time_wait(struct tcp_timewait_sock *tcptw, struct tcp_sock *tp)
298{
299 struct tcp_ao_info *ao_info = rcu_dereference_protected(tp->ao_info, 1);
300
301 if (ao_info) {
302 struct tcp_ao_key *key;
303 struct hlist_node *n;
304 int omem = 0;
305
306 hlist_for_each_entry_safe(key, n, &ao_info->head, node) {
307 omem += tcp_ao_sizeof_key(key);
308 }
309
310 refcount_inc(&ao_info->refcnt);
311 atomic_sub(omem, &(((struct sock *)tp)->sk_omem_alloc));
312 rcu_assign_pointer(tcptw->ao_info, ao_info);
313 } else {
314 tcptw->ao_info = NULL;
315 }
316}
317
318/* 4 tuple and ISNs are expected in NBO */
319static int tcp_v4_ao_calc_key(struct tcp_ao_key *mkt, u8 *key,
320 __be32 saddr, __be32 daddr,
321 __be16 sport, __be16 dport,
322 __be32 sisn, __be32 disn)
323{
324 /* See RFC5926 3.1.1 */
325 struct kdf_input_block {
326 u8 counter;
327 u8 label[6];
328 struct tcp4_ao_context ctx;
329 __be16 outlen;
330 } __packed * tmp;
331 struct tcp_sigpool hp;
332 int err;
333
334 err = tcp_sigpool_start(mkt->tcp_sigpool_id, &hp);
335 if (err)
336 return err;
337
338 tmp = hp.scratch;
339 tmp->counter = 1;
340 memcpy(tmp->label, "TCP-AO", 6);
341 tmp->ctx.saddr = saddr;
342 tmp->ctx.daddr = daddr;
343 tmp->ctx.sport = sport;
344 tmp->ctx.dport = dport;
345 tmp->ctx.sisn = sisn;
346 tmp->ctx.disn = disn;
347 tmp->outlen = htons(tcp_ao_digest_size(mkt) * 8); /* in bits */
348
349 err = tcp_ao_calc_traffic_key(mkt, key, tmp, sizeof(*tmp), &hp);
350 tcp_sigpool_end(&hp);
351
352 return err;
353}
354
355int tcp_v4_ao_calc_key_sk(struct tcp_ao_key *mkt, u8 *key,
356 const struct sock *sk,
357 __be32 sisn, __be32 disn, bool send)
358{
359 if (send)
360 return tcp_v4_ao_calc_key(mkt, key, sk->sk_rcv_saddr,
361 sk->sk_daddr, htons(sk->sk_num),
362 sk->sk_dport, sisn, disn);
363 else
364 return tcp_v4_ao_calc_key(mkt, key, sk->sk_daddr,
365 sk->sk_rcv_saddr, sk->sk_dport,
366 htons(sk->sk_num), disn, sisn);
367}
368
369static int tcp_ao_calc_key_sk(struct tcp_ao_key *mkt, u8 *key,
370 const struct sock *sk,
371 __be32 sisn, __be32 disn, bool send)
372{
373 if (mkt->family == AF_INET)
374 return tcp_v4_ao_calc_key_sk(mkt, key, sk, sisn, disn, send);
375#if IS_ENABLED(CONFIG_IPV6)
376 else if (mkt->family == AF_INET6)
377 return tcp_v6_ao_calc_key_sk(mkt, key, sk, sisn, disn, send);
378#endif
379 else
380 return -EOPNOTSUPP;
381}
382
383int tcp_v4_ao_calc_key_rsk(struct tcp_ao_key *mkt, u8 *key,
384 struct request_sock *req)
385{
386 struct inet_request_sock *ireq = inet_rsk(req);
387
388 return tcp_v4_ao_calc_key(mkt, key,
389 ireq->ir_loc_addr, ireq->ir_rmt_addr,
390 htons(ireq->ir_num), ireq->ir_rmt_port,
391 htonl(tcp_rsk(req)->snt_isn),
392 htonl(tcp_rsk(req)->rcv_isn));
393}
394
395static int tcp_v4_ao_calc_key_skb(struct tcp_ao_key *mkt, u8 *key,
396 const struct sk_buff *skb,
397 __be32 sisn, __be32 disn)
398{
399 const struct iphdr *iph = ip_hdr(skb);
400 const struct tcphdr *th = tcp_hdr(skb);
401
402 return tcp_v4_ao_calc_key(mkt, key, iph->saddr, iph->daddr,
403 th->source, th->dest, sisn, disn);
404}
405
406static int tcp_ao_calc_key_skb(struct tcp_ao_key *mkt, u8 *key,
407 const struct sk_buff *skb,
408 __be32 sisn, __be32 disn, int family)
409{
410 if (family == AF_INET)
411 return tcp_v4_ao_calc_key_skb(mkt, key, skb, sisn, disn);
412#if IS_ENABLED(CONFIG_IPV6)
413 else if (family == AF_INET6)
414 return tcp_v6_ao_calc_key_skb(mkt, key, skb, sisn, disn);
415#endif
416 return -EAFNOSUPPORT;
417}
418
419static int tcp_v4_ao_hash_pseudoheader(struct tcp_sigpool *hp,
420 __be32 daddr, __be32 saddr,
421 int nbytes)
422{
423 struct tcp4_pseudohdr *bp;
424 struct scatterlist sg;
425
426 bp = hp->scratch;
427 bp->saddr = saddr;
428 bp->daddr = daddr;
429 bp->pad = 0;
430 bp->protocol = IPPROTO_TCP;
431 bp->len = cpu_to_be16(nbytes);
432
433 sg_init_one(&sg, bp, sizeof(*bp));
434 ahash_request_set_crypt(hp->req, &sg, NULL, sizeof(*bp));
435 return crypto_ahash_update(hp->req);
436}
437
438static int tcp_ao_hash_pseudoheader(unsigned short int family,
439 const struct sock *sk,
440 const struct sk_buff *skb,
441 struct tcp_sigpool *hp, int nbytes)
442{
443 const struct tcphdr *th = tcp_hdr(skb);
444
445 /* TODO: Can we rely on checksum being zero to mean outbound pkt? */
446 if (!th->check) {
447 if (family == AF_INET)
448 return tcp_v4_ao_hash_pseudoheader(hp, sk->sk_daddr,
449 sk->sk_rcv_saddr, skb->len);
450#if IS_ENABLED(CONFIG_IPV6)
451 else if (family == AF_INET6)
452 return tcp_v6_ao_hash_pseudoheader(hp, &sk->sk_v6_daddr,
453 &sk->sk_v6_rcv_saddr, skb->len);
454#endif
455 else
456 return -EAFNOSUPPORT;
457 }
458
459 if (family == AF_INET) {
460 const struct iphdr *iph = ip_hdr(skb);
461
462 return tcp_v4_ao_hash_pseudoheader(hp, iph->daddr,
463 iph->saddr, skb->len);
464#if IS_ENABLED(CONFIG_IPV6)
465 } else if (family == AF_INET6) {
466 const struct ipv6hdr *iph = ipv6_hdr(skb);
467
468 return tcp_v6_ao_hash_pseudoheader(hp, &iph->daddr,
469 &iph->saddr, skb->len);
470#endif
471 }
472 return -EAFNOSUPPORT;
473}
474
475u32 tcp_ao_compute_sne(u32 next_sne, u32 next_seq, u32 seq)
476{
477 u32 sne = next_sne;
478
479 if (before(seq, next_seq)) {
480 if (seq > next_seq)
481 sne--;
482 } else {
483 if (seq < next_seq)
484 sne++;
485 }
486
487 return sne;
488}
489
490/* tcp_ao_hash_sne(struct tcp_sigpool *hp)
491 * @hp - used for hashing
492 * @sne - sne value
493 */
494static int tcp_ao_hash_sne(struct tcp_sigpool *hp, u32 sne)
495{
496 struct scatterlist sg;
497 __be32 *bp;
498
499 bp = (__be32 *)hp->scratch;
500 *bp = htonl(sne);
501
502 sg_init_one(&sg, bp, sizeof(*bp));
503 ahash_request_set_crypt(hp->req, &sg, NULL, sizeof(*bp));
504 return crypto_ahash_update(hp->req);
505}
506
507static int tcp_ao_hash_header(struct tcp_sigpool *hp,
508 const struct tcphdr *th,
509 bool exclude_options, u8 *hash,
510 int hash_offset, int hash_len)
511{
512 int err, len = th->doff << 2;
513 struct scatterlist sg;
514 u8 *hdr = hp->scratch;
515
516 /* We are not allowed to change tcphdr, make a local copy */
517 if (exclude_options) {
518 len = sizeof(*th) + sizeof(struct tcp_ao_hdr) + hash_len;
519 memcpy(hdr, th, sizeof(*th));
520 memcpy(hdr + sizeof(*th),
521 (u8 *)th + hash_offset - sizeof(struct tcp_ao_hdr),
522 sizeof(struct tcp_ao_hdr));
523 memset(hdr + sizeof(*th) + sizeof(struct tcp_ao_hdr),
524 0, hash_len);
525 ((struct tcphdr *)hdr)->check = 0;
526 } else {
527 len = th->doff << 2;
528 memcpy(hdr, th, len);
529 /* zero out tcp-ao hash */
530 ((struct tcphdr *)hdr)->check = 0;
531 memset(hdr + hash_offset, 0, hash_len);
532 }
533
534 sg_init_one(&sg, hdr, len);
535 ahash_request_set_crypt(hp->req, &sg, NULL, len);
536 err = crypto_ahash_update(hp->req);
537 WARN_ON_ONCE(err != 0);
538 return err;
539}
540
541int tcp_ao_hash_hdr(unsigned short int family, char *ao_hash,
542 struct tcp_ao_key *key, const u8 *tkey,
543 const union tcp_ao_addr *daddr,
544 const union tcp_ao_addr *saddr,
545 const struct tcphdr *th, u32 sne)
546{
547 int tkey_len = tcp_ao_digest_size(key);
548 int hash_offset = ao_hash - (char *)th;
549 struct tcp_sigpool hp;
550 void *hash_buf = NULL;
551
552 hash_buf = kmalloc(tkey_len, GFP_ATOMIC);
553 if (!hash_buf)
554 goto clear_hash_noput;
555
556 if (tcp_sigpool_start(key->tcp_sigpool_id, &hp))
557 goto clear_hash_noput;
558
559 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp.req), tkey, tkey_len))
560 goto clear_hash;
561
562 if (crypto_ahash_init(hp.req))
563 goto clear_hash;
564
565 if (tcp_ao_hash_sne(&hp, sne))
566 goto clear_hash;
567 if (family == AF_INET) {
568 if (tcp_v4_ao_hash_pseudoheader(&hp, daddr->a4.s_addr,
569 saddr->a4.s_addr, th->doff * 4))
570 goto clear_hash;
571#if IS_ENABLED(CONFIG_IPV6)
572 } else if (family == AF_INET6) {
573 if (tcp_v6_ao_hash_pseudoheader(&hp, &daddr->a6,
574 &saddr->a6, th->doff * 4))
575 goto clear_hash;
576#endif
577 } else {
578 WARN_ON_ONCE(1);
579 goto clear_hash;
580 }
581 if (tcp_ao_hash_header(&hp, th,
582 !!(key->keyflags & TCP_AO_KEYF_EXCLUDE_OPT),
583 ao_hash, hash_offset, tcp_ao_maclen(key)))
584 goto clear_hash;
585 ahash_request_set_crypt(hp.req, NULL, hash_buf, 0);
586 if (crypto_ahash_final(hp.req))
587 goto clear_hash;
588
589 memcpy(ao_hash, hash_buf, tcp_ao_maclen(key));
590 tcp_sigpool_end(&hp);
591 kfree(hash_buf);
592 return 0;
593
594clear_hash:
595 tcp_sigpool_end(&hp);
596clear_hash_noput:
597 memset(ao_hash, 0, tcp_ao_maclen(key));
598 kfree(hash_buf);
599 return 1;
600}
601
602int tcp_ao_hash_skb(unsigned short int family,
603 char *ao_hash, struct tcp_ao_key *key,
604 const struct sock *sk, const struct sk_buff *skb,
605 const u8 *tkey, int hash_offset, u32 sne)
606{
607 const struct tcphdr *th = tcp_hdr(skb);
608 int tkey_len = tcp_ao_digest_size(key);
609 struct tcp_sigpool hp;
610 void *hash_buf = NULL;
611
612 hash_buf = kmalloc(tkey_len, GFP_ATOMIC);
613 if (!hash_buf)
614 goto clear_hash_noput;
615
616 if (tcp_sigpool_start(key->tcp_sigpool_id, &hp))
617 goto clear_hash_noput;
618
619 if (crypto_ahash_setkey(crypto_ahash_reqtfm(hp.req), tkey, tkey_len))
620 goto clear_hash;
621
622 /* For now use sha1 by default. Depends on alg in tcp_ao_key */
623 if (crypto_ahash_init(hp.req))
624 goto clear_hash;
625
626 if (tcp_ao_hash_sne(&hp, sne))
627 goto clear_hash;
628 if (tcp_ao_hash_pseudoheader(family, sk, skb, &hp, skb->len))
629 goto clear_hash;
630 if (tcp_ao_hash_header(&hp, th,
631 !!(key->keyflags & TCP_AO_KEYF_EXCLUDE_OPT),
632 ao_hash, hash_offset, tcp_ao_maclen(key)))
633 goto clear_hash;
634 if (tcp_sigpool_hash_skb_data(&hp, skb, th->doff << 2))
635 goto clear_hash;
636 ahash_request_set_crypt(hp.req, NULL, hash_buf, 0);
637 if (crypto_ahash_final(hp.req))
638 goto clear_hash;
639
640 memcpy(ao_hash, hash_buf, tcp_ao_maclen(key));
641 tcp_sigpool_end(&hp);
642 kfree(hash_buf);
643 return 0;
644
645clear_hash:
646 tcp_sigpool_end(&hp);
647clear_hash_noput:
648 memset(ao_hash, 0, tcp_ao_maclen(key));
649 kfree(hash_buf);
650 return 1;
651}
652
653int tcp_v4_ao_hash_skb(char *ao_hash, struct tcp_ao_key *key,
654 const struct sock *sk, const struct sk_buff *skb,
655 const u8 *tkey, int hash_offset, u32 sne)
656{
657 return tcp_ao_hash_skb(AF_INET, ao_hash, key, sk, skb,
658 tkey, hash_offset, sne);
659}
660
661int tcp_v4_ao_synack_hash(char *ao_hash, struct tcp_ao_key *ao_key,
662 struct request_sock *req, const struct sk_buff *skb,
663 int hash_offset, u32 sne)
664{
665 void *hash_buf = NULL;
666 int err;
667
668 hash_buf = kmalloc(tcp_ao_digest_size(ao_key), GFP_ATOMIC);
669 if (!hash_buf)
670 return -ENOMEM;
671
672 err = tcp_v4_ao_calc_key_rsk(ao_key, hash_buf, req);
673 if (err)
674 goto out;
675
676 err = tcp_ao_hash_skb(AF_INET, ao_hash, ao_key, req_to_sk(req), skb,
677 hash_buf, hash_offset, sne);
678out:
679 kfree(hash_buf);
680 return err;
681}
682
683struct tcp_ao_key *tcp_v4_ao_lookup_rsk(const struct sock *sk,
684 struct request_sock *req,
685 int sndid, int rcvid)
686{
687 struct inet_request_sock *ireq = inet_rsk(req);
688 union tcp_ao_addr *addr = (union tcp_ao_addr *)&ireq->ir_rmt_addr;
689 int l3index;
690
691 l3index = l3mdev_master_ifindex_by_index(sock_net(sk), ireq->ir_iif);
692 return tcp_ao_do_lookup(sk, l3index, addr, AF_INET, sndid, rcvid);
693}
694
695struct tcp_ao_key *tcp_v4_ao_lookup(const struct sock *sk, struct sock *addr_sk,
696 int sndid, int rcvid)
697{
698 int l3index = l3mdev_master_ifindex_by_index(sock_net(sk),
699 addr_sk->sk_bound_dev_if);
700 union tcp_ao_addr *addr = (union tcp_ao_addr *)&addr_sk->sk_daddr;
701
702 return tcp_ao_do_lookup(sk, l3index, addr, AF_INET, sndid, rcvid);
703}
704
705int tcp_ao_prepare_reset(const struct sock *sk, struct sk_buff *skb,
706 const struct tcp_ao_hdr *aoh, int l3index, u32 seq,
707 struct tcp_ao_key **key, char **traffic_key,
708 bool *allocated_traffic_key, u8 *keyid, u32 *sne)
709{
710 const struct tcphdr *th = tcp_hdr(skb);
711 struct tcp_ao_info *ao_info;
712
713 *allocated_traffic_key = false;
714 /* If there's no socket - than initial sisn/disn are unknown.
715 * Drop the segment. RFC5925 (7.7) advises to require graceful
716 * restart [RFC4724]. Alternatively, the RFC5925 advises to
717 * save/restore traffic keys before/after reboot.
718 * Linux TCP-AO support provides TCP_AO_ADD_KEY and TCP_AO_REPAIR
719 * options to restore a socket post-reboot.
720 */
721 if (!sk)
722 return -ENOTCONN;
723
724 if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_NEW_SYN_RECV)) {
725 unsigned int family = READ_ONCE(sk->sk_family);
726 union tcp_ao_addr *addr;
727 __be32 disn, sisn;
728
729 if (sk->sk_state == TCP_NEW_SYN_RECV) {
730 struct request_sock *req = inet_reqsk(sk);
731
732 sisn = htonl(tcp_rsk(req)->rcv_isn);
733 disn = htonl(tcp_rsk(req)->snt_isn);
734 *sne = tcp_ao_compute_sne(0, tcp_rsk(req)->snt_isn, seq);
735 } else {
736 sisn = th->seq;
737 disn = 0;
738 }
739 if (IS_ENABLED(CONFIG_IPV6) && family == AF_INET6)
740 addr = (union tcp_md5_addr *)&ipv6_hdr(skb)->saddr;
741 else
742 addr = (union tcp_md5_addr *)&ip_hdr(skb)->saddr;
743#if IS_ENABLED(CONFIG_IPV6)
744 if (family == AF_INET6 && ipv6_addr_v4mapped(&sk->sk_v6_daddr))
745 family = AF_INET;
746#endif
747
748 sk = sk_const_to_full_sk(sk);
749 ao_info = rcu_dereference(tcp_sk(sk)->ao_info);
750 if (!ao_info)
751 return -ENOENT;
752 *key = tcp_ao_do_lookup(sk, l3index, addr, family,
753 -1, aoh->rnext_keyid);
754 if (!*key)
755 return -ENOENT;
756 *traffic_key = kmalloc(tcp_ao_digest_size(*key), GFP_ATOMIC);
757 if (!*traffic_key)
758 return -ENOMEM;
759 *allocated_traffic_key = true;
760 if (tcp_ao_calc_key_skb(*key, *traffic_key, skb,
761 sisn, disn, family))
762 return -1;
763 *keyid = (*key)->rcvid;
764 } else {
765 struct tcp_ao_key *rnext_key;
766 u32 snd_basis;
767
768 if (sk->sk_state == TCP_TIME_WAIT) {
769 ao_info = rcu_dereference(tcp_twsk(sk)->ao_info);
770 snd_basis = tcp_twsk(sk)->tw_snd_nxt;
771 } else {
772 ao_info = rcu_dereference(tcp_sk(sk)->ao_info);
773 snd_basis = tcp_sk(sk)->snd_una;
774 }
775 if (!ao_info)
776 return -ENOENT;
777
778 *key = tcp_ao_established_key(ao_info, aoh->rnext_keyid, -1);
779 if (!*key)
780 return -ENOENT;
781 *traffic_key = snd_other_key(*key);
782 rnext_key = READ_ONCE(ao_info->rnext_key);
783 *keyid = rnext_key->rcvid;
784 *sne = tcp_ao_compute_sne(READ_ONCE(ao_info->snd_sne),
785 snd_basis, seq);
786 }
787 return 0;
788}
789
790int tcp_ao_transmit_skb(struct sock *sk, struct sk_buff *skb,
791 struct tcp_ao_key *key, struct tcphdr *th,
792 __u8 *hash_location)
793{
794 struct tcp_skb_cb *tcb = TCP_SKB_CB(skb);
795 struct tcp_sock *tp = tcp_sk(sk);
796 struct tcp_ao_info *ao;
797 void *tkey_buf = NULL;
798 u8 *traffic_key;
799 u32 sne;
800
801 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
802 lockdep_sock_is_held(sk));
803 traffic_key = snd_other_key(key);
804 if (unlikely(tcb->tcp_flags & TCPHDR_SYN)) {
805 __be32 disn;
806
807 if (!(tcb->tcp_flags & TCPHDR_ACK)) {
808 disn = 0;
809 tkey_buf = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
810 if (!tkey_buf)
811 return -ENOMEM;
812 traffic_key = tkey_buf;
813 } else {
814 disn = ao->risn;
815 }
816 tp->af_specific->ao_calc_key_sk(key, traffic_key,
817 sk, ao->lisn, disn, true);
818 }
819 sne = tcp_ao_compute_sne(READ_ONCE(ao->snd_sne), READ_ONCE(tp->snd_una),
820 ntohl(th->seq));
821 tp->af_specific->calc_ao_hash(hash_location, key, sk, skb, traffic_key,
822 hash_location - (u8 *)th, sne);
823 kfree(tkey_buf);
824 return 0;
825}
826
827static struct tcp_ao_key *tcp_ao_inbound_lookup(unsigned short int family,
828 const struct sock *sk, const struct sk_buff *skb,
829 int sndid, int rcvid, int l3index)
830{
831 if (family == AF_INET) {
832 const struct iphdr *iph = ip_hdr(skb);
833
834 return tcp_ao_do_lookup(sk, l3index,
835 (union tcp_ao_addr *)&iph->saddr,
836 AF_INET, sndid, rcvid);
837 } else {
838 const struct ipv6hdr *iph = ipv6_hdr(skb);
839
840 return tcp_ao_do_lookup(sk, l3index,
841 (union tcp_ao_addr *)&iph->saddr,
842 AF_INET6, sndid, rcvid);
843 }
844}
845
846void tcp_ao_syncookie(struct sock *sk, const struct sk_buff *skb,
847 struct request_sock *req, unsigned short int family)
848{
849 struct tcp_request_sock *treq = tcp_rsk(req);
850 const struct tcphdr *th = tcp_hdr(skb);
851 const struct tcp_ao_hdr *aoh;
852 struct tcp_ao_key *key;
853 int l3index;
854
855 /* treq->af_specific is used to perform TCP_AO lookup
856 * in tcp_create_openreq_child().
857 */
858#if IS_ENABLED(CONFIG_IPV6)
859 if (family == AF_INET6)
860 treq->af_specific = &tcp_request_sock_ipv6_ops;
861 else
862#endif
863 treq->af_specific = &tcp_request_sock_ipv4_ops;
864
865 treq->used_tcp_ao = false;
866
867 if (tcp_parse_auth_options(th, NULL, &aoh) || !aoh)
868 return;
869
870 l3index = l3mdev_master_ifindex_by_index(sock_net(sk), inet_rsk(req)->ir_iif);
871 key = tcp_ao_inbound_lookup(family, sk, skb, -1, aoh->keyid, l3index);
872 if (!key)
873 /* Key not found, continue without TCP-AO */
874 return;
875
876 treq->ao_rcv_next = aoh->keyid;
877 treq->ao_keyid = aoh->rnext_keyid;
878 treq->used_tcp_ao = true;
879}
880
881static enum skb_drop_reason
882tcp_ao_verify_hash(const struct sock *sk, const struct sk_buff *skb,
883 unsigned short int family, struct tcp_ao_info *info,
884 const struct tcp_ao_hdr *aoh, struct tcp_ao_key *key,
885 u8 *traffic_key, u8 *phash, u32 sne, int l3index)
886{
887 u8 maclen = aoh->length - sizeof(struct tcp_ao_hdr);
888 const struct tcphdr *th = tcp_hdr(skb);
889 void *hash_buf = NULL;
890
891 if (maclen != tcp_ao_maclen(key)) {
892 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOBAD);
893 atomic64_inc(&info->counters.pkt_bad);
894 atomic64_inc(&key->pkt_bad);
895 tcp_hash_fail("AO hash wrong length", family, skb,
896 "%u != %d L3index: %d", maclen,
897 tcp_ao_maclen(key), l3index);
898 return SKB_DROP_REASON_TCP_AOFAILURE;
899 }
900
901 hash_buf = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
902 if (!hash_buf)
903 return SKB_DROP_REASON_NOT_SPECIFIED;
904
905 /* XXX: make it per-AF callback? */
906 tcp_ao_hash_skb(family, hash_buf, key, sk, skb, traffic_key,
907 (phash - (u8 *)th), sne);
908 if (memcmp(phash, hash_buf, maclen)) {
909 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOBAD);
910 atomic64_inc(&info->counters.pkt_bad);
911 atomic64_inc(&key->pkt_bad);
912 tcp_hash_fail("AO hash mismatch", family, skb,
913 "L3index: %d", l3index);
914 kfree(hash_buf);
915 return SKB_DROP_REASON_TCP_AOFAILURE;
916 }
917 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOGOOD);
918 atomic64_inc(&info->counters.pkt_good);
919 atomic64_inc(&key->pkt_good);
920 kfree(hash_buf);
921 return SKB_NOT_DROPPED_YET;
922}
923
924enum skb_drop_reason
925tcp_inbound_ao_hash(struct sock *sk, const struct sk_buff *skb,
926 unsigned short int family, const struct request_sock *req,
927 int l3index, const struct tcp_ao_hdr *aoh)
928{
929 const struct tcphdr *th = tcp_hdr(skb);
930 u8 *phash = (u8 *)(aoh + 1); /* hash goes just after the header */
931 struct tcp_ao_info *info;
932 enum skb_drop_reason ret;
933 struct tcp_ao_key *key;
934 __be32 sisn, disn;
935 u8 *traffic_key;
936 u32 sne = 0;
937
938 info = rcu_dereference(tcp_sk(sk)->ao_info);
939 if (!info) {
940 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOKEYNOTFOUND);
941 tcp_hash_fail("AO key not found", family, skb,
942 "keyid: %u L3index: %d", aoh->keyid, l3index);
943 return SKB_DROP_REASON_TCP_AOUNEXPECTED;
944 }
945
946 if (unlikely(th->syn)) {
947 sisn = th->seq;
948 disn = 0;
949 }
950
951 /* Fast-path */
952 if (likely((1 << sk->sk_state) & TCP_AO_ESTABLISHED)) {
953 enum skb_drop_reason err;
954 struct tcp_ao_key *current_key;
955
956 /* Check if this socket's rnext_key matches the keyid in the
957 * packet. If not we lookup the key based on the keyid
958 * matching the rcvid in the mkt.
959 */
960 key = READ_ONCE(info->rnext_key);
961 if (key->rcvid != aoh->keyid) {
962 key = tcp_ao_established_key(info, -1, aoh->keyid);
963 if (!key)
964 goto key_not_found;
965 }
966
967 /* Delayed retransmitted SYN */
968 if (unlikely(th->syn && !th->ack))
969 goto verify_hash;
970
971 sne = tcp_ao_compute_sne(info->rcv_sne, tcp_sk(sk)->rcv_nxt,
972 ntohl(th->seq));
973 /* Established socket, traffic key are cached */
974 traffic_key = rcv_other_key(key);
975 err = tcp_ao_verify_hash(sk, skb, family, info, aoh, key,
976 traffic_key, phash, sne, l3index);
977 if (err)
978 return err;
979 current_key = READ_ONCE(info->current_key);
980 /* Key rotation: the peer asks us to use new key (RNext) */
981 if (unlikely(aoh->rnext_keyid != current_key->sndid)) {
982 /* If the key is not found we do nothing. */
983 key = tcp_ao_established_key(info, aoh->rnext_keyid, -1);
984 if (key)
985 /* pairs with tcp_ao_del_cmd */
986 WRITE_ONCE(info->current_key, key);
987 }
988 return SKB_NOT_DROPPED_YET;
989 }
990
991 /* Lookup key based on peer address and keyid.
992 * current_key and rnext_key must not be used on tcp listen
993 * sockets as otherwise:
994 * - request sockets would race on those key pointers
995 * - tcp_ao_del_cmd() allows async key removal
996 */
997 key = tcp_ao_inbound_lookup(family, sk, skb, -1, aoh->keyid, l3index);
998 if (!key)
999 goto key_not_found;
1000
1001 if (th->syn && !th->ack)
1002 goto verify_hash;
1003
1004 if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_NEW_SYN_RECV)) {
1005 /* Make the initial syn the likely case here */
1006 if (unlikely(req)) {
1007 sne = tcp_ao_compute_sne(0, tcp_rsk(req)->rcv_isn,
1008 ntohl(th->seq));
1009 sisn = htonl(tcp_rsk(req)->rcv_isn);
1010 disn = htonl(tcp_rsk(req)->snt_isn);
1011 } else if (unlikely(th->ack && !th->syn)) {
1012 /* Possible syncookie packet */
1013 sisn = htonl(ntohl(th->seq) - 1);
1014 disn = htonl(ntohl(th->ack_seq) - 1);
1015 sne = tcp_ao_compute_sne(0, ntohl(sisn),
1016 ntohl(th->seq));
1017 } else if (unlikely(!th->syn)) {
1018 /* no way to figure out initial sisn/disn - drop */
1019 return SKB_DROP_REASON_TCP_FLAGS;
1020 }
1021 } else if ((1 << sk->sk_state) & (TCPF_SYN_SENT | TCPF_SYN_RECV)) {
1022 disn = info->lisn;
1023 if (th->syn || th->rst)
1024 sisn = th->seq;
1025 else
1026 sisn = info->risn;
1027 } else {
1028 WARN_ONCE(1, "TCP-AO: Unexpected sk_state %d", sk->sk_state);
1029 return SKB_DROP_REASON_TCP_AOFAILURE;
1030 }
1031verify_hash:
1032 traffic_key = kmalloc(tcp_ao_digest_size(key), GFP_ATOMIC);
1033 if (!traffic_key)
1034 return SKB_DROP_REASON_NOT_SPECIFIED;
1035 tcp_ao_calc_key_skb(key, traffic_key, skb, sisn, disn, family);
1036 ret = tcp_ao_verify_hash(sk, skb, family, info, aoh, key,
1037 traffic_key, phash, sne, l3index);
1038 kfree(traffic_key);
1039 return ret;
1040
1041key_not_found:
1042 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPAOKEYNOTFOUND);
1043 atomic64_inc(&info->counters.key_not_found);
1044 tcp_hash_fail("Requested by the peer AO key id not found",
1045 family, skb, "L3index: %d", l3index);
1046 return SKB_DROP_REASON_TCP_AOKEYNOTFOUND;
1047}
1048
1049static int tcp_ao_cache_traffic_keys(const struct sock *sk,
1050 struct tcp_ao_info *ao,
1051 struct tcp_ao_key *ao_key)
1052{
1053 u8 *traffic_key = snd_other_key(ao_key);
1054 int ret;
1055
1056 ret = tcp_ao_calc_key_sk(ao_key, traffic_key, sk,
1057 ao->lisn, ao->risn, true);
1058 if (ret)
1059 return ret;
1060
1061 traffic_key = rcv_other_key(ao_key);
1062 ret = tcp_ao_calc_key_sk(ao_key, traffic_key, sk,
1063 ao->lisn, ao->risn, false);
1064 return ret;
1065}
1066
1067void tcp_ao_connect_init(struct sock *sk)
1068{
1069 struct tcp_sock *tp = tcp_sk(sk);
1070 struct tcp_ao_info *ao_info;
1071 union tcp_ao_addr *addr;
1072 struct tcp_ao_key *key;
1073 int family, l3index;
1074
1075 ao_info = rcu_dereference_protected(tp->ao_info,
1076 lockdep_sock_is_held(sk));
1077 if (!ao_info)
1078 return;
1079
1080 /* Remove all keys that don't match the peer */
1081 family = sk->sk_family;
1082 if (family == AF_INET)
1083 addr = (union tcp_ao_addr *)&sk->sk_daddr;
1084#if IS_ENABLED(CONFIG_IPV6)
1085 else if (family == AF_INET6)
1086 addr = (union tcp_ao_addr *)&sk->sk_v6_daddr;
1087#endif
1088 else
1089 return;
1090 l3index = l3mdev_master_ifindex_by_index(sock_net(sk),
1091 sk->sk_bound_dev_if);
1092
1093 hlist_for_each_entry_rcu(key, &ao_info->head, node) {
1094 if (!tcp_ao_key_cmp(key, l3index, addr, key->prefixlen, family, -1, -1))
1095 continue;
1096
1097 if (key == ao_info->current_key)
1098 ao_info->current_key = NULL;
1099 if (key == ao_info->rnext_key)
1100 ao_info->rnext_key = NULL;
1101 hlist_del_rcu(&key->node);
1102 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1103 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1104 }
1105
1106 key = tp->af_specific->ao_lookup(sk, sk, -1, -1);
1107 if (key) {
1108 /* if current_key or rnext_key were not provided,
1109 * use the first key matching the peer
1110 */
1111 if (!ao_info->current_key)
1112 ao_info->current_key = key;
1113 if (!ao_info->rnext_key)
1114 ao_info->rnext_key = key;
1115 tp->tcp_header_len += tcp_ao_len_aligned(key);
1116
1117 ao_info->lisn = htonl(tp->write_seq);
1118 ao_info->snd_sne = 0;
1119 } else {
1120 /* Can't happen: tcp_connect() verifies that there's
1121 * at least one tcp-ao key that matches the remote peer.
1122 */
1123 WARN_ON_ONCE(1);
1124 rcu_assign_pointer(tp->ao_info, NULL);
1125 kfree(ao_info);
1126 }
1127}
1128
1129void tcp_ao_established(struct sock *sk)
1130{
1131 struct tcp_ao_info *ao;
1132 struct tcp_ao_key *key;
1133
1134 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
1135 lockdep_sock_is_held(sk));
1136 if (!ao)
1137 return;
1138
1139 hlist_for_each_entry_rcu(key, &ao->head, node)
1140 tcp_ao_cache_traffic_keys(sk, ao, key);
1141}
1142
1143void tcp_ao_finish_connect(struct sock *sk, struct sk_buff *skb)
1144{
1145 struct tcp_ao_info *ao;
1146 struct tcp_ao_key *key;
1147
1148 ao = rcu_dereference_protected(tcp_sk(sk)->ao_info,
1149 lockdep_sock_is_held(sk));
1150 if (!ao)
1151 return;
1152
1153 WRITE_ONCE(ao->risn, tcp_hdr(skb)->seq);
1154 ao->rcv_sne = 0;
1155
1156 hlist_for_each_entry_rcu(key, &ao->head, node)
1157 tcp_ao_cache_traffic_keys(sk, ao, key);
1158}
1159
1160int tcp_ao_copy_all_matching(const struct sock *sk, struct sock *newsk,
1161 struct request_sock *req, struct sk_buff *skb,
1162 int family)
1163{
1164 struct tcp_ao_key *key, *new_key, *first_key;
1165 struct tcp_ao_info *new_ao, *ao;
1166 struct hlist_node *key_head;
1167 int l3index, ret = -ENOMEM;
1168 union tcp_ao_addr *addr;
1169 bool match = false;
1170
1171 ao = rcu_dereference(tcp_sk(sk)->ao_info);
1172 if (!ao)
1173 return 0;
1174
1175 /* New socket without TCP-AO on it */
1176 if (!tcp_rsk_used_ao(req))
1177 return 0;
1178
1179 new_ao = tcp_ao_alloc_info(GFP_ATOMIC);
1180 if (!new_ao)
1181 return -ENOMEM;
1182 new_ao->lisn = htonl(tcp_rsk(req)->snt_isn);
1183 new_ao->risn = htonl(tcp_rsk(req)->rcv_isn);
1184 new_ao->ao_required = ao->ao_required;
1185 new_ao->accept_icmps = ao->accept_icmps;
1186
1187 if (family == AF_INET) {
1188 addr = (union tcp_ao_addr *)&newsk->sk_daddr;
1189#if IS_ENABLED(CONFIG_IPV6)
1190 } else if (family == AF_INET6) {
1191 addr = (union tcp_ao_addr *)&newsk->sk_v6_daddr;
1192#endif
1193 } else {
1194 ret = -EAFNOSUPPORT;
1195 goto free_ao;
1196 }
1197 l3index = l3mdev_master_ifindex_by_index(sock_net(newsk),
1198 newsk->sk_bound_dev_if);
1199
1200 hlist_for_each_entry_rcu(key, &ao->head, node) {
1201 if (tcp_ao_key_cmp(key, l3index, addr, key->prefixlen, family, -1, -1))
1202 continue;
1203
1204 new_key = tcp_ao_copy_key(newsk, key);
1205 if (!new_key)
1206 goto free_and_exit;
1207
1208 tcp_ao_cache_traffic_keys(newsk, new_ao, new_key);
1209 tcp_ao_link_mkt(new_ao, new_key);
1210 match = true;
1211 }
1212
1213 if (!match) {
1214 /* RFC5925 (7.4.1) specifies that the TCP-AO status
1215 * of a connection is determined on the initial SYN.
1216 * At this point the connection was TCP-AO enabled, so
1217 * it can't switch to being unsigned if peer's key
1218 * disappears on the listening socket.
1219 */
1220 ret = -EKEYREJECTED;
1221 goto free_and_exit;
1222 }
1223
1224 if (!static_key_fast_inc_not_disabled(&tcp_ao_needed.key.key)) {
1225 ret = -EUSERS;
1226 goto free_and_exit;
1227 }
1228
1229 key_head = rcu_dereference(hlist_first_rcu(&new_ao->head));
1230 first_key = hlist_entry_safe(key_head, struct tcp_ao_key, node);
1231
1232 key = tcp_ao_established_key(new_ao, tcp_rsk(req)->ao_keyid, -1);
1233 if (key)
1234 new_ao->current_key = key;
1235 else
1236 new_ao->current_key = first_key;
1237
1238 /* set rnext_key */
1239 key = tcp_ao_established_key(new_ao, -1, tcp_rsk(req)->ao_rcv_next);
1240 if (key)
1241 new_ao->rnext_key = key;
1242 else
1243 new_ao->rnext_key = first_key;
1244
1245 sk_gso_disable(newsk);
1246 rcu_assign_pointer(tcp_sk(newsk)->ao_info, new_ao);
1247
1248 return 0;
1249
1250free_and_exit:
1251 hlist_for_each_entry_safe(key, key_head, &new_ao->head, node) {
1252 hlist_del(&key->node);
1253 tcp_sigpool_release(key->tcp_sigpool_id);
1254 atomic_sub(tcp_ao_sizeof_key(key), &newsk->sk_omem_alloc);
1255 kfree_sensitive(key);
1256 }
1257free_ao:
1258 kfree(new_ao);
1259 return ret;
1260}
1261
1262static bool tcp_ao_can_set_current_rnext(struct sock *sk)
1263{
1264 /* There aren't current/rnext keys on TCP_LISTEN sockets */
1265 if (sk->sk_state == TCP_LISTEN)
1266 return false;
1267 return true;
1268}
1269
1270static int tcp_ao_verify_ipv4(struct sock *sk, struct tcp_ao_add *cmd,
1271 union tcp_ao_addr **addr)
1272{
1273 struct sockaddr_in *sin = (struct sockaddr_in *)&cmd->addr;
1274 struct inet_sock *inet = inet_sk(sk);
1275
1276 if (sin->sin_family != AF_INET)
1277 return -EINVAL;
1278
1279 /* Currently matching is not performed on port (or port ranges) */
1280 if (sin->sin_port != 0)
1281 return -EINVAL;
1282
1283 /* Check prefix and trailing 0's in addr */
1284 if (cmd->prefix != 0) {
1285 __be32 mask;
1286
1287 if (ntohl(sin->sin_addr.s_addr) == INADDR_ANY)
1288 return -EINVAL;
1289 if (cmd->prefix > 32)
1290 return -EINVAL;
1291
1292 mask = inet_make_mask(cmd->prefix);
1293 if (sin->sin_addr.s_addr & ~mask)
1294 return -EINVAL;
1295
1296 /* Check that MKT address is consistent with socket */
1297 if (ntohl(inet->inet_daddr) != INADDR_ANY &&
1298 (inet->inet_daddr & mask) != sin->sin_addr.s_addr)
1299 return -EINVAL;
1300 } else {
1301 if (ntohl(sin->sin_addr.s_addr) != INADDR_ANY)
1302 return -EINVAL;
1303 }
1304
1305 *addr = (union tcp_ao_addr *)&sin->sin_addr;
1306 return 0;
1307}
1308
1309static int tcp_ao_parse_crypto(struct tcp_ao_add *cmd, struct tcp_ao_key *key)
1310{
1311 unsigned int syn_tcp_option_space;
1312 bool is_kdf_aes_128_cmac = false;
1313 struct crypto_ahash *tfm;
1314 struct tcp_sigpool hp;
1315 void *tmp_key = NULL;
1316 int err;
1317
1318 /* RFC5926, 3.1.1.2. KDF_AES_128_CMAC */
1319 if (!strcmp("cmac(aes128)", cmd->alg_name)) {
1320 strscpy(cmd->alg_name, "cmac(aes)", sizeof(cmd->alg_name));
1321 is_kdf_aes_128_cmac = (cmd->keylen != 16);
1322 tmp_key = kmalloc(cmd->keylen, GFP_KERNEL);
1323 if (!tmp_key)
1324 return -ENOMEM;
1325 }
1326
1327 key->maclen = cmd->maclen ?: 12; /* 12 is the default in RFC5925 */
1328
1329 /* Check: maclen + tcp-ao header <= (MAX_TCP_OPTION_SPACE - mss
1330 * - tstamp (including sackperm)
1331 * - wscale),
1332 * see tcp_syn_options(), tcp_synack_options(), commit 33ad798c924b.
1333 *
1334 * In order to allow D-SACK with TCP-AO, the header size should be:
1335 * (MAX_TCP_OPTION_SPACE - TCPOLEN_TSTAMP_ALIGNED
1336 * - TCPOLEN_SACK_BASE_ALIGNED
1337 * - 2 * TCPOLEN_SACK_PERBLOCK) = 8 (maclen = 4),
1338 * see tcp_established_options().
1339 *
1340 * RFC5925, 2.2:
1341 * Typical MACs are 96-128 bits (12-16 bytes), but any length
1342 * that fits in the header of the segment being authenticated
1343 * is allowed.
1344 *
1345 * RFC5925, 7.6:
1346 * TCP-AO continues to consume 16 bytes in non-SYN segments,
1347 * leaving a total of 24 bytes for other options, of which
1348 * the timestamp consumes 10. This leaves 14 bytes, of which 10
1349 * are used for a single SACK block. When two SACK blocks are used,
1350 * such as to handle D-SACK, a smaller TCP-AO MAC would be required
1351 * to make room for the additional SACK block (i.e., to leave 18
1352 * bytes for the D-SACK variant of the SACK option) [RFC2883].
1353 * Note that D-SACK is not supportable in TCP MD5 in the presence
1354 * of timestamps, because TCP MD5’s MAC length is fixed and too
1355 * large to leave sufficient option space.
1356 */
1357 syn_tcp_option_space = MAX_TCP_OPTION_SPACE;
1358 syn_tcp_option_space -= TCPOLEN_MSS_ALIGNED;
1359 syn_tcp_option_space -= TCPOLEN_TSTAMP_ALIGNED;
1360 syn_tcp_option_space -= TCPOLEN_WSCALE_ALIGNED;
1361 if (tcp_ao_len_aligned(key) > syn_tcp_option_space) {
1362 err = -EMSGSIZE;
1363 goto err_kfree;
1364 }
1365
1366 key->keylen = cmd->keylen;
1367 memcpy(key->key, cmd->key, cmd->keylen);
1368
1369 err = tcp_sigpool_start(key->tcp_sigpool_id, &hp);
1370 if (err)
1371 goto err_kfree;
1372
1373 tfm = crypto_ahash_reqtfm(hp.req);
1374 if (is_kdf_aes_128_cmac) {
1375 void *scratch = hp.scratch;
1376 struct scatterlist sg;
1377
1378 memcpy(tmp_key, cmd->key, cmd->keylen);
1379 sg_init_one(&sg, tmp_key, cmd->keylen);
1380
1381 /* Using zero-key of 16 bytes as described in RFC5926 */
1382 memset(scratch, 0, 16);
1383 err = crypto_ahash_setkey(tfm, scratch, 16);
1384 if (err)
1385 goto err_pool_end;
1386
1387 err = crypto_ahash_init(hp.req);
1388 if (err)
1389 goto err_pool_end;
1390
1391 ahash_request_set_crypt(hp.req, &sg, key->key, cmd->keylen);
1392 err = crypto_ahash_update(hp.req);
1393 if (err)
1394 goto err_pool_end;
1395
1396 err |= crypto_ahash_final(hp.req);
1397 if (err)
1398 goto err_pool_end;
1399 key->keylen = 16;
1400 }
1401
1402 err = crypto_ahash_setkey(tfm, key->key, key->keylen);
1403 if (err)
1404 goto err_pool_end;
1405
1406 tcp_sigpool_end(&hp);
1407 kfree_sensitive(tmp_key);
1408
1409 if (tcp_ao_maclen(key) > key->digest_size)
1410 return -EINVAL;
1411
1412 return 0;
1413
1414err_pool_end:
1415 tcp_sigpool_end(&hp);
1416err_kfree:
1417 kfree_sensitive(tmp_key);
1418 return err;
1419}
1420
1421#if IS_ENABLED(CONFIG_IPV6)
1422static int tcp_ao_verify_ipv6(struct sock *sk, struct tcp_ao_add *cmd,
1423 union tcp_ao_addr **paddr,
1424 unsigned short int *family)
1425{
1426 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd->addr;
1427 struct in6_addr *addr = &sin6->sin6_addr;
1428 u8 prefix = cmd->prefix;
1429
1430 if (sin6->sin6_family != AF_INET6)
1431 return -EINVAL;
1432
1433 /* Currently matching is not performed on port (or port ranges) */
1434 if (sin6->sin6_port != 0)
1435 return -EINVAL;
1436
1437 /* Check prefix and trailing 0's in addr */
1438 if (cmd->prefix != 0 && ipv6_addr_v4mapped(addr)) {
1439 __be32 addr4 = addr->s6_addr32[3];
1440 __be32 mask;
1441
1442 if (prefix > 32 || ntohl(addr4) == INADDR_ANY)
1443 return -EINVAL;
1444
1445 mask = inet_make_mask(prefix);
1446 if (addr4 & ~mask)
1447 return -EINVAL;
1448
1449 /* Check that MKT address is consistent with socket */
1450 if (!ipv6_addr_any(&sk->sk_v6_daddr)) {
1451 __be32 daddr4 = sk->sk_v6_daddr.s6_addr32[3];
1452
1453 if (!ipv6_addr_v4mapped(&sk->sk_v6_daddr))
1454 return -EINVAL;
1455 if ((daddr4 & mask) != addr4)
1456 return -EINVAL;
1457 }
1458
1459 *paddr = (union tcp_ao_addr *)&addr->s6_addr32[3];
1460 *family = AF_INET;
1461 return 0;
1462 } else if (cmd->prefix != 0) {
1463 struct in6_addr pfx;
1464
1465 if (ipv6_addr_any(addr) || prefix > 128)
1466 return -EINVAL;
1467
1468 ipv6_addr_prefix(&pfx, addr, prefix);
1469 if (ipv6_addr_cmp(&pfx, addr))
1470 return -EINVAL;
1471
1472 /* Check that MKT address is consistent with socket */
1473 if (!ipv6_addr_any(&sk->sk_v6_daddr) &&
1474 !ipv6_prefix_equal(&sk->sk_v6_daddr, addr, prefix))
1475
1476 return -EINVAL;
1477 } else {
1478 if (!ipv6_addr_any(addr))
1479 return -EINVAL;
1480 }
1481
1482 *paddr = (union tcp_ao_addr *)addr;
1483 return 0;
1484}
1485#else
1486static int tcp_ao_verify_ipv6(struct sock *sk, struct tcp_ao_add *cmd,
1487 union tcp_ao_addr **paddr,
1488 unsigned short int *family)
1489{
1490 return -EOPNOTSUPP;
1491}
1492#endif
1493
1494static struct tcp_ao_info *setsockopt_ao_info(struct sock *sk)
1495{
1496 if (sk_fullsock(sk)) {
1497 return rcu_dereference_protected(tcp_sk(sk)->ao_info,
1498 lockdep_sock_is_held(sk));
1499 } else if (sk->sk_state == TCP_TIME_WAIT) {
1500 return rcu_dereference_protected(tcp_twsk(sk)->ao_info,
1501 lockdep_sock_is_held(sk));
1502 }
1503 return ERR_PTR(-ESOCKTNOSUPPORT);
1504}
1505
1506static struct tcp_ao_info *getsockopt_ao_info(struct sock *sk)
1507{
1508 if (sk_fullsock(sk))
1509 return rcu_dereference(tcp_sk(sk)->ao_info);
1510 else if (sk->sk_state == TCP_TIME_WAIT)
1511 return rcu_dereference(tcp_twsk(sk)->ao_info);
1512
1513 return ERR_PTR(-ESOCKTNOSUPPORT);
1514}
1515
1516#define TCP_AO_KEYF_ALL (TCP_AO_KEYF_IFINDEX | TCP_AO_KEYF_EXCLUDE_OPT)
1517#define TCP_AO_GET_KEYF_VALID (TCP_AO_KEYF_IFINDEX)
1518
1519static struct tcp_ao_key *tcp_ao_key_alloc(struct sock *sk,
1520 struct tcp_ao_add *cmd)
1521{
1522 const char *algo = cmd->alg_name;
1523 unsigned int digest_size;
1524 struct crypto_ahash *tfm;
1525 struct tcp_ao_key *key;
1526 struct tcp_sigpool hp;
1527 int err, pool_id;
1528 size_t size;
1529
1530 /* Force null-termination of alg_name */
1531 cmd->alg_name[ARRAY_SIZE(cmd->alg_name) - 1] = '\0';
1532
1533 /* RFC5926, 3.1.1.2. KDF_AES_128_CMAC */
1534 if (!strcmp("cmac(aes128)", algo))
1535 algo = "cmac(aes)";
1536
1537 /* Full TCP header (th->doff << 2) should fit into scratch area,
1538 * see tcp_ao_hash_header().
1539 */
1540 pool_id = tcp_sigpool_alloc_ahash(algo, 60);
1541 if (pool_id < 0)
1542 return ERR_PTR(pool_id);
1543
1544 err = tcp_sigpool_start(pool_id, &hp);
1545 if (err)
1546 goto err_free_pool;
1547
1548 tfm = crypto_ahash_reqtfm(hp.req);
1549 digest_size = crypto_ahash_digestsize(tfm);
1550 tcp_sigpool_end(&hp);
1551
1552 size = sizeof(struct tcp_ao_key) + (digest_size << 1);
1553 key = sock_kmalloc(sk, size, GFP_KERNEL);
1554 if (!key) {
1555 err = -ENOMEM;
1556 goto err_free_pool;
1557 }
1558
1559 key->tcp_sigpool_id = pool_id;
1560 key->digest_size = digest_size;
1561 return key;
1562
1563err_free_pool:
1564 tcp_sigpool_release(pool_id);
1565 return ERR_PTR(err);
1566}
1567
1568static int tcp_ao_add_cmd(struct sock *sk, unsigned short int family,
1569 sockptr_t optval, int optlen)
1570{
1571 struct tcp_ao_info *ao_info;
1572 union tcp_ao_addr *addr;
1573 struct tcp_ao_key *key;
1574 struct tcp_ao_add cmd;
1575 int ret, l3index = 0;
1576 bool first = false;
1577
1578 if (optlen < sizeof(cmd))
1579 return -EINVAL;
1580
1581 ret = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1582 if (ret)
1583 return ret;
1584
1585 if (cmd.keylen > TCP_AO_MAXKEYLEN)
1586 return -EINVAL;
1587
1588 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1589 return -EINVAL;
1590
1591 if (family == AF_INET)
1592 ret = tcp_ao_verify_ipv4(sk, &cmd, &addr);
1593 else
1594 ret = tcp_ao_verify_ipv6(sk, &cmd, &addr, &family);
1595 if (ret)
1596 return ret;
1597
1598 if (cmd.keyflags & ~TCP_AO_KEYF_ALL)
1599 return -EINVAL;
1600
1601 if (cmd.set_current || cmd.set_rnext) {
1602 if (!tcp_ao_can_set_current_rnext(sk))
1603 return -EINVAL;
1604 }
1605
1606 if (cmd.ifindex && !(cmd.keyflags & TCP_AO_KEYF_IFINDEX))
1607 return -EINVAL;
1608
1609 /* For cmd.tcp_ifindex = 0 the key will apply to the default VRF */
1610 if (cmd.keyflags & TCP_AO_KEYF_IFINDEX && cmd.ifindex) {
1611 int bound_dev_if = READ_ONCE(sk->sk_bound_dev_if);
1612 struct net_device *dev;
1613
1614 rcu_read_lock();
1615 dev = dev_get_by_index_rcu(sock_net(sk), cmd.ifindex);
1616 if (dev && netif_is_l3_master(dev))
1617 l3index = dev->ifindex;
1618 rcu_read_unlock();
1619
1620 if (!dev || !l3index)
1621 return -EINVAL;
1622
1623 if (!bound_dev_if || bound_dev_if != cmd.ifindex) {
1624 /* tcp_ao_established_key() doesn't expect having
1625 * non peer-matching key on an established TCP-AO
1626 * connection.
1627 */
1628 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE)))
1629 return -EINVAL;
1630 }
1631
1632 /* It's still possible to bind after adding keys or even
1633 * re-bind to a different dev (with CAP_NET_RAW).
1634 * So, no reason to return error here, rather try to be
1635 * nice and warn the user.
1636 */
1637 if (bound_dev_if && bound_dev_if != cmd.ifindex)
1638 net_warn_ratelimited("AO key ifindex %d != sk bound ifindex %d\n",
1639 cmd.ifindex, bound_dev_if);
1640 }
1641
1642 /* Don't allow keys for peers that have a matching TCP-MD5 key */
1643 if (cmd.keyflags & TCP_AO_KEYF_IFINDEX) {
1644 /* Non-_exact version of tcp_md5_do_lookup() will
1645 * as well match keys that aren't bound to a specific VRF
1646 * (that will make them match AO key with
1647 * sysctl_tcp_l3dev_accept = 1
1648 */
1649 if (tcp_md5_do_lookup(sk, l3index, addr, family))
1650 return -EKEYREJECTED;
1651 } else {
1652 if (tcp_md5_do_lookup_any_l3index(sk, addr, family))
1653 return -EKEYREJECTED;
1654 }
1655
1656 ao_info = setsockopt_ao_info(sk);
1657 if (IS_ERR(ao_info))
1658 return PTR_ERR(ao_info);
1659
1660 if (!ao_info) {
1661 ao_info = tcp_ao_alloc_info(GFP_KERNEL);
1662 if (!ao_info)
1663 return -ENOMEM;
1664 first = true;
1665 } else {
1666 /* Check that neither RecvID nor SendID match any
1667 * existing key for the peer, RFC5925 3.1:
1668 * > The IDs of MKTs MUST NOT overlap where their
1669 * > TCP connection identifiers overlap.
1670 */
1671 if (__tcp_ao_do_lookup(sk, l3index, addr, family, cmd.prefix, -1, cmd.rcvid))
1672 return -EEXIST;
1673 if (__tcp_ao_do_lookup(sk, l3index, addr, family,
1674 cmd.prefix, cmd.sndid, -1))
1675 return -EEXIST;
1676 }
1677
1678 key = tcp_ao_key_alloc(sk, &cmd);
1679 if (IS_ERR(key)) {
1680 ret = PTR_ERR(key);
1681 goto err_free_ao;
1682 }
1683
1684 INIT_HLIST_NODE(&key->node);
1685 memcpy(&key->addr, addr, (family == AF_INET) ? sizeof(struct in_addr) :
1686 sizeof(struct in6_addr));
1687 key->prefixlen = cmd.prefix;
1688 key->family = family;
1689 key->keyflags = cmd.keyflags;
1690 key->sndid = cmd.sndid;
1691 key->rcvid = cmd.rcvid;
1692 key->l3index = l3index;
1693 atomic64_set(&key->pkt_good, 0);
1694 atomic64_set(&key->pkt_bad, 0);
1695
1696 ret = tcp_ao_parse_crypto(&cmd, key);
1697 if (ret < 0)
1698 goto err_free_sock;
1699
1700 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))) {
1701 tcp_ao_cache_traffic_keys(sk, ao_info, key);
1702 if (first) {
1703 ao_info->current_key = key;
1704 ao_info->rnext_key = key;
1705 }
1706 }
1707
1708 tcp_ao_link_mkt(ao_info, key);
1709 if (first) {
1710 if (!static_branch_inc(&tcp_ao_needed.key)) {
1711 ret = -EUSERS;
1712 goto err_free_sock;
1713 }
1714 sk_gso_disable(sk);
1715 rcu_assign_pointer(tcp_sk(sk)->ao_info, ao_info);
1716 }
1717
1718 if (cmd.set_current)
1719 WRITE_ONCE(ao_info->current_key, key);
1720 if (cmd.set_rnext)
1721 WRITE_ONCE(ao_info->rnext_key, key);
1722 return 0;
1723
1724err_free_sock:
1725 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1726 tcp_sigpool_release(key->tcp_sigpool_id);
1727 kfree_sensitive(key);
1728err_free_ao:
1729 if (first)
1730 kfree(ao_info);
1731 return ret;
1732}
1733
1734static int tcp_ao_delete_key(struct sock *sk, struct tcp_ao_info *ao_info,
1735 bool del_async, struct tcp_ao_key *key,
1736 struct tcp_ao_key *new_current,
1737 struct tcp_ao_key *new_rnext)
1738{
1739 int err;
1740
1741 hlist_del_rcu(&key->node);
1742
1743 /* Support for async delete on listening sockets: as they don't
1744 * need current_key/rnext_key maintaining, we don't need to check
1745 * them and we can just free all resources in RCU fashion.
1746 */
1747 if (del_async) {
1748 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1749 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1750 return 0;
1751 }
1752
1753 /* At this moment another CPU could have looked this key up
1754 * while it was unlinked from the list. Wait for RCU grace period,
1755 * after which the key is off-list and can't be looked up again;
1756 * the rx path [just before RCU came] might have used it and set it
1757 * as current_key (very unlikely).
1758 * Free the key with next RCU grace period (in case it was
1759 * current_key before tcp_ao_current_rnext() might have
1760 * changed it in forced-delete).
1761 */
1762 synchronize_rcu();
1763 if (new_current)
1764 WRITE_ONCE(ao_info->current_key, new_current);
1765 if (new_rnext)
1766 WRITE_ONCE(ao_info->rnext_key, new_rnext);
1767
1768 if (unlikely(READ_ONCE(ao_info->current_key) == key ||
1769 READ_ONCE(ao_info->rnext_key) == key)) {
1770 err = -EBUSY;
1771 goto add_key;
1772 }
1773
1774 atomic_sub(tcp_ao_sizeof_key(key), &sk->sk_omem_alloc);
1775 call_rcu(&key->rcu, tcp_ao_key_free_rcu);
1776
1777 return 0;
1778add_key:
1779 hlist_add_head_rcu(&key->node, &ao_info->head);
1780 return err;
1781}
1782
1783#define TCP_AO_DEL_KEYF_ALL (TCP_AO_KEYF_IFINDEX)
1784static int tcp_ao_del_cmd(struct sock *sk, unsigned short int family,
1785 sockptr_t optval, int optlen)
1786{
1787 struct tcp_ao_key *key, *new_current = NULL, *new_rnext = NULL;
1788 int err, addr_len, l3index = 0;
1789 struct tcp_ao_info *ao_info;
1790 union tcp_ao_addr *addr;
1791 struct tcp_ao_del cmd;
1792 __u8 prefix;
1793 u16 port;
1794
1795 if (optlen < sizeof(cmd))
1796 return -EINVAL;
1797
1798 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1799 if (err)
1800 return err;
1801
1802 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1803 return -EINVAL;
1804
1805 if (cmd.set_current || cmd.set_rnext) {
1806 if (!tcp_ao_can_set_current_rnext(sk))
1807 return -EINVAL;
1808 }
1809
1810 if (cmd.keyflags & ~TCP_AO_DEL_KEYF_ALL)
1811 return -EINVAL;
1812
1813 /* No sanity check for TCP_AO_KEYF_IFINDEX as if a VRF
1814 * was destroyed, there still should be a way to delete keys,
1815 * that were bound to that l3intf. So, fail late at lookup stage
1816 * if there is no key for that ifindex.
1817 */
1818 if (cmd.ifindex && !(cmd.keyflags & TCP_AO_KEYF_IFINDEX))
1819 return -EINVAL;
1820
1821 ao_info = setsockopt_ao_info(sk);
1822 if (IS_ERR(ao_info))
1823 return PTR_ERR(ao_info);
1824 if (!ao_info)
1825 return -ENOENT;
1826
1827 /* For sockets in TCP_CLOSED it's possible set keys that aren't
1828 * matching the future peer (address/VRF/etc),
1829 * tcp_ao_connect_init() will choose a correct matching MKT
1830 * if there's any.
1831 */
1832 if (cmd.set_current) {
1833 new_current = tcp_ao_established_key(ao_info, cmd.current_key, -1);
1834 if (!new_current)
1835 return -ENOENT;
1836 }
1837 if (cmd.set_rnext) {
1838 new_rnext = tcp_ao_established_key(ao_info, -1, cmd.rnext);
1839 if (!new_rnext)
1840 return -ENOENT;
1841 }
1842 if (cmd.del_async && sk->sk_state != TCP_LISTEN)
1843 return -EINVAL;
1844
1845 if (family == AF_INET) {
1846 struct sockaddr_in *sin = (struct sockaddr_in *)&cmd.addr;
1847
1848 addr = (union tcp_ao_addr *)&sin->sin_addr;
1849 addr_len = sizeof(struct in_addr);
1850 port = ntohs(sin->sin_port);
1851 } else {
1852 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&cmd.addr;
1853 struct in6_addr *addr6 = &sin6->sin6_addr;
1854
1855 if (ipv6_addr_v4mapped(addr6)) {
1856 addr = (union tcp_ao_addr *)&addr6->s6_addr32[3];
1857 addr_len = sizeof(struct in_addr);
1858 family = AF_INET;
1859 } else {
1860 addr = (union tcp_ao_addr *)addr6;
1861 addr_len = sizeof(struct in6_addr);
1862 }
1863 port = ntohs(sin6->sin6_port);
1864 }
1865 prefix = cmd.prefix;
1866
1867 /* Currently matching is not performed on port (or port ranges) */
1868 if (port != 0)
1869 return -EINVAL;
1870
1871 /* We could choose random present key here for current/rnext
1872 * but that's less predictable. Let's be strict and don't
1873 * allow removing a key that's in use. RFC5925 doesn't
1874 * specify how-to coordinate key removal, but says:
1875 * "It is presumed that an MKT affecting a particular
1876 * connection cannot be destroyed during an active connection"
1877 */
1878 hlist_for_each_entry_rcu(key, &ao_info->head, node) {
1879 if (cmd.sndid != key->sndid ||
1880 cmd.rcvid != key->rcvid)
1881 continue;
1882
1883 if (family != key->family ||
1884 prefix != key->prefixlen ||
1885 memcmp(addr, &key->addr, addr_len))
1886 continue;
1887
1888 if ((cmd.keyflags & TCP_AO_KEYF_IFINDEX) !=
1889 (key->keyflags & TCP_AO_KEYF_IFINDEX))
1890 continue;
1891
1892 if (key->l3index != l3index)
1893 continue;
1894
1895 if (key == new_current || key == new_rnext)
1896 continue;
1897
1898 return tcp_ao_delete_key(sk, ao_info, cmd.del_async, key,
1899 new_current, new_rnext);
1900 }
1901 return -ENOENT;
1902}
1903
1904/* cmd.ao_required makes a socket TCP-AO only.
1905 * Don't allow any md5 keys for any l3intf on the socket together with it.
1906 * Restricting it early in setsockopt() removes a check for
1907 * ao_info->ao_required on inbound tcp segment fast-path.
1908 */
1909static int tcp_ao_required_verify(struct sock *sk)
1910{
1911#ifdef CONFIG_TCP_MD5SIG
1912 const struct tcp_md5sig_info *md5sig;
1913
1914 if (!static_branch_unlikely(&tcp_md5_needed.key))
1915 return 0;
1916
1917 md5sig = rcu_dereference_check(tcp_sk(sk)->md5sig_info,
1918 lockdep_sock_is_held(sk));
1919 if (!md5sig)
1920 return 0;
1921
1922 if (rcu_dereference_check(hlist_first_rcu(&md5sig->head),
1923 lockdep_sock_is_held(sk)))
1924 return 1;
1925#endif
1926 return 0;
1927}
1928
1929static int tcp_ao_info_cmd(struct sock *sk, unsigned short int family,
1930 sockptr_t optval, int optlen)
1931{
1932 struct tcp_ao_key *new_current = NULL, *new_rnext = NULL;
1933 struct tcp_ao_info *ao_info;
1934 struct tcp_ao_info_opt cmd;
1935 bool first = false;
1936 int err;
1937
1938 if (optlen < sizeof(cmd))
1939 return -EINVAL;
1940
1941 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
1942 if (err)
1943 return err;
1944
1945 if (cmd.set_current || cmd.set_rnext) {
1946 if (!tcp_ao_can_set_current_rnext(sk))
1947 return -EINVAL;
1948 }
1949
1950 if (cmd.reserved != 0 || cmd.reserved2 != 0)
1951 return -EINVAL;
1952
1953 ao_info = setsockopt_ao_info(sk);
1954 if (IS_ERR(ao_info))
1955 return PTR_ERR(ao_info);
1956 if (!ao_info) {
1957 if (!((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE)))
1958 return -EINVAL;
1959 ao_info = tcp_ao_alloc_info(GFP_KERNEL);
1960 if (!ao_info)
1961 return -ENOMEM;
1962 first = true;
1963 }
1964
1965 if (cmd.ao_required && tcp_ao_required_verify(sk))
1966 return -EKEYREJECTED;
1967
1968 /* For sockets in TCP_CLOSED it's possible set keys that aren't
1969 * matching the future peer (address/port/VRF/etc),
1970 * tcp_ao_connect_init() will choose a correct matching MKT
1971 * if there's any.
1972 */
1973 if (cmd.set_current) {
1974 new_current = tcp_ao_established_key(ao_info, cmd.current_key, -1);
1975 if (!new_current) {
1976 err = -ENOENT;
1977 goto out;
1978 }
1979 }
1980 if (cmd.set_rnext) {
1981 new_rnext = tcp_ao_established_key(ao_info, -1, cmd.rnext);
1982 if (!new_rnext) {
1983 err = -ENOENT;
1984 goto out;
1985 }
1986 }
1987 if (cmd.set_counters) {
1988 atomic64_set(&ao_info->counters.pkt_good, cmd.pkt_good);
1989 atomic64_set(&ao_info->counters.pkt_bad, cmd.pkt_bad);
1990 atomic64_set(&ao_info->counters.key_not_found, cmd.pkt_key_not_found);
1991 atomic64_set(&ao_info->counters.ao_required, cmd.pkt_ao_required);
1992 atomic64_set(&ao_info->counters.dropped_icmp, cmd.pkt_dropped_icmp);
1993 }
1994
1995 ao_info->ao_required = cmd.ao_required;
1996 ao_info->accept_icmps = cmd.accept_icmps;
1997 if (new_current)
1998 WRITE_ONCE(ao_info->current_key, new_current);
1999 if (new_rnext)
2000 WRITE_ONCE(ao_info->rnext_key, new_rnext);
2001 if (first) {
2002 if (!static_branch_inc(&tcp_ao_needed.key)) {
2003 err = -EUSERS;
2004 goto out;
2005 }
2006 sk_gso_disable(sk);
2007 rcu_assign_pointer(tcp_sk(sk)->ao_info, ao_info);
2008 }
2009 return 0;
2010out:
2011 if (first)
2012 kfree(ao_info);
2013 return err;
2014}
2015
2016int tcp_parse_ao(struct sock *sk, int cmd, unsigned short int family,
2017 sockptr_t optval, int optlen)
2018{
2019 if (WARN_ON_ONCE(family != AF_INET && family != AF_INET6))
2020 return -EAFNOSUPPORT;
2021
2022 switch (cmd) {
2023 case TCP_AO_ADD_KEY:
2024 return tcp_ao_add_cmd(sk, family, optval, optlen);
2025 case TCP_AO_DEL_KEY:
2026 return tcp_ao_del_cmd(sk, family, optval, optlen);
2027 case TCP_AO_INFO:
2028 return tcp_ao_info_cmd(sk, family, optval, optlen);
2029 default:
2030 WARN_ON_ONCE(1);
2031 return -EINVAL;
2032 }
2033}
2034
2035int tcp_v4_parse_ao(struct sock *sk, int cmd, sockptr_t optval, int optlen)
2036{
2037 return tcp_parse_ao(sk, cmd, AF_INET, optval, optlen);
2038}
2039
2040/* tcp_ao_copy_mkts_to_user(ao_info, optval, optlen)
2041 *
2042 * @ao_info: struct tcp_ao_info on the socket that
2043 * socket getsockopt(TCP_AO_GET_KEYS) is executed on
2044 * @optval: pointer to array of tcp_ao_getsockopt structures in user space.
2045 * Must be != NULL.
2046 * @optlen: pointer to size of tcp_ao_getsockopt structure.
2047 * Must be != NULL.
2048 *
2049 * Return value: 0 on success, a negative error number otherwise.
2050 *
2051 * optval points to an array of tcp_ao_getsockopt structures in user space.
2052 * optval[0] is used as both input and output to getsockopt. It determines
2053 * which keys are returned by the kernel.
2054 * optval[0].nkeys is the size of the array in user space. On return it contains
2055 * the number of keys matching the search criteria.
2056 * If tcp_ao_getsockopt::get_all is set, then all keys in the socket are
2057 * returned, otherwise only keys matching <addr, prefix, sndid, rcvid>
2058 * in optval[0] are returned.
2059 * optlen is also used as both input and output. The user provides the size
2060 * of struct tcp_ao_getsockopt in user space, and the kernel returns the size
2061 * of the structure in kernel space.
2062 * The size of struct tcp_ao_getsockopt may differ between user and kernel.
2063 * There are three cases to consider:
2064 * * If usize == ksize, then keys are copied verbatim.
2065 * * If usize < ksize, then the userspace has passed an old struct to a
2066 * newer kernel. The rest of the trailing bytes in optval[0]
2067 * (ksize - usize) are interpreted as 0 by the kernel.
2068 * * If usize > ksize, then the userspace has passed a new struct to an
2069 * older kernel. The trailing bytes unknown to the kernel (usize - ksize)
2070 * are checked to ensure they are zeroed, otherwise -E2BIG is returned.
2071 * On return the kernel fills in min(usize, ksize) in each entry of the array.
2072 * The layout of the fields in the user and kernel structures is expected to
2073 * be the same (including in the 32bit vs 64bit case).
2074 */
2075static int tcp_ao_copy_mkts_to_user(struct tcp_ao_info *ao_info,
2076 sockptr_t optval, sockptr_t optlen)
2077{
2078 struct tcp_ao_getsockopt opt_in, opt_out;
2079 struct tcp_ao_key *key, *current_key;
2080 bool do_address_matching = true;
2081 union tcp_ao_addr *addr = NULL;
2082 int err, l3index, user_len;
2083 unsigned int max_keys; /* maximum number of keys to copy to user */
2084 size_t out_offset = 0;
2085 size_t bytes_to_write; /* number of bytes to write to user level */
2086 u32 matched_keys; /* keys from ao_info matched so far */
2087 int optlen_out;
2088 __be16 port = 0;
2089
2090 if (copy_from_sockptr(&user_len, optlen, sizeof(int)))
2091 return -EFAULT;
2092
2093 if (user_len <= 0)
2094 return -EINVAL;
2095
2096 memset(&opt_in, 0, sizeof(struct tcp_ao_getsockopt));
2097 err = copy_struct_from_sockptr(&opt_in, sizeof(opt_in),
2098 optval, user_len);
2099 if (err < 0)
2100 return err;
2101
2102 if (opt_in.pkt_good || opt_in.pkt_bad)
2103 return -EINVAL;
2104 if (opt_in.keyflags & ~TCP_AO_GET_KEYF_VALID)
2105 return -EINVAL;
2106 if (opt_in.ifindex && !(opt_in.keyflags & TCP_AO_KEYF_IFINDEX))
2107 return -EINVAL;
2108
2109 if (opt_in.reserved != 0)
2110 return -EINVAL;
2111
2112 max_keys = opt_in.nkeys;
2113 l3index = (opt_in.keyflags & TCP_AO_KEYF_IFINDEX) ? opt_in.ifindex : -1;
2114
2115 if (opt_in.get_all || opt_in.is_current || opt_in.is_rnext) {
2116 if (opt_in.get_all && (opt_in.is_current || opt_in.is_rnext))
2117 return -EINVAL;
2118 do_address_matching = false;
2119 }
2120
2121 switch (opt_in.addr.ss_family) {
2122 case AF_INET: {
2123 struct sockaddr_in *sin;
2124 __be32 mask;
2125
2126 sin = (struct sockaddr_in *)&opt_in.addr;
2127 port = sin->sin_port;
2128 addr = (union tcp_ao_addr *)&sin->sin_addr;
2129
2130 if (opt_in.prefix > 32)
2131 return -EINVAL;
2132
2133 if (ntohl(sin->sin_addr.s_addr) == INADDR_ANY &&
2134 opt_in.prefix != 0)
2135 return -EINVAL;
2136
2137 mask = inet_make_mask(opt_in.prefix);
2138 if (sin->sin_addr.s_addr & ~mask)
2139 return -EINVAL;
2140
2141 break;
2142 }
2143 case AF_INET6: {
2144 struct sockaddr_in6 *sin6;
2145 struct in6_addr *addr6;
2146
2147 sin6 = (struct sockaddr_in6 *)&opt_in.addr;
2148 addr = (union tcp_ao_addr *)&sin6->sin6_addr;
2149 addr6 = &sin6->sin6_addr;
2150 port = sin6->sin6_port;
2151
2152 /* We don't have to change family and @addr here if
2153 * ipv6_addr_v4mapped() like in key adding:
2154 * tcp_ao_key_cmp() does it. Do the sanity checks though.
2155 */
2156 if (opt_in.prefix != 0) {
2157 if (ipv6_addr_v4mapped(addr6)) {
2158 __be32 mask, addr4 = addr6->s6_addr32[3];
2159
2160 if (opt_in.prefix > 32 ||
2161 ntohl(addr4) == INADDR_ANY)
2162 return -EINVAL;
2163 mask = inet_make_mask(opt_in.prefix);
2164 if (addr4 & ~mask)
2165 return -EINVAL;
2166 } else {
2167 struct in6_addr pfx;
2168
2169 if (ipv6_addr_any(addr6) ||
2170 opt_in.prefix > 128)
2171 return -EINVAL;
2172
2173 ipv6_addr_prefix(&pfx, addr6, opt_in.prefix);
2174 if (ipv6_addr_cmp(&pfx, addr6))
2175 return -EINVAL;
2176 }
2177 } else if (!ipv6_addr_any(addr6)) {
2178 return -EINVAL;
2179 }
2180 break;
2181 }
2182 case 0:
2183 if (!do_address_matching)
2184 break;
2185 fallthrough;
2186 default:
2187 return -EAFNOSUPPORT;
2188 }
2189
2190 if (!do_address_matching) {
2191 /* We could just ignore those, but let's do stricter checks */
2192 if (addr || port)
2193 return -EINVAL;
2194 if (opt_in.prefix || opt_in.sndid || opt_in.rcvid)
2195 return -EINVAL;
2196 }
2197
2198 bytes_to_write = min_t(int, user_len, sizeof(struct tcp_ao_getsockopt));
2199 matched_keys = 0;
2200 /* May change in RX, while we're dumping, pre-fetch it */
2201 current_key = READ_ONCE(ao_info->current_key);
2202
2203 hlist_for_each_entry_rcu(key, &ao_info->head, node) {
2204 if (opt_in.get_all)
2205 goto match;
2206
2207 if (opt_in.is_current || opt_in.is_rnext) {
2208 if (opt_in.is_current && key == current_key)
2209 goto match;
2210 if (opt_in.is_rnext && key == ao_info->rnext_key)
2211 goto match;
2212 continue;
2213 }
2214
2215 if (tcp_ao_key_cmp(key, l3index, addr, opt_in.prefix,
2216 opt_in.addr.ss_family,
2217 opt_in.sndid, opt_in.rcvid) != 0)
2218 continue;
2219match:
2220 matched_keys++;
2221 if (matched_keys > max_keys)
2222 continue;
2223
2224 memset(&opt_out, 0, sizeof(struct tcp_ao_getsockopt));
2225
2226 if (key->family == AF_INET) {
2227 struct sockaddr_in *sin_out = (struct sockaddr_in *)&opt_out.addr;
2228
2229 sin_out->sin_family = key->family;
2230 sin_out->sin_port = 0;
2231 memcpy(&sin_out->sin_addr, &key->addr, sizeof(struct in_addr));
2232 } else {
2233 struct sockaddr_in6 *sin6_out = (struct sockaddr_in6 *)&opt_out.addr;
2234
2235 sin6_out->sin6_family = key->family;
2236 sin6_out->sin6_port = 0;
2237 memcpy(&sin6_out->sin6_addr, &key->addr, sizeof(struct in6_addr));
2238 }
2239 opt_out.sndid = key->sndid;
2240 opt_out.rcvid = key->rcvid;
2241 opt_out.prefix = key->prefixlen;
2242 opt_out.keyflags = key->keyflags;
2243 opt_out.is_current = (key == current_key);
2244 opt_out.is_rnext = (key == ao_info->rnext_key);
2245 opt_out.nkeys = 0;
2246 opt_out.maclen = key->maclen;
2247 opt_out.keylen = key->keylen;
2248 opt_out.ifindex = key->l3index;
2249 opt_out.pkt_good = atomic64_read(&key->pkt_good);
2250 opt_out.pkt_bad = atomic64_read(&key->pkt_bad);
2251 memcpy(&opt_out.key, key->key, key->keylen);
2252 tcp_sigpool_algo(key->tcp_sigpool_id, opt_out.alg_name, 64);
2253
2254 /* Copy key to user */
2255 if (copy_to_sockptr_offset(optval, out_offset,
2256 &opt_out, bytes_to_write))
2257 return -EFAULT;
2258 out_offset += user_len;
2259 }
2260
2261 optlen_out = (int)sizeof(struct tcp_ao_getsockopt);
2262 if (copy_to_sockptr(optlen, &optlen_out, sizeof(int)))
2263 return -EFAULT;
2264
2265 out_offset = offsetof(struct tcp_ao_getsockopt, nkeys);
2266 if (copy_to_sockptr_offset(optval, out_offset,
2267 &matched_keys, sizeof(u32)))
2268 return -EFAULT;
2269
2270 return 0;
2271}
2272
2273int tcp_ao_get_mkts(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2274{
2275 struct tcp_ao_info *ao_info;
2276
2277 ao_info = setsockopt_ao_info(sk);
2278 if (IS_ERR(ao_info))
2279 return PTR_ERR(ao_info);
2280 if (!ao_info)
2281 return -ENOENT;
2282
2283 return tcp_ao_copy_mkts_to_user(ao_info, optval, optlen);
2284}
2285
2286int tcp_ao_get_sock_info(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2287{
2288 struct tcp_ao_info_opt out, in = {};
2289 struct tcp_ao_key *current_key;
2290 struct tcp_ao_info *ao;
2291 int err, len;
2292
2293 if (copy_from_sockptr(&len, optlen, sizeof(int)))
2294 return -EFAULT;
2295
2296 if (len <= 0)
2297 return -EINVAL;
2298
2299 /* Copying this "in" only to check ::reserved, ::reserved2,
2300 * that may be needed to extend (struct tcp_ao_info_opt) and
2301 * what getsockopt() provides in future.
2302 */
2303 err = copy_struct_from_sockptr(&in, sizeof(in), optval, len);
2304 if (err)
2305 return err;
2306
2307 if (in.reserved != 0 || in.reserved2 != 0)
2308 return -EINVAL;
2309
2310 ao = setsockopt_ao_info(sk);
2311 if (IS_ERR(ao))
2312 return PTR_ERR(ao);
2313 if (!ao)
2314 return -ENOENT;
2315
2316 memset(&out, 0, sizeof(out));
2317 out.ao_required = ao->ao_required;
2318 out.accept_icmps = ao->accept_icmps;
2319 out.pkt_good = atomic64_read(&ao->counters.pkt_good);
2320 out.pkt_bad = atomic64_read(&ao->counters.pkt_bad);
2321 out.pkt_key_not_found = atomic64_read(&ao->counters.key_not_found);
2322 out.pkt_ao_required = atomic64_read(&ao->counters.ao_required);
2323 out.pkt_dropped_icmp = atomic64_read(&ao->counters.dropped_icmp);
2324
2325 current_key = READ_ONCE(ao->current_key);
2326 if (current_key) {
2327 out.set_current = 1;
2328 out.current_key = current_key->sndid;
2329 }
2330 if (ao->rnext_key) {
2331 out.set_rnext = 1;
2332 out.rnext = ao->rnext_key->rcvid;
2333 }
2334
2335 if (copy_to_sockptr(optval, &out, min_t(int, len, sizeof(out))))
2336 return -EFAULT;
2337
2338 return 0;
2339}
2340
2341int tcp_ao_set_repair(struct sock *sk, sockptr_t optval, unsigned int optlen)
2342{
2343 struct tcp_sock *tp = tcp_sk(sk);
2344 struct tcp_ao_repair cmd;
2345 struct tcp_ao_key *key;
2346 struct tcp_ao_info *ao;
2347 int err;
2348
2349 if (optlen < sizeof(cmd))
2350 return -EINVAL;
2351
2352 err = copy_struct_from_sockptr(&cmd, sizeof(cmd), optval, optlen);
2353 if (err)
2354 return err;
2355
2356 if (!tp->repair)
2357 return -EPERM;
2358
2359 ao = setsockopt_ao_info(sk);
2360 if (IS_ERR(ao))
2361 return PTR_ERR(ao);
2362 if (!ao)
2363 return -ENOENT;
2364
2365 WRITE_ONCE(ao->lisn, cmd.snt_isn);
2366 WRITE_ONCE(ao->risn, cmd.rcv_isn);
2367 WRITE_ONCE(ao->snd_sne, cmd.snd_sne);
2368 WRITE_ONCE(ao->rcv_sne, cmd.rcv_sne);
2369
2370 hlist_for_each_entry_rcu(key, &ao->head, node)
2371 tcp_ao_cache_traffic_keys(sk, ao, key);
2372
2373 return 0;
2374}
2375
2376int tcp_ao_get_repair(struct sock *sk, sockptr_t optval, sockptr_t optlen)
2377{
2378 struct tcp_sock *tp = tcp_sk(sk);
2379 struct tcp_ao_repair opt;
2380 struct tcp_ao_info *ao;
2381 int len;
2382
2383 if (copy_from_sockptr(&len, optlen, sizeof(int)))
2384 return -EFAULT;
2385
2386 if (len <= 0)
2387 return -EINVAL;
2388
2389 if (!tp->repair)
2390 return -EPERM;
2391
2392 rcu_read_lock();
2393 ao = getsockopt_ao_info(sk);
2394 if (IS_ERR_OR_NULL(ao)) {
2395 rcu_read_unlock();
2396 return ao ? PTR_ERR(ao) : -ENOENT;
2397 }
2398
2399 opt.snt_isn = ao->lisn;
2400 opt.rcv_isn = ao->risn;
2401 opt.snd_sne = READ_ONCE(ao->snd_sne);
2402 opt.rcv_sne = READ_ONCE(ao->rcv_sne);
2403 rcu_read_unlock();
2404
2405 if (copy_to_sockptr(optval, &opt, min_t(int, len, sizeof(opt))))
2406 return -EFAULT;
2407 return 0;
2408}