Linux Audio

Check our new training course

Loading...
v6.13.7
  1/* SPDX-License-Identifier: GPL-2.0 */
  2// Copyright (c) 2018 Covalent IO, Inc. http://covalent.io
  3
  4#include <stddef.h>
  5#include <stdbool.h>
  6#include <string.h>
  7#include <linux/bpf.h>
  8#include <linux/if_ether.h>
  9#include <linux/in.h>
 10#include <linux/ip.h>
 11#include <linux/ipv6.h>
 12#include <linux/pkt_cls.h>
 13#include <linux/tcp.h>
 14#include <sys/socket.h>
 15#include <bpf/bpf_helpers.h>
 16#include <bpf/bpf_endian.h>
 17
 
 18char _license[] SEC("license") = "GPL";
 19
 20/* Fill 'tuple' with L3 info, and attempt to find L4. On fail, return NULL. */
 21static struct bpf_sock_tuple *get_tuple(void *data, __u64 nh_off,
 22					void *data_end, __u16 eth_proto,
 23					bool *ipv4)
 24{
 25	struct bpf_sock_tuple *result;
 26	__u64 ihl_len = 0;
 27	__u8 proto = 0;
 
 28
 29	if (eth_proto == bpf_htons(ETH_P_IP)) {
 30		struct iphdr *iph = (struct iphdr *)(data + nh_off);
 31
 32		if (iph + 1 > data_end)
 33			return NULL;
 34		ihl_len = iph->ihl * 4;
 35		proto = iph->protocol;
 36		*ipv4 = true;
 37		result = (struct bpf_sock_tuple *)&iph->saddr;
 38	} else if (eth_proto == bpf_htons(ETH_P_IPV6)) {
 39		struct ipv6hdr *ip6h = (struct ipv6hdr *)(data + nh_off);
 40
 41		if (ip6h + 1 > data_end)
 42			return NULL;
 43		ihl_len = sizeof(*ip6h);
 44		proto = ip6h->nexthdr;
 45		*ipv4 = true;
 46		result = (struct bpf_sock_tuple *)&ip6h->saddr;
 47	}
 48
 49	if (data + nh_off + ihl_len > data_end || proto != IPPROTO_TCP)
 50		return NULL;
 51
 52	return result;
 53}
 54
 55SEC("?tc")
 56int sk_lookup_success(struct __sk_buff *skb)
 57{
 58	void *data_end = (void *)(long)skb->data_end;
 59	void *data = (void *)(long)skb->data;
 60	struct ethhdr *eth = (struct ethhdr *)(data);
 61	struct bpf_sock_tuple *tuple;
 62	struct bpf_sock *sk;
 63	size_t tuple_len;
 64	bool ipv4;
 65
 66	if (eth + 1 > data_end)
 67		return TC_ACT_SHOT;
 68
 69	tuple = get_tuple(data, sizeof(*eth), data_end, eth->h_proto, &ipv4);
 70	if (!tuple || tuple + sizeof *tuple > data_end)
 71		return TC_ACT_SHOT;
 72
 73	tuple_len = ipv4 ? sizeof(tuple->ipv4) : sizeof(tuple->ipv6);
 74	sk = bpf_sk_lookup_tcp(skb, tuple, tuple_len, BPF_F_CURRENT_NETNS, 0);
 75	bpf_printk("sk=%d\n", sk ? 1 : 0);
 76	if (sk)
 77		bpf_sk_release(sk);
 78	return sk ? TC_ACT_OK : TC_ACT_UNSPEC;
 79}
 80
 81SEC("?tc")
 82int sk_lookup_success_simple(struct __sk_buff *skb)
 83{
 84	struct bpf_sock_tuple tuple = {};
 85	struct bpf_sock *sk;
 86
 87	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
 88	if (sk)
 89		bpf_sk_release(sk);
 90	return 0;
 91}
 92
 93SEC("?tc")
 94int err_use_after_free(struct __sk_buff *skb)
 95{
 96	struct bpf_sock_tuple tuple = {};
 97	struct bpf_sock *sk;
 98	__u32 family = 0;
 99
100	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
101	if (sk) {
102		bpf_sk_release(sk);
103		family = sk->family;
104	}
105	return family;
106}
107
108SEC("?tc")
109int err_modify_sk_pointer(struct __sk_buff *skb)
110{
111	struct bpf_sock_tuple tuple = {};
112	struct bpf_sock *sk;
 
113
114	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
115	if (sk) {
116		sk += 1;
117		bpf_sk_release(sk);
118	}
119	return 0;
120}
121
122SEC("?tc")
123int err_modify_sk_or_null_pointer(struct __sk_buff *skb)
124{
125	struct bpf_sock_tuple tuple = {};
126	struct bpf_sock *sk;
 
127
128	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
129	sk += 1;
130	if (sk)
131		bpf_sk_release(sk);
132	return 0;
133}
134
135SEC("?tc")
136int err_no_release(struct __sk_buff *skb)
137{
138	struct bpf_sock_tuple tuple = {};
139
140	bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
141	return 0;
142}
143
144SEC("?tc")
145int err_release_twice(struct __sk_buff *skb)
146{
147	struct bpf_sock_tuple tuple = {};
148	struct bpf_sock *sk;
149
150	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
151	bpf_sk_release(sk);
152	bpf_sk_release(sk);
153	return 0;
154}
155
156SEC("?tc")
157int err_release_unchecked(struct __sk_buff *skb)
158{
159	struct bpf_sock_tuple tuple = {};
160	struct bpf_sock *sk;
161
162	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
163	bpf_sk_release(sk);
164	return 0;
165}
166
167void lookup_no_release(struct __sk_buff *skb)
168{
169	struct bpf_sock_tuple tuple = {};
170	bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
171}
172
173SEC("?tc")
174int err_no_release_subcall(struct __sk_buff *skb)
175{
176	lookup_no_release(skb);
177	return 0;
178}
v5.9
  1/* SPDX-License-Identifier: GPL-2.0 */
  2// Copyright (c) 2018 Covalent IO, Inc. http://covalent.io
  3
  4#include <stddef.h>
  5#include <stdbool.h>
  6#include <string.h>
  7#include <linux/bpf.h>
  8#include <linux/if_ether.h>
  9#include <linux/in.h>
 10#include <linux/ip.h>
 11#include <linux/ipv6.h>
 12#include <linux/pkt_cls.h>
 13#include <linux/tcp.h>
 14#include <sys/socket.h>
 15#include <bpf/bpf_helpers.h>
 16#include <bpf/bpf_endian.h>
 17
 18int _version SEC("version") = 1;
 19char _license[] SEC("license") = "GPL";
 20
 21/* Fill 'tuple' with L3 info, and attempt to find L4. On fail, return NULL. */
 22static struct bpf_sock_tuple *get_tuple(void *data, __u64 nh_off,
 23					void *data_end, __u16 eth_proto,
 24					bool *ipv4)
 25{
 26	struct bpf_sock_tuple *result;
 
 27	__u8 proto = 0;
 28	__u64 ihl_len;
 29
 30	if (eth_proto == bpf_htons(ETH_P_IP)) {
 31		struct iphdr *iph = (struct iphdr *)(data + nh_off);
 32
 33		if (iph + 1 > data_end)
 34			return NULL;
 35		ihl_len = iph->ihl * 4;
 36		proto = iph->protocol;
 37		*ipv4 = true;
 38		result = (struct bpf_sock_tuple *)&iph->saddr;
 39	} else if (eth_proto == bpf_htons(ETH_P_IPV6)) {
 40		struct ipv6hdr *ip6h = (struct ipv6hdr *)(data + nh_off);
 41
 42		if (ip6h + 1 > data_end)
 43			return NULL;
 44		ihl_len = sizeof(*ip6h);
 45		proto = ip6h->nexthdr;
 46		*ipv4 = true;
 47		result = (struct bpf_sock_tuple *)&ip6h->saddr;
 48	}
 49
 50	if (data + nh_off + ihl_len > data_end || proto != IPPROTO_TCP)
 51		return NULL;
 52
 53	return result;
 54}
 55
 56SEC("classifier/sk_lookup_success")
 57int bpf_sk_lookup_test0(struct __sk_buff *skb)
 58{
 59	void *data_end = (void *)(long)skb->data_end;
 60	void *data = (void *)(long)skb->data;
 61	struct ethhdr *eth = (struct ethhdr *)(data);
 62	struct bpf_sock_tuple *tuple;
 63	struct bpf_sock *sk;
 64	size_t tuple_len;
 65	bool ipv4;
 66
 67	if (eth + 1 > data_end)
 68		return TC_ACT_SHOT;
 69
 70	tuple = get_tuple(data, sizeof(*eth), data_end, eth->h_proto, &ipv4);
 71	if (!tuple || tuple + sizeof *tuple > data_end)
 72		return TC_ACT_SHOT;
 73
 74	tuple_len = ipv4 ? sizeof(tuple->ipv4) : sizeof(tuple->ipv6);
 75	sk = bpf_sk_lookup_tcp(skb, tuple, tuple_len, BPF_F_CURRENT_NETNS, 0);
 76	bpf_printk("sk=%d\n", sk ? 1 : 0);
 77	if (sk)
 78		bpf_sk_release(sk);
 79	return sk ? TC_ACT_OK : TC_ACT_UNSPEC;
 80}
 81
 82SEC("classifier/sk_lookup_success_simple")
 83int bpf_sk_lookup_test1(struct __sk_buff *skb)
 84{
 85	struct bpf_sock_tuple tuple = {};
 86	struct bpf_sock *sk;
 87
 88	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
 89	if (sk)
 90		bpf_sk_release(sk);
 91	return 0;
 92}
 93
 94SEC("classifier/fail_use_after_free")
 95int bpf_sk_lookup_uaf(struct __sk_buff *skb)
 96{
 97	struct bpf_sock_tuple tuple = {};
 98	struct bpf_sock *sk;
 99	__u32 family = 0;
100
101	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
102	if (sk) {
103		bpf_sk_release(sk);
104		family = sk->family;
105	}
106	return family;
107}
108
109SEC("classifier/fail_modify_sk_pointer")
110int bpf_sk_lookup_modptr(struct __sk_buff *skb)
111{
112	struct bpf_sock_tuple tuple = {};
113	struct bpf_sock *sk;
114	__u32 family;
115
116	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
117	if (sk) {
118		sk += 1;
119		bpf_sk_release(sk);
120	}
121	return 0;
122}
123
124SEC("classifier/fail_modify_sk_or_null_pointer")
125int bpf_sk_lookup_modptr_or_null(struct __sk_buff *skb)
126{
127	struct bpf_sock_tuple tuple = {};
128	struct bpf_sock *sk;
129	__u32 family;
130
131	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
132	sk += 1;
133	if (sk)
134		bpf_sk_release(sk);
135	return 0;
136}
137
138SEC("classifier/fail_no_release")
139int bpf_sk_lookup_test2(struct __sk_buff *skb)
140{
141	struct bpf_sock_tuple tuple = {};
142
143	bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
144	return 0;
145}
146
147SEC("classifier/fail_release_twice")
148int bpf_sk_lookup_test3(struct __sk_buff *skb)
149{
150	struct bpf_sock_tuple tuple = {};
151	struct bpf_sock *sk;
152
153	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
154	bpf_sk_release(sk);
155	bpf_sk_release(sk);
156	return 0;
157}
158
159SEC("classifier/fail_release_unchecked")
160int bpf_sk_lookup_test4(struct __sk_buff *skb)
161{
162	struct bpf_sock_tuple tuple = {};
163	struct bpf_sock *sk;
164
165	sk = bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
166	bpf_sk_release(sk);
167	return 0;
168}
169
170void lookup_no_release(struct __sk_buff *skb)
171{
172	struct bpf_sock_tuple tuple = {};
173	bpf_sk_lookup_tcp(skb, &tuple, sizeof(tuple), BPF_F_CURRENT_NETNS, 0);
174}
175
176SEC("classifier/fail_no_release_subcall")
177int bpf_sk_lookup_test5(struct __sk_buff *skb)
178{
179	lookup_no_release(skb);
180	return 0;
181}