1 // SPDX-License-Identifier: GPL-2.0
6 #include <linux/pkt_cls.h>
9 #include <linux/if_ether.h>
10 #include <linux/icmp.h>
12 #include <linux/ipv6.h>
13 #include <linux/tcp.h>
14 #include <linux/udp.h>
15 #include <linux/if_packet.h>
16 #include <sys/socket.h>
17 #include <linux/if_tunnel.h>
18 #include <linux/mpls.h>
19 #include <bpf/bpf_helpers.h>
20 #include <bpf/bpf_endian.h>
22 #define PROG(F) PROG_(F, _##F)
23 #define PROG_(NUM, NAME) SEC("flow_dissector") int flow_dissector_##NUM
25 #define FLOW_CONTINUE_SADDR 0x7f00007f /* 127.0.0.127 */
27 /* These are the identifiers of the BPF programs that will be used in tail
28 * calls. Name is limited to 16 characters, with the terminating character and
29 * bpf_func_ above, we have only 6 to work with, anything after will be cropped.
33 #define IPV6OP 2 /* Destination/Hop-by-Hop Options IPv6 Ext. Header */
34 #define IPV6FR 3 /* Fragmentation IPv6 Extension Header */
40 #define IP_OFFSET 0x1FFF
42 #define IP6_OFFSET 0xFFF8
46 __be16 h_vlan_encapsulated_proto;
58 __be32 identification;
62 __uint(type, BPF_MAP_TYPE_PROG_ARRAY);
63 __uint(max_entries, MAX_PROG);
64 __uint(key_size, sizeof(__u32));
65 __uint(value_size, sizeof(__u32));
66 } jmp_table SEC(".maps");
69 __uint(type, BPF_MAP_TYPE_HASH);
70 __uint(max_entries, 1024);
72 __type(value, struct bpf_flow_keys);
73 } last_dissection SEC(".maps");
75 static __always_inline int export_flow_keys(struct bpf_flow_keys *keys,
78 __u32 key = (__u32)(keys->sport) << 16 | keys->dport;
79 struct bpf_flow_keys val;
81 memcpy(&val, keys, sizeof(val));
82 bpf_map_update_elem(&last_dissection, &key, &val, BPF_ANY);
86 #define IPV6_FLOWLABEL_MASK __bpf_constant_htonl(0x000FFFFF)
87 static inline __be32 ip6_flowlabel(const struct ipv6hdr *hdr)
89 return *(__be32 *)hdr & IPV6_FLOWLABEL_MASK;
92 static __always_inline void *bpf_flow_dissect_get_header(struct __sk_buff *skb,
96 void *data_end = (void *)(long)skb->data_end;
97 void *data = (void *)(long)skb->data;
98 __u16 thoff = skb->flow_keys->thoff;
101 /* Verifies this variable offset does not overflow */
102 if (thoff > (USHRT_MAX - hdr_size))
106 if (hdr + hdr_size <= data_end)
109 if (bpf_skb_load_bytes(skb, thoff, buffer, hdr_size))
115 /* Dispatches on ETHERTYPE */
116 static __always_inline int parse_eth_proto(struct __sk_buff *skb, __be16 proto)
118 struct bpf_flow_keys *keys = skb->flow_keys;
121 case bpf_htons(ETH_P_IP):
122 bpf_tail_call_static(skb, &jmp_table, IP);
124 case bpf_htons(ETH_P_IPV6):
125 bpf_tail_call_static(skb, &jmp_table, IPV6);
127 case bpf_htons(ETH_P_MPLS_MC):
128 case bpf_htons(ETH_P_MPLS_UC):
129 bpf_tail_call_static(skb, &jmp_table, MPLS);
131 case bpf_htons(ETH_P_8021Q):
132 case bpf_htons(ETH_P_8021AD):
133 bpf_tail_call_static(skb, &jmp_table, VLAN);
136 /* Protocol not supported */
137 return export_flow_keys(keys, BPF_DROP);
140 return export_flow_keys(keys, BPF_DROP);
143 SEC("flow_dissector")
144 int _dissect(struct __sk_buff *skb)
146 struct bpf_flow_keys *keys = skb->flow_keys;
148 if (keys->n_proto == bpf_htons(ETH_P_IP)) {
149 /* IP traffic from FLOW_CONTINUE_SADDR falls-back to
152 struct iphdr *iph, _iph;
154 iph = bpf_flow_dissect_get_header(skb, sizeof(*iph), &_iph);
155 if (iph && iph->ihl == 5 &&
156 iph->saddr == bpf_htonl(FLOW_CONTINUE_SADDR)) {
157 return BPF_FLOW_DISSECTOR_CONTINUE;
161 return parse_eth_proto(skb, keys->n_proto);
164 /* Parses on IPPROTO_* */
165 static __always_inline int parse_ip_proto(struct __sk_buff *skb, __u8 proto)
167 struct bpf_flow_keys *keys = skb->flow_keys;
168 void *data_end = (void *)(long)skb->data_end;
169 struct icmphdr *icmp, _icmp;
170 struct gre_hdr *gre, _gre;
171 struct ethhdr *eth, _eth;
172 struct tcphdr *tcp, _tcp;
173 struct udphdr *udp, _udp;
177 icmp = bpf_flow_dissect_get_header(skb, sizeof(*icmp), &_icmp);
179 return export_flow_keys(keys, BPF_DROP);
180 return export_flow_keys(keys, BPF_OK);
182 keys->is_encap = true;
183 if (keys->flags & BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP)
184 return export_flow_keys(keys, BPF_OK);
186 return parse_eth_proto(skb, bpf_htons(ETH_P_IP));
188 keys->is_encap = true;
189 if (keys->flags & BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP)
190 return export_flow_keys(keys, BPF_OK);
192 return parse_eth_proto(skb, bpf_htons(ETH_P_IPV6));
194 gre = bpf_flow_dissect_get_header(skb, sizeof(*gre), &_gre);
196 return export_flow_keys(keys, BPF_DROP);
198 if (bpf_htons(gre->flags & GRE_VERSION))
199 /* Only inspect standard GRE packets with version 0 */
200 return export_flow_keys(keys, BPF_OK);
202 keys->thoff += sizeof(*gre); /* Step over GRE Flags and Proto */
203 if (GRE_IS_CSUM(gre->flags))
204 keys->thoff += 4; /* Step over chksum and Padding */
205 if (GRE_IS_KEY(gre->flags))
206 keys->thoff += 4; /* Step over key */
207 if (GRE_IS_SEQ(gre->flags))
208 keys->thoff += 4; /* Step over sequence number */
210 keys->is_encap = true;
211 if (keys->flags & BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP)
212 return export_flow_keys(keys, BPF_OK);
214 if (gre->proto == bpf_htons(ETH_P_TEB)) {
215 eth = bpf_flow_dissect_get_header(skb, sizeof(*eth),
218 return export_flow_keys(keys, BPF_DROP);
220 keys->thoff += sizeof(*eth);
222 return parse_eth_proto(skb, eth->h_proto);
224 return parse_eth_proto(skb, gre->proto);
227 tcp = bpf_flow_dissect_get_header(skb, sizeof(*tcp), &_tcp);
229 return export_flow_keys(keys, BPF_DROP);
232 return export_flow_keys(keys, BPF_DROP);
234 if ((__u8 *)tcp + (tcp->doff << 2) > data_end)
235 return export_flow_keys(keys, BPF_DROP);
237 keys->sport = tcp->source;
238 keys->dport = tcp->dest;
239 return export_flow_keys(keys, BPF_OK);
241 case IPPROTO_UDPLITE:
242 udp = bpf_flow_dissect_get_header(skb, sizeof(*udp), &_udp);
244 return export_flow_keys(keys, BPF_DROP);
246 keys->sport = udp->source;
247 keys->dport = udp->dest;
248 return export_flow_keys(keys, BPF_OK);
250 return export_flow_keys(keys, BPF_DROP);
253 return export_flow_keys(keys, BPF_DROP);
256 static __always_inline int parse_ipv6_proto(struct __sk_buff *skb, __u8 nexthdr)
258 struct bpf_flow_keys *keys = skb->flow_keys;
261 case IPPROTO_HOPOPTS:
262 case IPPROTO_DSTOPTS:
263 bpf_tail_call_static(skb, &jmp_table, IPV6OP);
265 case IPPROTO_FRAGMENT:
266 bpf_tail_call_static(skb, &jmp_table, IPV6FR);
269 return parse_ip_proto(skb, nexthdr);
272 return export_flow_keys(keys, BPF_DROP);
275 PROG(IP)(struct __sk_buff *skb)
277 void *data_end = (void *)(long)skb->data_end;
278 struct bpf_flow_keys *keys = skb->flow_keys;
279 void *data = (void *)(long)skb->data;
280 struct iphdr *iph, _iph;
283 iph = bpf_flow_dissect_get_header(skb, sizeof(*iph), &_iph);
285 return export_flow_keys(keys, BPF_DROP);
287 /* IP header cannot be smaller than 20 bytes */
289 return export_flow_keys(keys, BPF_DROP);
291 keys->addr_proto = ETH_P_IP;
292 keys->ipv4_src = iph->saddr;
293 keys->ipv4_dst = iph->daddr;
294 keys->ip_proto = iph->protocol;
296 keys->thoff += iph->ihl << 2;
297 if (data + keys->thoff > data_end)
298 return export_flow_keys(keys, BPF_DROP);
300 if (iph->frag_off & bpf_htons(IP_MF | IP_OFFSET)) {
301 keys->is_frag = true;
302 if (iph->frag_off & bpf_htons(IP_OFFSET)) {
303 /* From second fragment on, packets do not have headers
308 keys->is_first_frag = true;
309 /* No need to parse fragmented packet unless
310 * explicitly asked for.
313 BPF_FLOW_DISSECTOR_F_PARSE_1ST_FRAG))
319 return export_flow_keys(keys, BPF_OK);
321 return parse_ip_proto(skb, iph->protocol);
324 PROG(IPV6)(struct __sk_buff *skb)
326 struct bpf_flow_keys *keys = skb->flow_keys;
327 struct ipv6hdr *ip6h, _ip6h;
329 ip6h = bpf_flow_dissect_get_header(skb, sizeof(*ip6h), &_ip6h);
331 return export_flow_keys(keys, BPF_DROP);
333 keys->addr_proto = ETH_P_IPV6;
334 memcpy(&keys->ipv6_src, &ip6h->saddr, 2*sizeof(ip6h->saddr));
336 keys->thoff += sizeof(struct ipv6hdr);
337 keys->ip_proto = ip6h->nexthdr;
338 keys->flow_label = ip6_flowlabel(ip6h);
340 if (keys->flow_label && keys->flags & BPF_FLOW_DISSECTOR_F_STOP_AT_FLOW_LABEL)
341 return export_flow_keys(keys, BPF_OK);
343 return parse_ipv6_proto(skb, ip6h->nexthdr);
346 PROG(IPV6OP)(struct __sk_buff *skb)
348 struct bpf_flow_keys *keys = skb->flow_keys;
349 struct ipv6_opt_hdr *ip6h, _ip6h;
351 ip6h = bpf_flow_dissect_get_header(skb, sizeof(*ip6h), &_ip6h);
353 return export_flow_keys(keys, BPF_DROP);
355 /* hlen is in 8-octets and does not include the first 8 bytes
358 keys->thoff += (1 + ip6h->hdrlen) << 3;
359 keys->ip_proto = ip6h->nexthdr;
361 return parse_ipv6_proto(skb, ip6h->nexthdr);
364 PROG(IPV6FR)(struct __sk_buff *skb)
366 struct bpf_flow_keys *keys = skb->flow_keys;
367 struct frag_hdr *fragh, _fragh;
369 fragh = bpf_flow_dissect_get_header(skb, sizeof(*fragh), &_fragh);
371 return export_flow_keys(keys, BPF_DROP);
373 keys->thoff += sizeof(*fragh);
374 keys->is_frag = true;
375 keys->ip_proto = fragh->nexthdr;
377 if (!(fragh->frag_off & bpf_htons(IP6_OFFSET))) {
378 keys->is_first_frag = true;
380 /* No need to parse fragmented packet unless
381 * explicitly asked for.
383 if (!(keys->flags & BPF_FLOW_DISSECTOR_F_PARSE_1ST_FRAG))
384 return export_flow_keys(keys, BPF_OK);
386 return export_flow_keys(keys, BPF_OK);
389 return parse_ipv6_proto(skb, fragh->nexthdr);
392 PROG(MPLS)(struct __sk_buff *skb)
394 struct bpf_flow_keys *keys = skb->flow_keys;
395 struct mpls_label *mpls, _mpls;
397 mpls = bpf_flow_dissect_get_header(skb, sizeof(*mpls), &_mpls);
399 return export_flow_keys(keys, BPF_DROP);
401 return export_flow_keys(keys, BPF_OK);
404 PROG(VLAN)(struct __sk_buff *skb)
406 struct bpf_flow_keys *keys = skb->flow_keys;
407 struct vlan_hdr *vlan, _vlan;
409 /* Account for double-tagging */
410 if (keys->n_proto == bpf_htons(ETH_P_8021AD)) {
411 vlan = bpf_flow_dissect_get_header(skb, sizeof(*vlan), &_vlan);
413 return export_flow_keys(keys, BPF_DROP);
415 if (vlan->h_vlan_encapsulated_proto != bpf_htons(ETH_P_8021Q))
416 return export_flow_keys(keys, BPF_DROP);
418 keys->nhoff += sizeof(*vlan);
419 keys->thoff += sizeof(*vlan);
422 vlan = bpf_flow_dissect_get_header(skb, sizeof(*vlan), &_vlan);
424 return export_flow_keys(keys, BPF_DROP);
426 keys->nhoff += sizeof(*vlan);
427 keys->thoff += sizeof(*vlan);
428 /* Only allow 8021AD + 8021Q double tagging and no triple tagging.*/
429 if (vlan->h_vlan_encapsulated_proto == bpf_htons(ETH_P_8021AD) ||
430 vlan->h_vlan_encapsulated_proto == bpf_htons(ETH_P_8021Q))
431 return export_flow_keys(keys, BPF_DROP);
433 keys->n_proto = vlan->h_vlan_encapsulated_proto;
434 return parse_eth_proto(skb, vlan->h_vlan_encapsulated_proto);
437 char __license[] SEC("license") = "GPL";