1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (c) 2008-2009 Patrick McHardy <kaber@trash.net>
4 * Copyright (c) 2016 Pablo Neira Ayuso <pablo@netfilter.org>
6 * Development of this code funded by Astaro AG (http://www.astaro.com/)
9 #include <linux/kernel.h>
10 #include <linux/init.h>
11 #include <linux/module.h>
12 #include <linux/netlink.h>
13 #include <linux/netfilter.h>
14 #include <linux/netfilter/nf_tables.h>
15 #include <net/netfilter/nf_tables.h>
16 #include <net/netfilter/nf_conntrack.h>
17 #include <net/netfilter/nf_conntrack_acct.h>
18 #include <net/netfilter/nf_conntrack_tuple.h>
19 #include <net/netfilter/nf_conntrack_helper.h>
20 #include <net/netfilter/nf_conntrack_ecache.h>
21 #include <net/netfilter/nf_conntrack_labels.h>
22 #include <net/netfilter/nf_conntrack_timeout.h>
23 #include <net/netfilter/nf_conntrack_l4proto.h>
24 #include <net/netfilter/nf_conntrack_expect.h>
27 enum nft_ct_keys key:8;
28 enum ip_conntrack_dir dir:8;
36 struct nft_ct_helper_obj {
37 struct nf_conntrack_helper *helper4;
38 struct nf_conntrack_helper *helper6;
42 #ifdef CONFIG_NF_CONNTRACK_ZONES
43 static DEFINE_PER_CPU(struct nf_conn *, nft_ct_pcpu_template);
44 static unsigned int nft_ct_pcpu_template_refcnt __read_mostly;
45 static DEFINE_MUTEX(nft_ct_pcpu_mutex);
48 static u64 nft_ct_get_eval_counter(const struct nf_conn_counter *c,
50 enum ip_conntrack_dir d)
52 if (d < IP_CT_DIR_MAX)
53 return k == NFT_CT_BYTES ? atomic64_read(&c[d].bytes) :
54 atomic64_read(&c[d].packets);
56 return nft_ct_get_eval_counter(c, k, IP_CT_DIR_ORIGINAL) +
57 nft_ct_get_eval_counter(c, k, IP_CT_DIR_REPLY);
60 static void nft_ct_get_eval(const struct nft_expr *expr,
61 struct nft_regs *regs,
62 const struct nft_pktinfo *pkt)
64 const struct nft_ct *priv = nft_expr_priv(expr);
65 u32 *dest = ®s->data[priv->dreg];
66 enum ip_conntrack_info ctinfo;
67 const struct nf_conn *ct;
68 const struct nf_conn_help *help;
69 const struct nf_conntrack_tuple *tuple;
70 const struct nf_conntrack_helper *helper;
73 ct = nf_ct_get(pkt->skb, &ctinfo);
78 state = NF_CT_STATE_BIT(ctinfo);
79 else if (ctinfo == IP_CT_UNTRACKED)
80 state = NF_CT_STATE_UNTRACKED_BIT;
82 state = NF_CT_STATE_INVALID_BIT;
93 case NFT_CT_DIRECTION:
94 nft_reg_store8(dest, CTINFO2DIR(ctinfo));
99 #ifdef CONFIG_NF_CONNTRACK_MARK
104 #ifdef CONFIG_NF_CONNTRACK_SECMARK
109 case NFT_CT_EXPIRATION:
110 *dest = jiffies_to_msecs(nf_ct_expires(ct));
113 if (ct->master == NULL)
115 help = nfct_help(ct->master);
118 helper = rcu_dereference(help->helper);
121 strncpy((char *)dest, helper->name, NF_CT_HELPER_NAME_LEN);
123 #ifdef CONFIG_NF_CONNTRACK_LABELS
124 case NFT_CT_LABELS: {
125 struct nf_conn_labels *labels = nf_ct_labels_find(ct);
128 memcpy(dest, labels->bits, NF_CT_LABELS_MAX_SIZE);
130 memset(dest, 0, NF_CT_LABELS_MAX_SIZE);
136 const struct nf_conn_acct *acct = nf_conn_acct_find(ct);
140 count = nft_ct_get_eval_counter(acct->counter,
141 priv->key, priv->dir);
142 memcpy(dest, &count, sizeof(count));
145 case NFT_CT_AVGPKT: {
146 const struct nf_conn_acct *acct = nf_conn_acct_find(ct);
147 u64 avgcnt = 0, bcnt = 0, pcnt = 0;
150 pcnt = nft_ct_get_eval_counter(acct->counter,
151 NFT_CT_PKTS, priv->dir);
152 bcnt = nft_ct_get_eval_counter(acct->counter,
153 NFT_CT_BYTES, priv->dir);
155 avgcnt = div64_u64(bcnt, pcnt);
158 memcpy(dest, &avgcnt, sizeof(avgcnt));
161 case NFT_CT_L3PROTOCOL:
162 nft_reg_store8(dest, nf_ct_l3num(ct));
164 case NFT_CT_PROTOCOL:
165 nft_reg_store8(dest, nf_ct_protonum(ct));
167 #ifdef CONFIG_NF_CONNTRACK_ZONES
169 const struct nf_conntrack_zone *zone = nf_ct_zone(ct);
172 if (priv->dir < IP_CT_DIR_MAX)
173 zoneid = nf_ct_zone_id(zone, priv->dir);
177 nft_reg_store16(dest, zoneid);
182 *dest = nf_ct_get_id(ct);
188 tuple = &ct->tuplehash[priv->dir].tuple;
191 memcpy(dest, tuple->src.u3.all,
192 nf_ct_l3num(ct) == NFPROTO_IPV4 ? 4 : 16);
195 memcpy(dest, tuple->dst.u3.all,
196 nf_ct_l3num(ct) == NFPROTO_IPV4 ? 4 : 16);
198 case NFT_CT_PROTO_SRC:
199 nft_reg_store16(dest, (__force u16)tuple->src.u.all);
201 case NFT_CT_PROTO_DST:
202 nft_reg_store16(dest, (__force u16)tuple->dst.u.all);
205 if (nf_ct_l3num(ct) != NFPROTO_IPV4)
207 *dest = (__force __u32)tuple->src.u3.ip;
210 if (nf_ct_l3num(ct) != NFPROTO_IPV4)
212 *dest = (__force __u32)tuple->dst.u3.ip;
215 if (nf_ct_l3num(ct) != NFPROTO_IPV6)
217 memcpy(dest, tuple->src.u3.ip6, sizeof(struct in6_addr));
220 if (nf_ct_l3num(ct) != NFPROTO_IPV6)
222 memcpy(dest, tuple->dst.u3.ip6, sizeof(struct in6_addr));
229 regs->verdict.code = NFT_BREAK;
232 #ifdef CONFIG_NF_CONNTRACK_ZONES
233 static void nft_ct_set_zone_eval(const struct nft_expr *expr,
234 struct nft_regs *regs,
235 const struct nft_pktinfo *pkt)
237 struct nf_conntrack_zone zone = { .dir = NF_CT_DEFAULT_ZONE_DIR };
238 const struct nft_ct *priv = nft_expr_priv(expr);
239 struct sk_buff *skb = pkt->skb;
240 enum ip_conntrack_info ctinfo;
241 u16 value = nft_reg_load16(®s->data[priv->sreg]);
244 ct = nf_ct_get(skb, &ctinfo);
245 if (ct) /* already tracked */
251 case IP_CT_DIR_ORIGINAL:
252 zone.dir = NF_CT_ZONE_DIR_ORIG;
254 case IP_CT_DIR_REPLY:
255 zone.dir = NF_CT_ZONE_DIR_REPL;
261 ct = this_cpu_read(nft_ct_pcpu_template);
263 if (likely(refcount_read(&ct->ct_general.use) == 1)) {
264 refcount_inc(&ct->ct_general.use);
265 nf_ct_zone_add(ct, &zone);
267 /* previous skb got queued to userspace, allocate temporary
268 * one until percpu template can be reused.
270 ct = nf_ct_tmpl_alloc(nft_net(pkt), &zone, GFP_ATOMIC);
272 regs->verdict.code = NF_DROP;
277 nf_ct_set(skb, ct, IP_CT_NEW);
281 static void nft_ct_set_eval(const struct nft_expr *expr,
282 struct nft_regs *regs,
283 const struct nft_pktinfo *pkt)
285 const struct nft_ct *priv = nft_expr_priv(expr);
286 struct sk_buff *skb = pkt->skb;
287 #if defined(CONFIG_NF_CONNTRACK_MARK) || defined(CONFIG_NF_CONNTRACK_SECMARK)
288 u32 value = regs->data[priv->sreg];
290 enum ip_conntrack_info ctinfo;
293 ct = nf_ct_get(skb, &ctinfo);
294 if (ct == NULL || nf_ct_is_template(ct))
298 #ifdef CONFIG_NF_CONNTRACK_MARK
300 if (ct->mark != value) {
302 nf_conntrack_event_cache(IPCT_MARK, ct);
306 #ifdef CONFIG_NF_CONNTRACK_SECMARK
308 if (ct->secmark != value) {
310 nf_conntrack_event_cache(IPCT_SECMARK, ct);
314 #ifdef CONFIG_NF_CONNTRACK_LABELS
316 nf_connlabels_replace(ct,
317 ®s->data[priv->sreg],
318 ®s->data[priv->sreg],
319 NF_CT_LABELS_MAX_SIZE / sizeof(u32));
322 #ifdef CONFIG_NF_CONNTRACK_EVENTS
323 case NFT_CT_EVENTMASK: {
324 struct nf_conntrack_ecache *e = nf_ct_ecache_find(ct);
325 u32 ctmask = regs->data[priv->sreg];
328 if (e->ctmask != ctmask)
333 if (ctmask && !nf_ct_is_confirmed(ct))
334 nf_ct_ecache_ext_add(ct, ctmask, 0, GFP_ATOMIC);
343 static const struct nla_policy nft_ct_policy[NFTA_CT_MAX + 1] = {
344 [NFTA_CT_DREG] = { .type = NLA_U32 },
345 [NFTA_CT_KEY] = { .type = NLA_U32 },
346 [NFTA_CT_DIRECTION] = { .type = NLA_U8 },
347 [NFTA_CT_SREG] = { .type = NLA_U32 },
350 #ifdef CONFIG_NF_CONNTRACK_ZONES
351 static void nft_ct_tmpl_put_pcpu(void)
356 for_each_possible_cpu(cpu) {
357 ct = per_cpu(nft_ct_pcpu_template, cpu);
361 per_cpu(nft_ct_pcpu_template, cpu) = NULL;
365 static bool nft_ct_tmpl_alloc_pcpu(void)
367 struct nf_conntrack_zone zone = { .id = 0 };
371 if (nft_ct_pcpu_template_refcnt)
374 for_each_possible_cpu(cpu) {
375 tmp = nf_ct_tmpl_alloc(&init_net, &zone, GFP_KERNEL);
377 nft_ct_tmpl_put_pcpu();
381 per_cpu(nft_ct_pcpu_template, cpu) = tmp;
388 static int nft_ct_get_init(const struct nft_ctx *ctx,
389 const struct nft_expr *expr,
390 const struct nlattr * const tb[])
392 struct nft_ct *priv = nft_expr_priv(expr);
396 priv->key = ntohl(nla_get_be32(tb[NFTA_CT_KEY]));
397 priv->dir = IP_CT_DIR_MAX;
399 case NFT_CT_DIRECTION:
400 if (tb[NFTA_CT_DIRECTION] != NULL)
406 #ifdef CONFIG_NF_CONNTRACK_MARK
409 #ifdef CONFIG_NF_CONNTRACK_SECMARK
412 case NFT_CT_EXPIRATION:
413 if (tb[NFTA_CT_DIRECTION] != NULL)
417 #ifdef CONFIG_NF_CONNTRACK_LABELS
419 if (tb[NFTA_CT_DIRECTION] != NULL)
421 len = NF_CT_LABELS_MAX_SIZE;
425 if (tb[NFTA_CT_DIRECTION] != NULL)
427 len = NF_CT_HELPER_NAME_LEN;
430 case NFT_CT_L3PROTOCOL:
431 case NFT_CT_PROTOCOL:
432 /* For compatibility, do not report error if NFTA_CT_DIRECTION
433 * attribute is specified.
439 if (tb[NFTA_CT_DIRECTION] == NULL)
442 switch (ctx->family) {
444 len = sizeof_field(struct nf_conntrack_tuple,
449 len = sizeof_field(struct nf_conntrack_tuple,
453 return -EAFNOSUPPORT;
458 if (tb[NFTA_CT_DIRECTION] == NULL)
461 len = sizeof_field(struct nf_conntrack_tuple, src.u3.ip);
465 if (tb[NFTA_CT_DIRECTION] == NULL)
468 len = sizeof_field(struct nf_conntrack_tuple, src.u3.ip6);
470 case NFT_CT_PROTO_SRC:
471 case NFT_CT_PROTO_DST:
472 if (tb[NFTA_CT_DIRECTION] == NULL)
474 len = sizeof_field(struct nf_conntrack_tuple, src.u.all);
481 #ifdef CONFIG_NF_CONNTRACK_ZONES
493 if (tb[NFTA_CT_DIRECTION] != NULL) {
494 priv->dir = nla_get_u8(tb[NFTA_CT_DIRECTION]);
496 case IP_CT_DIR_ORIGINAL:
497 case IP_CT_DIR_REPLY:
505 err = nft_parse_register_store(ctx, tb[NFTA_CT_DREG], &priv->dreg, NULL,
506 NFT_DATA_VALUE, len);
510 err = nf_ct_netns_get(ctx->net, ctx->family);
514 if (priv->key == NFT_CT_BYTES ||
515 priv->key == NFT_CT_PKTS ||
516 priv->key == NFT_CT_AVGPKT)
517 nf_ct_set_acct(ctx->net, true);
522 static void __nft_ct_set_destroy(const struct nft_ctx *ctx, struct nft_ct *priv)
525 #ifdef CONFIG_NF_CONNTRACK_LABELS
527 nf_connlabels_put(ctx->net);
530 #ifdef CONFIG_NF_CONNTRACK_ZONES
532 mutex_lock(&nft_ct_pcpu_mutex);
533 if (--nft_ct_pcpu_template_refcnt == 0)
534 nft_ct_tmpl_put_pcpu();
535 mutex_unlock(&nft_ct_pcpu_mutex);
543 static int nft_ct_set_init(const struct nft_ctx *ctx,
544 const struct nft_expr *expr,
545 const struct nlattr * const tb[])
547 struct nft_ct *priv = nft_expr_priv(expr);
551 priv->dir = IP_CT_DIR_MAX;
552 priv->key = ntohl(nla_get_be32(tb[NFTA_CT_KEY]));
554 #ifdef CONFIG_NF_CONNTRACK_MARK
556 if (tb[NFTA_CT_DIRECTION])
558 len = sizeof_field(struct nf_conn, mark);
561 #ifdef CONFIG_NF_CONNTRACK_LABELS
563 if (tb[NFTA_CT_DIRECTION])
565 len = NF_CT_LABELS_MAX_SIZE;
566 err = nf_connlabels_get(ctx->net, (len * BITS_PER_BYTE) - 1);
571 #ifdef CONFIG_NF_CONNTRACK_ZONES
573 mutex_lock(&nft_ct_pcpu_mutex);
574 if (!nft_ct_tmpl_alloc_pcpu()) {
575 mutex_unlock(&nft_ct_pcpu_mutex);
578 nft_ct_pcpu_template_refcnt++;
579 mutex_unlock(&nft_ct_pcpu_mutex);
583 #ifdef CONFIG_NF_CONNTRACK_EVENTS
584 case NFT_CT_EVENTMASK:
585 if (tb[NFTA_CT_DIRECTION])
590 #ifdef CONFIG_NF_CONNTRACK_SECMARK
592 if (tb[NFTA_CT_DIRECTION])
601 if (tb[NFTA_CT_DIRECTION]) {
602 priv->dir = nla_get_u8(tb[NFTA_CT_DIRECTION]);
604 case IP_CT_DIR_ORIGINAL:
605 case IP_CT_DIR_REPLY:
614 err = nft_parse_register_load(tb[NFTA_CT_SREG], &priv->sreg, len);
618 err = nf_ct_netns_get(ctx->net, ctx->family);
625 __nft_ct_set_destroy(ctx, priv);
629 static void nft_ct_get_destroy(const struct nft_ctx *ctx,
630 const struct nft_expr *expr)
632 nf_ct_netns_put(ctx->net, ctx->family);
635 static void nft_ct_set_destroy(const struct nft_ctx *ctx,
636 const struct nft_expr *expr)
638 struct nft_ct *priv = nft_expr_priv(expr);
640 __nft_ct_set_destroy(ctx, priv);
641 nf_ct_netns_put(ctx->net, ctx->family);
644 static int nft_ct_get_dump(struct sk_buff *skb, const struct nft_expr *expr)
646 const struct nft_ct *priv = nft_expr_priv(expr);
648 if (nft_dump_register(skb, NFTA_CT_DREG, priv->dreg))
649 goto nla_put_failure;
650 if (nla_put_be32(skb, NFTA_CT_KEY, htonl(priv->key)))
651 goto nla_put_failure;
660 case NFT_CT_PROTO_SRC:
661 case NFT_CT_PROTO_DST:
662 if (nla_put_u8(skb, NFTA_CT_DIRECTION, priv->dir))
663 goto nla_put_failure;
669 if (priv->dir < IP_CT_DIR_MAX &&
670 nla_put_u8(skb, NFTA_CT_DIRECTION, priv->dir))
671 goto nla_put_failure;
683 static bool nft_ct_get_reduce(struct nft_regs_track *track,
684 const struct nft_expr *expr)
686 const struct nft_ct *priv = nft_expr_priv(expr);
687 const struct nft_ct *ct;
689 if (!nft_reg_track_cmp(track, expr, priv->dreg)) {
690 nft_reg_track_update(track, expr, priv->dreg, priv->len);
694 ct = nft_expr_priv(track->regs[priv->dreg].selector);
695 if (priv->key != ct->key) {
696 nft_reg_track_update(track, expr, priv->dreg, priv->len);
700 if (!track->regs[priv->dreg].bitwise)
703 return nft_expr_reduce_bitwise(track, expr);
706 static int nft_ct_set_dump(struct sk_buff *skb, const struct nft_expr *expr)
708 const struct nft_ct *priv = nft_expr_priv(expr);
710 if (nft_dump_register(skb, NFTA_CT_SREG, priv->sreg))
711 goto nla_put_failure;
712 if (nla_put_be32(skb, NFTA_CT_KEY, htonl(priv->key)))
713 goto nla_put_failure;
717 if (priv->dir < IP_CT_DIR_MAX &&
718 nla_put_u8(skb, NFTA_CT_DIRECTION, priv->dir))
719 goto nla_put_failure;
731 static struct nft_expr_type nft_ct_type;
732 static const struct nft_expr_ops nft_ct_get_ops = {
733 .type = &nft_ct_type,
734 .size = NFT_EXPR_SIZE(sizeof(struct nft_ct)),
735 .eval = nft_ct_get_eval,
736 .init = nft_ct_get_init,
737 .destroy = nft_ct_get_destroy,
738 .dump = nft_ct_get_dump,
739 .reduce = nft_ct_get_reduce,
742 static bool nft_ct_set_reduce(struct nft_regs_track *track,
743 const struct nft_expr *expr)
747 for (i = 0; i < NFT_REG32_NUM; i++) {
748 if (!track->regs[i].selector)
751 if (track->regs[i].selector->ops != &nft_ct_get_ops)
754 __nft_reg_track_cancel(track, i);
760 static const struct nft_expr_ops nft_ct_set_ops = {
761 .type = &nft_ct_type,
762 .size = NFT_EXPR_SIZE(sizeof(struct nft_ct)),
763 .eval = nft_ct_set_eval,
764 .init = nft_ct_set_init,
765 .destroy = nft_ct_set_destroy,
766 .dump = nft_ct_set_dump,
767 .reduce = nft_ct_set_reduce,
770 #ifdef CONFIG_NF_CONNTRACK_ZONES
771 static const struct nft_expr_ops nft_ct_set_zone_ops = {
772 .type = &nft_ct_type,
773 .size = NFT_EXPR_SIZE(sizeof(struct nft_ct)),
774 .eval = nft_ct_set_zone_eval,
775 .init = nft_ct_set_init,
776 .destroy = nft_ct_set_destroy,
777 .dump = nft_ct_set_dump,
778 .reduce = nft_ct_set_reduce,
782 static const struct nft_expr_ops *
783 nft_ct_select_ops(const struct nft_ctx *ctx,
784 const struct nlattr * const tb[])
786 if (tb[NFTA_CT_KEY] == NULL)
787 return ERR_PTR(-EINVAL);
789 if (tb[NFTA_CT_DREG] && tb[NFTA_CT_SREG])
790 return ERR_PTR(-EINVAL);
792 if (tb[NFTA_CT_DREG])
793 return &nft_ct_get_ops;
795 if (tb[NFTA_CT_SREG]) {
796 #ifdef CONFIG_NF_CONNTRACK_ZONES
797 if (nla_get_be32(tb[NFTA_CT_KEY]) == htonl(NFT_CT_ZONE))
798 return &nft_ct_set_zone_ops;
800 return &nft_ct_set_ops;
803 return ERR_PTR(-EINVAL);
806 static struct nft_expr_type nft_ct_type __read_mostly = {
808 .select_ops = nft_ct_select_ops,
809 .policy = nft_ct_policy,
810 .maxattr = NFTA_CT_MAX,
811 .owner = THIS_MODULE,
814 static void nft_notrack_eval(const struct nft_expr *expr,
815 struct nft_regs *regs,
816 const struct nft_pktinfo *pkt)
818 struct sk_buff *skb = pkt->skb;
819 enum ip_conntrack_info ctinfo;
822 ct = nf_ct_get(pkt->skb, &ctinfo);
823 /* Previously seen (loopback or untracked)? Ignore. */
824 if (ct || ctinfo == IP_CT_UNTRACKED)
827 nf_ct_set(skb, ct, IP_CT_UNTRACKED);
830 static struct nft_expr_type nft_notrack_type;
831 static const struct nft_expr_ops nft_notrack_ops = {
832 .type = &nft_notrack_type,
833 .size = NFT_EXPR_SIZE(0),
834 .eval = nft_notrack_eval,
835 .reduce = NFT_REDUCE_READONLY,
838 static struct nft_expr_type nft_notrack_type __read_mostly = {
840 .ops = &nft_notrack_ops,
841 .owner = THIS_MODULE,
844 #ifdef CONFIG_NF_CONNTRACK_TIMEOUT
846 nft_ct_timeout_parse_policy(void *timeouts,
847 const struct nf_conntrack_l4proto *l4proto,
848 struct net *net, const struct nlattr *attr)
853 tb = kcalloc(l4proto->ctnl_timeout.nlattr_max + 1, sizeof(*tb),
859 ret = nla_parse_nested_deprecated(tb,
860 l4proto->ctnl_timeout.nlattr_max,
862 l4proto->ctnl_timeout.nla_policy,
867 ret = l4proto->ctnl_timeout.nlattr_to_obj(tb, net, timeouts);
874 struct nft_ct_timeout_obj {
875 struct nf_ct_timeout *timeout;
879 static void nft_ct_timeout_obj_eval(struct nft_object *obj,
880 struct nft_regs *regs,
881 const struct nft_pktinfo *pkt)
883 const struct nft_ct_timeout_obj *priv = nft_obj_data(obj);
884 struct nf_conn *ct = (struct nf_conn *)skb_nfct(pkt->skb);
885 struct nf_conn_timeout *timeout;
886 const unsigned int *values;
888 if (priv->l4proto != pkt->tprot)
891 if (!ct || nf_ct_is_template(ct) || nf_ct_is_confirmed(ct))
894 timeout = nf_ct_timeout_find(ct);
896 timeout = nf_ct_timeout_ext_add(ct, priv->timeout, GFP_ATOMIC);
898 regs->verdict.code = NF_DROP;
903 rcu_assign_pointer(timeout->timeout, priv->timeout);
905 /* adjust the timeout as per 'new' state. ct is unconfirmed,
906 * so the current timestamp must not be added.
908 values = nf_ct_timeout_data(timeout);
910 nf_ct_refresh(ct, pkt->skb, values[0]);
913 static int nft_ct_timeout_obj_init(const struct nft_ctx *ctx,
914 const struct nlattr * const tb[],
915 struct nft_object *obj)
917 struct nft_ct_timeout_obj *priv = nft_obj_data(obj);
918 const struct nf_conntrack_l4proto *l4proto;
919 struct nf_ct_timeout *timeout;
920 int l3num = ctx->family;
924 if (!tb[NFTA_CT_TIMEOUT_L4PROTO] ||
925 !tb[NFTA_CT_TIMEOUT_DATA])
928 if (tb[NFTA_CT_TIMEOUT_L3PROTO])
929 l3num = ntohs(nla_get_be16(tb[NFTA_CT_TIMEOUT_L3PROTO]));
931 l4num = nla_get_u8(tb[NFTA_CT_TIMEOUT_L4PROTO]);
932 priv->l4proto = l4num;
934 l4proto = nf_ct_l4proto_find(l4num);
936 if (l4proto->l4proto != l4num) {
941 timeout = kzalloc(sizeof(struct nf_ct_timeout) +
942 l4proto->ctnl_timeout.obj_size, GFP_KERNEL);
943 if (timeout == NULL) {
948 ret = nft_ct_timeout_parse_policy(&timeout->data, l4proto, ctx->net,
949 tb[NFTA_CT_TIMEOUT_DATA]);
951 goto err_free_timeout;
953 timeout->l3num = l3num;
954 timeout->l4proto = l4proto;
956 ret = nf_ct_netns_get(ctx->net, ctx->family);
958 goto err_free_timeout;
960 priv->timeout = timeout;
969 static void nft_ct_timeout_obj_destroy(const struct nft_ctx *ctx,
970 struct nft_object *obj)
972 struct nft_ct_timeout_obj *priv = nft_obj_data(obj);
973 struct nf_ct_timeout *timeout = priv->timeout;
975 nf_ct_untimeout(ctx->net, timeout);
976 nf_ct_netns_put(ctx->net, ctx->family);
977 kfree(priv->timeout);
980 static int nft_ct_timeout_obj_dump(struct sk_buff *skb,
981 struct nft_object *obj, bool reset)
983 const struct nft_ct_timeout_obj *priv = nft_obj_data(obj);
984 const struct nf_ct_timeout *timeout = priv->timeout;
985 struct nlattr *nest_params;
988 if (nla_put_u8(skb, NFTA_CT_TIMEOUT_L4PROTO, timeout->l4proto->l4proto) ||
989 nla_put_be16(skb, NFTA_CT_TIMEOUT_L3PROTO, htons(timeout->l3num)))
992 nest_params = nla_nest_start(skb, NFTA_CT_TIMEOUT_DATA);
996 ret = timeout->l4proto->ctnl_timeout.obj_to_nlattr(skb, &timeout->data);
999 nla_nest_end(skb, nest_params);
1003 static const struct nla_policy nft_ct_timeout_policy[NFTA_CT_TIMEOUT_MAX + 1] = {
1004 [NFTA_CT_TIMEOUT_L3PROTO] = {.type = NLA_U16 },
1005 [NFTA_CT_TIMEOUT_L4PROTO] = {.type = NLA_U8 },
1006 [NFTA_CT_TIMEOUT_DATA] = {.type = NLA_NESTED },
1009 static struct nft_object_type nft_ct_timeout_obj_type;
1011 static const struct nft_object_ops nft_ct_timeout_obj_ops = {
1012 .type = &nft_ct_timeout_obj_type,
1013 .size = sizeof(struct nft_ct_timeout_obj),
1014 .eval = nft_ct_timeout_obj_eval,
1015 .init = nft_ct_timeout_obj_init,
1016 .destroy = nft_ct_timeout_obj_destroy,
1017 .dump = nft_ct_timeout_obj_dump,
1020 static struct nft_object_type nft_ct_timeout_obj_type __read_mostly = {
1021 .type = NFT_OBJECT_CT_TIMEOUT,
1022 .ops = &nft_ct_timeout_obj_ops,
1023 .maxattr = NFTA_CT_TIMEOUT_MAX,
1024 .policy = nft_ct_timeout_policy,
1025 .owner = THIS_MODULE,
1027 #endif /* CONFIG_NF_CONNTRACK_TIMEOUT */
1029 static int nft_ct_helper_obj_init(const struct nft_ctx *ctx,
1030 const struct nlattr * const tb[],
1031 struct nft_object *obj)
1033 struct nft_ct_helper_obj *priv = nft_obj_data(obj);
1034 struct nf_conntrack_helper *help4, *help6;
1035 char name[NF_CT_HELPER_NAME_LEN];
1036 int family = ctx->family;
1039 if (!tb[NFTA_CT_HELPER_NAME] || !tb[NFTA_CT_HELPER_L4PROTO])
1042 priv->l4proto = nla_get_u8(tb[NFTA_CT_HELPER_L4PROTO]);
1046 nla_strscpy(name, tb[NFTA_CT_HELPER_NAME], sizeof(name));
1048 if (tb[NFTA_CT_HELPER_L3PROTO])
1049 family = ntohs(nla_get_be16(tb[NFTA_CT_HELPER_L3PROTO]));
1056 if (ctx->family == NFPROTO_IPV6)
1059 help4 = nf_conntrack_helper_try_module_get(name, family,
1063 if (ctx->family == NFPROTO_IPV4)
1066 help6 = nf_conntrack_helper_try_module_get(name, family,
1069 case NFPROTO_NETDEV:
1070 case NFPROTO_BRIDGE:
1072 help4 = nf_conntrack_helper_try_module_get(name, NFPROTO_IPV4,
1074 help6 = nf_conntrack_helper_try_module_get(name, NFPROTO_IPV6,
1078 return -EAFNOSUPPORT;
1081 /* && is intentional; only error if INET found neither ipv4 or ipv6 */
1082 if (!help4 && !help6)
1085 priv->helper4 = help4;
1086 priv->helper6 = help6;
1088 err = nf_ct_netns_get(ctx->net, ctx->family);
1090 goto err_put_helper;
1096 nf_conntrack_helper_put(priv->helper4);
1098 nf_conntrack_helper_put(priv->helper6);
1102 static void nft_ct_helper_obj_destroy(const struct nft_ctx *ctx,
1103 struct nft_object *obj)
1105 struct nft_ct_helper_obj *priv = nft_obj_data(obj);
1108 nf_conntrack_helper_put(priv->helper4);
1110 nf_conntrack_helper_put(priv->helper6);
1112 nf_ct_netns_put(ctx->net, ctx->family);
1115 static void nft_ct_helper_obj_eval(struct nft_object *obj,
1116 struct nft_regs *regs,
1117 const struct nft_pktinfo *pkt)
1119 const struct nft_ct_helper_obj *priv = nft_obj_data(obj);
1120 struct nf_conn *ct = (struct nf_conn *)skb_nfct(pkt->skb);
1121 struct nf_conntrack_helper *to_assign = NULL;
1122 struct nf_conn_help *help;
1125 nf_ct_is_confirmed(ct) ||
1126 nf_ct_is_template(ct) ||
1127 priv->l4proto != nf_ct_protonum(ct))
1130 switch (nf_ct_l3num(ct)) {
1132 to_assign = priv->helper4;
1135 to_assign = priv->helper6;
1145 if (test_bit(IPS_HELPER_BIT, &ct->status))
1148 help = nf_ct_helper_ext_add(ct, GFP_ATOMIC);
1150 rcu_assign_pointer(help->helper, to_assign);
1151 set_bit(IPS_HELPER_BIT, &ct->status);
1155 static int nft_ct_helper_obj_dump(struct sk_buff *skb,
1156 struct nft_object *obj, bool reset)
1158 const struct nft_ct_helper_obj *priv = nft_obj_data(obj);
1159 const struct nf_conntrack_helper *helper;
1162 if (priv->helper4 && priv->helper6) {
1163 family = NFPROTO_INET;
1164 helper = priv->helper4;
1165 } else if (priv->helper6) {
1166 family = NFPROTO_IPV6;
1167 helper = priv->helper6;
1169 family = NFPROTO_IPV4;
1170 helper = priv->helper4;
1173 if (nla_put_string(skb, NFTA_CT_HELPER_NAME, helper->name))
1176 if (nla_put_u8(skb, NFTA_CT_HELPER_L4PROTO, priv->l4proto))
1179 if (nla_put_be16(skb, NFTA_CT_HELPER_L3PROTO, htons(family)))
1185 static const struct nla_policy nft_ct_helper_policy[NFTA_CT_HELPER_MAX + 1] = {
1186 [NFTA_CT_HELPER_NAME] = { .type = NLA_STRING,
1187 .len = NF_CT_HELPER_NAME_LEN - 1 },
1188 [NFTA_CT_HELPER_L3PROTO] = { .type = NLA_U16 },
1189 [NFTA_CT_HELPER_L4PROTO] = { .type = NLA_U8 },
1192 static struct nft_object_type nft_ct_helper_obj_type;
1193 static const struct nft_object_ops nft_ct_helper_obj_ops = {
1194 .type = &nft_ct_helper_obj_type,
1195 .size = sizeof(struct nft_ct_helper_obj),
1196 .eval = nft_ct_helper_obj_eval,
1197 .init = nft_ct_helper_obj_init,
1198 .destroy = nft_ct_helper_obj_destroy,
1199 .dump = nft_ct_helper_obj_dump,
1202 static struct nft_object_type nft_ct_helper_obj_type __read_mostly = {
1203 .type = NFT_OBJECT_CT_HELPER,
1204 .ops = &nft_ct_helper_obj_ops,
1205 .maxattr = NFTA_CT_HELPER_MAX,
1206 .policy = nft_ct_helper_policy,
1207 .owner = THIS_MODULE,
1210 struct nft_ct_expect_obj {
1218 static int nft_ct_expect_obj_init(const struct nft_ctx *ctx,
1219 const struct nlattr * const tb[],
1220 struct nft_object *obj)
1222 struct nft_ct_expect_obj *priv = nft_obj_data(obj);
1224 if (!tb[NFTA_CT_EXPECT_L4PROTO] ||
1225 !tb[NFTA_CT_EXPECT_DPORT] ||
1226 !tb[NFTA_CT_EXPECT_TIMEOUT] ||
1227 !tb[NFTA_CT_EXPECT_SIZE])
1230 priv->l3num = ctx->family;
1231 if (tb[NFTA_CT_EXPECT_L3PROTO])
1232 priv->l3num = ntohs(nla_get_be16(tb[NFTA_CT_EXPECT_L3PROTO]));
1234 priv->l4proto = nla_get_u8(tb[NFTA_CT_EXPECT_L4PROTO]);
1235 priv->dport = nla_get_be16(tb[NFTA_CT_EXPECT_DPORT]);
1236 priv->timeout = nla_get_u32(tb[NFTA_CT_EXPECT_TIMEOUT]);
1237 priv->size = nla_get_u8(tb[NFTA_CT_EXPECT_SIZE]);
1239 return nf_ct_netns_get(ctx->net, ctx->family);
1242 static void nft_ct_expect_obj_destroy(const struct nft_ctx *ctx,
1243 struct nft_object *obj)
1245 nf_ct_netns_put(ctx->net, ctx->family);
1248 static int nft_ct_expect_obj_dump(struct sk_buff *skb,
1249 struct nft_object *obj, bool reset)
1251 const struct nft_ct_expect_obj *priv = nft_obj_data(obj);
1253 if (nla_put_be16(skb, NFTA_CT_EXPECT_L3PROTO, htons(priv->l3num)) ||
1254 nla_put_u8(skb, NFTA_CT_EXPECT_L4PROTO, priv->l4proto) ||
1255 nla_put_be16(skb, NFTA_CT_EXPECT_DPORT, priv->dport) ||
1256 nla_put_u32(skb, NFTA_CT_EXPECT_TIMEOUT, priv->timeout) ||
1257 nla_put_u8(skb, NFTA_CT_EXPECT_SIZE, priv->size))
1263 static void nft_ct_expect_obj_eval(struct nft_object *obj,
1264 struct nft_regs *regs,
1265 const struct nft_pktinfo *pkt)
1267 const struct nft_ct_expect_obj *priv = nft_obj_data(obj);
1268 struct nf_conntrack_expect *exp;
1269 enum ip_conntrack_info ctinfo;
1270 struct nf_conn_help *help;
1271 enum ip_conntrack_dir dir;
1272 u16 l3num = priv->l3num;
1275 ct = nf_ct_get(pkt->skb, &ctinfo);
1276 if (!ct || nf_ct_is_confirmed(ct) || nf_ct_is_template(ct)) {
1277 regs->verdict.code = NFT_BREAK;
1280 dir = CTINFO2DIR(ctinfo);
1282 help = nfct_help(ct);
1284 help = nf_ct_helper_ext_add(ct, GFP_ATOMIC);
1286 regs->verdict.code = NF_DROP;
1290 if (help->expecting[NF_CT_EXPECT_CLASS_DEFAULT] >= priv->size) {
1291 regs->verdict.code = NFT_BREAK;
1294 if (l3num == NFPROTO_INET)
1295 l3num = nf_ct_l3num(ct);
1297 exp = nf_ct_expect_alloc(ct);
1299 regs->verdict.code = NF_DROP;
1302 nf_ct_expect_init(exp, NF_CT_EXPECT_CLASS_DEFAULT, l3num,
1303 &ct->tuplehash[!dir].tuple.src.u3,
1304 &ct->tuplehash[!dir].tuple.dst.u3,
1305 priv->l4proto, NULL, &priv->dport);
1306 exp->timeout.expires = jiffies + priv->timeout * HZ;
1308 if (nf_ct_expect_related(exp, 0) != 0)
1309 regs->verdict.code = NF_DROP;
1312 static const struct nla_policy nft_ct_expect_policy[NFTA_CT_EXPECT_MAX + 1] = {
1313 [NFTA_CT_EXPECT_L3PROTO] = { .type = NLA_U16 },
1314 [NFTA_CT_EXPECT_L4PROTO] = { .type = NLA_U8 },
1315 [NFTA_CT_EXPECT_DPORT] = { .type = NLA_U16 },
1316 [NFTA_CT_EXPECT_TIMEOUT] = { .type = NLA_U32 },
1317 [NFTA_CT_EXPECT_SIZE] = { .type = NLA_U8 },
1320 static struct nft_object_type nft_ct_expect_obj_type;
1322 static const struct nft_object_ops nft_ct_expect_obj_ops = {
1323 .type = &nft_ct_expect_obj_type,
1324 .size = sizeof(struct nft_ct_expect_obj),
1325 .eval = nft_ct_expect_obj_eval,
1326 .init = nft_ct_expect_obj_init,
1327 .destroy = nft_ct_expect_obj_destroy,
1328 .dump = nft_ct_expect_obj_dump,
1331 static struct nft_object_type nft_ct_expect_obj_type __read_mostly = {
1332 .type = NFT_OBJECT_CT_EXPECT,
1333 .ops = &nft_ct_expect_obj_ops,
1334 .maxattr = NFTA_CT_EXPECT_MAX,
1335 .policy = nft_ct_expect_policy,
1336 .owner = THIS_MODULE,
1339 static int __init nft_ct_module_init(void)
1343 BUILD_BUG_ON(NF_CT_LABELS_MAX_SIZE > NFT_REG_SIZE);
1345 err = nft_register_expr(&nft_ct_type);
1349 err = nft_register_expr(&nft_notrack_type);
1353 err = nft_register_obj(&nft_ct_helper_obj_type);
1357 err = nft_register_obj(&nft_ct_expect_obj_type);
1360 #ifdef CONFIG_NF_CONNTRACK_TIMEOUT
1361 err = nft_register_obj(&nft_ct_timeout_obj_type);
1367 #ifdef CONFIG_NF_CONNTRACK_TIMEOUT
1369 nft_unregister_obj(&nft_ct_expect_obj_type);
1372 nft_unregister_obj(&nft_ct_helper_obj_type);
1374 nft_unregister_expr(&nft_notrack_type);
1376 nft_unregister_expr(&nft_ct_type);
1380 static void __exit nft_ct_module_exit(void)
1382 #ifdef CONFIG_NF_CONNTRACK_TIMEOUT
1383 nft_unregister_obj(&nft_ct_timeout_obj_type);
1385 nft_unregister_obj(&nft_ct_expect_obj_type);
1386 nft_unregister_obj(&nft_ct_helper_obj_type);
1387 nft_unregister_expr(&nft_notrack_type);
1388 nft_unregister_expr(&nft_ct_type);
1391 module_init(nft_ct_module_init);
1392 module_exit(nft_ct_module_exit);
1394 MODULE_LICENSE("GPL");
1395 MODULE_AUTHOR("Patrick McHardy <kaber@trash.net>");
1396 MODULE_ALIAS_NFT_EXPR("ct");
1397 MODULE_ALIAS_NFT_EXPR("notrack");
1398 MODULE_ALIAS_NFT_OBJ(NFT_OBJECT_CT_HELPER);
1399 MODULE_ALIAS_NFT_OBJ(NFT_OBJECT_CT_TIMEOUT);
1400 MODULE_ALIAS_NFT_OBJ(NFT_OBJECT_CT_EXPECT);
1401 MODULE_DESCRIPTION("Netfilter nf_tables conntrack module");