netfilter: hashlimit: do not use indirect calls during gc
authorFlorian Westphal <fw@strlen.de>
Tue, 7 Jan 2020 11:25:10 +0000 (12:25 +0100)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Sat, 21 Mar 2020 07:11:46 +0000 (08:11 +0100)
[ Upstream commit 28b3a4270c0fc064557e409111f2a678e64b6fa7 ]

no need, just use a simple boolean to indicate we want to reap all
entries.

Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
Signed-off-by: Sasha Levin <sashal@kernel.org>
net/netfilter/xt_hashlimit.c

index 1b68a13..7a2c4b8 100644 (file)
@@ -358,21 +358,7 @@ static int htable_create(struct net *net, struct hashlimit_cfg3 *cfg,
        return 0;
 }
 
-static bool select_all(const struct xt_hashlimit_htable *ht,
-                      const struct dsthash_ent *he)
-{
-       return true;
-}
-
-static bool select_gc(const struct xt_hashlimit_htable *ht,
-                     const struct dsthash_ent *he)
-{
-       return time_after_eq(jiffies, he->expires);
-}
-
-static void htable_selective_cleanup(struct xt_hashlimit_htable *ht,
-                       bool (*select)(const struct xt_hashlimit_htable *ht,
-                                     const struct dsthash_ent *he))
+static void htable_selective_cleanup(struct xt_hashlimit_htable *ht, bool select_all)
 {
        unsigned int i;
 
@@ -382,7 +368,7 @@ static void htable_selective_cleanup(struct xt_hashlimit_htable *ht,
 
                spin_lock_bh(&ht->lock);
                hlist_for_each_entry_safe(dh, n, &ht->hash[i], node) {
-                       if ((*select)(ht, dh))
+                       if (time_after_eq(jiffies, dh->expires) || select_all)
                                dsthash_free(ht, dh);
                }
                spin_unlock_bh(&ht->lock);
@@ -396,7 +382,7 @@ static void htable_gc(struct work_struct *work)
 
        ht = container_of(work, struct xt_hashlimit_htable, gc_work.work);
 
-       htable_selective_cleanup(ht, select_gc);
+       htable_selective_cleanup(ht, false);
 
        queue_delayed_work(system_power_efficient_wq,
                           &ht->gc_work, msecs_to_jiffies(ht->cfg.gc_interval));
@@ -420,7 +406,7 @@ static void htable_destroy(struct xt_hashlimit_htable *hinfo)
 {
        cancel_delayed_work_sync(&hinfo->gc_work);
        htable_remove_proc_entry(hinfo);
-       htable_selective_cleanup(hinfo, select_all);
+       htable_selective_cleanup(hinfo, true);
        kfree(hinfo->name);
        vfree(hinfo);
 }