|
@@ -285,16 +285,16 @@ int __nf_ct_try_assign_helper(struct nf_conn *ct, struct nf_conn *tmpl,
|
|
|
EXPORT_SYMBOL_GPL(__nf_ct_try_assign_helper);
|
|
|
|
|
|
/* appropriate ct lock protecting must be taken by caller */
|
|
|
-static inline int unhelp(struct nf_conntrack_tuple_hash *i,
|
|
|
- const struct nf_conntrack_helper *me)
|
|
|
+static int unhelp(struct nf_conn *ct, void *me)
|
|
|
{
|
|
|
- struct nf_conn *ct = nf_ct_tuplehash_to_ctrack(i);
|
|
|
struct nf_conn_help *help = nfct_help(ct);
|
|
|
|
|
|
if (help && rcu_dereference_raw(help->helper) == me) {
|
|
|
nf_conntrack_event(IPCT_HELPER, ct);
|
|
|
RCU_INIT_POINTER(help->helper, NULL);
|
|
|
}
|
|
|
+
|
|
|
+ /* We are not intended to delete this conntrack. */
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
@@ -437,33 +437,10 @@ out:
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(nf_conntrack_helper_register);
|
|
|
|
|
|
-static void __nf_conntrack_helper_unregister(struct nf_conntrack_helper *me,
|
|
|
- struct net *net)
|
|
|
-{
|
|
|
- struct nf_conntrack_tuple_hash *h;
|
|
|
- const struct hlist_nulls_node *nn;
|
|
|
- int cpu;
|
|
|
-
|
|
|
- /* Get rid of expecteds, set helpers to NULL. */
|
|
|
- for_each_possible_cpu(cpu) {
|
|
|
- struct ct_pcpu *pcpu = per_cpu_ptr(net->ct.pcpu_lists, cpu);
|
|
|
-
|
|
|
- spin_lock_bh(&pcpu->lock);
|
|
|
- hlist_nulls_for_each_entry(h, nn, &pcpu->unconfirmed, hnnode)
|
|
|
- unhelp(h, me);
|
|
|
- spin_unlock_bh(&pcpu->lock);
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
void nf_conntrack_helper_unregister(struct nf_conntrack_helper *me)
|
|
|
{
|
|
|
- struct nf_conntrack_tuple_hash *h;
|
|
|
struct nf_conntrack_expect *exp;
|
|
|
const struct hlist_node *next;
|
|
|
- const struct hlist_nulls_node *nn;
|
|
|
- unsigned int last_hsize;
|
|
|
- spinlock_t *lock;
|
|
|
- struct net *net;
|
|
|
unsigned int i;
|
|
|
|
|
|
mutex_lock(&nf_ct_helper_mutex);
|
|
@@ -491,26 +468,7 @@ void nf_conntrack_helper_unregister(struct nf_conntrack_helper *me)
|
|
|
}
|
|
|
spin_unlock_bh(&nf_conntrack_expect_lock);
|
|
|
|
|
|
- rtnl_lock();
|
|
|
- for_each_net(net)
|
|
|
- __nf_conntrack_helper_unregister(me, net);
|
|
|
- rtnl_unlock();
|
|
|
-
|
|
|
- local_bh_disable();
|
|
|
-restart:
|
|
|
- last_hsize = nf_conntrack_htable_size;
|
|
|
- for (i = 0; i < last_hsize; i++) {
|
|
|
- lock = &nf_conntrack_locks[i % CONNTRACK_LOCKS];
|
|
|
- nf_conntrack_lock(lock);
|
|
|
- if (last_hsize != nf_conntrack_htable_size) {
|
|
|
- spin_unlock(lock);
|
|
|
- goto restart;
|
|
|
- }
|
|
|
- hlist_nulls_for_each_entry(h, nn, &nf_conntrack_hash[i], hnnode)
|
|
|
- unhelp(h, me);
|
|
|
- spin_unlock(lock);
|
|
|
- }
|
|
|
- local_bh_enable();
|
|
|
+ nf_ct_iterate_destroy(unhelp, me);
|
|
|
}
|
|
|
EXPORT_SYMBOL_GPL(nf_conntrack_helper_unregister);
|
|
|
|