/home/lenb/linux-2.6 branch 'acpi-2.6.12'
[sfrench/cifs-2.6.git] / net / ipv4 / netfilter / ip_conntrack_core.c
1 /* Connection state tracking for netfilter.  This is separated from,
2    but required by, the NAT layer; it can also be used by an iptables
3    extension. */
4
5 /* (C) 1999-2001 Paul `Rusty' Russell  
6  * (C) 2002-2004 Netfilter Core Team <coreteam@netfilter.org>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  *
12  * 23 Apr 2001: Harald Welte <laforge@gnumonks.org>
13  *      - new API and handling of conntrack/nat helpers
14  *      - now capable of multiple expectations for one master
15  * 16 Jul 2002: Harald Welte <laforge@gnumonks.org>
16  *      - add usage/reference counts to ip_conntrack_expect
17  *      - export ip_conntrack[_expect]_{find_get,put} functions
18  * */
19
20 #include <linux/config.h>
21 #include <linux/types.h>
22 #include <linux/icmp.h>
23 #include <linux/ip.h>
24 #include <linux/netfilter.h>
25 #include <linux/netfilter_ipv4.h>
26 #include <linux/module.h>
27 #include <linux/skbuff.h>
28 #include <linux/proc_fs.h>
29 #include <linux/vmalloc.h>
30 #include <net/checksum.h>
31 #include <net/ip.h>
32 #include <linux/stddef.h>
33 #include <linux/sysctl.h>
34 #include <linux/slab.h>
35 #include <linux/random.h>
36 #include <linux/jhash.h>
37 #include <linux/err.h>
38 #include <linux/percpu.h>
39 #include <linux/moduleparam.h>
40
41 /* ip_conntrack_lock protects the main hash table, protocol/helper/expected
42    registrations, conntrack timers*/
43 #define ASSERT_READ_LOCK(x)
44 #define ASSERT_WRITE_LOCK(x)
45
46 #include <linux/netfilter_ipv4/ip_conntrack.h>
47 #include <linux/netfilter_ipv4/ip_conntrack_protocol.h>
48 #include <linux/netfilter_ipv4/ip_conntrack_helper.h>
49 #include <linux/netfilter_ipv4/ip_conntrack_core.h>
50 #include <linux/netfilter_ipv4/listhelp.h>
51
52 #define IP_CONNTRACK_VERSION    "2.1"
53
54 #if 0
55 #define DEBUGP printk
56 #else
57 #define DEBUGP(format, args...)
58 #endif
59
60 DEFINE_RWLOCK(ip_conntrack_lock);
61
62 /* ip_conntrack_standalone needs this */
63 atomic_t ip_conntrack_count = ATOMIC_INIT(0);
64
65 void (*ip_conntrack_destroyed)(struct ip_conntrack *conntrack) = NULL;
66 LIST_HEAD(ip_conntrack_expect_list);
67 struct ip_conntrack_protocol *ip_ct_protos[MAX_IP_CT_PROTO];
68 static LIST_HEAD(helpers);
69 unsigned int ip_conntrack_htable_size = 0;
70 int ip_conntrack_max;
71 struct list_head *ip_conntrack_hash;
72 static kmem_cache_t *ip_conntrack_cachep;
73 static kmem_cache_t *ip_conntrack_expect_cachep;
74 struct ip_conntrack ip_conntrack_untracked;
75 unsigned int ip_ct_log_invalid;
76 static LIST_HEAD(unconfirmed);
77 static int ip_conntrack_vmalloc;
78
79 DEFINE_PER_CPU(struct ip_conntrack_stat, ip_conntrack_stat);
80
81 void 
82 ip_conntrack_put(struct ip_conntrack *ct)
83 {
84         IP_NF_ASSERT(ct);
85         nf_conntrack_put(&ct->ct_general);
86 }
87
88 static int ip_conntrack_hash_rnd_initted;
89 static unsigned int ip_conntrack_hash_rnd;
90
91 static u_int32_t
92 hash_conntrack(const struct ip_conntrack_tuple *tuple)
93 {
94 #if 0
95         dump_tuple(tuple);
96 #endif
97         return (jhash_3words(tuple->src.ip,
98                              (tuple->dst.ip ^ tuple->dst.protonum),
99                              (tuple->src.u.all | (tuple->dst.u.all << 16)),
100                              ip_conntrack_hash_rnd) % ip_conntrack_htable_size);
101 }
102
103 int
104 ip_ct_get_tuple(const struct iphdr *iph,
105                 const struct sk_buff *skb,
106                 unsigned int dataoff,
107                 struct ip_conntrack_tuple *tuple,
108                 const struct ip_conntrack_protocol *protocol)
109 {
110         /* Never happen */
111         if (iph->frag_off & htons(IP_OFFSET)) {
112                 printk("ip_conntrack_core: Frag of proto %u.\n",
113                        iph->protocol);
114                 return 0;
115         }
116
117         tuple->src.ip = iph->saddr;
118         tuple->dst.ip = iph->daddr;
119         tuple->dst.protonum = iph->protocol;
120         tuple->dst.dir = IP_CT_DIR_ORIGINAL;
121
122         return protocol->pkt_to_tuple(skb, dataoff, tuple);
123 }
124
125 int
126 ip_ct_invert_tuple(struct ip_conntrack_tuple *inverse,
127                    const struct ip_conntrack_tuple *orig,
128                    const struct ip_conntrack_protocol *protocol)
129 {
130         inverse->src.ip = orig->dst.ip;
131         inverse->dst.ip = orig->src.ip;
132         inverse->dst.protonum = orig->dst.protonum;
133         inverse->dst.dir = !orig->dst.dir;
134
135         return protocol->invert_tuple(inverse, orig);
136 }
137
138
139 /* ip_conntrack_expect helper functions */
140 static void unlink_expect(struct ip_conntrack_expect *exp)
141 {
142         ASSERT_WRITE_LOCK(&ip_conntrack_lock);
143         IP_NF_ASSERT(!timer_pending(&exp->timeout));
144         list_del(&exp->list);
145         CONNTRACK_STAT_INC(expect_delete);
146         exp->master->expecting--;
147 }
148
149 static void expectation_timed_out(unsigned long ul_expect)
150 {
151         struct ip_conntrack_expect *exp = (void *)ul_expect;
152
153         write_lock_bh(&ip_conntrack_lock);
154         unlink_expect(exp);
155         write_unlock_bh(&ip_conntrack_lock);
156         ip_conntrack_expect_put(exp);
157 }
158
159 /* If an expectation for this connection is found, it gets delete from
160  * global list then returned. */
161 static struct ip_conntrack_expect *
162 find_expectation(const struct ip_conntrack_tuple *tuple)
163 {
164         struct ip_conntrack_expect *i;
165
166         list_for_each_entry(i, &ip_conntrack_expect_list, list) {
167                 /* If master is not in hash table yet (ie. packet hasn't left
168                    this machine yet), how can other end know about expected?
169                    Hence these are not the droids you are looking for (if
170                    master ct never got confirmed, we'd hold a reference to it
171                    and weird things would happen to future packets). */
172                 if (ip_ct_tuple_mask_cmp(tuple, &i->tuple, &i->mask)
173                     && is_confirmed(i->master)
174                     && del_timer(&i->timeout)) {
175                         unlink_expect(i);
176                         return i;
177                 }
178         }
179         return NULL;
180 }
181
182 /* delete all expectations for this conntrack */
183 static void remove_expectations(struct ip_conntrack *ct)
184 {
185         struct ip_conntrack_expect *i, *tmp;
186
187         /* Optimization: most connection never expect any others. */
188         if (ct->expecting == 0)
189                 return;
190
191         list_for_each_entry_safe(i, tmp, &ip_conntrack_expect_list, list) {
192                 if (i->master == ct && del_timer(&i->timeout)) {
193                         unlink_expect(i);
194                         ip_conntrack_expect_put(i);
195                 }
196         }
197 }
198
199 static void
200 clean_from_lists(struct ip_conntrack *ct)
201 {
202         unsigned int ho, hr;
203         
204         DEBUGP("clean_from_lists(%p)\n", ct);
205         ASSERT_WRITE_LOCK(&ip_conntrack_lock);
206
207         ho = hash_conntrack(&ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple);
208         hr = hash_conntrack(&ct->tuplehash[IP_CT_DIR_REPLY].tuple);
209         LIST_DELETE(&ip_conntrack_hash[ho], &ct->tuplehash[IP_CT_DIR_ORIGINAL]);
210         LIST_DELETE(&ip_conntrack_hash[hr], &ct->tuplehash[IP_CT_DIR_REPLY]);
211
212         /* Destroy all pending expectations */
213         remove_expectations(ct);
214 }
215
216 static void
217 destroy_conntrack(struct nf_conntrack *nfct)
218 {
219         struct ip_conntrack *ct = (struct ip_conntrack *)nfct;
220         struct ip_conntrack_protocol *proto;
221
222         DEBUGP("destroy_conntrack(%p)\n", ct);
223         IP_NF_ASSERT(atomic_read(&nfct->use) == 0);
224         IP_NF_ASSERT(!timer_pending(&ct->timeout));
225
226         /* To make sure we don't get any weird locking issues here:
227          * destroy_conntrack() MUST NOT be called with a write lock
228          * to ip_conntrack_lock!!! -HW */
229         proto = ip_ct_find_proto(ct->tuplehash[IP_CT_DIR_REPLY].tuple.dst.protonum);
230         if (proto && proto->destroy)
231                 proto->destroy(ct);
232
233         if (ip_conntrack_destroyed)
234                 ip_conntrack_destroyed(ct);
235
236         write_lock_bh(&ip_conntrack_lock);
237         /* Expectations will have been removed in clean_from_lists,
238          * except TFTP can create an expectation on the first packet,
239          * before connection is in the list, so we need to clean here,
240          * too. */
241         remove_expectations(ct);
242
243         /* We overload first tuple to link into unconfirmed list. */
244         if (!is_confirmed(ct)) {
245                 BUG_ON(list_empty(&ct->tuplehash[IP_CT_DIR_ORIGINAL].list));
246                 list_del(&ct->tuplehash[IP_CT_DIR_ORIGINAL].list);
247         }
248
249         CONNTRACK_STAT_INC(delete);
250         write_unlock_bh(&ip_conntrack_lock);
251
252         if (ct->master)
253                 ip_conntrack_put(ct->master);
254
255         DEBUGP("destroy_conntrack: returning ct=%p to slab\n", ct);
256         kmem_cache_free(ip_conntrack_cachep, ct);
257         atomic_dec(&ip_conntrack_count);
258 }
259
260 static void death_by_timeout(unsigned long ul_conntrack)
261 {
262         struct ip_conntrack *ct = (void *)ul_conntrack;
263
264         write_lock_bh(&ip_conntrack_lock);
265         /* Inside lock so preempt is disabled on module removal path.
266          * Otherwise we can get spurious warnings. */
267         CONNTRACK_STAT_INC(delete_list);
268         clean_from_lists(ct);
269         write_unlock_bh(&ip_conntrack_lock);
270         ip_conntrack_put(ct);
271 }
272
273 static inline int
274 conntrack_tuple_cmp(const struct ip_conntrack_tuple_hash *i,
275                     const struct ip_conntrack_tuple *tuple,
276                     const struct ip_conntrack *ignored_conntrack)
277 {
278         ASSERT_READ_LOCK(&ip_conntrack_lock);
279         return tuplehash_to_ctrack(i) != ignored_conntrack
280                 && ip_ct_tuple_equal(tuple, &i->tuple);
281 }
282
283 static struct ip_conntrack_tuple_hash *
284 __ip_conntrack_find(const struct ip_conntrack_tuple *tuple,
285                     const struct ip_conntrack *ignored_conntrack)
286 {
287         struct ip_conntrack_tuple_hash *h;
288         unsigned int hash = hash_conntrack(tuple);
289
290         ASSERT_READ_LOCK(&ip_conntrack_lock);
291         list_for_each_entry(h, &ip_conntrack_hash[hash], list) {
292                 if (conntrack_tuple_cmp(h, tuple, ignored_conntrack)) {
293                         CONNTRACK_STAT_INC(found);
294                         return h;
295                 }
296                 CONNTRACK_STAT_INC(searched);
297         }
298
299         return NULL;
300 }
301
302 /* Find a connection corresponding to a tuple. */
303 struct ip_conntrack_tuple_hash *
304 ip_conntrack_find_get(const struct ip_conntrack_tuple *tuple,
305                       const struct ip_conntrack *ignored_conntrack)
306 {
307         struct ip_conntrack_tuple_hash *h;
308
309         read_lock_bh(&ip_conntrack_lock);
310         h = __ip_conntrack_find(tuple, ignored_conntrack);
311         if (h)
312                 atomic_inc(&tuplehash_to_ctrack(h)->ct_general.use);
313         read_unlock_bh(&ip_conntrack_lock);
314
315         return h;
316 }
317
318 /* Confirm a connection given skb; places it in hash table */
319 int
320 __ip_conntrack_confirm(struct sk_buff **pskb)
321 {
322         unsigned int hash, repl_hash;
323         struct ip_conntrack *ct;
324         enum ip_conntrack_info ctinfo;
325
326         ct = ip_conntrack_get(*pskb, &ctinfo);
327
328         /* ipt_REJECT uses ip_conntrack_attach to attach related
329            ICMP/TCP RST packets in other direction.  Actual packet
330            which created connection will be IP_CT_NEW or for an
331            expected connection, IP_CT_RELATED. */
332         if (CTINFO2DIR(ctinfo) != IP_CT_DIR_ORIGINAL)
333                 return NF_ACCEPT;
334
335         hash = hash_conntrack(&ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple);
336         repl_hash = hash_conntrack(&ct->tuplehash[IP_CT_DIR_REPLY].tuple);
337
338         /* We're not in hash table, and we refuse to set up related
339            connections for unconfirmed conns.  But packet copies and
340            REJECT will give spurious warnings here. */
341         /* IP_NF_ASSERT(atomic_read(&ct->ct_general.use) == 1); */
342
343         /* No external references means noone else could have
344            confirmed us. */
345         IP_NF_ASSERT(!is_confirmed(ct));
346         DEBUGP("Confirming conntrack %p\n", ct);
347
348         write_lock_bh(&ip_conntrack_lock);
349
350         /* See if there's one in the list already, including reverse:
351            NAT could have grabbed it without realizing, since we're
352            not in the hash.  If there is, we lost race. */
353         if (!LIST_FIND(&ip_conntrack_hash[hash],
354                        conntrack_tuple_cmp,
355                        struct ip_conntrack_tuple_hash *,
356                        &ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple, NULL)
357             && !LIST_FIND(&ip_conntrack_hash[repl_hash],
358                           conntrack_tuple_cmp,
359                           struct ip_conntrack_tuple_hash *,
360                           &ct->tuplehash[IP_CT_DIR_REPLY].tuple, NULL)) {
361                 /* Remove from unconfirmed list */
362                 list_del(&ct->tuplehash[IP_CT_DIR_ORIGINAL].list);
363
364                 list_prepend(&ip_conntrack_hash[hash],
365                              &ct->tuplehash[IP_CT_DIR_ORIGINAL]);
366                 list_prepend(&ip_conntrack_hash[repl_hash],
367                              &ct->tuplehash[IP_CT_DIR_REPLY]);
368                 /* Timer relative to confirmation time, not original
369                    setting time, otherwise we'd get timer wrap in
370                    weird delay cases. */
371                 ct->timeout.expires += jiffies;
372                 add_timer(&ct->timeout);
373                 atomic_inc(&ct->ct_general.use);
374                 set_bit(IPS_CONFIRMED_BIT, &ct->status);
375                 CONNTRACK_STAT_INC(insert);
376                 write_unlock_bh(&ip_conntrack_lock);
377                 return NF_ACCEPT;
378         }
379
380         CONNTRACK_STAT_INC(insert_failed);
381         write_unlock_bh(&ip_conntrack_lock);
382
383         return NF_DROP;
384 }
385
386 /* Returns true if a connection correspondings to the tuple (required
387    for NAT). */
388 int
389 ip_conntrack_tuple_taken(const struct ip_conntrack_tuple *tuple,
390                          const struct ip_conntrack *ignored_conntrack)
391 {
392         struct ip_conntrack_tuple_hash *h;
393
394         read_lock_bh(&ip_conntrack_lock);
395         h = __ip_conntrack_find(tuple, ignored_conntrack);
396         read_unlock_bh(&ip_conntrack_lock);
397
398         return h != NULL;
399 }
400
401 /* There's a small race here where we may free a just-assured
402    connection.  Too bad: we're in trouble anyway. */
403 static inline int unreplied(const struct ip_conntrack_tuple_hash *i)
404 {
405         return !(test_bit(IPS_ASSURED_BIT, &tuplehash_to_ctrack(i)->status));
406 }
407
408 static int early_drop(struct list_head *chain)
409 {
410         /* Traverse backwards: gives us oldest, which is roughly LRU */
411         struct ip_conntrack_tuple_hash *h;
412         struct ip_conntrack *ct = NULL;
413         int dropped = 0;
414
415         read_lock_bh(&ip_conntrack_lock);
416         h = LIST_FIND_B(chain, unreplied, struct ip_conntrack_tuple_hash *);
417         if (h) {
418                 ct = tuplehash_to_ctrack(h);
419                 atomic_inc(&ct->ct_general.use);
420         }
421         read_unlock_bh(&ip_conntrack_lock);
422
423         if (!ct)
424                 return dropped;
425
426         if (del_timer(&ct->timeout)) {
427                 death_by_timeout((unsigned long)ct);
428                 dropped = 1;
429                 CONNTRACK_STAT_INC(early_drop);
430         }
431         ip_conntrack_put(ct);
432         return dropped;
433 }
434
435 static inline int helper_cmp(const struct ip_conntrack_helper *i,
436                              const struct ip_conntrack_tuple *rtuple)
437 {
438         return ip_ct_tuple_mask_cmp(rtuple, &i->tuple, &i->mask);
439 }
440
441 static struct ip_conntrack_helper *ip_ct_find_helper(const struct ip_conntrack_tuple *tuple)
442 {
443         return LIST_FIND(&helpers, helper_cmp,
444                          struct ip_conntrack_helper *,
445                          tuple);
446 }
447
448 /* Allocate a new conntrack: we return -ENOMEM if classification
449    failed due to stress.  Otherwise it really is unclassifiable. */
450 static struct ip_conntrack_tuple_hash *
451 init_conntrack(const struct ip_conntrack_tuple *tuple,
452                struct ip_conntrack_protocol *protocol,
453                struct sk_buff *skb)
454 {
455         struct ip_conntrack *conntrack;
456         struct ip_conntrack_tuple repl_tuple;
457         size_t hash;
458         struct ip_conntrack_expect *exp;
459
460         if (!ip_conntrack_hash_rnd_initted) {
461                 get_random_bytes(&ip_conntrack_hash_rnd, 4);
462                 ip_conntrack_hash_rnd_initted = 1;
463         }
464
465         hash = hash_conntrack(tuple);
466
467         if (ip_conntrack_max
468             && atomic_read(&ip_conntrack_count) >= ip_conntrack_max) {
469                 /* Try dropping from this hash chain. */
470                 if (!early_drop(&ip_conntrack_hash[hash])) {
471                         if (net_ratelimit())
472                                 printk(KERN_WARNING
473                                        "ip_conntrack: table full, dropping"
474                                        " packet.\n");
475                         return ERR_PTR(-ENOMEM);
476                 }
477         }
478
479         if (!ip_ct_invert_tuple(&repl_tuple, tuple, protocol)) {
480                 DEBUGP("Can't invert tuple.\n");
481                 return NULL;
482         }
483
484         conntrack = kmem_cache_alloc(ip_conntrack_cachep, GFP_ATOMIC);
485         if (!conntrack) {
486                 DEBUGP("Can't allocate conntrack.\n");
487                 return ERR_PTR(-ENOMEM);
488         }
489
490         memset(conntrack, 0, sizeof(*conntrack));
491         atomic_set(&conntrack->ct_general.use, 1);
492         conntrack->ct_general.destroy = destroy_conntrack;
493         conntrack->tuplehash[IP_CT_DIR_ORIGINAL].tuple = *tuple;
494         conntrack->tuplehash[IP_CT_DIR_REPLY].tuple = repl_tuple;
495         if (!protocol->new(conntrack, skb)) {
496                 kmem_cache_free(ip_conntrack_cachep, conntrack);
497                 return NULL;
498         }
499         /* Don't set timer yet: wait for confirmation */
500         init_timer(&conntrack->timeout);
501         conntrack->timeout.data = (unsigned long)conntrack;
502         conntrack->timeout.function = death_by_timeout;
503
504         write_lock_bh(&ip_conntrack_lock);
505         exp = find_expectation(tuple);
506
507         if (exp) {
508                 DEBUGP("conntrack: expectation arrives ct=%p exp=%p\n",
509                         conntrack, exp);
510                 /* Welcome, Mr. Bond.  We've been expecting you... */
511                 __set_bit(IPS_EXPECTED_BIT, &conntrack->status);
512                 conntrack->master = exp->master;
513 #if CONFIG_IP_NF_CONNTRACK_MARK
514                 conntrack->mark = exp->master->mark;
515 #endif
516                 nf_conntrack_get(&conntrack->master->ct_general);
517                 CONNTRACK_STAT_INC(expect_new);
518         } else {
519                 conntrack->helper = ip_ct_find_helper(&repl_tuple);
520
521                 CONNTRACK_STAT_INC(new);
522         }
523
524         /* Overload tuple linked list to put us in unconfirmed list. */
525         list_add(&conntrack->tuplehash[IP_CT_DIR_ORIGINAL].list, &unconfirmed);
526
527         atomic_inc(&ip_conntrack_count);
528         write_unlock_bh(&ip_conntrack_lock);
529
530         if (exp) {
531                 if (exp->expectfn)
532                         exp->expectfn(conntrack, exp);
533                 ip_conntrack_expect_put(exp);
534         }
535
536         return &conntrack->tuplehash[IP_CT_DIR_ORIGINAL];
537 }
538
539 /* On success, returns conntrack ptr, sets skb->nfct and ctinfo */
540 static inline struct ip_conntrack *
541 resolve_normal_ct(struct sk_buff *skb,
542                   struct ip_conntrack_protocol *proto,
543                   int *set_reply,
544                   unsigned int hooknum,
545                   enum ip_conntrack_info *ctinfo)
546 {
547         struct ip_conntrack_tuple tuple;
548         struct ip_conntrack_tuple_hash *h;
549         struct ip_conntrack *ct;
550
551         IP_NF_ASSERT((skb->nh.iph->frag_off & htons(IP_OFFSET)) == 0);
552
553         if (!ip_ct_get_tuple(skb->nh.iph, skb, skb->nh.iph->ihl*4, 
554                                 &tuple,proto))
555                 return NULL;
556
557         /* look for tuple match */
558         h = ip_conntrack_find_get(&tuple, NULL);
559         if (!h) {
560                 h = init_conntrack(&tuple, proto, skb);
561                 if (!h)
562                         return NULL;
563                 if (IS_ERR(h))
564                         return (void *)h;
565         }
566         ct = tuplehash_to_ctrack(h);
567
568         /* It exists; we have (non-exclusive) reference. */
569         if (DIRECTION(h) == IP_CT_DIR_REPLY) {
570                 *ctinfo = IP_CT_ESTABLISHED + IP_CT_IS_REPLY;
571                 /* Please set reply bit if this packet OK */
572                 *set_reply = 1;
573         } else {
574                 /* Once we've had two way comms, always ESTABLISHED. */
575                 if (test_bit(IPS_SEEN_REPLY_BIT, &ct->status)) {
576                         DEBUGP("ip_conntrack_in: normal packet for %p\n",
577                                ct);
578                         *ctinfo = IP_CT_ESTABLISHED;
579                 } else if (test_bit(IPS_EXPECTED_BIT, &ct->status)) {
580                         DEBUGP("ip_conntrack_in: related packet for %p\n",
581                                ct);
582                         *ctinfo = IP_CT_RELATED;
583                 } else {
584                         DEBUGP("ip_conntrack_in: new packet for %p\n",
585                                ct);
586                         *ctinfo = IP_CT_NEW;
587                 }
588                 *set_reply = 0;
589         }
590         skb->nfct = &ct->ct_general;
591         skb->nfctinfo = *ctinfo;
592         return ct;
593 }
594
595 /* Netfilter hook itself. */
596 unsigned int ip_conntrack_in(unsigned int hooknum,
597                              struct sk_buff **pskb,
598                              const struct net_device *in,
599                              const struct net_device *out,
600                              int (*okfn)(struct sk_buff *))
601 {
602         struct ip_conntrack *ct;
603         enum ip_conntrack_info ctinfo;
604         struct ip_conntrack_protocol *proto;
605         int set_reply;
606         int ret;
607
608         /* Previously seen (loopback or untracked)?  Ignore. */
609         if ((*pskb)->nfct) {
610                 CONNTRACK_STAT_INC(ignore);
611                 return NF_ACCEPT;
612         }
613
614         /* Never happen */
615         if ((*pskb)->nh.iph->frag_off & htons(IP_OFFSET)) {
616                 if (net_ratelimit()) {
617                 printk(KERN_ERR "ip_conntrack_in: Frag of proto %u (hook=%u)\n",
618                        (*pskb)->nh.iph->protocol, hooknum);
619                 }
620                 return NF_DROP;
621         }
622
623         /* FIXME: Do this right please. --RR */
624         (*pskb)->nfcache |= NFC_UNKNOWN;
625
626 /* Doesn't cover locally-generated broadcast, so not worth it. */
627 #if 0
628         /* Ignore broadcast: no `connection'. */
629         if ((*pskb)->pkt_type == PACKET_BROADCAST) {
630                 printk("Broadcast packet!\n");
631                 return NF_ACCEPT;
632         } else if (((*pskb)->nh.iph->daddr & htonl(0x000000FF)) 
633                    == htonl(0x000000FF)) {
634                 printk("Should bcast: %u.%u.%u.%u->%u.%u.%u.%u (sk=%p, ptype=%u)\n",
635                        NIPQUAD((*pskb)->nh.iph->saddr),
636                        NIPQUAD((*pskb)->nh.iph->daddr),
637                        (*pskb)->sk, (*pskb)->pkt_type);
638         }
639 #endif
640
641         proto = ip_ct_find_proto((*pskb)->nh.iph->protocol);
642
643         /* It may be an special packet, error, unclean...
644          * inverse of the return code tells to the netfilter
645          * core what to do with the packet. */
646         if (proto->error != NULL 
647             && (ret = proto->error(*pskb, &ctinfo, hooknum)) <= 0) {
648                 CONNTRACK_STAT_INC(error);
649                 CONNTRACK_STAT_INC(invalid);
650                 return -ret;
651         }
652
653         if (!(ct = resolve_normal_ct(*pskb, proto,&set_reply,hooknum,&ctinfo))) {
654                 /* Not valid part of a connection */
655                 CONNTRACK_STAT_INC(invalid);
656                 return NF_ACCEPT;
657         }
658
659         if (IS_ERR(ct)) {
660                 /* Too stressed to deal. */
661                 CONNTRACK_STAT_INC(drop);
662                 return NF_DROP;
663         }
664
665         IP_NF_ASSERT((*pskb)->nfct);
666
667         ret = proto->packet(ct, *pskb, ctinfo);
668         if (ret < 0) {
669                 /* Invalid: inverse of the return code tells
670                  * the netfilter core what to do*/
671                 nf_conntrack_put((*pskb)->nfct);
672                 (*pskb)->nfct = NULL;
673                 CONNTRACK_STAT_INC(invalid);
674                 return -ret;
675         }
676
677         if (set_reply)
678                 set_bit(IPS_SEEN_REPLY_BIT, &ct->status);
679
680         return ret;
681 }
682
683 int invert_tuplepr(struct ip_conntrack_tuple *inverse,
684                    const struct ip_conntrack_tuple *orig)
685 {
686         return ip_ct_invert_tuple(inverse, orig, 
687                                   ip_ct_find_proto(orig->dst.protonum));
688 }
689
690 /* Would two expected things clash? */
691 static inline int expect_clash(const struct ip_conntrack_expect *a,
692                                const struct ip_conntrack_expect *b)
693 {
694         /* Part covered by intersection of masks must be unequal,
695            otherwise they clash */
696         struct ip_conntrack_tuple intersect_mask
697                 = { { a->mask.src.ip & b->mask.src.ip,
698                       { a->mask.src.u.all & b->mask.src.u.all } },
699                     { a->mask.dst.ip & b->mask.dst.ip,
700                       { a->mask.dst.u.all & b->mask.dst.u.all },
701                       a->mask.dst.protonum & b->mask.dst.protonum } };
702
703         return ip_ct_tuple_mask_cmp(&a->tuple, &b->tuple, &intersect_mask);
704 }
705
706 static inline int expect_matches(const struct ip_conntrack_expect *a,
707                                  const struct ip_conntrack_expect *b)
708 {
709         return a->master == b->master
710                 && ip_ct_tuple_equal(&a->tuple, &b->tuple)
711                 && ip_ct_tuple_equal(&a->mask, &b->mask);
712 }
713
714 /* Generally a bad idea to call this: could have matched already. */
715 void ip_conntrack_unexpect_related(struct ip_conntrack_expect *exp)
716 {
717         struct ip_conntrack_expect *i;
718
719         write_lock_bh(&ip_conntrack_lock);
720         /* choose the the oldest expectation to evict */
721         list_for_each_entry_reverse(i, &ip_conntrack_expect_list, list) {
722                 if (expect_matches(i, exp) && del_timer(&i->timeout)) {
723                         unlink_expect(i);
724                         write_unlock_bh(&ip_conntrack_lock);
725                         ip_conntrack_expect_put(i);
726                         return;
727                 }
728         }
729         write_unlock_bh(&ip_conntrack_lock);
730 }
731
732 struct ip_conntrack_expect *ip_conntrack_expect_alloc(struct ip_conntrack *me)
733 {
734         struct ip_conntrack_expect *new;
735
736         new = kmem_cache_alloc(ip_conntrack_expect_cachep, GFP_ATOMIC);
737         if (!new) {
738                 DEBUGP("expect_related: OOM allocating expect\n");
739                 return NULL;
740         }
741         new->master = me;
742         atomic_inc(&new->master->ct_general.use);
743         atomic_set(&new->use, 1);
744         return new;
745 }
746
747 void ip_conntrack_expect_put(struct ip_conntrack_expect *exp)
748 {
749         if (atomic_dec_and_test(&exp->use)) {
750                 ip_conntrack_put(exp->master);
751                 kmem_cache_free(ip_conntrack_expect_cachep, exp);
752         }
753 }
754
755 static void ip_conntrack_expect_insert(struct ip_conntrack_expect *exp)
756 {
757         atomic_inc(&exp->use);
758         exp->master->expecting++;
759         list_add(&exp->list, &ip_conntrack_expect_list);
760
761         init_timer(&exp->timeout);
762         exp->timeout.data = (unsigned long)exp;
763         exp->timeout.function = expectation_timed_out;
764         exp->timeout.expires = jiffies + exp->master->helper->timeout * HZ;
765         add_timer(&exp->timeout);
766
767         CONNTRACK_STAT_INC(expect_create);
768 }
769
770 /* Race with expectations being used means we could have none to find; OK. */
771 static void evict_oldest_expect(struct ip_conntrack *master)
772 {
773         struct ip_conntrack_expect *i;
774
775         list_for_each_entry_reverse(i, &ip_conntrack_expect_list, list) {
776                 if (i->master == master) {
777                         if (del_timer(&i->timeout)) {
778                                 unlink_expect(i);
779                                 ip_conntrack_expect_put(i);
780                         }
781                         break;
782                 }
783         }
784 }
785
786 static inline int refresh_timer(struct ip_conntrack_expect *i)
787 {
788         if (!del_timer(&i->timeout))
789                 return 0;
790
791         i->timeout.expires = jiffies + i->master->helper->timeout*HZ;
792         add_timer(&i->timeout);
793         return 1;
794 }
795
796 int ip_conntrack_expect_related(struct ip_conntrack_expect *expect)
797 {
798         struct ip_conntrack_expect *i;
799         int ret;
800
801         DEBUGP("ip_conntrack_expect_related %p\n", related_to);
802         DEBUGP("tuple: "); DUMP_TUPLE(&expect->tuple);
803         DEBUGP("mask:  "); DUMP_TUPLE(&expect->mask);
804
805         write_lock_bh(&ip_conntrack_lock);
806         list_for_each_entry(i, &ip_conntrack_expect_list, list) {
807                 if (expect_matches(i, expect)) {
808                         /* Refresh timer: if it's dying, ignore.. */
809                         if (refresh_timer(i)) {
810                                 ret = 0;
811                                 goto out;
812                         }
813                 } else if (expect_clash(i, expect)) {
814                         ret = -EBUSY;
815                         goto out;
816                 }
817         }
818
819         /* Will be over limit? */
820         if (expect->master->helper->max_expected && 
821             expect->master->expecting >= expect->master->helper->max_expected)
822                 evict_oldest_expect(expect->master);
823
824         ip_conntrack_expect_insert(expect);
825         ret = 0;
826 out:
827         write_unlock_bh(&ip_conntrack_lock);
828         return ret;
829 }
830
831 /* Alter reply tuple (maybe alter helper).  This is for NAT, and is
832    implicitly racy: see __ip_conntrack_confirm */
833 void ip_conntrack_alter_reply(struct ip_conntrack *conntrack,
834                               const struct ip_conntrack_tuple *newreply)
835 {
836         write_lock_bh(&ip_conntrack_lock);
837         /* Should be unconfirmed, so not in hash table yet */
838         IP_NF_ASSERT(!is_confirmed(conntrack));
839
840         DEBUGP("Altering reply tuple of %p to ", conntrack);
841         DUMP_TUPLE(newreply);
842
843         conntrack->tuplehash[IP_CT_DIR_REPLY].tuple = *newreply;
844         if (!conntrack->master && conntrack->expecting == 0)
845                 conntrack->helper = ip_ct_find_helper(newreply);
846         write_unlock_bh(&ip_conntrack_lock);
847 }
848
849 int ip_conntrack_helper_register(struct ip_conntrack_helper *me)
850 {
851         BUG_ON(me->timeout == 0);
852         write_lock_bh(&ip_conntrack_lock);
853         list_prepend(&helpers, me);
854         write_unlock_bh(&ip_conntrack_lock);
855
856         return 0;
857 }
858
859 static inline int unhelp(struct ip_conntrack_tuple_hash *i,
860                          const struct ip_conntrack_helper *me)
861 {
862         if (tuplehash_to_ctrack(i)->helper == me)
863                 tuplehash_to_ctrack(i)->helper = NULL;
864         return 0;
865 }
866
867 void ip_conntrack_helper_unregister(struct ip_conntrack_helper *me)
868 {
869         unsigned int i;
870         struct ip_conntrack_expect *exp, *tmp;
871
872         /* Need write lock here, to delete helper. */
873         write_lock_bh(&ip_conntrack_lock);
874         LIST_DELETE(&helpers, me);
875
876         /* Get rid of expectations */
877         list_for_each_entry_safe(exp, tmp, &ip_conntrack_expect_list, list) {
878                 if (exp->master->helper == me && del_timer(&exp->timeout)) {
879                         unlink_expect(exp);
880                         ip_conntrack_expect_put(exp);
881                 }
882         }
883         /* Get rid of expecteds, set helpers to NULL. */
884         LIST_FIND_W(&unconfirmed, unhelp, struct ip_conntrack_tuple_hash*, me);
885         for (i = 0; i < ip_conntrack_htable_size; i++)
886                 LIST_FIND_W(&ip_conntrack_hash[i], unhelp,
887                             struct ip_conntrack_tuple_hash *, me);
888         write_unlock_bh(&ip_conntrack_lock);
889
890         /* Someone could be still looking at the helper in a bh. */
891         synchronize_net();
892 }
893
894 static inline void ct_add_counters(struct ip_conntrack *ct,
895                                    enum ip_conntrack_info ctinfo,
896                                    const struct sk_buff *skb)
897 {
898 #ifdef CONFIG_IP_NF_CT_ACCT
899         if (skb) {
900                 ct->counters[CTINFO2DIR(ctinfo)].packets++;
901                 ct->counters[CTINFO2DIR(ctinfo)].bytes += 
902                                         ntohs(skb->nh.iph->tot_len);
903         }
904 #endif
905 }
906
907 /* Refresh conntrack for this many jiffies and do accounting (if skb != NULL) */
908 void ip_ct_refresh_acct(struct ip_conntrack *ct, 
909                         enum ip_conntrack_info ctinfo,
910                         const struct sk_buff *skb,
911                         unsigned long extra_jiffies)
912 {
913         IP_NF_ASSERT(ct->timeout.data == (unsigned long)ct);
914
915         /* If not in hash table, timer will not be active yet */
916         if (!is_confirmed(ct)) {
917                 ct->timeout.expires = extra_jiffies;
918                 ct_add_counters(ct, ctinfo, skb);
919         } else {
920                 write_lock_bh(&ip_conntrack_lock);
921                 /* Need del_timer for race avoidance (may already be dying). */
922                 if (del_timer(&ct->timeout)) {
923                         ct->timeout.expires = jiffies + extra_jiffies;
924                         add_timer(&ct->timeout);
925                 }
926                 ct_add_counters(ct, ctinfo, skb);
927                 write_unlock_bh(&ip_conntrack_lock);
928         }
929 }
930
931 /* Returns new sk_buff, or NULL */
932 struct sk_buff *
933 ip_ct_gather_frags(struct sk_buff *skb, u_int32_t user)
934 {
935         skb_orphan(skb);
936
937         local_bh_disable(); 
938         skb = ip_defrag(skb, user);
939         local_bh_enable();
940
941         if (skb) {
942                 ip_send_check(skb->nh.iph);
943                 skb->nfcache |= NFC_ALTERED;
944         }
945         return skb;
946 }
947
948 /* Used by ipt_REJECT. */
949 static void ip_conntrack_attach(struct sk_buff *nskb, struct sk_buff *skb)
950 {
951         struct ip_conntrack *ct;
952         enum ip_conntrack_info ctinfo;
953
954         /* This ICMP is in reverse direction to the packet which caused it */
955         ct = ip_conntrack_get(skb, &ctinfo);
956         
957         if (CTINFO2DIR(ctinfo) == IP_CT_DIR_ORIGINAL)
958                 ctinfo = IP_CT_RELATED + IP_CT_IS_REPLY;
959         else
960                 ctinfo = IP_CT_RELATED;
961
962         /* Attach to new skbuff, and increment count */
963         nskb->nfct = &ct->ct_general;
964         nskb->nfctinfo = ctinfo;
965         nf_conntrack_get(nskb->nfct);
966 }
967
968 static inline int
969 do_iter(const struct ip_conntrack_tuple_hash *i,
970         int (*iter)(struct ip_conntrack *i, void *data),
971         void *data)
972 {
973         return iter(tuplehash_to_ctrack(i), data);
974 }
975
976 /* Bring out ya dead! */
977 static struct ip_conntrack_tuple_hash *
978 get_next_corpse(int (*iter)(struct ip_conntrack *i, void *data),
979                 void *data, unsigned int *bucket)
980 {
981         struct ip_conntrack_tuple_hash *h = NULL;
982
983         write_lock_bh(&ip_conntrack_lock);
984         for (; *bucket < ip_conntrack_htable_size; (*bucket)++) {
985                 h = LIST_FIND_W(&ip_conntrack_hash[*bucket], do_iter,
986                                 struct ip_conntrack_tuple_hash *, iter, data);
987                 if (h)
988                         break;
989         }
990         if (!h)
991                 h = LIST_FIND_W(&unconfirmed, do_iter,
992                                 struct ip_conntrack_tuple_hash *, iter, data);
993         if (h)
994                 atomic_inc(&tuplehash_to_ctrack(h)->ct_general.use);
995         write_unlock_bh(&ip_conntrack_lock);
996
997         return h;
998 }
999
1000 void
1001 ip_ct_iterate_cleanup(int (*iter)(struct ip_conntrack *i, void *), void *data)
1002 {
1003         struct ip_conntrack_tuple_hash *h;
1004         unsigned int bucket = 0;
1005
1006         while ((h = get_next_corpse(iter, data, &bucket)) != NULL) {
1007                 struct ip_conntrack *ct = tuplehash_to_ctrack(h);
1008                 /* Time to push up daises... */
1009                 if (del_timer(&ct->timeout))
1010                         death_by_timeout((unsigned long)ct);
1011                 /* ... else the timer will get him soon. */
1012
1013                 ip_conntrack_put(ct);
1014         }
1015 }
1016
1017 /* Fast function for those who don't want to parse /proc (and I don't
1018    blame them). */
1019 /* Reversing the socket's dst/src point of view gives us the reply
1020    mapping. */
1021 static int
1022 getorigdst(struct sock *sk, int optval, void __user *user, int *len)
1023 {
1024         struct inet_sock *inet = inet_sk(sk);
1025         struct ip_conntrack_tuple_hash *h;
1026         struct ip_conntrack_tuple tuple;
1027         
1028         IP_CT_TUPLE_U_BLANK(&tuple);
1029         tuple.src.ip = inet->rcv_saddr;
1030         tuple.src.u.tcp.port = inet->sport;
1031         tuple.dst.ip = inet->daddr;
1032         tuple.dst.u.tcp.port = inet->dport;
1033         tuple.dst.protonum = IPPROTO_TCP;
1034
1035         /* We only do TCP at the moment: is there a better way? */
1036         if (strcmp(sk->sk_prot->name, "TCP")) {
1037                 DEBUGP("SO_ORIGINAL_DST: Not a TCP socket\n");
1038                 return -ENOPROTOOPT;
1039         }
1040
1041         if ((unsigned int) *len < sizeof(struct sockaddr_in)) {
1042                 DEBUGP("SO_ORIGINAL_DST: len %u not %u\n",
1043                        *len, sizeof(struct sockaddr_in));
1044                 return -EINVAL;
1045         }
1046
1047         h = ip_conntrack_find_get(&tuple, NULL);
1048         if (h) {
1049                 struct sockaddr_in sin;
1050                 struct ip_conntrack *ct = tuplehash_to_ctrack(h);
1051
1052                 sin.sin_family = AF_INET;
1053                 sin.sin_port = ct->tuplehash[IP_CT_DIR_ORIGINAL]
1054                         .tuple.dst.u.tcp.port;
1055                 sin.sin_addr.s_addr = ct->tuplehash[IP_CT_DIR_ORIGINAL]
1056                         .tuple.dst.ip;
1057
1058                 DEBUGP("SO_ORIGINAL_DST: %u.%u.%u.%u %u\n",
1059                        NIPQUAD(sin.sin_addr.s_addr), ntohs(sin.sin_port));
1060                 ip_conntrack_put(ct);
1061                 if (copy_to_user(user, &sin, sizeof(sin)) != 0)
1062                         return -EFAULT;
1063                 else
1064                         return 0;
1065         }
1066         DEBUGP("SO_ORIGINAL_DST: Can't find %u.%u.%u.%u/%u-%u.%u.%u.%u/%u.\n",
1067                NIPQUAD(tuple.src.ip), ntohs(tuple.src.u.tcp.port),
1068                NIPQUAD(tuple.dst.ip), ntohs(tuple.dst.u.tcp.port));
1069         return -ENOENT;
1070 }
1071
1072 static struct nf_sockopt_ops so_getorigdst = {
1073         .pf             = PF_INET,
1074         .get_optmin     = SO_ORIGINAL_DST,
1075         .get_optmax     = SO_ORIGINAL_DST+1,
1076         .get            = &getorigdst,
1077 };
1078
1079 static int kill_all(struct ip_conntrack *i, void *data)
1080 {
1081         return 1;
1082 }
1083
1084 static void free_conntrack_hash(void)
1085 {
1086         if (ip_conntrack_vmalloc)
1087                 vfree(ip_conntrack_hash);
1088         else
1089                 free_pages((unsigned long)ip_conntrack_hash, 
1090                            get_order(sizeof(struct list_head)
1091                                      * ip_conntrack_htable_size));
1092 }
1093
1094 /* Mishearing the voices in his head, our hero wonders how he's
1095    supposed to kill the mall. */
1096 void ip_conntrack_cleanup(void)
1097 {
1098         ip_ct_attach = NULL;
1099         /* This makes sure all current packets have passed through
1100            netfilter framework.  Roll on, two-stage module
1101            delete... */
1102         synchronize_net();
1103  
1104  i_see_dead_people:
1105         ip_ct_iterate_cleanup(kill_all, NULL);
1106         if (atomic_read(&ip_conntrack_count) != 0) {
1107                 schedule();
1108                 goto i_see_dead_people;
1109         }
1110         /* wait until all references to ip_conntrack_untracked are dropped */
1111         while (atomic_read(&ip_conntrack_untracked.ct_general.use) > 1)
1112                 schedule();
1113
1114         kmem_cache_destroy(ip_conntrack_cachep);
1115         kmem_cache_destroy(ip_conntrack_expect_cachep);
1116         free_conntrack_hash();
1117         nf_unregister_sockopt(&so_getorigdst);
1118 }
1119
1120 static int hashsize;
1121 module_param(hashsize, int, 0400);
1122
1123 int __init ip_conntrack_init(void)
1124 {
1125         unsigned int i;
1126         int ret;
1127
1128         /* Idea from tcp.c: use 1/16384 of memory.  On i386: 32MB
1129          * machine has 256 buckets.  >= 1GB machines have 8192 buckets. */
1130         if (hashsize) {
1131                 ip_conntrack_htable_size = hashsize;
1132         } else {
1133                 ip_conntrack_htable_size
1134                         = (((num_physpages << PAGE_SHIFT) / 16384)
1135                            / sizeof(struct list_head));
1136                 if (num_physpages > (1024 * 1024 * 1024 / PAGE_SIZE))
1137                         ip_conntrack_htable_size = 8192;
1138                 if (ip_conntrack_htable_size < 16)
1139                         ip_conntrack_htable_size = 16;
1140         }
1141         ip_conntrack_max = 8 * ip_conntrack_htable_size;
1142
1143         printk("ip_conntrack version %s (%u buckets, %d max)"
1144                " - %Zd bytes per conntrack\n", IP_CONNTRACK_VERSION,
1145                ip_conntrack_htable_size, ip_conntrack_max,
1146                sizeof(struct ip_conntrack));
1147
1148         ret = nf_register_sockopt(&so_getorigdst);
1149         if (ret != 0) {
1150                 printk(KERN_ERR "Unable to register netfilter socket option\n");
1151                 return ret;
1152         }
1153
1154         /* AK: the hash table is twice as big than needed because it
1155            uses list_head.  it would be much nicer to caches to use a
1156            single pointer list head here. */
1157         ip_conntrack_vmalloc = 0; 
1158         ip_conntrack_hash 
1159                 =(void*)__get_free_pages(GFP_KERNEL, 
1160                                          get_order(sizeof(struct list_head)
1161                                                    *ip_conntrack_htable_size));
1162         if (!ip_conntrack_hash) { 
1163                 ip_conntrack_vmalloc = 1;
1164                 printk(KERN_WARNING "ip_conntrack: falling back to vmalloc.\n");
1165                 ip_conntrack_hash = vmalloc(sizeof(struct list_head)
1166                                             * ip_conntrack_htable_size);
1167         }
1168         if (!ip_conntrack_hash) {
1169                 printk(KERN_ERR "Unable to create ip_conntrack_hash\n");
1170                 goto err_unreg_sockopt;
1171         }
1172
1173         ip_conntrack_cachep = kmem_cache_create("ip_conntrack",
1174                                                 sizeof(struct ip_conntrack), 0,
1175                                                 0, NULL, NULL);
1176         if (!ip_conntrack_cachep) {
1177                 printk(KERN_ERR "Unable to create ip_conntrack slab cache\n");
1178                 goto err_free_hash;
1179         }
1180
1181         ip_conntrack_expect_cachep = kmem_cache_create("ip_conntrack_expect",
1182                                         sizeof(struct ip_conntrack_expect),
1183                                         0, 0, NULL, NULL);
1184         if (!ip_conntrack_expect_cachep) {
1185                 printk(KERN_ERR "Unable to create ip_expect slab cache\n");
1186                 goto err_free_conntrack_slab;
1187         }
1188
1189         /* Don't NEED lock here, but good form anyway. */
1190         write_lock_bh(&ip_conntrack_lock);
1191         for (i = 0; i < MAX_IP_CT_PROTO; i++)
1192                 ip_ct_protos[i] = &ip_conntrack_generic_protocol;
1193         /* Sew in builtin protocols. */
1194         ip_ct_protos[IPPROTO_TCP] = &ip_conntrack_protocol_tcp;
1195         ip_ct_protos[IPPROTO_UDP] = &ip_conntrack_protocol_udp;
1196         ip_ct_protos[IPPROTO_ICMP] = &ip_conntrack_protocol_icmp;
1197         write_unlock_bh(&ip_conntrack_lock);
1198
1199         for (i = 0; i < ip_conntrack_htable_size; i++)
1200                 INIT_LIST_HEAD(&ip_conntrack_hash[i]);
1201
1202         /* For use by ipt_REJECT */
1203         ip_ct_attach = ip_conntrack_attach;
1204
1205         /* Set up fake conntrack:
1206             - to never be deleted, not in any hashes */
1207         atomic_set(&ip_conntrack_untracked.ct_general.use, 1);
1208         /*  - and look it like as a confirmed connection */
1209         set_bit(IPS_CONFIRMED_BIT, &ip_conntrack_untracked.status);
1210
1211         return ret;
1212
1213 err_free_conntrack_slab:
1214         kmem_cache_destroy(ip_conntrack_cachep);
1215 err_free_hash:
1216         free_conntrack_hash();
1217 err_unreg_sockopt:
1218         nf_unregister_sockopt(&so_getorigdst);
1219
1220         return -ENOMEM;
1221 }