2 * x_tables core - Backend for {ip,ip6,arp}_tables
4 * Copyright (C) 2006-2006 Harald Welte <laforge@netfilter.org>
6 * Based on existing ip_tables code which is
7 * Copyright (C) 1999 Paul `Rusty' Russell & Michael J. Neuling
8 * Copyright (C) 2000-2005 Netfilter Core Team <coreteam@netfilter.org>
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License version 2 as
12 * published by the Free Software Foundation.
16 #include <linux/kernel.h>
17 #include <linux/socket.h>
18 #include <linux/net.h>
19 #include <linux/proc_fs.h>
20 #include <linux/seq_file.h>
21 #include <linux/string.h>
22 #include <linux/vmalloc.h>
23 #include <linux/mutex.h>
25 #include <net/net_namespace.h>
27 #include <linux/netfilter/x_tables.h>
28 #include <linux/netfilter_arp.h>
31 MODULE_LICENSE("GPL");
32 MODULE_AUTHOR("Harald Welte <laforge@netfilter.org>");
33 MODULE_DESCRIPTION("[ip,ip6,arp]_tables backend module");
35 #define SMP_ALIGN(x) (((x) + SMP_CACHE_BYTES-1) & ~(SMP_CACHE_BYTES-1))
38 struct compat_delta
*next
;
45 struct list_head match
;
46 struct list_head target
;
48 struct mutex compat_mutex
;
49 struct compat_delta
*compat_offsets
;
53 static struct xt_af
*xt
;
55 #ifdef DEBUG_IP_FIREWALL_USER
56 #define duprintf(format, args...) printk(format , ## args)
58 #define duprintf(format, args...)
61 static const char *const xt_prefix
[NPROTO
] = {
67 /* Registration hooks for targets. */
69 xt_register_target(struct xt_target
*target
)
71 int ret
, af
= target
->family
;
73 ret
= mutex_lock_interruptible(&xt
[af
].mutex
);
76 list_add(&target
->list
, &xt
[af
].target
);
77 mutex_unlock(&xt
[af
].mutex
);
80 EXPORT_SYMBOL(xt_register_target
);
83 xt_unregister_target(struct xt_target
*target
)
85 int af
= target
->family
;
87 mutex_lock(&xt
[af
].mutex
);
88 list_del(&target
->list
);
89 mutex_unlock(&xt
[af
].mutex
);
91 EXPORT_SYMBOL(xt_unregister_target
);
94 xt_register_targets(struct xt_target
*target
, unsigned int n
)
99 for (i
= 0; i
< n
; i
++) {
100 err
= xt_register_target(&target
[i
]);
108 xt_unregister_targets(target
, i
);
111 EXPORT_SYMBOL(xt_register_targets
);
114 xt_unregister_targets(struct xt_target
*target
, unsigned int n
)
118 for (i
= 0; i
< n
; i
++)
119 xt_unregister_target(&target
[i
]);
121 EXPORT_SYMBOL(xt_unregister_targets
);
124 xt_register_match(struct xt_match
*match
)
126 int ret
, af
= match
->family
;
128 ret
= mutex_lock_interruptible(&xt
[af
].mutex
);
132 list_add(&match
->list
, &xt
[af
].match
);
133 mutex_unlock(&xt
[af
].mutex
);
137 EXPORT_SYMBOL(xt_register_match
);
140 xt_unregister_match(struct xt_match
*match
)
142 int af
= match
->family
;
144 mutex_lock(&xt
[af
].mutex
);
145 list_del(&match
->list
);
146 mutex_unlock(&xt
[af
].mutex
);
148 EXPORT_SYMBOL(xt_unregister_match
);
151 xt_register_matches(struct xt_match
*match
, unsigned int n
)
156 for (i
= 0; i
< n
; i
++) {
157 err
= xt_register_match(&match
[i
]);
165 xt_unregister_matches(match
, i
);
168 EXPORT_SYMBOL(xt_register_matches
);
171 xt_unregister_matches(struct xt_match
*match
, unsigned int n
)
175 for (i
= 0; i
< n
; i
++)
176 xt_unregister_match(&match
[i
]);
178 EXPORT_SYMBOL(xt_unregister_matches
);
182 * These are weird, but module loading must not be done with mutex
183 * held (since they will register), and we have to have a single
184 * function to use try_then_request_module().
187 /* Find match, grabs ref. Returns ERR_PTR() on error. */
188 struct xt_match
*xt_find_match(int af
, const char *name
, u8 revision
)
193 if (mutex_lock_interruptible(&xt
[af
].mutex
) != 0)
194 return ERR_PTR(-EINTR
);
196 list_for_each_entry(m
, &xt
[af
].match
, list
) {
197 if (strcmp(m
->name
, name
) == 0) {
198 if (m
->revision
== revision
) {
199 if (try_module_get(m
->me
)) {
200 mutex_unlock(&xt
[af
].mutex
);
204 err
= -EPROTOTYPE
; /* Found something. */
207 mutex_unlock(&xt
[af
].mutex
);
210 EXPORT_SYMBOL(xt_find_match
);
212 /* Find target, grabs ref. Returns ERR_PTR() on error. */
213 struct xt_target
*xt_find_target(int af
, const char *name
, u8 revision
)
218 if (mutex_lock_interruptible(&xt
[af
].mutex
) != 0)
219 return ERR_PTR(-EINTR
);
221 list_for_each_entry(t
, &xt
[af
].target
, list
) {
222 if (strcmp(t
->name
, name
) == 0) {
223 if (t
->revision
== revision
) {
224 if (try_module_get(t
->me
)) {
225 mutex_unlock(&xt
[af
].mutex
);
229 err
= -EPROTOTYPE
; /* Found something. */
232 mutex_unlock(&xt
[af
].mutex
);
235 EXPORT_SYMBOL(xt_find_target
);
237 struct xt_target
*xt_request_find_target(int af
, const char *name
, u8 revision
)
239 struct xt_target
*target
;
241 target
= try_then_request_module(xt_find_target(af
, name
, revision
),
242 "%st_%s", xt_prefix
[af
], name
);
243 if (IS_ERR(target
) || !target
)
247 EXPORT_SYMBOL_GPL(xt_request_find_target
);
249 static int match_revfn(int af
, const char *name
, u8 revision
, int *bestp
)
251 const struct xt_match
*m
;
254 list_for_each_entry(m
, &xt
[af
].match
, list
) {
255 if (strcmp(m
->name
, name
) == 0) {
256 if (m
->revision
> *bestp
)
257 *bestp
= m
->revision
;
258 if (m
->revision
== revision
)
265 static int target_revfn(int af
, const char *name
, u8 revision
, int *bestp
)
267 const struct xt_target
*t
;
270 list_for_each_entry(t
, &xt
[af
].target
, list
) {
271 if (strcmp(t
->name
, name
) == 0) {
272 if (t
->revision
> *bestp
)
273 *bestp
= t
->revision
;
274 if (t
->revision
== revision
)
281 /* Returns true or false (if no such extension at all) */
282 int xt_find_revision(int af
, const char *name
, u8 revision
, int target
,
285 int have_rev
, best
= -1;
287 if (mutex_lock_interruptible(&xt
[af
].mutex
) != 0) {
292 have_rev
= target_revfn(af
, name
, revision
, &best
);
294 have_rev
= match_revfn(af
, name
, revision
, &best
);
295 mutex_unlock(&xt
[af
].mutex
);
297 /* Nothing at all? Return 0 to try loading module. */
305 *err
= -EPROTONOSUPPORT
;
308 EXPORT_SYMBOL_GPL(xt_find_revision
);
310 int xt_check_match(const struct xt_match
*match
, unsigned short family
,
311 unsigned int size
, const char *table
, unsigned int hook_mask
,
312 unsigned short proto
, int inv_proto
)
314 if (XT_ALIGN(match
->matchsize
) != size
) {
315 printk("%s_tables: %s match: invalid size %Zu != %u\n",
316 xt_prefix
[family
], match
->name
,
317 XT_ALIGN(match
->matchsize
), size
);
320 if (match
->table
&& strcmp(match
->table
, table
)) {
321 printk("%s_tables: %s match: only valid in %s table, not %s\n",
322 xt_prefix
[family
], match
->name
, match
->table
, table
);
325 if (match
->hooks
&& (hook_mask
& ~match
->hooks
) != 0) {
326 printk("%s_tables: %s match: bad hook_mask %u/%u\n",
327 xt_prefix
[family
], match
->name
, hook_mask
, match
->hooks
);
330 if (match
->proto
&& (match
->proto
!= proto
|| inv_proto
)) {
331 printk("%s_tables: %s match: only valid for protocol %u\n",
332 xt_prefix
[family
], match
->name
, match
->proto
);
337 EXPORT_SYMBOL_GPL(xt_check_match
);
340 int xt_compat_add_offset(int af
, unsigned int offset
, short delta
)
342 struct compat_delta
*tmp
;
344 tmp
= kmalloc(sizeof(struct compat_delta
), GFP_KERNEL
);
348 tmp
->offset
= offset
;
351 if (xt
[af
].compat_offsets
) {
352 tmp
->next
= xt
[af
].compat_offsets
->next
;
353 xt
[af
].compat_offsets
->next
= tmp
;
355 xt
[af
].compat_offsets
= tmp
;
360 EXPORT_SYMBOL_GPL(xt_compat_add_offset
);
362 void xt_compat_flush_offsets(int af
)
364 struct compat_delta
*tmp
, *next
;
366 if (xt
[af
].compat_offsets
) {
367 for (tmp
= xt
[af
].compat_offsets
; tmp
; tmp
= next
) {
371 xt
[af
].compat_offsets
= NULL
;
374 EXPORT_SYMBOL_GPL(xt_compat_flush_offsets
);
376 short xt_compat_calc_jump(int af
, unsigned int offset
)
378 struct compat_delta
*tmp
;
381 for (tmp
= xt
[af
].compat_offsets
, delta
= 0; tmp
; tmp
= tmp
->next
)
382 if (tmp
->offset
< offset
)
386 EXPORT_SYMBOL_GPL(xt_compat_calc_jump
);
388 int xt_compat_match_offset(const struct xt_match
*match
)
390 u_int16_t csize
= match
->compatsize
? : match
->matchsize
;
391 return XT_ALIGN(match
->matchsize
) - COMPAT_XT_ALIGN(csize
);
393 EXPORT_SYMBOL_GPL(xt_compat_match_offset
);
395 int xt_compat_match_from_user(struct xt_entry_match
*m
, void **dstptr
,
398 const struct xt_match
*match
= m
->u
.kernel
.match
;
399 struct compat_xt_entry_match
*cm
= (struct compat_xt_entry_match
*)m
;
400 int pad
, off
= xt_compat_match_offset(match
);
401 u_int16_t msize
= cm
->u
.user
.match_size
;
404 memcpy(m
, cm
, sizeof(*cm
));
405 if (match
->compat_from_user
)
406 match
->compat_from_user(m
->data
, cm
->data
);
408 memcpy(m
->data
, cm
->data
, msize
- sizeof(*cm
));
409 pad
= XT_ALIGN(match
->matchsize
) - match
->matchsize
;
411 memset(m
->data
+ match
->matchsize
, 0, pad
);
414 m
->u
.user
.match_size
= msize
;
420 EXPORT_SYMBOL_GPL(xt_compat_match_from_user
);
422 int xt_compat_match_to_user(struct xt_entry_match
*m
, void __user
**dstptr
,
425 const struct xt_match
*match
= m
->u
.kernel
.match
;
426 struct compat_xt_entry_match __user
*cm
= *dstptr
;
427 int off
= xt_compat_match_offset(match
);
428 u_int16_t msize
= m
->u
.user
.match_size
- off
;
430 if (copy_to_user(cm
, m
, sizeof(*cm
)) ||
431 put_user(msize
, &cm
->u
.user
.match_size
) ||
432 copy_to_user(cm
->u
.user
.name
, m
->u
.kernel
.match
->name
,
433 strlen(m
->u
.kernel
.match
->name
) + 1))
436 if (match
->compat_to_user
) {
437 if (match
->compat_to_user((void __user
*)cm
->data
, m
->data
))
440 if (copy_to_user(cm
->data
, m
->data
, msize
- sizeof(*cm
)))
448 EXPORT_SYMBOL_GPL(xt_compat_match_to_user
);
449 #endif /* CONFIG_COMPAT */
451 int xt_check_target(const struct xt_target
*target
, unsigned short family
,
452 unsigned int size
, const char *table
, unsigned int hook_mask
,
453 unsigned short proto
, int inv_proto
)
455 if (XT_ALIGN(target
->targetsize
) != size
) {
456 printk("%s_tables: %s target: invalid size %Zu != %u\n",
457 xt_prefix
[family
], target
->name
,
458 XT_ALIGN(target
->targetsize
), size
);
461 if (target
->table
&& strcmp(target
->table
, table
)) {
462 printk("%s_tables: %s target: only valid in %s table, not %s\n",
463 xt_prefix
[family
], target
->name
, target
->table
, table
);
466 if (target
->hooks
&& (hook_mask
& ~target
->hooks
) != 0) {
467 printk("%s_tables: %s target: bad hook_mask %u/%u\n",
468 xt_prefix
[family
], target
->name
, hook_mask
,
472 if (target
->proto
&& (target
->proto
!= proto
|| inv_proto
)) {
473 printk("%s_tables: %s target: only valid for protocol %u\n",
474 xt_prefix
[family
], target
->name
, target
->proto
);
479 EXPORT_SYMBOL_GPL(xt_check_target
);
482 int xt_compat_target_offset(const struct xt_target
*target
)
484 u_int16_t csize
= target
->compatsize
? : target
->targetsize
;
485 return XT_ALIGN(target
->targetsize
) - COMPAT_XT_ALIGN(csize
);
487 EXPORT_SYMBOL_GPL(xt_compat_target_offset
);
489 void xt_compat_target_from_user(struct xt_entry_target
*t
, void **dstptr
,
492 const struct xt_target
*target
= t
->u
.kernel
.target
;
493 struct compat_xt_entry_target
*ct
= (struct compat_xt_entry_target
*)t
;
494 int pad
, off
= xt_compat_target_offset(target
);
495 u_int16_t tsize
= ct
->u
.user
.target_size
;
498 memcpy(t
, ct
, sizeof(*ct
));
499 if (target
->compat_from_user
)
500 target
->compat_from_user(t
->data
, ct
->data
);
502 memcpy(t
->data
, ct
->data
, tsize
- sizeof(*ct
));
503 pad
= XT_ALIGN(target
->targetsize
) - target
->targetsize
;
505 memset(t
->data
+ target
->targetsize
, 0, pad
);
508 t
->u
.user
.target_size
= tsize
;
513 EXPORT_SYMBOL_GPL(xt_compat_target_from_user
);
515 int xt_compat_target_to_user(struct xt_entry_target
*t
, void __user
**dstptr
,
518 const struct xt_target
*target
= t
->u
.kernel
.target
;
519 struct compat_xt_entry_target __user
*ct
= *dstptr
;
520 int off
= xt_compat_target_offset(target
);
521 u_int16_t tsize
= t
->u
.user
.target_size
- off
;
523 if (copy_to_user(ct
, t
, sizeof(*ct
)) ||
524 put_user(tsize
, &ct
->u
.user
.target_size
) ||
525 copy_to_user(ct
->u
.user
.name
, t
->u
.kernel
.target
->name
,
526 strlen(t
->u
.kernel
.target
->name
) + 1))
529 if (target
->compat_to_user
) {
530 if (target
->compat_to_user((void __user
*)ct
->data
, t
->data
))
533 if (copy_to_user(ct
->data
, t
->data
, tsize
- sizeof(*ct
)))
541 EXPORT_SYMBOL_GPL(xt_compat_target_to_user
);
544 struct xt_table_info
*xt_alloc_table_info(unsigned int size
)
546 struct xt_table_info
*newinfo
;
549 /* Pedantry: prevent them from hitting BUG() in vmalloc.c --RR */
550 if ((SMP_ALIGN(size
) >> PAGE_SHIFT
) + 2 > num_physpages
)
553 newinfo
= kzalloc(XT_TABLE_INFO_SZ
, GFP_KERNEL
);
557 newinfo
->size
= size
;
559 for_each_possible_cpu(cpu
) {
560 if (size
<= PAGE_SIZE
)
561 newinfo
->entries
[cpu
] = kmalloc_node(size
,
565 newinfo
->entries
[cpu
] = vmalloc_node(size
,
568 if (newinfo
->entries
[cpu
] == NULL
) {
569 xt_free_table_info(newinfo
);
576 EXPORT_SYMBOL(xt_alloc_table_info
);
578 void xt_free_table_info(struct xt_table_info
*info
)
582 for_each_possible_cpu(cpu
) {
583 if (info
->size
<= PAGE_SIZE
)
584 kfree(info
->entries
[cpu
]);
586 vfree(info
->entries
[cpu
]);
590 EXPORT_SYMBOL(xt_free_table_info
);
592 /* Find table by name, grabs mutex & ref. Returns ERR_PTR() on error. */
593 struct xt_table
*xt_find_table_lock(struct net
*net
, int af
, const char *name
)
597 if (mutex_lock_interruptible(&xt
[af
].mutex
) != 0)
598 return ERR_PTR(-EINTR
);
600 list_for_each_entry(t
, &net
->xt
.tables
[af
], list
)
601 if (strcmp(t
->name
, name
) == 0 && try_module_get(t
->me
))
603 mutex_unlock(&xt
[af
].mutex
);
606 EXPORT_SYMBOL_GPL(xt_find_table_lock
);
608 void xt_table_unlock(struct xt_table
*table
)
610 mutex_unlock(&xt
[table
->af
].mutex
);
612 EXPORT_SYMBOL_GPL(xt_table_unlock
);
615 void xt_compat_lock(int af
)
617 mutex_lock(&xt
[af
].compat_mutex
);
619 EXPORT_SYMBOL_GPL(xt_compat_lock
);
621 void xt_compat_unlock(int af
)
623 mutex_unlock(&xt
[af
].compat_mutex
);
625 EXPORT_SYMBOL_GPL(xt_compat_unlock
);
628 struct xt_table_info
*
629 xt_replace_table(struct xt_table
*table
,
630 unsigned int num_counters
,
631 struct xt_table_info
*newinfo
,
634 struct xt_table_info
*oldinfo
, *private;
636 /* Do the substitution. */
637 write_lock_bh(&table
->lock
);
638 private = table
->private;
639 /* Check inside lock: is the old number correct? */
640 if (num_counters
!= private->number
) {
641 duprintf("num_counters != table->private->number (%u/%u)\n",
642 num_counters
, private->number
);
643 write_unlock_bh(&table
->lock
);
648 table
->private = newinfo
;
649 newinfo
->initial_entries
= oldinfo
->initial_entries
;
650 write_unlock_bh(&table
->lock
);
654 EXPORT_SYMBOL_GPL(xt_replace_table
);
656 struct xt_table
*xt_register_table(struct net
*net
, struct xt_table
*table
,
657 struct xt_table_info
*bootstrap
,
658 struct xt_table_info
*newinfo
)
661 struct xt_table_info
*private;
664 /* Don't add one object to multiple lists. */
665 table
= kmemdup(table
, sizeof(struct xt_table
), GFP_KERNEL
);
671 ret
= mutex_lock_interruptible(&xt
[table
->af
].mutex
);
675 /* Don't autoload: we'd eat our tail... */
676 list_for_each_entry(t
, &net
->xt
.tables
[table
->af
], list
) {
677 if (strcmp(t
->name
, table
->name
) == 0) {
683 /* Simplifies replace_table code. */
684 table
->private = bootstrap
;
685 rwlock_init(&table
->lock
);
686 if (!xt_replace_table(table
, 0, newinfo
, &ret
))
689 private = table
->private;
690 duprintf("table->private->number = %u\n", private->number
);
692 /* save number of initial entries */
693 private->initial_entries
= private->number
;
695 list_add(&table
->list
, &net
->xt
.tables
[table
->af
]);
696 mutex_unlock(&xt
[table
->af
].mutex
);
700 mutex_unlock(&xt
[table
->af
].mutex
);
706 EXPORT_SYMBOL_GPL(xt_register_table
);
708 void *xt_unregister_table(struct xt_table
*table
)
710 struct xt_table_info
*private;
712 mutex_lock(&xt
[table
->af
].mutex
);
713 private = table
->private;
714 list_del(&table
->list
);
715 mutex_unlock(&xt
[table
->af
].mutex
);
720 EXPORT_SYMBOL_GPL(xt_unregister_table
);
722 #ifdef CONFIG_PROC_FS
723 struct xt_names_priv
{
724 struct seq_net_private p
;
727 static void *xt_table_seq_start(struct seq_file
*seq
, loff_t
*pos
)
729 struct xt_names_priv
*priv
= seq
->private;
730 struct net
*net
= seq_file_net(seq
);
733 mutex_lock(&xt
[af
].mutex
);
734 return seq_list_start(&net
->xt
.tables
[af
], *pos
);
737 static void *xt_table_seq_next(struct seq_file
*seq
, void *v
, loff_t
*pos
)
739 struct xt_names_priv
*priv
= seq
->private;
740 struct net
*net
= seq_file_net(seq
);
743 return seq_list_next(v
, &net
->xt
.tables
[af
], pos
);
746 static void xt_table_seq_stop(struct seq_file
*seq
, void *v
)
748 struct xt_names_priv
*priv
= seq
->private;
751 mutex_unlock(&xt
[af
].mutex
);
754 static int xt_table_seq_show(struct seq_file
*seq
, void *v
)
756 struct xt_table
*table
= list_entry(v
, struct xt_table
, list
);
758 if (strlen(table
->name
))
759 return seq_printf(seq
, "%s\n", table
->name
);
764 static const struct seq_operations xt_table_seq_ops
= {
765 .start
= xt_table_seq_start
,
766 .next
= xt_table_seq_next
,
767 .stop
= xt_table_seq_stop
,
768 .show
= xt_table_seq_show
,
771 static int xt_table_open(struct inode
*inode
, struct file
*file
)
774 struct xt_names_priv
*priv
;
776 ret
= seq_open_net(inode
, file
, &xt_table_seq_ops
,
777 sizeof(struct xt_names_priv
));
779 priv
= ((struct seq_file
*)file
->private_data
)->private;
780 priv
->af
= (unsigned long)PDE(inode
)->data
;
785 static const struct file_operations xt_table_ops
= {
786 .owner
= THIS_MODULE
,
787 .open
= xt_table_open
,
790 .release
= seq_release_net
,
793 static void *xt_match_seq_start(struct seq_file
*seq
, loff_t
*pos
)
795 struct proc_dir_entry
*pde
= (struct proc_dir_entry
*)seq
->private;
796 u_int16_t af
= (unsigned long)pde
->data
;
798 mutex_lock(&xt
[af
].mutex
);
799 return seq_list_start(&xt
[af
].match
, *pos
);
802 static void *xt_match_seq_next(struct seq_file
*seq
, void *v
, loff_t
*pos
)
804 struct proc_dir_entry
*pde
= (struct proc_dir_entry
*)seq
->private;
805 u_int16_t af
= (unsigned long)pde
->data
;
807 return seq_list_next(v
, &xt
[af
].match
, pos
);
810 static void xt_match_seq_stop(struct seq_file
*seq
, void *v
)
812 struct proc_dir_entry
*pde
= seq
->private;
813 u_int16_t af
= (unsigned long)pde
->data
;
815 mutex_unlock(&xt
[af
].mutex
);
818 static int xt_match_seq_show(struct seq_file
*seq
, void *v
)
820 struct xt_match
*match
= list_entry(v
, struct xt_match
, list
);
822 if (strlen(match
->name
))
823 return seq_printf(seq
, "%s\n", match
->name
);
828 static const struct seq_operations xt_match_seq_ops
= {
829 .start
= xt_match_seq_start
,
830 .next
= xt_match_seq_next
,
831 .stop
= xt_match_seq_stop
,
832 .show
= xt_match_seq_show
,
835 static int xt_match_open(struct inode
*inode
, struct file
*file
)
839 ret
= seq_open(file
, &xt_match_seq_ops
);
841 struct seq_file
*seq
= file
->private_data
;
843 seq
->private = PDE(inode
);
848 static const struct file_operations xt_match_ops
= {
849 .owner
= THIS_MODULE
,
850 .open
= xt_match_open
,
853 .release
= seq_release
,
856 static void *xt_target_seq_start(struct seq_file
*seq
, loff_t
*pos
)
858 struct proc_dir_entry
*pde
= (struct proc_dir_entry
*)seq
->private;
859 u_int16_t af
= (unsigned long)pde
->data
;
861 mutex_lock(&xt
[af
].mutex
);
862 return seq_list_start(&xt
[af
].target
, *pos
);
865 static void *xt_target_seq_next(struct seq_file
*seq
, void *v
, loff_t
*pos
)
867 struct proc_dir_entry
*pde
= (struct proc_dir_entry
*)seq
->private;
868 u_int16_t af
= (unsigned long)pde
->data
;
870 return seq_list_next(v
, &xt
[af
].target
, pos
);
873 static void xt_target_seq_stop(struct seq_file
*seq
, void *v
)
875 struct proc_dir_entry
*pde
= seq
->private;
876 u_int16_t af
= (unsigned long)pde
->data
;
878 mutex_unlock(&xt
[af
].mutex
);
881 static int xt_target_seq_show(struct seq_file
*seq
, void *v
)
883 struct xt_target
*target
= list_entry(v
, struct xt_target
, list
);
885 if (strlen(target
->name
))
886 return seq_printf(seq
, "%s\n", target
->name
);
891 static const struct seq_operations xt_target_seq_ops
= {
892 .start
= xt_target_seq_start
,
893 .next
= xt_target_seq_next
,
894 .stop
= xt_target_seq_stop
,
895 .show
= xt_target_seq_show
,
898 static int xt_target_open(struct inode
*inode
, struct file
*file
)
902 ret
= seq_open(file
, &xt_target_seq_ops
);
904 struct seq_file
*seq
= file
->private_data
;
906 seq
->private = PDE(inode
);
911 static const struct file_operations xt_target_ops
= {
912 .owner
= THIS_MODULE
,
913 .open
= xt_target_open
,
916 .release
= seq_release
,
919 #define FORMAT_TABLES "_tables_names"
920 #define FORMAT_MATCHES "_tables_matches"
921 #define FORMAT_TARGETS "_tables_targets"
923 #endif /* CONFIG_PROC_FS */
925 int xt_proto_init(struct net
*net
, int af
)
927 #ifdef CONFIG_PROC_FS
928 char buf
[XT_FUNCTION_MAXNAMELEN
];
929 struct proc_dir_entry
*proc
;
936 #ifdef CONFIG_PROC_FS
937 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
938 strlcat(buf
, FORMAT_TABLES
, sizeof(buf
));
939 proc
= proc_create_data(buf
, 0440, net
->proc_net
, &xt_table_ops
,
940 (void *)(unsigned long)af
);
944 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
945 strlcat(buf
, FORMAT_MATCHES
, sizeof(buf
));
946 proc
= proc_create_data(buf
, 0440, net
->proc_net
, &xt_match_ops
,
947 (void *)(unsigned long)af
);
949 goto out_remove_tables
;
951 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
952 strlcat(buf
, FORMAT_TARGETS
, sizeof(buf
));
953 proc
= proc_create_data(buf
, 0440, net
->proc_net
, &xt_target_ops
,
954 (void *)(unsigned long)af
);
956 goto out_remove_matches
;
961 #ifdef CONFIG_PROC_FS
963 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
964 strlcat(buf
, FORMAT_MATCHES
, sizeof(buf
));
965 proc_net_remove(net
, buf
);
968 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
969 strlcat(buf
, FORMAT_TABLES
, sizeof(buf
));
970 proc_net_remove(net
, buf
);
975 EXPORT_SYMBOL_GPL(xt_proto_init
);
977 void xt_proto_fini(struct net
*net
, int af
)
979 #ifdef CONFIG_PROC_FS
980 char buf
[XT_FUNCTION_MAXNAMELEN
];
982 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
983 strlcat(buf
, FORMAT_TABLES
, sizeof(buf
));
984 proc_net_remove(net
, buf
);
986 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
987 strlcat(buf
, FORMAT_TARGETS
, sizeof(buf
));
988 proc_net_remove(net
, buf
);
990 strlcpy(buf
, xt_prefix
[af
], sizeof(buf
));
991 strlcat(buf
, FORMAT_MATCHES
, sizeof(buf
));
992 proc_net_remove(net
, buf
);
993 #endif /*CONFIG_PROC_FS*/
995 EXPORT_SYMBOL_GPL(xt_proto_fini
);
997 static int __net_init
xt_net_init(struct net
*net
)
1001 for (i
= 0; i
< NPROTO
; i
++)
1002 INIT_LIST_HEAD(&net
->xt
.tables
[i
]);
1006 static struct pernet_operations xt_net_ops
= {
1007 .init
= xt_net_init
,
1010 static int __init
xt_init(void)
1014 xt
= kmalloc(sizeof(struct xt_af
) * NPROTO
, GFP_KERNEL
);
1018 for (i
= 0; i
< NPROTO
; i
++) {
1019 mutex_init(&xt
[i
].mutex
);
1020 #ifdef CONFIG_COMPAT
1021 mutex_init(&xt
[i
].compat_mutex
);
1022 xt
[i
].compat_offsets
= NULL
;
1024 INIT_LIST_HEAD(&xt
[i
].target
);
1025 INIT_LIST_HEAD(&xt
[i
].match
);
1027 rv
= register_pernet_subsys(&xt_net_ops
);
1033 static void __exit
xt_fini(void)
1035 unregister_pernet_subsys(&xt_net_ops
);
1039 module_init(xt_init
);
1040 module_exit(xt_fini
);