2 * netsniff-ng - the packet sniffing beast
3 * By Daniel Borkmann <daniel@netsniff-ng.org>
4 * Copyright 2009 - 2012 Daniel Borkmann.
5 * Copyright 2009, 2010 Emmanuel Roullit.
6 * Copyright 1990-1996 The Regents of the University of
7 * California. All rights reserved. (3-clause BSD license)
8 * Subject to the GPL, version 2.
13 #include <arpa/inet.h>
14 #include <sys/types.h>
23 /* This is a bug in libpcap, they actually use 'unsigned long' instead
25 #define EXTRACT_SHORT(packet) \
26 ((unsigned short) ntohs(*(unsigned short *) packet))
27 #define EXTRACT_LONG(packet) \
28 (ntohl(*(unsigned long *) packet))
30 # define BPF_MEMWORDS 16
33 #define BPF_LD_B (BPF_LD | BPF_B)
34 #define BPF_LD_H (BPF_LD | BPF_H)
35 #define BPF_LD_W (BPF_LD | BPF_W)
36 #define BPF_LDX_B (BPF_LDX | BPF_B)
37 #define BPF_LDX_W (BPF_LDX | BPF_W)
38 #define BPF_JMP_JA (BPF_JMP | BPF_JA)
39 #define BPF_JMP_JEQ (BPF_JMP | BPF_JEQ)
40 #define BPF_JMP_JGT (BPF_JMP | BPF_JGT)
41 #define BPF_JMP_JGE (BPF_JMP | BPF_JGE)
42 #define BPF_JMP_JSET (BPF_JMP | BPF_JSET)
43 #define BPF_ALU_ADD (BPF_ALU | BPF_ADD)
44 #define BPF_ALU_SUB (BPF_ALU | BPF_SUB)
45 #define BPF_ALU_MUL (BPF_ALU | BPF_MUL)
46 #define BPF_ALU_DIV (BPF_ALU | BPF_DIV)
47 #define BPF_ALU_MOD (BPF_ALU | BPF_MOD)
48 #define BPF_ALU_NEG (BPF_ALU | BPF_NEG)
49 #define BPF_ALU_AND (BPF_ALU | BPF_AND)
50 #define BPF_ALU_OR (BPF_ALU | BPF_OR)
51 #define BPF_ALU_XOR (BPF_ALU | BPF_XOR)
52 #define BPF_ALU_LSH (BPF_ALU | BPF_LSH)
53 #define BPF_ALU_RSH (BPF_ALU | BPF_RSH)
54 #define BPF_MISC_TAX (BPF_MISC | BPF_TAX)
55 #define BPF_MISC_TXA (BPF_MISC | BPF_TXA)
57 static const char *op_table
[] = {
66 [BPF_JMP_JEQ
] = "jeq",
67 [BPF_JMP_JGT
] = "jgt",
68 [BPF_JMP_JGE
] = "jge",
69 [BPF_JMP_JSET
] = "jset",
70 [BPF_ALU_ADD
] = "add",
71 [BPF_ALU_SUB
] = "sub",
72 [BPF_ALU_MUL
] = "mul",
73 [BPF_ALU_DIV
] = "div",
74 [BPF_ALU_MOD
] = "mod",
75 [BPF_ALU_NEG
] = "neg",
76 [BPF_ALU_AND
] = "and",
78 [BPF_ALU_XOR
] = "xor",
79 [BPF_ALU_LSH
] = "lsh",
80 [BPF_ALU_RSH
] = "rsh",
82 [BPF_MISC_TAX
] = "tax",
83 [BPF_MISC_TXA
] = "txa",
86 void bpf_dump_op_table(void)
89 for (i
= 0; i
< array_size(op_table
); ++i
) {
91 printf("%s\n", op_table
[i
]);
95 static const char *bpf_dump_linux_k(uint32_t k
)
100 /* Linux specific arguments */
101 case (SKF_AD_OFF
+ SKF_AD_PROTOCOL
):
103 case (SKF_AD_OFF
+ SKF_AD_PKTTYPE
):
105 case (SKF_AD_OFF
+ SKF_AD_IFINDEX
):
107 case (SKF_AD_OFF
+ SKF_AD_NLATTR
):
109 case (SKF_AD_OFF
+ SKF_AD_NLATTR_NEST
):
111 case (SKF_AD_OFF
+ SKF_AD_MARK
):
113 case (SKF_AD_OFF
+ SKF_AD_QUEUE
):
115 case (SKF_AD_OFF
+ SKF_AD_HATYPE
):
117 case (SKF_AD_OFF
+ SKF_AD_RXHASH
):
119 case (SKF_AD_OFF
+ SKF_AD_CPU
):
121 case (SKF_AD_OFF
+ SKF_AD_VLAN_TAG
):
123 case (SKF_AD_OFF
+ SKF_AD_VLAN_TAG_PRESENT
):
128 static char *bpf_dump(const struct sock_filter bpf
, int n
)
131 const char *fmt
, *op
;
132 static char image
[256];
142 case BPF_RET
| BPF_K
:
143 op
= op_table
[BPF_RET
];
146 case BPF_RET
| BPF_A
:
147 op
= op_table
[BPF_RET
];
150 case BPF_LD_W
| BPF_ABS
:
151 op
= op_table
[BPF_LD_W
];
152 fmt
= bpf_dump_linux_k(bpf
.k
);
154 case BPF_LD_H
| BPF_ABS
:
155 op
= op_table
[BPF_LD_H
];
156 fmt
= bpf_dump_linux_k(bpf
.k
);
158 case BPF_LD_B
| BPF_ABS
:
159 op
= op_table
[BPF_LD_B
];
160 fmt
= bpf_dump_linux_k(bpf
.k
);
162 case BPF_LD_W
| BPF_LEN
:
163 op
= op_table
[BPF_LD_W
];
166 case BPF_LD_W
| BPF_IND
:
167 op
= op_table
[BPF_LD_W
];
170 case BPF_LD_H
| BPF_IND
:
171 op
= op_table
[BPF_LD_H
];
174 case BPF_LD_B
| BPF_IND
:
175 op
= op_table
[BPF_LD_B
];
178 case BPF_LD
| BPF_IMM
:
179 op
= op_table
[BPF_LD_W
];
182 case BPF_LDX
| BPF_IMM
:
183 op
= op_table
[BPF_LDX
];
186 case BPF_LDX_B
| BPF_MSH
:
187 op
= op_table
[BPF_LDX_B
];
188 fmt
= "4*([%d]&0xf)";
190 case BPF_LD
| BPF_MEM
:
191 op
= op_table
[BPF_LD_W
];
194 case BPF_LDX
| BPF_MEM
:
195 op
= op_table
[BPF_LDX
];
199 op
= op_table
[BPF_ST
];
203 op
= op_table
[BPF_STX
];
207 op
= op_table
[BPF_JMP_JA
];
211 case BPF_JMP_JGT
| BPF_K
:
212 op
= op_table
[BPF_JMP_JGT
];
215 case BPF_JMP_JGE
| BPF_K
:
216 op
= op_table
[BPF_JMP_JGE
];
219 case BPF_JMP_JEQ
| BPF_K
:
220 op
= op_table
[BPF_JMP_JEQ
];
223 case BPF_JMP_JSET
| BPF_K
:
224 op
= op_table
[BPF_JMP_JSET
];
227 case BPF_JMP_JGT
| BPF_X
:
228 op
= op_table
[BPF_JMP_JGT
];
231 case BPF_JMP_JGE
| BPF_X
:
232 op
= op_table
[BPF_JMP_JGE
];
235 case BPF_JMP_JEQ
| BPF_X
:
236 op
= op_table
[BPF_JMP_JEQ
];
239 case BPF_JMP_JSET
| BPF_X
:
240 op
= op_table
[BPF_JMP_JSET
];
243 case BPF_ALU_ADD
| BPF_X
:
244 op
= op_table
[BPF_ALU_ADD
];
247 case BPF_ALU_SUB
| BPF_X
:
248 op
= op_table
[BPF_ALU_SUB
];
251 case BPF_ALU_MUL
| BPF_X
:
252 op
= op_table
[BPF_ALU_MUL
];
255 case BPF_ALU_DIV
| BPF_X
:
256 op
= op_table
[BPF_ALU_DIV
];
259 case BPF_ALU_MOD
| BPF_X
:
260 op
= op_table
[BPF_ALU_MOD
];
263 case BPF_ALU_AND
| BPF_X
:
264 op
= op_table
[BPF_ALU_AND
];
267 case BPF_ALU_OR
| BPF_X
:
268 op
= op_table
[BPF_ALU_OR
];
271 case BPF_ALU_XOR
| BPF_X
:
272 op
= op_table
[BPF_ALU_XOR
];
275 case BPF_ALU_LSH
| BPF_X
:
276 op
= op_table
[BPF_ALU_LSH
];
279 case BPF_ALU_RSH
| BPF_X
:
280 op
= op_table
[BPF_ALU_RSH
];
283 case BPF_ALU_ADD
| BPF_K
:
284 op
= op_table
[BPF_ALU_ADD
];
287 case BPF_ALU_SUB
| BPF_K
:
288 op
= op_table
[BPF_ALU_SUB
];
291 case BPF_ALU_MUL
| BPF_K
:
292 op
= op_table
[BPF_ALU_MUL
];
295 case BPF_ALU_DIV
| BPF_K
:
296 op
= op_table
[BPF_ALU_DIV
];
299 case BPF_ALU_MOD
| BPF_K
:
300 op
= op_table
[BPF_ALU_MOD
];
303 case BPF_ALU_AND
| BPF_K
:
304 op
= op_table
[BPF_ALU_AND
];
307 case BPF_ALU_OR
| BPF_K
:
308 op
= op_table
[BPF_ALU_OR
];
311 case BPF_ALU_XOR
| BPF_K
:
312 op
= op_table
[BPF_ALU_XOR
];
315 case BPF_ALU_LSH
| BPF_K
:
316 op
= op_table
[BPF_ALU_LSH
];
319 case BPF_ALU_RSH
| BPF_K
:
320 op
= op_table
[BPF_ALU_RSH
];
324 op
= op_table
[BPF_ALU_NEG
];
328 op
= op_table
[BPF_MISC_TAX
];
332 op
= op_table
[BPF_MISC_TXA
];
337 slprintf(operand
, sizeof(operand
), fmt
, v
);
338 slprintf(image
, sizeof(image
),
339 (BPF_CLASS(bpf
.code
) == BPF_JMP
&&
340 BPF_OP(bpf
.code
) != BPF_JA
) ?
341 " L%d: %s %s, L%d, L%d" : " L%d: %s %s",
342 n
, op
, operand
, n
+ 1 + bpf
.jt
, n
+ 1 + bpf
.jf
);
347 void bpf_dump_all(struct sock_fprog
*bpf
)
350 for (i
= 0; i
< bpf
->len
; ++i
)
351 printf("%s\n", bpf_dump(bpf
->filter
[i
], i
));
354 void bpf_attach_to_sock(int sock
, struct sock_fprog
*bpf
)
358 if (bpf
->filter
[0].code
== BPF_RET
&&
359 bpf
->filter
[0].k
== 0xFFFFFFFF)
362 ret
= setsockopt(sock
, SOL_SOCKET
, SO_ATTACH_FILTER
,
365 panic("Cannot attach filter to socket!\n");
368 void bpf_detach_from_sock(int sock
)
372 ret
= setsockopt(sock
, SOL_SOCKET
, SO_DETACH_FILTER
,
373 &empty
, sizeof(empty
));
375 panic("Cannot detach filter from socket!\n");
378 void enable_kernel_bpf_jit_compiler(void)
382 char *file
= "/proc/sys/net/core/bpf_jit_enable";
384 fd
= open(file
, O_WRONLY
);
388 ret
= write(fd
, "1", strlen("1"));
395 int bpf_validate(const struct sock_fprog
*bpf
)
398 const struct sock_filter
*p
;
405 for (i
= 0; i
< bpf
->len
; ++i
) {
407 switch (BPF_CLASS(p
->code
)) {
408 /* Check that memory operations use valid addresses. */
411 switch (BPF_MODE(p
->code
)) {
417 /* There's no maximum packet data size
418 * in userland. The runtime packet length
423 if (p
->k
>= BPF_MEMWORDS
)
434 if (p
->k
>= BPF_MEMWORDS
)
438 switch (BPF_OP(p
->code
)) {
451 /* Check for constant division by 0 (undefined
454 if (BPF_RVAL(p
->code
) == BPF_K
&& p
->k
== 0)
462 /* Check that jumps are within the code block,
463 * and that unconditional branches don't go
464 * backwards as a result of an overflow.
465 * Unconditional branches have a 32-bit offset,
466 * so they could overflow; we check to make
467 * sure they don't. Conditional branches have
468 * an 8-bit offset, and the from address is <=
469 * BPF_MAXINSNS, and we assume that BPF_MAXINSNS
470 * is sufficiently small that adding 255 to it
473 * We know that len is <= BPF_MAXINSNS, and we
474 * assume that BPF_MAXINSNS is < the maximum size
475 * of a u_int, so that i + 1 doesn't overflow.
477 * For userland, we don't know that the from
478 * or len are <= BPF_MAXINSNS, but we know that
479 * from <= len, and, except on a 64-bit system,
480 * it's unlikely that len, if it truly reflects
481 * the size of the program we've been handed,
482 * will be anywhere near the maximum size of
483 * a u_int. We also don't check for backward
484 * branches, as we currently support them in
485 * userland for the protochain operation.
488 switch (BPF_OP(p
->code
)) {
490 if (from
+ p
->k
>= bpf
->len
)
497 if (from
+ p
->jt
>= bpf
->len
||
498 from
+ p
->jf
>= bpf
->len
)
514 return BPF_CLASS(bpf
->filter
[bpf
->len
- 1].code
) == BPF_RET
;
517 uint32_t bpf_run_filter(const struct sock_fprog
* fcode
, uint8_t * packet
,
520 /* XXX: caplen == len */
523 struct sock_filter
*bpf
;
524 int32_t mem
[BPF_MEMWORDS
];
526 if (fcode
== NULL
|| fcode
->filter
== NULL
|| fcode
->len
== 0)
539 case BPF_RET
| BPF_K
:
540 return (uint32_t) bpf
->k
;
541 case BPF_RET
| BPF_A
:
543 case BPF_LD_W
| BPF_ABS
:
544 /* No Linux extensions supported here! */
546 if (k
+ sizeof(int32_t) > plen
)
548 A
= EXTRACT_LONG(&packet
[k
]);
550 case BPF_LD_H
| BPF_ABS
:
551 /* No Linux extensions supported here! */
553 if (k
+ sizeof(short) > plen
)
555 A
= EXTRACT_SHORT(&packet
[k
]);
557 case BPF_LD_B
| BPF_ABS
:
558 /* No Linux extensions supported here! */
564 case BPF_LD_W
| BPF_LEN
:
567 case BPF_LDX_W
| BPF_LEN
:
570 case BPF_LD_W
| BPF_IND
:
572 if (k
+ sizeof(int32_t) > plen
)
574 A
= EXTRACT_LONG(&packet
[k
]);
576 case BPF_LD_H
| BPF_IND
:
578 if (k
+ sizeof(short) > plen
)
580 A
= EXTRACT_SHORT(&packet
[k
]);
582 case BPF_LD_B
| BPF_IND
:
588 case BPF_LDX_B
| BPF_MSH
:
592 X
= (packet
[bpf
->k
] & 0xf) << 2;
594 case BPF_LD
| BPF_IMM
:
597 case BPF_LDX
| BPF_IMM
:
600 case BPF_LD
| BPF_MEM
:
603 case BPF_LDX
| BPF_MEM
:
615 case BPF_JMP_JGT
| BPF_K
:
616 bpf
+= (A
> bpf
->k
) ? bpf
->jt
: bpf
->jf
;
618 case BPF_JMP_JGE
| BPF_K
:
619 bpf
+= (A
>= bpf
->k
) ? bpf
->jt
: bpf
->jf
;
621 case BPF_JMP_JEQ
| BPF_K
:
622 bpf
+= (A
== bpf
->k
) ? bpf
->jt
: bpf
->jf
;
624 case BPF_JMP_JSET
| BPF_K
:
625 bpf
+= (A
& bpf
->k
) ? bpf
->jt
: bpf
->jf
;
627 case BPF_JMP_JGT
| BPF_X
:
628 bpf
+= (A
> X
) ? bpf
->jt
: bpf
->jf
;
630 case BPF_JMP_JGE
| BPF_X
:
631 bpf
+= (A
>= X
) ? bpf
->jt
: bpf
->jf
;
633 case BPF_JMP_JEQ
| BPF_X
:
634 bpf
+= (A
== X
) ? bpf
->jt
: bpf
->jf
;
636 case BPF_JMP_JSET
| BPF_X
:
637 bpf
+= (A
& X
) ? bpf
->jt
: bpf
->jf
;
639 case BPF_ALU_ADD
| BPF_X
:
642 case BPF_ALU_SUB
| BPF_X
:
645 case BPF_ALU_MUL
| BPF_X
:
648 case BPF_ALU_DIV
| BPF_X
:
653 case BPF_ALU_MOD
| BPF_X
:
658 case BPF_ALU_AND
| BPF_X
:
661 case BPF_ALU_OR
| BPF_X
:
664 case BPF_ALU_XOR
| BPF_X
:
667 case BPF_ALU_LSH
| BPF_X
:
670 case BPF_ALU_RSH
| BPF_X
:
673 case BPF_ALU_ADD
| BPF_K
:
676 case BPF_ALU_SUB
| BPF_K
:
679 case BPF_ALU_MUL
| BPF_K
:
682 case BPF_ALU_DIV
| BPF_K
:
685 case BPF_ALU_MOD
| BPF_K
:
688 case BPF_ALU_AND
| BPF_K
:
691 case BPF_ALU_OR
| BPF_K
:
694 case BPF_ALU_XOR
| BPF_K
:
697 case BPF_ALU_LSH
| BPF_K
:
700 case BPF_ALU_RSH
| BPF_K
:
716 void bpf_parse_rules(char *rulefile
, struct sock_fprog
*bpf
)
720 struct sock_filter sf_single
= { 0x06, 0, 0, 0xFFFFFFFF };
723 if (rulefile
== NULL
) {
725 bpf
->filter
= xmalloc(sizeof(sf_single
));
726 fmemcpy(&bpf
->filter
[0], &sf_single
, sizeof(sf_single
));
730 fp
= fopen(rulefile
, "r");
732 panic("Cannot read BPF rule file!\n");
734 fmemset(buff
, 0, sizeof(buff
));
735 while (fgets(buff
, sizeof(buff
), fp
) != NULL
) {
736 buff
[sizeof(buff
) - 1] = 0;
737 if (buff
[0] != '{') {
738 fmemset(buff
, 0, sizeof(buff
));
742 fmemset(&sf_single
, 0, sizeof(sf_single
));
743 ret
= sscanf(buff
, "{ 0x%x, %u, %u, 0x%08x },",
744 (unsigned int *) &sf_single
.code
,
745 (unsigned int *) &sf_single
.jt
,
746 (unsigned int *) &sf_single
.jf
,
747 (unsigned int *) &sf_single
.k
);
749 panic("BPF syntax error!\n");
752 bpf
->filter
= xrealloc(bpf
->filter
, 1,
753 bpf
->len
* sizeof(sf_single
));
755 fmemcpy(&bpf
->filter
[bpf
->len
- 1], &sf_single
,
757 fmemset(buff
, 0, sizeof(buff
));
762 if (bpf_validate(bpf
) == 0)
763 panic("This is not a valid BPF program!\n");