too long ago again
[lartc.git] / 2.4routing.sgml
blob5749e61403810df4db296a5e92be0c0c8fc12499
1 <!doctype linuxdoc system>
3 <!-- $Id$
4 -->
6 <article>
8 <!-- Title information -->
10 <title>Linux Advanced Routing &amp; Traffic Control HOWTO
11 <author>Netherlabs BV (bert hubert &lt;bert.hubert@netherlabs.nl&gt;)&nl;
12 Gregory Maxwell &lt;greg@linuxpower.cx&gt; &nl;
13 Remco van Mook &lt;remco@virtu.nl&gt; &nl;
14 Martijn van Oosterhout &lt;kleptog@cupid.suninternet.com&gt; &nl;
15 Paul B Schroeder &lt;paulsch@us.ibm.com&gt; &nl;
16 Jasper Spaans &lt;jasper@spaans.ds9a.nl&gt; &nl;
17 howto@ds9a.nl
18 <date>v0.9.0 $Date$
19 <abstract>
20 A very hands-on approach to iproute2, traffic shaping and a bit of netfilter
21 </abstract>
23 <!-- Table of contents -->
24 <toc>
26 <!-- Begin the document -->
28 <sect>Dedication
29 <p>
30 This document is dedicated to lots of people, and is my attempt to do
31 something back. To list but a few:
32 <p>
33 <itemize>
34 <item>Rusty Russell
35 <item>Alexey N. Kuznetsov
36 <item>The good folks from Google
37 <item>The staff of Casema Internet
38 </itemize>
40 <sect>Introduction
41 <p>
42 Welcome, gentle reader.
43 <p>
44 This document hopes to enlighten you on how to do more with Linux 2.2/2.4
45 routing. Unbeknownst to most users, you already run tools which allow you to
46 do spectacular things. Commands like 'route' and 'ifconfig' are actually
47 very thin wrappers for the very powerful iproute2 infrastructure.
48 <p>
49 I hope that this HOWTO will become as readable as the ones by Rusty Russell
50 of (amongst other things) netfilter fame.
52 You can always reach us by writing to the <url name="HOWTO team"
53 url="mailto:HOWTO@ds9a.nl">. However, please consider posting to the mailing
54 list (see the relevant section) if you have questions which are not directly
55 related to this HOWTO.
57 Before losing your way in this HOWTO, if all you want to do is simple
58 traffic shaping, skip everything and head to the 'Other possibilties'
59 chapter, and read about CBQ.init.
62 <sect1>Disclaimer &amp; License
63 <p>
64 This document is distributed in the hope that it will be useful,
65 but WITHOUT ANY WARRANTY; without even the implied warranty of
66 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
68 In short, if your STM-64 backbone breaks down and distributes pornography to
69 your most esteemed customers - it's never our fault. Sorry.
71 Copyright (c) 2001 by bert hubert, Gregory Maxwell, Martijn van
72 Oosterhout, Remco can Mook, Paul B. Schroeder and others. This material may
73 be distributed only subject to the terms and conditions set forth in the
74 Open Publication License, v1.0 or later (the latest version is presently
75 available at http://www.opencontent.org/openpub/).
77 Please freely copy and distribute (sell or give away) this document in any
78 format. It's requested that corrections and/or comments be fowarded to the
79 document maintainer.
81 It is also requested that if you publish this HOWTO in hardcopy that you
82 send the authors some samples for 'review purposes' :-)
84 <sect1>Prior knowledge
85 <p>
86 As the title implies, this is the 'Advanced' HOWTO. While by no means rocket
87 science, some prior knowledge is assumed.
89 Here are some other references which might help teach you more:
90 <descrip>
91 <tag><url
92 url="http://netfilter.samba.org/unreliable-guides/networking-concepts-HOWTO/index.html"
93 name="Rusty Russell's networking-concepts-HOWTO"></tag>
94 Very nice introduction, explaining what a network is, and how it is
95 connected to other networks
96 <tag>Linux Networking-HOWTO (Previously the Net-3 HOWTO)</tag>
97 Great stuff, although very verbose. It teaches you a lot of stuff that's
98 already configured if you are able to connect to the Internet.
99 Should be located in <file>/usr/doc/HOWTO/NET3-4-HOWTO.txt</file> but can be also be found
100 <url url="http://www.linuxports.com/howto/networking"
101 name="online">
102 </descrip>
104 <sect1>What Linux can do for you
106 A small list of things that are possible:
108 <itemize>
109 <item>Throttle bandwidth for certain computers
110 <item>Throttle bandwidth TO certain computers
111 <item>Help you to fairly share your bandwidth
112 <item>Protect your network from DoS attacks
113 <item>Protect the Internet from your customers
114 <item>Multiplex several servers as one, for load balancing or
115 enhanced availability
116 <item>Restrict access to your computers
117 <item>Limit access of your users to other hosts
118 <item>Do routing based on user id (yes!), MAC address, source IP
119 address, port, type of service, time of day or content
120 </itemize>
122 Currently, not many people are using these advanced features. This is for
123 several reasons. While the provided documentation is verbose, it is not very
124 hands-on. Traffic control is almost undocumented.
125 <sect1>Housekeeping notes
127 There are several things which should be noted about this document. While I
128 wrote most of it, I really don't want it to stay that way. I am a strong
129 believer in Open Source, so I encourage you to send feedback, updates,
130 patches etcetera. Do not hesitate to inform me of typos or plain old errors.
131 If my English sounds somewhat wooden, please realise that I'm not a native
132 speaker. Feel free to send suggestions.
134 If you feel to you are better qualified to maintain a section, or think that
135 you can author and maintain new sections, you are welcome to do so. The SGML
136 of this HOWTO is available via CVS, I very much envision more people
137 working on it.
139 In aid of this, you will find lots of FIXME notices. Patches are always
140 welcome! Wherever you find a FIXME, you should know that you are treading in
141 unknown territory. This is not to say that there are no errors elsewhere,
142 but be extra careful. If you have validated something, please let us know so
143 we can remove the FIXME notice.
145 About this HOWTO, I will take some liberties along the road. For example, I
146 postulate a 10Mbit Internet connection, while I know full well that those
147 are not very common.
148 <sect1>Access, CVS &amp; submitting updates
150 The canonical location for the HOWTO is <url
151 url="http://www.ds9a.nl/lartc" name="here">.
153 We now have anonymous CVS access available to the world at large. This is
154 good in a number of ways. You can easily upgrade to newer versions of this
155 HOWTO and submitting patches is no work at all.
157 Furthermore, it allows the authors to work on the source independently,
158 which is good too.
160 <tscreen><verb>
161 $ export CVSROOT=:pserver:anon@outpost.ds9a.nl:/var/cvsroot
162 $ cvs login
163 CVS password: [enter 'cvs' (without 's)]
164 $ cvs co 2.4routing
165 cvs server: Updating 2.4routing
166 U 2.4routing/2.4routing.sgml
167 </verb></tscreen>
169 If you spot an error, or want to add something, just fix it locally, and run
170 cvs diff -u, and send the result off to us.
172 A Makefile is supplied which should help you create postscript, dvi, pdf,
173 html and plain text. You may need to install sgml-tools, ghostscript and
174 tetex to get all formats.
176 <sect1>Mailing list
178 <label id="MLIST">
179 The authors receive an increasing amount of mail about this HOWTO. Because
180 of the clear interest of the community, it has been decided to start a
181 mailinglist where people can talk to each other about Advanced Routing and
182 Traffic Control. You can subscribe to the list
183 <url url="http://mailman.ds9a.nl/mailman/listinfo/lartc" name="here">.
185 It should be pointed out that the authors are very hesitant of answering
186 questions not asked on the list. We would like the archive of the list to
187 become some kind of knowledge base. If you have a question, please search
188 the archive, and then post to the mailinglist.
190 <sect1>Layout of this document
192 We will be doing interesting stuff almost immediately, which also means that
193 there will initially be parts that are explained incompletely or are not
194 perfect. Please gloss over these parts and assume that all will become clear.
196 Routing and filtering are two distinct things. Filtering is documented very
197 well by Rusty's HOWTOs, available here:
199 <itemize>
200 <item><url url="http://netfilter.samba.org/unreliable-guides/"
201 name="Rusty's Remarkably Unreliable Guides">
202 </itemize>
204 We will be focusing mostly on what is possible by combining netfilter and
205 iproute2.
206 <sect>Introduction to iproute2
207 <sect1>Why iproute2?
209 Most Linux distributions, and most UNIX's, currently use the
210 venerable 'arp', 'ifconfig' and 'route' commands. While these tools work,
211 they show some unexpected behaviour under Linux 2.2 and up. For example, GRE
212 tunnels are an integral part of routing these days, but require completely
213 different tools.
215 With iproute2, tunnels are an integral part of the tool set.
217 The 2.2 and above Linux kernels include a completely redesigned network
218 subsystem. This new networking code brings Linux performance and a feature
219 set with little competition in the general OS arena. In fact, the new
220 routing, filtering, and classifying code is more featureful than the one
221 provided by many dedicated routers and firewalls and traffic shaping
222 products.
224 As new networking concepts have been invented, people have found ways to
225 plaster them on top of the existing framework in existing OSes. This
226 constant layering of cruft has lead to networking code that is filled with
227 strange behaviour, much like most human languages. In the past, Linux
228 emulated SunOS's handling of many of these things, which was not ideal.
230 This new framework makes it possible to clearly express features
231 previously beyond Linux's reach.
233 <sect1>iproute2 tour
235 Linux has a sophisticated system for bandwidth provisioning called Traffic
236 Control. This system supports various method for classifying, prioritizing,
237 sharing, and limiting both inbound and outbound traffic.
240 We'll start off with a tiny tour of iproute2 possibilities.
241 <sect1>Prerequisites
243 You should make sure that you have the userland tools installed. This
244 package is called 'iproute' on both RedHat and Debian, and may otherwise be
245 found at <tt>ftp://ftp.inr.ac.ru/ip-routing/iproute2-2.2.4-now-ss??????.tar.gz"</tt>.
247 You can also try <url name="here" url="ftp://ftp.inr.ac.ru/ip-routing/iproute2-current.tar.gz">
248 for the latest version.
250 Some parts of iproute require you to have certain kernel options enabled. It
251 should also be noted that all releases of RedHat up to and including 6.2
252 come without most of the traffic control features in the default kernel.
254 RedHat 7.2 has everything in by default.
256 Also make sure that you have netlink support, should you choose to roll your
257 own kernel. Iproute2 needs it.
259 <sect1>Exploring your current configuration
261 This may come as a surprise, but iproute2 is already configured! The current
262 commands <tt>ifconfig</tt> and <tt>route</tt> are already using the advanced
263 syscalls, but mostly with very default (ie. boring) settings.
265 The <tt>ip</tt> tool is central, and we'll ask it to display our interfaces
266 for us.
267 <sect2><tt>ip</tt> shows us our links
269 <tscreen><verb>
270 [ahu@home ahu]$ ip link list
271 1: lo: <LOOPBACK,UP> mtu 3924 qdisc noqueue
272 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
273 2: dummy: <BROADCAST,NOARP> mtu 1500 qdisc noop
274 link/ether 00:00:00:00:00:00 brd ff:ff:ff:ff:ff:ff
275 3: eth0: <BROADCAST,MULTICAST,PROMISC,UP> mtu 1400 qdisc pfifo_fast qlen 100
276 link/ether 48:54:e8:2a:47:16 brd ff:ff:ff:ff:ff:ff
277 4: eth1: <BROADCAST,MULTICAST,PROMISC,UP> mtu 1500 qdisc pfifo_fast qlen 100
278 link/ether 00:e0:4c:39:24:78 brd ff:ff:ff:ff:ff:ff
279 3764: ppp0: <POINTOPOINT,MULTICAST,NOARP,UP> mtu 1492 qdisc pfifo_fast qlen 10
280 link/ppp
282 </verb></tscreen>
283 <p>Your mileage may vary, but this is what it shows on my NAT router at
284 home. I'll only explain part of the output as not everything is directly
285 relevant.
287 We first see the loopback interface. While your computer may function
288 somewhat without one, I'd advise against it. The MTU size (Maximum Transfer
289 Unit) is 3924 octets, and it is not supposed to queue. Which makes sense
290 because the loopback interface is a figment of your kernel's imagination.
292 I'll skip the dummy interface for now, and it may not be present on your
293 computer. Then there are my two physical network interfaces, one at the side
294 of my cable modem, the other one serves my home ethernet segment.
295 Furthermore, we see a ppp0 interface.
297 Note the absence of IP addresses. iproute disconnects the concept of 'links'
298 and 'IP addresses'. With IP aliasing, the concept of 'the' IP address had
299 become quite irrelevant anyhow.
301 It does show us the MAC addresses though, the hardware identifier of our
302 ethernet interfaces.
303 <sect2><tt>ip</tt> shows us our IP addresses
305 <tscreen><verb>
306 [ahu@home ahu]$ ip address show
307 1: lo: <LOOPBACK,UP> mtu 3924 qdisc noqueue
308 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
309 inet 127.0.0.1/8 brd 127.255.255.255 scope host lo
310 2: dummy: <BROADCAST,NOARP> mtu 1500 qdisc noop
311 link/ether 00:00:00:00:00:00 brd ff:ff:ff:ff:ff:ff
312 3: eth0: <BROADCAST,MULTICAST,PROMISC,UP> mtu 1400 qdisc pfifo_fast qlen 100
313 link/ether 48:54:e8:2a:47:16 brd ff:ff:ff:ff:ff:ff
314 inet 10.0.0.1/8 brd 10.255.255.255 scope global eth0
315 4: eth1: <BROADCAST,MULTICAST,PROMISC,UP> mtu 1500 qdisc pfifo_fast qlen 100
316 link/ether 00:e0:4c:39:24:78 brd ff:ff:ff:ff:ff:ff
317 3764: ppp0: <POINTOPOINT,MULTICAST,NOARP,UP> mtu 1492 qdisc pfifo_fast qlen 10
318 link/ppp
319 inet 212.64.94.251 peer 212.64.94.1/32 scope global ppp0
320 </verb></tscreen>
322 This contains more information. It shows all our addresses, and to which
323 cards they belong. 'inet' stands for Internet (IPv4). There are lots of other
324 address families, but these don't concern us right now.
326 Let's examine eth0 somewhat closer. It says that it is related to the inet
327 address '10.0.0.1/8'. What does this mean? The /8 stands for the number of
328 bits that are in the Network Address. There are 32 bits, so we have 24 bits
329 left that are part of our network. The first 8 bits of 10.0.0.1 correspond
330 to 10.0.0.0, our Network Address, and our netmask is 255.0.0.0.
332 The other bits are connected to this interface, so 10.250.3.13 is directly
333 available on eth0, as is 10.0.0.1 for example.
335 With ppp0, the same concept goes, though the numbers are different. Its
336 address is 212.64.94.251, without a subnet mask. This means that we have a
337 point-to-point connection and that every address, with the exception of
338 212.64.94.251, is remote. There is more information, however. It tells us
339 that on the other side of the link there is, yet again, only one address,
340 212.64.94.1. The /32 tells us that there are no 'network bits'.
342 It is absolutely vital that you grasp these concepts. Refer to the
343 documentation mentioned at the beginning of this HOWTO if you have trouble.
345 You may also note 'qdisc', which stands for Queueing Discipline. This will
346 become vital later on.
348 <sect2><tt>ip</tt> shows us our routes
350 Well, we now know how to find 10.x.y.z addresses, and we are able to reach
351 212.64.94.1. This is not enough however, so we need instructions on how to
352 reach the world. The Internet is available via our ppp connection, and it
353 appears that 212.64.94.1 is willing to spread our packets around the
354 world, and deliver results back to us.
356 <tscreen><verb>
357 [ahu@home ahu]$ ip route show
358 212.64.94.1 dev ppp0 proto kernel scope link src 212.64.94.251
359 10.0.0.0/8 dev eth0 proto kernel scope link src 10.0.0.1
360 127.0.0.0/8 dev lo scope link
361 default via 212.64.94.1 dev ppp0
362 </verb></tscreen>
364 This is pretty much self explanatory. The first 4 lines of output explicitly
365 state what was already implied by <tt>ip address show</tt>, the last line
366 tells us that the rest of the world can be found via 212.64.94.1, our
367 default gateway. We can see that it is a gateway because of the word
368 via, which tells us that we need to send packets to 212.64.94.1, and that it
369 will take care of things.
371 For reference, this is what the old 'route' utility shows us:
372 <tscreen><verb>
373 [ahu@home ahu]$ route -n
374 Kernel IP routing table
375 Destination Gateway Genmask Flags Metric Ref Use
376 Iface
377 212.64.94.1 0.0.0.0 255.255.255.255 UH 0 0 0 ppp0
378 10.0.0.0 0.0.0.0 255.0.0.0 U 0 0 0 eth0
379 127.0.0.0 0.0.0.0 255.0.0.0 U 0 0 0 lo
380 0.0.0.0 212.64.94.1 0.0.0.0 UG 0 0 0 ppp0
381 </verb></tscreen>
383 <sect1>ARP
385 ARP is the Address Resolution Protocol as described in
386 <url url="http://www.faqs.org/rfcs/rfc826.html" name="RFC 826">.
387 ARP is used by a networked machine to resolve the hardware location/address of
388 another machine on the same
389 local network. Machines on the Internet are generally known by their names
390 which resolve to IP
391 addresses. This is how a machine on the foo.com network is able to communicate
392 with another machine which is on the bar.net network. An IP address, though,
393 cannot tell you the physical location of a machine. This is where ARP comes
394 into the picture.
396 Let's take a very simple example. Suppose I have a network composed of several
397 machines. Two of the machines which are currently on my network are foo
398 with an IP address of 10.0.0.1 and bar with an IP address of 10.0.0.2.
399 Now foo wants to ping bar to see that he is alive, but alas, foo has no idea
400 where bar is. So when foo decides to ping bar he will need to send
401 out an ARP request.
402 This ARP request is akin to foo shouting out on the network "Bar (10.0.0.2)!
403 Where are you?" As a result of this every machine on the network will hear
404 foo shouting, but only bar (10.0.0.2) will respond. Bar will then send an
405 ARP reply directly back to foo which is akin
406 bar saying,
407 "Foo (10.0.0.1) I am here at 00:60:94:E9:08:12." After this simple transaction
408 that's used to locate his friend on the network, foo is able to communicate
409 with bar until he (his arp cache) forgets where bar is (typically after
410 15 minutes on Unix).
412 Now let's see how this works.
413 You can view your machines current arp/neighbor cache/table like so:
414 <tscreen><verb>
415 [root@espa041 /home/src/iputils]# ip neigh show
416 9.3.76.42 dev eth0 lladdr 00:60:08:3f:e9:f9 nud reachable
417 9.3.76.1 dev eth0 lladdr 00:06:29:21:73:c8 nud reachable
418 </verb></tscreen>
420 As you can see my machine espa041 (9.3.76.41) knows where to find espa042
421 (9.3.76.42) and
422 espagate (9.3.76.1). Now let's add another machine to the arp cache.
424 <tscreen><verb>
425 [root@espa041 /home/paulsch/.gnome-desktop]# ping -c 1 espa043
426 PING espa043.austin.ibm.com (9.3.76.43) from 9.3.76.41 : 56(84) bytes of data.
427 64 bytes from 9.3.76.43: icmp_seq=0 ttl=255 time=0.9 ms
429 --- espa043.austin.ibm.com ping statistics ---
430 1 packets transmitted, 1 packets received, 0% packet loss
431 round-trip min/avg/max = 0.9/0.9/0.9 ms
433 [root@espa041 /home/src/iputils]# ip neigh show
434 9.3.76.43 dev eth0 lladdr 00:06:29:21:80:20 nud reachable
435 9.3.76.42 dev eth0 lladdr 00:60:08:3f:e9:f9 nud reachable
436 9.3.76.1 dev eth0 lladdr 00:06:29:21:73:c8 nud reachable
437 </verb></tscreen>
439 As a result of espa041 trying to contact espa043, espa043's hardware
440 address/location has now been added to the arp/neighbor cache.
441 So until the entry for
442 espa043 times out (as a result of no communication between the two) espa041
443 knows where to find espa043 and has no need to send an ARP request.
445 Now let's delete espa043 from our arp cache:
447 <tscreen><verb>
448 [root@espa041 /home/src/iputils]# ip neigh delete 9.3.76.43 dev eth0
449 [root@espa041 /home/src/iputils]# ip neigh show
450 9.3.76.43 dev eth0 nud failed
451 9.3.76.42 dev eth0 lladdr 00:60:08:3f:e9:f9 nud reachable
452 9.3.76.1 dev eth0 lladdr 00:06:29:21:73:c8 nud stale
453 </verb></tscreen>
455 Now espa041 has again forgotten where to find espa043 and will need to send
456 another ARP request the next time he needs to communicate with espa043.
457 You can also see from the above output that espagate (9.3.76.1) has been
458 changed to the "stale" state. This means that the location shown is still
459 valid, but it will have to be confirmed at the first transaction to that
460 machine.
462 <sect>Rules - routing policy database
464 If you have a large router, you may well cater for the needs of different
465 people, who should be served differently. The routing policy database allows
466 you to do this by having multiple sets of routing tables.
468 If you want to use this feature, make sure that your kernel is compiled with
469 the "IP: advanced router" and "IP: policy routing" features.
471 When the kernel needs to make a routing decision, it finds out which table
472 needs to be consulted. By default, there are three tables. The old 'route'
473 tool modifies the main and local tables, as does the ip tool (by default).
475 The default rules:
476 <tscreen><verb>
477 [ahu@home ahu]$ ip rule list
478 0: from all lookup local
479 32766: from all lookup main
480 32767: from all lookup default
481 </verb></tscreen>
483 This lists the priority of all rules. We see that all rules apply to all
484 packets ('from all'). We've seen the 'main' table before, it is output by
485 <tt>ip route ls</tt>, but the 'local' and 'default' table are new.
487 If we want to do fancy things, we generate rules which point to different
488 tables which allow us to override system wide routing rules.
490 For the exact semantics on what the kernel does when there are more matching
491 rules, see Alexey's ip-cref documentation.
493 <sect1>Simple source policy routing
495 Let's take a real example once again, I have 2 (actually 3, about time I
496 returned them) cable modems, connected to a Linux NAT ('masquerading')
497 router. People living here pay me to use the Internet. Suppose one of my
498 house mates only visits hotmail and wants to pay less. This is fine with me,
499 but they'll end up using the low-end cable modem.
501 The 'fast' cable modem is known as 212.64.94.251 and is a PPP link to
502 212.64.94.1. The 'slow' cable modem is known by various ip addresses,
503 212.64.78.148 in this example and is a link to 195.96.98.253.
505 The local table:
506 <tscreen><verb>
507 [ahu@home ahu]$ ip route list table local
508 broadcast 127.255.255.255 dev lo proto kernel scope link src 127.0.0.1
509 local 10.0.0.1 dev eth0 proto kernel scope host src 10.0.0.1
510 broadcast 10.0.0.0 dev eth0 proto kernel scope link src 10.0.0.1
511 local 212.64.94.251 dev ppp0 proto kernel scope host src 212.64.94.251
512 broadcast 10.255.255.255 dev eth0 proto kernel scope link src 10.0.0.1
513 broadcast 127.0.0.0 dev lo proto kernel scope link src 127.0.0.1
514 local 212.64.78.148 dev ppp2 proto kernel scope host src 212.64.78.148
515 local 127.0.0.1 dev lo proto kernel scope host src 127.0.0.1
516 local 127.0.0.0/8 dev lo proto kernel scope host src 127.0.0.1
517 </verb></tscreen>
519 Lots of obvious things, but things that need to be specified somewhere.
520 Well, here they are. The default table is empty.
522 Let's view the 'main' table:
523 <tscreen><verb>
524 [ahu@home ahu]$ ip route list table main
525 195.96.98.253 dev ppp2 proto kernel scope link src 212.64.78.148
526 212.64.94.1 dev ppp0 proto kernel scope link src 212.64.94.251
527 10.0.0.0/8 dev eth0 proto kernel scope link src 10.0.0.1
528 127.0.0.0/8 dev lo scope link
529 default via 212.64.94.1 dev ppp0
530 </verb></tscreen>
532 We now generate a new rule which we call 'John', for our hypothetical
533 house mate. Although we can work with pure numbers, it's far easier if we add
534 our tables to <file>/etc/iproute2/rt_tables</file>.
536 <tscreen><verb>
537 # echo 200 John >> /etc/iproute2/rt_tables
538 # ip rule add from 10.0.0.10 table John
539 # ip rule ls
540 0: from all lookup local
541 32765: from 10.0.0.10 lookup John
542 32766: from all lookup main
543 32767: from all lookup default
544 </verb></tscreen>
546 Now all that is left is to generate John's table, and flush the route cache:
547 <tscreen><verb>
548 # ip route add default via 195.96.98.253 dev ppp2 table John
549 # ip route flush cache
550 </verb></tscreen>
552 And we are done. It is left as an exercise for the reader to implement this
553 in ip-up.
554 <sect>GRE and other tunnels
556 There are 3 kinds of tunnels in Linux. There's IP in IP tunneling, GRE tunneling and tunnels that live outside the kernel (like, for example PPTP).
557 <sect1>A few general remarks about tunnels:
559 Tunnels can be used to do some very unusual and very cool stuff. They can
560 also make things go horribly wrong when you don't configure them right.
561 Don't point your default route to a tunnel device unless you know
562 <bf>exactly</bf> what you are doing :-). Furthermore, tunneling increases
563 overhead, because it needs an extra set of IP headers. Typically this is 20
564 bytes per packet, so if the normal packet size (MTU) on a network is 1500
565 bytes, a packet that is sent through a tunnel can only be 1480 bytes big.
566 This is not necessarily a problem, but be sure to read up on IP packet
567 fragmentation/reassembly when you plan to connect large networks with
568 tunnels. Oh, and of course, the fastest way to dig a tunnel is to dig at
569 both sides.
571 <sect1>IP in IP tunneling
573 This kind of tunneling has been available in Linux for a long time. It requires 2 kernel modules,
574 ipip.o and new_tunnel.o.
576 Let's say you have 3 networks: Internal networks A and B, and intermediate network C (or let's say, Internet).
577 So we have network A:
579 <tscreen><verb>
580 network 10.0.1.0
581 netmask 255.255.255.0
582 router 10.0.1.1
583 </verb></tscreen>
584 The router has address 172.16.17.18 on network C.
586 and network B:
587 <tscreen><verb>
588 network 10.0.2.0
589 netmask 255.255.255.0
590 router 10.0.2.1
591 </verb></tscreen>
592 The router has address 172.19.20.21 on network C.
594 As far as network C is concerned, we assume that it will pass any packet sent
595 from A to B and vice versa. You might even use the Internet for this.
597 Here's what you do:
599 First, make sure the modules are installed:
601 <tscreen><verb>
602 insmod ipip.o
603 insmod new_tunnel.o
604 </verb></tscreen>
605 Then, on the router of network A, you do the following:
606 <tscreen><verb>
607 ifconfig tunl0 10.0.1.1 pointopoint 172.19.20.21
608 route add -net 10.0.2.0 netmask 255.255.255.0 dev tunl0
609 </verb></tscreen>
610 And on the router of network B:
611 <tscreen><verb>
612 ifconfig tunl0 10.0.2.1 pointopoint 172.16.17.18
613 route add -net 10.0.1.0 netmask 255.255.255.0 dev tunl0
614 </verb></tscreen>
615 And if you're finished with your tunnel:
616 <tscreen><verb>
617 ifconfig tunl0 down
618 </verb></tscreen>
619 Presto, you're done. You can't forward broadcast or IPv6 traffic through
620 an IP-in-IP tunnel, though. You just connect 2 IPv4 networks that normally wouldn't be able to talk to each other, that's all. As far as compatibility goes, this code has been around a long time, so it's compatible all the way back to 1.3 kernels. Linux IP-in-IP tunneling doesn't work with other Operating Systems or routers, as far as I know. It's simple, it works. Use it if you have to, otherwise use GRE.
622 <sect1>GRE tunneling
624 GRE is a tunneling protocol that was originally developed by Cisco, and it
625 can do a few more things than IP-in-IP tunneling. For example, you can also
626 transport multicast traffic and IPv6 through a GRE tunnel.
628 In Linux, you'll need the ip_gre.o module.
630 <sect2>IPv4 Tunneling
632 Let's do IPv4 tunneling first:
634 Let's say you have 3 networks: Internal networks A and B, and intermediate network C (or let's say, Internet).
636 So we have network A:
637 <tscreen><verb>
638 network 10.0.1.0
639 netmask 255.255.255.0
640 router 10.0.1.1
641 </verb></tscreen>
642 The router has address 172.16.17.18 on network C.
643 Let's call this network neta (ok, hardly original)
645 and network B:
646 <tscreen><verb>
647 network 10.0.2.0
648 netmask 255.255.255.0
649 router 10.0.2.1
650 </verb></tscreen>
651 The router has address 172.19.20.21 on network C.
652 Let's call this network netb (still not original)
654 As far as network C is concerned, we assume that it will pass any packet sent
655 from A to B and vice versa. How and why, we do not care.
657 On the router of network A, you do the following:
658 <tscreen><verb>
659 ip tunnel add netb mode gre remote 172.19.20.21 local 172.16.17.18 ttl 255
660 ip link set netb up
661 ip addr add 10.0.1.1 dev netb
662 ip route add 10.0.2.0/24 dev netb
663 </verb></tscreen>
665 Let's discuss this for a bit. In line 1, we added a tunnel device, and
666 called it netb (which is kind of obvious because that's where we want it to
667 go). Furthermore we told it to use the GRE protocol (mode gre), that the
668 remote address is 172.19.20.21 (the router at the other end), that our
669 tunneling packets should originate from 172.16.17.18 (which allows your
670 router to have several IP addresses on network C and let you decide which
671 one to use for tunneling) and that the TTL field of the packet should be set
672 to 255 (ttl 255).
674 The second line enables the device.
676 In the third line we gave the newly born interface netb the address
677 10.0.1.1. This is OK for smaller networks, but when you're starting up a
678 mining expedition (LOTS of tunnels), you might want to consider using
679 another IP range for tunneling interfaces (in this example, you could use
680 10.0.3.0).
682 <p>In the fourth line we set the route for network B. Note the different notation for the netmask. If you're not familiar with this notation, here's how it works: you write out the netmask in binary form, and you count all the ones. If you don't know how to do that, just remember that 255.0.0.0 is /8, 255.255.0.0 is /16 and 255.255.255.0 is /24. Oh, and 255.255.254.0 is /23, in case you were wondering.
684 But enough about this, let's go on with the router of network B.
685 <tscreen><verb>
686 ip tunnel add neta mode gre remote 172.16.17.18 local 172.19.20.21 ttl 255
687 ip link set neta up
688 ip addr add 10.0.2.1 dev neta
689 ip route add 10.0.1.0/24 dev neta
690 </verb></tscreen>
691 And when you want to remove the tunnel on router A:
692 <tscreen><verb>
693 ip link set netb down
694 ip tunnel del netb
695 </verb></tscreen>
696 Of course, you can replace netb with neta for router B.
698 <sect2>IPv6 Tunneling
700 See Section 6 for a short bit about IPv6 Addresses.
702 On with the tunnels.
704 Let's assume that you have the following IPv6 network, and you want to connect it to 6bone, or a friend.
706 <tscreen><verb>
707 Network 3ffe:406:5:1:5:a:2:1/96
708 </verb></tscreen>
709 Your IPv4 address is 172.16.17.18, and the 6bone router has IPv4 address 172.22.23.24.
711 <tscreen><verb>
712 ip tunnel add sixbone mode sit remote 172.22.23.24 local 172.16.17.18 ttl 255
713 ip link set sixbone up
714 ip addr add 3ffe:406:5:1:5:a:2:1/96 dev sixbone
715 ip route add 3ffe::/15 dev sixbone
716 </verb></tscreen>
718 Let's discuss this. In the first line, we created a tunnel device called sixbone. We gave it mode sit (which is IPv6 in IPv4 tunneling) and told it where to go to (remote) and where to come from (local). TTL is set to maximum, 255. Next, we made the device active (up). After that, we added our own network address, and set a route for 3ffe::/15 (which is currently all of 6bone) through the tunnel.
720 GRE tunnels are currently the preferred type of tunneling. It's a standard that is also widely adopted outside the Linux community and therefore a Good Thing.
722 <sect1>Userland tunnels
724 There are literally dozens of implementations of tunneling outside the kernel. Best known are of course PPP and PPTP, but there are lots more (some proprietary, some secure, some that don't even use IP) and that is really beyond the scope of this HOWTO.
726 <sect>IPv6 tunneling with Cisco and/or 6bone
728 By Marco Davids &lt;marco@sara.nl&gt;
730 NOTE to maintainer:
732 As far as I am concerned, this IPv6-IPv4 tunneling is not per definition
733 GRE tunneling. You could tunnel IPv6 over IPv4 by means of GRE tunnel devices
734 (GRE tunnels ANY to IPv4), but the device used here ("sit") only tunnels
735 IPv6 over IPv4 and is therefore something different.
737 <sect1>IPv6 Tunneling
739 This is another application of the tunneling capabilities of Linux. It is
740 popular among the IPv6 early adopters, or pioneers if you like.
741 The 'hands-on' example described below is certainly not the only way
742 to do IPv6 tunneling. However, it is the method that is often used to tunnel
743 between Linux and a Cisco IPv6 capable router and experience tells us that
744 this is just the thing many people are after. Ten to one this applies to
745 you too ;-)
747 A short bit about IPv6 addresses:
749 IPv6 addresses are, compared to IPv4 addresses, really big: 128 bits
750 against 32 bits. And this provides us just with the thing we need: many, many
751 IP-addresses: 340,282,266,920,938,463,463,374,607,431,768,211,465 to be
752 precise. Apart from this, IPv6 (or IPng, for IP Next Generation) is supposed
753 to provide for smaller routing tables on the Internet's backbone routers,
754 simpler configuration of equipment, better security at the IP level and
755 better support for QoS.
757 An example: 2002:836b:9820:0000:0000:0000:836b:9886
759 Writing down IPv6 addresses can be quite a burden. Therefore, to make
760 life easier there are some rules:
762 <itemize>
763 <item>
764 Don't use leading zeroes. Same as in IPv4.
766 <item>Use colons to separate every 16 bits or two bytes.
768 <item>When you have lots of consecutive zeroes,
769 you can write this down as ::. You can only do this once in an
770 address and only for quantities of 16 bits, though.
771 </itemize>
773 The address 2002:836b:9820:0000:0000:0000:836b:9886 can be written down
774 as 2002:836b:9820::836b:9886, which is somewhat friendlier.
776 Another example, the address 3ffe:0000:0000:0000:0000:0020:34A1:F32C can be
777 written down as 3ffe::20:34A1:F32C, which is a lot shorter.
779 IPv6 is intended to be the successor of the current IPv4. Because it
780 is relatively new technology, there is no worldwide native IPv6 network
781 yet. To be able to move forward swiftly, the 6bone was introduced.
783 Native IPv6 networks are connected to each other by encapsulating the IPv6
784 protocol in IPv4 packets and sending them over the existing IPv4 infrastructure
785 from one IPv6 site to another.
787 That is precisely where the tunnel steps in.
789 To be able to use IPv6, we should have a kernel that supports it. There
790 are many good documents on how to achieve this. But it all comes down to
791 a few steps:
792 <itemize>
793 <item>Get yourself a recent Linux distribution, with suitable glibc.
794 <item>Then get yourself an up-to-date kernel source.
795 </itemize>
796 If you are all set, then you can go ahead and compile an IPv6 capable
797 kernel:
798 <itemize>
799 <item>Go to /usr/src/linux and type:
800 <item>make menuconfig
801 <item>Choose "Networking Options"
802 <item>Select "The IPv6 protocol", "IPv6: enable EUI-64 token format", "IPv6:
803 disable provider based addresses"
804 </itemize>
805 HINT: Don't go for the 'module' option. Often this won't work well.
807 In other words, compile IPv6 as 'built-in' in your kernel.
808 You can then save your config like usual and go ahead with compiling
809 the kernel.
811 HINT: Before doing so, consider editing the Makefile:
812 EXTRAVERSION = -x ; --> ; EXTRAVERSION = -x-IPv6
814 There is a lot of good documentation about compiling and installing
815 a kernel, however this document is about something else. If you run into
816 problems at this stage, go and look for documentation about compiling a
817 Linux kernel according to your own specifications.
819 The file /usr/src/linux/README might be a good start.
820 After you acomplished all this, and rebooted with your brand new kernel,
821 you might want to issue an '/sbin/ifconfig -a' and notice the brand
822 new 'sit0-device'. SIT stands for Simple Internet Transition. You may give
823 yourself a compliment; you are now one major step closer to IP, the Next
824 Generation ;-)
826 Now on to the next step. You want to connect your host, or maybe even
827 your entire LAN to another IPv6 capable network. This might be the "6bone"
828 that is setup especially for this particular purpose.
830 Let's assume that you have the following IPv6 network: 3ffe:604:6:8::/64 and
831 you want to connect it to 6bone, or a friend. Please note that the /64
832 subnet notation works just like with regular IP adresses.
834 Your IPv4 address is 145.100.24.181 and the 6bone router has IPv4 address
835 145.100.1.5
836 <tscreen><verb>
837 # ip tunnel add sixbone mode sit remote 145.100.1.5 [local 145.100.24.181 ttl 255]
838 # ip link set sixbone up
839 # ip addr add 3FFE:604:6:7::2/126 dev sixbone
840 # ip route add 3ffe::0/16 dev sixbone
841 </verb></tscreen>
843 Let's discuss this. In the first line, we created a tunnel device called
844 sixbone. We gave it mode sit (which is IPv6 in IPv4 tunneling) and told it
845 where to go to (remote) and where to come from (local). TTL is set to
846 maximum, 255.
848 Next, we made the device active (up). After that, we added our own network
849 address, and set a route for 3ffe::/15 (which is currently all of 6bone)
850 through the tunnel. If the particular machine you run this on is your IPv6
851 gateway, then consider adding the following lines:
853 <tscreen><verb>
854 # echo 1 >/proc/sys/net/ipv6/conf/all/forwarding
855 # /usr/local/sbin/radvd
856 </verb></tscreen>
857 The latter, radvd is -like zebra- a router advertisement daemon, to
858 support IPv6's autoconfiguration features. Search for it with your favourite
859 search-engine if you like.
860 You can check things like this:
862 <tscreen><verb>
863 # /sbin/ip -f inet6 addr
864 </verb></tscreen>
866 If you happen to have radvd running on your IPv6 gateway and boot your
867 IPv6 capable Linux on a machine on your local LAN, you would be able to
868 enjoy the benefits of IPv6 autoconfiguration:
869 <tscreen><verb>
870 # /sbin/ip -f inet6 addr
871 1: lo: &lt;LOOPBACK,UP&gt; mtu 3924 qdisc noqueue inet6 ::1/128 scope host
873 3: eth0: &lt;BROADCAST,MULTICAST,UP&gt; mtu 1500 qdisc pfifo_fast qlen 100
874 inet6 3ffe:604:6:8:5054:4cff:fe01:e3d6/64 scope global dynamic
875 valid_lft forever preferred_lft 604646sec inet6 fe80::5054:4cff:fe01:e3d6/10
876 scope link
877 </verb></tscreen>
879 You could go ahead and configure your bind for IPv6 addresses. The A
880 type has an equivalent for IPv6: AAAA. The in-addr.arpa's equivalent is:
881 ip6.int. There's a lot of information available on this topic.
883 There is an increasing number of IPv6-aware applications available,
884 including secure shell, telnet, inetd, Mozilla the browser, Apache the
885 websever and a lot of others. But this is all outside the scope of this
886 Routing document ;-)
888 On the Cisco side the configuration would be something like this:
889 <tscreen><verb>
891 interface Tunnel1
892 description IPv6 tunnel
893 no ip address
894 no ip directed-broadcast
895 ipv6 enable
896 ipv6 address 3FFE:604:6:7::1/126
897 tunnel source Serial0
898 tunnel destination 145.100.24.181
899 tunnel mode ipv6ip
901 ipv6 route 3FFE:604:6:8::/64 Tunnel1
902 </verb></tscreen>
903 But if you don't have a Cisco at your disposal, try one of the many
904 IPv6 tunnel brokers available on the Internet. They are willing to configure
905 their Cisco with an extra tunnel for you. Mostly by means of a friendly
906 web interface. Search for "ipv6 tunnel broker" on your favourite search engine.
908 <sect>IPsec: secure IP over the Internet
910 FIXME: editor vacancy.
911 In the meantime, see: <url url="http://www.freeswan.org/" name="The
912 FreeS/WAN project">. Another IPSec implementation for Linux is Cerberus,
913 by NIST. However, their web pages have not been updated in over a year,
914 and their version tended to trail well behind the current Linux kernel.
915 USAGI, an alternative IPv6 implementation for Linux, also includes an
916 IPSec implementation, but that might only be for IPv6.
918 <sect>Multicast routing
920 FIXME: Editor Vacancy!
922 The Multicast-HOWTO is ancient (relatively-speaking) and may be inaccurate
923 or misleading in places, for that reason.
925 Before you can do any multicast routing, you need to configure the Linux
926 kernel to support the type of multicast routing you want to do. This, in
927 turn, requires you to decide what type of multicast routing you expect to
928 be using. There are essentially four "common" types - DVMRP (the Multicast
929 version of the RIP unicast protocol), MOSPF (the same, but for OSPF), PIM-SM
930 ("Protocol Independent Multicasting - Sparse Mode", which assumes that users
931 of any multicast group are spread out, rather than clumped) and PIM-DM (the
932 same, but "Dense Mode", which assumes that there will be significant clumps
933 of users of the same multicast group).
935 In the Linux kernel, you will notice that these options don't appear. This is
936 because the protocol itself is handled by a routing application, such as
937 Zebra, mrouted, or pimd. However, you still have to have a good idea of which
938 you're going to use, to select the right options in the kernel.
940 For all multicast routing, you will definitely need to enable "multicasting"
941 and "multicast routing". For DVMRP and MOSPF, this is sufficient. If you are
942 going to use PIM, you must also enable PIMv1 or PIMv2, depending on whether
943 the network you are connecting to uses version 1 or 2 of the PIM protocol.
945 Once you have all that sorted out, and your new Linux kernel compiled, you
946 will see that the IP protocols listed, at boot time, now include IGMP. This
947 is a protocol for managing multicast groups. At the time of writing, Linux
948 supports IGMP versions 1 and 2 only, although version 3 does exist and has
949 been documented. This doesn't really affect us that much, as IGMPv3 is still
950 new enough that the extra capabilities of IGMPv3 aren't going to be that
951 much use. Because IGMP deals with groups, only the features present in the
952 simplest version of IGMP over the entire group are going to be used. For the
953 most part, that will be IGMPv2, although IGMPv1 is sill going to be
954 encountered.
956 So far, so good. We've enabled multicasting. Now, we have to tell the Linux
957 kernel to actually do something with it, so we can start routing. This means
958 adding the Multicast virtual network to the router table:
960 ip route add 224.0.0.0/4 dev eth0
962 (Assuming, of course, that you're multicasting over eth0! Substitute the
963 device of your choice, for this.)
965 Now, tell Linux to forward packets...
967 echo 1 > /proc/sys/net/ipv4/ip_forward
969 At this point, you may be wondering if this is ever going to do anything. So,
970 to test our connection, we ping the default group, 224.0.0.1, to see if anyone
971 is alive. All machines on your LAN with multicasting enabled <em>should</em>
972 respond, but nothing else. You'll notice that none of the machines that
973 respond have an IP address of 224.0.0.1. What a surprise! :) This is a group
974 address (a "broadcast" to subscribers), and all members of the group will
975 respond with their own address, not the group address.
977 ping -c 2 224.0.0.1
979 At this point, you're ready to do actual multicast routing. Well, assuming
980 that you have two networks to route between.
982 (To Be Continued!)
984 <sect>Queueing Disciplines for Bandwidth Management
986 Now, when I discovered this, it <em>really</em> blew me away. Linux 2.2/2.4
987 comes with everything to manage bandwidth in ways comparable to high-end
988 dedicated bandwidth management systems.
990 Linux even goes far beyond what Frame and ATM provide.
992 Just to prevent confusion, tc uses the following rules for bandwith
993 specification:
994 <verb>
995 mbps = 1024 kbps = 1024 * 1024 bps => byte/s
996 mbit = 1024 kbit => kilo bit/s.
997 mb = 1024 kb = 1024 * 1024 b => byte
998 mbit = 1024 kbit => kilo bit.
999 </verb>
1000 Internally, the number is stored in bps and b.
1002 But when tc prints the rate, it uses following :
1003 <verb>
1004 1Mbit = 1024 Kbit = 1024 * 1024 bps => bit/s
1005 </verb>
1007 <sect1>Queues and Queueing Disciplines explained
1009 With queueing we determine the way in which data is <em>sent</em>. It is
1010 important to realise that we can only shape data that we transmit.
1012 With the way the Internet works, we have no direct control of what people
1013 send us. It's a bit like your (physical!) mailbox at home. There is no way
1014 you can influence the world to modify the amount of mail they send you,
1015 short of contacting everybody.
1017 However, the Internet is mostly based on TCP/IP which has a few features
1018 that help us. TCP/IP has no way of knowing the capacity of the network
1019 between two hosts, so it just starts sending data faster and faster ('slow
1020 start') and when packets start getting lost, because there is no room to
1021 send them, it will slow down. In fact it is a bit smarter than this, but
1022 more about that later.
1024 This is the equivalent of not reading half of your mail, and hoping that
1025 people will stop sending it to you. With the difference that it works for
1026 the Internet :-)
1028 If you have a router and wish to prevent certain hosts within your network
1029 from downloading too fast, you need to do your shaping on the *inner* interface
1030 of your router, the one that sends data to your own computers.
1032 You also have to be sure you are controlling the bottleneck of the link.
1033 If you have a 100Mbit NIC and you have a router that has a 256kbit link,
1034 you have to make sure you are not sending more data than your router can
1035 handle. Othewise, it will be the router who is controlling the link and
1036 shaping the available bandwith. We need to 'own the queue' so to speak, and
1037 be the slowest link in the chain. Luckily this is easily possible.
1039 <sect1>Simple, classless Queueing Disciplines
1041 As said, with queueing disciplines, we change the way data is sent.
1042 Classless queueing disciplines are those that, by and large accept data and
1043 only reschedule, delay or drop it.
1045 These can be used to shape traffic for an entire interface, without any
1046 subdivisions. It is vital that you understand this part of queueing before
1047 we go on the the classful qdisc-containing-qdiscs!
1049 By far the most widely used discipline is the pfifo_fast qdisc - this is the
1050 default. This also explains why these advanced features are so robust. They
1051 are nothing more than 'just another queue'.
1053 Each of these queues has specific strengths and weaknesses. Not all of them
1054 may be as well tested.
1056 <sect2>pfifo_fast
1058 This queue is, as the name says, First In, First Out, which means that no
1059 packet receives special treatment. At least, not quite. This queue has 3 so
1060 called 'bands'. Within each band, FIFO rules apply. However, as long as
1061 there are packets waiting in band 0, band 1 won't be processed. Same goes
1062 for band 1 and band 2.
1064 The kernel honors the so called Type of Service flag of packets, and takes
1065 care to insert 'minimum delay' packets in band 0.
1067 Do not confuse this classless simple qdisc with the classful PRIO one!
1068 Although they behave similarly, pfifo_fast is classless and you cannot add
1069 other qdiscs to it with the tc command.
1071 <sect3>Parameters &amp; usage
1073 You can't configure the pfifo_fast qdisc as it is the hardwired default.
1074 This is how it is configured by default:
1075 <descrip>
1076 <tag>priomap</tag>
1077 Determines how packet priorities, as assigned by the kernel, map to bands.
1078 Mapping occurs based on the TOS octet of the packet, which looks like this:
1080 <tscreen><verb>
1081 0 1 2 3 4 5 6 7
1082 +-----+-----+-----+-----+-----+-----+-----+-----+
1083 | | | |
1084 | PRECEDENCE | TOS | MBZ |
1085 | | | |
1086 +-----+-----+-----+-----+-----+-----+-----+-----+
1087 </verb></tscreen>
1089 The four TOS bits (the 'TOS field') are defined as:
1090 <tscreen><verb>
1091 Binary Decimcal Meaning
1092 -----------------------------------------
1093 1000 8 Minimize delay (md)
1094 0100 4 Maximize throughput (mt)
1095 0010 2 Maximize reliability (mr)
1096 0001 1 Minimize monetary cost (mmc)
1097 0000 0 Normal Service
1098 </verb></tscreen>
1100 As there is 1 bit to the right of these four bits, the actual value of the
1101 TOS field is double the value of the TOS bits. Tcpdump -v -v shows you the
1102 value of the entire TOS field, not just the four bits. It is the value you
1103 see in the first column of this table:
1105 <verb>
1106 TOS Bits Means Linux Priority Band
1107 ------------------------------------------------------------
1108 0x0 0 Normal Service 0 Best Effort 1
1109 0x2 1 Minimize Monetary Cost 1 Filler 2
1110 0x4 2 Maximize Reliability 0 Best Effort 1
1111 0x6 3 mmc+mr 0 Best Effort 1
1112 0x8 4 Maximize Throughput 2 Bulk 2
1113 0xa 5 mmc+mt 2 Bulk 2
1114 0xc 6 mr+mt 2 Bulk 2
1115 0xe 7 mmc+mr+mt 2 Bulk 2
1116 0x10 8 Minimize Delay 6 Interactive 0
1117 0x12 9 mmc+md 6 Interactive 0
1118 0x14 10 mr+md 6 Interactive 0
1119 0x16 11 mmc+mr+md 6 Interactive 0
1120 0x18 12 mt+md 4 Int. Bulk 1
1121 0x1a 13 mmc+mt+md 4 Int. Bulk 1
1122 0x1c 14 mr+mt+md 4 Int. Bulk 1
1123 0x1e 15 mmc+mr+mt+md 4 Int. Bulk 1
1124 </verb>
1126 Lots of numbers. The second column contains the value of the relevant four
1127 TOS bits, followed by their translated meaning. For example, 15 stands for a
1128 packet wanting Minimal Montetary Cost, Maximum Reliability, Maximum
1129 Throughput AND Minimum Delay. I would call this a 'Dutch Packet'.
1131 The fourth column lists the way the Linux kernel interprets the TOS bits, by
1132 showing to which Priority they are mapped.
1134 The last column shows the result of the default priomap. On the commandline,
1135 the default priomap looks like this:
1136 <tscreen>
1137 1, 2, 2, 2, 1, 2, 0, 0 , 1, 1, 1, 1, 1, 1, 1, 1
1138 </tscreen>
1140 This means that priority 4, for example, gets mapped to band number 1. The
1141 priomap also allows you to list higher priorities (&gt; 7) which do not
1142 correspond to TOS mappings, but which are set by other means.
1144 This table from RFC 1349 (read it for more details) tells you how
1145 applications might very well set their TOS bits:
1146 <verb>
1147 TELNET 1000 (minimize delay)
1149 Control 1000 (minimize delay)
1150 Data 0100 (maximize throughput)
1152 TFTP 1000 (minimize delay)
1154 SMTP
1155 Command phase 1000 (minimize delay)
1156 DATA phase 0100 (maximize throughput)
1158 Domain Name Service
1159 UDP Query 1000 (minimize delay)
1160 TCP Query 0000
1161 Zone Transfer 0100 (maximize throughput)
1163 NNTP 0001 (minimize monetary cost)
1165 ICMP
1166 Errors 0000
1167 Requests 0000 (mostly)
1168 Responses <same as request> (mostly)
1169 </verb>
1171 <tag>txqueuelen</tag>
1172 The length of this queue is gleaned from the interface configuration, which
1173 you can see and set with ifconfig and ip. To set the queue length to 10,
1174 execute: ifconfig eth0 txqueuelen 10
1176 You can't set this parameter with tc!
1177 </descrip>
1178 <sect2>Token Bucket Filter
1180 The Token Bucket Filter (TBF) is a simple qdisc that only passes packets
1181 arriving at a rate which is not exceeding some administratively set rate, but
1182 with the possibility to allow short bursts in excess of this rate.
1184 TBF is very precise, network- and processor friendly. It should be your
1185 first choice if you simply want to slow an interface down!
1187 The TBF implementation consists of a buffer (bucket), constantly filled by
1188 some virtual pieces of information called tokens, at a specific rate (token
1189 rate). The most important parameter of the bucket is its size, that is the
1190 number of tokens it can store.
1192 Each arriving token collects one incoming data packet from the data queue
1193 and is then deleted from the bucket. Associating this algorithm
1194 with the two flows -- token and data, gives us three possible scenarios:
1196 <itemize>
1197 <item> The data arrives in TBF at a rate that's <em>equal</em> to the rate
1198 of incoming tokens. In this case each incoming packet has its matching token
1199 and passes the queue without delay.
1201 <item> The data arrives in TBF at a rate that's <em>smaller</em> than the
1202 token rate. Only a part of the tokens are deleted at output of each data packet
1203 that's sent out the queue, so the tokens accumulate, up to the bucket size.
1204 The unused tokens can then be used to send data a a speed that's exceeding the
1205 standard token rate, in case short data bursts occur.
1207 <item> The data arrives in TBF at a rate <em>bigger</em> than the token rate.
1208 This means that the bucket will soon be devoid of tokens, which causes the
1209 TBF to throttle itself for a while. This is called an 'overlimit situation'.
1210 If packets keep coming in, packets will start to get dropped.
1211 </itemize>
1213 The last scenario is very important, because it allows to
1214 administratively shape the bandwidth available to data that's passing
1215 the filter.
1217 The accumulation of tokens allows a short burst of overlimit data to be
1218 still passed without loss, but any lasting overload will cause packets to be
1219 constantly delayed, and then dropped.
1221 Please note that in the actual implementation, tokens correspond to bytes,
1222 not packets.
1223 <sect3>Parameters &amp; usage
1225 Even though you will probably not need to change them, tbf has some knobs
1226 available. First the parameters that are always available:
1227 <descrip>
1228 <tag>limit or latency</tag>
1229 Limit is the number of bytes that can be queued waiting for tokens to become
1230 available. You can also specify this the other way around by setting the
1231 latency parameter, which specifies the maximum amount of time a packet can
1232 sit in the TBF. The latter calculation takes into account the size of the
1233 bucket, the rate and possibly the peakrate (if set).
1235 <tag>burst/buffer/maxburst</tag>
1236 Size of the bucket, in bytes. This is the maximum amount of bytes that
1237 tokens can be available for instantaneously. In general, larger shaping
1238 rates require a larger buffer. For 10mbit/s on Intel, you need at least
1239 10kbyte buffer if you want to reach your configured rate!
1241 If your buffer is too small, packets may be dropped because more tokens
1242 arrive per timer tick than fit in your bucket.
1243 <tag>mpu</tag>
1244 A zero-sized packet does not use zero bandwidth. For ethernet, no packet
1245 uses less than 64 bytes. The Minimum Packet Unit determines the minimal
1246 token usage for a packet.
1247 <tag>rate</tag>
1248 The speedknob. See remarks above about limits!
1249 </descrip>
1251 If the bucket contains tokens and is allowed to empty, by default it does so
1252 at infinite speed. If this is unacceptable, use the following parameters:
1254 <descrip>
1255 <tag>peakrate</tag>
1256 If tokens are available, and packets arrive, they are sent out immediately
1257 by default, at 'lightspeed' so to speak. That may not be what you want,
1258 especially if you have a large bucket.
1260 The peakrate can be used to specify how quickly the bucket is allowed to be
1261 depleted. If doing everything by the book, this is achieved by releasing a
1262 packet, and then wait just long enough, and release the next. We calculated
1263 our waits so we send just at peakrate.
1265 However, due to de default 10ms timer resolution of Unix, with 10.000 bits
1266 average packets, we are limited to 1mbit/s of peakrate!
1268 <tag>mtu/minburst</tag>
1269 The 1mbit/s peakrate is not very useful if your regular rate is more than
1270 that. A higher peakrate is possible by sending out more packets per
1271 timertick, which effectively means that we create a second bucket!
1273 This second bucket defaults to a single packet, which is not a bucket at
1274 all.
1276 To calculate the maximum possible peakrate, multiply the configured mtu by
1277 100 (or more correctly, HZ, which is 100 on intel, 1024 on Alpha).
1279 </descrip>
1280 <sect3>Sample configuration
1282 A simple but *very* useful configuration is this:
1283 <verb>
1284 # tc qdisc add dev ppp0 root tbf rate 220kbit latency 50ms burst 1540
1285 </verb>
1287 Ok, why is this useful? If you have a networking device with a large queue,
1288 like a DSL modem or a cablemodem, and you talk to it over a fast device,
1289 like over an ethernet interface, you will find that uploading absolutely
1290 destroys interactivity.
1292 This is because uploading will fill the queue in the modem, which is
1293 probably *huge* because this helps actually achieving good data throughput
1294 uploading. But this is not what you want, you want to have the queue not too
1295 big so interactivity remains and you can still do other stuff while sending
1296 data.
1298 The line above slows down sending to a rate that does not lead to a queue in
1299 the modem - the queue will be in Linux, where we can control it to a limited
1300 size.
1302 Change 220kbit to your uplink's *actual* speed, minus a few percent. If you
1303 have a really fast modem, raise 'burst' a bit.
1304 <sect2>Stochastic Fairness Queueing
1306 Stochastic Fairness Queueing (SFQ) is a simple implementation of the fair
1307 queueing algorithms family. It's less accurate than others, but it also
1308 requires less calculations while being almost perfectly fair.
1310 The key word in SFQ is conversation (or flow), which mostly corresponds to a
1311 TCP session or a UDP stream. Traffic is divided into a pretty large number
1312 of FIFO queues, one for each conversation. Traffic is then sent in a round
1313 robin fashion, giving each session the chance to send data in turn.
1315 This leads to very fair behaviour and disallows any single conversation from
1316 drowning out the rest. SFQ is called 'Stochastic' because it doesn't really
1317 allocate a queue for each session, it has an algorithm which divides traffic
1318 over a limited number of queues using a hashing algorithm.
1320 Because of the hash, multiple sessions might end up in the same bucket, which
1321 would halve each session's chance of sending a packet, thus halving the
1322 effective speed available. To prevent this situation from becoming
1323 noticeable, SFQ changes its hashing algorithm quite often so that any two
1324 colliding sessions will only do so for a small number of seconds.
1326 It is important to note that SFQ is only useful in case your actual outgoing
1327 interface is really full! If it isn't then there will be no queue on your
1328 linux machine and hence no effect. Later on we will describe how to combine
1329 SFQ with other qdiscs to get a best-of-both worlds situation.
1331 Specifically, setting SFQ on the ethernet interface heading to your
1332 cablemodem or DSL router is pointless without further shaping!
1333 <sect3>Parameters &amp; usage
1335 The SFQ is pretty much selftuning:
1336 <descrip>
1337 <tag>perturb</tag>
1338 Reconfigure hashing once this many seconds. If unset, hash will never be
1339 reconfigured. Not recommended. 10 seconds is probably a good value.
1340 <tag>quantum</tag>
1341 Amount of bytes a stream is allowed to dequeue before the next queue gets a
1342 turn. Defaults to 1 maximum sized packet (MTU-sized). Do not set below the
1343 MTU!
1344 </descrip>
1345 <sect3>Sample configuration
1347 If you have a device which has identical link speed and actual available
1348 rate, like a phone modem, this configuration will help promote fairness:
1349 <verb>
1350 # tc qdisc add dev ppp0 root sfq perturb 10
1351 # tc -s -d qdisc ls
1352 qdisc sfq 800c: dev ppp0 quantum 1514b limit 128p flows 128/1024 perturb 10sec
1353 Sent 4812 bytes 62 pkts (dropped 0, overlimits 0)
1354 </verb>
1356 The number 800c: is the automatically assigned handle number, limit means
1357 that 128 packets can wait in this queue. There are 1024 hashbuckets
1358 available for accounting, of which 128 can be active at a time (no more
1359 packets fit in the queue!) Once every 10 seconds, the hashes are
1360 reconfigured.
1362 <sect1>Advice for when to use which queue
1364 Summarizing, these are the simple queues that actually manage traffic by
1365 reordering, slowing or dropping packets.
1367 The following tips may help in chosing which queue to use. It mentions some
1368 qdiscs described in the 'Advanced &amp; less common queueing disciplines'.
1370 <itemize>
1371 <item>
1372 To purely slow down outgoing traffic, use the Token Bucket Filter. Works up
1373 to huge bandwidths, if you scale the bucket.
1374 <item>
1375 If your link is truly full and you want to make sure that no single session
1376 can dominate your outgoing bandwidth, use Stochastical Fairness Queueing.
1377 <item>
1378 If you have a big backbone and know what you are doing, consider Random
1379 Early Drop (see Advanced chapter).
1380 <item>
1381 To 'shape' incoming traffic which you are not forwarding, use the Ingress
1382 Policer. Incoming shaping is called 'policing', by the way, not 'shaping'.
1383 <item>
1384 If you *are* forwarding it, use a TBF on the interface you are forwarding
1385 the data to. Unless you want to shape traffic that may go out over several
1386 interfaces, in which case the only common factor is the incoming interface.
1387 In that case use the Ingress Policer.
1388 <item>
1389 If you don't want to shape, but only want to see if your interface is so
1390 loaded that it has to queue, use the pfifo queue (not pfifo_fast). It lacks
1391 internal bands but does account the size of its backlog.
1392 <item>
1393 Finally - you can also do 'social shaping'. You may not always be able to
1394 use technology to achieve what you want. Users experience technical
1395 constraints as hostile. A kind word may also help with getting your
1396 bandwidth to be divided right!
1397 </itemize>
1398 <sect1>Terminology
1400 To properly understand more complicated configurations it is necessary to
1401 explain a few concepts first. Because of the complexity and he relative
1402 youth of the subject, a lot of different words are used when people in fact
1403 mean the same thing.
1405 The following is loosely based on draft-ietf-diffserv-model-06.txt, 'An
1406 Informal Management Model for Diffserv Routers'. It can currently be found
1407 at http://www.ietf.org/internet-drafts/draft-ietf-diffserv-model-06.txt.
1409 Read it for the strict definitions of the terms used.
1410 <descrip>
1411 <tag>Queueing Discipline</tag>
1412 An algorithm that manages the queue of a device, either incoming (ingress)
1413 or outgoing (egress).
1414 <tag>Classless qdisc</tag>
1415 A qdisc with no configurable internal subdivisions.
1416 <tag>Classful qdisc</tag>
1417 A classful qdisc contains multiple classes. Each of these classes contains a
1418 further qdisc, which may again be classful, but need not be. According to
1419 the strict definition, pfifo_fast *is* classful, because it contains three
1420 bands which are, in fact, classes. However, from the user's configuration
1421 perspective, it is classless as the classes can't be touched with the tc
1422 tool.
1423 <tag>Classes</tag>
1424 A classful qdisc may have many classes, which each are internal to the
1425 qdisc. Each of these classes may contain a real qdisc.
1426 <tag>Classifier</tag>
1427 Each classful qdisc needs to determine to which class it needs to send a
1428 packet. This is done using the classifier.
1429 <tag>Filter</tag>
1430 Classification can be performed using filters. A filter contains a number of
1431 conditions which if matched, make the filter match.
1432 <tag>Scheduling</tag>
1433 A qdisc may, with the help of a classifier, decide that some packets need to
1434 go out earlier than others. This process is called Scheduling, and is
1435 performed for example by the pfifo_fast qdisc mentioned earlier. Scheduling
1436 is also called 'reordering', but this is confusing.
1437 <tag>Shaping</tag>
1438 The process of delaying packets before they go out to make traffic confirm
1439 to a configured maximum rate. Shaping is performed on egress. Colloquially,
1440 dropping packets to slow traffic down is also often called Shaping.
1441 <tag>Policing</tag>
1442 Delaying or dropping packets in order to make traffic stay below a
1443 configured bandwidth. In Linux, policing can only drop a packet and not
1444 delay it - there is no 'ingress queue'.
1445 <tag>Work-Conserving</tag>
1446 A work-conserving qdisc always delivers a packet if one is available. In
1447 other words, it never delays a packet if the network adaptor is ready to
1448 send one (in the case of an egress qdisc).
1449 <tag>non-Work-Conserving</tag>
1450 Some queues, like for example the Token Bucket Filter, may need to hold on
1451 to a packet for a certain time in order to limit the bandwidth. This means
1452 that they sometimes refuse to give up a packet, even though they have one
1453 available.
1454 </descrip>
1456 Now that we have our terminology straight, let's see where all these things
1457 are.
1459 <verb>
1460 Userspace programs
1463 +---------------+-----------------------------------------+
1464 | Y |
1465 | -------> IP Stack |
1466 | | | |
1467 | | Y |
1468 | | Y |
1469 | ^ | |
1470 | | / ----------> Forwarding -> |
1471 | ^ / | |
1472 | |/ Y |
1473 | | | |
1474 | ^ Y /-qdisc1-\ |
1475 | | Egress /--qdisc2--\ |
1476 --->->Ingress Classifier ---qdisc3---- | ->
1477 | Qdisc \__qdisc4__/ |
1478 | \-qdiscN_/ |
1480 +----------------------------------------------------------+
1481 </verb>
1482 Thanks to Jamal Hadi Salim for this ascii representation.
1484 The big block represents the kernel. The leftmost arrow represents traffic
1485 entering your machine from the network. It is then fed to the Ingress
1486 Qdisc which may apply Filters to a packet, and decide to drop it. This
1487 is called 'Policing'.
1489 This happens at a very early stage, before it has seen a lot of the kernel.
1490 It is therefore a very good place to drop traffic very early, without
1491 consuming a lot of CPU power.
1493 If the packet is allowed to continue, it may be destined for a local
1494 application, in which case it enters the IP stack in order to be processed,
1495 and handed over to a userspace program. The packet may also be forwarded
1496 without entering an application, in which case it is destined for egress.
1497 Userspace programs may also deliver data, which is then examined and
1498 forwarded to the Egress Classifier.
1500 There it is investigated and enqueued to any of a number of qdiscs. In the
1501 unconfigured default case, there is only one egress qdisc installed, the
1502 pfifo_fast, which always receives the packet. This is called 'enqueueing'.
1504 The packet now sits in the qdisc, waiting for the kernel to ask for
1505 it for transmission over the network interface. This is called 'dequeueing'.
1507 This picture also holds in case there is only one network adaptor - the
1508 arrows entering and leaving the kernel should not be taken too literally.
1509 Each network adaptor has both ingress and egress hooks.
1511 <sect1>Classful Queueing Disciplines
1513 Classful qdiscs are very useful if you have different kinds of traffic which
1514 should have differing treatment. One of the classful qdiscs is called 'CBQ'
1515 , 'Class Based Queueing' and it is so widely mentioned that people identify
1516 queueing with classes solely with CBQ, but this is not the case.
1518 CBQ is merely the oldest kid on the block - and also the most complex one.
1519 It may not always do what you want. This may come as something of a shock
1520 to many who fell for the 'sendmail effect', which teaches us that any
1521 complex technology which doesn't come with documentation must be the best
1522 available.
1524 More about CBQ and its alternatives shortly.
1525 <sect2>Flow within classful qdiscs &amp; classes
1527 When traffic enters a classful qdisc, it needs to be sent to any of the
1528 classes within - it needs to be 'classified'. To determine what to do with a
1529 packet, the so called 'filters' are consulted. It is important to know that
1530 the filters are called from within a qdisc, and not the other way around!
1532 The filters attached to that qdisc then return with a decision, and the
1533 qdisc uses this to enqueue the packet into one of the classes. Each subclass
1534 may try other filters to see if further instructions apply. If not, the
1535 class enqueues the packet to the qdisc it contains.
1537 Besides containing other qdiscs, most classful qdiscs also perform shaping.
1538 This is useful to perform both packet scheduling (with SFQ, for example) and
1539 rate control. You need this in cases where you have a high speed
1540 interface (for example, ethernet) to a slower device (a cable modem).
1542 If you were only to run SFQ, nothing would happen, as packets enter &amp;
1543 leave your router without delay: the output interface is far faster than
1544 your actual link speed. There is no queue to schedule then.
1546 <sect2>The qdisc family: roots, handles, siblings and parents
1548 Each interface has one egress 'root qdisc', by default the earlier mentioned
1549 classless pfifo_fast queueing discipline. Each qdisc can be assigned a
1550 handle, which can be used by later configuration statements to refer to that
1551 qdisc. Besides an egress qdisc, an interface may also have an ingress, which
1552 polices traffic coming in.
1554 The handles of these qdiscs consist of two parts, a major number and a minor
1555 number. It is habitual to name the root qdisc '1:', which is equal to '1:0'.
1556 The minor number of a qdisc is always 0.
1558 Classes need to have the same major number as their parent.
1559 <sect3>How filters are used to classify traffic
1561 Recapping, a typical hierarchy might look like this:
1562 <verb>
1563 root 1:
1565 _1:1_
1566 / | \
1567 / | \
1568 / | \
1569 10: 11: 12:
1570 / \ / \
1571 10:1 10:2 12:1 12:2
1572 </verb>
1574 But don't let this tree fool you! You should *not* imagine the kernel to be
1575 at the apex of the tree and the network below, that is just not the case.
1576 Packets get enqueued and dequeued at the root qdisc, which is the only thing
1577 the kernel talks to.
1579 A packet might get classified in a chain like this:
1581 1: -> 1:1 -> 12: -> 12:2
1583 The packet now resides in a queue in a qdisc attached to class 12:2. In this
1584 example, a filter was attached to each 'node' in the tree, each chosing a
1585 branch to take next. This can make sense. However, this is also possible:
1587 1: -> 12:2
1589 In this case, a filter attached to the root decided to send the packet
1590 directly to 12:2.
1592 <sect3>How packets are dequeued to the hardware
1594 When the kernel decides that it needs to extract packets to send to the
1595 interface, the root qdisc 1: gets a dequeue request, which is passed to
1596 1:1, which is in turn passed to 10:, 11: and 12:, which each query their
1597 siblings, and try to dequeue() from them. In this case, the kernel needs to
1598 walk the entire tree, because only 12:2 contains a packet.
1600 In short, nested classes ONLY talk to their parent qdiscs, never to an
1601 interface. Only the root qdisc gets dequeued by the kernel!
1603 The upshot of this is that classes never get dequeued faster than their
1604 parents allow. And this is exactly what we want: this way we can have SFQ in
1605 an inner class, which doesn't do any shaping, only scheduling, and have a
1606 shaping outer qdisc, which does the shaping.
1607 <sect2>The PRIO qdisc
1609 The PRIO qdisc doesn't actually shape, it only subdivides traffic based on
1610 how you configured your filters. You can consider the PRIO qdisc a kind
1611 of pfifo_fast on stereoids, whereby each band is a separate class instead of
1612 a simple FIFO.
1614 When a packet is enqueued to the PRIO qdisc, a class is chosen based on the
1615 filter commands you gave. By default, three classes are created. These
1616 classes by default contain pure FIFO qdiscs with no internal
1617 structure, but you can replace these by any qdisc you have available.
1619 Whenever a packet needs to be dequeued, class :1 is tried first. Higher
1620 classes are only used if lower bands all did not give up a packet.
1622 This qdisc is very useful in case you want to prioritize certain kinds of
1623 traffic without using only TOS-flags but using all the power of the tc
1624 filters. It can also contain more all qdiscs, whereas pfifo_fast is limited
1625 to simple fifo qdiscs.
1627 Because it doesn't actually shape, the same warning as for SFQ holds: either
1628 use it only if your physical link is really full or wrap it inside a
1629 classful qdisc that does shape. The last holds for almost all cablemodems
1630 and DSL devices.
1632 In formal words, the PRIO qdisc is a Work-Conserving scheduler.
1633 <sect3>PRIO parameters &amp; usage
1635 The following parameters are recognized by tc:
1636 <descrip>
1637 <tag>bands</tag>
1638 Number of bands to create. Each band is in fact a class. If you change this
1639 number, you must also change:
1640 <tag>priomap</tag>
1641 If you do not provide tc filters to classify traffic, the PRIO qdisc looks
1642 at the TC_PRIO priority to decide how to enqueue traffic.
1644 This works just like with the pfifo_fast qdisc mentioned earlier, see there
1645 for lots of detail.
1646 </descrip>
1647 The bands are classes, and are called major:1 to major:3 by default, so if
1648 your PRIO qdisc is called 12:, tc filter traffic to 12:1 to grant it more
1649 priority.
1651 Reiterating, band 0 goes to minor number 1! Band 1 to minor number 2, etc.
1652 <sect3>Sample configuration
1654 We will create this tree:
1655 <verb>
1656 root 1: prio
1657 / | \
1658 1:1 1:2 1:3
1659 | | |
1660 10: 20: 30:
1661 sfq tbf sfq
1662 band 0 1 2
1663 </verb>
1665 Bulk traffic will go to 30:, interactive traffic to 20: or 10:.
1667 Commandlines:
1668 <verb>
1669 # tc qdisc add dev eth0 root handle 1: prio
1670 ## This *instantly* creates classes 1:1, 1:2, 1:3
1672 # tc qdisc add dev eth0 parent 1:1 handle 10: sfq
1673 # tc qdisc add dev eth0 parent 1:2 handle 20: tbf rate 20kbit buffer 1600 limit 3000
1674 # tc qdisc add dev eth0 parent 1:3 handle 30: sfq
1675 </verb>
1677 Now lets's see what we created:
1678 <verb>
1679 # tc -s qdisc ls dev eth0
1680 qdisc sfq 30: quantum 1514b
1681 Sent 0 bytes 0 pkts (dropped 0, overlimits 0)
1683 qdisc tbf 20: rate 20Kbit burst 1599b lat 667.6ms
1684 Sent 0 bytes 0 pkts (dropped 0, overlimits 0)
1686 qdisc sfq 10: quantum 1514b
1687 Sent 132 bytes 2 pkts (dropped 0, overlimits 0)
1689 qdisc prio 1: bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
1690 Sent 174 bytes 3 pkts (dropped 0, overlimits 0)
1691 </verb>
1692 As you can see, band 0 has already had some traffic, and one packet was sent
1693 while running this command!
1695 We now do some bulk data transfer with a tool that properly sets TOS flags,
1696 and take another look:
1697 <verb>
1698 # scp tc ahu@10.0.0.11:./
1699 ahu@10.0.0.11's password:
1700 tc 100% |*****************************| 353 KB 00:00
1701 # tc -s qdisc ls dev eth0
1702 qdisc sfq 30: quantum 1514b
1703 Sent 384228 bytes 274 pkts (dropped 0, overlimits 0)
1705 qdisc tbf 20: rate 20Kbit burst 1599b lat 667.6ms
1706 Sent 2640 bytes 20 pkts (dropped 0, overlimits 0)
1708 qdisc sfq 10: quantum 1514b
1709 Sent 2230 bytes 31 pkts (dropped 0, overlimits 0)
1711 qdisc prio 1: bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
1712 Sent 389140 bytes 326 pkts (dropped 0, overlimits 0)
1713 </verb>
1714 As you can see, all traffic went to handle 30:, which is the lowest priority
1715 band, just as intended. Now to verify that interactive traffic goes to
1716 higher bands, we create some interactive traffic:
1718 <verb>
1719 # tc -s qdisc ls dev eth0
1720 qdisc sfq 30: quantum 1514b
1721 Sent 384228 bytes 274 pkts (dropped 0, overlimits 0)
1723 qdisc tbf 20: rate 20Kbit burst 1599b lat 667.6ms
1724 Sent 2640 bytes 20 pkts (dropped 0, overlimits 0)
1726 qdisc sfq 10: quantum 1514b
1727 Sent 14926 bytes 193 pkts (dropped 0, overlimits 0)
1729 qdisc prio 1: bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
1730 Sent 401836 bytes 488 pkts (dropped 0, overlimits 0)
1731 </verb>
1733 It worked - all additional traffic has gone to 10:, which is our highest
1734 priority qdisc. No traffic was sent to the lowest priority, which previously
1735 received our entire scp.
1737 <sect2>The famous CBQ qdisc
1739 As said before, CBQ is the most complex qdisc available, the most hyped, the
1740 least understood, and probably the trickiest one to get right. This is not
1741 because the authors are evil or incompetent, far from it, it's just that the
1742 CBQ algorithm isn't all that precise and doesn't really match the way Linux
1743 works.
1745 Besides being classful, CBQ is also a shaper and it is in that aspect that
1746 it really doesn't work very well. It should work like this. If you try to
1747 shape a 10mbit/s connection to 1mbit/s, the link should be idle 90% of the
1748 time. If it isn't, we need to throttle so that it IS idle 90% of the time.
1750 This is pretty hard to measure, so CBQ instead derives the idle time from
1751 the number of microseconds that elapse between requests from the hardware
1752 layer for more data. Combined, this can be used to approximate how full or
1753 empty the link is.
1755 This is rather circumspect and doesn't always arrive at proper results. For
1756 example, what if the actual link speed of an interface that is not really
1757 able to transmit the full 100mbit/s of data, perhaps because of a badly
1758 implemented driver? A PCMCIA network card will also never achieve 100mbit/s
1759 because of the way the bus is designed - again, how do we calculate the idle
1760 time?
1762 It gets even worse if we consider not-quite-real network devices like PPP
1763 over Ethernet or PPTP over TCP/IP. The effective bandwidth in that case is
1764 probably determined by the efficiency of pipes to userspace - which is huge.
1766 People who have done measurements discover that CBQ is not always very
1767 accurate and sometimes completely misses the mark.
1769 In many circumstances however it works well. With the documentation provided
1770 here, you should be able to configure it to work well in most cases.
1771 <sect3>CBQ shaping in detail
1773 As said before, CBQ works by making sure that the link is idle just long
1774 enough to bring down the real bandwidth to the configured rate. To do so, it
1775 calculates the time that should pass between average packets.
1777 During operations, the effective idletime is measured using an exponential
1778 weighted moving average (EWMA), which considers recent packets to be
1779 exponentially more important than past ones. The unix loadaverage is
1780 calculated in the same way.
1782 The calculated idle time is substracted from the EWMA measured one, the
1783 resulting number is called 'avgidle'. A perfectly loaded link has an avgidle
1784 of zero: packets arrive exactly once every calculated interval.
1786 An overloaded link has a negative avgidle and if it gets too negative, CBQ
1787 shuts down for a while and is then 'overlimit'.
1789 Conversely, an idle link might amass a huge avgidle, which would then allow
1790 infinite bandwidths after a few hours of silence. To prevent this, avgidle is
1791 capped at maxidle.
1793 If overlimit, in theory, the CBQ could throttle itself for exactly the
1794 amount of time that was calculated to pass between packets, and then pass
1795 one packet, and throttle again. But see the 'minburst' parameter below.
1797 These are parameters you can specify in order to configure shaping:
1798 <descrip>
1799 <tag>avpkt</tag>
1800 Average size of a packet, measured in bytes. Needed for calculating maxidle,
1801 which is derived from maxburst, which is specified in packets.
1802 <tag>bandwidth</tag>
1803 The physical bandwidth of your device, needed for idle time
1804 calculations.
1805 <tag>cell</tag>
1806 The time a packet takes to be transmitted over a device may grow in steps,
1807 based on the packet size. An 800 and an 806 size packet may take just as long
1808 to send, for example - this sets the granularity. Most often set to '8'.
1809 Must be an integral power of two.
1810 <tag>maxburst</tag>
1811 This number of packets is used to calculate maxidle so that when avgidle is
1812 at maxidle, this number of average packets can be burst before avgidle drops
1813 to 0. Set it higher to be more tolerant of bursts. You can't set maxidle
1814 directly, only via this parameter.
1815 <tag>minburst</tag>
1816 As mentioned before, CBQ needs to throttle in case of overlimit. The ideal
1817 solution is to do so for exactly the calculated idle time, and pass 1
1818 packet. However, Unix kernels generally have a hard time scheduling events
1819 shorter than 10ms, so it is better to throttle for a longer period, and then
1820 pass minburst packets in one go, and then sleep minburst times longer.
1822 The time to wait is called the offtime. Higher values of minburst lead to
1823 more accurate shaping in the long term, but to bigger bursts at millisecond
1824 timescales.
1825 <tag>minidle</tag>
1826 If avgidle is below 0, we are overlimits and need to wait until avgidle will
1827 be big enough to send one packet. To prevent a sudden burst from shutting
1828 down the link for a prolonged period of time, avgidle is reset to minidle if
1829 it gets too low.
1831 Minidle is specified in negative microseconds, so 10 means that avgidle is
1832 capped at -10us.
1833 <tag>mpu</tag>
1834 Mininum packet size - needed because even a zero size packet is padded
1835 to 64 bytes on ethernet, and so takes a certain time to transmit. CBQ needs
1836 to know this to accurately calculate the idle time.
1837 <tag>rate</tag>
1838 Desired rate of traffic leaving this qdisc - this is the 'speed knob'!
1839 </descrip>
1841 Internally, CBQ has a lot of finetuning. For example, classes which are
1842 known not to have data enqueued to them aren't queried. Overlimit classes
1843 are penalized by lowering their effective priority. All very smart &amp;
1844 complicated.
1846 <sect3>CBQ classful behaviour
1848 Besides shaping, using the aforementioned idletime approximations, CBQ also
1849 acts like the PRIO queue in the sense that classes can have differing
1850 priorities and that lower priority numbers will be polled before the higher
1851 priority ones.
1853 Each time a packet is requested by the hardware layer to be sent out to the
1854 network, a weighted round robin process ('WRR') starts, beginning with the
1855 lower priority classes.
1857 These are then grouped and queried if they have data available. If so, it is
1858 returned. After a class has been allowed to dequeue a number of bytes, the
1859 next class within that priority is tried.
1861 The following parameters control the WRR process:
1862 <descrip>
1863 <tag>allot</tag>
1864 When the outer cbq is asked for a packet to send out on the interface, it
1865 will try all inner qdiscs (in the classes) in turn, in order of
1866 the 'priority' parameter. Each time a class gets its turn, it can only send out
1867 a limited amount of data. 'Allot' is the base unit of this amount. See
1868 the 'weight' parameter for more information.
1870 <tag>prio</tag>
1871 The CBQ can also act like the PRIO device. Inner classes with lower priority
1872 are tried first and as long as they have traffic, other classes are not
1873 polled for traffic.
1875 <tag>weight</tag>
1876 Weight helps in the Weighted Round Robin process. Each class gets a chance
1877 to send in turn. If you have classes with significantly more bandwidth than
1878 other classes, it makes sense to allow them to send more data in one round
1879 than the others.
1881 A CBQ adds up all weights under a class, and normalizes them, so you can use
1882 arbitrary numbers: only the ratios are important. People have been
1883 using 'rate/10' as a rule of thumb and it appears to work well. The renormalized
1884 weight is multiplied by the 'allot' parameter to determine how much data can
1885 be sent in one round.
1886 </descrip>
1888 Please note that all classes within an CBQ hierarchy need to share the same
1889 major number!
1890 <sect3>CBQ parameters that determine link sharing &amp; borrowing
1892 Besides purely limiting certain kinds of traffic, it is also possible to
1893 specify which classes can borrow capacity from other classes or, conversely,
1894 lend out bandwidth.
1896 <descrip>
1897 <tag>Isolated/sharing</tag>
1898 A class that is configured with 'isolated' will not lend out bandwidth to
1899 sibling classes. Use this if you have competing or mutually-unfriendly
1900 agencies on your link who do want to give eachother freebies.
1902 The control program tc also knows about 'sharing', which is the reverse
1903 of 'isolated'.
1904 <tag>bounded/borrow</tag>
1905 A class can also be 'bounded', which means that it will not try to borrow
1906 bandwidth from sibling classes. tc also knows about 'borrow', which is the
1907 reverse of 'bounded'.
1908 </descrip>
1909 A typical situation might be where you have two agencies on your link which
1910 are both 'isolated' and 'bounded', which means that they are really limited
1911 to their assigned rate, and also won't allow each other to borrow.
1913 Within such an agency class, there might be other classes which are allowed
1914 to swap bandwidth.
1915 <sect3>Sample configuration
1917 This configuration limits webserver traffic to 5mbit and smtp traffic to 3
1918 mbit. Together, they may not get more than 6mbit. We have a 100mbit NIC and
1919 the classes may borrow bandwidth from each other.
1920 <verb>
1921 # tc qdisc add dev eth0 root handle 1:0 cbq bandwidth 100Mbit \
1922 avpkt 1000 cell 8
1923 # tc class add dev eth0 parent 1:0 classid 1:1 cbq bandwidth 100Mbit \
1924 rate 6Mbit weight 0.6Mbit prio 8 allot 1514 cell 8 maxburst 20 \
1925 avpkt 1000 bounded
1926 </verb>
1927 This part installs the root and the customary 1:0 class. The 1:1 class is
1928 bounded, so the total bandwidth can't exceed 6mbit.
1930 As said before, CBQ requires a *lot* of knobs. All parameters are explained
1931 above, however. The corresponding HTB configuration is lots simpler.
1933 <verb>
1934 # tc class add dev eth0 parent 1:1 classid 1:3 cbq bandwidth 100Mbit \
1935 rate 5Mbit weight 0.5Mbit prio 5 allot 1514 cell 8 maxburst 20 \
1936 avpkt 1000
1937 # tc class add dev eth0 parent 1:1 classid 1:4 cbq bandwidth 100Mbit \
1938 rate 3Mbit weight 0.3Mbit prio 5 allot 1514 cell 8 maxburst 20 \
1939 avpkt 1000
1940 </verb>
1942 These are our two classes. Note how we scale the weight with the configured
1943 rate. Both classes are not bounded, but they are connected to class 1:1
1944 which is bounded. So the sum of bandwith of the 2 classes will never be
1945 more than 6mbit. The classid's need to be within the same major number as
1946 the parent CBQ, by the way!
1948 <verb>
1949 # tc qdisc add dev eth0 parent 1:3 handle 30: sfq
1950 # tc qdisc add dev eth0 parent 1:4 handle 40: sfq
1951 </verb>
1953 Both classes have a FIFO qdisc by default. But we replaced these with an SFQ
1954 queue so each flow of data is treated equally.
1955 <verb>
1956 # tc filter add dev eth0 parent 1:0 protocol ip prio 1 u32 match ip \
1957 sport 80 0xffff flowid 1:3
1958 # tc filter add dev eth0 parent 1:0 protocol ip prio 1 u32 match ip \
1959 sport 25 0xffff flowid 1:4
1960 </verb>
1962 These commands, attached directly to the root, send traffic to the right
1963 qdiscs.
1965 Note that we use 'tc class add' to CREATE classes within a qdisc, but that
1966 we use 'tc qdisc add' to actually add qdiscs to these classes.
1968 You may wonder what happens to traffic that is not classified by any of the
1969 two rules. It appears that in this case, data will then be processed within
1970 1:0, and be unlimited.
1972 If smtp+web together try to exceed the set limit of 6mbit/s, bandwidth will
1973 be divided according to the weight parameter, giving 5/8 of traffic to the
1974 webserver and 3/8 to the mailserver.
1976 With this configuratien you can also say that webserver traffic will always
1977 get at minimum 5/8 * 6 mbit = 3.75 mbit.
1978 <sect3>Other CBQ parameters: split &amp; defmap
1980 As said before, a classful qdisc needs to call filters to determine
1981 which class a packet will be enqueued to.
1983 Besides calling the filter, CBQ offers other options, defmap &amp; split.
1984 This is pretty complicated to understand, and it is not vital. But as this
1985 is the only known place where defmap &amp; split are properly explained, I'm
1986 doing my best.
1988 As you will often want to filter on the Type of Service field only, a special
1989 syntax is provided. Whenever the CBQ needs to figure out where a packet
1990 needs to be enqueued, it checks if this node is a 'split node'. If so, one
1991 of the sub-qdiscs has indicated that it wishes to receive all packets with
1992 a certain configured priority, as might be derived from the TOS field, or
1993 socket options set by applications.
1995 The packets' priority bits are or-ed with the defmap field to see if a match
1996 exists. In other words, this is a short-hand way of creating a very fast
1997 filter, which only matches certain priorities. A defmap of ff (hex) will
1998 match everything, a map of 0 nothing. A sample configuration may help make
1999 things clearer:
2001 <verb>
2002 # tc qdisc add dev eth1 root handle 1: cbq bandwidth 10Mbit allot 1514 \
2003 cell 8 avpkt 1000 mpu 64
2005 # tc class add dev eth1 parent 1:0 classid 1:1 cbq bandwidth 10Mbit \
2006 rate 10Mbit allot 1514 cell 8 weight 1Mbit prio 8 maxburst 20 \
2007 avpkt 1000
2008 </verb>
2009 Standard CBQ preamble. I never get used to the sheer amount of numbers
2010 required!
2012 Defmap refers to TC_PRIO bits, which are defined as follows:
2014 <verb>
2015 TC_PRIO.. Num Corresponds to TOS
2016 -------------------------------------------------
2017 BESTEFFORT 0 Maximize Reliablity
2018 FILLER 1 Minimize Cost
2019 BULK 2 Maximize Throughput (0x8)
2020 INTERACTIVE_BULK 4
2021 INTERACTIVE 6 Minimize Delay (0x10)
2022 CONTROL 7
2023 </verb>
2025 The TC_PRIO.. number corresponds to bits, counted from the right. See the
2026 pfifo_fast section for more details how TOS bits are converted to
2027 priorities.
2029 Now the interactive and the bulk classes:
2031 <verb>
2032 # tc class add dev eth1 parent 1:1 classid 1:2 cbq bandwidth 10Mbit \
2033 rate 1Mbit allot 1514 cell 8 weight 100Kbit prio 3 maxburst 20 \
2034 avpkt 1000 split 1:0 defmap c0
2036 # tc class add dev eth1 parent 1:1 classid 1:3 cbq bandwidth 10Mbit \
2037 rate 8Mbit allot 1514 cell 8 weight 800Kbit prio 7 maxburst 20 \
2038 avpkt 1000 split 1:0 defmap 3f
2039 </verb>
2041 The 'split qdisc' is 1:0, which is where the choice will be made. C0 is
2042 binary for 11000000, 3F for 00111111, so these two together will match
2043 everything. The first class matches bits 7 & 6, and thus corresponds
2044 to 'interactive' and 'control' traffic. The second class matches the rest.
2046 Node 1:0 now has a table like this:
2047 <verb>
2048 priority send to
2049 0 1:3
2050 1 1:3
2051 2 1:3
2052 3 1:3
2053 4 1:3
2054 5 1:3
2055 6 1:2
2056 7 1:2
2057 </verb>
2059 For additional fun, you can also pass a 'change mask', which indicates
2060 exactly which priorities you wish to change. You only need to use this if you
2061 are running 'tc class change'. For example, to add best effort traffic to
2062 1:2, we could run this:
2064 <verb>
2065 # tc class change dev eth1 classid 1:2 cbq defmap 01/01
2066 </verb>
2068 The priority map over at 1:0 now looks like this:
2070 <verb>
2071 priority send to
2072 0 1:2
2073 1 1:3
2074 2 1:3
2075 3 1:3
2076 4 1:3
2077 5 1:3
2078 6 1:2
2079 7 1:2
2080 </verb>
2082 FIXME: did not test 'tc class change', only looked at the source.
2083 <sect2>Hierarchical Token Bucket
2085 Martin Devera (&lt;devik&gt;) rightly realised that CBQ is complex and does
2086 not seem optimized for many typical situations. His Hierarchial approach is
2087 well suited for setups where you have a fixed amount of bandwidth which you
2088 want to divide for different purposes, giving each purpose a guaranteed
2089 bandwidth, with the possibility of specifying how much bandwidth can be
2090 borrowed.
2092 HTB works just like CBQ but does not resort to idle time calculations to
2093 shape. Instead, it is a classful Token Bucket Filter - hence the name. It
2094 has only a few parameters, which are well documented on his
2095 <url url="http://luxik.cdi.cz/~devik/qos/htb/"
2096 name="site">.
2098 As your HTB configuration gets more complex, your configuration scales
2099 well. With CBQ it is already complex even in simple cases! HTB is not yet a
2100 part of the standard kernel, but it should soon be!
2102 If you are in a position to patch your kernel, by all means consider HTB.
2103 <sect3>Sample configuration
2105 Functionally almost identical to the CBQ sample configuration above:
2107 <tscreen><verb>
2108 # tc qdisc add dev eth0 root handle 1: htb default 30
2110 # tc class add dev eth0 parent 1: classid 1:1 htb rate 6mbit burst 15k
2112 # tc class add dev eth0 parent 1:1 classid 1:10 htb rate 5mbit burst 15k
2113 # tc class add dev eth0 parent 1:1 classid 1:20 htb rate 3mbit ceil 6mbit burst 15k
2114 # tc class add dev eth0 parent 1:1 classid 1:30 htb rate 1kbit ceil 6mbit burst 15k
2115 </verb></tscreen>
2117 The author then recommends SFQ for beneath these classes:
2118 <tscreen><verb>
2119 # tc qdisc add dev eth0 parent 1:10 handle 10: sfq perturb 10
2120 # tc qdisc add dev eth0 parent 1:20 handle 20: sfq perturb 10
2121 # tc qdisc add dev eth0 parent 1:30 handle 30: sfq perturb 10
2122 </verb></tscreen>
2124 Add the filters which direct traffic to the right classes:
2125 <tscreen><verb>
2126 # U32="tc filter add dev eth0 protocol ip parent 1:0 prio 1 u32"
2127 # $U32 match ip dport 80 0xffff flowid 1:10
2128 # $U32 match ip sport 25 0xffff flowid 1:20
2129 </verb></tscreen>
2130 And that's it - no unsightly unexplained numbers, no undocumented
2131 parameters.
2133 HTB certainly looks wonderful - if 10: and 20: both have their guaranteed
2134 bandwidth, and more is left to divide, they borrow in a 5:3 ratio, just as
2135 you would expect.
2137 Unclassified traffic gets routed to 30:, which has little bandwidth of its
2138 own but can borrow everything that is left over. Because we chose SFQ
2139 internally, we get fairness thrown in for free!
2141 <sect1>Classifying packets with filters
2143 To determine which class shall process a packet, the so-called 'classifier
2144 chain' is called each time a choice needs to be made. This chain consists of
2145 all filters attached to the classful qdisc that needs to decide.
2147 To reiterate the tree, which is not a tree:
2148 <verb>
2149 root 1:
2151 _1:1_
2152 / | \
2153 / | \
2154 / | \
2155 10: 11: 12:
2156 / \ / \
2157 10:1 10:2 12:1 12:2
2158 </verb>
2160 When enqueueing a packet, at each branch the filter chain is consulted for a
2161 relevant instruction. A typical setup might be to have a filter in 1:1 that
2162 directs a packet to 12: and a filter on 12: that sends the packet to 12:2.
2164 You might also attach this latter rule to 1:1, but you can make efficiency
2165 gains by having more specific tests lower in the chain.
2167 You can't filter a packet 'upwards', by the way. Also, with HTB, you should
2168 attach all filters to the root!
2170 And again - packets are only enqueued downwards! When they are dequeued,
2171 they go up again, where the interface lives. They do NOT fall off the end of
2172 the tree to the network adaptor!
2174 <sect2>Some simple filtering examples
2176 As explained in the Classifier chapter, you can match on literally anything,
2177 using a very complicated syntax. To start, we will show how to do the
2178 obvious things, which luckily are quite easy.
2180 Let's say we have a PRIO qdisc called '10:' which contains three classes, and
2181 we want to assign all traffic from and to port 22 to the highest priority
2182 band, the filters would be:
2184 <tscreen><verb>
2185 # tc filter add dev eth0 protocol ip parent 10: prio 1 u32 match \
2186 ip dport 22 0xffff flowid 10:1
2187 # tc filter add dev eth0 protocol ip parent 10: prio 1 u32 match \
2188 ip sport 80 0xffff flowid 10:1
2189 # tc filter add dev eth0 protocol ip parent 10: prio 2 flowid 10:2
2190 </verb></tscreen>
2192 What does this say? It says: attach to eth0, node 10: a priority 1 u32
2193 filter that matches on IP destination port 22 *exactly* and send it to band
2194 10:1. And it then repeats the same for source port 80. The last command says
2195 that anything unmatched so far should go to band 10:2, the next-highest
2196 priority.
2198 You need to add 'eth0', or whatever your interface is called, because each
2199 interface has a unique namespace of handles.
2201 To select on an IP address, use this:
2202 <tscreen><verb>
2203 # tc filter add dev eth0 parent 10:0 protocol ip prio 1 u32 \
2204 match ip dst 4.3.2.1/32 flowid 10:1
2205 # tc filter add dev eth0 parent 10:0 protocol ip prio 1 u32 \
2206 match ip src 1.2.3.4/32 flowid 10:1
2207 # tc filter add dev eth0 protocol ip parent 10: prio 2 \
2208 flowid 10:2
2209 </verb></tscreen>
2211 This assigns traffic to 4.3.2.1 and traffic from 1.2.3.4 to the highest
2212 priority queue, and the rest to the next-highest one.
2214 You can concatenate matches, to match on traffic from 1.2.3.4 and from port
2215 80, do this:
2216 <tscreen><verb>
2217 # tc filter add dev eth0 parent 10:0 protocol ip prio 1 u32 match ip src 4.3.2.1/32
2218 match ip sport 80 0xffff flowid 10:1
2219 </verb></tscreen>
2221 <sect2>All the filtering commands you will normally need
2223 Most shaping commands presented here start with this preamble:
2224 <verb>
2225 # tc filter add dev eth0 parent 1:0 protocol ip prio 1 u32 ..
2226 </verb>
2227 These are the so called 'u32' matches, which can match on ANY part of a
2228 packet.
2229 <descrip>
2230 <tag>On source/destination address</tag>
2231 Source mask 'match ip src 1.2.3.0/24', destination mask 'match ip dst
2232 4.3.2.0/24'. To match a single host, use /32, or omit the mask.
2233 <tag>On source/destination port, all IP protocols</tag>
2234 Source: 'match ip sport 80 0xffff', 'match ip dport 0xffff'
2235 <tag>On ip protocol (tcp, udp, icmp, gre, ipsec)</tag>
2236 Use the numbers from /etc/protocols, for example, icmp is 1: 'match ip
2237 protocol 1 0xff'.
2238 <tag>On fwmark</tag>
2239 You can mark packets with either ipchains and have that mark survive routing
2240 across interfaces. This is really useful to for example only shape traffic on
2241 eth1 that came in on eth0. Syntax:
2242 # tc filter add dev eth1 protocol ip parent 1:0 prio 1 handle 6 fw flowid 1:1
2243 Note that this is not a u32 match!
2245 You can place a mark like this:
2246 <verb>
2247 # iptables -A PREROUTING -t mangle -i eth0 -j MARK --set-mark 6
2248 </verb>
2249 The number 6 is arbitrary.
2251 If you don't want to understand the full tc filter syntax, just use
2252 iptables, and only learn to select on fwmark.
2253 <tag>On the TOS field</tag>
2254 To select interactive, minimum delay traffic:
2255 <verb>
2256 # tc filter add dev ppp0 parent 1:0 protocol ip prio 10 u32 \
2257 match ip tos 0x10 0xff \
2258 flowid 1:4
2259 </verb>
2260 Use 0x08 0xff for bulk traffic.
2261 </descrip>
2263 For more filtering commands, see the Advanced Filters chapter.
2264 <sect>Loadsharing over multiple interfaces
2266 There are several ways of doing this. One of the easiest and straightforward
2267 ways is 'TEQL' - "True" (or "trivial") link equalizer. Like most things
2268 having to do with queueing, loadsharing goes both ways. Both ends of a link
2269 may need to participate for full effect.
2271 Imagine this situation:
2273 <tscreen><verb>
2274 +-------+ eth1 +-------+
2275 | |==========| |
2276 'network 1' ----| A | | B |---- 'network 2'
2277 | |==========| |
2278 +-------+ eth2 +-------+
2279 </verb></tscreen>
2281 A and B are routers, and for the moment we'll assume both run Linux. If
2282 traffic is going from network 1 to network 2, router A needs to distribute
2283 the packets over both links to B. Router B needs to be configured to accept
2284 this. Same goes the other way around, when packets go from network 2 to
2285 network 1, router B needs to send the packets over both eth1 and eth2.
2287 The distributing part is done by a 'TEQL' device, like this (it couldn't be
2288 easier):
2290 <tscreen><verb>
2291 # tc qdisc add dev eth1 root teql0
2292 # tc qdisc add dev eth2 root teql0
2293 # ip link set dev teql0 up
2294 </verb></tscreen>
2296 Don't forget the 'ip link set up' command!
2298 This needs to be done on both hosts. The device teql0 is basically a
2299 roundrobbin distributor over eth1 and eth2, for sending packets. No data
2300 ever comes in over an teql device, that just appears on the 'raw' eth1 and
2301 eth2.
2303 But now we just have devices, we also need proper routing. One way to do
2304 this is to assign a /31 network to both links, and a /31 to the teql0 device
2305 as well:
2307 FIXME: does this need something like 'nobroadcast'? A /31 is too small to
2308 house a network address and a broadcast address - if this doesn't work as
2309 planned, try a /30, and adjust the ip adresses accordingly. You might even
2310 try to make eth1 and eth2 do without an IP address!
2312 On router A:
2313 <tscreen><verb>
2314 # ip addr add dev eth1 10.0.0.0/31
2315 # ip addr add dev eth2 10.0.0.2/31
2316 # ip addr add dev teql0 10.0.0.4/31
2317 </verb></tscreen>
2319 On router B:
2320 <tscreen><verb>
2321 # ip addr add dev eth1 10.0.0.1/31
2322 # ip addr add dev eth2 10.0.0.3/31
2323 # ip addr add dev teql0 10.0.0.5/31
2324 </verb></tscreen>
2326 Router A should now be able to ping 10.0.0.1, 10.0.0.3 and 10.0.0.5 over the
2327 2 real links and the 1 equalized device. Router B should be able to ping
2328 10.0.0.0, 10.0.0.2 and 10.0.0.4 over the links.
2330 If this works, Router A should make 10.0.0.5 its route for reaching network
2331 2, and Router B should make 10.0.0.4 its route for reaching network 1. For
2332 the special case where network 1 is your network at home, and network 2 is
2333 the Internet, Router A should make 10.0.0.5 its default gateway.
2335 <sect1>Caveats
2337 Nothing is as easy as it seems. eth1 and eth2 on both router A and B need to
2338 have return path filtering turned off, because they will otherwise drop
2339 packets destined for ip addresses other than their own:
2341 <tscreen><verb>
2342 # echo 0 > /proc/net/ipv4/conf/eth1/rp_filter
2343 # echo 0 > /proc/net/ipv4/conf/eth2/rp_filter
2344 </verb></tscreen>
2346 Then there is the nasty problem of packet reordering. Let's say 6 packets
2347 need to be sent from A to B - eth1 might get 1, 3 and 5. eth2 would then do
2348 2, 4 and 6. In an ideal world, router B would receive this in order, 1, 2,
2349 3, 4, 5, 6. But the possibility is very real that the kernel gets it like
2350 this: 2, 1, 4, 3, 6, 5. The problem is that this confuses TCP/IP. While not
2351 a problem for links carrying many different TCP/IP sessions, you won't be
2352 able to to a bundle multiple links and get to ftp a single file lots faster,
2353 except when your receiving or sending OS is Linux, which is not easily
2354 shaken by some simple reordering.
2356 However, for lots of applications, link loadbalancing is a great idea.
2359 <sect>Netfilter &amp; iproute - marking packets
2361 So far we've seen how iproute works, and netfilter was mentioned a few
2362 times. This would be a good time to browse through <url name="Rusty's Remarkably
2363 Unreliable Guides"
2364 url="http://netfilter.samba.org/unreliable-guides/">. Netfilter itself
2365 can be found <url name="here"
2366 url="http://netfilter.filewatcher.org/">.
2368 Netfilter allows us to filter packets, or mangle their headers. One special
2369 feature is that we can mark a packet with a number. This is done with the
2370 --set-mark facility.
2372 As an example, this command marks all packets destined for port 25, outgoing
2373 mail:
2375 <tscreen><verb>
2376 # iptables -A PREROUTING -i eth0 -t mangle -p tcp --dport 25 \
2377 -j MARK --set-mark 1
2378 </verb></tscreen>
2380 Let's say that we have multiple connections, one that is fast (and
2381 expensive, per megabyte) and one that is slower, but flat fee. We would most
2382 certainly like outgoing mail to go via the cheap route.
2384 We've already marked the packets with a '1', we now instruct the routing
2385 policy database to act on this:
2387 <tscreen><verb>
2388 # echo 201 mail.out >> /etc/iproute2/rt_tables
2389 # ip rule add fwmark 1 table mail.out
2390 # ip rule ls
2391 0: from all lookup local
2392 32764: from all fwmark 1 lookup mail.out
2393 32766: from all lookup main
2394 32767: from all lookup default
2395 </verb></tscreen>
2397 Now we generate the mail.out table with a route to the slow but cheap link:
2398 <tscreen><verb>
2399 # /sbin/ip route add default via 195.96.98.253 dev ppp0 table mail.out
2400 </verb></tscreen>
2402 And we are done. Should we want to make exceptions, there are lots of ways
2403 to achieve this. We can modify the netfilter statement to exclude certain
2404 hosts, or we can insert a rule with a lower priority that points to the main
2405 table for our excepted hosts.
2407 We can also use this feature to honour TOS bits by marking packets with a
2408 different type of service with different numbers, and creating rules to act
2409 on that. This way you can even dedicate, say, an ISDN line to interactive
2410 sessions.
2412 Needless to say, this also works fine on a host that's doing NAT
2413 ('masquerading').
2415 IMPORTANT: We received a report that MASQ and SNAT at least collide
2416 with marking packets. Rusty Russell explains it in
2417 <url
2418 url="http://lists.samba.org/pipermail/netfilter/2000-November/006089.html"
2419 name="this posting">. Turn off the reverse path filter to make it work
2420 properly.
2422 Note: to mark packets, you need to have some options enabled in your
2423 kernel:
2425 <tscreen><verb>
2426 IP: advanced router (CONFIG_IP_ADVANCED_ROUTER) [Y/n/?]
2427 IP: policy routing (CONFIG_IP_MULTIPLE_TABLES) [Y/n/?]
2428 IP: use netfilter MARK value as routing key (CONFIG_IP_ROUTE_FWMARK) [Y/n/?]
2429 </verb></tscreen>
2431 See also <ref id="SQUID" name="Transparent web-caching using netfilter, iproute2, ipchains and squid">
2432 in the Cookbook.
2433 <sect>Advanced filters for (re-)classifying packets
2435 As explained in the section on classful queueing disciplines, filters are
2436 needed to classify packets into any of the sub-queues. These filters are
2437 called from within the classful qdisc.
2439 Here is an incomplete list of classifiers available:
2440 <descrip>
2441 <tag>fw</tag>
2442 Bases the decision on how the firewall has marked the packet. This can be
2443 the easy way out if you don't want to learn tc filter syntax. See the
2444 Queueing chapter for details.
2446 <tag>u32</tag>
2447 Bases the decision on fields within the packet (i.e. source IP address, etc)
2449 <tag>route</tag>
2450 Bases the decision on which route the packet will be routed by
2452 <tag>rsvp, rsvp6</tag>
2453 Routes packets based on <url
2454 url="http://www.isi.edu/div7/rsvp/overview.html" name="RSVP ">. Only useful
2455 on networks you control - the Internet does not respect RSVP.
2457 <tag>tcindex</tag>
2458 Used in the DSMARK qdisc, see the relevant section.
2459 </descrip>
2461 Note that in general there are many ways in which you can classify packet
2462 and that it generally comes down to preference as to which system you wish
2463 to use.
2465 Classifiers in general accept a few arguments in common. They are listed
2466 here for convenience:
2468 <descrip>
2469 <tag>protocol</tag>
2470 The protocol this classifier will accept. Generally you will only be
2471 accepting only IP traffic. Required.
2473 <tag>parent</tag>
2474 The handle this classifier is to be attached to. This handle must be
2475 an already existing class. Required.
2477 <tag>prio</tag>
2478 The priority of this classifier. Lower numbers get tested first.
2480 <tag>handle</tag>
2481 This handle means different things to different filters.
2483 </descrip>
2485 All the following sections will assume you are trying to shape the traffic
2486 going to <tt>HostA</tt>. They will assume that the root class has been
2487 configured on 1: and that the class you want to send the selected traffic to
2488 is 1:1.
2491 <sect1>The "u32" classifier
2493 The U32 filter is the most advanced filter available in the current
2494 implementation. It entirely based on hashing tables, which make it
2495 robust when there are many filter rules.
2497 In its simplest form the U32 filter is a list of records, each
2498 consisting of two fields: a selector and an action. The selectors,
2499 described below, are compared with the currently processed IP packet
2500 until the first match occurs, and then the associated action is performed.
2501 The simplest type of action would be directing the packet into defined
2502 CBQ class.
2504 The commandline of <tt>tc filter</tt> program, used to configure the filter,
2505 consists of three parts: filter specification, a selector and an action.
2506 The filter specification can be defined as:
2508 <tscreen><verb>
2509 tc filter add dev IF [ protocol PROTO ]
2510 [ (preference|priority) PRIO ]
2511 [ parent CBQ ]
2512 </verb></tscreen>
2514 The <tt>protocol</tt> field describes protocol that the filter will be
2515 applied to. We will only discuss case of <tt>ip</tt> protocol. The
2516 <tt>preference</tt> field (<tt>priority</tt> can be used alternatively)
2517 sets the priority of currently defined filter. This is important, since
2518 you can have several filters (lists of rules) with different priorities.
2519 Each list will be passed in the order the rules were added, then list with
2520 lower priority (higher preference number) will be processed. The <tt>parent</tt>
2521 field defines the CBQ tree top (e.g. 1:0), the filter should be attached
2524 The options decribed above apply to all filters, not only U32.
2526 <sect2>U32 selector
2528 The U32 selector contains definition of the pattern, that will be matched
2529 to the currently processed packet. Precisely, it defines which bits are
2530 to be matched in the packet header and nothing more, but this simple
2531 method is very powerful. Let's take a look at the following examples,
2532 taken directly from a pretty complex, real-world filter:
2534 <tscreen><verb>
2535 # tc filter add dev eth0 protocol ip parent 1:0 pref 10 u32 \
2536 match u32 00100000 00ff0000 at 0 flowid 1:10
2537 </verb></tscreen>
2540 For now, leave the first line alone - all these parameters describe
2541 the filter's hash tables. Focus on the selector line, containing
2542 <tt>match</tt> keyword. This selector will match to IP headers, whose
2543 second byte will be 0x10 (0010). As you can guess, the 00ff number is
2544 the match mask, telling the filter exactly which bits to match. Here
2545 it's 0xff, so the byte will match if it's exactly 0x10. The <tt>at</tt>
2546 keyword means that the match is to be started at specified offset (in
2547 bytes) -- in this case it's beginning of the packet. Translating all
2548 that to human language, the packet will match if its Type of Service
2549 field will have `low delay' bits set. Let's analyze another rule:
2551 <tscreen><verb>
2552 # tc filter add dev eth0 protocol ip parent 1:0 pref 10 u32 \
2553 match u32 00000016 0000ffff at nexthdr+0 flowid 1:10
2554 </verb></tscreen>
2557 The <tt>nexthdr</tt> option means next header encapsulated in the IP packet,
2558 i.e. header of upper-layer protocol. The match will also start here
2559 at the beginning of the next header. The match should occur in the
2560 second, 32-bit word of the header. In TCP and UDP protocols this field
2561 contains packet's destination port. The number is given in big-endian
2562 format, i.e. older bits first, so we simply read 0x0016 as 22 decimal,
2563 which stands for SSH service if this was TCP. As you guess, this match
2564 is ambigous without a context, and we will discuss this later.
2567 Having understood all the above, we will find the following selector
2568 quite easy to read: <tt>match c0a80100 ffffff00 at 16</tt>. What we
2569 got here is a three byte match at 17-th byte, counting from the IP
2570 header start. This will match for packets with destination address
2571 anywhere in 192.168.1/24 network. After analyzing the examples, we
2572 can summarize what we have learnt.
2574 <sect2>General selectors
2577 General selectors define the pattern, mask and offset the pattern
2578 will be matched to the packet contents. Using the general selectors
2579 you can match virtually any single bit in the IP (or upper layer)
2580 header. They are more difficult to write and read, though, than
2581 specific selectors that described below. The general selector syntax
2584 <tscreen><verb>
2585 match [ u32 | u16 | u8 ] PATTERN MASK [ at OFFSET | nexthdr+OFFSET]
2586 </verb></tscreen>
2589 One of the keywords <tt>u32</tt>, <tt>u16</tt> or <tt>u8</tt> specifies
2590 length of the pattern in bits. PATTERN and MASK should follow, of length
2591 defined by the previous keyword. The OFFSET parameter is the offset,
2592 in bytes, to start matching. If <tt>nexthdr+</tt> keyword is given,
2593 the offset is relative to start of the upper layer header.
2596 Some examples:
2598 <tscreen><verb>
2599 # tc filter add dev ppp14 parent 1:0 prio 10 u32 \
2600 match u8 64 0xff at 8 \
2601 flowid 1:4
2602 </verb></tscreen>
2605 Packet will match to this rule, if its time to live (TTL) is 64.
2606 TTL is the field starting just after 8-th byte of the IP header.
2608 <tscreen><verb>
2609 # tc filter add dev ppp14 parent 1:0 prio 10 u32 \
2610 match u8 0x10 0xff at nexthdr+13 \
2611 protocol tcp \
2612 flowid 1:3
2613 </verb></tscreen>
2615 FIXME: it has been pointed out that this syntax does not work currently.
2617 Use this to match ACKs on packets smaller than 64 bytes:
2619 <tscreen><verb>
2620 ## match acks the hard way,
2621 ## IP protocol 6,
2622 ## IP header length 0x5(32 bit words),
2623 ## IP Total length 0x34 (ACK + 12 bytes of TCP options)
2624 ## TCP ack set (bit 5, offset 33)
2625 # tc filter add dev ppp14 parent 1:0 protocol ip prio 10 u32 \
2626 match ip protocol 6 0xff \
2627 match u8 0x05 0x0f at 0 \
2628 match u16 0x0000 0xffc0 at 2 \
2629 match u8 0x10 0xff at 33 \
2630 flowid 1:3
2631 </verb></tscreen>
2635 This rule will only match TCP packets with ACK bit set, and no further
2636 payload. Here we can see an example of using two selectors, the final result
2637 will be logical AND of their results. If we take a look at TCP header
2638 diagram, we can see that the ACK bit is second older bit (0x10) in the 14-th
2639 byte of the TCP header (<tt>at nexthdr+13</tt>). As for the second
2640 selector, if we'd like to make our life harder, we could write <tt>match u8
2641 0x06 0xff at 9</tt> instead of using the specific selector <tt>protocol
2642 tcp</tt>, because 6 is the number of TCP protocol, present in 10-th byte of
2643 the IP header. On the other hand, in this example we couldn't use any
2644 specific selector for the first match - simply because there's no specific
2645 selector to match TCP ACK bits.
2647 <sect2>Specific selectors
2649 The following table contains a list of all specific selectors
2650 the author of this section has found in the <tt>tc</tt> program
2651 source code. They simply make your life easier and increase readability
2652 of your filter's configuration.
2654 FIXME: table placeholder - the table is in separate file ,,selector.html''
2656 FIXME: it's also still in Polish :-(
2658 FIXME: must be sgml'ized
2660 Some examples:
2663 <tscreen><verb>
2664 # tc filter add dev ppp0 parent 1:0 prio 10 u32 \
2665 match ip tos 0x10 0xff \
2666 flowid 1:4
2667 </verb></tscreen>
2669 FIXME: tcp dst match does not work as described below:
2671 The above rule will match packets which have the TOS field set to 0x10.
2672 The TOS field starts at second byte of the packet and is one byte big,
2673 so we could write an equivalent general selector: <tt>match u8 0x10 0xff
2674 at 1</tt>. This gives us hint to the internals of U32 filter -- the
2675 specific rules are always translated to general ones, and in this
2676 form they are stored in the kernel memory. This leads to another conclusion
2677 -- the <tt>tcp</tt> and <tt>udp</tt> selectors are exactly the same
2678 and this is why you can't use single <tt>match tcp dst 53 0xffff</tt>
2679 selector to match TCP packets sent to given port -- they will also
2680 match UDP packets sent to this port. You must remember to also specify
2681 the protocol and end up with the following rule:
2683 <tscreen><verb>
2684 # tc filter add dev ppp0 parent 1:0 prio 10 u32 \
2685 match tcp dst 53 0xffff \
2686 match ip protocol 0x6 0xff \
2687 flowid 1:2
2688 </verb></tscreen>
2690 <!--
2691 TODO:
2693 describe more options
2695 match
2696 offset
2697 hashkey
2698 classid | flowid
2699 divisor
2700 order
2701 link
2703 sample
2704 police
2708 <sect1>The "route" classifier
2711 This classifier filters based on the results of the routing tables. When a
2712 packet that is traversing through the classes reaches one that is marked
2713 with the "route" filter, it splits the packets up based on information in
2714 the routing table.
2716 <tscreen><verb>
2717 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 route
2718 </verb></tscreen>
2720 Here we add a route classifier onto the parent node 1:0 with priority 100.
2721 When a packet reaches this node (which, since it is the root, will happen
2722 immediately) it will consult the routing table and if one matches will
2723 send it to the given class and give it a priority of 100. Then, to finally
2724 kick it into action, you add the appropriate routing entry:
2726 The trick here is to define 'realm' based on either destination or source.
2727 The way to do it is like this:
2729 <tscreen><verb>
2730 # ip route add Host/Network via Gateway dev Device realm RealmNumber
2731 </verb></tscreen>
2733 For instance, we can define our destination network 192.168.10.0 with a realm
2734 number 10:
2736 <tscreen><verb>
2737 # ip route add 192.168.10.0/24 via 192.168.10.1 dev eth1 realm 10
2738 </verb></tscreen>
2740 When adding route filters, we can use realm numbers to represent the
2741 networks or hosts and specify how the routes match the filters.
2743 <tscreen><verb>
2744 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 \
2745 route to 10 classid 1:10
2746 </verb></tscreen>
2748 The above rule says packets going to the network 192.168.10.0 match class id
2749 1:10.
2751 Route filter can also be used to match source routes. For example, there is
2752 a subnetwork attached to the Linux router on eth2.
2754 <tscreen><verb>
2755 # ip route add 192.168.2.0/24 dev eth2 realm 2
2756 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 \
2757 route from 2 classid 1:2
2758 </verb></tscreen>
2760 Here the filter specifies that packets from the subnetwork 192.168.2.0
2761 (realm 2) will match class id 1:2.
2763 <sect1>Policing filters
2765 To make even more complicated setups possible, you can have filters that
2766 only match up to a certain bandwidth. You can declare a filter to entirely
2767 cease matching above a certain rate, or only to not match only the bandwidth
2768 exceeding a certain rate.
2770 So if you decided to police at 4mbit/s, but 5mbit/s of traffic is present,
2771 you can stop matching either the entire 5mbit/s, or only not match 1mbit/s,
2772 and do send 4mbit/s to the configured class.
2774 If bandwidth exceeds the configured rate, you can drop a packet, reclassify
2775 it, or see if another filter will match it.
2777 <sect2>Ways to police
2779 There are basically two ways to police. If you compiled the kernel
2780 with 'Estimators', the kernel can measure for each filter how much traffic
2781 it is passing, more or less. These estimators are very easy on the CPU, as
2782 they simply count 25 times per second how many data has been passed, and
2783 calculate the bitrate from that.
2785 The other way works again via a Token Bucket Filter, this time living within
2786 your filter. The TBF only matches traffic UP TO your configured bandwidth,
2787 if more is offered, only the excess is subject to the configured overlimit
2788 action.
2790 <sect3>With the kernel estimator
2792 This is very simple and has only one parameter: avrate. Either the flow
2793 remains below avrate, and the filter classifies the traffic to the classid
2794 configured, or your rate exceeds it in which case the specified action is
2795 taken, which is 'reclassify' by default.
2797 The kernel uses an Exponential Weighted Moving Average for your bandwidth
2798 which makes it less sensitive to short bursts.
2800 <sect3>With Token Bucket Filter
2802 Uses the following parameters:
2803 <itemize>
2804 <item>buffer/maxburst
2805 <item>mtu/minburst
2806 <item>mpu
2807 <item>rate
2808 </itemize>
2810 Which behave mostly identical to those described in the Token Bucket Filter
2811 section. Please note however that if you set the mtu of a TBF policer too
2812 low, *no* packets will pass, whereas the egress TBF qdisc will just pass
2813 them slower.
2815 Another difference is that a policer can only let a packet pass, or drop it.
2816 It cannot delay hold on to it in order to delay it.
2817 <sect2>Overlimit actions
2819 If your filter decides that it is overlimit, it can take 'actions'.
2820 Currently, three actions are available:
2821 <descrip>
2822 <tag>continue</tag>
2823 Causes this filter not to match, but perhaps other filters will.
2824 <tag>drop</tag>
2825 This is a very fierce option which simply discards traffic exceeding a
2826 certain rate. It is often used in the ingress policer and has limited uses.
2827 For example, you may have a nameserver that falls over if offered more than
2828 5mbit/s of packets, in which case an ingress filter could be used to make
2829 sure no more is ever offered.
2830 <tag>Pass/OK</tag>
2831 Pass on traffic ok. Might be used to disable a complicated filter, but leave
2832 it in place.
2833 <tag>reclassify</tag>
2834 Most often comes down to reclassification to Best Effort. This is the
2835 default action.
2836 </descrip>
2838 <sect2>Examples
2840 The only real example known is mentioned in the 'Protecting your host
2841 from SYN floods' section.
2843 FIXME: if you have used this, please share your experience with us
2845 <sect1>Hashing filters for very fast massive filtering
2847 If you have a need for thousands of rules, for example if you have a lot of
2848 clients or computers, all with different QoS specifications, you may find
2849 that the kernel spends a lot of time matching all those rules.
2851 By default, all filters reside in one big chain which is matched in
2852 descending order of priority. If you have 1000 rules, 1000 checks may be
2853 needed to determine what to do with a packet.
2855 Matching would go much quicker if you would have 256 chains with each four
2856 rules - if you could divide packets over those 256 chains, so that the right
2857 rule will be there.
2859 Hashing makes this possible. Let's say you have 1024 cablemodem customers in
2860 your network, with IP addresses ranging from 1.2.0.0 to 1.2.3.255, and each
2861 has to go in another bin, for example 'lite', 'regular' and 'premium'. You
2862 would then have 1024 rules like this:
2864 <verb>
2865 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2866 1.2.0.0 classid 1:1
2867 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2868 1.2.0.1 classid 1:1
2870 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2871 1.2.3.254 classid 1:3
2872 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2873 1.2.3.255 classid 1:2
2874 </verb>
2876 To speed this up, we can use the last part of the IP address as a 'hash
2877 key'. We then get 256 tables, the first of which looks like this:
2878 <verb>
2879 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2880 1.2.0.0 classid 1:1
2881 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2882 1.2.1.0 classid 1:1
2883 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2884 1.2.2.0 classid 1:3
2885 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2886 1.2.3.0 classid 1:2
2887 </verb>
2889 The next one starts like this:
2890 <verb>
2891 # tc filter add dev eth1 parent 1:0 protocol ip prio 100 match ip src \
2892 1.2.0.1 classid 1:1
2894 </verb>
2896 This way, only four checks are needed at most, two on average.
2898 Configuration is pretty complicated, but very worth it by the time you have
2899 this many rules. First we make a filter root, then we create a table with
2900 256 entries:
2901 <verb>
2902 # tc filter add dev eth1 parent 1:0 prio 5 protocol ip u32
2903 # tc filter add dev eth1 parent 1:0 prio 5 handle 2: protocol ip u32 divisor 256
2904 </verb>
2906 Now we add some rules to entries in the created table:
2908 <verb>
2909 # tc filter add dev eth1 protocol ip parent 1:0 prio 5 u32 ht 2:7b: \
2910 match ip src 1.2.0.123 flowid 1:1
2911 # tc filter add dev eth1 protocol ip parent 1:0 prio 5 u32 ht 2:7b: \
2912 match ip src 1.2.1.123 flowid 1:2
2913 # tc filter add dev eth1 protocol ip parent 1:0 prio 5 u32 ht 2:7b: \
2914 match ip src 1.2.3.123 flowid 1:3
2915 # tc filter add dev eth1 protocol ip parent 1:0 prio 5 u32 ht 2:7b: \
2916 match ip src 1.2.4.123 flowid 1:2
2917 </verb>
2918 This is entry 123, which contains matches for 1.2.0.123, 1.2.1.123,
2919 1.2.2.123, 1.2.3.123, and sends them to 1:1, 1:2, 1:3 and 1:2 respectively.
2920 Note that we need to specify our hash bucket in hex, 0x7b is 123.
2922 Next create a 'hashing filter' that directs traffic to the right entry in
2923 the hashing table:
2924 <verb>
2925 # tc filter add dev eth1 protocol ip parent 1:0 prio 5 u32 ht 800:: \
2926 match ip src 1.2.0.0/16 \
2927 hashkey mask 0x000000ff at 12 \
2928 link 2:
2929 </verb>
2930 Ok, some numbers need explaining. The default hash table is called 800:: and
2931 all filtering starts there. Then we select the source address, which lives
2932 as position 12, 13, 14 and 15 in the IP header, and indicate that we are
2933 only interested in the last part. This we send to hash table 2:, which we
2934 created earlier.
2936 It is quite complicated, but it does work in practice and performance will
2937 be staggering. Note that this example could be improved to the ideal case
2938 where each chain contains 1 filter!
2939 <sect>Kernel network parameters
2940 <p>
2941 The kernel has lots of parameters which
2942 can be tuned for different circumstances. While, as usual, the default
2943 parameters serve 99% of installations very well, we don't call this the
2944 Advanced HOWTO for the fun of it!
2946 The interesting bits are in /proc/sys/net, take a look there. Not everything
2947 will be documented here initially, but we're working on it.
2949 In the meantime you may want to have a look at the Linux-Kernel sources;
2950 read the file <file>Documentation/filesystems/proc.txt</file>. Most of the
2951 features are explained there.
2953 (FIXME)
2955 <sect1>Reverse Path Filtering
2957 By default, routers route everything, even packets which 'obviously' don't
2958 belong on your network. A common example is private IP space escaping onto
2959 the Internet. If you have an interface with a route of 195.96.96.0/24 to it,
2960 you do not expect packets from 212.64.94.1 to arrive there.
2962 Lots of people will want to turn this feature off, so the kernel hackers
2963 have made it easy. There are files in <file>/proc</file> where you can tell
2964 the kernel to do this for you. The method is called "Reverse Path
2965 Filtering". Basically, if the reply to this packet wouldn't go out the
2966 interface this packet came in, then this is a bogus packet and should be
2967 ignored.
2969 The following fragment will turn this on for all current and future
2970 interfaces.
2972 <tscreen><verb>
2973 # for i in /proc/sys/net/ipv4/conf/*/rp_filter ; do
2974 &gt; echo 2 > $i
2975 &gt; done
2976 </verb></tscreen>
2978 Going by the example above, if a packet arrived on the Linux router on eth1
2979 claiming to come from the Office+ISP subnet, it would be dropped. Similarly,
2980 if a packet came from the Office subnet, claiming to be from somewhere
2981 outside your firewall, it would be dropped also.
2983 The above is full reverse path filtering. The default is to only filter
2984 based on IPs that are on directly connected networks. This is because the
2985 full filtering breaks in the case of asymmetric routing (where packets come
2986 in one way and go out another, like satellite traffic, or if you have
2987 dynamic (bgp, ospf, rip) routes in your network. The data comes down
2988 through the satellite dish and replies go back through normal land-lines).
2990 If this exception applies to you (and you'll probably know if it does) you
2991 can simply turn off the <file>rp_filter</file> on the interface where the
2992 satellite data comes in. If you want to see if any packets are being
2993 dropped, the <file>log_martians</file> file in the same directory will tell
2994 the kernel to log them to your syslog.
2996 <tscreen><verb>
2997 # echo 1 >/proc/sys/net/ipv4/conf/<interfacename>/log_martians
2998 </verb></tscreen>
3000 FIXME: is setting the conf/{default,all}/* files enough? - martijn
3002 <sect1>Obscure settings
3004 Ok, there are a lot of parameters which can be modified. We try to list them
3005 all. Also documented (partly) in <file>Documentation/ip-sysctl.txt</file>.
3007 Some of these settings have different defaults based on whether you
3008 answered 'Yes' to 'Configure as router and not host' while compiling your
3009 kernel.
3011 <sect2>Generic ipv4
3013 As a generic note, most rate limiting features don't work on loopback, so
3014 don't test them locally. The limits are supplied in 'jiffies', and are
3015 enforced using the earlier mentioned token bucket filter.
3017 The kernel has an internal clock which runs at 'HZ' ticks (or 'jiffies') per
3018 second. On intel, 'HZ' is mostly 100. So setting a *_rate file to, say 50,
3019 would allow for 2 packets per second. The token bucket filter is also
3020 configured to allow for a burst of at most 6 packets, if enough tokens have
3021 been earned.
3023 Several entries in the following list have been copied from
3024 /usr/src/linux/Documentation/networking/ip-sysctl.txt, written by Alexey
3025 Kuznetsov &lt;kuznet@ms2.inr.ac.ru&gt; and Andi Kleen &lt;ak@muc.de&gt;
3026 <descrip>
3027 <tag>/proc/sys/net/ipv4/icmp_destunreach_rate</tag>
3028 If the kernel decides that it can't deliver a packet, it will drop it, and
3029 send the source of the packet an ICMP notice to this effect.
3030 <tag>/proc/sys/net/ipv4/icmp_echo_ignore_all</tag>
3031 Don't act on echo packets at all. Please don't set this by default, but if
3032 you are used as a relay in a DoS attack, it may be useful.
3033 <tag>/proc/sys/net/ipv4/icmp_echo_ignore_broadcasts [Useful]</tag>
3034 If you ping the broadcast address of a network, all hosts are supposed to
3035 respond. This makes for a dandy denial-of-service tool. Set this to 1 to
3036 ignore these broadcast messages.
3037 <tag>/proc/sys/net/ipv4/icmp_echoreply_rate</tag>
3038 The rate at which echo replies are sent to any one destination.
3039 <tag>/proc/sys/net/ipv4/icmp_ignore_bogus_error_responses</tag>
3040 Set this to ignore ICMP errors caused by hosts in the network reacting badly
3041 to frames sent to what they perceive to be the broadcast address.
3042 <tag>/proc/sys/net/ipv4/icmp_paramprob_rate</tag>
3043 A relatively unknown ICMP message, which is sent in response to incorrect
3044 packets with broken IP or TCP headers. With this file you can control the
3045 rate at which it is sent.
3046 <tag>/proc/sys/net/ipv4/icmp_timeexceed_rate</tag>
3047 This the famous cause of the 'Solaris middle star' in traceroutes. Limits
3048 number of ICMP Time Exceeded messages sent.
3049 <tag>/proc/sys/net/ipv4/igmp_max_memberships</tag>
3050 Maximum number of listening igmp (multicast) sockets on the host.
3051 FIXME: Is this true?
3052 <tag>/proc/sys/net/ipv4/inet_peer_gc_maxtime</tag>
3053 FIXME: Add a little explanation about the inet peer storage?&nl;
3054 Minimum interval between garbage collection passes. This interval is in
3055 effect under low (or absent) memory pressure on the pool. Measured in
3056 jiffies.
3057 <tag>/proc/sys/net/ipv4/inet_peer_gc_mintime</tag>
3058 Minimum interval between garbage collection passes. This interval is in
3059 effect under high memory pressure on the pool. Measured in jiffies.
3060 <tag>/proc/sys/net/ipv4/inet_peer_maxttl</tag>
3061 Maximum time-to-live of entries. Unused entries will expire after this
3062 period of time if there is no memory pressure on the pool (i.e. when the
3063 number of entries in the pool is very small). Measured in jiffies.
3064 <tag>/proc/sys/net/ipv4/inet_peer_minttl</tag>
3065 Minimum time-to-live of entries. Should be enough to cover fragment
3066 time-to-live on the reassembling side. This minimum time-to-live
3067 is guaranteed if the pool size is less than inet_peer_threshold.
3068 Measured in jiffies.
3069 <tag>/proc/sys/net/ipv4/inet_peer_threshold</tag>
3070 The approximate size of the INET peer storage. Starting from this threshold
3071 entries will be thrown aggressively. This threshold also determines
3072 entries' time-to-live and time intervals between garbage collection passes.
3073 More entries, less time-to-live, less GC interval.
3074 <tag>/proc/sys/net/ipv4/ip_autoconfig</tag>
3075 This file contains the number one if the host received its IP configuration by
3076 RARP, BOOTP, DHCP or a similar mechanism. Otherwise it is zero.
3077 <tag>/proc/sys/net/ipv4/ip_default_ttl</tag>
3078 Time To Live of packets. Set to a safe 64. Raise it if you have a huge
3079 network. Don't do so for fun - routing loops cause much more damage that
3080 way. You might even consider lowering it in some circumstances.
3081 <tag>/proc/sys/net/ipv4/ip_dynaddr</tag>
3082 You need to set this if you use dial-on-demand with a dynamic interface
3083 address. Once your demand interface comes up, any local TCP sockets which haven't seen replies will be rebound to have the right address. This solves the problem that the
3084 connection that brings up your interface itself does not work, but the
3085 second try does.
3086 <tag>/proc/sys/net/ipv4/ip_forward</tag>
3087 If the kernel should attempt to forward packets. Off by default.
3088 <tag>/proc/sys/net/ipv4/ip_local_port_range</tag>
3089 Range of local ports for outgoing connections. Actually quite small by
3090 default, 1024 to 4999.
3091 <tag>/proc/sys/net/ipv4/ip_no_pmtu_disc</tag>
3092 Set this if you want to disable Path MTU discovery - a technique to
3093 determine the largest Maximum Transfer Unit possible on your path. See also
3094 the section on Path MTU discovery in the cookbook chapter.
3095 <tag>/proc/sys/net/ipv4/ipfrag_high_thresh</tag>
3096 Maximum memory used to reassemble IP fragments. When
3097 ipfrag_high_thresh bytes of memory is allocated for this purpose,
3098 the fragment handler will toss packets until ipfrag_low_thresh
3099 is reached.
3100 <tag>/proc/sys/net/ipv4/ip_nonlocal_bind</tag>
3101 Set this if you want your applications to be able to bind to an address
3102 which doesn't belong to a device on your system. This can be useful when
3103 your machine is on a non-permanent (or even dynamic) link, so your services
3104 are able to start up and bind to a specific address when your link is down.
3105 <tag>/proc/sys/net/ipv4/ipfrag_low_thresh</tag>
3106 Minimum memory used to reassemble IP fragments.
3107 <tag>/proc/sys/net/ipv4/ipfrag_time</tag>
3108 Time in seconds to keep an IP fragment in memory.
3109 <tag>/proc/sys/net/ipv4/tcp_abort_on_overflow</tag>
3110 A boolean flag controlling the behaviour under lots of incoming connections.
3111 When enabled, this causes the kernel to actively send RST packets when a
3112 service is overloaded.
3113 <tag>/proc/sys/net/ipv4/tcp_fin_timeout</tag>
3114 Time to hold socket in state FIN-WAIT-2, if it was closed by our side. Peer
3115 can be broken and never close its side, or even died unexpectedly. Default
3116 value is 60sec. Usual value used in 2.2 was 180 seconds, you may restore it,
3117 but remember that if your machine is even underloaded WEB server, you risk
3118 to overflow memory with kilotons of dead sockets, FIN-WAIT-2 sockets are
3119 less dangerous than FIN-WAIT-1, because they eat maximum 1.5K of memory, but
3120 they tend to live longer. Cf. tcp_max_orphans.
3121 <tag>/proc/sys/net/ipv4/tcp_keepalive_time</tag>
3122 How often TCP sends out keepalive messages when keepalive is enabled. &nl;
3123 Default: 2hours.
3124 <tag>/proc/sys/net/ipv4/tcp_keepalive_intvl</tag>
3125 How frequent probes are retransmitted, when a probe isn't acknowledged. &nl;
3126 Default: 75 seconds.
3127 <tag>/proc/sys/net/ipv4/tcp_keepalive_probes</tag>
3128 How many keepalive probes TCP will send, until it decides that the
3129 connection is broken. &nl;
3130 Default value: 9. &nl;
3131 Multiplied with tcp_keepalive_intvl, this gives the time a link can be
3132 nonresponsive after a keepalive has been sent.
3133 <tag>/proc/sys/net/ipv4/tcp_max_orphans</tag>
3134 Maximal number of TCP sockets not attached to any user file handle, held by
3135 system. If this number is exceeded orphaned connections are reset
3136 immediately and warning is printed. This limit exists only to prevent simple
3137 DoS attacks, you _must_ not rely on this or lower the limit artificially,
3138 but rather increase it (probably, after increasing installed memory), if
3139 network conditions require more than default value, and tune network
3140 services to linger and kill such states more aggressively. Let me remind you
3141 again: each orphan eats up to ~64K of unswappable memory.
3142 <tag>/proc/sys/net/ipv4/tcp_orphan_retries</tag>
3143 How may times to retry before killing TCP connection, closed by our side.
3144 Default value 7 corresponds to ~50sec-16min depending on RTO. If your machine
3145 is a loaded WEB server, you should think about lowering this value, such
3146 sockets may consume significant resources. Cf. tcp_max_orphans.
3147 <tag>/proc/sys/net/ipv4/tcp_max_syn_backlog</tag>
3148 Maximal number of remembered connection requests, which still did not
3149 receive an acknowledgement from connecting client. Default value is 1024 for
3150 systems with more than 128Mb of memory, and 128 for low memory machines. If
3151 server suffers of overload, try to increase this number. Warning! If you
3152 make it greater than 1024, it would be better to change TCP_SYNQ_HSIZE in
3153 include/net/tcp.h to keep TCP_SYNQ_HSIZE*16<=tcp_max_syn_backlog and to
3154 recompile kernel.
3155 <tag>/proc/sys/net/ipv4/tcp_max_tw_buckets</tag>
3156 Maximal number of timewait sockets held by system simultaneously. If this
3157 number is exceeded time-wait socket is immediately destroyed and warning is
3158 printed. This limit exists only to prevent simple DoS attacks, you _must_
3159 not lower the limit artificially, but rather increase it (probably, after
3160 increasing installed memory), if network conditions require more than
3161 default value.
3162 <tag>/proc/sys/net/ipv4/tcp_retrans_collapse</tag>
3163 Bug-to-bug compatibility with some broken printers.
3164 On retransmit try to send bigger packets to work around bugs in
3165 certain TCP stacks.
3166 <tag>/proc/sys/net/ipv4/tcp_retries1</tag>
3167 How many times to retry before deciding that something is wrong
3168 and it is necessary to report this suspection to network layer.
3169 Minimal RFC value is 3, it is default, which corresponds
3170 to ~3sec-8min depending on RTO.
3171 <tag>/proc/sys/net/ipv4/tcp_retries2</tag>
3172 How may times to retry before killing alive TCP connection.
3173 <url url="http://www.ietf.org/rfc/rfc1122.txt" name="RFC 1122">
3174 says that the limit should be longer than 100 sec.
3175 It is too small number. Default value 15 corresponds to ~13-30min
3176 depending on RTO.
3177 <tag>/proc/sys/net/ipv4/tcp_rfc1337</tag>
3178 This boolean enables a fix for 'time-wait assassination hazards in tcp', described
3179 in RFC 1337. If enabled, this causes the kernel to drop RST packets for
3180 sockets in the time-wait state.&nl;
3181 Default: 0
3182 <tag>/proc/sys/net/ipv4/tcp_sack</tag>
3183 Use Selective ACK which can be used to signify that specific packets are
3184 missing - therefore helping fast recovery.
3185 <tag>/proc/sys/net/ipv4/tcp_stdurg</tag>
3186 Use the Host requirements interpretation of the TCP urg pointer
3187 field. &nl;
3188 Most hosts use the older BSD interpretation, so if you turn this on
3189 Linux might not communicate correctly with them. &nl;
3190 Default: FALSE
3191 <tag>/proc/sys/net/ipv4/tcp_syn_retries</tag>
3192 Number of SYN packets the kernel will send before giving up on the new
3193 connection.
3194 <tag>/proc/sys/net/ipv4/tcp_synack_retries</tag>
3195 To open the other side of the connection, the kernel sends a SYN with a
3196 piggybacked ACK on it, to acknowledge the earlier received SYN. This is part
3197 2 of the threeway handshake. This setting determines the number of SYN+ACK
3198 packets sent before the kernel gives up on the connection.
3199 <tag>/proc/sys/net/ipv4/tcp_timestamps</tag>
3200 Timestamps are used, amongst other things, to protect against wrapping
3201 sequence numbers. A 1 gigabit link might conceivably re-encounter a previous
3202 sequence number with an out-of-line value, because it was of a previous
3203 generation. The timestamp will let it recognise this 'ancient packet'.
3204 <tag>/proc/sys/net/ipv4/tcp_tw_recycle</tag>
3205 Enable fast recycling TIME-WAIT sockets. Default value is 1.
3206 It should not be changed without advice/request of technical experts.
3207 <tag>/proc/sys/net/ipv4/tcp_window_scaling</tag>
3208 TCP/IP normally allows windows up to 65535 bytes big. For really fast
3209 networks, this may not be enough. The window scaling options allows for
3210 almost gigabyte windows, which is good for high bandwidth*delay products.
3212 </descrip>
3213 <sect2>Per device settings
3215 DEV can either stand for a real interface, or for 'all' or 'default'.
3216 Default also changes settings for interfaces yet to be created.
3217 <descrip>
3218 <tag>/proc/sys/net/ipv4/conf/DEV/accept_redirects</tag>
3219 If a router decides that you are using it for a wrong purpose (ie, it needs
3220 to resend your packet on the same interface), it will send us a ICMP
3221 Redirect. This is a slight security risk however, so you may want to turn it
3222 off, or use secure redirects.
3223 <tag>/proc/sys/net/ipv4/conf/DEV/accept_source_route</tag>
3224 Not used very much anymore. You used to be able to give a packet a list of
3225 IP addresses it should visit on its way. Linux can be made to honor this IP
3226 option.
3227 <tag>/proc/sys/net/ipv4/conf/DEV/bootp_relay</tag>
3228 Accept packets with source address 0.b.c.d with destinations not to this host
3229 as local ones. It is supposed that a BOOTP relay daemon will catch and forward
3230 such packets.
3232 The default is 0, since this feature is not implemented yet (kernel version
3233 2.2.12).
3234 <tag>/proc/sys/net/ipv4/conf/DEV/forwarding</tag>
3235 Enable or disable IP forwarding on this interface.
3236 <tag>/proc/sys/net/ipv4/conf/DEV/log_martians</tag>
3237 See the section on reverse path filters.
3238 <tag>/proc/sys/net/ipv4/conf/DEV/mc_forwarding</tag>
3239 If we do multicast forwarding on this interface
3240 <tag>/proc/sys/net/ipv4/conf/DEV/proxy_arp</tag>
3241 If you set this to 1, this interface will respond to ARP requests for
3242 addresses the kernel has routes to. Can be very useful when building 'ip
3243 pseudo bridges'. Do take care that your netmasks are very correct before
3244 enabling this! Also be aware that the rp_filter, mentioned elsewhere, also
3245 operates on ARP queries!
3246 <tag>/proc/sys/net/ipv4/conf/DEV/rp_filter</tag>
3247 See the section on reverse path filters.
3248 <tag>/proc/sys/net/ipv4/conf/DEV/secure_redirects</tag>
3249 Accept ICMP redirect messages only for gateways, listed in default gateway
3250 list. Enabled by default.
3251 <tag>/proc/sys/net/ipv4/conf/DEV/send_redirects</tag>
3252 If we send the above mentioned redirects.
3253 <tag>/proc/sys/net/ipv4/conf/DEV/shared_media</tag>
3254 If it is not set the kernel does not assume that different subnets on this
3255 device can communicate directly. Default setting is 'yes'.
3256 <tag>/proc/sys/net/ipv4/conf/DEV/tag</tag>
3257 FIXME: fill this in
3259 </descrip>
3261 <sect2> Neighbor policy
3263 Dev can either stand for a real interface, or for 'all' or 'default'.
3264 Default also changes settings for interfaces yet to be created.
3265 <descrip>
3266 <tag>/proc/sys/net/ipv4/neigh/DEV/anycast_delay</tag>
3267 Maximum for random delay of answers to neighbor solicitation messages in
3268 jiffies (1/100 sec). Not yet implemented (Linux does not have anycast support
3269 yet).
3270 <tag>/proc/sys/net/ipv4/neigh/DEV/app_solicit</tag>
3271 Determines the number of requests to send to the user level ARP daemon. Use 0
3272 to turn off.
3273 <tag>/proc/sys/net/ipv4/neigh/DEV/base_reachable_time</tag>
3274 A base value used for computing the random reachable time value as specified
3275 in RFC2461.
3276 <tag>/proc/sys/net/ipv4/neigh/DEV/delay_first_probe_time</tag>
3277 Delay for the first time probe if the neighbor is reachable. (see
3278 gc_stale_time)
3279 <tag>/proc/sys/net/ipv4/neigh/DEV/gc_stale_time</tag>
3280 Determines how often to check for stale ARP entries. After an ARP entry is
3281 stale it will be resolved again (which is useful when an IP address migrates
3282 to another machine). When ucast_solicit is greater than 0 it first tries to
3283 send an ARP packet directly to the known host When that fails and
3284 mcast_solicit is greater than 0, an ARP request is broadcasted.
3285 <tag>/proc/sys/net/ipv4/neigh/DEV/locktime</tag>
3286 An ARP/neighbor entry is only replaced with a new one if the old is at least
3287 locktime old. This prevents ARP cache thrashing.
3288 <tag>/proc/sys/net/ipv4/neigh/DEV/mcast_solicit</tag>
3289 Maximum number of retries for multicast solicitation.
3290 <tag>/proc/sys/net/ipv4/neigh/DEV/proxy_delay</tag>
3291 Maximum time (real time is random [0..proxytime]) before answering to an ARP
3292 request for which we have an proxy ARP entry. In some cases, this is used to
3293 prevent network flooding.
3294 <tag>/proc/sys/net/ipv4/neigh/DEV/proxy_qlen</tag>
3295 Maximum queue length of the delayed proxy arp timer. (see proxy_delay).
3296 <tag>/proc/sys/net/ipv4/neigh/DEV/retrans_time</tag>
3297 The time, expressed in jiffies (1/100 sec), between retransmitted Neighbor
3298 Solicitation messages. Used for address resolution and to determine if a
3299 neighbor is unreachable.
3300 <tag>/proc/sys/net/ipv4/neigh/DEV/ucast_solicit</tag>
3301 Maximum number of retries for unicast solicitation.
3302 <tag>/proc/sys/net/ipv4/neigh/DEV/unres_qlen</tag>
3303 Maximum queue length for a pending arp request - the number of packets which
3304 are accepted from other layers while the ARP address is still resolved.
3306 <tag>Internet QoS: Architectures and Mechanisms for Quality of Service,
3307 Zheng Wang, ISBN 1-55860-608-4</tag> Hardcover textbook covering topics
3308 related to Quality of Service. Good for understanding basic concepts.
3310 </descrip>
3312 <sect2>Routing settings
3314 <descrip>
3315 <tag>/proc/sys/net/ipv4/route/error_burst</tag>
3316 These parameters are used to limit the warning messages written to the kernel
3317 log from the routing code. The higher the error_cost factor is, the fewer
3318 messages will be written. Error_burst controls when messages will be dropped.
3319 The default settings limit warning messages to one every five seconds.
3320 <tag>/proc/sys/net/ipv4/route/error_cost</tag>
3321 These parameters are used to limit the warning messages written to the kernel
3322 log from the routing code. The higher the error_cost factor is, the fewer
3323 messages will be written. Error_burst controls when messages will be dropped.
3324 The default settings limit warning messages to one every five seconds.
3325 <tag>/proc/sys/net/ipv4/route/flush</tag>
3326 Writing to this file results in a flush of the routing cache.
3327 <tag>/proc/sys/net/ipv4/route/gc_elasticity</tag>
3328 Values to control the frequency and behavior of the garbage collection
3329 algorithm for the routing cache. This can be important for when doing
3330 failover. At least gc_timeout seconds will elapse before Linux will skip
3331 to another route because the previous one has died. By default set to 300,
3332 you may want to lower it if you want to have a speedy failover.
3334 Also see <url
3335 url="http://mailman.ds9a.nl/pipermail/lartc/2002q1/002667.html" name="this
3336 post"> by Ard van Breemen.
3338 <tag>/proc/sys/net/ipv4/route/gc_interval</tag>
3339 See /proc/sys/net/ipv4/route/gc_elasticity.
3340 <tag>/proc/sys/net/ipv4/route/gc_min_interval</tag>
3341 See /proc/sys/net/ipv4/route/gc_elasticity.
3342 <tag>/proc/sys/net/ipv4/route/gc_thresh</tag>
3343 See /proc/sys/net/ipv4/route/gc_elasticity.
3344 <tag>/proc/sys/net/ipv4/route/gc_timeout</tag>
3345 See /proc/sys/net/ipv4/route/gc_elasticity.
3346 <tag>/proc/sys/net/ipv4/route/max_delay</tag>
3347 Delays for flushing the routing cache.
3348 <tag>/proc/sys/net/ipv4/route/max_size</tag>
3349 Maximum size of the routing cache. Old entries will be purged once the cache
3350 reached has this size.
3351 <tag>/proc/sys/net/ipv4/route/min_adv_mss</tag>
3352 FIXME: fill this in
3353 <tag>/proc/sys/net/ipv4/route/min_delay</tag>
3354 Delays for flushing the routing cache.
3355 <tag>/proc/sys/net/ipv4/route/min_pmtu</tag>
3356 FIXME: fill this in
3357 <tag>/proc/sys/net/ipv4/route/mtu_expires</tag>
3358 FIXME: fill this in
3359 <tag>/proc/sys/net/ipv4/route/redirect_load</tag>
3360 Factors which determine if more ICPM redirects should be sent to a specific
3361 host. No redirects will be sent once the load limit or the maximum number of
3362 redirects has been reached.
3363 <tag>/proc/sys/net/ipv4/route/redirect_number</tag>
3364 See /proc/sys/net/ipv4/route/redirect_load.
3365 <tag>/proc/sys/net/ipv4/route/redirect_silence</tag>
3366 Timeout for redirects. After this period redirects will be sent again, even if
3367 this has been stopped, because the load or number limit has been reached.
3368 </descrip>
3371 <sect>Advanced &amp; less common queueing disciplines
3373 Should you find that you have needs not addressed by the queues mentioned
3374 earlier, the kernel contains some other more specialized queues mentioned here.
3375 <sect1>bfifo/pfifo
3377 These classless queues are even simpler than pfifo_fast in that they lack
3378 the internal bands - all traffic is really equal. They have one important
3379 benefit though, they have some statistics. So even if you don't need shaping
3380 or prioritizing, you can use this qdisc to determine the backlog on your
3381 interface.
3383 pfifo has a length measured in packets, bfifo in bytes.
3384 <sect2>Parameters &amp; usage
3386 <descrip>
3387 <tag>limit</tag>
3388 Specifies the length of the queue. Measured in bytes for bfifo, in packets
3389 for pfifo. Defaults to the interface txqueuelen (see pfifo_fast chapter)
3390 packets long or txqueuelen*mtu bytes for bfifo.
3391 </descrip>
3392 <sect1> Clark-Shenker-Zhang algorithm (CSZ)
3394 This is so theoretical that not even Alexey (the main CBQ author) claims to
3395 understand it. From his source:
3397 "David D. Clark, Scott Shenker and Lixia Zhang
3398 Supporting Real-Time Applications in an Integrated Services Packet
3399 Network: Architecture and Mechanism.
3401 As I understand it, the main idea is to create WFQ flows for each guaranteed
3402 service and to allocate the rest of bandwith to dummy flow-0. Flow-0
3403 comprises the predictive services and the best effort traffic; it is handled
3404 by a priority scheduler with the highest priority band allocated for
3405 predictive services, and the rest --- to the best effort packets.
3407 Note that in CSZ flows are NOT limited to their bandwidth. It is supposed
3408 that the flow passed admission control at the edge of the QoS network and it
3409 doesn't need further shaping. Any attempt to improve the flow or to shape it
3410 to a token bucket at intermediate hops will introduce undesired delays and
3411 raise jitter.
3413 At the moment CSZ is the only scheduler that provides true guaranteed
3414 service. Another schemes (including CBQ) do not provide guaranteed delay and
3415 randomize jitter."
3417 Does not currently seem like a good canidate to use, unless you've read and
3418 understand the article mentioned.
3419 <sect1>DSMARK
3421 Esteve Camps Chust &lt;marvin@grn.es&gt;&nl;
3422 This text is an extract from my thesis on "QoS Support in Linux", September 2000.&nl;
3424 Source documents:&nl;
3425 <itemize>
3426 <item><url url="http://ica1www.epfl.ch/~almesber" name="Draft-almesberger-wajhak-diffserv-linux-01.txt">.
3427 <item>Examples in iproute2 distribution.
3428 <item><url url="http://www.qosforum.com/white-papers/qosprot_v3.pdf" name="White Paper-QoS protocols and architectures"> and
3429 <url url="http://www.qosforum.com/docs/faq" name="IP QoS Frequently Asked Questions"> both by <em>Quality of Service Forum</em>.
3430 </itemize>
3432 This chapter was written by Esteve Camps &lt;esteve@hades.udg.es&gt;.
3433 <sect2>Introduction
3436 First of all, first of all, it would be a great idea for you to read RFCs
3437 written about this (RFC2474, RFC2475, RFC2597 and RFC2598) at <url
3438 url="http://www.ietf.org/html.charters/diffserv-charter.html" name="IETF
3439 DiffServ working Group web site"> and <url
3440 url="http://ica1www.epfl.ch/~almesber" name="Werner Almesberger web site">
3441 (he wrote the code to support Differentiated Services on Linux).
3443 <sect2>What is Dsmark related to?
3445 Dsmark is a queueing discipline that offers the capabilities needed in
3446 Differentiated Services (also called DiffServ or, simply, DS). DiffServ is
3447 one of two actual QoS architectures (the other one is called Integrated
3448 Services) that is based on a value carried by packets in the DS field of the
3449 IP header.
3452 One of the first solutions in IP designed to offer some QoS level was
3453 the Type of Service field (TOS byte) in IP header. By changing that value,
3454 we could choose a high/low level of throughput, delay or reliability.
3455 But this didn't provide sufficient flexibility to the needs of new
3456 services (such as real-time applications, interactive applications and
3457 others). After this, new architectures appeared. One of these was DiffServ
3458 which kept TOS bits and renamed DS field.
3459 <sect2>Differentiated Services guidelines
3461 Differentiated Services is group-oriented. I mean, we don't know anything
3462 about flows (this will be the Integrated Services purpose); we know about
3463 flow aggregations and we will apply different behaviours depending on which
3464 aggregation a packet belongs to.
3467 When a packet arrives to an edge node (entry node to a DiffServ domain)
3468 entering to a DiffServ Domain we'll have to policy, shape and/or mark those
3469 packets (marking refers to assigning a value to the DS field. It's just like the
3470 cows :-) ). This will be the mark/value that the internal/core nodes on our
3471 DiffServ Domain will look at to determine which behaviour or QoS level
3472 apply.
3475 As you can deduce, Differentiated Services involves a domain on which
3476 all DS rules will have to be applied. In fact you can think &dquot;I
3477 will classify all the packets entering my domain. Once they enter my
3478 domain they will be subjected to the rules that my classification dictates
3479 and every traversed node will apply that QoS level&dquot;.
3481 In fact, you can apply your own policies into your local domains, but some
3482 <em>Service Level Agreements</em> should be considered when connecting to
3483 other DS domains.
3486 At this point, you maybe have a lot of questions. DiffServ is more than I've
3487 explained. In fact, you can understand that I can not resume more than 3
3488 RFC's in just 50 lines :-).
3490 <sect2>Working with Dsmark
3493 As the DiffServ bibliography specifies, we differentiate boundary nodes and
3494 interior nodes. These are two important points in the traffic path. Both
3495 types perform a classification when the packets arrive. Its result may be
3496 used in different places along the DS process before the packet is released
3497 to the network. It's just because of this that the diffserv code supplies an
3498 structure called sk_buff, including a new field called skb-&gt;tc_index
3499 where we'll store the result of initial classification that may be used in
3500 several points in DS treatment.
3503 The skb-&gt;tc_index value will be initially set by the DSMARK qdisc,
3504 retrieving it from the DS field in IP header of every received packet.
3505 Besides, cls_tcindex classifier will read all or part of skb-&gt;tcindex
3506 value and use it to select classes.
3509 But, first of all, take a look at DSMARK qdisc command and its parameters:
3510 <tscreen><verb>
3511 ... dsmark indices INDICES [ default_index DEFAULT_INDEX ] [ set_tc_index ]
3512 </verb></tscreen>
3513 What do these parameters mean?
3514 <itemize>
3515 <item><bf>indices</bf>: size of table of (mask,value) pairs. Maximum value is 2^n, where n&gt=0.
3516 <item><bf>Default_index</bf>: the default table entry index if classifier finds no match.
3517 <item><bf>Set_tc_index</bf>: instructs dsmark discipline to retrieve the DS field and store it onto skb-&gt;tc_index.
3518 </itemize>
3519 Let's see the DSMARK process.
3521 <sect2>How SCH_DSMARK works.
3523 This qdisc will apply the next steps:
3524 <itemize>
3525 <item>If we have declared set_tc_index option in qdisc command, DS field is retrieved and stored onto
3526 skb-&gt;tc_index variable.
3527 <item>Classifier is invoked. The classifier will be executed and it will return a class ID that will be stored in
3528 skb-&gt;tc_index variable.If no filter matches are found, we consider the default_index option to be the
3529 classId to store. If neither set_tc_index nor default_index has been declared results may be
3530 impredictibles.
3531 <item>After been sent to internal qdisc's where you can reuse the result of the filter, the classid returned by
3532 the internal qdisc is stored into skb-&gt;tc_index. We will use this value in the future to index a mask-
3533 value table. The final result to assign to the packet will be that resulting from next operation:
3534 <tscreen><verb>
3535 New_Ds_field = ( Old_DS_field & mask ) | value
3536 </verb></tscreen>
3538 <item>Thus, new value will result from "anding" ds_field and mask values and next, this result "ORed" with
3539 value parameter. See next diagram to understand all this process:
3540 </itemize>
3541 <tscreen>
3542 <verb>
3543 skb-&gt;ihp-&gt;tos
3544 - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - &gt;
3545 | | ^
3546 | -- If you declare set_tc_index, we set DS | | &lt;-----May change
3547 | value into skb-&gt;tc_index variable | |O DS field
3548 | A| |R
3549 +-|-+ +------+ +---+-+ Internal +-+ +---N|-----|----+
3550 | | | | tc |---&gt;| | |--&gt; . . . --&gt;| | | D| | |
3551 | | |-----&gt;|index |---&gt;| | | Qdisc | |----&gt;| v | |
3552 | | | |filter|---&gt;| | | +---------------+ | ----&gt;(mask,value) |
3553 --&gt;| O | +------+ +-|-+--------------^----+ / | (. , .) |
3554 | | | ^ | | | | (. , .) |
3555 | | +----------|---------|----------------|-------|--+ (. , .) |
3556 | | sch_dsmark | | | | |
3557 +-|------------|---------|----------------|-------|------------------+
3558 | | | &lt;- tc_index -&gt; | |
3559 | |(read) | may change | | &lt;--------------Index to the
3560 | | | | | (mask,value)
3561 v | v v | pairs table
3562 - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -&gt;
3563 skb-&gt;tc_index
3564 </verb>
3565 </tscreen>
3567 How to do marking? Just change the mask and value of the class you want to remark. See next line of code:
3568 <tscreen>
3569 tc class change dev eth0 classid 1:1 dsmark mask 0x3 value 0xb8
3570 </tscreen>
3571 This changes the (mask,value) pair in hash table, to remark packets belonging to class 1:1.You have to "change" this values
3572 because of default values that (mask,value) gets initially (see table below).
3574 Now, we'll explain how TC_INDEX filter works and how fits into this. Besides, TCINDEX filter can be
3575 used in other configurations rather than those including DS services.
3578 <sect2>TC_INDEX Filter
3580 This is the basic command to declare a TC_INDEX filter:
3581 <tscreen>
3582 <verb>
3583 ... tcindex [ hash SIZE ] [ mask MASK ] [ shift SHIFT ]
3584 [ pass_on | fall_through ]
3585 [ classid CLASSID ] [ police POLICE_SPEC ]
3586 </verb>
3587 </tscreen>
3588 Next, we show the example used to explain TC_INDEX operation mode. Pay attention to bolded words:
3589 &nl;&nl;
3590 tc qdisc add dev eth0 handle 1:0 root dsmark indices 64 <bf>set_tc_index</bf>&nl;
3591 tc filter add dev eth0 parent 1:0 protocol ip prio 1 tcindex <bf>mask 0xfc shift 2</bf>&nl;
3592 tc qdisc add dev eth0 parent 1:0 handle 2:0 cbq bandwidth 10Mbit cell 8 avpkt 1000 mpu 64&nl;
3593 # EF traffic class&nl;
3594 tc class add dev eth0 parent 2:0 classid 2:1 cbq bandwidth 10Mbit rate 1500Kbit avpkt 1000 prio 1 bounded isolated allot 1514 weight 1 maxburst 10&nl;
3595 # Packet fifo qdisc for EF traffic&nl;
3596 tc qdisc add dev eth0 parent 2:1 pfifo limit 5&nl;
3597 tc filter add dev eth0 parent 2:0 protocol ip prio 1 <bf>handle 0x2e</bf> tcindex <bf>classid 2:1 pass_on</bf>&nl;
3598 &nl;&nl;
3599 (This code is not complete. It's just an extract from EFCBQ example included in iproute2 distribution).
3601 First of all, suppose we receive a packet marked as EF . If you read RFC2598, you'll see that DSCP
3602 recommended value for EF traffic is 101110. This means that DS field will be 10111000 (remember that
3603 less signifiant bits in TOS byte are not used in DS) or 0xb8 in hexadecimal codification.
3605 <tscreen>
3606 <verb>
3607 TC INDEX
3608 FILTER
3609 +---+ +-------+ +---+-+ +------+ +-+ +-------+
3610 | | | | | | | |FILTER| +-+ +-+ | | | |
3611 | |-----&gt;| MASK | -&gt; | | | -&gt; |HANDLE|-&gt;| | | | -&gt; | | -&gt; | |
3612 | | . | =0xfc | | | | |0x2E | | +----+ | | | | |
3613 | | . | | | | | +------+ +--------+ | | | |
3614 | | . | | | | | | | | |
3615 --&gt;| | . | SHIFT | | | | | | | |--&gt;
3616 | | . | =2 | | | +----------------------------+ | | |
3617 | | | | | | CBQ 2:0 | | |
3618 | | +-------+ +---+--------------------------------+ | |
3619 | | | |
3620 | +-------------------------------------------------------------+ |
3621 | DSMARK 1:0 |
3622 +-------------------------------------------------------------------------+
3624 </verb>
3625 </tscreen>
3627 The packet arrives, then, set with 0xb8 value at DS field. As we explained before, dsmark qdisc identified
3628 by 1:0 id in the example, retrieves DS field and store it in skb-&gt;tc_index variable.
3629 Next step in the example will correspond to the filter associated to this qdisc (second line in the example).
3630 This will perform next operations:
3631 <tscreen>
3632 <verb>
3633 Value1 = skb->tc_index & MASK
3634 Key = Value1 >> SHIFT
3635 </verb>
3636 </tscreen>
3639 In the example, MASK=0xFC i SHIFT=2.
3640 <tscreen>
3641 <verb>
3642 Value1 = 10111000 & 11111100 = 10111000
3643 Key = 10111000 >> 2 = 00101110 -> 0x2E in hexadecimal
3644 </verb>
3645 </tscreen>
3648 The returned value will correspond to a qdisc interal filter handle (in the example, identifier 2:0). If a
3649 filter with this id exists, policing and metering conditions will be verified (in case that filter includes this)
3650 and the classid will be returned (in our example, classid 2:1) and stored in skb-&gt;tc_index variable.
3653 But if any filter with that identifier is found, the result will depend on fall_through flag declaration. If so,
3654 value key is returned as classid. If not, an error is returned and process continues with the rest filters. Be
3655 careful if you use fall_through flag; this can be done if a simple relation exists between values
3656 &nl;of skb-&gt;tc_index variable and class id's.
3659 The latest parameters to comment on are hash and pass_on. The first one
3660 relates to hash table size. Pass_on will be used to indicate that if no classid
3661 equal to the result of this filter is found, try next filter.
3662 The default action is fall_through (look at next table).
3665 Finally, let's see which possible values can be set to all this TCINDEX parameters:
3666 <tscreen>
3667 <verb>
3668 TC Name Value Default
3669 -----------------------------------------------------------------
3670 Hash 1...0x10000 Implementation dependent
3671 Mask 0...0xffff 0xffff
3672 Shift 0...15 0
3673 Fall through / Pass_on Flag Fall_through
3674 Classid Major:minor None
3675 Police ..... None
3676 </verb>
3677 </tscreen>
3679 This kind of filter is very powerful. It's necessary to explore all possibilities. Besides, this filter is not only used in DiffServ configurations.
3680 You can use it as any other kind of filter.
3682 I recommend you to look at all DiffServ examples included in iproute2 distribution. I promise I will try to
3683 complement this text as soon as I can. Besides, all I have explained is the result of a lot of tests.
3684 I would thank you tell me if I'm wrong in any point.
3685 <sect1>Ingress qdisc
3686 <p>
3687 All qdiscs discussed so far are egress qdiscs. Each interface however can
3688 also have an ingress qdisc which is not used to send packets
3689 out to the network adaptor. Instead, it allows you to apply tc filters to
3690 packets coming in over the interface, regardless of whether they have a local
3691 destination or are to be forwarded.
3693 As the tc filters contain a full Token Bucket Filter implementation, and are
3694 also able to match on the kernel flow estimator, there is a lot of
3695 functionality available. This effectively allows you to police incoming
3696 traffic, before it even enters the IP stack.
3698 <sect2>Parameters &amp; usage
3700 The ingress qdisc itself does not require any parameters. It differs from
3701 other qdiscs in that it does not occupy the root of a device. Attach it like
3702 this:
3703 <verb>
3704 # tc qdisc add dev eth0 ingress
3705 </verb>
3706 This allows you to have other, sending, qdiscs on your device besides the
3707 ingress qdisc.
3709 For a contrived example how the ingress qdisc could be used, see the
3710 Cookbook.
3712 <sect1>Random Early Detection (RED)
3714 This section is meant as an introduction to backbone routing, which often
3715 involves &lt;100 megabit bandwidths, which requires a different approach than
3716 your ADSL modem at home.
3718 The normal behaviour of router queues on the Internet is called tail-drop.
3719 Tail-drop works by queueing up to a certain amount, then dropping all traffic
3720 that 'spills over'. This is very unfair, and also leads to retransmit
3721 synchronisation. When retransmit synchronisation occurs, the sudden burst
3722 of drops from a router that has reached its fill will cause a delayed burst
3723 of retransmits, which will over fill the congested router again.
3725 In order to cope with transient congestion on links, backbone routers will
3726 often implement large queues. Unfortunately, while these queues are good for
3727 throughput, they can substantially increase latency and cause TCP
3728 connections to behave very bursty during congestion.
3730 These issues with tail-drop are becoming increasingly troublesome on the
3731 Internet because the use of network unfriendly applications is increasing.
3732 The Linux kernel offers us RED, short for Random Early Detect, also called
3733 Random Early Drop, as that is how it works.
3735 RED isn't a cure-all for this, applications which inappropriately fail to
3736 implement exponential backoff still get an unfair share of the bandwidth,
3737 however, with RED they do not cause as much harm to the throughput and
3738 latency of other connections.
3740 RED statistically drops packets from flows before it reaches its hard
3741 limit. This causes a congested backbone link to slow more gracefully, and
3742 prevents retransmit synchronisation. This also helps TCP find its 'fair'
3743 speed faster by allowing some packets to get dropped sooner keeping queue
3744 sizes low and latency under control. The probability of a packet being
3745 dropped from a particular connection is proportional to its bandwidth usage
3746 rather than the number of packets it transmits.
3748 RED is a good queue for backbones, where you can't afford the
3749 complexity of per-session state tracking needed by fairness queueing.
3751 In order to use RED, you must decide on three parameters: Min, Max, and
3752 burst. Min sets the minimum queue size in bytes before dropping will begin,
3753 Max is a soft maximum that the algorithm will attempt to stay under, and
3754 burst sets the maximum number of packets that can 'burst through'.
3756 You should set the min by calculating that highest acceptable base queueing
3757 latency you wish, and multiply it by your bandwidth. For instance, on my
3758 64kbit/s ISDN link, I might want a base queueing latency of 200ms so I set
3759 min to 1600 bytes. Setting min too small will degrade throughput and too
3760 large will degrade latency. Setting a small min is not a replacement for
3761 reducing the MTU on a slow link to improve interactive response.
3763 You should make max at least twice min to prevent synchronisation. On slow
3764 links with small min's it might be wise to make max perhaps four or
3765 more times large then min.
3767 Burst controls how the RED algorithm responds to bursts. Burst must be set
3768 larger then min/avpkt. Experimentally, I've found (min+min+max)/(3*avpkt) to
3769 work okay.
3771 Additionally, you need to set limit and avpkt. Limit is a safety value, after
3772 there are limit bytes in the queue, RED 'turns into' tail-drop. I typical set
3773 limit to eight times max. Avpkt should be your average packet size. 1000
3774 works okay on high speed Internet links with a 1500byte MTU.
3776 Read <url url="http://www.aciri.org/floyd/papers/red/red.html"
3777 name="the paper on RED queueing"> by Sally Floyd and Van Jacobson for technical
3778 information.
3779 <sect1>Generic Random Early Detection
3781 Not a lot is known about GRED. It looks like GRED with several internal
3782 queues, whereby the internal queue is chosen based on the Diffserv tcindex
3783 field. According to a slide found <url
3784 url="http://www.davin.ottawa.on.ca/ols/img22.htm" name="here">, it contains
3785 the capabilities of Cisco's 'Distributed Weighted RED', as well as Dave
3786 Clark's RIO.
3788 Each virtual queue can have its own Drop Parameters specified.
3790 FIXME: get Jamal or Werner to tell us more
3792 <sect1>VC/ATM emulation
3794 This is quite a major effort by Werner Almesberger to allow you to build
3795 Virtual Circuits over TCP/IP sockets. A Virtual Circuit is a concept from
3796 ATM network theory.
3798 For more information, see the <url url="http://linux-atm.sourceforge.net/"
3799 name="ATM on Linux homepage">.
3801 <sect1>Weighted Round Robin (WRR)
3803 This qdisc is not included in the standard kernels but can be downloaded from
3804 <url url="http://wipl-wrr.dkik.dk/wrr/">.
3805 Currently the qdisc is only tested with Linux 2.2 kernels but it will
3806 probably work with 2.4/2.5 kernels too.
3808 The WRR qdisc distributes bandwidth between its classes using the weighted
3809 round robin scheme. That is, like the CBQ qdisc it contains classes
3810 into which arbitrary qdiscs can be plugged. All classes which have sufficient
3811 demand will get bandwidth proportional to the weights associated with the classes.
3812 The weights can be set manually using the <tt>tc</tt> program. But they
3813 can also be made automatically decreasing for classes transferring much data.
3815 The qdisc has a built-in classifier which assigns packets coming from or
3816 sent to different machines to different classes. Either the MAC or IP and
3817 either source or destination addresses can be used. The MAC address can only
3818 be used when the Linux box is acting as an ethernet bridge, however. The
3819 classes are automatically assigned to machines based on the packets seen.
3821 The qdisc can be very useful at sites such as dorms where a lot of unrelated
3822 individuals share an Internet connection. A set of scripts setting up a
3823 relevant behavior for such a site is a central part of the WRR distribution.
3825 <sect>Cookbook
3827 This section contains 'cookbook' entries which may help you solve problems.
3828 A cookbook is no replacement for understanding however, so try and comprehend
3829 what is going on.
3831 <sect1>Running multiple sites with different SLAs
3833 You can do this in several ways. Apache has some support for this with a
3834 module, but we'll show how Linux can do this for you, and do so for other
3835 services as well. These commands are stolen from a presentation by Jamal
3836 Hadi that's referenced below.
3838 Let's say we have two customers, with http, ftp and streaming audio, and we
3839 want to sell them a limited amount of bandwidth. We do so on the server itself.
3841 Customer A should have at most 2 megabits, customer B has paid for 5
3842 megabits. We separate our customers by creating virtual IP addresses on our
3843 server.
3845 <tscreen><verb>
3846 # ip address add 188.177.166.1 dev eth0
3847 # ip address add 188.177.166.2 dev eth0
3848 </verb></tscreen>
3850 It is up to you to attach the different servers to the right IP address. All
3851 popular daemons have support for this.
3853 We first attach a CBQ qdisc to eth0:
3854 <tscreen><verb>
3855 # tc qdisc add dev eth0 root handle 1: cbq bandwidth 10Mbit cell 8 avpkt 1000 \
3856 mpu 64
3857 </verb></tscreen>
3859 We then create classes for our customers:
3861 <tscreen><verb>
3862 # tc class add dev eth0 parent 1:0 classid 1:1 cbq bandwidth 10Mbit rate \
3863 2MBit avpkt 1000 prio 5 bounded isolated allot 1514 weight 1 maxburst 21
3864 # tc class add dev eth0 parent 1:0 classid 1:2 cbq bandwidth 10Mbit rate \
3865 5Mbit avpkt 1000 prio 5 bounded isolated allot 1514 weight 1 maxburst 21
3866 </verb></tscreen>
3868 Then we add filters for our two classes:
3869 <tscreen><verb>
3870 ##FIXME: Why this line, what does it do?, what is a divisor?:
3871 ##FIXME: A divisor has something to do with a hash table, and the number of
3872 ## buckets - ahu
3873 # tc filter add dev eth0 parent 1:0 protocol ip prio 5 handle 1: u32 divisor 1
3874 # tc filter add dev eth0 parent 1:0 prio 5 u32 match ip src 188.177.166.1
3875 flowid 1:1
3876 # tc filter add dev eth0 parent 1:0 prio 5 u32 match ip src 188.177.166.2
3877 flowid 1:2
3878 </verb></tscreen>
3880 And we're done.
3882 FIXME: why no token bucket filter? is there a default pfifo_fast fallback
3883 somewhere?
3885 <sect1>Protecting your host from SYN floods
3887 From Alexey's iproute documentation, adapted to netfilter and with more
3888 plausible paths. If you use this, take care to adjust the numbers to
3889 reasonable values for your system.
3891 If you want to protect an entire network, skip this script, which is best
3892 suited for a single host.
3894 It appears that you need the very latest version of the iproute2 tools to
3895 get this to work with 2.4.0.
3897 <tscreen><verb>
3898 #! /bin/sh -x
3900 # sample script on using the ingress capabilities
3901 # this script shows how one can rate limit incoming SYNs
3902 # Useful for TCP-SYN attack protection. You can use
3903 # IPchains to have more powerful additions to the SYN (eg
3904 # in addition the subnet)
3906 #path to various utilities;
3907 #change to reflect yours.
3909 TC=/sbin/tc
3910 IP=/sbin/ip
3911 IPTABLES=/sbin/iptables
3912 INDEV=eth2
3914 # tag all incoming SYN packets through $INDEV as mark value 1
3915 ############################################################
3916 $iptables -A PREROUTING -i $INDEV -t mangle -p tcp --syn \
3917 -j MARK --set-mark 1
3918 ############################################################
3920 # install the ingress qdisc on the ingress interface
3921 ############################################################
3922 $TC qdisc add dev $INDEV handle ffff: ingress
3923 ############################################################
3927 # SYN packets are 40 bytes (320 bits) so three SYNs equals
3928 # 960 bits (approximately 1kbit); so we rate limit below
3929 # the incoming SYNs to 3/sec (not very useful really; but
3930 #serves to show the point - JHS
3931 ############################################################
3932 $TC filter add dev $INDEV parent ffff: protocol ip prio 50 handle 1 fw \
3933 police rate 1kbit burst 40 mtu 9k drop flowid :1
3934 ############################################################
3938 echo "---- qdisc parameters Ingress ----------"
3939 $TC qdisc ls dev $INDEV
3940 echo "---- Class parameters Ingress ----------"
3941 $TC class ls dev $INDEV
3942 echo "---- filter parameters Ingress ----------"
3943 $TC filter ls dev $INDEV parent ffff:
3945 #deleting the ingress qdisc
3946 #$TC qdisc del $INDEV ingress
3947 </verb></tscreen>
3948 <sect1>Ratelimit ICMP to prevent dDoS
3950 Recently, distributed denial of service attacks have become a major nuisance
3951 on the Internet. By properly filtering and ratelimiting your network, you can
3952 both prevent becoming a casualty or the cause of these attacks.
3954 You should filter your networks so that you do not allow non-local IP source
3955 addressed packets to leave your network. This stops people from anonymously
3956 sending junk to the Internet.
3958 <!-- FIXME: netfilter one liner. Is there a netfilter one-liner? Martijn -->
3961 Rate limiting goes much as shown earlier. To refresh your memory, our
3962 ASCIIgram again:
3964 <tscreen><verb>
3965 [The Internet] ---<E3, T3, whatever>--- [Linux router] --- [Office+ISP]
3966 eth1 eth0
3967 </verb></tscreen>
3969 We first set up the prerequisite parts:
3971 <tscreen><verb>
3972 # tc qdisc add dev eth0 root handle 10: cbq bandwidth 10Mbit avpkt 1000
3973 # tc class add dev eth0 parent 10:0 classid 10:1 cbq bandwidth 10Mbit rate \
3974 10Mbit allot 1514 prio 5 maxburst 20 avpkt 1000
3975 </verb></tscreen>
3977 If you have 100Mbit, or more, interfaces, adjust these numbers. Now you need
3978 to determine how much ICMP traffic you want to allow. You can perform
3979 measurements with tcpdump, by having it write to a file for a while, and
3980 seeing how much ICMP passes your network. Do not forget to raise the
3981 snapshot length!
3983 If measurement is impractical, you might want to choose 5% of your available
3984 bandwidth. Let's set up our class:
3985 <tscreen><verb>
3986 # tc class add dev eth0 parent 10:1 classid 10:100 cbq bandwidth 10Mbit rate \
3987 100Kbit allot 1514 weight 800Kbit prio 5 maxburst 20 avpkt 250 \
3988 bounded
3989 </verb></tscreen>
3991 This limits at 100Kbit. Now we need a filter to assign ICMP traffic to this
3992 class:
3993 <tscreen><verb>
3994 # tc filter add dev eth0 parent 10:0 protocol ip prio 100 u32 match ip
3995 protocol 1 0xFF flowid 10:100
3997 </verb></tscreen>
3999 <sect1>Prioritizing interactive traffic
4001 If lots of data is coming down your link, or going up for that matter, and
4002 you are trying to do some maintenance via telnet or ssh, this may not go too
4003 well. Other packets are blocking your keystrokes. Wouldn't it be great if
4004 there were a way for your interactive packets to sneak past the bulk
4005 traffic? Linux can do this for you!
4007 As before, we need to handle traffic going both ways. Evidently, this works
4008 best if there are Linux boxes on both ends of your link, although other
4009 UNIX's are able to do this. Consult your local Solaris/BSD guru for this.
4011 The standard pfifo_fast scheduler has 3 different 'bands'. Traffic in band 0
4012 is transmitted first, after which traffic in band 1 and 2 gets considered.
4013 It is vital that our interactive traffic be in band 0!
4015 We blatantly adapt from the (soon to be obsolete) ipchains HOWTO:
4017 There are four seldom-used bits in the IP header, called the Type of Service
4018 (TOS) bits. They effect the way packets are treated; the four bits are
4019 "Minimum Delay", "Maximum Throughput", "Maximum Reliability" and "Minimum
4020 Cost". Only one of these bits is allowed to be set. Rob van Nieuwkerk, the
4021 author of the ipchains TOS-mangling code, puts it as follows:
4023 <tscreen>
4024 Especially the "Minimum Delay" is important for me. I switch it on for
4025 "interactive" packets in my upstream (Linux) router. I'm
4026 behind a 33k6 modem link. Linux prioritizes packets in 3 queues. This
4027 way I get acceptable interactive performance while doing bulk
4028 downloads at the same time.
4029 </tscreen>
4031 The most common use is to set telnet & ftp control connections to "Minimum
4032 Delay" and FTP data to "Maximum Throughput". This would be
4033 done as follows, on your upstream router:
4035 <tscreen><verb>
4036 # iptables -A PREROUTING -t mangle -p tcp --sport telnet \
4037 -j TOS --set-tos Minimize-Delay
4038 # iptables -A PREROUTING -t mangle -p tcp --sport ftp \
4039 -j TOS --set-tos Minimize-Delay
4040 # iptables -A PREROUTING -t mangle -p tcp --sport ftp-data \
4041 -j TOS --set-tos Maximize-Throughput
4042 </verb></tscreen>
4044 Now, this only works for data going from your telnet foreign host to your
4045 local computer. The other way around appears to be done for you, ie, telnet,
4046 ssh & friends all set the TOS field on outgoing packets automatically.
4048 Should you have an application that does not do this, you can always do it
4049 with netfilter. On your local box:
4051 <tscreen><verb>
4052 # iptables -A OUTPUT -t mangle -p tcp --dport telnet \
4053 -j TOS --set-tos Minimize-Delay
4054 # iptables -A OUTPUT -t mangle -p tcp --dport ftp \
4055 -j TOS --set-tos Minimize-Delay
4056 # iptables -A OUTPUT -t mangle -p tcp --dport ftp-data \
4057 -j TOS --set-tos Maximize-Throughput
4058 </verb></tscreen>
4060 <sect1>Transparent web-caching using netfilter, iproute2, ipchains and squid
4062 <label id="SQUID">
4063 This section was sent in by reader Ram Narula from Internet for Education
4064 (Thailand).
4066 The regular technique in accomplishing this in Linux
4067 is probably with use of ipchains AFTER making sure
4068 that the "outgoing" port 80(web) traffic gets routed through
4069 the server running squid.
4071 There are 3 common methods to make sure "outgoing"
4072 port 80 traffic gets routed to the server running squid
4073 and 4th one is being introduced here.
4075 <descrip>
4076 <tag>Making the gateway router do it.</tag>
4077 If you can tell your gateway router to
4078 match packets that has outgoing destination port
4079 of 80 to be sent to the IP address of squid server.
4083 This would put additional load on the router and
4084 some commercial routers might not even support this.
4085 <tag>Using a Layer 4 switch.</tag>
4086 Layer 4 switches can handle this without any problem.
4090 The cost for this equipment is usually very high. Typical
4091 layer 4 switch would normally cost more than
4092 a typical router+good linux server.
4093 <tag>Using cache server as network's gateway.</tag>
4094 You can force ALL traffic through cache server.
4098 This is quite risky because Squid does
4099 utilize lots of cpu power which might
4100 result in slower over-all network performance
4101 or the server itself might crash and no one on the
4102 network will be able to access the Internet if
4103 that occurs.
4106 <tag>Linux+NetFilter router.</tag>
4107 By using NetFilter another technique can be implemented
4108 which is using NetFilter for "mark"ing the packets
4109 with destination port 80 and using iproute2 to
4110 route the "mark"ed packets to the Squid server.
4111 </descrip>
4112 <tscreen><verb>
4113 |----------------|
4114 | Implementation |
4115 |----------------|
4117 Addresses used
4118 10.0.0.1 naret (NetFilter server)
4119 10.0.0.2 silom (Squid server)
4120 10.0.0.3 donmuang (Router connected to the Internet)
4121 10.0.0.4 kaosarn (other server on network)
4122 10.0.0.5 RAS
4123 10.0.0.0/24 main network
4124 10.0.0.0/19 total network
4126 |---------------|
4127 |Network diagram|
4128 |---------------|
4130 Internet
4132 donmuang
4134 ------------hub/switch----------
4135 | | | |
4136 naret silom kaosarn RAS etc.
4137 </verb></tscreen>
4138 First, make all traffic pass through naret by making
4139 sure it is the default gateway except for silom.
4140 Silom's default gateway has to be donmuang (10.0.0.3) or
4141 this would create web traffic loop.
4145 (all servers on my network had 10.0.0.1 as the default gateway
4146 which was the former IP address of donmuang router so what I did
4147 was changed the IP address of donmuang to 10.0.0.3 and gave
4148 naret ip address of 10.0.0.1)
4150 <tscreen><verb>
4151 Silom
4152 -----
4153 -setup squid and ipchains
4154 </verb></tscreen>
4157 Setup Squid server on silom, make sure it does support
4158 transparent caching/proxying, the default port is usually
4159 3128, so all traffic for port 80 has to be redirected to port
4160 3128 locally. This can be done by using ipchains with the following:
4162 <tscreen><verb>
4163 silom# ipchains -N allow1
4164 silom# ipchains -A allow1 -p TCP -s 10.0.0.0/19 -d 0/0 80 -j REDIRECT 3128
4165 silom# ipchains -I input -j allow1
4166 </verb></tscreen>
4168 <p>
4170 Or, in netfilter lingo:
4171 <tscreen><verb>
4172 silom# iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT --to-port 3128
4173 </verb></tscreen>
4175 (note: you might have other entries as well)
4178 For more information on setting Squid server please refer
4179 to Squid faq page on <url
4180 url="http://squid.nlanr.net" name="http://squid.nlanr.net">).
4184 Make sure ip forwarding is enabled on this server and the default
4185 gateway for this server is donmuang router (NOT naret).
4189 <tscreen><verb>
4190 Naret
4191 -----
4192 -setup iptables and iproute2
4193 -disable icmp REDIRECT messages (if needed)
4194 </verb></tscreen>
4196 <enum>
4197 <item>"Mark" packets of destination port 80 with value 2
4198 <tscreen><verb>
4199 naret# iptables -A PREROUTING -i eth0 -t mangle -p tcp --dport 80 \
4200 -j MARK --set-mark 2
4201 </verb></tscreen>
4202 </item>
4203 <item>Setup iproute2 so it will route packets with "mark" 2 to silom
4204 <tscreen><verb>
4205 naret# echo 202 www.out >> /etc/iproute2/rt_tables
4206 naret# ip rule add fwmark 2 table www.out
4207 naret# ip route add default via 10.0.0.2 dev eth0 table www.out
4208 naret# ip route flush cache
4210 </verb></tscreen>
4212 If donmuang and naret is on the same subnet then
4213 naret should not send out icmp REDIRECT messages.
4214 In this case it is, so icmp REDIRECTs has to be
4215 disabled by:
4216 <tscreen><verb>
4217 naret# echo 0 > /proc/sys/net/ipv4/conf/all/send_redirects
4218 naret# echo 0 > /proc/sys/net/ipv4/conf/default/send_redirects
4219 naret# echo 0 > /proc/sys/net/ipv4/conf/eth0/send_redirects
4221 </verb></tscreen>
4222 </item>
4223 </enum>
4225 The setup is complete, check the configuration
4227 <tscreen><verb>
4228 On naret:
4230 naret# iptables -t mangle -L
4231 Chain PREROUTING (policy ACCEPT)
4232 target prot opt source destination
4233 MARK tcp -- anywhere anywhere tcp dpt:www MARK set 0x2
4235 Chain OUTPUT (policy ACCEPT)
4236 target prot opt source destination
4238 naret# ip rule ls
4239 0: from all lookup local
4240 32765: from all fwmark 2 lookup www.out
4241 32766: from all lookup main
4242 32767: from all lookup default
4244 naret# ip route list table www.out
4245 default via 203.114.224.8 dev eth0
4247 naret# ip route
4248 10.0.0.1 dev eth0 scope link
4249 10.0.0.0/24 dev eth0 proto kernel scope link src 10.0.0.1
4250 127.0.0.0/8 dev lo scope link
4251 default via 10.0.0.3 dev eth0
4253 (make sure silom belongs to one of the above lines, in this case
4254 it's the line with 10.0.0.0/24)
4256 |------|
4257 |-DONE-|
4258 |------|
4261 </verb></tscreen>
4262 <sect2>Traffic flow diagram after implementation
4264 <tscreen><verb>
4266 |-----------------------------------------|
4267 |Traffic flow diagram after implementation|
4268 |-----------------------------------------|
4270 INTERNET
4274 -----------------donmuang router---------------------
4275 /\ /\ ||
4276 || || ||
4277 || \/ ||
4278 naret silom ||
4279 *destination port 80 traffic=========>(cache) ||
4280 /\ || ||
4281 || \/ \/
4282 \\===================================kaosarn, RAS, etc.
4284 </verb></tscreen>
4286 Note that the network is asymmetric as there is one extra hop on
4287 general outgoing path.
4289 <tscreen><verb>
4290 Here is run down for packet traversing the network from kaosarn
4291 to and from the Internet.
4293 For web/http traffic:
4294 kaosarn http request->naret->silom->donmuang->internet
4295 http replies from Internet->donmuang->silom->kaosarn
4297 For non-web/http requests(eg. telnet):
4298 kaosarn outgoing data->naret->donmuang->internet
4299 incoming data from Internet->donmuang->kaosarn
4300 </verb></tscreen>
4302 <sect1>Circumventing Path MTU Discovery issues with per route MTU settings
4304 For sending bulk data, the Internet generally works better when using larger
4305 packets. Each packet implies a routing decision, when sending a 1 megabyte
4306 file, this can either mean around 700 packets when using packets that are as
4307 large as possible, or 4000 if using the smallest default.
4309 However, not all parts of the Internet support full 1460 bytes of payload
4310 per packet. It is therefore necessary to try and find the largest packet
4311 that will 'fit', in order to optimize a connection.
4313 This process is called 'Path MTU Discovery', where MTU stands for 'Maximum
4314 Transfer Unit.'
4316 When a router encounters a packet that's too big too send in one piece, AND
4317 it has been flagged with the "Don't Fragment" bit, it returns an ICMP
4318 message stating that it was forced to drop a packet because of this. The
4319 sending host acts on this hint by sending smaller packets, and by iterating
4320 it can find the optimum packet size for a connection over a certain path.
4322 This used to work well until the Internet was discovered by hooligans who do
4323 their best to disrupt communications. This in turn lead administrators to
4324 either block or shape ICMP traffic in a misguided attempt to improve
4325 security or robustness of their Internet service.
4327 What has happened now is that Path MTU Discovery is working less and less
4328 well and fails for certain routes, which leads to strange TCP/IP sessions
4329 which die after a while.
4331 Although I have no proof for this, two sites who I used to have this problem
4332 with both run Alteon Acedirectors before the affected systems - perhaps
4333 somebody more knowledgeable can provide clues as to why this happens.
4335 <sect2>Solution
4337 When you encounter sites that suffer from this problem, you can disable Path
4338 MTU discovery by setting it manually. Koos van den Hout, slightly edited,
4339 writes:
4341 <tscreen>
4343 The following problem: I set the mtu/mru of my leased line running ppp to
4344 296 because it's only 33k6 and I cannot influence the queueing on the
4345 other side. At 296, the response to a keypress is within a reasonable
4346 timeframe.
4348 And, on my side I have a masqrouter running (of course) Linux.
4350 Recently I split 'server' and 'router' so most applications are run on a
4351 different machine than the routing happens on.
4353 I then had trouble logging into irc. Big panic! Some digging did find
4354 out that I got connected to irc, even showed up as 'connected' on irc
4355 but I did not receive the motd from irc. I checked what could be wrong
4356 and noted that I already had some previous trouble reaching certain
4357 websites related to the MTU, since I had no trouble reaching them when
4358 the MTU was 1500, the problem just showed when the MTU was set to 296.
4359 Since irc servers block about every kind of traffic not needed for their
4360 immediate operation, they also block icmp.
4362 I managed to convince the operators of a webserver that this was the cause
4363 of a problem, but the irc server operators were not going to fix this.
4365 So, I had to make sure outgoing masqueraded traffic started with the lower
4366 mtu of the outside link. But I want local ethernet traffic to have the
4367 normal mtu (for things like nfs traffic).
4369 Solution:
4370 <tscreen><verb>
4371 ip route add default via 10.0.0.1 mtu 296
4372 </verb></tscreen>
4374 (10.0.0.1 being the default gateway, the inside address of the
4375 masquerading router)
4376 </tscreen>
4378 In general, it is possible to override PMTU Discovery by setting specific
4379 routes. For example, if only a certain subnet is giving problems, this
4380 should help:
4382 <tscreen><verb>
4383 ip route add 195.96.96.0/24 via 10.0.0.1 mtu 1000
4384 </verb></tscreen>
4385 <sect1>Circumventing Path MTU Discovery issues with MSS Clamping (for ADSL,
4386 cable, PPPoE &amp; PPtP users)
4388 As explained above, Path MTU Discovery doesn't work as well as it should
4389 anymore. If you know for a fact that a hop somewhere in your network has a
4390 limited (&lt;1500) MTU, you cannot rely on PMTU Discovery finding this out.
4392 Besides MTU, there is yet another way to set the maximum packet size, the so
4393 called Maximum Segment Size. This is a field in the TCP Options part of a
4394 SYN packet.
4396 Recent Linux kernels, and a few pppoe drivers (notably, the excellent
4397 Roaring Penguin one), feature the possibility to 'clamp the MSS'.
4399 The good thing about this is that by setting the MSS value, you are telling
4400 the remote side unequivocally 'do not ever try to send me packets bigger
4401 than this value'. No ICMP traffic is needed to get this to work.
4403 The bad thing is that it's an obvious hack - it breaks 'end to end' by
4404 modifying packets. Having said that, we use this trick in many places and it
4405 works like a charm.
4407 In order for this to work you need at least iptables-1.2.1a and Linux 2.4.3
4408 or higher. The basic commandline is:
4409 <tscreen><verb>
4410 # iptables -A FORWARD -p tcp --tcp-flags SYN,RST SYN -j TCPMSS --clamp-mss-to-pmtu
4411 </verb></tscreen>
4413 This calculates the proper MSS for your link. If you are feeling brave, or
4414 think that you know best, you can also do something like this:
4416 <tscreen><verb>
4417 # iptables -A FORWARD -p tcp --tcp-flags SYN,RST SYN -j TCPMSS --set-mss 128
4418 </verb></tscreen>
4420 This sets the MSS of passing SYN packets to 128. Use this if you have VoIP
4421 with tiny packets, and huge http packets which are causing chopping in your
4422 voice calls.
4423 <sect1>The Ultimate Traffic Conditioner: Low Latency, Fast Up &amp; Downloads
4425 Note: This script has recently been upgraded and previously only worked for
4426 Linux clients in your network! So you might want to update if you have
4427 Windows machines or Macs in your network and noticed that they were not able
4428 to download faster while others were uploading.
4430 I attempted to create the holy grail:
4431 <descrip>
4432 <tag>Maintain low latency for interfactive traffic at all times</tag>
4433 This means that downloading or uploading files should not disturb SSH or
4434 even telnet. These are the most important things, even 200ms latency is
4435 sluggish to work over.
4436 <tag>Allow 'surfing' at reasonable speeds while up or downloading</tag>
4437 Even though http is 'bulk' traffic, other traffic should not drown it out
4438 too much.
4439 <tag>Make sure uploads don't harm downloads, and the other way around</tag>
4440 This is a much observed phenomenon where upstream traffic simply destroys
4441 download speed.
4442 </descrip>
4443 It turns out that all this is possible, at the cost of a tiny bit of
4444 bandwidth. The reason that uploads, downloads and ssh hurt eachother is the
4445 presence of large queues in many domestic access devices like cable or DSL
4446 modems.
4448 The next section explains in depth what causes the delays, and how we can
4449 fix them. You can safely skip it and head straight for the script if you
4450 don't care how the magic is performed.
4451 <sect2>Why it doesn't work well by default
4453 ISPs know that they are benchmarked solely on how fast people can download.
4454 Besides available bandwidth, download speed is influenced heavily by packet
4455 loss, which seriously hampers TCP/IP performance. Large queues can help
4456 prevent packetloss, and speed up downloads. So ISPs configure large queues.
4458 These large queues however damage interactivity. A keystroke must first
4459 travel the upstream queue, which may be seconds (!) long and go to your
4460 remote host. It is then displayed, which leads to a packet coming back, which
4461 must then traverse the downstream queue, located at your ISP, before it
4462 appears on your screen.
4464 This HOWTO teaches you how to mangle and process the queue in many ways, but
4465 sadly, not all queues are accessible to us. The queue over at the ISP is
4466 completely off-limits, whereas the upstream queue probably lives inside your
4467 cable modem or DSL device. You may or may not be able to configure it. Most
4468 probably not.
4470 So, what next? As we can't control either of those queues, they must be
4471 eliminated, and moved to your Linux router. Luckily this is possible.
4473 <descrip>
4474 <tag>Limit upload speed</tag>
4475 By limiting our upload speed to slightly less than the truly available rate,
4476 no queues are built up in our modem. The queue is now moved to Linux.
4477 <tag>Limit download speed</tag>
4478 This is slightly trickier as we can't really influence how fast the internet
4479 ships us data. We can however drop packets that are coming in too fast,
4480 which causes TCP/IP to slow down to just the rate we want. Because we don't
4481 want to drop traffic unnecessarily, we configure a 'burst' size we allow at
4482 higher speed.
4483 </descrip>
4485 Now, once we have done this, we have eliminated the downstream queue totally
4486 (except for short bursts), and gain the ability to manage the upstream queue
4487 with all the power Linux offers.
4489 What remains to be done is to make sure interactive traffic jumps to the
4490 front of the upstream queue. To make sure that uploads don't hurt downloads,
4491 we also move ACK packets to the front of the queue. This is what normally
4492 causes the huge slowdown observed when generating bulk traffic both ways.
4493 The ACKnowledgements for downstream traffic must compete with upstream
4494 traffic, and get delayed in the process.
4496 If we do all this we get the following measurements using an excellent ADSL
4497 connection from xs4all in the Netherlands:
4499 <verb>
4500 Baseline latency:
4501 round-trip min/avg/max = 14.4/17.1/21.7 ms
4503 Without traffic conditioner, while downloading:
4504 round-trip min/avg/max = 560.9/573.6/586.4 ms
4506 Without traffic conditioner, while uploading:
4507 round-trip min/avg/max = 2041.4/2332.1/2427.6 ms
4509 With conditioner, during 220kbit/s upload:
4510 round-trip min/avg/max = 15.7/51.8/79.9 ms
4512 With conditioner, during 850kbit/s download:
4513 round-trip min/avg/max = 20.4/46.9/74.0 ms
4515 When uploading, downloads proceed at ~80% of the available speed. Uploads
4516 at around 90%. Latency then jumps to 850 ms, still figuring out why.
4517 </verb>
4519 What you can expect from this script depends a lot on your actual uplink
4520 speed. When uploading at full speed, there will always be a single packet
4521 ahead of your keystroke. That is the lower limit to the latency you can
4522 achieve - divide your MTU by your upstream speed to calculate. Typical
4523 values will be somewhat higher than that. Lower your MTU for better effects!
4525 Next, two versions of this script, one with Devik's excellent HTB, the other
4526 with CBQ which is in each Linux kernel, unlike HTB. Both are tested and work
4527 well.
4528 <sect2>The actual script (CBQ)
4530 Works on all kernels. Within the CBQ
4531 qdisc we place two Stochastic Fairness Queues that make sure that multiple
4532 bulk streams don't drown each other out.
4534 Downstream traffic is policed using a tc filter containing a Token Bucket
4535 Filter.
4537 You might improve on this script by adding 'bounded' to the line that starts
4538 with 'tc class add .. classid 1:20'. If you lowered your MTU, also lower the
4539 allot &amp; avpkt numbers!
4541 <verb>
4542 #!/bin/bash
4544 # The Ultimate Setup For Your Internet Connection At Home
4547 # Set the following values to somewhat less than your actual download
4548 # and uplink speed. In kilobits
4549 DOWNLINK=800
4550 UPLINK=220
4551 DEV=ppp0
4553 # clean existing down- and uplink qdiscs, hide errors
4554 tc qdisc del dev $DEV root 2> /dev/null > /dev/null
4555 tc qdisc del dev $DEV ingress 2> /dev/null > /dev/null
4557 ###### uplink
4559 # install root CBQ
4561 tc qdisc add dev $DEV root handle 1: cbq avpkt 1000 bandwidth 10mbit
4563 # shape everything at $UPLINK speed - this prevents huge queues in your
4564 # DSL modem which destroy latency:
4565 # main class
4567 tc class add dev $DEV parent 1: classid 1:1 cbq rate ${UPLINK}kbit \
4568 allot 1500 prio 5 bounded isolated
4570 # high prio class 1:10:
4572 tc class add dev $DEV parent 1:1 classid 1:10 cbq rate ${UPLINK}kbit \
4573 allot 1600 prio 1 avpkt 1000
4575 # bulk and default class 1:20 - gets slightly less traffic,
4576 # and a lower priority:
4578 tc class add dev $DEV parent 1:1 classid 1:20 cbq rate $[9*$UPLINK/10]kbit \
4579 allot 1600 prio 2 avpkt 1000
4581 # both get Stochastic Fairness:
4582 tc qdisc add dev $DEV parent 1:10 handle 10: sfq perturb 10
4583 tc qdisc add dev $DEV parent 1:20 handle 20: sfq perturb 10
4585 # start filters
4586 # TOS Minimum Delay (ssh, NOT scp) in 1:10:
4587 tc filter add dev $DEV parent 1:0 protocol ip prio 10 u32 \
4588 match ip tos 0x10 0xff flowid 1:10
4590 # ICMP (ip protocol 1) in the interactive class 1:10 so we
4591 # can do measurements & impress our friends:
4592 tc filter add dev $DEV parent 1:0 protocol ip prio 11 u32 \
4593 match ip protocol 1 0xff flowid 1:10
4595 # To speed up downloads while an upload is going on, put ACK packets in
4596 # the interactive class:
4598 tc filter add dev $DEV parent 1: protocol ip prio 12 u32 \
4599 match ip protocol 6 0xff \
4600 match u8 0x05 0x0f at 0 \
4601 match u16 0x0000 0xffc0 at 2 \
4602 match u8 0x10 0xff at 33 \
4603 flowid 1:10
4605 # rest is 'non-interactive' ie 'bulk' and ends up in 1:20
4607 tc filter add dev $DEV parent 1: protocol ip prio 13 u32 \
4608 match ip dst 0.0.0.0/0 flowid 1:20
4610 ########## downlink #############
4611 # slow downloads down to somewhat less than the real speed to prevent
4612 # queuing at our ISP. Tune to see how high you can set it.
4613 # ISPs tend to have *huge* queues to make sure big downloads are fast
4615 # attach ingress policer:
4617 tc qdisc add dev $DEV handle ffff: ingress
4619 # filter *everything* to it (0.0.0.0/0), drop everything that's
4620 # coming in too fast:
4622 tc filter add dev $DEV parent ffff: protocol ip prio 50 u32 match ip src \
4623 0.0.0.0/0 police rate ${DOWNLINK}kbit burst 10k drop flowid :1
4624 </verb>
4625 If you want this script to be run by ppp on connect, copy it to
4626 /etc/ppp/ip-up.d.
4628 If the last two lines give an error, update your tc tool to a newer version!
4629 <sect2>The actual script (HTB)
4631 The following script achieves all goals using the wonderful HTB queue, see
4632 the relevant chapter. Well worth patching your kernel for!
4633 <verb>
4634 #!/bin/bash
4636 # The Ultimate Setup For Your Internet Connection At Home
4639 # Set the following values to somewhat less than your actual download
4640 # and uplink speed. In kilobits
4641 DOWNLINK=800
4642 UPLINK=220
4643 DEV=ppp0
4645 # clean existing down- and uplink qdiscs, hide errors
4646 tc qdisc del dev $DEV root 2> /dev/null > /dev/null
4647 tc qdisc del dev $DEV ingress 2> /dev/null > /dev/null
4649 ###### uplink
4651 # install root HTB, point default traffic to 1:20:
4653 tc qdisc add dev $DEV root handle 1: htb default 20
4655 # shape everything at $UPLINK speed - this prevents huge queues in your
4656 # DSL modem which destroy latency:
4658 tc class add dev $DEV parent 1: classid 1:1 htb rate ${UPLINK}kbit burst 6k
4660 # high prio class 1:10:
4662 tc class add dev $DEV parent 1:1 classid 1:10 htb rate ${UPLINK}kbit \
4663 burst 6k prio 1
4665 # bulk & default class 1:20 - gets slightly less traffic,
4666 # and a lower priority:
4668 tc class add dev $DEV parent 1:1 classid 1:20 htb rate $[9*$UPLINK/10]kbit \
4669 burst 6k prio 2
4671 # both get Stochastic Fairness:
4672 tc qdisc add dev $DEV parent 1:10 handle 10: sfq perturb 10
4673 tc qdisc add dev $DEV parent 1:20 handle 20: sfq perturb 10
4675 # TOS Minimum Delay (ssh, NOT scp) in 1:10:
4676 tc filter add dev $DEV parent 1:0 protocol ip prio 10 u32 \
4677 match ip tos 0x10 0xff flowid 1:10
4679 # ICMP (ip protocol 1) in the interactive class 1:10 so we
4680 # can do measurements & impress our friends:
4681 tc filter add dev $DEV parent 1:0 protocol ip prio 10 u32 \
4682 match ip protocol 1 0xff flowid 1:10
4684 # To speed up downloads while an upload is going on, put ACK packets in
4685 # the interactive class:
4687 tc filter add dev $DEV parent 1: protocol ip prio 10 u32 \
4688 match ip protocol 6 0xff \
4689 match u8 0x05 0x0f at 0 \
4690 match u16 0x0000 0xffc0 at 2 \
4691 match u8 0x10 0xff at 33 \
4692 flowid 1:10
4694 # rest is 'non-interactive' ie 'bulk' and ends up in 1:20
4697 ########## downlink #############
4698 # slow downloads down to somewhat less than the real speed to prevent
4699 # queuing at our ISP. Tune to see how high you can set it.
4700 # ISPs tend to have *huge* queues to make sure big downloads are fast
4702 # attach ingress policer:
4704 tc qdisc add dev $DEV handle ffff: ingress
4706 # filter *everything* to it (0.0.0.0/0), drop everything that's
4707 # coming in too fast:
4709 tc filter add dev $DEV parent ffff: protocol ip prio 50 u32 match ip src \
4710 0.0.0.0/0 police rate ${DOWNLINK}kbit burst 10k drop flowid :1
4711 </verb>
4713 If you want this script to be run by ppp on connect, copy it to
4714 /etc/ppp/ip-up.d.
4716 If the last two lines give an error, update your tc tool to a newer version!
4717 <sect>Building bridges, and pseudo-bridges with Proxy ARP
4719 Bridges are devices which can be installed in a network without any
4720 reconfiguration. A network switch is basically a many-port bridge. A bridge
4721 is often a 2-port switch. Linux does however support multiple interfaces in
4722 a bridge, making it a true switch.
4724 Bridges are often deployed when confronted with a broken network that needs
4725 to be fixed without any alterations. Because the bridge is a layer-2 device,
4726 one layer below IP, routers and servers are not aware of its existence.
4727 This means that you can transparently block or modify certain packets, or do
4728 shaping.
4730 Another good thing is that a bridge can often be replaced by a cross cable
4731 or a hub, should it break down.
4733 The bad news is that a bridge can cause great confusion unless it is very
4734 well documented. It does not appear in traceroutes, but somehow packets
4735 disappear or get changed from point A to point B ('this network is
4736 HAUNTED!'). You should also wonder if an organization that 'does not want to
4737 change anything' is doing the right thing.
4739 The Linux 2.4/2.5 bridge is documented on
4741 <url url=" http://bridge.sourceforge.net/" name="this page">.
4743 <sect1>State of bridging and iptables
4745 As of Linux 2.4.14, bridging and iptables do not 'see' each other without
4746 help. If you bridge packets from eth0 to eth1, they do not 'pass' by
4747 iptables. This means that you cannot do filtering, or NAT or mangling or
4748 whatever.
4750 There are several projects going on to fix this, the truly right one is by
4751 the author of the Linux 2.4 bridging code, Lennert Buytenhek. He recently
4752 informed us that as of bridge-nf 0.0.2 (see the url above), the code is
4753 stable and usable in production environments. He is now asking the kernel
4754 people if and how the patch can be merged, stay tuned!
4756 <sect1>Bridging and shaping
4758 This does work as advertised. Be sure to figure out which side each
4759 interface is on, otherwise you might be shaping outbound traffic in your
4760 internal interface, which won't work. Use tcpdump if needed.
4762 <sect1>Pseudo-bridges with Proxy-ARP
4764 If you just want to implement a Pseudo-bridge, skip down a few sections
4765 to 'Implementing it', but it is wise to read a bit about how it works in
4766 practice.
4768 A Pseudo-bridge works a bit differently. By default, a bridge passes packets
4769 unaltered from one interface to the other. It only looks at the hardware
4770 address of packets to determine what goes where. This in turn means that you
4771 can bridge traffic that Linux does not understand, as long as it has an
4772 hardware address it does.
4774 A 'Pseudo-bridge' works differently and looks more like a hidden router than
4775 a bridge, but like a bridge, it has little impact on network design.
4777 An advantage of the fact that it is not a brige lies in the fact that
4778 packets really pass through the kernel, and can be filtered, changed,
4779 redirected or rerouted.
4781 A real bridge can also be made to perform these feats, but it needs special
4782 code, like the Ethernet Frame Diverter, or the above mentioned patch.
4784 Another advantage of a pseudo-bridge is that it does not pass packets it
4785 does not understand - thus cleaning your network of a lot of cruft. In cases
4786 where you need this cruft (like SAP packets, or Netbeui), use a real bridge.
4787 <sect2>ARP &amp; Proxy-ARP
4789 When a host wants to talk to another host on the same physical network
4790 segment, it sends out an Address Resolution Protocol packet, which, somewhat
4791 simplified, reads like this 'who has 10.0.0.1, tell 10.0.0.7'. In response
4792 to this, 10.0.0.1 replies with a short 'here' packet.
4794 10.0.0.7 then sends packets to the hardware address mentioned in the 'here'
4795 packet. It caches this hardware address for a relatively long time, and
4796 after the cache expires, it reasks the question.
4798 When building a Pseudo-bridge, we instruct the bridge to reply to these ARP
4799 packets, which causes the hosts in the network to send its packets to the
4800 bridge. The brige then processes these packets, and sends them to the
4801 relevant interface.
4803 So, in short, whenever a host on one side of the bridge asks for the
4804 hardware address of a host on the other, the bridge replies with a packet
4805 that says 'hand it to me'.
4807 This way, all data traffic gets transmitted to the right place, and always
4808 passes through the bridge.
4809 <sect2>Implementing it
4811 In the bad old days, it used to be possible to instruct the Linux Kernel to
4812 perform 'proxy-ARP' for just any subnet. So, to configure a pseudo-bridge,
4813 you would have to specify both the proper routes to both sides of the bridge
4814 AND create matching proxy-ARP rules. This is bad in that it requires a lot
4815 of typing, but also because it easily allows you to make mistakes which make
4816 your bridge respond to ARP queries for networks it does not know how to
4817 route.
4819 With Linux 2.4/2.5 (and possibly 2.2), this possibility has been withdrawn and
4820 has been replaced by a flag in the /proc directory, called 'proxy_arp'. The
4821 procedure for building a pseudo-bridge is then:
4823 <enum>
4824 <item>Assign an IP address to both interfaces, the 'left' and the 'right'
4826 <item>Create routes so your machine knows which hosts reside on the left,
4827 and which on the right
4828 <item>Turn on proxy-ARP on both interfaces, echo 1 >
4829 /proc/sys/net/ipv4/conf/ethL/proxy_arp, echo 1 >
4830 /proc/sys/net/ipv4/conf/ethR/proxy_arp, where L and R stand for the numbers
4831 of your interfaces on the left and on the right side
4832 </enum>
4834 Also, do not forget to turn on the ip_forwarding flag! When converting from
4835 a true bridge, you may find that this flag was turned off as it is not
4836 needed when bridging.
4838 Another thing you might note when converting is that you need to clear the
4839 arp cache of computers in the network - the arp cache might contain old
4840 pre-bridge hardware addresses which are no longer correct.
4842 On a Cisco, this is done using the command 'clear arp-cache', under
4843 Linux, use 'arp -d ip.address'. You can also wait for the cache to expire
4844 manually, which can take rather long.
4846 You may also discover that your network was misconfigured if you are/were of
4847 the habit of specifying routes without netmasks. To explain, some versions
4848 of route may have guessed your netmask right in the past, or guessed wrong
4849 without you noticing. When doing surgical routing like described above, it
4850 is *vital* that you check your netmasks!
4851 <sect>Dynamic routing - OSPF and BGP
4853 Once your network starts to get really big, or you start to consider 'the
4854 internet' as your network, you need tools which dynamically route your data.
4855 Sites are often connected to each other with multiple links, and more are
4856 popping up all the time.
4858 The Internet has mostly standardised on OSPF and BGP4 (rfc1771). Linux
4859 supports both, by way of <tt>gated</tt> and <tt>zebra</tt>
4861 While currently not within the scope of this document, we would like to
4862 point you to the definitive works:
4864 Overview:
4866 Cisco Systems
4867 <url
4868 url="http://www.cisco.com/univercd/cc/td/doc/cisintwk/idg4/nd2003.htm"
4869 name="Designing large-scale IP Internetworks">
4873 For OSPF:
4875 Moy, John T.
4876 "OSPF. The anatomy of an Internet routing protocol"
4877 Addison Wesley. Reading, MA. 1998.
4879 Halabi has also written a good guide to OSPF routing design, but this
4880 appears to have been dropped from the Cisco web site.
4883 For BGP:
4885 Halabi, Bassam
4886 "Internet routing architectures"
4887 Cisco Press (New Riders Publishing). Indianapolis, IN. 1997.
4889 also
4891 Cisco Systems
4893 <url
4894 url="http://www.cisco.com/univercd/cc/td/doc/cisintwk/ics/icsbgp4.htm"
4895 name="Using the Border Gateway Protocol for interdomain routing">
4898 Although the examples are Cisco-specific, they are remarkably similar
4899 to the configuration language in Zebra :-)
4900 <sect>Other possibilities
4902 This chapter is a list of projects having to do with advanced Linux routing
4903 &amp; traffic shaping. Some of these links may deserve chapters of their
4904 own, some are documented very well of themselves, and don't need more HOWTO.
4906 <descrip>
4907 <tag>802.1Q VLAN Implementation for Linux <url url="http://scry.wanfear.com/~greear/vlan.html"
4908 name="(site)"></tag>
4910 VLANs are a very cool way to segregate your
4911 networks in a more virtual than physical way. Good information on VLANs can
4912 be found <url
4913 url="ftp://ftp.netlab.ohio-state.edu/pub/jain/courses/cis788-97/virtual_lans/index.htm"
4914 name="here">. With this implementation, you can have your Linux box talk
4915 VLANs with machines like Cisco Catalyst, 3Com: {Corebuilder, Netbuilder II,
4916 SuperStack II switch 630}, Extreme Ntwks Summit 48, Foundry: {ServerIronXL,
4917 FastIron}.
4919 A great HOWTO about VLANs can be found <url
4920 url="http://scry.wanfear.com/~greear/vlan/cisco_howto.html" name="here">.
4922 Update: has been included in the kernel as of 2.4.14 (perhaps 13).
4923 <tag>Alternate 802.1Q VLAN Implementation for Linux <url
4924 url="http://vlan.sourceforge.net "
4925 name="(site)"></tag>
4926 Alternative VLAN implementation for linux. This project was started out of
4927 disagreement with the 'established' VLAN project's architecture and coding
4928 style, resulting in a cleaner overall design.
4930 <tag>Linux Virtual Server <url url="http://www.LinuxVirtualServer.org/"
4931 name="(site)"></tag>
4933 These people are brilliant. The Linux Virtual Server is a highly scalable and
4934 highly available server built on a cluster of real servers, with the load
4935 balancer running on the Linux operating system. The architecture of the
4936 cluster is transparent to end users. End users only see a single virtual
4937 server.
4939 In short whatever you need to loadbalance, at whatever level of traffic, LVS
4940 will have a way of doing it. Some of their techniques are positively evil!
4941 For example, they let several machines have the same IP address on a
4942 segment, but turn off ARP on them. Only the LVS machine does ARP - it then
4943 decides which of the backend hosts should handle an incoming packet, and
4944 sends it directly to the right MAC address of the backend server. Outgoing
4945 traffic will flow directly to the router, and not via the LVS machine, which
4946 does therefor not need to see your 5Gbit/s of content flowing to the world,
4947 and cannot be a bottleneck.
4949 The LVS is implemented as a kernel patch in Linux 2.0 and 2.2, but as a
4950 Netfilter module in 2.4/2.5, so it does not need kernel patches! Their 2.4
4951 support is still in early development, so beat on it and give feedback or
4952 send patches.
4954 <tag>CBQ.init <url url="ftp://ftp.equinox.gu.net/pub/linux/cbq/"
4955 name="(site)"></tag>
4956 Configuring CBQ can be a bit daunting, especially if all you want to do is
4957 shape some computers behind a router. CBQ.init can help you configure Linux
4958 with a simplified syntax.
4960 For example, if you want all computers in your 192.168.1.0/24 subnet
4961 (on 10mbit eth1) to be limited to 28kbit/s download speed, put
4962 this in the CBQ.init configuration file:
4964 <tscreen><verb>
4965 DEVICE=eth1,10Mbit,1Mbit
4966 RATE=28Kbit
4967 WEIGHT=2Kbit
4968 PRIO=5
4969 RULE=192.168.1.0/24
4970 </verb></tscreen>
4972 By all means use this program if the 'how and why' don't interest you.
4973 We're using CBQ.init in production and it works very well. It can even do
4974 some more advanced things, like time dependent shaping. The documentation is
4975 embedded in the script, which explains why you can't find a README.
4977 <tag>Chronox easy shaping scripts <url url="http://www.chronox.de"
4978 name="(site)"></tag>
4980 Stephan Mueller (smueller@chronox.de) wrote two useful scripts, 'limit.conn'
4981 and 'shaper'. The first one allows you to easily throttle a single download
4982 session, like this:
4984 <tscreen><verb>
4985 # limit.conn -s SERVERIP -p SERVERPORT -l LIMIT
4986 </verb></tscreen>
4988 It works on Linux 2.2 and 2.4/2.5.
4990 The second script is more complicated, and can be used to make lots of
4991 different queues based on iptables rules, which are used to mark packets
4992 which are then shaped.
4994 <tag>Virtual Router
4995 Redundancy Protocol implementation <url url="http://w3.arobas.net/~jetienne/vrrpd/index.html"
4996 name="(site)"></tag>
4998 This is purely for redundancy. Two machines with their own IP address and
4999 MAC Address together create a third IP Address and MAC Address, which is
5000 virtual. Originally intended purely for routers, which need constant MAC
5001 addresses, it also works for other servers.
5003 The beauty of this approach is the incredibly easy configuration. No kernel
5004 compiling or patching required, all userspace.
5006 Just run this on all machines participating in a service:
5007 <tscreen><verb>
5008 # vrrpd -i eth0 -v 50 10.0.0.22
5009 </verb></tscreen>
5011 And you are in business! 10.0.0.22 is now carried by one of your servers,
5012 probably the first one to run the vrrp daemon. Now disconnect that computer
5013 from the network and very rapidly one of the other computers will assume the
5014 10.0.0.22 address, as well as the MAC address.
5016 I tried this over here and had it up and running in 1 minute. For some
5017 strange reason it decided to drop my default gateway, but the -n flag
5018 prevented that.
5020 This is a 'live' failover:
5022 <tscreen><verb>
5023 64 bytes from 10.0.0.22: icmp_seq=3 ttl=255 time=0.2 ms
5024 64 bytes from 10.0.0.22: icmp_seq=4 ttl=255 time=0.2 ms
5025 64 bytes from 10.0.0.22: icmp_seq=5 ttl=255 time=16.8 ms
5026 64 bytes from 10.0.0.22: icmp_seq=6 ttl=255 time=1.8 ms
5027 64 bytes from 10.0.0.22: icmp_seq=7 ttl=255 time=1.7 ms
5028 </verb></tscreen>
5030 Not *one* ping packet was lost! Just after packet 4, I disconnected my P200
5031 from the network, and my 486 took over, which you can see from the higher
5032 latency.
5033 </descrip>
5034 <sect>Further reading
5036 <descrip>
5037 <tag><url url="http://snafu.freedom.org/linux2.2/iproute-notes.html"
5038 name="http://snafu.freedom.org/linux2.2/iproute-notes.html"></tag>
5039 Contains lots of technical information, comments from the kernel
5040 <tag><url url="http://www.davin.ottawa.on.ca/ols/"
5041 name="http://www.davin.ottawa.on.ca/ols/"></tag>
5042 Slides by Jamal Hadi Salim, one of the authors of Linux traffic control
5043 <tag><url url="http://defiant.coinet.com/iproute2/ip-cref/"
5044 name="http://defiant.coinet.com/iproute2/ip-cref/"></tag>
5045 HTML version of Alexeys LaTeX documentation - explains part of iproute2 in
5046 great detail
5047 <tag><url url="http://www.aciri.org/floyd/cbq.html"
5048 name="http://www.aciri.org/floyd/cbq.html"></tag>
5049 Sally Floyd has a good page on CBQ, including her original papers. None of
5050 it is Linux specific, but it does a fair job discussing the theory and uses
5051 of CBQ.
5052 Very technical stuff, but good reading for those so inclined.
5054 <tag>Differentiated Services on Linux</tag>
5055 This <url url="ftp://icaftp.epfl.ch/pub/linux/diffserv/misc/dsid-01.txt.gz"
5056 name="document"> by Werner Almesberger, Jamal Hadi Salim and Alexey
5057 Kuznetsov describes DiffServ facilities in the Linux kernel, amongst which
5058 are TBF, GRED, the DSMARK qdisc and the tcindex classifyer.
5061 <tag><url url="http://ceti.pl/~kravietz/cbq/NET4_tc.html"
5062 name="http://ceti.pl/~kravietz/cbq/NET4_tc.html"></tag>
5063 Yet another HOWTO, this time in Polish! You can copy/paste command lines
5064 however, they work just the same in every language. The author is
5065 cooperating with us and may soon author sections of this HOWTO.
5067 <tag><url
5068 url="http://www.cisco.com/univercd/cc/td/doc/product/software/ios111/cc111/car.htm"
5069 name="IOS Committed Access Rate"></tag>
5070 <label id="CAR">
5071 From the helpful folks of Cisco who have the laudable habit of putting
5072 their documentation online. Cisco syntax is different but the concepts are
5073 the same, except that we can do more and do it without routers the price of
5074 cars :-)
5076 <tag>Docum experimental site<url url="http://www.docum.org"
5077 name="(site)"></tag>
5078 Stef Coene is busy convincing his boss to sell Linux support, and so he is
5079 experimenting a lot, especially with managing bandwidth. His site has a lot
5080 of practical information, examples, tests and also points out some CBQ/tc bugs.
5082 <tag>TCP/IP Illustrated, volume 1, W. Richard Stevens, ISBN 0-201-63346-9</tag>
5083 Required reading if you truly want to understand TCP/IP. Entertaining as
5084 well.
5086 </descrip>
5087 <sect>Acknowledgements
5088 <p>
5089 It is our goal to list everybody who has contributed to this HOWTO, or
5090 helped us demystify how things work. While there are currently no plans
5091 for a Netfilter type scoreboard, we do like to recognise the people who are
5092 helping.
5094 <itemize>
5095 <item>Juanjo Alins &lt;juanjo%mat.upc.es&gt;
5096 <item>Michael T. Babcock &lt;mbabcock@fibrespeed.net&gt;
5097 <item>Ard van Breemen &lt;ard%kwaak.net&gt;
5098 <item>Ron Brinker &lt;service%emcis.com&gt;
5099 <item>?ukasz Bromirski &lt;L.Bromirski@prosys.com.pl&gt;
5100 <item>Lennert Buytenhek &lt;buytenh@gnu.org&gt;
5101 <item>Esteve Camps &lt;esteve@hades.udg.es&gt;
5102 <item>Stef Coene &lt;stef.coene@docum.org&gt;
5103 <item>Don Cohen &lt;don-lartc%isis.cs3-inc.com&gt;
5104 <item>Jonathan Corbet &lt;lwn%lwn.net&gt;
5105 <item>Gerry Creager N5JXS &lt;gerry%cs.tamu.edu&gt;
5106 <item>Marco Davids &lt;marco@sara.nl&gt;
5107 <item>Jonathan Day &lt;jd9812@my-deja.com&gt;
5108 <item>Martin Devera aka devik &lt;devik@cdi.cz&gt;
5109 <item>Stephan "Kobold" Gehring &lt;Stephan.Gehring@bechtle.de&gt;
5110 <item>Jacek Glinkowski &lt;jglinkow%hns.com&gt;
5111 <item>Andrea Glorioso &lt;sama%perchetopi.org&gt;
5112 <item>Nadeem Hasan &lt;nhasan@usa.net&gt;
5113 <item>Erik Hensema &lt;erik%hensema.xs4all.nl&gt;
5114 <item>Vik Heyndrickx &lt;vik.heyndrickx@edchq.com&gt;
5115 <item>Spauldo Da Hippie &lt;spauldo%usa.net&gt;
5116 <item>Koos van den Hout &lt;koos@kzdoos.xs4all.nl&gt;
5117 <item>Stefan Huelbrock &lt;shuelbrock%datasystems.de&gt;
5118 <item>Alexander W. Janssen &lt;yalla%ynfonatic.de&gt;
5119 <item>Gareth John &lt;gdjohn%zepler.org&gt;
5120 <item>Martin Josefsson &lt;gandalf%wlug.westbo.se&gt;
5121 <item>Andi Kleen &lt;ak%suse.de&gt;
5122 <item>Andreas J. Koenig &lt;andreas.koenig%anima.de&gt;
5123 <item>Pawel Krawczyk &lt;kravietz%alfa.ceti.pl&gt;
5124 <item>Amit Kucheria &lt;amitk@ittc.ku.edu&gt;
5125 <item>Edmund Lau &lt;edlau%ucf.ics.uci.edu&gt;
5126 <item>Philippe Latu &lt;philippe.latu%linux-france.org&gt;
5127 <item>Arthur van Leeuwen &lt;arthurvl%sci.kun.nl&gt;
5128 <item>Jason Lunz &lt;j@cc.gatech.edu&gt;
5129 <item>Stuart Lynne &lt;sl@fireplug.net&gt;
5130 <item>Alexey Mahotkin &lt;alexm@formulabez.ru&gt;
5131 <item>Predrag Malicevic &lt;pmalic@ieee.org&gt;
5132 <item>Andreas Mohr &lt;andi%lisas.de&gt;
5133 <item>Andrew Morton &lt;akpm@zip.com.au&gt;
5134 <item>Wim van der Most
5135 <item>Stephan Mueller &lt;smueller@chronox.de&gt;
5136 <item>Togan Muftuoglu &lt;toganm%yahoo.com&gt;
5137 <item>Chris Murray &lt;cmurray@stargate.ca&gt;
5138 <item>Patrick Nagelschmidt &lt;dto%gmx.net&gt;
5139 <item>Ram Narula &lt;ram@princess1.net&gt;
5140 <item>Jorge Novo &lt;jnovo@educanet.net&gt;
5141 <item>Patrik &lt;ph@kurd.nu&gt;
5142 <item>Lutz Pre&szlig;ler &lt;Lutz.Pressler%SerNet.DE&gt;
5143 <item>Jason Pyeron &lt;jason%pyeron.com&gt;
5144 <item>Rusty Russell &lt;rusty%rustcorp.com.au&gt;
5145 <item>Mihai RUSU &lt;dizzy%roedu.net&gt;
5146 <item>Jamal Hadi Salim &lt;hadi%cyberus.ca&gt;
5147 <item>David Sauer &lt;davids%penguin.cz&gt;
5148 <item>Sheharyar Suleman Shaikh &lt;sss23@drexel.edu&gt;
5149 <item>Stewart Shields &lt;MourningBlade%bigfoot.com&gt;
5150 <item>Nick Silberstein &lt;nhsilber%yahoo.com&gt;
5151 <item>Konrads Smelkov &lt;konrads@interbaltika.com&gt;
5152 <item>Andreas Steinmetz &lt;ast%domdv.de&gt;
5153 <item>Jason Tackaberry &lt;tack@linux.com&gt;
5154 <item>Charles Tassell &lt;ctassell%isn.net&gt;
5155 <item>Glen Turner &lt;glen.turner%aarnet.edu.au&gt;
5156 <item>Tea Sponsor: Eric Veldhuyzen &lt;eric%terra.nu&gt;
5157 <item>Song Wang &lt;wsong@ece.uci.edu&gt;
5158 </itemize>
5160 </article>