1 /* $KAME: altq_cbq.c,v 1.20 2004/04/17 10:54:48 kjc Exp $ */
2 /* $DragonFly: src/sys/net/altq/altq_cbq.c,v 1.7 2008/05/14 11:59:23 sephe Exp $ */
5 * Copyright (c) Sun Microsystems, Inc. 1993-1998 All rights reserved.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
18 * 3. All advertising materials mentioning features or use of this software
19 * must display the following acknowledgement:
20 * This product includes software developed by the SMCC Technology
21 * Development Group at Sun Microsystems, Inc.
23 * 4. The name of the Sun Microsystems, Inc nor may not be used to endorse or
24 * promote products derived from this software without specific prior
27 * SUN MICROSYSTEMS DOES NOT CLAIM MERCHANTABILITY OF THIS SOFTWARE OR THE
28 * SUITABILITY OF THIS SOFTWARE FOR ANY PARTICULAR PURPOSE. The software is
29 * provided "as is" without express or implied warranty of any kind.
31 * These notices must be retained in any copies of any part of this software.
36 #include "opt_inet6.h"
38 #ifdef ALTQ_CBQ /* cbq is enabled by ALTQ_CBQ option in opt_altq.h */
40 #include <sys/param.h>
41 #include <sys/malloc.h>
43 #include <sys/socket.h>
44 #include <sys/systm.h>
46 #include <sys/callout.h>
47 #include <sys/errno.h>
49 #include <sys/thread.h>
52 #include <net/ifq_var.h>
53 #include <netinet/in.h>
55 #include <net/pf/pfvar.h>
56 #include <net/altq/altq.h>
57 #include <net/altq/altq_cbq.h>
59 #include <sys/thread2.h>
62 * Forward Declarations.
64 static int cbq_class_destroy(cbq_state_t
*, struct rm_class
*);
65 static struct rm_class
*clh_to_clp(cbq_state_t
*, uint32_t);
66 static int cbq_clear_interface(cbq_state_t
*);
67 static int cbq_request(struct ifaltq
*, int, void *);
68 static int cbq_enqueue(struct ifaltq
*, struct mbuf
*,
69 struct altq_pktattr
*);
70 static struct mbuf
*cbq_dequeue(struct ifaltq
*, struct mbuf
*, int);
71 static void cbqrestart(struct ifaltq
*);
72 static void get_class_stats(class_stats_t
*, struct rm_class
*);
73 static void cbq_purge(cbq_state_t
*);
77 * cbq_class_destroy(cbq_mod_state_t *, struct rm_class *) - This
78 * function destroys a given traffic class. Before destroying
79 * the class, all traffic for that class is released.
82 cbq_class_destroy(cbq_state_t
*cbqp
, struct rm_class
*cl
)
86 /* delete the class */
87 rmc_delete_class(&cbqp
->ifnp
, cl
);
90 * free the class handle
92 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++)
93 if (cbqp
->cbq_class_tbl
[i
] == cl
)
94 cbqp
->cbq_class_tbl
[i
] = NULL
;
96 if (cl
== cbqp
->ifnp
.root_
)
97 cbqp
->ifnp
.root_
= NULL
;
98 if (cl
== cbqp
->ifnp
.default_
)
99 cbqp
->ifnp
.default_
= NULL
;
103 /* convert class handle to class pointer */
104 static struct rm_class
*
105 clh_to_clp(cbq_state_t
*cbqp
, uint32_t chandle
)
113 * first, try optimistically the slot matching the lower bits of
114 * the handle. if it fails, do the linear table search.
116 i
= chandle
% CBQ_MAX_CLASSES
;
117 if ((cl
= cbqp
->cbq_class_tbl
[i
]) != NULL
&&
118 cl
->stats_
.handle
== chandle
)
120 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++)
121 if ((cl
= cbqp
->cbq_class_tbl
[i
]) != NULL
&&
122 cl
->stats_
.handle
== chandle
)
128 cbq_clear_interface(cbq_state_t
*cbqp
)
133 /* clear out the classes now */
136 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++) {
137 if ((cl
= cbqp
->cbq_class_tbl
[i
]) != NULL
) {
138 if (is_a_parent_class(cl
))
141 cbq_class_destroy(cbqp
, cl
);
142 cbqp
->cbq_class_tbl
[i
] = NULL
;
143 if (cl
== cbqp
->ifnp
.root_
)
144 cbqp
->ifnp
.root_
= NULL
;
145 if (cl
== cbqp
->ifnp
.default_
)
146 cbqp
->ifnp
.default_
= NULL
;
156 cbq_request(struct ifaltq
*ifq
, int req
, void *arg
)
158 cbq_state_t
*cbqp
= (cbq_state_t
*)ifq
->altq_disc
;
170 /* copy the stats info in rm_class to class_states_t */
172 get_class_stats(class_stats_t
*statsp
, struct rm_class
*cl
)
174 statsp
->xmit_cnt
= cl
->stats_
.xmit_cnt
;
175 statsp
->drop_cnt
= cl
->stats_
.drop_cnt
;
176 statsp
->over
= cl
->stats_
.over
;
177 statsp
->borrows
= cl
->stats_
.borrows
;
178 statsp
->overactions
= cl
->stats_
.overactions
;
179 statsp
->delays
= cl
->stats_
.delays
;
181 statsp
->depth
= cl
->depth_
;
182 statsp
->priority
= cl
->pri_
;
183 statsp
->maxidle
= cl
->maxidle_
;
184 statsp
->minidle
= cl
->minidle_
;
185 statsp
->offtime
= cl
->offtime_
;
186 statsp
->qmax
= qlimit(cl
->q_
);
187 statsp
->ns_per_byte
= cl
->ns_per_byte_
;
188 statsp
->wrr_allot
= cl
->w_allotment_
;
189 statsp
->qcnt
= qlen(cl
->q_
);
190 statsp
->avgidle
= cl
->avgidle_
;
192 statsp
->qtype
= qtype(cl
->q_
);
194 if (q_is_red(cl
->q_
))
195 red_getstats(cl
->red_
, &statsp
->red
[0]);
198 if (q_is_rio(cl
->q_
))
199 rio_getstats((rio_t
*)cl
->red_
, &statsp
->red
[0]);
204 cbq_pfattach(struct pf_altq
*a
, struct ifaltq
*ifq
)
206 return altq_attach(ifq
, ALTQT_CBQ
, a
->altq_disc
,
207 cbq_enqueue
, cbq_dequeue
, cbq_request
, NULL
, NULL
);
211 cbq_add_altq(struct pf_altq
*a
)
216 if ((ifp
= ifunit(a
->ifname
)) == NULL
)
218 if (!ifq_is_ready(&ifp
->if_snd
))
221 /* allocate and initialize cbq_state_t */
222 cbqp
= kmalloc(sizeof(*cbqp
), M_ALTQ
, M_WAITOK
| M_ZERO
);
223 callout_init(&cbqp
->cbq_callout
);
225 cbqp
->ifnp
.ifq_
= &ifp
->if_snd
; /* keep the ifq */
226 ifq_purge(&ifp
->if_snd
);
228 /* keep the state in pf_altq */
235 cbq_remove_altq(struct pf_altq
*a
)
239 if ((cbqp
= a
->altq_disc
) == NULL
)
243 cbq_clear_interface(cbqp
);
245 if (cbqp
->ifnp
.default_
)
246 cbq_class_destroy(cbqp
, cbqp
->ifnp
.default_
);
247 if (cbqp
->ifnp
.root_
)
248 cbq_class_destroy(cbqp
, cbqp
->ifnp
.root_
);
250 /* deallocate cbq_state_t */
257 cbq_add_queue_locked(struct pf_altq
*a
, cbq_state_t
*cbqp
)
259 struct rm_class
*borrow
, *parent
;
261 struct cbq_opts
*opts
;
264 KKASSERT(a
->qid
!= 0);
267 * find a free slot in the class table. if the slot matching
268 * the lower bits of qid is free, use this slot. otherwise,
269 * use the first free slot.
271 i
= a
->qid
% CBQ_MAX_CLASSES
;
272 if (cbqp
->cbq_class_tbl
[i
] != NULL
) {
273 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++)
274 if (cbqp
->cbq_class_tbl
[i
] == NULL
)
276 if (i
== CBQ_MAX_CLASSES
)
280 opts
= &a
->pq_u
.cbq_opts
;
281 /* check parameters */
282 if (a
->priority
>= CBQ_MAXPRI
)
285 /* Get pointers to parent and borrow classes. */
286 parent
= clh_to_clp(cbqp
, a
->parent_qid
);
287 if (opts
->flags
& CBQCLF_BORROW
)
293 * A class must borrow from it's parent or it can not
294 * borrow at all. Hence, borrow can be null.
296 if (parent
== NULL
&& (opts
->flags
& CBQCLF_ROOTCLASS
) == 0) {
297 kprintf("cbq_add_queue: no parent class!\n");
301 if ((borrow
!= parent
) && (borrow
!= NULL
)) {
302 kprintf("cbq_add_class: borrow class != parent\n");
309 switch (opts
->flags
& CBQCLF_CLASSMASK
) {
310 case CBQCLF_ROOTCLASS
:
313 if (cbqp
->ifnp
.root_
)
316 case CBQCLF_DEFCLASS
:
317 if (cbqp
->ifnp
.default_
)
325 /* more than two flags bits set */
330 * create a class. if this is a root class, initialize the
333 if ((opts
->flags
& CBQCLF_CLASSMASK
) == CBQCLF_ROOTCLASS
) {
334 rmc_init(cbqp
->ifnp
.ifq_
, &cbqp
->ifnp
, opts
->ns_per_byte
,
335 cbqrestart
, a
->qlimit
, RM_MAXQUEUED
,
336 opts
->maxidle
, opts
->minidle
, opts
->offtime
,
338 cl
= cbqp
->ifnp
.root_
;
340 cl
= rmc_newclass(a
->priority
,
341 &cbqp
->ifnp
, opts
->ns_per_byte
,
342 rmc_delay_action
, a
->qlimit
, parent
, borrow
,
343 opts
->maxidle
, opts
->minidle
, opts
->offtime
,
344 opts
->pktsize
, opts
->flags
);
349 /* return handle to user space. */
350 cl
->stats_
.handle
= a
->qid
;
351 cl
->stats_
.depth
= cl
->depth_
;
353 /* save the allocated class */
354 cbqp
->cbq_class_tbl
[i
] = cl
;
356 if ((opts
->flags
& CBQCLF_CLASSMASK
) == CBQCLF_DEFCLASS
)
357 cbqp
->ifnp
.default_
= cl
;
363 cbq_add_queue(struct pf_altq
*a
)
372 /* XXX not MP safe */
373 if ((cbqp
= a
->altq_disc
) == NULL
)
375 ifq
= cbqp
->ifnp
.ifq_
;
378 error
= cbq_add_queue_locked(a
, cbqp
);
385 cbq_remove_queue_locked(struct pf_altq
*a
, cbq_state_t
*cbqp
)
390 if ((cl
= clh_to_clp(cbqp
, a
->qid
)) == NULL
)
393 /* if we are a parent class, then return an error. */
394 if (is_a_parent_class(cl
))
397 /* delete the class */
398 rmc_delete_class(&cbqp
->ifnp
, cl
);
401 * free the class handle
403 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++)
404 if (cbqp
->cbq_class_tbl
[i
] == cl
) {
405 cbqp
->cbq_class_tbl
[i
] = NULL
;
406 if (cl
== cbqp
->ifnp
.root_
)
407 cbqp
->ifnp
.root_
= NULL
;
408 if (cl
== cbqp
->ifnp
.default_
)
409 cbqp
->ifnp
.default_
= NULL
;
417 cbq_remove_queue(struct pf_altq
*a
)
423 /* XXX not MP safe */
424 if ((cbqp
= a
->altq_disc
) == NULL
)
426 ifq
= cbqp
->ifnp
.ifq_
;
429 error
= cbq_remove_queue_locked(a
, cbqp
);
436 cbq_getqstats(struct pf_altq
*a
, void *ubuf
, int *nbytes
)
444 if (*nbytes
< sizeof(stats
))
447 /* XXX not MP safe */
448 if ((cbqp
= altq_lookup(a
->ifname
, ALTQT_CBQ
)) == NULL
)
450 ifq
= cbqp
->ifnp
.ifq_
;
454 if ((cl
= clh_to_clp(cbqp
, a
->qid
)) == NULL
) {
459 get_class_stats(&stats
, cl
);
463 if ((error
= copyout((caddr_t
)&stats
, ubuf
, sizeof(stats
))) != 0)
465 *nbytes
= sizeof(stats
);
471 * cbq_enqueue(struct ifaltq *ifq, struct mbuf *m, struct altq_pktattr *pattr)
472 * - Queue data packets.
474 * cbq_enqueue is set to ifp->if_altqenqueue and called by an upper
475 * layer (e.g. ether_output). cbq_enqueue queues the given packet
476 * to the cbq, then invokes the driver's start routine.
478 * Returns: 0 if the queueing is successful.
479 * ENOBUFS if a packet dropping occurred as a result of
484 cbq_enqueue(struct ifaltq
*ifq
, struct mbuf
*m
, struct altq_pktattr
*pktattr
)
486 cbq_state_t
*cbqp
= (cbq_state_t
*)ifq
->altq_disc
;
490 /* grab class set by classifier */
491 if ((m
->m_flags
& M_PKTHDR
) == 0) {
492 /* should not happen */
493 if_printf(ifq
->altq_ifp
, "altq: packet does not have pkthdr\n");
497 if (m
->m_pkthdr
.fw_flags
& ALTQ_MBUF_TAGGED
)
498 cl
= clh_to_clp(cbqp
, m
->m_pkthdr
.altq_qid
);
502 cl
= cbqp
->ifnp
.default_
;
511 if (rmc_queue_packet(cl
, m
) != 0) {
512 /* drop occurred. some mbuf was freed in rmc_queue_packet. */
513 PKTCNTR_ADD(&cl
->stats_
.drop_cnt
, len
);
518 /* successfully queued. */
526 cbq_dequeue(struct ifaltq
*ifq
, struct mbuf
*mpolled
, int op
)
528 cbq_state_t
*cbqp
= (cbq_state_t
*)ifq
->altq_disc
;
532 m
= rmc_dequeue_next(&cbqp
->ifnp
, op
);
534 if (m
&& op
== ALTDQ_REMOVE
) {
535 --cbqp
->cbq_qlen
; /* decrement # of packets in cbq */
538 /* Update the class. */
539 rmc_update_class_util(&cbqp
->ifnp
);
542 KKASSERT(mpolled
== NULL
|| mpolled
== m
);
548 * cbqrestart(queue_t *) - Restart sending of data.
549 * called from rmc_restart in a critical section via timeout after waking up
555 cbqrestart(struct ifaltq
*ifq
)
559 ALTQ_ASSERT_LOCKED(ifq
);
561 if (!ifq_is_enabled(ifq
))
562 /* cbq must have been detached */
565 if ((cbqp
= (cbq_state_t
*)ifq
->altq_disc
) == NULL
)
566 /* should not happen */
569 if (cbqp
->cbq_qlen
> 0) {
570 struct ifnet
*ifp
= ifq
->altq_ifp
;
572 /* Release the altq lock to avoid deadlock */
575 ifnet_serialize_tx(ifp
);
576 if (ifp
->if_start
&& (ifp
->if_flags
& IFF_OACTIVE
) == 0)
577 (*ifp
->if_start
)(ifp
);
578 ifnet_deserialize_tx(ifp
);
585 cbq_purge(cbq_state_t
*cbqp
)
589 for (i
= 0; i
< CBQ_MAX_CLASSES
; i
++) {
590 if ((cl
= cbqp
->cbq_class_tbl
[i
]) != NULL
)
593 if (ifq_is_enabled(cbqp
->ifnp
.ifq_
))
594 cbqp
->ifnp
.ifq_
->ifq_len
= 0;
597 #endif /* ALTQ_CBQ */