UBIFS: Implement ->migratepage()
[linux-2.6/btrfs-unstable.git] / include / net / fq_impl.h
blob163f3ed0f05a741ed0f18933895f48a7c91bf754
1 /*
2 * Copyright (c) 2016 Qualcomm Atheros, Inc
4 * GPL v2
6 * Based on net/sched/sch_fq_codel.c
7 */
8 #ifndef __NET_SCHED_FQ_IMPL_H
9 #define __NET_SCHED_FQ_IMPL_H
11 #include <net/fq.h>
13 /* functions that are embedded into includer */
15 static struct sk_buff *fq_flow_dequeue(struct fq *fq,
16 struct fq_flow *flow)
18 struct fq_tin *tin = flow->tin;
19 struct fq_flow *i;
20 struct sk_buff *skb;
22 lockdep_assert_held(&fq->lock);
24 skb = __skb_dequeue(&flow->queue);
25 if (!skb)
26 return NULL;
28 tin->backlog_bytes -= skb->len;
29 tin->backlog_packets--;
30 flow->backlog -= skb->len;
31 fq->backlog--;
33 if (flow->backlog == 0) {
34 list_del_init(&flow->backlogchain);
35 } else {
36 i = flow;
38 list_for_each_entry_continue(i, &fq->backlogs, backlogchain)
39 if (i->backlog < flow->backlog)
40 break;
42 list_move_tail(&flow->backlogchain,
43 &i->backlogchain);
46 return skb;
49 static struct sk_buff *fq_tin_dequeue(struct fq *fq,
50 struct fq_tin *tin,
51 fq_tin_dequeue_t dequeue_func)
53 struct fq_flow *flow;
54 struct list_head *head;
55 struct sk_buff *skb;
57 lockdep_assert_held(&fq->lock);
59 begin:
60 head = &tin->new_flows;
61 if (list_empty(head)) {
62 head = &tin->old_flows;
63 if (list_empty(head))
64 return NULL;
67 flow = list_first_entry(head, struct fq_flow, flowchain);
69 if (flow->deficit <= 0) {
70 flow->deficit += fq->quantum;
71 list_move_tail(&flow->flowchain,
72 &tin->old_flows);
73 goto begin;
76 skb = dequeue_func(fq, tin, flow);
77 if (!skb) {
78 /* force a pass through old_flows to prevent starvation */
79 if ((head == &tin->new_flows) &&
80 !list_empty(&tin->old_flows)) {
81 list_move_tail(&flow->flowchain, &tin->old_flows);
82 } else {
83 list_del_init(&flow->flowchain);
84 flow->tin = NULL;
86 goto begin;
89 flow->deficit -= skb->len;
90 tin->tx_bytes += skb->len;
91 tin->tx_packets++;
93 return skb;
96 static struct fq_flow *fq_flow_classify(struct fq *fq,
97 struct fq_tin *tin,
98 struct sk_buff *skb,
99 fq_flow_get_default_t get_default_func)
101 struct fq_flow *flow;
102 u32 hash;
103 u32 idx;
105 lockdep_assert_held(&fq->lock);
107 hash = skb_get_hash_perturb(skb, fq->perturbation);
108 idx = reciprocal_scale(hash, fq->flows_cnt);
109 flow = &fq->flows[idx];
111 if (flow->tin && flow->tin != tin) {
112 flow = get_default_func(fq, tin, idx, skb);
113 tin->collisions++;
114 fq->collisions++;
117 if (!flow->tin)
118 tin->flows++;
120 return flow;
123 static void fq_recalc_backlog(struct fq *fq,
124 struct fq_tin *tin,
125 struct fq_flow *flow)
127 struct fq_flow *i;
129 if (list_empty(&flow->backlogchain))
130 list_add_tail(&flow->backlogchain, &fq->backlogs);
132 i = flow;
133 list_for_each_entry_continue_reverse(i, &fq->backlogs,
134 backlogchain)
135 if (i->backlog > flow->backlog)
136 break;
138 list_move(&flow->backlogchain, &i->backlogchain);
141 static void fq_tin_enqueue(struct fq *fq,
142 struct fq_tin *tin,
143 struct sk_buff *skb,
144 fq_skb_free_t free_func,
145 fq_flow_get_default_t get_default_func)
147 struct fq_flow *flow;
149 lockdep_assert_held(&fq->lock);
151 flow = fq_flow_classify(fq, tin, skb, get_default_func);
153 flow->tin = tin;
154 flow->backlog += skb->len;
155 tin->backlog_bytes += skb->len;
156 tin->backlog_packets++;
157 fq->backlog++;
159 fq_recalc_backlog(fq, tin, flow);
161 if (list_empty(&flow->flowchain)) {
162 flow->deficit = fq->quantum;
163 list_add_tail(&flow->flowchain,
164 &tin->new_flows);
167 __skb_queue_tail(&flow->queue, skb);
169 if (fq->backlog > fq->limit) {
170 flow = list_first_entry_or_null(&fq->backlogs,
171 struct fq_flow,
172 backlogchain);
173 if (!flow)
174 return;
176 skb = fq_flow_dequeue(fq, flow);
177 if (!skb)
178 return;
180 free_func(fq, flow->tin, flow, skb);
182 flow->tin->overlimit++;
183 fq->overlimit++;
187 static void fq_flow_reset(struct fq *fq,
188 struct fq_flow *flow,
189 fq_skb_free_t free_func)
191 struct sk_buff *skb;
193 while ((skb = fq_flow_dequeue(fq, flow)))
194 free_func(fq, flow->tin, flow, skb);
196 if (!list_empty(&flow->flowchain))
197 list_del_init(&flow->flowchain);
199 if (!list_empty(&flow->backlogchain))
200 list_del_init(&flow->backlogchain);
202 flow->tin = NULL;
204 WARN_ON_ONCE(flow->backlog);
207 static void fq_tin_reset(struct fq *fq,
208 struct fq_tin *tin,
209 fq_skb_free_t free_func)
211 struct list_head *head;
212 struct fq_flow *flow;
214 for (;;) {
215 head = &tin->new_flows;
216 if (list_empty(head)) {
217 head = &tin->old_flows;
218 if (list_empty(head))
219 break;
222 flow = list_first_entry(head, struct fq_flow, flowchain);
223 fq_flow_reset(fq, flow, free_func);
226 WARN_ON_ONCE(tin->backlog_bytes);
227 WARN_ON_ONCE(tin->backlog_packets);
230 static void fq_flow_init(struct fq_flow *flow)
232 INIT_LIST_HEAD(&flow->flowchain);
233 INIT_LIST_HEAD(&flow->backlogchain);
234 __skb_queue_head_init(&flow->queue);
237 static void fq_tin_init(struct fq_tin *tin)
239 INIT_LIST_HEAD(&tin->new_flows);
240 INIT_LIST_HEAD(&tin->old_flows);
243 static int fq_init(struct fq *fq, int flows_cnt)
245 int i;
247 memset(fq, 0, sizeof(fq[0]));
248 INIT_LIST_HEAD(&fq->backlogs);
249 spin_lock_init(&fq->lock);
250 fq->flows_cnt = max_t(u32, flows_cnt, 1);
251 fq->perturbation = prandom_u32();
252 fq->quantum = 300;
253 fq->limit = 8192;
255 fq->flows = kcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL);
256 if (!fq->flows)
257 return -ENOMEM;
259 for (i = 0; i < fq->flows_cnt; i++)
260 fq_flow_init(&fq->flows[i]);
262 return 0;
265 static void fq_reset(struct fq *fq,
266 fq_skb_free_t free_func)
268 int i;
270 for (i = 0; i < fq->flows_cnt; i++)
271 fq_flow_reset(fq, &fq->flows[i], free_func);
273 kfree(fq->flows);
274 fq->flows = NULL;
277 #endif