2 * Copyright (c) 2003-2008 Fabrice Bellard
3 * Copyright (c) 2009 Red Hat, Inc.
5 * Permission is hereby granted, free of charge, to any person obtaining a copy
6 * of this software and associated documentation files (the "Software"), to deal
7 * in the Software without restriction, including without limitation the rights
8 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9 * copies of the Software, and to permit persons to whom the Software is
10 * furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 #include "net/queue.h"
25 #include "qemu-queue.h"
27 /* The delivery handler may only return zero if it will call
28 * qemu_net_queue_flush() when it determines that it is once again able
29 * to deliver packets. It must also call qemu_net_queue_purge() in its
32 * If a sent callback is provided to send(), the caller must handle a
33 * zero return from the delivery handler by not sending any more packets
34 * until we have invoked the callback. Only in that case will we queue
37 * If a sent callback isn't provided, we just drop the packet to avoid
42 QTAILQ_ENTRY(NetPacket
) entry
;
43 VLANClientState
*sender
;
46 NetPacketSent
*sent_cb
;
51 NetPacketDeliver
*deliver
;
52 NetPacketDeliverIOV
*deliver_iov
;
55 QTAILQ_HEAD(packets
, NetPacket
) packets
;
57 unsigned delivering
: 1;
60 NetQueue
*qemu_new_net_queue(NetPacketDeliver
*deliver
,
61 NetPacketDeliverIOV
*deliver_iov
,
66 queue
= qemu_mallocz(sizeof(NetQueue
));
68 queue
->deliver
= deliver
;
69 queue
->deliver_iov
= deliver_iov
;
70 queue
->opaque
= opaque
;
72 QTAILQ_INIT(&queue
->packets
);
74 queue
->delivering
= 0;
79 void qemu_del_net_queue(NetQueue
*queue
)
81 NetPacket
*packet
, *next
;
83 QTAILQ_FOREACH_SAFE(packet
, &queue
->packets
, entry
, next
) {
84 QTAILQ_REMOVE(&queue
->packets
, packet
, entry
);
91 static ssize_t
qemu_net_queue_append(NetQueue
*queue
,
92 VLANClientState
*sender
,
96 NetPacketSent
*sent_cb
)
100 packet
= qemu_malloc(sizeof(NetPacket
) + size
);
101 packet
->sender
= sender
;
102 packet
->flags
= flags
;
104 packet
->sent_cb
= sent_cb
;
105 memcpy(packet
->data
, buf
, size
);
107 QTAILQ_INSERT_TAIL(&queue
->packets
, packet
, entry
);
112 static ssize_t
qemu_net_queue_append_iov(NetQueue
*queue
,
113 VLANClientState
*sender
,
115 const struct iovec
*iov
,
117 NetPacketSent
*sent_cb
)
123 for (i
= 0; i
< iovcnt
; i
++) {
124 max_len
+= iov
[i
].iov_len
;
127 packet
= qemu_malloc(sizeof(NetPacket
) + max_len
);
128 packet
->sender
= sender
;
129 packet
->sent_cb
= sent_cb
;
130 packet
->flags
= flags
;
133 for (i
= 0; i
< iovcnt
; i
++) {
134 size_t len
= iov
[i
].iov_len
;
136 memcpy(packet
->data
+ packet
->size
, iov
[i
].iov_base
, len
);
140 QTAILQ_INSERT_TAIL(&queue
->packets
, packet
, entry
);
145 static ssize_t
qemu_net_queue_deliver(NetQueue
*queue
,
146 VLANClientState
*sender
,
153 queue
->delivering
= 1;
154 ret
= queue
->deliver(sender
, flags
, data
, size
, queue
->opaque
);
155 queue
->delivering
= 0;
160 static ssize_t
qemu_net_queue_deliver_iov(NetQueue
*queue
,
161 VLANClientState
*sender
,
163 const struct iovec
*iov
,
168 queue
->delivering
= 1;
169 ret
= queue
->deliver_iov(sender
, flags
, iov
, iovcnt
, queue
->opaque
);
170 queue
->delivering
= 0;
175 ssize_t
qemu_net_queue_send(NetQueue
*queue
,
176 VLANClientState
*sender
,
180 NetPacketSent
*sent_cb
)
184 if (queue
->delivering
) {
185 return qemu_net_queue_append(queue
, sender
, flags
, data
, size
, NULL
);
188 ret
= qemu_net_queue_deliver(queue
, sender
, flags
, data
, size
);
189 if (ret
== 0 && sent_cb
!= NULL
) {
190 qemu_net_queue_append(queue
, sender
, flags
, data
, size
, sent_cb
);
194 qemu_net_queue_flush(queue
);
199 ssize_t
qemu_net_queue_send_iov(NetQueue
*queue
,
200 VLANClientState
*sender
,
202 const struct iovec
*iov
,
204 NetPacketSent
*sent_cb
)
208 if (queue
->delivering
) {
209 return qemu_net_queue_append_iov(queue
, sender
, flags
, iov
, iovcnt
, NULL
);
212 ret
= qemu_net_queue_deliver_iov(queue
, sender
, flags
, iov
, iovcnt
);
213 if (ret
== 0 && sent_cb
!= NULL
) {
214 qemu_net_queue_append_iov(queue
, sender
, flags
, iov
, iovcnt
, sent_cb
);
218 qemu_net_queue_flush(queue
);
223 void qemu_net_queue_purge(NetQueue
*queue
, VLANClientState
*from
)
225 NetPacket
*packet
, *next
;
227 QTAILQ_FOREACH_SAFE(packet
, &queue
->packets
, entry
, next
) {
228 if (packet
->sender
== from
) {
229 QTAILQ_REMOVE(&queue
->packets
, packet
, entry
);
235 void qemu_net_queue_flush(NetQueue
*queue
)
237 while (!QTAILQ_EMPTY(&queue
->packets
)) {
241 packet
= QTAILQ_FIRST(&queue
->packets
);
242 QTAILQ_REMOVE(&queue
->packets
, packet
, entry
);
244 ret
= qemu_net_queue_deliver(queue
,
249 if (ret
== 0 && packet
->sent_cb
!= NULL
) {
250 QTAILQ_INSERT_HEAD(&queue
->packets
, packet
, entry
);
254 if (packet
->sent_cb
) {
255 packet
->sent_cb(packet
->sender
, ret
);