2 * netmap access for qemu
4 * Copyright (c) 2012-2013 Luigi Rizzo
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
26 #include "qemu/osdep.h"
27 #include <sys/ioctl.h>
29 #define NETMAP_WITH_LIBS
30 #include <net/netmap.h>
31 #include <net/netmap_user.h>
36 #include "qemu/error-report.h"
37 #include "qapi/error.h"
39 #include "qemu/cutils.h"
40 #include "qemu/main-loop.h"
42 typedef struct NetmapState
{
45 char ifname
[IFNAMSIZ
];
46 struct netmap_ring
*tx
;
47 struct netmap_ring
*rx
;
50 struct iovec iov
[IOV_MAX
];
51 int vnet_hdr_len
; /* Current virtio-net header length. */
55 #define pkt_copy bcopy
57 /* A fast copy routine only for multiples of 64 bytes, non overlapped. */
59 pkt_copy(const void *_src
, void *_dst
, int l
)
61 const uint64_t *src
= _src
;
63 if (unlikely(l
>= 1024)) {
67 for (; l
> 0; l
-= 64) {
78 #endif /* __FreeBSD__ */
81 * Open a netmap device. We assume there is only one queue
82 * (which is the case for the VALE bridge).
84 static struct nm_desc
*netmap_open(const NetdevNetmapOptions
*nm_opts
,
90 memset(&req
, 0, sizeof(req
));
92 nmd
= nm_open(nm_opts
->ifname
, &req
, NETMAP_NO_TX_POLL
,
95 error_setg_errno(errp
, errno
, "Failed to nm_open() %s",
103 static void netmap_send(void *opaque
);
104 static void netmap_writable(void *opaque
);
106 /* Set the event-loop handlers for the netmap backend. */
107 static void netmap_update_fd_handler(NetmapState
*s
)
109 qemu_set_fd_handler(s
->nmd
->fd
,
110 s
->read_poll
? netmap_send
: NULL
,
111 s
->write_poll
? netmap_writable
: NULL
,
115 /* Update the read handler. */
116 static void netmap_read_poll(NetmapState
*s
, bool enable
)
118 if (s
->read_poll
!= enable
) { /* Do nothing if not changed. */
119 s
->read_poll
= enable
;
120 netmap_update_fd_handler(s
);
124 /* Update the write handler. */
125 static void netmap_write_poll(NetmapState
*s
, bool enable
)
127 if (s
->write_poll
!= enable
) {
128 s
->write_poll
= enable
;
129 netmap_update_fd_handler(s
);
133 static void netmap_poll(NetClientState
*nc
, bool enable
)
135 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
137 if (s
->read_poll
!= enable
|| s
->write_poll
!= enable
) {
138 s
->write_poll
= enable
;
139 s
->read_poll
= enable
;
140 netmap_update_fd_handler(s
);
145 * The fd_write() callback, invoked if the fd is marked as
146 * writable after a poll. Unregister the handler and flush any
149 static void netmap_writable(void *opaque
)
151 NetmapState
*s
= opaque
;
153 netmap_write_poll(s
, false);
154 qemu_flush_queued_packets(&s
->nc
);
157 static ssize_t
netmap_receive_iov(NetClientState
*nc
,
158 const struct iovec
*iov
, int iovcnt
)
160 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
161 struct netmap_ring
*ring
= s
->tx
;
162 unsigned int tail
= ring
->tail
;
170 last
= i
= ring
->head
;
172 if (nm_ring_space(ring
) < iovcnt
) {
173 /* Not enough netmap slots. Tell the kernel that we have seen the new
174 * available slots (so that it notifies us again when it has more
175 * ones), but without publishing any new slots to be processed
176 * (e.g., we don't advance ring->head). */
178 netmap_write_poll(s
, true);
182 for (j
= 0; j
< iovcnt
; j
++) {
183 int iov_frag_size
= iov
[j
].iov_len
;
187 totlen
+= iov_frag_size
;
189 /* Split each iovec fragment over more netmap slots, if
191 while (iov_frag_size
) {
192 nm_frag_size
= MIN(iov_frag_size
, ring
->nr_buf_size
);
194 if (unlikely(i
== tail
)) {
195 /* We ran out of netmap slots while splitting the
198 netmap_write_poll(s
, true);
202 idx
= ring
->slot
[i
].buf_idx
;
203 dst
= (uint8_t *)NETMAP_BUF(ring
, idx
);
205 ring
->slot
[i
].len
= nm_frag_size
;
206 ring
->slot
[i
].flags
= NS_MOREFRAG
;
207 pkt_copy(iov
[j
].iov_base
+ offset
, dst
, nm_frag_size
);
210 i
= nm_ring_next(ring
, i
);
212 offset
+= nm_frag_size
;
213 iov_frag_size
-= nm_frag_size
;
216 /* The last slot must not have NS_MOREFRAG set. */
217 ring
->slot
[last
].flags
&= ~NS_MOREFRAG
;
219 /* Now update ring->head and ring->cur to publish the new slots and
220 * the new wakeup point. */
221 ring
->head
= ring
->cur
= i
;
223 ioctl(s
->nmd
->fd
, NIOCTXSYNC
, NULL
);
228 static ssize_t
netmap_receive(NetClientState
*nc
,
229 const uint8_t *buf
, size_t size
)
233 iov
.iov_base
= (void *)buf
;
236 return netmap_receive_iov(nc
, &iov
, 1);
239 /* Complete a previous send (backend --> guest) and enable the
241 static void netmap_send_completed(NetClientState
*nc
, ssize_t len
)
243 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
245 netmap_read_poll(s
, true);
248 static void netmap_send(void *opaque
)
250 NetmapState
*s
= opaque
;
251 struct netmap_ring
*ring
= s
->rx
;
252 unsigned int tail
= ring
->tail
;
254 /* Keep sending while there are available slots in the netmap
255 RX ring and the forwarding path towards the peer is open. */
256 while (ring
->head
!= tail
) {
257 uint32_t i
= ring
->head
;
263 /* Get a (possibly multi-slot) packet. */
265 idx
= ring
->slot
[i
].buf_idx
;
266 morefrag
= (ring
->slot
[i
].flags
& NS_MOREFRAG
);
267 s
->iov
[iovcnt
].iov_base
= (void *)NETMAP_BUF(ring
, idx
);
268 s
->iov
[iovcnt
].iov_len
= ring
->slot
[i
].len
;
270 i
= nm_ring_next(ring
, i
);
271 } while (i
!= tail
&& morefrag
);
273 /* Advance ring->cur to tell the kernel that we have seen the slots. */
276 if (unlikely(morefrag
)) {
277 /* This is a truncated packet, so we can stop without releasing the
278 * incomplete slots by updating ring->head. We will hopefully
279 * re-read the complete packet the next time we are called. */
283 iovsize
= qemu_sendv_packet_async(&s
->nc
, s
->iov
, iovcnt
,
284 netmap_send_completed
);
286 /* Release the slots to the kernel. */
290 /* The peer does not receive anymore. Packet is queued, stop
291 * reading from the backend until netmap_send_completed(). */
292 netmap_read_poll(s
, false);
298 /* Flush and close. */
299 static void netmap_cleanup(NetClientState
*nc
)
301 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
303 qemu_purge_queued_packets(nc
);
305 netmap_poll(nc
, false);
310 /* Offloading manipulation support callbacks. */
311 static int netmap_fd_set_vnet_hdr_len(NetmapState
*s
, int len
)
315 /* Issue a NETMAP_BDG_VNET_HDR command to change the virtio-net header
316 * length for the netmap adapter associated to 's->ifname'.
318 memset(&req
, 0, sizeof(req
));
319 pstrcpy(req
.nr_name
, sizeof(req
.nr_name
), s
->ifname
);
320 req
.nr_version
= NETMAP_API
;
321 req
.nr_cmd
= NETMAP_BDG_VNET_HDR
;
324 return ioctl(s
->nmd
->fd
, NIOCREGIF
, &req
);
327 static bool netmap_has_vnet_hdr_len(NetClientState
*nc
, int len
)
329 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
330 int prev_len
= s
->vnet_hdr_len
;
332 /* Check that we can set the new length. */
333 if (netmap_fd_set_vnet_hdr_len(s
, len
)) {
337 /* Restore the previous length. */
338 if (netmap_fd_set_vnet_hdr_len(s
, prev_len
)) {
339 error_report("Failed to restore vnet-hdr length %d on %s: %s",
340 prev_len
, s
->ifname
, strerror(errno
));
347 /* A netmap interface that supports virtio-net headers always
348 * supports UFO, so we use this callback also for the has_ufo hook. */
349 static bool netmap_has_vnet_hdr(NetClientState
*nc
)
351 return netmap_has_vnet_hdr_len(nc
, sizeof(struct virtio_net_hdr
));
354 static void netmap_set_vnet_hdr_len(NetClientState
*nc
, int len
)
356 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
359 err
= netmap_fd_set_vnet_hdr_len(s
, len
);
361 error_report("Unable to set vnet-hdr length %d on %s: %s",
362 len
, s
->ifname
, strerror(errno
));
364 /* Keep track of the current length. */
365 s
->vnet_hdr_len
= len
;
369 static void netmap_set_offload(NetClientState
*nc
, int csum
, int tso4
, int tso6
,
370 int ecn
, int ufo
, int uso4
, int uso6
)
372 NetmapState
*s
= DO_UPCAST(NetmapState
, nc
, nc
);
374 /* Setting a virtio-net header length greater than zero automatically
375 * enables the offloadings. */
376 if (!s
->vnet_hdr_len
) {
377 netmap_set_vnet_hdr_len(nc
, sizeof(struct virtio_net_hdr
));
381 /* NetClientInfo methods */
382 static NetClientInfo net_netmap_info
= {
383 .type
= NET_CLIENT_DRIVER_NETMAP
,
384 .size
= sizeof(NetmapState
),
385 .receive
= netmap_receive
,
386 .receive_iov
= netmap_receive_iov
,
388 .cleanup
= netmap_cleanup
,
389 .has_ufo
= netmap_has_vnet_hdr
,
390 .has_vnet_hdr
= netmap_has_vnet_hdr
,
391 .has_vnet_hdr_len
= netmap_has_vnet_hdr_len
,
392 .set_offload
= netmap_set_offload
,
393 .set_vnet_hdr_len
= netmap_set_vnet_hdr_len
,
396 /* The exported init function
398 * ... -net netmap,ifname="..."
400 int net_init_netmap(const Netdev
*netdev
,
401 const char *name
, NetClientState
*peer
, Error
**errp
)
403 const NetdevNetmapOptions
*netmap_opts
= &netdev
->u
.netmap
;
409 nmd
= netmap_open(netmap_opts
, &err
);
411 error_propagate(errp
, err
);
414 /* Create the object. */
415 nc
= qemu_new_net_client(&net_netmap_info
, peer
, "netmap", name
);
416 s
= DO_UPCAST(NetmapState
, nc
, nc
);
418 s
->tx
= NETMAP_TXRING(nmd
->nifp
, 0);
419 s
->rx
= NETMAP_RXRING(nmd
->nifp
, 0);
421 pstrcpy(s
->ifname
, sizeof(s
->ifname
), netmap_opts
->ifname
);
422 netmap_read_poll(s
, true); /* Initially only poll for reads. */