2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
4 * PAPR Inter-VM Logical Lan, aka ibmveth
6 * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
28 #include "qemu/osdep.h"
32 #include "qemu/module.h"
35 #include "hw/ppc/spapr.h"
36 #include "hw/ppc/spapr_vio.h"
37 #include "sysemu/sysemu.h"
43 #define MAX_PACKET_SIZE 65536
45 /* Compatibility flags for migration */
46 #define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT 0
47 #define SPAPRVLAN_FLAG_RX_BUF_POOLS (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
53 typedef uint64_t vlan_bd_t
;
55 #define VLAN_BD_VALID 0x8000000000000000ULL
56 #define VLAN_BD_TOGGLE 0x4000000000000000ULL
57 #define VLAN_BD_NO_CSUM 0x0200000000000000ULL
58 #define VLAN_BD_CSUM_GOOD 0x0100000000000000ULL
59 #define VLAN_BD_LEN_MASK 0x00ffffff00000000ULL
60 #define VLAN_BD_LEN(bd) (((bd) & VLAN_BD_LEN_MASK) >> 32)
61 #define VLAN_BD_ADDR_MASK 0x00000000ffffffffULL
62 #define VLAN_BD_ADDR(bd) ((bd) & VLAN_BD_ADDR_MASK)
64 #define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
65 (((len) << 32) & VLAN_BD_LEN_MASK) | \
66 (addr & VLAN_BD_ADDR_MASK))
68 #define VLAN_RXQC_TOGGLE 0x80
69 #define VLAN_RXQC_VALID 0x40
70 #define VLAN_RXQC_NO_CSUM 0x02
71 #define VLAN_RXQC_CSUM_GOOD 0x01
73 #define VLAN_RQ_ALIGNMENT 16
74 #define VLAN_RXQ_BD_OFF 0
75 #define VLAN_FILTER_BD_OFF 8
76 #define VLAN_RX_BDS_OFF 16
78 * The final 8 bytes of the buffer list is a counter of frames dropped
79 * because there was not a buffer in the buffer list capable of holding
80 * the frame. We must avoid it, or the operating system will report garbage
83 #define VLAN_RX_BDS_LEN (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
84 #define VLAN_MAX_BUFS (VLAN_RX_BDS_LEN / 8)
86 #define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
87 #define VIO_SPAPR_VLAN_DEVICE(obj) \
88 OBJECT_CHECK(SpaprVioVlan, (obj), TYPE_VIO_SPAPR_VLAN_DEVICE)
90 #define RX_POOL_MAX_BDS 4096
91 #define RX_MAX_POOLS 5
96 vlan_bd_t bds
[RX_POOL_MAX_BDS
];
99 typedef struct SpaprVioVlan
{
106 uint32_t add_buf_ptr
, use_buf_ptr
, rx_bufs
;
108 QEMUTimer
*rxp_timer
;
109 uint32_t compat_flags
; /* Compatibility flags for migration */
110 RxBufPool
*rx_pool
[RX_MAX_POOLS
]; /* Receive buffer descriptor pools */
113 static int spapr_vlan_can_receive(NetClientState
*nc
)
115 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
117 return (dev
->isopen
&& dev
->rx_bufs
> 0);
121 * The last 8 bytes of the receive buffer list page (that has been
122 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
123 * a counter for frames that have been dropped because there was no
124 * suitable receive buffer available. This function is used to increase
125 * this counter by one.
127 static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan
*dev
)
131 cnt
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8);
132 vio_stq(&dev
->sdev
, dev
->buf_list
+ 4096 - 8, cnt
+ 1);
136 * Get buffer descriptor from one of our receive buffer pools
138 static vlan_bd_t
spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan
*dev
,
144 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
145 if (dev
->rx_pool
[pool
]->count
> 0 &&
146 dev
->rx_pool
[pool
]->bufsize
>= size
+ 8) {
150 if (pool
== RX_MAX_POOLS
) {
151 /* Failed to find a suitable buffer */
156 trace_spapr_vlan_get_rx_bd_from_pool_found(pool
,
157 dev
->rx_pool
[pool
]->count
,
160 /* Remove the buffer from the pool */
161 dev
->rx_pool
[pool
]->count
--;
162 bd
= dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
];
163 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
] = 0;
169 * Get buffer descriptor from the receive buffer list page that has been
170 * supplied by the guest with the H_REGISTER_LOGICAL_LAN call
172 static vlan_bd_t
spapr_vlan_get_rx_bd_from_page(SpaprVioVlan
*dev
,
175 int buf_ptr
= dev
->use_buf_ptr
;
180 if (buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
181 buf_ptr
= VLAN_RX_BDS_OFF
;
184 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ buf_ptr
);
186 trace_spapr_vlan_get_rx_bd_from_page(buf_ptr
, (uint64_t)bd
);
187 } while ((!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8)
188 && buf_ptr
!= dev
->use_buf_ptr
);
190 if (!(bd
& VLAN_BD_VALID
) || VLAN_BD_LEN(bd
) < size
+ 8) {
191 /* Failed to find a suitable buffer */
195 /* Remove the buffer from the pool */
196 dev
->use_buf_ptr
= buf_ptr
;
197 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->use_buf_ptr
, 0);
199 trace_spapr_vlan_get_rx_bd_from_page_found(dev
->use_buf_ptr
, dev
->rx_bufs
);
204 static ssize_t
spapr_vlan_receive(NetClientState
*nc
, const uint8_t *buf
,
207 SpaprVioVlan
*dev
= qemu_get_nic_opaque(nc
);
208 SpaprVioDevice
*sdev
= VIO_SPAPR_DEVICE(dev
);
209 vlan_bd_t rxq_bd
= vio_ldq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
);
214 trace_spapr_vlan_receive(sdev
->qdev
.id
, dev
->rx_bufs
);
221 spapr_vlan_record_dropped_rx_frame(dev
);
225 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
226 bd
= spapr_vlan_get_rx_bd_from_pool(dev
, size
);
228 bd
= spapr_vlan_get_rx_bd_from_page(dev
, size
);
231 spapr_vlan_record_dropped_rx_frame(dev
);
237 /* Transfer the packet data */
238 if (spapr_vio_dma_write(sdev
, VLAN_BD_ADDR(bd
) + 8, buf
, size
) < 0) {
242 trace_spapr_vlan_receive_dma_completed();
244 /* Update the receive queue */
245 control
= VLAN_RXQC_TOGGLE
| VLAN_RXQC_VALID
;
246 if (rxq_bd
& VLAN_BD_TOGGLE
) {
247 control
^= VLAN_RXQC_TOGGLE
;
250 handle
= vio_ldq(sdev
, VLAN_BD_ADDR(bd
));
251 vio_stq(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 8, handle
);
252 vio_stl(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 4, size
);
253 vio_sth(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
+ 2, 8);
254 vio_stb(sdev
, VLAN_BD_ADDR(rxq_bd
) + dev
->rxq_ptr
, control
);
256 trace_spapr_vlan_receive_wrote(dev
->rxq_ptr
,
257 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
259 vio_ldq(sdev
, VLAN_BD_ADDR(rxq_bd
) +
263 if (dev
->rxq_ptr
>= VLAN_BD_LEN(rxq_bd
)) {
265 vio_stq(sdev
, dev
->buf_list
+ VLAN_RXQ_BD_OFF
, rxq_bd
^ VLAN_BD_TOGGLE
);
268 if (sdev
->signal_state
& 1) {
269 qemu_irq_pulse(spapr_vio_qirq(sdev
));
275 static NetClientInfo net_spapr_vlan_info
= {
276 .type
= NET_CLIENT_DRIVER_NIC
,
277 .size
= sizeof(NICState
),
278 .can_receive
= spapr_vlan_can_receive
,
279 .receive
= spapr_vlan_receive
,
282 static void spapr_vlan_flush_rx_queue(void *opaque
)
284 SpaprVioVlan
*dev
= opaque
;
286 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
289 static void spapr_vlan_reset_rx_pool(RxBufPool
*rxp
)
292 * Use INT_MAX as bufsize so that unused buffers are moved to the end
293 * of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
295 rxp
->bufsize
= INT_MAX
;
297 memset(rxp
->bds
, 0, sizeof(rxp
->bds
));
300 static void spapr_vlan_reset(SpaprVioDevice
*sdev
)
302 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
309 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
310 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
311 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
315 memcpy(&dev
->nicconf
.macaddr
.a
, &dev
->perm_mac
.a
,
316 sizeof(dev
->nicconf
.macaddr
.a
));
317 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
320 static void spapr_vlan_realize(SpaprVioDevice
*sdev
, Error
**errp
)
322 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
324 qemu_macaddr_default_if_unset(&dev
->nicconf
.macaddr
);
326 memcpy(&dev
->perm_mac
.a
, &dev
->nicconf
.macaddr
.a
, sizeof(dev
->perm_mac
.a
));
328 dev
->nic
= qemu_new_nic(&net_spapr_vlan_info
, &dev
->nicconf
,
329 object_get_typename(OBJECT(sdev
)), sdev
->qdev
.id
, dev
);
330 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
332 dev
->rxp_timer
= timer_new_us(QEMU_CLOCK_VIRTUAL
, spapr_vlan_flush_rx_queue
,
336 static void spapr_vlan_instance_init(Object
*obj
)
338 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
341 device_add_bootindex_property(obj
, &dev
->nicconf
.bootindex
,
345 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
346 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
347 dev
->rx_pool
[i
] = g_new(RxBufPool
, 1);
348 spapr_vlan_reset_rx_pool(dev
->rx_pool
[i
]);
353 static void spapr_vlan_instance_finalize(Object
*obj
)
355 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(obj
);
358 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
359 for (i
= 0; i
< RX_MAX_POOLS
; i
++) {
360 g_free(dev
->rx_pool
[i
]);
361 dev
->rx_pool
[i
] = NULL
;
365 if (dev
->rxp_timer
) {
366 timer_del(dev
->rxp_timer
);
367 timer_free(dev
->rxp_timer
);
371 void spapr_vlan_create(SpaprVioBus
*bus
, NICInfo
*nd
)
375 dev
= qdev_create(&bus
->bus
, "spapr-vlan");
377 qdev_set_nic_properties(dev
, nd
);
379 qdev_init_nofail(dev
);
382 static int spapr_vlan_devnode(SpaprVioDevice
*dev
, void *fdt
, int node_off
)
384 SpaprVioVlan
*vdev
= VIO_SPAPR_VLAN_DEVICE(dev
);
385 uint8_t padded_mac
[8] = {0, 0};
388 /* Some old phyp versions give the mac address in an 8-byte
389 * property. The kernel driver (before 3.10) has an insane workaround;
390 * rather than doing the obvious thing and checking the property
391 * length, it checks whether the first byte has 0b10 in the low
392 * bits. If a correct 6-byte property has a different first byte
393 * the kernel will get the wrong mac address, overrunning its
394 * buffer in the process (read only, thank goodness).
396 * Here we return a 6-byte address unless that would break a pre-3.10
397 * driver. In that case we return a padded 8-byte address to allow the old
398 * workaround to succeed. */
399 if ((vdev
->nicconf
.macaddr
.a
[0] & 0x3) == 0x2) {
400 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
401 &vdev
->nicconf
.macaddr
, ETH_ALEN
);
403 memcpy(&padded_mac
[2], &vdev
->nicconf
.macaddr
, ETH_ALEN
);
404 ret
= fdt_setprop(fdt
, node_off
, "local-mac-address",
405 padded_mac
, sizeof(padded_mac
));
411 ret
= fdt_setprop_cell(fdt
, node_off
, "ibm,mac-address-filters", 0);
419 static int check_bd(SpaprVioVlan
*dev
, vlan_bd_t bd
,
420 target_ulong alignment
)
422 if ((VLAN_BD_ADDR(bd
) % alignment
)
423 || (VLAN_BD_LEN(bd
) % alignment
)) {
427 if (!spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
428 VLAN_BD_LEN(bd
), DMA_DIRECTION_FROM_DEVICE
)
429 || !spapr_vio_dma_valid(&dev
->sdev
, VLAN_BD_ADDR(bd
),
430 VLAN_BD_LEN(bd
), DMA_DIRECTION_TO_DEVICE
)) {
437 static target_ulong
h_register_logical_lan(PowerPCCPU
*cpu
,
438 SpaprMachineState
*spapr
,
442 target_ulong reg
= args
[0];
443 target_ulong buf_list
= args
[1];
444 target_ulong rec_queue
= args
[2];
445 target_ulong filter_list
= args
[3];
446 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
447 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
448 vlan_bd_t filter_list_bd
;
455 hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
456 "H_FREE_LOGICAL_LAN\n");
460 if (check_bd(dev
, VLAN_VALID_BD(buf_list
, SPAPR_TCE_PAGE_SIZE
),
461 SPAPR_TCE_PAGE_SIZE
) < 0) {
462 hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx
"\n", buf_list
);
466 filter_list_bd
= VLAN_VALID_BD(filter_list
, SPAPR_TCE_PAGE_SIZE
);
467 if (check_bd(dev
, filter_list_bd
, SPAPR_TCE_PAGE_SIZE
) < 0) {
468 hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx
"\n", filter_list
);
472 if (!(rec_queue
& VLAN_BD_VALID
)
473 || (check_bd(dev
, rec_queue
, VLAN_RQ_ALIGNMENT
) < 0)) {
474 hcall_dprintf("Bad receive queue\n");
478 dev
->buf_list
= buf_list
;
479 sdev
->signal_state
= 0;
481 rec_queue
&= ~VLAN_BD_TOGGLE
;
483 /* Initialize the buffer list */
484 vio_stq(sdev
, buf_list
, rec_queue
);
485 vio_stq(sdev
, buf_list
+ 8, filter_list_bd
);
486 spapr_vio_dma_set(sdev
, buf_list
+ VLAN_RX_BDS_OFF
, 0,
487 SPAPR_TCE_PAGE_SIZE
- VLAN_RX_BDS_OFF
);
488 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
489 dev
->use_buf_ptr
= VLAN_RX_BDS_OFF
- 8;
493 /* Initialize the receive queue */
494 spapr_vio_dma_set(sdev
, VLAN_BD_ADDR(rec_queue
), 0, VLAN_BD_LEN(rec_queue
));
497 qemu_flush_queued_packets(qemu_get_queue(dev
->nic
));
503 static target_ulong
h_free_logical_lan(PowerPCCPU
*cpu
,
504 SpaprMachineState
*spapr
,
505 target_ulong opcode
, target_ulong
*args
)
507 target_ulong reg
= args
[0];
508 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
509 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
516 hcall_dprintf("H_FREE_LOGICAL_LAN called without "
517 "H_REGISTER_LOGICAL_LAN\n");
521 spapr_vlan_reset(sdev
);
526 * Used for qsort, this function compares two RxBufPools by size.
528 static int rx_pool_size_compare(const void *p1
, const void *p2
)
530 const RxBufPool
*pool1
= *(RxBufPool
**)p1
;
531 const RxBufPool
*pool2
= *(RxBufPool
**)p2
;
533 if (pool1
->bufsize
< pool2
->bufsize
) {
536 return pool1
->bufsize
> pool2
->bufsize
;
540 * Search for a matching buffer pool with exact matching size,
541 * or return -1 if no matching pool has been found.
543 static int spapr_vlan_get_rx_pool_id(SpaprVioVlan
*dev
, int size
)
547 for (pool
= 0; pool
< RX_MAX_POOLS
; pool
++) {
548 if (dev
->rx_pool
[pool
]->bufsize
== size
) {
557 * Enqueuing receive buffer by adding it to one of our receive buffer pools
559 static target_long
spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan
*dev
,
562 int size
= VLAN_BD_LEN(buf
);
565 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
568 * No matching pool found? Try to use a new one. If the guest used all
569 * pools before, but changed the size of one pool in the meantime, we might
570 * need to recycle that pool here (if it's empty already). Thus scan
571 * all buffer pools now, starting with the last (likely empty) one.
573 for (pool
= RX_MAX_POOLS
- 1; pool
>= 0 ; pool
--) {
574 if (dev
->rx_pool
[pool
]->count
== 0) {
575 dev
->rx_pool
[pool
]->bufsize
= size
;
577 * Sort pools by size so that spapr_vlan_receive()
578 * can later find the smallest buffer pool easily.
580 qsort(dev
->rx_pool
, RX_MAX_POOLS
, sizeof(dev
->rx_pool
[0]),
581 rx_pool_size_compare
);
582 pool
= spapr_vlan_get_rx_pool_id(dev
, size
);
583 trace_spapr_vlan_add_rxbuf_to_pool_create(pool
,
589 /* Still no usable pool? Give up */
590 if (pool
< 0 || dev
->rx_pool
[pool
]->count
>= RX_POOL_MAX_BDS
) {
594 trace_spapr_vlan_add_rxbuf_to_pool(pool
, VLAN_BD_LEN(buf
),
595 dev
->rx_pool
[pool
]->count
);
597 dev
->rx_pool
[pool
]->bds
[dev
->rx_pool
[pool
]->count
++] = buf
;
603 * This is the old way of enqueuing receive buffers: Add it to the rx queue
604 * page that has been supplied by the guest (which is quite limited in size).
606 static target_long
spapr_vlan_add_rxbuf_to_page(SpaprVioVlan
*dev
,
611 if (dev
->rx_bufs
>= VLAN_MAX_BUFS
) {
616 dev
->add_buf_ptr
+= 8;
617 if (dev
->add_buf_ptr
>= VLAN_RX_BDS_LEN
+ VLAN_RX_BDS_OFF
) {
618 dev
->add_buf_ptr
= VLAN_RX_BDS_OFF
;
621 bd
= vio_ldq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
);
622 } while (bd
& VLAN_BD_VALID
);
624 vio_stq(&dev
->sdev
, dev
->buf_list
+ dev
->add_buf_ptr
, buf
);
626 trace_spapr_vlan_add_rxbuf_to_page(dev
->add_buf_ptr
, dev
->rx_bufs
, buf
);
631 static target_ulong
h_add_logical_lan_buffer(PowerPCCPU
*cpu
,
632 SpaprMachineState
*spapr
,
636 target_ulong reg
= args
[0];
637 target_ulong buf
= args
[1];
638 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
639 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
642 trace_spapr_vlan_h_add_logical_lan_buffer(reg
, buf
);
645 hcall_dprintf("Bad device\n");
649 if ((check_bd(dev
, buf
, 4) < 0)
650 || (VLAN_BD_LEN(buf
) < 16)) {
651 hcall_dprintf("Bad buffer enqueued\n");
659 if (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) {
660 ret
= spapr_vlan_add_rxbuf_to_pool(dev
, buf
);
662 ret
= spapr_vlan_add_rxbuf_to_page(dev
, buf
);
671 * Give guest some more time to add additional RX buffers before we
672 * flush the receive queue, so that e.g. fragmented IP packets can
673 * be passed to the guest in one go later (instead of passing single
674 * fragments if there is only one receive buffer available).
676 timer_mod(dev
->rxp_timer
, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL
) + 500);
681 static target_ulong
h_send_logical_lan(PowerPCCPU
*cpu
,
682 SpaprMachineState
*spapr
,
683 target_ulong opcode
, target_ulong
*args
)
685 target_ulong reg
= args
[0];
686 target_ulong
*bufs
= args
+ 1;
687 target_ulong continue_token
= args
[7];
688 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
689 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
695 trace_spapr_vlan_h_send_logical_lan(reg
, continue_token
);
701 trace_spapr_vlan_h_send_logical_lan_rxbufs(dev
->rx_bufs
);
707 if (continue_token
) {
708 return H_HARDWARE
; /* FIXME actually handle this */
712 for (i
= 0; i
< 6; i
++) {
713 trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs
[i
]);
714 if (!(bufs
[i
] & VLAN_BD_VALID
)) {
717 total_len
+= VLAN_BD_LEN(bufs
[i
]);
721 trace_spapr_vlan_h_send_logical_lan_total(nbufs
, total_len
);
723 if (total_len
== 0) {
727 if (total_len
> MAX_PACKET_SIZE
) {
728 /* Don't let the guest force too large an allocation */
732 lbuf
= alloca(total_len
);
734 for (i
= 0; i
< nbufs
; i
++) {
735 ret
= spapr_vio_dma_read(sdev
, VLAN_BD_ADDR(bufs
[i
]),
736 p
, VLAN_BD_LEN(bufs
[i
]));
741 p
+= VLAN_BD_LEN(bufs
[i
]);
744 qemu_send_packet(qemu_get_queue(dev
->nic
), lbuf
, total_len
);
749 static target_ulong
h_multicast_ctrl(PowerPCCPU
*cpu
, SpaprMachineState
*spapr
,
750 target_ulong opcode
, target_ulong
*args
)
752 target_ulong reg
= args
[0];
753 SpaprVioDevice
*dev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
762 static target_ulong
h_change_logical_lan_mac(PowerPCCPU
*cpu
,
763 SpaprMachineState
*spapr
,
767 target_ulong reg
= args
[0];
768 target_ulong macaddr
= args
[1];
769 SpaprVioDevice
*sdev
= spapr_vio_find_by_reg(spapr
->vio_bus
, reg
);
770 SpaprVioVlan
*dev
= VIO_SPAPR_VLAN_DEVICE(sdev
);
773 for (i
= 0; i
< ETH_ALEN
; i
++) {
774 dev
->nicconf
.macaddr
.a
[ETH_ALEN
- i
- 1] = macaddr
& 0xff;
778 qemu_format_nic_info_str(qemu_get_queue(dev
->nic
), dev
->nicconf
.macaddr
.a
);
783 static Property spapr_vlan_properties
[] = {
784 DEFINE_SPAPR_PROPERTIES(SpaprVioVlan
, sdev
),
785 DEFINE_NIC_PROPERTIES(SpaprVioVlan
, nicconf
),
786 DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan
,
787 compat_flags
, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT
, true),
788 DEFINE_PROP_END_OF_LIST(),
791 static bool spapr_vlan_rx_buffer_pools_needed(void *opaque
)
793 SpaprVioVlan
*dev
= opaque
;
795 return (dev
->compat_flags
& SPAPRVLAN_FLAG_RX_BUF_POOLS
) != 0;
798 static const VMStateDescription vmstate_rx_buffer_pool
= {
799 .name
= "spapr_llan/rx_buffer_pool",
801 .minimum_version_id
= 1,
802 .needed
= spapr_vlan_rx_buffer_pools_needed
,
803 .fields
= (VMStateField
[]) {
804 VMSTATE_INT32(bufsize
, RxBufPool
),
805 VMSTATE_INT32(count
, RxBufPool
),
806 VMSTATE_UINT64_ARRAY(bds
, RxBufPool
, RX_POOL_MAX_BDS
),
807 VMSTATE_END_OF_LIST()
811 static const VMStateDescription vmstate_rx_pools
= {
812 .name
= "spapr_llan/rx_pools",
814 .minimum_version_id
= 1,
815 .needed
= spapr_vlan_rx_buffer_pools_needed
,
816 .fields
= (VMStateField
[]) {
817 VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool
, SpaprVioVlan
,
819 vmstate_rx_buffer_pool
, RxBufPool
),
820 VMSTATE_END_OF_LIST()
824 static const VMStateDescription vmstate_spapr_llan
= {
825 .name
= "spapr_llan",
827 .minimum_version_id
= 1,
828 .fields
= (VMStateField
[]) {
829 VMSTATE_SPAPR_VIO(sdev
, SpaprVioVlan
),
831 VMSTATE_BOOL(isopen
, SpaprVioVlan
),
832 VMSTATE_UINT64(buf_list
, SpaprVioVlan
),
833 VMSTATE_UINT32(add_buf_ptr
, SpaprVioVlan
),
834 VMSTATE_UINT32(use_buf_ptr
, SpaprVioVlan
),
835 VMSTATE_UINT32(rx_bufs
, SpaprVioVlan
),
836 VMSTATE_UINT64(rxq_ptr
, SpaprVioVlan
),
838 VMSTATE_END_OF_LIST()
840 .subsections
= (const VMStateDescription
* []) {
846 static void spapr_vlan_class_init(ObjectClass
*klass
, void *data
)
848 DeviceClass
*dc
= DEVICE_CLASS(klass
);
849 SpaprVioDeviceClass
*k
= VIO_SPAPR_DEVICE_CLASS(klass
);
851 k
->realize
= spapr_vlan_realize
;
852 k
->reset
= spapr_vlan_reset
;
853 k
->devnode
= spapr_vlan_devnode
;
854 k
->dt_name
= "l-lan";
855 k
->dt_type
= "network";
856 k
->dt_compatible
= "IBM,l-lan";
857 k
->signal_mask
= 0x1;
858 set_bit(DEVICE_CATEGORY_NETWORK
, dc
->categories
);
859 dc
->props
= spapr_vlan_properties
;
860 k
->rtce_window_size
= 0x10000000;
861 dc
->vmsd
= &vmstate_spapr_llan
;
864 static const TypeInfo spapr_vlan_info
= {
865 .name
= TYPE_VIO_SPAPR_VLAN_DEVICE
,
866 .parent
= TYPE_VIO_SPAPR_DEVICE
,
867 .instance_size
= sizeof(SpaprVioVlan
),
868 .class_init
= spapr_vlan_class_init
,
869 .instance_init
= spapr_vlan_instance_init
,
870 .instance_finalize
= spapr_vlan_instance_finalize
,
873 static void spapr_vlan_register_types(void)
875 spapr_register_hypercall(H_REGISTER_LOGICAL_LAN
, h_register_logical_lan
);
876 spapr_register_hypercall(H_FREE_LOGICAL_LAN
, h_free_logical_lan
);
877 spapr_register_hypercall(H_SEND_LOGICAL_LAN
, h_send_logical_lan
);
878 spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER
,
879 h_add_logical_lan_buffer
);
880 spapr_register_hypercall(H_MULTICAST_CTRL
, h_multicast_ctrl
);
881 spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC
,
882 h_change_logical_lan_mac
);
883 type_register_static(&spapr_vlan_info
);
886 type_init(spapr_vlan_register_types
)