2 * QEMU Freescale eTSEC Emulator
4 * Copyright (c) 2011-2013 AdaCore
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 #include "qemu/osdep.h"
25 #include "net/checksum.h"
28 #include "registers.h"
30 /* #define ETSEC_RING_DEBUG */
31 /* #define HEX_DUMP */
32 /* #define DEBUG_BD */
34 #ifdef ETSEC_RING_DEBUG
35 static const int debug_etsec
= 1;
37 static const int debug_etsec
;
40 #define RING_DEBUG(fmt, ...) do { \
42 qemu_log(fmt , ## __VA_ARGS__); \
48 static void print_tx_bd_flags(uint16_t flags
)
50 qemu_log(" Ready: %d\n", !!(flags
& BD_TX_READY
));
51 qemu_log(" PAD/CRC: %d\n", !!(flags
& BD_TX_PADCRC
));
52 qemu_log(" Wrap: %d\n", !!(flags
& BD_WRAP
));
53 qemu_log(" Interrupt: %d\n", !!(flags
& BD_INTERRUPT
));
54 qemu_log(" Last in frame: %d\n", !!(flags
& BD_LAST
));
55 qemu_log(" Tx CRC: %d\n", !!(flags
& BD_TX_TC
));
56 qemu_log(" User-defined preamble / defer: %d\n",
57 !!(flags
& BD_TX_PREDEF
));
58 qemu_log(" Huge frame enable / Late collision: %d\n",
59 !!(flags
& BD_TX_HFELC
));
60 qemu_log(" Control frame / Retransmission Limit: %d\n",
61 !!(flags
& BD_TX_CFRL
));
62 qemu_log(" Retry count: %d\n",
63 (flags
>> BD_TX_RC_OFFSET
) & BD_TX_RC_MASK
);
64 qemu_log(" Underrun / TCP/IP off-load enable: %d\n",
65 !!(flags
& BD_TX_TOEUN
));
66 qemu_log(" Truncation: %d\n", !!(flags
& BD_TX_TR
));
69 static void print_rx_bd_flags(uint16_t flags
)
71 qemu_log(" Empty: %d\n", !!(flags
& BD_RX_EMPTY
));
72 qemu_log(" Receive software ownership: %d\n", !!(flags
& BD_RX_RO1
));
73 qemu_log(" Wrap: %d\n", !!(flags
& BD_WRAP
));
74 qemu_log(" Interrupt: %d\n", !!(flags
& BD_INTERRUPT
));
75 qemu_log(" Last in frame: %d\n", !!(flags
& BD_LAST
));
76 qemu_log(" First in frame: %d\n", !!(flags
& BD_RX_FIRST
));
77 qemu_log(" Miss: %d\n", !!(flags
& BD_RX_MISS
));
78 qemu_log(" Broadcast: %d\n", !!(flags
& BD_RX_BROADCAST
));
79 qemu_log(" Multicast: %d\n", !!(flags
& BD_RX_MULTICAST
));
80 qemu_log(" Rx frame length violation: %d\n", !!(flags
& BD_RX_LG
));
81 qemu_log(" Rx non-octet aligned frame: %d\n", !!(flags
& BD_RX_NO
));
82 qemu_log(" Short frame: %d\n", !!(flags
& BD_RX_SH
));
83 qemu_log(" Rx CRC Error: %d\n", !!(flags
& BD_RX_CR
));
84 qemu_log(" Overrun: %d\n", !!(flags
& BD_RX_OV
));
85 qemu_log(" Truncation: %d\n", !!(flags
& BD_RX_TR
));
89 static void print_bd(eTSEC_rxtx_bd bd
, int mode
, uint32_t index
)
91 qemu_log("eTSEC %s Data Buffer Descriptor (%u)\n",
92 mode
== eTSEC_TRANSMIT
? "Transmit" : "Receive",
94 qemu_log(" Flags : 0x%04x\n", bd
.flags
);
95 if (mode
== eTSEC_TRANSMIT
) {
96 print_tx_bd_flags(bd
.flags
);
98 print_rx_bd_flags(bd
.flags
);
100 qemu_log(" Length : 0x%04x\n", bd
.length
);
101 qemu_log(" Pointer : 0x%08x\n", bd
.bufptr
);
104 #endif /* DEBUG_BD */
106 static void read_buffer_descriptor(eTSEC
*etsec
,
112 RING_DEBUG("READ Buffer Descriptor @ 0x" TARGET_FMT_plx
"\n", addr
);
113 cpu_physical_memory_read(addr
,
115 sizeof(eTSEC_rxtx_bd
));
117 if (etsec
->regs
[DMACTRL
].value
& DMACTRL_LE
) {
118 bd
->flags
= lduw_le_p(&bd
->flags
);
119 bd
->length
= lduw_le_p(&bd
->length
);
120 bd
->bufptr
= ldl_le_p(&bd
->bufptr
);
122 bd
->flags
= lduw_be_p(&bd
->flags
);
123 bd
->length
= lduw_be_p(&bd
->length
);
124 bd
->bufptr
= ldl_be_p(&bd
->bufptr
);
128 static void write_buffer_descriptor(eTSEC
*etsec
,
134 if (etsec
->regs
[DMACTRL
].value
& DMACTRL_LE
) {
135 stw_le_p(&bd
->flags
, bd
->flags
);
136 stw_le_p(&bd
->length
, bd
->length
);
137 stl_le_p(&bd
->bufptr
, bd
->bufptr
);
139 stw_be_p(&bd
->flags
, bd
->flags
);
140 stw_be_p(&bd
->length
, bd
->length
);
141 stl_be_p(&bd
->bufptr
, bd
->bufptr
);
144 RING_DEBUG("Write Buffer Descriptor @ 0x" TARGET_FMT_plx
"\n", addr
);
145 cpu_physical_memory_write(addr
,
147 sizeof(eTSEC_rxtx_bd
));
150 static void ievent_set(eTSEC
*etsec
,
153 etsec
->regs
[IEVENT
].value
|= flags
;
155 if ((flags
& IEVENT_TXB
&& etsec
->regs
[IMASK
].value
& IMASK_TXBEN
)
156 || (flags
& IEVENT_TXF
&& etsec
->regs
[IMASK
].value
& IMASK_TXFEN
)) {
157 qemu_irq_raise(etsec
->tx_irq
);
158 RING_DEBUG("%s Raise Tx IRQ\n", __func__
);
161 if ((flags
& IEVENT_RXB
&& etsec
->regs
[IMASK
].value
& IMASK_RXBEN
)
162 || (flags
& IEVENT_RXF
&& etsec
->regs
[IMASK
].value
& IMASK_RXFEN
)) {
163 qemu_irq_raise(etsec
->rx_irq
);
164 RING_DEBUG("%s Raise Rx IRQ\n", __func__
);
168 static void tx_padding_and_crc(eTSEC
*etsec
, uint32_t min_frame_len
)
170 int add
= min_frame_len
- etsec
->tx_buffer_len
;
174 RING_DEBUG("pad:%u\n", add
);
175 etsec
->tx_buffer
= g_realloc(etsec
->tx_buffer
,
176 etsec
->tx_buffer_len
+ add
);
178 memset(etsec
->tx_buffer
+ etsec
->tx_buffer_len
, 0x0, add
);
179 etsec
->tx_buffer_len
+= add
;
182 /* Never add CRC in QEMU */
185 static void process_tx_fcb(eTSEC
*etsec
)
187 uint8_t flags
= (uint8_t)(*etsec
->tx_buffer
);
188 /* L3 header offset from start of frame */
189 uint8_t l3_header_offset
= (uint8_t)*(etsec
->tx_buffer
+ 3);
190 /* L4 header offset from start of L3 header */
191 uint8_t l4_header_offset
= (uint8_t)*(etsec
->tx_buffer
+ 2);
193 uint8_t *l3_header
= etsec
->tx_buffer
+ 8 + l3_header_offset
;
195 uint8_t *l4_header
= l3_header
+ l4_header_offset
;
197 /* if packet is IP4 and IP checksum is requested */
198 if (flags
& FCB_TX_IP
&& flags
& FCB_TX_CIP
) {
199 /* do IP4 checksum (TODO This function does TCP/UDP checksum
200 * but not sure if it also does IP4 checksum.) */
201 net_checksum_calculate(etsec
->tx_buffer
+ 8,
202 etsec
->tx_buffer_len
- 8);
204 /* TODO Check the correct usage of the PHCS field of the FCB in case the NPH
207 /* if packet is IP4 and TCP or UDP */
208 if (flags
& FCB_TX_IP
&& flags
& FCB_TX_TUP
) {
210 if (flags
& FCB_TX_UDP
) {
211 /* if checksum is requested */
212 if (flags
& FCB_TX_CTU
) {
213 /* do UDP checksum */
215 net_checksum_calculate(etsec
->tx_buffer
+ 8,
216 etsec
->tx_buffer_len
- 8);
218 /* set checksum field to 0 */
222 } else if (flags
& FCB_TX_CTU
) { /* if TCP and checksum is requested */
223 /* do TCP checksum */
224 net_checksum_calculate(etsec
->tx_buffer
+ 8,
225 etsec
->tx_buffer_len
- 8);
230 static void process_tx_bd(eTSEC
*etsec
,
233 uint8_t *tmp_buff
= NULL
;
234 hwaddr tbdbth
= (hwaddr
)(etsec
->regs
[TBDBPH
].value
& 0xF) << 32;
236 if (bd
->length
== 0) {
241 if (etsec
->tx_buffer_len
== 0) {
242 /* It's the first BD */
243 etsec
->first_bd
= *bd
;
246 /* TODO: if TxBD[TOE/UN] skip the Tx Frame Control Block*/
248 /* Load this Data Buffer */
249 etsec
->tx_buffer
= g_realloc(etsec
->tx_buffer
,
250 etsec
->tx_buffer_len
+ bd
->length
);
251 tmp_buff
= etsec
->tx_buffer
+ etsec
->tx_buffer_len
;
252 cpu_physical_memory_read(bd
->bufptr
+ tbdbth
, tmp_buff
, bd
->length
);
254 /* Update buffer length */
255 etsec
->tx_buffer_len
+= bd
->length
;
258 if (etsec
->tx_buffer_len
!= 0 && (bd
->flags
& BD_LAST
)) {
259 if (etsec
->regs
[MACCFG1
].value
& MACCFG1_TX_EN
) {
260 /* MAC Transmit enabled */
262 /* Process offload Tx FCB */
263 if (etsec
->first_bd
.flags
& BD_TX_TOEUN
) {
264 process_tx_fcb(etsec
);
267 if (etsec
->first_bd
.flags
& BD_TX_PADCRC
268 || etsec
->regs
[MACCFG2
].value
& MACCFG2_PADCRC
) {
270 /* Padding and CRC (Padding implies CRC) */
271 tx_padding_and_crc(etsec
, 64);
273 } else if (etsec
->first_bd
.flags
& BD_TX_TC
274 || etsec
->regs
[MACCFG2
].value
& MACCFG2_CRC_EN
) {
277 /* Never add CRC in QEMU */
280 #if defined(HEX_DUMP)
281 qemu_log("eTSEC Send packet size:%d\n", etsec
->tx_buffer_len
);
282 qemu_hexdump(etsec
->tx_buffer
, stderr
, "", etsec
->tx_buffer_len
);
283 #endif /* ETSEC_RING_DEBUG */
285 if (etsec
->first_bd
.flags
& BD_TX_TOEUN
) {
286 qemu_send_packet(qemu_get_queue(etsec
->nic
),
287 etsec
->tx_buffer
+ 8,
288 etsec
->tx_buffer_len
- 8);
290 qemu_send_packet(qemu_get_queue(etsec
->nic
),
292 etsec
->tx_buffer_len
);
297 etsec
->tx_buffer_len
= 0;
299 if (bd
->flags
& BD_INTERRUPT
) {
300 ievent_set(etsec
, IEVENT_TXF
);
303 if (bd
->flags
& BD_INTERRUPT
) {
304 ievent_set(etsec
, IEVENT_TXB
);
308 /* Update DB flags */
311 bd
->flags
&= ~BD_TX_READY
;
314 bd
->flags
&= ~BD_TX_PREDEF
;
316 /* Clear Late Collision */
317 bd
->flags
&= ~BD_TX_HFELC
;
319 /* Clear Retransmission Limit */
320 bd
->flags
&= ~BD_TX_CFRL
;
322 /* Clear Retry Count */
323 bd
->flags
&= ~(BD_TX_RC_MASK
<< BD_TX_RC_OFFSET
);
326 bd
->flags
&= ~BD_TX_TOEUN
;
328 /* Clear Truncation */
329 bd
->flags
&= ~BD_TX_TR
;
332 void etsec_walk_tx_ring(eTSEC
*etsec
, int ring_nbr
)
334 hwaddr ring_base
= 0;
339 if (!(etsec
->regs
[MACCFG1
].value
& MACCFG1_TX_EN
)) {
340 RING_DEBUG("%s: MAC Transmit not enabled\n", __func__
);
344 ring_base
= (hwaddr
)(etsec
->regs
[TBASEH
].value
& 0xF) << 32;
345 ring_base
+= etsec
->regs
[TBASE0
+ ring_nbr
].value
& ~0x7;
346 bd_addr
= etsec
->regs
[TBPTR0
+ ring_nbr
].value
& ~0x7;
349 read_buffer_descriptor(etsec
, bd_addr
, &bd
);
354 (bd_addr
- ring_base
) / sizeof(eTSEC_rxtx_bd
));
356 #endif /* DEBUG_BD */
358 /* Save flags before BD update */
361 if (!(bd_flags
& BD_TX_READY
)) {
365 process_tx_bd(etsec
, &bd
);
366 /* Write back BD after update */
367 write_buffer_descriptor(etsec
, bd_addr
, &bd
);
369 /* Wrap or next BD */
370 if (bd_flags
& BD_WRAP
) {
373 bd_addr
+= sizeof(eTSEC_rxtx_bd
);
377 /* Save the Buffer Descriptor Pointers to last bd that was not
378 * succesfully closed */
379 etsec
->regs
[TBPTR0
+ ring_nbr
].value
= bd_addr
;
381 /* Set transmit halt THLTx */
382 etsec
->regs
[TSTAT
].value
|= 1 << (31 - ring_nbr
);
385 static void fill_rx_bd(eTSEC
*etsec
,
391 hwaddr bufptr
= bd
->bufptr
+
392 ((hwaddr
)(etsec
->regs
[TBDBPH
].value
& 0xF) << 32);
393 uint8_t padd
[etsec
->rx_padding
];
396 RING_DEBUG("eTSEC fill Rx buffer @ 0x%016" HWADDR_PRIx
397 " size:%zu(padding + crc:%u) + fcb:%u\n",
398 bufptr
, *size
, etsec
->rx_padding
, etsec
->rx_fcb_size
);
402 /* This operation will only write FCB */
403 if (etsec
->rx_fcb_size
!= 0) {
405 cpu_physical_memory_write(bufptr
, etsec
->rx_fcb
, etsec
->rx_fcb_size
);
407 bufptr
+= etsec
->rx_fcb_size
;
408 bd
->length
+= etsec
->rx_fcb_size
;
409 etsec
->rx_fcb_size
= 0;
413 /* We remove padding from the computation of to_write because it is not
414 * allocated in the buffer.
416 to_write
= MIN(*size
- etsec
->rx_padding
,
417 etsec
->regs
[MRBLR
].value
- etsec
->rx_fcb_size
);
419 /* This operation can only write packet data and no padding */
421 cpu_physical_memory_write(bufptr
, *buf
, to_write
);
427 bd
->flags
&= ~BD_RX_EMPTY
;
428 bd
->length
+= to_write
;
431 if (*size
== etsec
->rx_padding
) {
432 /* The remaining bytes are only for padding which is not actually
433 * allocated in the data buffer.
436 rem
= MIN(etsec
->regs
[MRBLR
].value
- bd
->length
, etsec
->rx_padding
);
439 memset(padd
, 0x0, sizeof(padd
));
440 etsec
->rx_padding
-= rem
;
443 cpu_physical_memory_write(bufptr
, padd
, rem
);
448 static void rx_init_frame(eTSEC
*etsec
, const uint8_t *buf
, size_t size
)
450 uint32_t fcb_size
= 0;
451 uint8_t prsdep
= (etsec
->regs
[RCTRL
].value
>> RCTRL_PRSDEP_OFFSET
)
455 /* Prepend FCB (FCB size + RCTRL[PAL]) */
456 fcb_size
= 8 + ((etsec
->regs
[RCTRL
].value
>> 16) & 0x1F);
458 etsec
->rx_fcb_size
= fcb_size
;
460 /* TODO: fill_FCB(etsec); */
461 memset(etsec
->rx_fcb
, 0x0, sizeof(etsec
->rx_fcb
));
464 etsec
->rx_fcb_size
= 0;
467 g_free(etsec
->rx_buffer
);
469 /* Do not copy the frame for now */
470 etsec
->rx_buffer
= (uint8_t *)buf
;
471 etsec
->rx_buffer_len
= size
;
473 /* CRC padding (We don't have to compute the CRC) */
474 etsec
->rx_padding
= 4;
477 * Ensure that payload length + CRC length is at least 802.3
478 * minimum MTU size bytes long (64)
480 if (etsec
->rx_buffer_len
< 60) {
481 etsec
->rx_padding
+= 60 - etsec
->rx_buffer_len
;
484 etsec
->rx_first_in_frame
= 1;
485 etsec
->rx_remaining_data
= etsec
->rx_buffer_len
;
486 RING_DEBUG("%s: rx_buffer_len:%u rx_padding+crc:%u\n", __func__
,
487 etsec
->rx_buffer_len
, etsec
->rx_padding
);
490 ssize_t
etsec_rx_ring_write(eTSEC
*etsec
, const uint8_t *buf
, size_t size
)
492 int ring_nbr
= 0; /* Always use ring0 (no filer) */
494 if (etsec
->rx_buffer_len
!= 0) {
495 RING_DEBUG("%s: We can't receive now,"
496 " a buffer is already in the pipe\n", __func__
);
500 if (etsec
->regs
[RSTAT
].value
& 1 << (23 - ring_nbr
)) {
501 RING_DEBUG("%s: The ring is halted\n", __func__
);
505 if (etsec
->regs
[DMACTRL
].value
& DMACTRL_GRS
) {
506 RING_DEBUG("%s: Graceful receive stop\n", __func__
);
510 if (!(etsec
->regs
[MACCFG1
].value
& MACCFG1_RX_EN
)) {
511 RING_DEBUG("%s: MAC Receive not enabled\n", __func__
);
515 if ((etsec
->regs
[RCTRL
].value
& RCTRL_RSF
) && (size
< 60)) {
516 /* CRC is not in the packet yet, so short frame is below 60 bytes */
517 RING_DEBUG("%s: Drop short frame\n", __func__
);
521 rx_init_frame(etsec
, buf
, size
);
523 etsec_walk_rx_ring(etsec
, ring_nbr
);
528 void etsec_walk_rx_ring(eTSEC
*etsec
, int ring_nbr
)
530 hwaddr ring_base
= 0;
532 hwaddr start_bd_addr
= 0;
535 size_t remaining_data
;
540 if (etsec
->rx_buffer_len
== 0) {
541 /* No frame to send */
542 RING_DEBUG("No frame to send\n");
546 remaining_data
= etsec
->rx_remaining_data
+ etsec
->rx_padding
;
547 buf
= etsec
->rx_buffer
548 + (etsec
->rx_buffer_len
- etsec
->rx_remaining_data
);
549 size
= etsec
->rx_buffer_len
+ etsec
->rx_padding
;
551 ring_base
= (hwaddr
)(etsec
->regs
[RBASEH
].value
& 0xF) << 32;
552 ring_base
+= etsec
->regs
[RBASE0
+ ring_nbr
].value
& ~0x7;
553 start_bd_addr
= bd_addr
= etsec
->regs
[RBPTR0
+ ring_nbr
].value
& ~0x7;
556 read_buffer_descriptor(etsec
, bd_addr
, &bd
);
561 (bd_addr
- ring_base
) / sizeof(eTSEC_rxtx_bd
));
563 #endif /* DEBUG_BD */
565 /* Save flags before BD update */
568 if (bd_flags
& BD_RX_EMPTY
) {
569 fill_rx_bd(etsec
, &bd
, &buf
, &remaining_data
);
571 if (etsec
->rx_first_in_frame
) {
572 bd
.flags
|= BD_RX_FIRST
;
573 etsec
->rx_first_in_frame
= 0;
574 etsec
->rx_first_bd
= bd
;
578 if (remaining_data
== 0) {
586 /* NOTE: non-octet aligned frame is impossible in qemu */
588 if (size
>= etsec
->regs
[MAXFRM
].value
) {
589 /* frame length violation */
590 qemu_log("%s frame length violation: size:%zu MAXFRM:%d\n",
591 __func__
, size
, etsec
->regs
[MAXFRM
].value
);
593 bd
.flags
|= BD_RX_LG
;
598 bd
.flags
|= BD_RX_SH
;
601 /* TODO: Broadcast and Multicast */
603 if (bd
.flags
& BD_INTERRUPT
) {
605 etsec
->regs
[RSTAT
].value
|= 1 << (7 - ring_nbr
);
608 ievent_set(etsec
, IEVENT_RXF
);
612 if (bd
.flags
& BD_INTERRUPT
) {
614 ievent_set(etsec
, IEVENT_RXB
);
618 /* Write back BD after update */
619 write_buffer_descriptor(etsec
, bd_addr
, &bd
);
622 /* Wrap or next BD */
623 if (bd_flags
& BD_WRAP
) {
626 bd_addr
+= sizeof(eTSEC_rxtx_bd
);
628 } while (remaining_data
!= 0
629 && (bd_flags
& BD_RX_EMPTY
)
630 && bd_addr
!= start_bd_addr
);
633 etsec
->regs
[RBPTR0
+ ring_nbr
].value
= bd_addr
;
635 /* The frame is too large to fit in the Rx ring */
636 if (remaining_data
> 0) {
638 /* Set RSTAT[QHLTx] */
639 etsec
->regs
[RSTAT
].value
|= 1 << (23 - ring_nbr
);
641 /* Save remaining data to send the end of the frame when the ring will
644 etsec
->rx_remaining_data
= remaining_data
;
647 tmp_buf
= g_malloc(size
);
648 memcpy(tmp_buf
, etsec
->rx_buffer
, size
);
649 etsec
->rx_buffer
= tmp_buf
;
651 RING_DEBUG("no empty RxBD available any more\n");
653 etsec
->rx_buffer_len
= 0;
654 etsec
->rx_buffer
= NULL
;
655 if (etsec
->need_flush
) {
656 qemu_flush_queued_packets(qemu_get_queue(etsec
->nic
));
660 RING_DEBUG("eTSEC End of ring_write: remaining_data:%zu\n", remaining_data
);