1 /******************************************************************************
3 Copyright (c) 2006-2009, Myricom Inc.
6 Redistribution and use in source and binary forms, with or without
7 modification, are permitted provided that the following conditions are met:
9 1. Redistributions of source code must retain the above copyright notice,
10 this list of conditions and the following disclaimer.
12 2. Neither the name of the Myricom Inc, nor the names of its
13 contributors may be used to endorse or promote products derived from
14 this software without specific prior written permission.
16 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
17 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
20 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
21 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
22 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
23 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
24 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
25 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
26 POSSIBILITY OF SUCH DAMAGE.
28 ***************************************************************************/
30 #include <sys/cdefs.h>
31 /*__FBSDID("$FreeBSD: src/sys/dev/mxge/if_mxge.c,v 1.63 2009/06/26 11:45:06 rwatson Exp $");*/
33 #include <sys/param.h>
34 #include <sys/systm.h>
35 #include <sys/linker.h>
36 #include <sys/firmware.h>
37 #include <sys/endian.h>
38 #include <sys/in_cksum.h>
39 #include <sys/sockio.h>
41 #include <sys/malloc.h>
42 #include <sys/kernel.h>
43 #include <sys/module.h>
44 #include <sys/serialize.h>
45 #include <sys/socket.h>
46 #include <sys/sysctl.h>
48 /* count xmits ourselves, rather than via drbr */
51 #include <net/if_arp.h>
52 #include <net/ifq_var.h>
53 #include <net/ethernet.h>
54 #include <net/if_dl.h>
55 #include <net/if_media.h>
59 #include <net/if_types.h>
60 #include <net/vlan/if_vlan_var.h>
63 #include <netinet/in_systm.h>
64 #include <netinet/in.h>
65 #include <netinet/ip.h>
66 #include <netinet/tcp.h>
71 #include <bus/pci/pcireg.h>
72 #include <bus/pci/pcivar.h>
73 #include <bus/pci/pci_private.h> /* XXX for pci_cfg_restore */
75 #include <vm/vm.h> /* for pmap_mapdev() */
78 #if defined(__i386) || defined(__amd64)
79 #include <machine/specialreg.h>
82 #include <dev/netif/mxge/mxge_mcp.h>
83 #include <dev/netif/mxge/mcp_gen_header.h>
84 /*#define MXGE_FAKE_IFP*/
85 #include <dev/netif/mxge/if_mxge_var.h>
87 #include <sys/buf_ring.h>
93 static int mxge_nvidia_ecrc_enable
= 1;
94 static int mxge_force_firmware
= 0;
95 static int mxge_intr_coal_delay
= 30;
96 static int mxge_deassert_wait
= 1;
97 static int mxge_flow_control
= 1;
98 static int mxge_verbose
= 0;
99 static int mxge_lro_cnt
= 8;
100 static int mxge_ticks
;
101 static int mxge_max_slices
= 1;
102 static int mxge_rss_hash_type
= MXGEFW_RSS_HASH_TYPE_SRC_PORT
;
103 static int mxge_always_promisc
= 0;
105 /* static int mxge_initial_mtu = ETHERMTU_JUMBO; */
106 static int mxge_initial_mtu
= ETHERMTU
;
107 static char *mxge_fw_unaligned
= "mxge_ethp_z8e";
108 static char *mxge_fw_aligned
= "mxge_eth_z8e";
109 static char *mxge_fw_rss_aligned
= "mxge_rss_eth_z8e";
110 static char *mxge_fw_rss_unaligned
= "mxge_rss_ethp_z8e";
112 static int mxge_probe(device_t dev
);
113 static int mxge_attach(device_t dev
);
114 static int mxge_detach(device_t dev
);
115 static int mxge_shutdown(device_t dev
);
116 static void mxge_intr(void *arg
);
118 static device_method_t mxge_methods
[] =
120 /* Device interface */
121 DEVMETHOD(device_probe
, mxge_probe
),
122 DEVMETHOD(device_attach
, mxge_attach
),
123 DEVMETHOD(device_detach
, mxge_detach
),
124 DEVMETHOD(device_shutdown
, mxge_shutdown
),
128 static driver_t mxge_driver
=
132 sizeof(mxge_softc_t
),
135 static devclass_t mxge_devclass
;
137 /* Declare ourselves to be a child of the PCI bus.*/
138 DRIVER_MODULE(mxge
, pci
, mxge_driver
, mxge_devclass
, 0, 0);
139 MODULE_DEPEND(mxge
, firmware
, 1, 1, 1);
140 MODULE_DEPEND(mxge
, zlib
, 1, 1, 1);
142 static int mxge_load_firmware(mxge_softc_t
*sc
, int adopt
);
143 static int mxge_send_cmd(mxge_softc_t
*sc
, uint32_t cmd
, mxge_cmd_t
*data
);
144 static int mxge_close(mxge_softc_t
*sc
);
145 static int mxge_open(mxge_softc_t
*sc
);
146 static void mxge_tick(void *arg
);
148 /* XXX: we don't have Large Receive Offload support yet */
150 mxge_lro_rx(struct mxge_slice_state
*ss
, struct mbuf
*m_head
, uint32_t csum
)
159 mxge_lro_flush(struct mxge_slice_state
*ss
, struct lro_entry
*lro
)
166 mxge_probe(device_t dev
)
171 if ((pci_get_vendor(dev
) == MXGE_PCI_VENDOR_MYRICOM
) &&
172 ((pci_get_device(dev
) == MXGE_PCI_DEVICE_Z8E
) ||
173 (pci_get_device(dev
) == MXGE_PCI_DEVICE_Z8E_9
))) {
174 rev
= pci_get_revid(dev
);
176 case MXGE_PCI_REV_Z8E
:
177 device_set_desc(dev
, "Myri10G-PCIE-8A");
179 case MXGE_PCI_REV_Z8ES
:
180 device_set_desc(dev
, "Myri10G-PCIE-8B");
183 device_set_desc(dev
, "Myri10G-PCIE-8??");
184 device_printf(dev
, "Unrecognized rev %d NIC\n",
194 mxge_enable_wc(mxge_softc_t
*sc
)
197 #if defined(__i386) || defined(__amd64)
202 len
= rman_get_size(sc
->mem_res
);
203 err
= pmap_change_attr((vm_offset_t
) sc
->sram
,
204 len
, PAT_WRITE_COMBINING
);
206 device_printf(sc
->dev
, "pmap_change_attr failed, %d\n",
212 sc
->wc
= 0; /* TBD: PAT support */
217 /* callback to get our DMA address */
219 mxge_dmamap_callback(void *arg
, bus_dma_segment_t
*segs
, int nsegs
,
223 *(bus_addr_t
*) arg
= segs
->ds_addr
;
228 mxge_dma_alloc(mxge_softc_t
*sc
, mxge_dma_t
*dma
, size_t bytes
,
229 bus_size_t alignment
)
232 device_t dev
= sc
->dev
;
233 bus_size_t boundary
, maxsegsize
;
235 if (bytes
> 4096 && alignment
== 4096) {
243 /* allocate DMAable memory tags */
244 err
= bus_dma_tag_create(sc
->parent_dmat
, /* parent */
245 alignment
, /* alignment */
246 boundary
, /* boundary */
247 BUS_SPACE_MAXADDR
, /* low */
248 BUS_SPACE_MAXADDR
, /* high */
249 NULL
, NULL
, /* filter */
252 maxsegsize
, /* maxsegsize */
253 BUS_DMA_COHERENT
, /* flags */
254 &dma
->dmat
); /* tag */
256 device_printf(dev
, "couldn't alloc tag (err = %d)\n", err
);
260 /* allocate DMAable memory & map */
261 err
= bus_dmamem_alloc(dma
->dmat
, &dma
->addr
,
262 (BUS_DMA_WAITOK
| BUS_DMA_COHERENT
263 | BUS_DMA_ZERO
), &dma
->map
);
265 device_printf(dev
, "couldn't alloc mem (err = %d)\n", err
);
266 goto abort_with_dmat
;
269 /* load the memory */
270 err
= bus_dmamap_load(dma
->dmat
, dma
->map
, dma
->addr
, bytes
,
271 mxge_dmamap_callback
,
272 (void *)&dma
->bus_addr
, 0);
274 device_printf(dev
, "couldn't load map (err = %d)\n", err
);
280 bus_dmamem_free(dma
->dmat
, dma
->addr
, dma
->map
);
282 (void)bus_dma_tag_destroy(dma
->dmat
);
288 mxge_dma_free(mxge_dma_t
*dma
)
290 bus_dmamap_unload(dma
->dmat
, dma
->map
);
291 bus_dmamem_free(dma
->dmat
, dma
->addr
, dma
->map
);
292 (void)bus_dma_tag_destroy(dma
->dmat
);
296 * The eeprom strings on the lanaiX have the format
303 mxge_parse_strings(mxge_softc_t
*sc
)
305 #define MXGE_NEXT_STRING(p) while(ptr < limit && *ptr++)
310 ptr
= sc
->eeprom_strings
;
311 limit
= sc
->eeprom_strings
+ MXGE_EEPROM_STRINGS_SIZE
;
313 while (ptr
< limit
&& *ptr
!= '\0') {
314 if (memcmp(ptr
, "MAC=", 4) == 0) {
316 sc
->mac_addr_string
= ptr
;
317 for (i
= 0; i
< 6; i
++) {
319 if ((ptr
+ 2) > limit
)
321 sc
->mac_addr
[i
] = strtoul(ptr
, NULL
, 16);
324 } else if (memcmp(ptr
, "PC=", 3) == 0) {
326 strncpy(sc
->product_code_string
, ptr
,
327 sizeof (sc
->product_code_string
) - 1);
328 } else if (memcmp(ptr
, "SN=", 3) == 0) {
330 strncpy(sc
->serial_number_string
, ptr
,
331 sizeof (sc
->serial_number_string
) - 1);
333 MXGE_NEXT_STRING(ptr
);
340 device_printf(sc
->dev
, "failed to parse eeprom_strings\n");
345 #if defined __i386 || defined i386 || defined __i386__ || defined __x86_64__
347 mxge_enable_nvidia_ecrc(mxge_softc_t
*sc
)
350 unsigned long base
, off
;
352 device_t pdev
, mcp55
;
353 uint16_t vendor_id
, device_id
, word
;
354 uintptr_t bus
, slot
, func
, ivend
, idev
;
358 if (!mxge_nvidia_ecrc_enable
)
361 pdev
= device_get_parent(device_get_parent(sc
->dev
));
363 device_printf(sc
->dev
, "could not find parent?\n");
366 vendor_id
= pci_read_config(pdev
, PCIR_VENDOR
, 2);
367 device_id
= pci_read_config(pdev
, PCIR_DEVICE
, 2);
369 if (vendor_id
!= 0x10de)
374 if (device_id
== 0x005d) {
375 /* ck804, base address is magic */
377 } else if (device_id
>= 0x0374 && device_id
<= 0x378) {
378 /* mcp55, base address stored in chipset */
379 mcp55
= pci_find_bsf(0, 0, 0);
381 0x10de == pci_read_config(mcp55
, PCIR_VENDOR
, 2) &&
382 0x0369 == pci_read_config(mcp55
, PCIR_DEVICE
, 2)) {
383 word
= pci_read_config(mcp55
, 0x90, 2);
384 base
= ((unsigned long)word
& 0x7ffeU
) << 25;
391 Test below is commented because it is believed that doing
392 config read/write beyond 0xff will access the config space
393 for the next larger function. Uncomment this and remove
394 the hacky pmap_mapdev() way of accessing config space when
395 FreeBSD grows support for extended pcie config space access
398 /* See if we can, by some miracle, access the extended
400 val
= pci_read_config(pdev
, 0x178, 4);
401 if (val
!= 0xffffffff) {
403 pci_write_config(pdev
, 0x178, val
, 4);
407 /* Rather than using normal pci config space writes, we must
408 * map the Nvidia config space ourselves. This is because on
409 * opteron/nvidia class machine the 0xe000000 mapping is
410 * handled by the nvidia chipset, that means the internal PCI
411 * device (the on-chip northbridge), or the amd-8131 bridge
412 * and things behind them are not visible by this method.
415 BUS_READ_IVAR(device_get_parent(pdev
), pdev
,
417 BUS_READ_IVAR(device_get_parent(pdev
), pdev
,
418 PCI_IVAR_SLOT
, &slot
);
419 BUS_READ_IVAR(device_get_parent(pdev
), pdev
,
420 PCI_IVAR_FUNCTION
, &func
);
421 BUS_READ_IVAR(device_get_parent(pdev
), pdev
,
422 PCI_IVAR_VENDOR
, &ivend
);
423 BUS_READ_IVAR(device_get_parent(pdev
), pdev
,
424 PCI_IVAR_DEVICE
, &idev
);
427 + 0x00100000UL
* (unsigned long)bus
428 + 0x00001000UL
* (unsigned long)(func
431 /* map it into the kernel */
432 va
= pmap_mapdev(trunc_page((vm_paddr_t
)off
), PAGE_SIZE
);
436 device_printf(sc
->dev
, "pmap_kenter_temporary didn't\n");
439 /* get a pointer to the config space mapped into the kernel */
440 cfgptr
= va
+ (off
& PAGE_MASK
);
442 /* make sure that we can really access it */
443 vendor_id
= *(uint16_t *)(cfgptr
+ PCIR_VENDOR
);
444 device_id
= *(uint16_t *)(cfgptr
+ PCIR_DEVICE
);
445 if (! (vendor_id
== ivend
&& device_id
== idev
)) {
446 device_printf(sc
->dev
, "mapping failed: 0x%x:0x%x\n",
447 vendor_id
, device_id
);
448 pmap_unmapdev((vm_offset_t
)va
, PAGE_SIZE
);
452 ptr32
= (uint32_t*)(cfgptr
+ 0x178);
455 if (val
== 0xffffffff) {
456 device_printf(sc
->dev
, "extended mapping failed\n");
457 pmap_unmapdev((vm_offset_t
)va
, PAGE_SIZE
);
461 pmap_unmapdev((vm_offset_t
)va
, PAGE_SIZE
);
463 device_printf(sc
->dev
,
464 "Enabled ECRC on upstream Nvidia bridge "
466 (int)bus
, (int)slot
, (int)func
);
471 mxge_enable_nvidia_ecrc(mxge_softc_t
*sc
)
473 device_printf(sc
->dev
,
474 "Nforce 4 chipset on non-x86/amd64!?!?!\n");
481 mxge_dma_test(mxge_softc_t
*sc
, int test_type
)
484 bus_addr_t dmatest_bus
= sc
->dmabench_dma
.bus_addr
;
490 /* Run a small DMA test.
491 * The magic multipliers to the length tell the firmware
492 * to do DMA read, write, or read+write tests. The
493 * results are returned in cmd.data0. The upper 16
494 * bits of the return is the number of transfers completed.
495 * The lower 16 bits is the time in 0.5us ticks that the
496 * transfers took to complete.
499 len
= sc
->tx_boundary
;
501 cmd
.data0
= MXGE_LOWPART_TO_U32(dmatest_bus
);
502 cmd
.data1
= MXGE_HIGHPART_TO_U32(dmatest_bus
);
503 cmd
.data2
= len
* 0x10000;
504 status
= mxge_send_cmd(sc
, test_type
, &cmd
);
509 sc
->read_dma
= ((cmd
.data0
>>16) * len
* 2) /
510 (cmd
.data0
& 0xffff);
511 cmd
.data0
= MXGE_LOWPART_TO_U32(dmatest_bus
);
512 cmd
.data1
= MXGE_HIGHPART_TO_U32(dmatest_bus
);
513 cmd
.data2
= len
* 0x1;
514 status
= mxge_send_cmd(sc
, test_type
, &cmd
);
519 sc
->write_dma
= ((cmd
.data0
>>16) * len
* 2) /
520 (cmd
.data0
& 0xffff);
522 cmd
.data0
= MXGE_LOWPART_TO_U32(dmatest_bus
);
523 cmd
.data1
= MXGE_HIGHPART_TO_U32(dmatest_bus
);
524 cmd
.data2
= len
* 0x10001;
525 status
= mxge_send_cmd(sc
, test_type
, &cmd
);
530 sc
->read_write_dma
= ((cmd
.data0
>>16) * len
* 2 * 2) /
531 (cmd
.data0
& 0xffff);
534 if (status
!= 0 && test_type
!= MXGEFW_CMD_UNALIGNED_TEST
)
535 device_printf(sc
->dev
, "DMA %s benchmark failed: %d\n",
542 * The Lanai Z8E PCI-E interface achieves higher Read-DMA throughput
543 * when the PCI-E Completion packets are aligned on an 8-byte
544 * boundary. Some PCI-E chip sets always align Completion packets; on
545 * the ones that do not, the alignment can be enforced by enabling
546 * ECRC generation (if supported).
548 * When PCI-E Completion packets are not aligned, it is actually more
549 * efficient to limit Read-DMA transactions to 2KB, rather than 4KB.
551 * If the driver can neither enable ECRC nor verify that it has
552 * already been enabled, then it must use a firmware image which works
553 * around unaligned completion packets (ethp_z8e.dat), and it should
554 * also ensure that it never gives the device a Read-DMA which is
555 * larger than 2KB by setting the tx_boundary to 2KB. If ECRC is
556 * enabled, then the driver should use the aligned (eth_z8e.dat)
557 * firmware image, and set tx_boundary to 4KB.
561 mxge_firmware_probe(mxge_softc_t
*sc
)
563 device_t dev
= sc
->dev
;
567 sc
->tx_boundary
= 4096;
569 * Verify the max read request size was set to 4KB
570 * before trying the test with 4KB.
572 if (pci_find_extcap(dev
, PCIY_EXPRESS
, ®
) == 0) {
573 pectl
= pci_read_config(dev
, reg
+ 0x8, 2);
574 if ((pectl
& (5 << 12)) != (5 << 12)) {
575 device_printf(dev
, "Max Read Req. size != 4k (0x%x\n",
577 sc
->tx_boundary
= 2048;
582 * load the optimized firmware (which assumes aligned PCIe
583 * completions) in order to see if it works on this host.
585 sc
->fw_name
= mxge_fw_aligned
;
586 status
= mxge_load_firmware(sc
, 1);
592 * Enable ECRC if possible
594 mxge_enable_nvidia_ecrc(sc
);
597 * Run a DMA test which watches for unaligned completions and
598 * aborts on the first one seen.
601 status
= mxge_dma_test(sc
, MXGEFW_CMD_UNALIGNED_TEST
);
603 return 0; /* keep the aligned firmware */
606 device_printf(dev
, "DMA test failed: %d\n", status
);
607 if (status
== ENOSYS
)
608 device_printf(dev
, "Falling back to ethp! "
609 "Please install up to date fw\n");
614 mxge_select_firmware(mxge_softc_t
*sc
)
619 if (mxge_force_firmware
!= 0) {
620 if (mxge_force_firmware
== 1)
625 device_printf(sc
->dev
,
626 "Assuming %s completions (forced)\n",
627 aligned
? "aligned" : "unaligned");
631 /* if the PCIe link width is 4 or less, we can use the aligned
632 firmware and skip any checks */
633 if (sc
->link_width
!= 0 && sc
->link_width
<= 4) {
634 device_printf(sc
->dev
,
635 "PCIe x%d Link, expect reduced performance\n",
641 if (0 == mxge_firmware_probe(sc
))
646 sc
->fw_name
= mxge_fw_aligned
;
647 sc
->tx_boundary
= 4096;
649 sc
->fw_name
= mxge_fw_unaligned
;
650 sc
->tx_boundary
= 2048;
652 return (mxge_load_firmware(sc
, 0));
662 mxge_validate_firmware(mxge_softc_t
*sc
, const mcp_gen_header_t
*hdr
)
666 if (be32toh(hdr
->mcp_type
) != MCP_TYPE_ETH
) {
667 device_printf(sc
->dev
, "Bad firmware type: 0x%x\n",
668 be32toh(hdr
->mcp_type
));
672 /* save firmware version for sysctl */
673 strncpy(sc
->fw_version
, hdr
->version
, sizeof (sc
->fw_version
));
675 device_printf(sc
->dev
, "firmware id: %s\n", hdr
->version
);
677 ksscanf(sc
->fw_version
, "%d.%d.%d", &sc
->fw_ver_major
,
678 &sc
->fw_ver_minor
, &sc
->fw_ver_tiny
);
680 if (!(sc
->fw_ver_major
== MXGEFW_VERSION_MAJOR
681 && sc
->fw_ver_minor
== MXGEFW_VERSION_MINOR
)) {
682 device_printf(sc
->dev
, "Found firmware version %s\n",
684 device_printf(sc
->dev
, "Driver needs %d.%d\n",
685 MXGEFW_VERSION_MAJOR
, MXGEFW_VERSION_MINOR
);
694 z_alloc(void *nil
, u_int items
, u_int size
)
698 ptr
= kmalloc(items
* size
, M_TEMP
, M_NOWAIT
);
703 z_free(void *nil
, void *ptr
)
710 mxge_load_firmware_helper(mxge_softc_t
*sc
, uint32_t *limit
)
713 const mcp_gen_header_t
*hdr
;
720 fw
= firmware_image_load(sc
->fw_name
, NULL
);
722 device_printf(sc
->dev
, "Could not find firmware image %s\n",
727 /* setup zlib and decompress f/w */
728 bzero(&zs
, sizeof (zs
));
731 status
= inflateInit(&zs
);
732 if (status
!= Z_OK
) {
737 /* the uncompressed size is stored as the firmware version,
738 which would otherwise go unused */
739 fw_len
= (size_t) fw
->version
;
740 inflate_buffer
= kmalloc(fw_len
, M_TEMP
, M_NOWAIT
);
741 if (inflate_buffer
== NULL
)
743 zs
.avail_in
= fw
->datasize
;
744 zs
.next_in
= __DECONST(char *, fw
->data
);
745 zs
.avail_out
= fw_len
;
746 zs
.next_out
= inflate_buffer
;
747 status
= inflate(&zs
, Z_FINISH
);
748 if (status
!= Z_STREAM_END
) {
749 device_printf(sc
->dev
, "zlib %d\n", status
);
751 goto abort_with_buffer
;
754 fw_len
= fw
->fw_imglen
;
756 hdr_offset
= htobe32(*(const uint32_t *)
757 (fw
->fw_image
+ MCP_HEADER_PTR_OFFSET
));
758 if ((hdr_offset
& 3) || hdr_offset
+ sizeof(*hdr
) > fw_len
) {
759 device_printf(sc
->dev
, "Bad firmware file");
763 hdr
= (const void*)(fw
->fw_image
+ hdr_offset
);
765 status
= mxge_validate_firmware(sc
, hdr
);
769 /* Copy the inflated firmware to NIC SRAM. */
770 for (i
= 0; i
< fw_len
; i
+= 256) {
771 mxge_pio_copy(sc
->sram
+ MXGE_FW_OFFSET
+ i
,
773 min(256U, (unsigned)(fw_len
- i
)));
783 kfree(inflate_buffer
, M_TEMP
);
788 firmware_image_unload(fw
);
793 * Enable or disable periodic RDMAs from the host to make certain
794 * chipsets resend dropped PCIe messages
798 mxge_dummy_rdma(mxge_softc_t
*sc
, int enable
)
801 volatile uint32_t *confirm
;
802 volatile char *submit
;
803 uint32_t *buf
, dma_low
, dma_high
;
806 buf
= (uint32_t *)((unsigned long)(buf_bytes
+ 7) & ~7UL);
808 /* clear confirmation addr */
809 confirm
= (volatile uint32_t *)sc
->cmd
;
813 /* send an rdma command to the PCIe engine, and wait for the
814 response in the confirmation address. The firmware should
815 write a -1 there to indicate it is alive and well
818 dma_low
= MXGE_LOWPART_TO_U32(sc
->cmd_dma
.bus_addr
);
819 dma_high
= MXGE_HIGHPART_TO_U32(sc
->cmd_dma
.bus_addr
);
820 buf
[0] = htobe32(dma_high
); /* confirm addr MSW */
821 buf
[1] = htobe32(dma_low
); /* confirm addr LSW */
822 buf
[2] = htobe32(0xffffffff); /* confirm data */
823 dma_low
= MXGE_LOWPART_TO_U32(sc
->zeropad_dma
.bus_addr
);
824 dma_high
= MXGE_HIGHPART_TO_U32(sc
->zeropad_dma
.bus_addr
);
825 buf
[3] = htobe32(dma_high
); /* dummy addr MSW */
826 buf
[4] = htobe32(dma_low
); /* dummy addr LSW */
827 buf
[5] = htobe32(enable
); /* enable? */
830 submit
= (volatile char *)(sc
->sram
+ MXGEFW_BOOT_DUMMY_RDMA
);
832 mxge_pio_copy(submit
, buf
, 64);
837 while (*confirm
!= 0xffffffff && i
< 20) {
841 if (*confirm
!= 0xffffffff) {
842 device_printf(sc
->dev
, "dummy rdma %s failed (%p = 0x%x)",
843 (enable
? "enable" : "disable"), confirm
,
850 mxge_send_cmd(mxge_softc_t
*sc
, uint32_t cmd
, mxge_cmd_t
*data
)
853 char buf_bytes
[sizeof(*buf
) + 8];
854 volatile mcp_cmd_response_t
*response
= sc
->cmd
;
855 volatile char *cmd_addr
= sc
->sram
+ MXGEFW_ETH_CMD
;
856 uint32_t dma_low
, dma_high
;
857 int err
, sleep_total
= 0;
860 * We may be called during attach, before if_serializer is available.
861 * This is not a fast path, just check for NULL
864 if (sc
->ifp
->if_serializer
)
865 ASSERT_SERIALIZED(sc
->ifp
->if_serializer
);
867 /* ensure buf is aligned to 8 bytes */
868 buf
= (mcp_cmd_t
*)((unsigned long)(buf_bytes
+ 7) & ~7UL);
870 buf
->data0
= htobe32(data
->data0
);
871 buf
->data1
= htobe32(data
->data1
);
872 buf
->data2
= htobe32(data
->data2
);
873 buf
->cmd
= htobe32(cmd
);
874 dma_low
= MXGE_LOWPART_TO_U32(sc
->cmd_dma
.bus_addr
);
875 dma_high
= MXGE_HIGHPART_TO_U32(sc
->cmd_dma
.bus_addr
);
877 buf
->response_addr
.low
= htobe32(dma_low
);
878 buf
->response_addr
.high
= htobe32(dma_high
);
881 response
->result
= 0xffffffff;
883 mxge_pio_copy((volatile void *)cmd_addr
, buf
, sizeof (*buf
));
885 /* wait up to 20ms */
887 for (sleep_total
= 0; sleep_total
< 20; sleep_total
++) {
888 bus_dmamap_sync(sc
->cmd_dma
.dmat
,
889 sc
->cmd_dma
.map
, BUS_DMASYNC_POSTREAD
);
891 switch (be32toh(response
->result
)) {
893 data
->data0
= be32toh(response
->data
);
899 case MXGEFW_CMD_UNKNOWN
:
902 case MXGEFW_CMD_ERROR_UNALIGNED
:
905 case MXGEFW_CMD_ERROR_BUSY
:
909 device_printf(sc
->dev
,
911 "failed, result = %d\n",
912 cmd
, be32toh(response
->result
));
920 device_printf(sc
->dev
, "mxge: command %d timed out"
922 cmd
, be32toh(response
->result
));
927 mxge_adopt_running_firmware(mxge_softc_t
*sc
)
929 struct mcp_gen_header
*hdr
;
930 const size_t bytes
= sizeof (struct mcp_gen_header
);
934 /* find running firmware header */
935 hdr_offset
= htobe32(*(volatile uint32_t *)
936 (sc
->sram
+ MCP_HEADER_PTR_OFFSET
));
938 if ((hdr_offset
& 3) || hdr_offset
+ sizeof(*hdr
) > sc
->sram_size
) {
939 device_printf(sc
->dev
,
940 "Running firmware has bad header offset (%d)\n",
945 /* copy header of running firmware from SRAM to host memory to
946 * validate firmware */
947 hdr
= kmalloc(bytes
, M_DEVBUF
, M_NOWAIT
);
949 device_printf(sc
->dev
, "could not kmalloc firmware hdr\n");
952 bus_space_read_region_1(rman_get_bustag(sc
->mem_res
),
953 rman_get_bushandle(sc
->mem_res
),
954 hdr_offset
, (char *)hdr
, bytes
);
955 status
= mxge_validate_firmware(sc
, hdr
);
956 kfree(hdr
, M_DEVBUF
);
959 * check to see if adopted firmware has bug where adopting
960 * it will cause broadcasts to be filtered unless the NIC
961 * is kept in ALLMULTI mode
963 if (sc
->fw_ver_major
== 1 && sc
->fw_ver_minor
== 4 &&
964 sc
->fw_ver_tiny
>= 4 && sc
->fw_ver_tiny
<= 11) {
965 sc
->adopted_rx_filter_bug
= 1;
966 device_printf(sc
->dev
, "Adopting fw %d.%d.%d: "
967 "working around rx filter bug\n",
968 sc
->fw_ver_major
, sc
->fw_ver_minor
,
977 mxge_load_firmware(mxge_softc_t
*sc
, int adopt
)
979 volatile uint32_t *confirm
;
980 volatile char *submit
;
982 uint32_t *buf
, size
, dma_low
, dma_high
;
985 buf
= (uint32_t *)((unsigned long)(buf_bytes
+ 7) & ~7UL);
987 size
= sc
->sram_size
;
988 status
= mxge_load_firmware_helper(sc
, &size
);
992 /* Try to use the currently running firmware, if
994 status
= mxge_adopt_running_firmware(sc
);
996 device_printf(sc
->dev
,
997 "failed to adopt running firmware\n");
1000 device_printf(sc
->dev
,
1001 "Successfully adopted running firmware\n");
1002 if (sc
->tx_boundary
== 4096) {
1003 device_printf(sc
->dev
,
1004 "Using firmware currently running on NIC"
1006 device_printf(sc
->dev
,
1007 "performance consider loading optimized "
1010 sc
->fw_name
= mxge_fw_unaligned
;
1011 sc
->tx_boundary
= 2048;
1014 /* clear confirmation addr */
1015 confirm
= (volatile uint32_t *)sc
->cmd
;
1018 /* send a reload command to the bootstrap MCP, and wait for the
1019 response in the confirmation address. The firmware should
1020 write a -1 there to indicate it is alive and well
1023 dma_low
= MXGE_LOWPART_TO_U32(sc
->cmd_dma
.bus_addr
);
1024 dma_high
= MXGE_HIGHPART_TO_U32(sc
->cmd_dma
.bus_addr
);
1026 buf
[0] = htobe32(dma_high
); /* confirm addr MSW */
1027 buf
[1] = htobe32(dma_low
); /* confirm addr LSW */
1028 buf
[2] = htobe32(0xffffffff); /* confirm data */
1030 /* FIX: All newest firmware should un-protect the bottom of
1031 the sram before handoff. However, the very first interfaces
1032 do not. Therefore the handoff copy must skip the first 8 bytes
1034 /* where the code starts*/
1035 buf
[3] = htobe32(MXGE_FW_OFFSET
+ 8);
1036 buf
[4] = htobe32(size
- 8); /* length of code */
1037 buf
[5] = htobe32(8); /* where to copy to */
1038 buf
[6] = htobe32(0); /* where to jump to */
1040 submit
= (volatile char *)(sc
->sram
+ MXGEFW_BOOT_HANDOFF
);
1041 mxge_pio_copy(submit
, buf
, 64);
1046 while (*confirm
!= 0xffffffff && i
< 20) {
1049 bus_dmamap_sync(sc
->cmd_dma
.dmat
,
1050 sc
->cmd_dma
.map
, BUS_DMASYNC_POSTREAD
);
1052 if (*confirm
!= 0xffffffff) {
1053 device_printf(sc
->dev
,"handoff failed (%p = 0x%x)",
1062 mxge_update_mac_address(mxge_softc_t
*sc
)
1065 uint8_t *addr
= sc
->mac_addr
;
1069 cmd
.data0
= ((addr
[0] << 24) | (addr
[1] << 16)
1070 | (addr
[2] << 8) | addr
[3]);
1072 cmd
.data1
= ((addr
[4] << 8) | (addr
[5]));
1074 status
= mxge_send_cmd(sc
, MXGEFW_SET_MAC_ADDRESS
, &cmd
);
1079 mxge_change_pause(mxge_softc_t
*sc
, int pause
)
1085 status
= mxge_send_cmd(sc
, MXGEFW_ENABLE_FLOW_CONTROL
,
1088 status
= mxge_send_cmd(sc
, MXGEFW_DISABLE_FLOW_CONTROL
,
1092 device_printf(sc
->dev
, "Failed to set flow control mode\n");
1100 mxge_change_promisc(mxge_softc_t
*sc
, int promisc
)
1105 if( sc
->ifp
->if_serializer
)
1106 ASSERT_SERIALIZED(sc
->ifp
->if_serializer
);
1107 if (mxge_always_promisc
)
1111 status
= mxge_send_cmd(sc
, MXGEFW_ENABLE_PROMISC
,
1114 status
= mxge_send_cmd(sc
, MXGEFW_DISABLE_PROMISC
,
1118 device_printf(sc
->dev
, "Failed to set promisc mode\n");
1123 mxge_set_multicast_list(mxge_softc_t
*sc
)
1126 struct ifmultiaddr
*ifma
;
1127 struct ifnet
*ifp
= sc
->ifp
;
1130 if (ifp
->if_serializer
)
1131 ASSERT_SERIALIZED(ifp
->if_serializer
);
1133 /* This firmware is known to not support multicast */
1134 if (!sc
->fw_multicast_support
)
1137 /* Disable multicast filtering while we play with the lists*/
1138 err
= mxge_send_cmd(sc
, MXGEFW_ENABLE_ALLMULTI
, &cmd
);
1140 device_printf(sc
->dev
, "Failed MXGEFW_ENABLE_ALLMULTI,"
1141 " error status: %d\n", err
);
1145 if (sc
->adopted_rx_filter_bug
)
1148 if (ifp
->if_flags
& IFF_ALLMULTI
)
1149 /* request to disable multicast filtering, so quit here */
1152 /* Flush all the filters */
1154 err
= mxge_send_cmd(sc
, MXGEFW_LEAVE_ALL_MULTICAST_GROUPS
, &cmd
);
1156 device_printf(sc
->dev
,
1157 "Failed MXGEFW_LEAVE_ALL_MULTICAST_GROUPS"
1158 ", error status: %d\n", err
);
1162 /* Walk the multicast list, and add each address */
1164 LIST_FOREACH(ifma
, &ifp
->if_multiaddrs
, ifma_link
) {
1165 if (ifma
->ifma_addr
->sa_family
!= AF_LINK
)
1167 bcopy(LLADDR((struct sockaddr_dl
*)ifma
->ifma_addr
),
1169 bcopy(LLADDR((struct sockaddr_dl
*)ifma
->ifma_addr
) + 4,
1171 cmd
.data0
= htonl(cmd
.data0
);
1172 cmd
.data1
= htonl(cmd
.data1
);
1173 err
= mxge_send_cmd(sc
, MXGEFW_JOIN_MULTICAST_GROUP
, &cmd
);
1175 device_printf(sc
->dev
, "Failed "
1176 "MXGEFW_JOIN_MULTICAST_GROUP, error status:"
1178 /* abort, leaving multicast filtering off */
1182 /* Enable multicast filtering */
1183 err
= mxge_send_cmd(sc
, MXGEFW_DISABLE_ALLMULTI
, &cmd
);
1185 device_printf(sc
->dev
, "Failed MXGEFW_DISABLE_ALLMULTI"
1186 ", error status: %d\n", err
);
1191 mxge_max_mtu(mxge_softc_t
*sc
)
1196 if (MJUMPAGESIZE
- MXGEFW_PAD
> MXGEFW_MAX_MTU
)
1197 return MXGEFW_MAX_MTU
- MXGEFW_PAD
;
1199 /* try to set nbufs to see if it we can
1200 use virtually contiguous jumbos */
1202 status
= mxge_send_cmd(sc
, MXGEFW_CMD_ALWAYS_USE_N_BIG_BUFFERS
,
1205 return MXGEFW_MAX_MTU
- MXGEFW_PAD
;
1207 /* otherwise, we're limited to MJUMPAGESIZE */
1208 return MJUMPAGESIZE
- MXGEFW_PAD
;
1212 mxge_reset(mxge_softc_t
*sc
, int interrupts_setup
)
1214 struct mxge_slice_state
*ss
;
1215 mxge_rx_done_t
*rx_done
;
1216 volatile uint32_t *irq_claim
;
1220 /* try to send a reset command to the card to see if it
1222 memset(&cmd
, 0, sizeof (cmd
));
1223 status
= mxge_send_cmd(sc
, MXGEFW_CMD_RESET
, &cmd
);
1225 device_printf(sc
->dev
, "failed reset\n");
1229 mxge_dummy_rdma(sc
, 1);
1232 /* set the intrq size */
1233 cmd
.data0
= sc
->rx_ring_size
;
1234 status
= mxge_send_cmd(sc
, MXGEFW_CMD_SET_INTRQ_SIZE
, &cmd
);
1237 * Even though we already know how many slices are supported
1238 * via mxge_slice_probe(), MXGEFW_CMD_GET_MAX_RSS_QUEUES
1239 * has magic side effects, and must be called after a reset.
1240 * It must be called prior to calling any RSS related cmds,
1241 * including assigning an interrupt queue for anything but
1242 * slice 0. It must also be called *after*
1243 * MXGEFW_CMD_SET_INTRQ_SIZE, since the intrq size is used by
1244 * the firmware to compute offsets.
1247 if (sc
->num_slices
> 1) {
1248 /* ask the maximum number of slices it supports */
1249 status
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_MAX_RSS_QUEUES
,
1252 device_printf(sc
->dev
,
1253 "failed to get number of slices\n");
1257 * MXGEFW_CMD_ENABLE_RSS_QUEUES must be called prior
1258 * to setting up the interrupt queue DMA
1260 cmd
.data0
= sc
->num_slices
;
1261 cmd
.data1
= MXGEFW_SLICE_INTR_MODE_ONE_PER_SLICE
;
1262 #ifdef IFNET_BUF_RING
1263 cmd
.data1
|= MXGEFW_SLICE_ENABLE_MULTIPLE_TX_QUEUES
;
1265 status
= mxge_send_cmd(sc
, MXGEFW_CMD_ENABLE_RSS_QUEUES
,
1268 device_printf(sc
->dev
,
1269 "failed to set number of slices\n");
1275 if (interrupts_setup
) {
1276 /* Now exchange information about interrupts */
1277 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
1278 rx_done
= &sc
->ss
[slice
].rx_done
;
1279 memset(rx_done
->entry
, 0, sc
->rx_ring_size
);
1280 cmd
.data0
= MXGE_LOWPART_TO_U32(rx_done
->dma
.bus_addr
);
1281 cmd
.data1
= MXGE_HIGHPART_TO_U32(rx_done
->dma
.bus_addr
);
1283 status
|= mxge_send_cmd(sc
,
1284 MXGEFW_CMD_SET_INTRQ_DMA
,
1289 status
|= mxge_send_cmd(sc
,
1290 MXGEFW_CMD_GET_INTR_COAL_DELAY_OFFSET
, &cmd
);
1293 sc
->intr_coal_delay_ptr
= (volatile uint32_t *)(sc
->sram
+ cmd
.data0
);
1295 status
|= mxge_send_cmd(sc
, MXGEFW_CMD_GET_IRQ_ACK_OFFSET
, &cmd
);
1296 irq_claim
= (volatile uint32_t *)(sc
->sram
+ cmd
.data0
);
1299 status
|= mxge_send_cmd(sc
, MXGEFW_CMD_GET_IRQ_DEASSERT_OFFSET
,
1301 sc
->irq_deassert
= (volatile uint32_t *)(sc
->sram
+ cmd
.data0
);
1303 device_printf(sc
->dev
, "failed set interrupt parameters\n");
1308 *sc
->intr_coal_delay_ptr
= htobe32(sc
->intr_coal_delay
);
1311 /* run a DMA benchmark */
1312 (void) mxge_dma_test(sc
, MXGEFW_DMA_TEST
);
1314 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
1315 ss
= &sc
->ss
[slice
];
1317 ss
->irq_claim
= irq_claim
+ (2 * slice
);
1318 /* reset mcp/driver shared state back to 0 */
1319 ss
->rx_done
.idx
= 0;
1320 ss
->rx_done
.cnt
= 0;
1323 ss
->tx
.pkt_done
= 0;
1324 ss
->tx
.queue_active
= 0;
1325 ss
->tx
.activate
= 0;
1326 ss
->tx
.deactivate
= 0;
1331 ss
->rx_small
.cnt
= 0;
1332 ss
->lro_bad_csum
= 0;
1334 ss
->lro_flushed
= 0;
1335 if (ss
->fw_stats
!= NULL
) {
1336 ss
->fw_stats
->valid
= 0;
1337 ss
->fw_stats
->send_done_count
= 0;
1340 sc
->rdma_tags_available
= 15;
1341 status
= mxge_update_mac_address(sc
);
1342 mxge_change_promisc(sc
, sc
->ifp
->if_flags
& IFF_PROMISC
);
1343 mxge_change_pause(sc
, sc
->pause
);
1344 mxge_set_multicast_list(sc
);
1349 mxge_change_intr_coal(SYSCTL_HANDLER_ARGS
)
1352 unsigned int intr_coal_delay
;
1356 intr_coal_delay
= sc
->intr_coal_delay
;
1357 err
= sysctl_handle_int(oidp
, &intr_coal_delay
, arg2
, req
);
1361 if (intr_coal_delay
== sc
->intr_coal_delay
)
1364 if (intr_coal_delay
== 0 || intr_coal_delay
> 1000*1000)
1367 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
1368 *sc
->intr_coal_delay_ptr
= htobe32(intr_coal_delay
);
1369 sc
->intr_coal_delay
= intr_coal_delay
;
1371 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
1376 mxge_change_flow_control(SYSCTL_HANDLER_ARGS
)
1379 unsigned int enabled
;
1383 enabled
= sc
->pause
;
1384 err
= sysctl_handle_int(oidp
, &enabled
, arg2
, req
);
1388 if (enabled
== sc
->pause
)
1391 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
1392 err
= mxge_change_pause(sc
, enabled
);
1393 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
1398 mxge_change_lro_locked(mxge_softc_t
*sc
, int lro_cnt
)
1405 ifp
->if_capenable
&= ~IFCAP_LRO
;
1407 ifp
->if_capenable
|= IFCAP_LRO
;
1408 sc
->lro_cnt
= lro_cnt
;
1409 if (ifp
->if_flags
& IFF_RUNNING
) {
1411 err
= mxge_open(sc
);
1417 mxge_change_lro(SYSCTL_HANDLER_ARGS
)
1420 unsigned int lro_cnt
;
1424 lro_cnt
= sc
->lro_cnt
;
1425 err
= sysctl_handle_int(oidp
, &lro_cnt
, arg2
, req
);
1429 if (lro_cnt
== sc
->lro_cnt
)
1435 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
1436 err
= mxge_change_lro_locked(sc
, lro_cnt
);
1437 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
1442 mxge_handle_be32(SYSCTL_HANDLER_ARGS
)
1448 arg2
= be32toh(*(int *)arg1
);
1450 err
= sysctl_handle_int(oidp
, arg1
, arg2
, req
);
1456 mxge_rem_sysctls(mxge_softc_t
*sc
)
1458 struct mxge_slice_state
*ss
;
1461 if (sc
->slice_sysctl_tree
== NULL
)
1464 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
1465 ss
= &sc
->ss
[slice
];
1466 if (ss
== NULL
|| ss
->sysctl_tree
== NULL
)
1468 sysctl_ctx_free(&ss
->sysctl_ctx
);
1469 ss
->sysctl_tree
= NULL
;
1471 sysctl_ctx_free(&sc
->slice_sysctl_ctx
);
1472 sc
->slice_sysctl_tree
= NULL
;
1473 sysctl_ctx_free(&sc
->sysctl_ctx
);
1474 sc
->sysctl_tree
= NULL
;
1479 mxge_add_sysctls(mxge_softc_t
*sc
)
1481 struct sysctl_ctx_list
*ctx
;
1482 struct sysctl_oid_list
*children
;
1484 struct mxge_slice_state
*ss
;
1488 ctx
= &sc
->sysctl_ctx
;
1489 sysctl_ctx_init(ctx
);
1490 sc
->sysctl_tree
= SYSCTL_ADD_NODE(ctx
, SYSCTL_STATIC_CHILDREN(_hw
),
1492 device_get_nameunit(sc
->dev
),
1494 if (sc
->sysctl_tree
== NULL
) {
1495 device_printf(sc
->dev
, "can't add sysctl node\n");
1499 children
= SYSCTL_CHILDREN(sc
->sysctl_tree
);
1500 fw
= sc
->ss
[0].fw_stats
;
1502 /* random information */
1503 SYSCTL_ADD_STRING(ctx
, children
, OID_AUTO
,
1505 CTLFLAG_RD
, &sc
->fw_version
,
1506 0, "firmware version");
1507 SYSCTL_ADD_STRING(ctx
, children
, OID_AUTO
,
1509 CTLFLAG_RD
, &sc
->serial_number_string
,
1510 0, "serial number");
1511 SYSCTL_ADD_STRING(ctx
, children
, OID_AUTO
,
1513 CTLFLAG_RD
, &sc
->product_code_string
,
1515 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1517 CTLFLAG_RD
, &sc
->link_width
,
1519 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1521 CTLFLAG_RD
, &sc
->tx_boundary
,
1523 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1525 CTLFLAG_RD
, &sc
->wc
,
1526 0, "write combining PIO?");
1527 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1529 CTLFLAG_RD
, &sc
->read_dma
,
1530 0, "DMA Read speed in MB/s");
1531 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1533 CTLFLAG_RD
, &sc
->write_dma
,
1534 0, "DMA Write speed in MB/s");
1535 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1536 "read_write_dma_MBs",
1537 CTLFLAG_RD
, &sc
->read_write_dma
,
1538 0, "DMA concurrent Read/Write speed in MB/s");
1541 /* performance related tunables */
1542 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1544 CTLTYPE_INT
|CTLFLAG_RW
, sc
,
1545 0, mxge_change_intr_coal
,
1546 "I", "interrupt coalescing delay in usecs");
1548 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1549 "flow_control_enabled",
1550 CTLTYPE_INT
|CTLFLAG_RW
, sc
,
1551 0, mxge_change_flow_control
,
1552 "I", "interrupt coalescing delay in usecs");
1554 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1556 CTLFLAG_RW
, &mxge_deassert_wait
,
1557 0, "Wait for IRQ line to go low in ihandler");
1559 /* stats block from firmware is in network byte order.
1561 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1563 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->link_up
,
1564 0, mxge_handle_be32
,
1566 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1567 "rdma_tags_available",
1568 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->rdma_tags_available
,
1569 0, mxge_handle_be32
,
1570 "I", "rdma_tags_available");
1571 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1572 "dropped_bad_crc32",
1573 CTLTYPE_INT
|CTLFLAG_RD
,
1574 &fw
->dropped_bad_crc32
,
1575 0, mxge_handle_be32
,
1576 "I", "dropped_bad_crc32");
1577 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1579 CTLTYPE_INT
|CTLFLAG_RD
,
1580 &fw
->dropped_bad_phy
,
1581 0, mxge_handle_be32
,
1582 "I", "dropped_bad_phy");
1583 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1584 "dropped_link_error_or_filtered",
1585 CTLTYPE_INT
|CTLFLAG_RD
,
1586 &fw
->dropped_link_error_or_filtered
,
1587 0, mxge_handle_be32
,
1588 "I", "dropped_link_error_or_filtered");
1589 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1590 "dropped_link_overflow",
1591 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->dropped_link_overflow
,
1592 0, mxge_handle_be32
,
1593 "I", "dropped_link_overflow");
1594 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1595 "dropped_multicast_filtered",
1596 CTLTYPE_INT
|CTLFLAG_RD
,
1597 &fw
->dropped_multicast_filtered
,
1598 0, mxge_handle_be32
,
1599 "I", "dropped_multicast_filtered");
1600 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1601 "dropped_no_big_buffer",
1602 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->dropped_no_big_buffer
,
1603 0, mxge_handle_be32
,
1604 "I", "dropped_no_big_buffer");
1605 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1606 "dropped_no_small_buffer",
1607 CTLTYPE_INT
|CTLFLAG_RD
,
1608 &fw
->dropped_no_small_buffer
,
1609 0, mxge_handle_be32
,
1610 "I", "dropped_no_small_buffer");
1611 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1613 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->dropped_overrun
,
1614 0, mxge_handle_be32
,
1615 "I", "dropped_overrun");
1616 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1618 CTLTYPE_INT
|CTLFLAG_RD
,
1620 0, mxge_handle_be32
,
1621 "I", "dropped_pause");
1622 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1624 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->dropped_runt
,
1625 0, mxge_handle_be32
,
1626 "I", "dropped_runt");
1628 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1629 "dropped_unicast_filtered",
1630 CTLTYPE_INT
|CTLFLAG_RD
, &fw
->dropped_unicast_filtered
,
1631 0, mxge_handle_be32
,
1632 "I", "dropped_unicast_filtered");
1634 /* verbose printing? */
1635 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1637 CTLFLAG_RW
, &mxge_verbose
,
1638 0, "verbose printing");
1641 SYSCTL_ADD_PROC(ctx
, children
, OID_AUTO
,
1643 CTLTYPE_INT
|CTLFLAG_RW
, sc
,
1645 "I", "number of lro merge queues");
1648 /* add counters exported for debugging from all slices */
1649 sysctl_ctx_init(&sc
->slice_sysctl_ctx
);
1650 sc
->slice_sysctl_tree
=
1651 SYSCTL_ADD_NODE(&sc
->slice_sysctl_ctx
, children
, OID_AUTO
,
1652 "slice", CTLFLAG_RD
, 0, "");
1654 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
1655 ss
= &sc
->ss
[slice
];
1656 sysctl_ctx_init(&ss
->sysctl_ctx
);
1657 ctx
= &ss
->sysctl_ctx
;
1658 children
= SYSCTL_CHILDREN(sc
->slice_sysctl_tree
);
1659 ksprintf(slice_num
, "%d", slice
);
1661 SYSCTL_ADD_NODE(ctx
, children
, OID_AUTO
, slice_num
,
1663 children
= SYSCTL_CHILDREN(ss
->sysctl_tree
);
1664 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1666 CTLFLAG_RD
, &ss
->rx_small
.cnt
,
1668 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1670 CTLFLAG_RD
, &ss
->rx_big
.cnt
,
1672 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1673 "lro_flushed", CTLFLAG_RD
, &ss
->lro_flushed
,
1674 0, "number of lro merge queues flushed");
1676 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1677 "lro_queued", CTLFLAG_RD
, &ss
->lro_queued
,
1678 0, "number of frames appended to lro merge"
1681 #ifndef IFNET_BUF_RING
1682 /* only transmit from slice 0 for now */
1686 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1688 CTLFLAG_RD
, &ss
->tx
.req
,
1691 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1693 CTLFLAG_RD
, &ss
->tx
.done
,
1695 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1697 CTLFLAG_RD
, &ss
->tx
.pkt_done
,
1699 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1701 CTLFLAG_RD
, &ss
->tx
.stall
,
1703 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1705 CTLFLAG_RD
, &ss
->tx
.wake
,
1707 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1709 CTLFLAG_RD
, &ss
->tx
.defrag
,
1711 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1713 CTLFLAG_RD
, &ss
->tx
.queue_active
,
1714 0, "tx_queue_active");
1715 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1717 CTLFLAG_RD
, &ss
->tx
.activate
,
1719 SYSCTL_ADD_INT(ctx
, children
, OID_AUTO
,
1721 CTLFLAG_RD
, &ss
->tx
.deactivate
,
1722 0, "tx_deactivate");
1726 /* copy an array of mcp_kreq_ether_send_t's to the mcp. Copy
1727 backwards one at a time and handle ring wraps */
1730 mxge_submit_req_backwards(mxge_tx_ring_t
*tx
,
1731 mcp_kreq_ether_send_t
*src
, int cnt
)
1733 int idx
, starting_slot
;
1734 starting_slot
= tx
->req
;
1737 idx
= (starting_slot
+ cnt
) & tx
->mask
;
1738 mxge_pio_copy(&tx
->lanai
[idx
],
1739 &src
[cnt
], sizeof(*src
));
1745 * copy an array of mcp_kreq_ether_send_t's to the mcp. Copy
1746 * at most 32 bytes at a time, so as to avoid involving the software
1747 * pio handler in the nic. We re-write the first segment's flags
1748 * to mark them valid only after writing the entire chain
1752 mxge_submit_req(mxge_tx_ring_t
*tx
, mcp_kreq_ether_send_t
*src
,
1757 volatile uint32_t *dst_ints
;
1758 mcp_kreq_ether_send_t
*srcp
;
1759 volatile mcp_kreq_ether_send_t
*dstp
, *dst
;
1762 idx
= tx
->req
& tx
->mask
;
1764 last_flags
= src
->flags
;
1767 dst
= dstp
= &tx
->lanai
[idx
];
1770 if ((idx
+ cnt
) < tx
->mask
) {
1771 for (i
= 0; i
< (cnt
- 1); i
+= 2) {
1772 mxge_pio_copy(dstp
, srcp
, 2 * sizeof(*src
));
1773 wmb(); /* force write every 32 bytes */
1778 /* submit all but the first request, and ensure
1779 that it is submitted below */
1780 mxge_submit_req_backwards(tx
, src
, cnt
);
1784 /* submit the first request */
1785 mxge_pio_copy(dstp
, srcp
, sizeof(*src
));
1786 wmb(); /* barrier before setting valid flag */
1789 /* re-write the last 32-bits with the valid flags */
1790 src
->flags
= last_flags
;
1791 src_ints
= (uint32_t *)src
;
1793 dst_ints
= (volatile uint32_t *)dst
;
1795 *dst_ints
= *src_ints
;
1803 mxge_encap_tso(struct mxge_slice_state
*ss
, struct mbuf
*m
,
1804 int busdma_seg_cnt
, int ip_off
)
1807 mcp_kreq_ether_send_t
*req
;
1808 bus_dma_segment_t
*seg
;
1811 uint32_t low
, high_swapped
;
1812 int len
, seglen
, cum_len
, cum_len_next
;
1813 int next_is_first
, chop
, cnt
, rdma_count
, small
;
1814 uint16_t pseudo_hdr_offset
, cksum_offset
, mss
;
1815 uint8_t flags
, flags_next
;
1818 mss
= m
->m_pkthdr
.tso_segsz
;
1820 /* negative cum_len signifies to the
1821 * send loop that we are still in the
1822 * header portion of the TSO packet.
1825 /* ensure we have the ethernet, IP and TCP
1826 header together in the first mbuf, copy
1827 it to a scratch buffer if not */
1828 if (__predict_false(m
->m_len
< ip_off
+ sizeof (*ip
))) {
1829 m_copydata(m
, 0, ip_off
+ sizeof (*ip
),
1831 ip
= (struct ip
*)(ss
->scratch
+ ip_off
);
1833 ip
= (struct ip
*)(mtod(m
, char *) + ip_off
);
1835 if (__predict_false(m
->m_len
< ip_off
+ (ip
->ip_hl
<< 2)
1837 m_copydata(m
, 0, ip_off
+ (ip
->ip_hl
<< 2)
1838 + sizeof (*tcp
), ss
->scratch
);
1839 ip
= (struct ip
*)(mtod(m
, char *) + ip_off
);
1842 tcp
= (struct tcphdr
*)((char *)ip
+ (ip
->ip_hl
<< 2));
1843 cum_len
= -(ip_off
+ ((ip
->ip_hl
+ tcp
->th_off
) << 2));
1845 /* TSO implies checksum offload on this hardware */
1846 cksum_offset
= ip_off
+ (ip
->ip_hl
<< 2);
1847 flags
= MXGEFW_FLAGS_TSO_HDR
| MXGEFW_FLAGS_FIRST
;
1850 /* for TSO, pseudo_hdr_offset holds mss.
1851 * The firmware figures out where to put
1852 * the checksum by parsing the header. */
1853 pseudo_hdr_offset
= htobe16(mss
);
1860 /* "rdma_count" is the number of RDMAs belonging to the
1861 * current packet BEFORE the current send request. For
1862 * non-TSO packets, this is equal to "count".
1863 * For TSO packets, rdma_count needs to be reset
1864 * to 0 after a segment cut.
1866 * The rdma_count field of the send request is
1867 * the number of RDMAs of the packet starting at
1868 * that request. For TSO send requests with one ore more cuts
1869 * in the middle, this is the number of RDMAs starting
1870 * after the last cut in the request. All previous
1871 * segments before the last cut implicitly have 1 RDMA.
1873 * Since the number of RDMAs is not known beforehand,
1874 * it must be filled-in retroactively - after each
1875 * segmentation cut or at the end of the entire packet.
1878 while (busdma_seg_cnt
) {
1879 /* Break the busdma segment up into pieces*/
1880 low
= MXGE_LOWPART_TO_U32(seg
->ds_addr
);
1881 high_swapped
= htobe32(MXGE_HIGHPART_TO_U32(seg
->ds_addr
));
1885 flags_next
= flags
& ~MXGEFW_FLAGS_FIRST
;
1887 cum_len_next
= cum_len
+ seglen
;
1888 (req
-rdma_count
)->rdma_count
= rdma_count
+ 1;
1889 if (__predict_true(cum_len
>= 0)) {
1891 chop
= (cum_len_next
> mss
);
1892 cum_len_next
= cum_len_next
% mss
;
1893 next_is_first
= (cum_len_next
== 0);
1894 flags
|= chop
* MXGEFW_FLAGS_TSO_CHOP
;
1895 flags_next
|= next_is_first
*
1897 rdma_count
|= -(chop
| next_is_first
);
1898 rdma_count
+= chop
& !next_is_first
;
1899 } else if (cum_len_next
>= 0) {
1904 small
= (mss
<= MXGEFW_SEND_SMALL_SIZE
);
1905 flags_next
= MXGEFW_FLAGS_TSO_PLD
|
1906 MXGEFW_FLAGS_FIRST
|
1907 (small
* MXGEFW_FLAGS_SMALL
);
1910 req
->addr_high
= high_swapped
;
1911 req
->addr_low
= htobe32(low
);
1912 req
->pseudo_hdr_offset
= pseudo_hdr_offset
;
1914 req
->rdma_count
= 1;
1915 req
->length
= htobe16(seglen
);
1916 req
->cksum_offset
= cksum_offset
;
1917 req
->flags
= flags
| ((cum_len
& 1) *
1918 MXGEFW_FLAGS_ALIGN_ODD
);
1921 cum_len
= cum_len_next
;
1926 if (__predict_false(cksum_offset
> seglen
))
1927 cksum_offset
-= seglen
;
1930 if (__predict_false(cnt
> tx
->max_desc
))
1936 (req
-rdma_count
)->rdma_count
= rdma_count
;
1940 req
->flags
|= MXGEFW_FLAGS_TSO_LAST
;
1941 } while (!(req
->flags
& (MXGEFW_FLAGS_TSO_CHOP
| MXGEFW_FLAGS_FIRST
)));
1943 tx
->info
[((cnt
- 1) + tx
->req
) & tx
->mask
].flag
= 1;
1944 mxge_submit_req(tx
, tx
->req_list
, cnt
);
1945 #ifdef IFNET_BUF_RING
1946 if ((ss
->sc
->num_slices
> 1) && tx
->queue_active
== 0) {
1947 /* tell the NIC to start polling this slice */
1949 tx
->queue_active
= 1;
1957 bus_dmamap_unload(tx
->dmat
, tx
->info
[tx
->req
& tx
->mask
].map
);
1961 kprintf("tx->max_desc exceeded via TSO!\n");
1962 kprintf("mss = %d, %ld, %d!\n", mss
,
1963 (long)seg
- (long)tx
->seg_list
, tx
->max_desc
);
1970 #endif /* IFCAP_TSO4 */
1972 #ifdef MXGE_NEW_VLAN_API
1974 * We reproduce the software vlan tag insertion from
1975 * net/if_vlan.c:vlan_start() here so that we can advertise "hardware"
1976 * vlan tag insertion. We need to advertise this in order to have the
1977 * vlan interface respect our csum offload flags.
1979 static struct mbuf
*
1980 mxge_vlan_tag_insert(struct mbuf
*m
)
1982 struct ether_vlan_header
*evl
;
1984 M_PREPEND(m
, EVL_ENCAPLEN
, MB_DONTWAIT
);
1985 if (__predict_false(m
== NULL
))
1987 if (m
->m_len
< sizeof(*evl
)) {
1988 m
= m_pullup(m
, sizeof(*evl
));
1989 if (__predict_false(m
== NULL
))
1993 * Transform the Ethernet header into an Ethernet header
1994 * with 802.1Q encapsulation.
1996 evl
= mtod(m
, struct ether_vlan_header
*);
1997 bcopy((char *)evl
+ EVL_ENCAPLEN
,
1998 (char *)evl
, ETHER_HDR_LEN
- ETHER_TYPE_LEN
);
1999 evl
->evl_encap_proto
= htons(ETHERTYPE_VLAN
);
2000 evl
->evl_tag
= htons(m
->m_pkthdr
.ether_vlantag
);
2001 m
->m_flags
&= ~M_VLANTAG
;
2004 #endif /* MXGE_NEW_VLAN_API */
2007 mxge_encap(struct mxge_slice_state
*ss
, struct mbuf
*m
)
2010 mcp_kreq_ether_send_t
*req
;
2011 bus_dma_segment_t
*seg
;
2016 int cnt
, cum_len
, err
, i
, idx
, odd_flag
, ip_off
;
2017 uint16_t pseudo_hdr_offset
;
2018 uint8_t flags
, cksum_offset
;
2025 ip_off
= sizeof (struct ether_header
);
2026 #ifdef MXGE_NEW_VLAN_API
2027 if (m
->m_flags
& M_VLANTAG
) {
2028 m
= mxge_vlan_tag_insert(m
);
2029 if (__predict_false(m
== NULL
))
2031 ip_off
+= EVL_ENCAPLEN
;
2034 /* (try to) map the frame for DMA */
2035 idx
= tx
->req
& tx
->mask
;
2036 err
= bus_dmamap_load_mbuf_segment(tx
->dmat
, tx
->info
[idx
].map
,
2037 m
, tx
->seg_list
, 1, &cnt
,
2039 if (__predict_false(err
== EFBIG
)) {
2040 /* Too many segments in the chain. Try
2042 m_tmp
= m_defrag(m
, M_NOWAIT
);
2043 if (m_tmp
== NULL
) {
2048 err
= bus_dmamap_load_mbuf_segment(tx
->dmat
,
2050 m
, tx
->seg_list
, 1, &cnt
,
2053 if (__predict_false(err
!= 0)) {
2054 device_printf(sc
->dev
, "bus_dmamap_load_mbuf_segment returned %d"
2055 " packet len = %d\n", err
, m
->m_pkthdr
.len
);
2058 bus_dmamap_sync(tx
->dmat
, tx
->info
[idx
].map
,
2059 BUS_DMASYNC_PREWRITE
);
2060 tx
->info
[idx
].m
= m
;
2063 /* TSO is different enough, we handle it in another routine */
2064 if (m
->m_pkthdr
.csum_flags
& (CSUM_TSO
)) {
2065 mxge_encap_tso(ss
, m
, cnt
, ip_off
);
2072 pseudo_hdr_offset
= 0;
2073 flags
= MXGEFW_FLAGS_NO_TSO
;
2075 /* checksum offloading? */
2076 if (m
->m_pkthdr
.csum_flags
& (CSUM_DELAY_DATA
)) {
2077 /* ensure ip header is in first mbuf, copy
2078 it to a scratch buffer if not */
2079 if (__predict_false(m
->m_len
< ip_off
+ sizeof (*ip
))) {
2080 m_copydata(m
, 0, ip_off
+ sizeof (*ip
),
2082 ip
= (struct ip
*)(ss
->scratch
+ ip_off
);
2084 ip
= (struct ip
*)(mtod(m
, char *) + ip_off
);
2086 cksum_offset
= ip_off
+ (ip
->ip_hl
<< 2);
2087 pseudo_hdr_offset
= cksum_offset
+ m
->m_pkthdr
.csum_data
;
2088 pseudo_hdr_offset
= htobe16(pseudo_hdr_offset
);
2089 req
->cksum_offset
= cksum_offset
;
2090 flags
|= MXGEFW_FLAGS_CKSUM
;
2091 odd_flag
= MXGEFW_FLAGS_ALIGN_ODD
;
2095 if (m
->m_pkthdr
.len
< MXGEFW_SEND_SMALL_SIZE
)
2096 flags
|= MXGEFW_FLAGS_SMALL
;
2098 /* convert segments into a request list */
2101 req
->flags
= MXGEFW_FLAGS_FIRST
;
2102 for (i
= 0; i
< cnt
; i
++) {
2104 htobe32(MXGE_LOWPART_TO_U32(seg
->ds_addr
));
2106 htobe32(MXGE_HIGHPART_TO_U32(seg
->ds_addr
));
2107 req
->length
= htobe16(seg
->ds_len
);
2108 req
->cksum_offset
= cksum_offset
;
2109 if (cksum_offset
> seg
->ds_len
)
2110 cksum_offset
-= seg
->ds_len
;
2113 req
->pseudo_hdr_offset
= pseudo_hdr_offset
;
2114 req
->pad
= 0; /* complete solid 16-byte block */
2115 req
->rdma_count
= 1;
2116 req
->flags
|= flags
| ((cum_len
& 1) * odd_flag
);
2117 cum_len
+= seg
->ds_len
;
2123 /* pad runts to 60 bytes */
2127 htobe32(MXGE_LOWPART_TO_U32(sc
->zeropad_dma
.bus_addr
));
2129 htobe32(MXGE_HIGHPART_TO_U32(sc
->zeropad_dma
.bus_addr
));
2130 req
->length
= htobe16(60 - cum_len
);
2131 req
->cksum_offset
= 0;
2132 req
->pseudo_hdr_offset
= pseudo_hdr_offset
;
2133 req
->pad
= 0; /* complete solid 16-byte block */
2134 req
->rdma_count
= 1;
2135 req
->flags
|= flags
| ((cum_len
& 1) * odd_flag
);
2139 tx
->req_list
[0].rdma_count
= cnt
;
2141 /* print what the firmware will see */
2142 for (i
= 0; i
< cnt
; i
++) {
2143 kprintf("%d: addr: 0x%x 0x%x len:%d pso%d,"
2144 "cso:%d, flags:0x%x, rdma:%d\n",
2145 i
, (int)ntohl(tx
->req_list
[i
].addr_high
),
2146 (int)ntohl(tx
->req_list
[i
].addr_low
),
2147 (int)ntohs(tx
->req_list
[i
].length
),
2148 (int)ntohs(tx
->req_list
[i
].pseudo_hdr_offset
),
2149 tx
->req_list
[i
].cksum_offset
, tx
->req_list
[i
].flags
,
2150 tx
->req_list
[i
].rdma_count
);
2152 kprintf("--------------\n");
2154 tx
->info
[((cnt
- 1) + tx
->req
) & tx
->mask
].flag
= 1;
2155 mxge_submit_req(tx
, tx
->req_list
, cnt
);
2156 #ifdef IFNET_BUF_RING
2157 if ((ss
->sc
->num_slices
> 1) && tx
->queue_active
== 0) {
2158 /* tell the NIC to start polling this slice */
2160 tx
->queue_active
= 1;
2173 #ifdef IFNET_BUF_RING
2175 mxge_qflush(struct ifnet
*ifp
)
2177 mxge_softc_t
*sc
= ifp
->if_softc
;
2182 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
2183 tx
= &sc
->ss
[slice
].tx
;
2184 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
2185 while ((m
= buf_ring_dequeue_sc(tx
->br
)) != NULL
)
2187 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
2193 mxge_start_locked(struct mxge_slice_state
*ss
)
2204 while ((tx
->mask
- (tx
->req
- tx
->done
)) > tx
->max_desc
) {
2205 m
= drbr_dequeue(ifp
, tx
->br
);
2209 /* let BPF see it */
2212 /* give it to the nic */
2215 /* ran out of transmit slots */
2216 if (((ss
->if_flags
& IFF_OACTIVE
) == 0)
2217 && (!drbr_empty(ifp
, tx
->br
))) {
2218 ss
->if_flags
|= IFF_OACTIVE
;
2224 mxge_transmit_locked(struct mxge_slice_state
*ss
, struct mbuf
*m
)
2235 if ((ss
->if_flags
& (IFF_RUNNING
|IFF_OACTIVE
)) !=
2237 err
= drbr_enqueue(ifp
, tx
->br
, m
);
2241 if (drbr_empty(ifp
, tx
->br
) &&
2242 ((tx
->mask
- (tx
->req
- tx
->done
)) > tx
->max_desc
)) {
2243 /* let BPF see it */
2245 /* give it to the nic */
2247 } else if ((err
= drbr_enqueue(ifp
, tx
->br
, m
)) != 0) {
2250 if (!drbr_empty(ifp
, tx
->br
))
2251 mxge_start_locked(ss
);
2256 mxge_transmit(struct ifnet
*ifp
, struct mbuf
*m
)
2258 mxge_softc_t
*sc
= ifp
->if_softc
;
2259 struct mxge_slice_state
*ss
;
2265 slice
= m
->m_pkthdr
.flowid
;
2267 slice
&= (sc
->num_slices
- 1); /* num_slices always power of 2 */
2269 ss
= &sc
->ss
[slice
];
2272 if(lwkt_serialize_try(ifp
->if_serializer
)) {
2273 err
= mxge_transmit_locked(ss
, m
);
2274 lwkt_serialize_exit(ifp
->if_serializer
);
2276 err
= drbr_enqueue(ifp
, tx
->br
, m
);
2285 mxge_start_locked(struct mxge_slice_state
*ss
)
2295 while ((tx
->mask
- (tx
->req
- tx
->done
)) > tx
->max_desc
) {
2296 m
= ifq_dequeue(&ifp
->if_snd
, NULL
);
2300 /* let BPF see it */
2303 /* give it to the nic */
2306 /* ran out of transmit slots */
2307 if ((sc
->ifp
->if_flags
& IFF_OACTIVE
) == 0) {
2308 sc
->ifp
->if_flags
|= IFF_OACTIVE
;
2314 mxge_start(struct ifnet
*ifp
)
2316 mxge_softc_t
*sc
= ifp
->if_softc
;
2317 struct mxge_slice_state
*ss
;
2319 ASSERT_SERIALIZED(sc
->ifp
->if_serializer
);
2320 /* only use the first slice for now */
2322 mxge_start_locked(ss
);
2326 * copy an array of mcp_kreq_ether_recv_t's to the mcp. Copy
2327 * at most 32 bytes at a time, so as to avoid involving the software
2328 * pio handler in the nic. We re-write the first segment's low
2329 * DMA address to mark it valid only after we write the entire chunk
2333 mxge_submit_8rx(volatile mcp_kreq_ether_recv_t
*dst
,
2334 mcp_kreq_ether_recv_t
*src
)
2338 low
= src
->addr_low
;
2339 src
->addr_low
= 0xffffffff;
2340 mxge_pio_copy(dst
, src
, 4 * sizeof (*src
));
2342 mxge_pio_copy(dst
+ 4, src
+ 4, 4 * sizeof (*src
));
2344 src
->addr_low
= low
;
2345 dst
->addr_low
= low
;
2350 mxge_get_buf_small(struct mxge_slice_state
*ss
, bus_dmamap_t map
, int idx
)
2352 bus_dma_segment_t seg
;
2354 mxge_rx_ring_t
*rx
= &ss
->rx_small
;
2357 m
= m_gethdr(MB_DONTWAIT
, MT_DATA
);
2363 m
->m_len
= m
->m_pkthdr
.len
= MHLEN
;
2364 err
= bus_dmamap_load_mbuf_segment(rx
->dmat
, map
, m
,
2365 &seg
, 1, &cnt
, BUS_DMA_NOWAIT
);
2367 kprintf("can't dmamap small (%d)\n", err
);
2371 rx
->info
[idx
].m
= m
;
2372 rx
->shadow
[idx
].addr_low
=
2373 htobe32(MXGE_LOWPART_TO_U32(seg
.ds_addr
));
2374 rx
->shadow
[idx
].addr_high
=
2375 htobe32(MXGE_HIGHPART_TO_U32(seg
.ds_addr
));
2379 mxge_submit_8rx(&rx
->lanai
[idx
- 7], &rx
->shadow
[idx
- 7]);
2385 mxge_get_buf_big(struct mxge_slice_state
*ss
, bus_dmamap_t map
, int idx
)
2387 bus_dma_segment_t seg
[3];
2389 mxge_rx_ring_t
*rx
= &ss
->rx_big
;
2392 if (rx
->cl_size
== MCLBYTES
)
2393 m
= m_getcl(MB_DONTWAIT
, MT_DATA
, M_PKTHDR
);
2396 m
= m_getjcl(MB_DONTWAIT
, MT_DATA
, M_PKTHDR
, rx
->cl_size
);
2399 * XXX: allocate normal sized buffers for big buffers.
2400 * We should be fine as long as we don't get any jumbo frames
2402 m
= m_getcl(MB_DONTWAIT
, MT_DATA
, M_PKTHDR
);
2410 m
->m_pkthdr
.len
= 0;
2411 m
->m_len
= m
->m_pkthdr
.len
= rx
->mlen
;
2412 err
= bus_dmamap_load_mbuf_segment(rx
->dmat
, map
, m
,
2413 seg
, 1, &cnt
, BUS_DMA_NOWAIT
);
2415 kprintf("can't dmamap big (%d)\n", err
);
2419 rx
->info
[idx
].m
= m
;
2420 rx
->shadow
[idx
].addr_low
=
2421 htobe32(MXGE_LOWPART_TO_U32(seg
->ds_addr
));
2422 rx
->shadow
[idx
].addr_high
=
2423 htobe32(MXGE_HIGHPART_TO_U32(seg
->ds_addr
));
2425 #if MXGE_VIRT_JUMBOS
2426 for (i
= 1; i
< cnt
; i
++) {
2427 rx
->shadow
[idx
+ i
].addr_low
=
2428 htobe32(MXGE_LOWPART_TO_U32(seg
[i
].ds_addr
));
2429 rx
->shadow
[idx
+ i
].addr_high
=
2430 htobe32(MXGE_HIGHPART_TO_U32(seg
[i
].ds_addr
));
2435 for (i
= 0; i
< rx
->nbufs
; i
++) {
2436 if ((idx
& 7) == 7) {
2437 mxge_submit_8rx(&rx
->lanai
[idx
- 7],
2438 &rx
->shadow
[idx
- 7]);
2446 * Myri10GE hardware checksums are not valid if the sender
2447 * padded the frame with non-zero padding. This is because
2448 * the firmware just does a simple 16-bit 1s complement
2449 * checksum across the entire frame, excluding the first 14
2450 * bytes. It is best to simply to check the checksum and
2451 * tell the stack about it only if the checksum is good
2454 static inline uint16_t
2455 mxge_rx_csum(struct mbuf
*m
, int csum
)
2457 struct ether_header
*eh
;
2461 eh
= mtod(m
, struct ether_header
*);
2463 /* only deal with IPv4 TCP & UDP for now */
2464 if (__predict_false(eh
->ether_type
!= htons(ETHERTYPE_IP
)))
2466 ip
= (struct ip
*)(eh
+ 1);
2467 if (__predict_false(ip
->ip_p
!= IPPROTO_TCP
&&
2468 ip
->ip_p
!= IPPROTO_UDP
))
2471 c
= in_pseudo(ip
->ip_src
.s_addr
, ip
->ip_dst
.s_addr
,
2472 htonl(ntohs(csum
) + ntohs(ip
->ip_len
) +
2473 - (ip
->ip_hl
<< 2) + ip
->ip_p
));
2482 mxge_vlan_tag_remove(struct mbuf
*m
, uint32_t *csum
)
2484 struct ether_vlan_header
*evl
;
2485 struct ether_header
*eh
;
2488 evl
= mtod(m
, struct ether_vlan_header
*);
2489 eh
= mtod(m
, struct ether_header
*);
2492 * fix checksum by subtracting EVL_ENCAPLEN bytes
2493 * after what the firmware thought was the end of the ethernet
2497 /* put checksum into host byte order */
2498 *csum
= ntohs(*csum
);
2499 partial
= ntohl(*(uint32_t *)(mtod(m
, char *) + ETHER_HDR_LEN
));
2500 (*csum
) += ~partial
;
2501 (*csum
) += ((*csum
) < ~partial
);
2502 (*csum
) = ((*csum
) >> 16) + ((*csum
) & 0xFFFF);
2503 (*csum
) = ((*csum
) >> 16) + ((*csum
) & 0xFFFF);
2505 /* restore checksum to network byte order;
2506 later consumers expect this */
2507 *csum
= htons(*csum
);
2510 #ifdef MXGE_NEW_VLAN_API
2511 m
->m_pkthdr
.ether_vlantag
= ntohs(evl
->evl_tag
);
2515 mtag
= m_tag_alloc(MTAG_VLAN
, MTAG_VLAN_TAG
, sizeof(u_int
),
2519 VLAN_TAG_VALUE(mtag
) = ntohs(evl
->evl_tag
);
2520 m_tag_prepend(m
, mtag
);
2524 m
->m_flags
|= M_VLANTAG
;
2527 * Remove the 802.1q header by copying the Ethernet
2528 * addresses over it and adjusting the beginning of
2529 * the data in the mbuf. The encapsulated Ethernet
2530 * type field is already in place.
2532 bcopy((char *)evl
, (char *)evl
+ EVL_ENCAPLEN
,
2533 ETHER_HDR_LEN
- ETHER_TYPE_LEN
);
2534 m_adj(m
, EVL_ENCAPLEN
);
2539 mxge_rx_done_big(struct mxge_slice_state
*ss
, uint32_t len
, uint32_t csum
)
2544 struct ether_header
*eh
;
2546 bus_dmamap_t old_map
;
2548 uint16_t tcpudp_csum
;
2553 idx
= rx
->cnt
& rx
->mask
;
2554 rx
->cnt
+= rx
->nbufs
;
2555 /* save a pointer to the received mbuf */
2556 m
= rx
->info
[idx
].m
;
2557 /* try to replace the received mbuf */
2558 if (mxge_get_buf_big(ss
, rx
->extra_map
, idx
)) {
2559 /* drop the frame -- the old mbuf is re-cycled */
2564 /* unmap the received buffer */
2565 old_map
= rx
->info
[idx
].map
;
2566 bus_dmamap_sync(rx
->dmat
, old_map
, BUS_DMASYNC_POSTREAD
);
2567 bus_dmamap_unload(rx
->dmat
, old_map
);
2569 /* swap the bus_dmamap_t's */
2570 rx
->info
[idx
].map
= rx
->extra_map
;
2571 rx
->extra_map
= old_map
;
2573 /* mcp implicitly skips 1st 2 bytes so that packet is properly
2575 m
->m_data
+= MXGEFW_PAD
;
2577 m
->m_pkthdr
.rcvif
= ifp
;
2578 m
->m_len
= m
->m_pkthdr
.len
= len
;
2580 eh
= mtod(m
, struct ether_header
*);
2581 if (eh
->ether_type
== htons(ETHERTYPE_VLAN
)) {
2582 mxge_vlan_tag_remove(m
, &csum
);
2584 /* if the checksum is valid, mark it in the mbuf header */
2585 if (sc
->csum_flag
&& (0 == (tcpudp_csum
= mxge_rx_csum(m
, csum
)))) {
2586 if (sc
->lro_cnt
&& (0 == mxge_lro_rx(ss
, m
, csum
)))
2588 /* otherwise, it was a UDP frame, or a TCP frame which
2589 we could not do LRO on. Tell the stack that the
2591 m
->m_pkthdr
.csum_data
= 0xffff;
2592 m
->m_pkthdr
.csum_flags
= CSUM_PSEUDO_HDR
| CSUM_DATA_VALID
;
2595 /* flowid only valid if RSS hashing is enabled */
2596 if (sc
->num_slices
> 1) {
2597 m
->m_pkthdr
.flowid
= (ss
- sc
->ss
);
2598 m
->m_flags
|= M_FLOWID
;
2601 /* pass the frame up the stack */
2602 (*ifp
->if_input
)(ifp
, m
);
2606 mxge_rx_done_small(struct mxge_slice_state
*ss
, uint32_t len
, uint32_t csum
)
2610 struct ether_header
*eh
;
2613 bus_dmamap_t old_map
;
2615 uint16_t tcpudp_csum
;
2620 idx
= rx
->cnt
& rx
->mask
;
2622 /* save a pointer to the received mbuf */
2623 m
= rx
->info
[idx
].m
;
2624 /* try to replace the received mbuf */
2625 if (mxge_get_buf_small(ss
, rx
->extra_map
, idx
)) {
2626 /* drop the frame -- the old mbuf is re-cycled */
2631 /* unmap the received buffer */
2632 old_map
= rx
->info
[idx
].map
;
2633 bus_dmamap_sync(rx
->dmat
, old_map
, BUS_DMASYNC_POSTREAD
);
2634 bus_dmamap_unload(rx
->dmat
, old_map
);
2636 /* swap the bus_dmamap_t's */
2637 rx
->info
[idx
].map
= rx
->extra_map
;
2638 rx
->extra_map
= old_map
;
2640 /* mcp implicitly skips 1st 2 bytes so that packet is properly
2642 m
->m_data
+= MXGEFW_PAD
;
2644 m
->m_pkthdr
.rcvif
= ifp
;
2645 m
->m_len
= m
->m_pkthdr
.len
= len
;
2647 eh
= mtod(m
, struct ether_header
*);
2648 if (eh
->ether_type
== htons(ETHERTYPE_VLAN
)) {
2649 mxge_vlan_tag_remove(m
, &csum
);
2651 /* if the checksum is valid, mark it in the mbuf header */
2652 if (sc
->csum_flag
&& (0 == (tcpudp_csum
= mxge_rx_csum(m
, csum
)))) {
2653 if (sc
->lro_cnt
&& (0 == mxge_lro_rx(ss
, m
, csum
)))
2655 /* otherwise, it was a UDP frame, or a TCP frame which
2656 we could not do LRO on. Tell the stack that the
2658 m
->m_pkthdr
.csum_data
= 0xffff;
2659 m
->m_pkthdr
.csum_flags
= CSUM_PSEUDO_HDR
| CSUM_DATA_VALID
;
2662 /* flowid only valid if RSS hashing is enabled */
2663 if (sc
->num_slices
> 1) {
2664 m
->m_pkthdr
.flowid
= (ss
- sc
->ss
);
2665 m
->m_flags
|= M_FLOWID
;
2668 /* pass the frame up the stack */
2669 (*ifp
->if_input
)(ifp
, m
);
2673 mxge_clean_rx_done(struct mxge_slice_state
*ss
)
2675 mxge_rx_done_t
*rx_done
= &ss
->rx_done
;
2681 while (rx_done
->entry
[rx_done
->idx
].length
!= 0) {
2682 length
= ntohs(rx_done
->entry
[rx_done
->idx
].length
);
2683 rx_done
->entry
[rx_done
->idx
].length
= 0;
2684 checksum
= rx_done
->entry
[rx_done
->idx
].checksum
;
2685 if (length
<= (MHLEN
- MXGEFW_PAD
))
2686 mxge_rx_done_small(ss
, length
, checksum
);
2688 mxge_rx_done_big(ss
, length
, checksum
);
2690 rx_done
->idx
= rx_done
->cnt
& rx_done
->mask
;
2692 /* limit potential for livelock */
2693 if (__predict_false(++limit
> rx_done
->mask
/ 2))
2697 while (!SLIST_EMPTY(&ss
->lro_active
)) {
2698 struct lro_entry
*lro
= SLIST_FIRST(&ss
->lro_active
);
2699 SLIST_REMOVE_HEAD(&ss
->lro_active
, next
);
2700 mxge_lro_flush(ss
, lro
);
2707 mxge_tx_done(struct mxge_slice_state
*ss
, uint32_t mcp_idx
)
2718 ASSERT_SERIALIZED(ifp
->if_serializer
);
2719 while (tx
->pkt_done
!= mcp_idx
) {
2720 idx
= tx
->done
& tx
->mask
;
2722 m
= tx
->info
[idx
].m
;
2723 /* mbuf and DMA map only attached to the first
2726 ss
->obytes
+= m
->m_pkthdr
.len
;
2727 if (m
->m_flags
& M_MCAST
)
2730 tx
->info
[idx
].m
= NULL
;
2731 map
= tx
->info
[idx
].map
;
2732 bus_dmamap_unload(tx
->dmat
, map
);
2735 if (tx
->info
[idx
].flag
) {
2736 tx
->info
[idx
].flag
= 0;
2741 /* If we have space, clear IFF_OACTIVE to tell the stack that
2742 its OK to send packets */
2743 #ifdef IFNET_BUF_RING
2744 flags
= &ss
->if_flags
;
2746 flags
= &ifp
->if_flags
;
2748 if ((*flags
) & IFF_OACTIVE
&&
2749 tx
->req
- tx
->done
< (tx
->mask
+ 1)/4) {
2750 *(flags
) &= ~IFF_OACTIVE
;
2752 mxge_start_locked(ss
);
2754 #ifdef IFNET_BUF_RING
2755 if ((ss
->sc
->num_slices
> 1) && (tx
->req
== tx
->done
)) {
2756 /* let the NIC stop polling this queue, since there
2757 * are no more transmits pending */
2758 if (tx
->req
== tx
->done
) {
2760 tx
->queue_active
= 0;
2769 static struct mxge_media_type mxge_xfp_media_types
[] =
2771 {IFM_10G_CX4
, 0x7f, "10GBASE-CX4 (module)"},
2772 {IFM_10G_SR
, (1 << 7), "10GBASE-SR"},
2773 {IFM_10G_LR
, (1 << 6), "10GBASE-LR"},
2774 {0, (1 << 5), "10GBASE-ER"},
2775 {IFM_10G_LRM
, (1 << 4), "10GBASE-LRM"},
2776 {0, (1 << 3), "10GBASE-SW"},
2777 {0, (1 << 2), "10GBASE-LW"},
2778 {0, (1 << 1), "10GBASE-EW"},
2779 {0, (1 << 0), "Reserved"}
2781 static struct mxge_media_type mxge_sfp_media_types
[] =
2783 {0, (1 << 7), "Reserved"},
2784 {IFM_10G_LRM
, (1 << 6), "10GBASE-LRM"},
2785 {IFM_10G_LR
, (1 << 5), "10GBASE-LR"},
2786 {IFM_10G_SR
, (1 << 4), "10GBASE-SR"}
2790 mxge_set_media(mxge_softc_t
*sc
, int type
)
2792 sc
->media_flags
|= type
;
2793 ifmedia_add(&sc
->media
, sc
->media_flags
, 0, NULL
);
2794 ifmedia_set(&sc
->media
, sc
->media_flags
);
2799 * Determine the media type for a NIC. Some XFPs will identify
2800 * themselves only when their link is up, so this is initiated via a
2801 * link up interrupt. However, this can potentially take up to
2802 * several milliseconds, so it is run via the watchdog routine, rather
2803 * than in the interrupt handler itself. This need only be done
2804 * once, not each time the link is up.
2807 mxge_media_probe(mxge_softc_t
*sc
)
2812 struct mxge_media_type
*mxge_media_types
= NULL
;
2813 int i
, err
, ms
, mxge_media_type_entries
;
2816 sc
->need_media_probe
= 0;
2818 /* if we've already set a media type, we're done */
2819 if (sc
->media_flags
!= (IFM_ETHER
| IFM_AUTO
))
2823 * parse the product code to deterimine the interface type
2824 * (CX4, XFP, Quad Ribbon Fiber) by looking at the character
2825 * after the 3rd dash in the driver's cached copy of the
2826 * EEPROM's product code string.
2828 ptr
= sc
->product_code_string
;
2830 device_printf(sc
->dev
, "Missing product code\n");
2833 for (i
= 0; i
< 3; i
++, ptr
++) {
2834 ptr
= index(ptr
, '-');
2836 device_printf(sc
->dev
,
2837 "only %d dashes in PC?!?\n", i
);
2843 mxge_set_media(sc
, IFM_10G_CX4
);
2846 else if (*ptr
== 'Q') {
2847 /* -Q is Quad Ribbon Fiber */
2848 device_printf(sc
->dev
, "Quad Ribbon Fiber Media\n");
2849 /* FreeBSD has no media type for Quad ribbon fiber */
2855 mxge_media_types
= mxge_xfp_media_types
;
2856 mxge_media_type_entries
=
2857 sizeof (mxge_xfp_media_types
) /
2858 sizeof (mxge_xfp_media_types
[0]);
2859 byte
= MXGE_XFP_COMPLIANCE_BYTE
;
2863 if (*ptr
== 'S' || *(ptr
+1) == 'S') {
2864 /* -S or -2S is SFP+ */
2865 mxge_media_types
= mxge_sfp_media_types
;
2866 mxge_media_type_entries
=
2867 sizeof (mxge_sfp_media_types
) /
2868 sizeof (mxge_sfp_media_types
[0]);
2873 if (mxge_media_types
== NULL
) {
2874 device_printf(sc
->dev
, "Unknown media type: %c\n", *ptr
);
2879 * At this point we know the NIC has an XFP cage, so now we
2880 * try to determine what is in the cage by using the
2881 * firmware's XFP I2C commands to read the XFP 10GbE compilance
2882 * register. We read just one byte, which may take over
2886 cmd
.data0
= 0; /* just fetch 1 byte, not all 256 */
2888 err
= mxge_send_cmd(sc
, MXGEFW_CMD_I2C_READ
, &cmd
);
2889 if (err
== MXGEFW_CMD_ERROR_I2C_FAILURE
) {
2890 device_printf(sc
->dev
, "failed to read XFP\n");
2892 if (err
== MXGEFW_CMD_ERROR_I2C_ABSENT
) {
2893 device_printf(sc
->dev
, "Type R/S with no XFP!?!?\n");
2895 if (err
!= MXGEFW_CMD_OK
) {
2899 /* now we wait for the data to be cached */
2901 err
= mxge_send_cmd(sc
, MXGEFW_CMD_I2C_BYTE
, &cmd
);
2902 for (ms
= 0; (err
== EBUSY
) && (ms
< 50); ms
++) {
2905 err
= mxge_send_cmd(sc
, MXGEFW_CMD_I2C_BYTE
, &cmd
);
2907 if (err
!= MXGEFW_CMD_OK
) {
2908 device_printf(sc
->dev
, "failed to read %s (%d, %dms)\n",
2909 cage_type
, err
, ms
);
2913 if (cmd
.data0
== mxge_media_types
[0].bitmask
) {
2915 device_printf(sc
->dev
, "%s:%s\n", cage_type
,
2916 mxge_media_types
[0].name
);
2917 mxge_set_media(sc
, IFM_10G_CX4
);
2920 for (i
= 1; i
< mxge_media_type_entries
; i
++) {
2921 if (cmd
.data0
& mxge_media_types
[i
].bitmask
) {
2923 device_printf(sc
->dev
, "%s:%s\n",
2925 mxge_media_types
[i
].name
);
2927 mxge_set_media(sc
, mxge_media_types
[i
].flag
);
2931 device_printf(sc
->dev
, "%s media 0x%x unknown\n", cage_type
,
2938 mxge_intr(void *arg
)
2940 struct mxge_slice_state
*ss
= arg
;
2941 mxge_softc_t
*sc
= ss
->sc
;
2942 mcp_irq_data_t
*stats
= ss
->fw_stats
;
2943 mxge_tx_ring_t
*tx
= &ss
->tx
;
2944 mxge_rx_done_t
*rx_done
= &ss
->rx_done
;
2945 uint32_t send_done_count
;
2949 #ifndef IFNET_BUF_RING
2950 /* an interrupt on a non-zero slice is implicitly valid
2951 since MSI-X irqs are not shared */
2953 mxge_clean_rx_done(ss
);
2954 *ss
->irq_claim
= be32toh(3);
2959 /* make sure the DMA has finished */
2960 if (!stats
->valid
) {
2963 valid
= stats
->valid
;
2965 if (sc
->legacy_irq
) {
2966 /* lower legacy IRQ */
2967 *sc
->irq_deassert
= 0;
2968 if (!mxge_deassert_wait
)
2969 /* don't wait for conf. that irq is low */
2975 /* loop while waiting for legacy irq deassertion */
2977 /* check for transmit completes and receives */
2978 send_done_count
= be32toh(stats
->send_done_count
);
2979 while ((send_done_count
!= tx
->pkt_done
) ||
2980 (rx_done
->entry
[rx_done
->idx
].length
!= 0)) {
2981 if (send_done_count
!= tx
->pkt_done
)
2982 mxge_tx_done(ss
, (int)send_done_count
);
2983 mxge_clean_rx_done(ss
);
2984 send_done_count
= be32toh(stats
->send_done_count
);
2986 if (sc
->legacy_irq
&& mxge_deassert_wait
)
2988 } while (*((volatile uint8_t *) &stats
->valid
));
2990 /* fw link & error stats meaningful only on the first slice */
2991 if (__predict_false((ss
== sc
->ss
) && stats
->stats_updated
)) {
2992 if (sc
->link_state
!= stats
->link_up
) {
2993 sc
->link_state
= stats
->link_up
;
2994 if (sc
->link_state
) {
2995 sc
->ifp
->if_link_state
= LINK_STATE_UP
;
2996 if_link_state_change(sc
->ifp
);
2998 device_printf(sc
->dev
, "link up\n");
3000 sc
->ifp
->if_link_state
= LINK_STATE_DOWN
;
3001 if_link_state_change(sc
->ifp
);
3003 device_printf(sc
->dev
, "link down\n");
3005 sc
->need_media_probe
= 1;
3007 if (sc
->rdma_tags_available
!=
3008 be32toh(stats
->rdma_tags_available
)) {
3009 sc
->rdma_tags_available
=
3010 be32toh(stats
->rdma_tags_available
);
3011 device_printf(sc
->dev
, "RDMA timed out! %d tags "
3012 "left\n", sc
->rdma_tags_available
);
3015 if (stats
->link_down
) {
3016 sc
->down_cnt
+= stats
->link_down
;
3018 sc
->ifp
->if_link_state
= LINK_STATE_DOWN
;
3019 if_link_state_change(sc
->ifp
);
3023 /* check to see if we have rx token to pass back */
3025 *ss
->irq_claim
= be32toh(3);
3026 *(ss
->irq_claim
+ 1) = be32toh(3);
3030 mxge_init(void *arg
)
3037 mxge_free_slice_mbufs(struct mxge_slice_state
*ss
)
3039 struct lro_entry
*lro_entry
;
3042 while (!SLIST_EMPTY(&ss
->lro_free
)) {
3043 lro_entry
= SLIST_FIRST(&ss
->lro_free
);
3044 SLIST_REMOVE_HEAD(&ss
->lro_free
, next
);
3045 kfree(lro_entry
, M_DEVBUF
);
3048 for (i
= 0; i
<= ss
->rx_big
.mask
; i
++) {
3049 if (ss
->rx_big
.info
[i
].m
== NULL
)
3051 bus_dmamap_unload(ss
->rx_big
.dmat
,
3052 ss
->rx_big
.info
[i
].map
);
3053 m_freem(ss
->rx_big
.info
[i
].m
);
3054 ss
->rx_big
.info
[i
].m
= NULL
;
3057 for (i
= 0; i
<= ss
->rx_small
.mask
; i
++) {
3058 if (ss
->rx_small
.info
[i
].m
== NULL
)
3060 bus_dmamap_unload(ss
->rx_small
.dmat
,
3061 ss
->rx_small
.info
[i
].map
);
3062 m_freem(ss
->rx_small
.info
[i
].m
);
3063 ss
->rx_small
.info
[i
].m
= NULL
;
3066 /* transmit ring used only on the first slice */
3067 if (ss
->tx
.info
== NULL
)
3070 for (i
= 0; i
<= ss
->tx
.mask
; i
++) {
3071 ss
->tx
.info
[i
].flag
= 0;
3072 if (ss
->tx
.info
[i
].m
== NULL
)
3074 bus_dmamap_unload(ss
->tx
.dmat
,
3075 ss
->tx
.info
[i
].map
);
3076 m_freem(ss
->tx
.info
[i
].m
);
3077 ss
->tx
.info
[i
].m
= NULL
;
3082 mxge_free_mbufs(mxge_softc_t
*sc
)
3086 for (slice
= 0; slice
< sc
->num_slices
; slice
++)
3087 mxge_free_slice_mbufs(&sc
->ss
[slice
]);
3091 mxge_free_slice_rings(struct mxge_slice_state
*ss
)
3096 if (ss
->rx_done
.entry
!= NULL
)
3097 mxge_dma_free(&ss
->rx_done
.dma
);
3098 ss
->rx_done
.entry
= NULL
;
3100 if (ss
->tx
.req_bytes
!= NULL
)
3101 kfree(ss
->tx
.req_bytes
, M_DEVBUF
);
3102 ss
->tx
.req_bytes
= NULL
;
3104 if (ss
->tx
.seg_list
!= NULL
)
3105 kfree(ss
->tx
.seg_list
, M_DEVBUF
);
3106 ss
->tx
.seg_list
= NULL
;
3108 if (ss
->rx_small
.shadow
!= NULL
)
3109 kfree(ss
->rx_small
.shadow
, M_DEVBUF
);
3110 ss
->rx_small
.shadow
= NULL
;
3112 if (ss
->rx_big
.shadow
!= NULL
)
3113 kfree(ss
->rx_big
.shadow
, M_DEVBUF
);
3114 ss
->rx_big
.shadow
= NULL
;
3116 if (ss
->tx
.info
!= NULL
) {
3117 if (ss
->tx
.dmat
!= NULL
) {
3118 for (i
= 0; i
<= ss
->tx
.mask
; i
++) {
3119 bus_dmamap_destroy(ss
->tx
.dmat
,
3120 ss
->tx
.info
[i
].map
);
3122 bus_dma_tag_destroy(ss
->tx
.dmat
);
3124 kfree(ss
->tx
.info
, M_DEVBUF
);
3128 if (ss
->rx_small
.info
!= NULL
) {
3129 if (ss
->rx_small
.dmat
!= NULL
) {
3130 for (i
= 0; i
<= ss
->rx_small
.mask
; i
++) {
3131 bus_dmamap_destroy(ss
->rx_small
.dmat
,
3132 ss
->rx_small
.info
[i
].map
);
3134 bus_dmamap_destroy(ss
->rx_small
.dmat
,
3135 ss
->rx_small
.extra_map
);
3136 bus_dma_tag_destroy(ss
->rx_small
.dmat
);
3138 kfree(ss
->rx_small
.info
, M_DEVBUF
);
3140 ss
->rx_small
.info
= NULL
;
3142 if (ss
->rx_big
.info
!= NULL
) {
3143 if (ss
->rx_big
.dmat
!= NULL
) {
3144 for (i
= 0; i
<= ss
->rx_big
.mask
; i
++) {
3145 bus_dmamap_destroy(ss
->rx_big
.dmat
,
3146 ss
->rx_big
.info
[i
].map
);
3148 bus_dmamap_destroy(ss
->rx_big
.dmat
,
3149 ss
->rx_big
.extra_map
);
3150 bus_dma_tag_destroy(ss
->rx_big
.dmat
);
3152 kfree(ss
->rx_big
.info
, M_DEVBUF
);
3154 ss
->rx_big
.info
= NULL
;
3158 mxge_free_rings(mxge_softc_t
*sc
)
3162 for (slice
= 0; slice
< sc
->num_slices
; slice
++)
3163 mxge_free_slice_rings(&sc
->ss
[slice
]);
3167 mxge_alloc_slice_rings(struct mxge_slice_state
*ss
, int rx_ring_entries
,
3168 int tx_ring_entries
)
3170 mxge_softc_t
*sc
= ss
->sc
;
3176 /* allocate per-slice receive resources */
3178 ss
->rx_small
.mask
= ss
->rx_big
.mask
= rx_ring_entries
- 1;
3179 ss
->rx_done
.mask
= (2 * rx_ring_entries
) - 1;
3181 /* allocate the rx shadow rings */
3182 bytes
= rx_ring_entries
* sizeof (*ss
->rx_small
.shadow
);
3183 ss
->rx_small
.shadow
= kmalloc(bytes
, M_DEVBUF
, M_ZERO
|M_WAITOK
);
3184 if (ss
->rx_small
.shadow
== NULL
)
3187 bytes
= rx_ring_entries
* sizeof (*ss
->rx_big
.shadow
);
3188 ss
->rx_big
.shadow
= kmalloc(bytes
, M_DEVBUF
, M_ZERO
|M_WAITOK
);
3189 if (ss
->rx_big
.shadow
== NULL
)
3192 /* allocate the rx host info rings */
3193 bytes
= rx_ring_entries
* sizeof (*ss
->rx_small
.info
);
3194 ss
->rx_small
.info
= kmalloc(bytes
, M_DEVBUF
, M_ZERO
|M_WAITOK
);
3195 if (ss
->rx_small
.info
== NULL
)
3198 bytes
= rx_ring_entries
* sizeof (*ss
->rx_big
.info
);
3199 ss
->rx_big
.info
= kmalloc(bytes
, M_DEVBUF
, M_ZERO
|M_WAITOK
);
3200 if (ss
->rx_big
.info
== NULL
)
3203 /* allocate the rx busdma resources */
3204 err
= bus_dma_tag_create(sc
->parent_dmat
, /* parent */
3206 4096, /* boundary */
3207 BUS_SPACE_MAXADDR
, /* low */
3208 BUS_SPACE_MAXADDR
, /* high */
3209 NULL
, NULL
, /* filter */
3210 MHLEN
, /* maxsize */
3212 MHLEN
, /* maxsegsize */
3213 BUS_DMA_ALLOCNOW
, /* flags */
3214 &ss
->rx_small
.dmat
); /* tag */
3216 device_printf(sc
->dev
, "Err %d allocating rx_small dmat\n",
3221 err
= bus_dma_tag_create(sc
->parent_dmat
, /* parent */
3223 #if MXGE_VIRT_JUMBOS
3224 4096, /* boundary */
3228 BUS_SPACE_MAXADDR
, /* low */
3229 BUS_SPACE_MAXADDR
, /* high */
3230 NULL
, NULL
, /* filter */
3231 3*4096, /* maxsize */
3232 #if MXGE_VIRT_JUMBOS
3234 4096, /* maxsegsize*/
3237 MJUM9BYTES
, /* maxsegsize*/
3239 BUS_DMA_ALLOCNOW
, /* flags */
3240 &ss
->rx_big
.dmat
); /* tag */
3242 device_printf(sc
->dev
, "Err %d allocating rx_big dmat\n",
3246 for (i
= 0; i
<= ss
->rx_small
.mask
; i
++) {
3247 err
= bus_dmamap_create(ss
->rx_small
.dmat
, 0,
3248 &ss
->rx_small
.info
[i
].map
);
3250 device_printf(sc
->dev
, "Err %d rx_small dmamap\n",
3255 err
= bus_dmamap_create(ss
->rx_small
.dmat
, 0,
3256 &ss
->rx_small
.extra_map
);
3258 device_printf(sc
->dev
, "Err %d extra rx_small dmamap\n",
3263 for (i
= 0; i
<= ss
->rx_big
.mask
; i
++) {
3264 err
= bus_dmamap_create(ss
->rx_big
.dmat
, 0,
3265 &ss
->rx_big
.info
[i
].map
);
3267 device_printf(sc
->dev
, "Err %d rx_big dmamap\n",
3272 err
= bus_dmamap_create(ss
->rx_big
.dmat
, 0,
3273 &ss
->rx_big
.extra_map
);
3275 device_printf(sc
->dev
, "Err %d extra rx_big dmamap\n",
3280 /* now allocate TX resouces */
3282 #ifndef IFNET_BUF_RING
3283 /* only use a single TX ring for now */
3284 if (ss
!= ss
->sc
->ss
)
3288 ss
->tx
.mask
= tx_ring_entries
- 1;
3289 ss
->tx
.max_desc
= MIN(MXGE_MAX_SEND_DESC
, tx_ring_entries
/ 4);
3292 /* allocate the tx request copy block */
3294 sizeof (*ss
->tx
.req_list
) * (ss
->tx
.max_desc
+ 4);
3295 ss
->tx
.req_bytes
= kmalloc(bytes
, M_DEVBUF
, M_WAITOK
);
3296 if (ss
->tx
.req_bytes
== NULL
)
3298 /* ensure req_list entries are aligned to 8 bytes */
3299 ss
->tx
.req_list
= (mcp_kreq_ether_send_t
*)
3300 ((unsigned long)(ss
->tx
.req_bytes
+ 7) & ~7UL);
3302 /* allocate the tx busdma segment list */
3303 bytes
= sizeof (*ss
->tx
.seg_list
) * ss
->tx
.max_desc
;
3304 ss
->tx
.seg_list
= (bus_dma_segment_t
*)
3305 kmalloc(bytes
, M_DEVBUF
, M_WAITOK
);
3306 if (ss
->tx
.seg_list
== NULL
)
3309 /* allocate the tx host info ring */
3310 bytes
= tx_ring_entries
* sizeof (*ss
->tx
.info
);
3311 ss
->tx
.info
= kmalloc(bytes
, M_DEVBUF
, M_ZERO
|M_WAITOK
);
3312 if (ss
->tx
.info
== NULL
)
3315 /* allocate the tx busdma resources */
3316 err
= bus_dma_tag_create(sc
->parent_dmat
, /* parent */
3318 sc
->tx_boundary
, /* boundary */
3319 BUS_SPACE_MAXADDR
, /* low */
3320 BUS_SPACE_MAXADDR
, /* high */
3321 NULL
, NULL
, /* filter */
3322 65536 + 256, /* maxsize */
3323 ss
->tx
.max_desc
- 2, /* num segs */
3324 sc
->tx_boundary
, /* maxsegsz */
3325 BUS_DMA_ALLOCNOW
, /* flags */
3326 &ss
->tx
.dmat
); /* tag */
3329 device_printf(sc
->dev
, "Err %d allocating tx dmat\n",
3334 /* now use these tags to setup dmamaps for each slot
3336 for (i
= 0; i
<= ss
->tx
.mask
; i
++) {
3337 err
= bus_dmamap_create(ss
->tx
.dmat
, 0,
3338 &ss
->tx
.info
[i
].map
);
3340 device_printf(sc
->dev
, "Err %d tx dmamap\n",
3350 mxge_alloc_rings(mxge_softc_t
*sc
)
3354 int tx_ring_entries
, rx_ring_entries
;
3357 /* get ring sizes */
3358 err
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_SEND_RING_SIZE
, &cmd
);
3359 tx_ring_size
= cmd
.data0
;
3361 device_printf(sc
->dev
, "Cannot determine tx ring sizes\n");
3365 tx_ring_entries
= tx_ring_size
/ sizeof (mcp_kreq_ether_send_t
);
3366 rx_ring_entries
= sc
->rx_ring_size
/ sizeof (mcp_dma_addr_t
);
3367 ifq_set_maxlen(&sc
->ifp
->if_snd
, tx_ring_entries
- 1);
3368 ifq_set_ready(&sc
->ifp
->if_snd
);
3370 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
3371 err
= mxge_alloc_slice_rings(&sc
->ss
[slice
],
3380 mxge_free_rings(sc
);
3387 mxge_choose_params(int mtu
, int *big_buf_size
, int *cl_size
, int *nbufs
)
3389 int bufsize
= mtu
+ ETHER_HDR_LEN
+ EVL_ENCAPLEN
+ MXGEFW_PAD
;
3391 if (bufsize
< MCLBYTES
) {
3392 /* easy, everything fits in a single buffer */
3393 *big_buf_size
= MCLBYTES
;
3394 *cl_size
= MCLBYTES
;
3399 if (bufsize
< MJUMPAGESIZE
) {
3400 /* still easy, everything still fits in a single buffer */
3401 *big_buf_size
= MJUMPAGESIZE
;
3402 *cl_size
= MJUMPAGESIZE
;
3406 #if MXGE_VIRT_JUMBOS
3407 /* now we need to use virtually contiguous buffers */
3408 *cl_size
= MJUM9BYTES
;
3409 *big_buf_size
= 4096;
3410 *nbufs
= mtu
/ 4096 + 1;
3411 /* needs to be a power of two, so round up */
3415 *cl_size
= MJUM9BYTES
;
3416 *big_buf_size
= MJUM9BYTES
;
3422 mxge_slice_open(struct mxge_slice_state
*ss
, int nbufs
, int cl_size
)
3427 struct lro_entry
*lro_entry
;
3432 slice
= ss
- sc
->ss
;
3434 SLIST_INIT(&ss
->lro_free
);
3435 SLIST_INIT(&ss
->lro_active
);
3437 for (i
= 0; i
< sc
->lro_cnt
; i
++) {
3438 lro_entry
= (struct lro_entry
*)
3439 kmalloc(sizeof (*lro_entry
), M_DEVBUF
,
3441 if (lro_entry
== NULL
) {
3445 SLIST_INSERT_HEAD(&ss
->lro_free
, lro_entry
, next
);
3447 /* get the lanai pointers to the send and receive rings */
3450 #ifndef IFNET_BUF_RING
3451 /* We currently only send from the first slice */
3455 err
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_SEND_OFFSET
, &cmd
);
3457 (volatile mcp_kreq_ether_send_t
*)(sc
->sram
+ cmd
.data0
);
3458 ss
->tx
.send_go
= (volatile uint32_t *)
3459 (sc
->sram
+ MXGEFW_ETH_SEND_GO
+ 64 * slice
);
3460 ss
->tx
.send_stop
= (volatile uint32_t *)
3461 (sc
->sram
+ MXGEFW_ETH_SEND_STOP
+ 64 * slice
);
3462 #ifndef IFNET_BUF_RING
3466 err
|= mxge_send_cmd(sc
,
3467 MXGEFW_CMD_GET_SMALL_RX_OFFSET
, &cmd
);
3468 ss
->rx_small
.lanai
=
3469 (volatile mcp_kreq_ether_recv_t
*)(sc
->sram
+ cmd
.data0
);
3471 err
|= mxge_send_cmd(sc
, MXGEFW_CMD_GET_BIG_RX_OFFSET
, &cmd
);
3473 (volatile mcp_kreq_ether_recv_t
*)(sc
->sram
+ cmd
.data0
);
3476 device_printf(sc
->dev
,
3477 "failed to get ring sizes or locations\n");
3481 /* stock receive rings */
3482 for (i
= 0; i
<= ss
->rx_small
.mask
; i
++) {
3483 map
= ss
->rx_small
.info
[i
].map
;
3484 err
= mxge_get_buf_small(ss
, map
, i
);
3486 device_printf(sc
->dev
, "alloced %d/%d smalls\n",
3487 i
, ss
->rx_small
.mask
+ 1);
3491 for (i
= 0; i
<= ss
->rx_big
.mask
; i
++) {
3492 ss
->rx_big
.shadow
[i
].addr_low
= 0xffffffff;
3493 ss
->rx_big
.shadow
[i
].addr_high
= 0xffffffff;
3495 ss
->rx_big
.nbufs
= nbufs
;
3496 ss
->rx_big
.cl_size
= cl_size
;
3497 ss
->rx_big
.mlen
= ss
->sc
->ifp
->if_mtu
+ ETHER_HDR_LEN
+
3498 EVL_ENCAPLEN
+ MXGEFW_PAD
;
3499 for (i
= 0; i
<= ss
->rx_big
.mask
; i
+= ss
->rx_big
.nbufs
) {
3500 map
= ss
->rx_big
.info
[i
].map
;
3501 err
= mxge_get_buf_big(ss
, map
, i
);
3503 device_printf(sc
->dev
, "alloced %d/%d bigs\n",
3504 i
, ss
->rx_big
.mask
+ 1);
3512 mxge_open(mxge_softc_t
*sc
)
3515 int err
, big_bytes
, nbufs
, slice
, cl_size
, i
;
3517 volatile uint8_t *itable
;
3518 struct mxge_slice_state
*ss
;
3520 ASSERT_SERIALIZED(sc
->ifp
->if_serializer
);
3521 /* Copy the MAC address in case it was overridden */
3522 bcopy(IF_LLADDR(sc
->ifp
), sc
->mac_addr
, ETHER_ADDR_LEN
);
3524 err
= mxge_reset(sc
, 1);
3526 device_printf(sc
->dev
, "failed to reset\n");
3530 if (sc
->num_slices
> 1) {
3531 /* setup the indirection table */
3532 cmd
.data0
= sc
->num_slices
;
3533 err
= mxge_send_cmd(sc
, MXGEFW_CMD_SET_RSS_TABLE_SIZE
,
3536 err
|= mxge_send_cmd(sc
, MXGEFW_CMD_GET_RSS_TABLE_OFFSET
,
3539 device_printf(sc
->dev
,
3540 "failed to setup rss tables\n");
3544 /* just enable an identity mapping */
3545 itable
= sc
->sram
+ cmd
.data0
;
3546 for (i
= 0; i
< sc
->num_slices
; i
++)
3547 itable
[i
] = (uint8_t)i
;
3550 cmd
.data1
= mxge_rss_hash_type
;
3551 err
= mxge_send_cmd(sc
, MXGEFW_CMD_SET_RSS_ENABLE
, &cmd
);
3553 device_printf(sc
->dev
, "failed to enable slices\n");
3559 mxge_choose_params(sc
->ifp
->if_mtu
, &big_bytes
, &cl_size
, &nbufs
);
3562 err
= mxge_send_cmd(sc
, MXGEFW_CMD_ALWAYS_USE_N_BIG_BUFFERS
,
3564 /* error is only meaningful if we're trying to set
3565 MXGEFW_CMD_ALWAYS_USE_N_BIG_BUFFERS > 1 */
3566 if (err
&& nbufs
> 1) {
3567 device_printf(sc
->dev
,
3568 "Failed to set alway-use-n to %d\n",
3572 /* Give the firmware the mtu and the big and small buffer
3573 sizes. The firmware wants the big buf size to be a power
3574 of two. Luckily, FreeBSD's clusters are powers of two */
3575 cmd
.data0
= sc
->ifp
->if_mtu
+ ETHER_HDR_LEN
+ EVL_ENCAPLEN
;
3576 err
= mxge_send_cmd(sc
, MXGEFW_CMD_SET_MTU
, &cmd
);
3577 cmd
.data0
= MHLEN
- MXGEFW_PAD
;
3578 err
|= mxge_send_cmd(sc
, MXGEFW_CMD_SET_SMALL_BUFFER_SIZE
,
3580 cmd
.data0
= big_bytes
;
3581 err
|= mxge_send_cmd(sc
, MXGEFW_CMD_SET_BIG_BUFFER_SIZE
, &cmd
);
3584 device_printf(sc
->dev
, "failed to setup params\n");
3588 /* Now give him the pointer to the stats block */
3590 #ifdef IFNET_BUF_RING
3591 slice
< sc
->num_slices
;
3596 ss
= &sc
->ss
[slice
];
3598 MXGE_LOWPART_TO_U32(ss
->fw_stats_dma
.bus_addr
);
3600 MXGE_HIGHPART_TO_U32(ss
->fw_stats_dma
.bus_addr
);
3601 cmd
.data2
= sizeof(struct mcp_irq_data
);
3602 cmd
.data2
|= (slice
<< 16);
3603 err
|= mxge_send_cmd(sc
, MXGEFW_CMD_SET_STATS_DMA_V2
, &cmd
);
3607 bus
= sc
->ss
->fw_stats_dma
.bus_addr
;
3608 bus
+= offsetof(struct mcp_irq_data
, send_done_count
);
3609 cmd
.data0
= MXGE_LOWPART_TO_U32(bus
);
3610 cmd
.data1
= MXGE_HIGHPART_TO_U32(bus
);
3611 err
= mxge_send_cmd(sc
,
3612 MXGEFW_CMD_SET_STATS_DMA_OBSOLETE
,
3614 /* Firmware cannot support multicast without STATS_DMA_V2 */
3615 sc
->fw_multicast_support
= 0;
3617 sc
->fw_multicast_support
= 1;
3621 device_printf(sc
->dev
, "failed to setup params\n");
3625 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
3626 err
= mxge_slice_open(&sc
->ss
[slice
], nbufs
, cl_size
);
3628 device_printf(sc
->dev
, "couldn't open slice %d\n",
3634 /* Finally, start the firmware running */
3635 err
= mxge_send_cmd(sc
, MXGEFW_CMD_ETHERNET_UP
, &cmd
);
3637 device_printf(sc
->dev
, "Couldn't bring up link\n");
3640 #ifdef IFNET_BUF_RING
3641 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
3642 ss
= &sc
->ss
[slice
];
3643 ss
->if_flags
|= IFF_RUNNING
;
3644 ss
->if_flags
&= ~IFF_OACTIVE
;
3647 sc
->ifp
->if_flags
|= IFF_RUNNING
;
3648 sc
->ifp
->if_flags
&= ~IFF_OACTIVE
;
3649 callout_reset(&sc
->co_hdl
, mxge_ticks
, mxge_tick
, sc
);
3655 mxge_free_mbufs(sc
);
3661 mxge_close(mxge_softc_t
*sc
)
3664 int err
, old_down_cnt
;
3665 #ifdef IFNET_BUF_RING
3666 struct mxge_slice_state
*ss
;
3670 ASSERT_SERIALIZED(sc
->ifp
->if_serializer
);
3671 callout_stop(&sc
->co_hdl
);
3672 #ifdef IFNET_BUF_RING
3673 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
3674 ss
= &sc
->ss
[slice
];
3675 ss
->if_flags
&= ~IFF_RUNNING
;
3678 sc
->ifp
->if_flags
&= ~IFF_RUNNING
;
3679 old_down_cnt
= sc
->down_cnt
;
3681 err
= mxge_send_cmd(sc
, MXGEFW_CMD_ETHERNET_DOWN
, &cmd
);
3683 device_printf(sc
->dev
, "Couldn't bring down link\n");
3685 if (old_down_cnt
== sc
->down_cnt
) {
3686 /* wait for down irq */
3687 DELAY(10 * sc
->intr_coal_delay
);
3690 if (old_down_cnt
== sc
->down_cnt
) {
3691 device_printf(sc
->dev
, "never got down irq\n");
3694 mxge_free_mbufs(sc
);
3700 mxge_setup_cfg_space(mxge_softc_t
*sc
)
3702 device_t dev
= sc
->dev
;
3704 uint16_t cmd
, lnk
, pectl
;
3706 /* find the PCIe link width and set max read request to 4KB*/
3707 if (pci_find_extcap(dev
, PCIY_EXPRESS
, ®
) == 0) {
3708 lnk
= pci_read_config(dev
, reg
+ 0x12, 2);
3709 sc
->link_width
= (lnk
>> 4) & 0x3f;
3711 pectl
= pci_read_config(dev
, reg
+ 0x8, 2);
3712 pectl
= (pectl
& ~0x7000) | (5 << 12);
3713 pci_write_config(dev
, reg
+ 0x8, pectl
, 2);
3716 /* Enable DMA and Memory space access */
3717 pci_enable_busmaster(dev
);
3718 cmd
= pci_read_config(dev
, PCIR_COMMAND
, 2);
3719 cmd
|= PCIM_CMD_MEMEN
;
3720 pci_write_config(dev
, PCIR_COMMAND
, cmd
, 2);
3724 mxge_read_reboot(mxge_softc_t
*sc
)
3726 device_t dev
= sc
->dev
;
3729 /* find the vendor specific offset */
3730 if (pci_find_extcap(dev
, PCIY_VENDOR
, &vs
) != 0) {
3731 device_printf(sc
->dev
,
3732 "could not find vendor specific offset\n");
3733 return (uint32_t)-1;
3735 /* enable read32 mode */
3736 pci_write_config(dev
, vs
+ 0x10, 0x3, 1);
3737 /* tell NIC which register to read */
3738 pci_write_config(dev
, vs
+ 0x18, 0xfffffff0, 4);
3739 return (pci_read_config(dev
, vs
+ 0x14, 4));
3743 mxge_watchdog_reset(mxge_softc_t
*sc
, int slice
)
3745 struct pci_devinfo
*dinfo
;
3753 device_printf(sc
->dev
, "Watchdog reset!\n");
3756 * check to see if the NIC rebooted. If it did, then all of
3757 * PCI config space has been reset, and things like the
3758 * busmaster bit will be zero. If this is the case, then we
3759 * must restore PCI config space before the NIC can be used
3762 cmd
= pci_read_config(sc
->dev
, PCIR_COMMAND
, 2);
3763 if (cmd
== 0xffff) {
3765 * maybe the watchdog caught the NIC rebooting; wait
3766 * up to 100ms for it to finish. If it does not come
3767 * back, then give up
3770 cmd
= pci_read_config(sc
->dev
, PCIR_COMMAND
, 2);
3771 if (cmd
== 0xffff) {
3772 device_printf(sc
->dev
, "NIC disappeared!\n");
3776 if ((cmd
& PCIM_CMD_BUSMASTEREN
) == 0) {
3777 /* print the reboot status */
3778 reboot
= mxge_read_reboot(sc
);
3779 device_printf(sc
->dev
, "NIC rebooted, status = 0x%x\n",
3781 /* restore PCI configuration space */
3782 dinfo
= device_get_ivars(sc
->dev
);
3783 pci_cfg_restore(sc
->dev
, dinfo
);
3785 /* and redo any changes we made to our config space */
3786 mxge_setup_cfg_space(sc
);
3788 if (sc
->ifp
->if_flags
& IFF_RUNNING
) {
3790 err
= mxge_open(sc
);
3793 tx
= &sc
->ss
[slice
].tx
;
3794 device_printf(sc
->dev
,
3795 "NIC did not reboot, slice %d ring state:\n",
3797 device_printf(sc
->dev
,
3798 "tx.req=%d tx.done=%d, tx.queue_active=%d\n",
3799 tx
->req
, tx
->done
, tx
->queue_active
);
3800 device_printf(sc
->dev
, "tx.activate=%d tx.deactivate=%d\n",
3801 tx
->activate
, tx
->deactivate
);
3802 device_printf(sc
->dev
, "pkt_done=%d fw=%d\n",
3804 be32toh(sc
->ss
->fw_stats
->send_done_count
));
3805 device_printf(sc
->dev
, "not resetting\n");
3811 mxge_watchdog(mxge_softc_t
*sc
)
3814 uint32_t rx_pause
= be32toh(sc
->ss
->fw_stats
->dropped_pause
);
3817 /* see if we have outstanding transmits, which
3818 have been pending for more than mxge_ticks */
3820 #ifdef IFNET_BUF_RING
3821 (i
< sc
->num_slices
) && (err
== 0);
3823 (i
< 1) && (err
== 0);
3827 if (tx
->req
!= tx
->done
&&
3828 tx
->watchdog_req
!= tx
->watchdog_done
&&
3829 tx
->done
== tx
->watchdog_done
) {
3830 /* check for pause blocking before resetting */
3831 if (tx
->watchdog_rx_pause
== rx_pause
)
3832 err
= mxge_watchdog_reset(sc
, i
);
3834 device_printf(sc
->dev
, "Flow control blocking "
3835 "xmits, check link partner\n");
3838 tx
->watchdog_req
= tx
->req
;
3839 tx
->watchdog_done
= tx
->done
;
3840 tx
->watchdog_rx_pause
= rx_pause
;
3843 if (sc
->need_media_probe
)
3844 mxge_media_probe(sc
);
3849 mxge_update_stats(mxge_softc_t
*sc
)
3851 struct mxge_slice_state
*ss
;
3852 u_long ipackets
= 0;
3853 u_long opackets
= 0;
3854 #ifdef IFNET_BUF_RING
3862 for (slice
= 0; slice
< sc
->num_slices
; slice
++) {
3863 ss
= &sc
->ss
[slice
];
3864 ipackets
+= ss
->ipackets
;
3865 opackets
+= ss
->opackets
;
3866 #ifdef IFNET_BUF_RING
3867 obytes
+= ss
->obytes
;
3868 omcasts
+= ss
->omcasts
;
3869 odrops
+= ss
->tx
.br
->br_drops
;
3871 oerrors
+= ss
->oerrors
;
3873 sc
->ifp
->if_ipackets
= ipackets
;
3874 sc
->ifp
->if_opackets
= opackets
;
3875 #ifdef IFNET_BUF_RING
3876 sc
->ifp
->if_obytes
= obytes
;
3877 sc
->ifp
->if_omcasts
= omcasts
;
3878 sc
->ifp
->if_snd
.ifq_drops
= odrops
;
3880 sc
->ifp
->if_oerrors
= oerrors
;
3884 mxge_tick(void *arg
)
3886 mxge_softc_t
*sc
= arg
;
3889 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
3890 /* aggregate stats from different slices */
3891 mxge_update_stats(sc
);
3892 if (!sc
->watchdog_countdown
) {
3893 err
= mxge_watchdog(sc
);
3894 sc
->watchdog_countdown
= 4;
3896 sc
->watchdog_countdown
--;
3898 callout_reset(&sc
->co_hdl
, mxge_ticks
, mxge_tick
, sc
);
3899 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
3903 mxge_media_change(struct ifnet
*ifp
)
3909 mxge_change_mtu(mxge_softc_t
*sc
, int mtu
)
3911 struct ifnet
*ifp
= sc
->ifp
;
3912 int real_mtu
, old_mtu
;
3915 if (ifp
->if_serializer
)
3916 ASSERT_SERIALIZED(ifp
->if_serializer
);
3918 real_mtu
= mtu
+ ETHER_HDR_LEN
+ EVL_ENCAPLEN
;
3919 if ((real_mtu
> sc
->max_mtu
) || real_mtu
< 60)
3921 old_mtu
= ifp
->if_mtu
;
3923 if (ifp
->if_flags
& IFF_RUNNING
) {
3925 err
= mxge_open(sc
);
3927 ifp
->if_mtu
= old_mtu
;
3929 (void) mxge_open(sc
);
3936 mxge_media_status(struct ifnet
*ifp
, struct ifmediareq
*ifmr
)
3938 mxge_softc_t
*sc
= ifp
->if_softc
;
3943 ifmr
->ifm_status
= IFM_AVALID
;
3944 ifmr
->ifm_status
|= sc
->link_state
? IFM_ACTIVE
: 0;
3945 ifmr
->ifm_active
= IFM_AUTO
| IFM_ETHER
;
3946 ifmr
->ifm_active
|= sc
->link_state
? IFM_FDX
: 0;
3950 mxge_ioctl(struct ifnet
*ifp
, u_long command
, caddr_t data
, struct ucred
*cr
)
3952 mxge_softc_t
*sc
= ifp
->if_softc
;
3953 struct ifreq
*ifr
= (struct ifreq
*)data
;
3958 ASSERT_SERIALIZED(ifp
->if_serializer
);
3962 err
= ether_ioctl(ifp
, command
, data
);
3966 err
= mxge_change_mtu(sc
, ifr
->ifr_mtu
);
3973 if (ifp
->if_flags
& IFF_UP
) {
3974 if (!(ifp
->if_flags
& IFF_RUNNING
)) {
3975 err
= mxge_open(sc
);
3977 /* take care of promis can allmulti
3979 mxge_change_promisc(sc
,
3980 ifp
->if_flags
& IFF_PROMISC
);
3981 mxge_set_multicast_list(sc
);
3984 if (ifp
->if_flags
& IFF_RUNNING
) {
3992 mxge_set_multicast_list(sc
);
3996 mask
= ifr
->ifr_reqcap
^ ifp
->if_capenable
;
3997 if (mask
& IFCAP_TXCSUM
) {
3998 if (IFCAP_TXCSUM
& ifp
->if_capenable
) {
3999 ifp
->if_capenable
&= ~(IFCAP_TXCSUM
|IFCAP_TSO4
);
4000 ifp
->if_hwassist
&= ~(CSUM_TCP
| CSUM_UDP
4003 ifp
->if_capenable
|= IFCAP_TXCSUM
;
4004 ifp
->if_hwassist
|= (CSUM_TCP
| CSUM_UDP
);
4006 } else if (mask
& IFCAP_RXCSUM
) {
4007 if (IFCAP_RXCSUM
& ifp
->if_capenable
) {
4008 ifp
->if_capenable
&= ~IFCAP_RXCSUM
;
4011 ifp
->if_capenable
|= IFCAP_RXCSUM
;
4015 if (mask
& IFCAP_TSO4
) {
4016 if (IFCAP_TSO4
& ifp
->if_capenable
) {
4017 ifp
->if_capenable
&= ~IFCAP_TSO4
;
4018 ifp
->if_hwassist
&= ~CSUM_TSO
;
4019 } else if (IFCAP_TXCSUM
& ifp
->if_capenable
) {
4020 ifp
->if_capenable
|= IFCAP_TSO4
;
4021 ifp
->if_hwassist
|= CSUM_TSO
;
4023 kprintf("mxge requires tx checksum offload"
4024 " be enabled to use TSO\n");
4028 if (mask
& IFCAP_LRO
) {
4029 if (IFCAP_LRO
& ifp
->if_capenable
)
4030 err
= mxge_change_lro_locked(sc
, 0);
4032 err
= mxge_change_lro_locked(sc
, mxge_lro_cnt
);
4034 if (mask
& IFCAP_VLAN_HWTAGGING
)
4035 ifp
->if_capenable
^= IFCAP_VLAN_HWTAGGING
;
4036 VLAN_CAPABILITIES(ifp
);
4041 err
= ifmedia_ioctl(ifp
, (struct ifreq
*)data
,
4042 &sc
->media
, command
);
4052 mxge_fetch_tunables(mxge_softc_t
*sc
)
4055 TUNABLE_INT_FETCH("hw.mxge.max_slices", &mxge_max_slices
);
4056 TUNABLE_INT_FETCH("hw.mxge.flow_control_enabled",
4057 &mxge_flow_control
);
4058 TUNABLE_INT_FETCH("hw.mxge.intr_coal_delay",
4059 &mxge_intr_coal_delay
);
4060 TUNABLE_INT_FETCH("hw.mxge.nvidia_ecrc_enable",
4061 &mxge_nvidia_ecrc_enable
);
4062 TUNABLE_INT_FETCH("hw.mxge.force_firmware",
4063 &mxge_force_firmware
);
4064 TUNABLE_INT_FETCH("hw.mxge.deassert_wait",
4065 &mxge_deassert_wait
);
4066 TUNABLE_INT_FETCH("hw.mxge.verbose",
4068 TUNABLE_INT_FETCH("hw.mxge.ticks", &mxge_ticks
);
4069 TUNABLE_INT_FETCH("hw.mxge.lro_cnt", &sc
->lro_cnt
);
4070 TUNABLE_INT_FETCH("hw.mxge.always_promisc", &mxge_always_promisc
);
4071 TUNABLE_INT_FETCH("hw.mxge.rss_hash_type", &mxge_rss_hash_type
);
4072 TUNABLE_INT_FETCH("hw.mxge.initial_mtu", &mxge_initial_mtu
);
4073 if (sc
->lro_cnt
!= 0)
4074 mxge_lro_cnt
= sc
->lro_cnt
;
4078 if (mxge_intr_coal_delay
< 0 || mxge_intr_coal_delay
> 10*1000)
4079 mxge_intr_coal_delay
= 30;
4080 if (mxge_ticks
== 0)
4081 mxge_ticks
= hz
/ 2;
4082 sc
->pause
= mxge_flow_control
;
4083 if (mxge_rss_hash_type
< MXGEFW_RSS_HASH_TYPE_IPV4
4084 || mxge_rss_hash_type
> MXGEFW_RSS_HASH_TYPE_MAX
) {
4085 mxge_rss_hash_type
= MXGEFW_RSS_HASH_TYPE_SRC_PORT
;
4087 if (mxge_initial_mtu
> ETHERMTU_JUMBO
||
4088 mxge_initial_mtu
< ETHER_MIN_LEN
)
4089 mxge_initial_mtu
= ETHERMTU_JUMBO
;
4094 mxge_free_slices(mxge_softc_t
*sc
)
4096 struct mxge_slice_state
*ss
;
4103 for (i
= 0; i
< sc
->num_slices
; i
++) {
4105 if (ss
->fw_stats
!= NULL
) {
4106 mxge_dma_free(&ss
->fw_stats_dma
);
4107 ss
->fw_stats
= NULL
;
4108 #ifdef IFNET_BUF_RING
4109 if (ss
->tx
.br
!= NULL
) {
4110 drbr_free(ss
->tx
.br
, M_DEVBUF
);
4115 if (ss
->rx_done
.entry
!= NULL
) {
4116 mxge_dma_free(&ss
->rx_done
.dma
);
4117 ss
->rx_done
.entry
= NULL
;
4120 kfree(sc
->ss
, M_DEVBUF
);
4125 mxge_alloc_slices(mxge_softc_t
*sc
)
4128 struct mxge_slice_state
*ss
;
4130 int err
, i
, max_intr_slots
;
4132 err
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_RX_RING_SIZE
, &cmd
);
4134 device_printf(sc
->dev
, "Cannot determine rx ring size\n");
4137 sc
->rx_ring_size
= cmd
.data0
;
4138 max_intr_slots
= 2 * (sc
->rx_ring_size
/ sizeof (mcp_dma_addr_t
));
4140 bytes
= sizeof (*sc
->ss
) * sc
->num_slices
;
4141 sc
->ss
= kmalloc(bytes
, M_DEVBUF
, M_NOWAIT
| M_ZERO
);
4144 for (i
= 0; i
< sc
->num_slices
; i
++) {
4149 /* allocate per-slice rx interrupt queues */
4151 bytes
= max_intr_slots
* sizeof (*ss
->rx_done
.entry
);
4152 err
= mxge_dma_alloc(sc
, &ss
->rx_done
.dma
, bytes
, 4096);
4155 ss
->rx_done
.entry
= ss
->rx_done
.dma
.addr
;
4156 bzero(ss
->rx_done
.entry
, bytes
);
4159 * allocate the per-slice firmware stats; stats
4160 * (including tx) are used used only on the first
4163 #ifndef IFNET_BUF_RING
4168 bytes
= sizeof (*ss
->fw_stats
);
4169 err
= mxge_dma_alloc(sc
, &ss
->fw_stats_dma
,
4170 sizeof (*ss
->fw_stats
), 64);
4173 ss
->fw_stats
= (mcp_irq_data_t
*)ss
->fw_stats_dma
.addr
;
4174 #ifdef IFNET_BUF_RING
4175 ss
->tx
.br
= buf_ring_alloc(2048, M_DEVBUF
, M_WAITOK
,
4183 mxge_free_slices(sc
);
4188 mxge_slice_probe(mxge_softc_t
*sc
)
4192 int msix_cnt
, status
, max_intr_slots
;
4196 * don't enable multiple slices if they are not enabled,
4197 * or if this is not an SMP system
4200 if (mxge_max_slices
== 0 || mxge_max_slices
== 1 || ncpus
< 2)
4203 /* see how many MSI-X interrupts are available */
4204 msix_cnt
= pci_msix_count(sc
->dev
);
4208 /* now load the slice aware firmware see what it supports */
4209 old_fw
= sc
->fw_name
;
4210 if (old_fw
== mxge_fw_aligned
)
4211 sc
->fw_name
= mxge_fw_rss_aligned
;
4213 sc
->fw_name
= mxge_fw_rss_unaligned
;
4214 status
= mxge_load_firmware(sc
, 0);
4216 device_printf(sc
->dev
, "Falling back to a single slice\n");
4220 /* try to send a reset command to the card to see if it
4222 memset(&cmd
, 0, sizeof (cmd
));
4223 status
= mxge_send_cmd(sc
, MXGEFW_CMD_RESET
, &cmd
);
4225 device_printf(sc
->dev
, "failed reset\n");
4229 /* get rx ring size */
4230 status
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_RX_RING_SIZE
, &cmd
);
4232 device_printf(sc
->dev
, "Cannot determine rx ring size\n");
4235 max_intr_slots
= 2 * (cmd
.data0
/ sizeof (mcp_dma_addr_t
));
4237 /* tell it the size of the interrupt queues */
4238 cmd
.data0
= max_intr_slots
* sizeof (struct mcp_slot
);
4239 status
= mxge_send_cmd(sc
, MXGEFW_CMD_SET_INTRQ_SIZE
, &cmd
);
4241 device_printf(sc
->dev
, "failed MXGEFW_CMD_SET_INTRQ_SIZE\n");
4245 /* ask the maximum number of slices it supports */
4246 status
= mxge_send_cmd(sc
, MXGEFW_CMD_GET_MAX_RSS_QUEUES
, &cmd
);
4248 device_printf(sc
->dev
,
4249 "failed MXGEFW_CMD_GET_MAX_RSS_QUEUES\n");
4252 sc
->num_slices
= cmd
.data0
;
4253 if (sc
->num_slices
> msix_cnt
)
4254 sc
->num_slices
= msix_cnt
;
4256 if (mxge_max_slices
== -1) {
4257 /* cap to number of CPUs in system */
4258 if (sc
->num_slices
> ncpus
)
4259 sc
->num_slices
= ncpus
;
4261 if (sc
->num_slices
> mxge_max_slices
)
4262 sc
->num_slices
= mxge_max_slices
;
4264 /* make sure it is a power of two */
4265 while (sc
->num_slices
& (sc
->num_slices
- 1))
4269 device_printf(sc
->dev
, "using %d slices\n",
4275 sc
->fw_name
= old_fw
;
4276 (void) mxge_load_firmware(sc
, 0);
4280 mxge_add_msix_irqs(mxge_softc_t
*sc
)
4283 int count
, err
, i
, rid
;
4286 sc
->msix_table_res
= bus_alloc_resource_any(sc
->dev
, SYS_RES_MEMORY
,
4289 if (sc
->msix_table_res
== NULL
) {
4290 device_printf(sc
->dev
, "couldn't alloc MSIX table res\n");
4294 count
= sc
->num_slices
;
4295 err
= pci_alloc_msix(sc
->dev
, &count
);
4297 device_printf(sc
->dev
, "pci_alloc_msix: failed, wanted %d"
4298 "err = %d \n", sc
->num_slices
, err
);
4299 goto abort_with_msix_table
;
4301 if (count
< sc
->num_slices
) {
4302 device_printf(sc
->dev
, "pci_alloc_msix: need %d, got %d\n",
4303 count
, sc
->num_slices
);
4304 device_printf(sc
->dev
,
4305 "Try setting hw.mxge.max_slices to %d\n",
4308 goto abort_with_msix
;
4310 bytes
= sizeof (*sc
->msix_irq_res
) * sc
->num_slices
;
4311 sc
->msix_irq_res
= kmalloc(bytes
, M_DEVBUF
, M_NOWAIT
|M_ZERO
);
4312 if (sc
->msix_irq_res
== NULL
) {
4314 goto abort_with_msix
;
4317 for (i
= 0; i
< sc
->num_slices
; i
++) {
4319 sc
->msix_irq_res
[i
] = bus_alloc_resource_any(sc
->dev
,
4322 if (sc
->msix_irq_res
[i
] == NULL
) {
4323 device_printf(sc
->dev
, "couldn't allocate IRQ res"
4324 " for message %d\n", i
);
4326 goto abort_with_res
;
4330 bytes
= sizeof (*sc
->msix_ih
) * sc
->num_slices
;
4331 sc
->msix_ih
= kmalloc(bytes
, M_DEVBUF
, M_NOWAIT
|M_ZERO
);
4333 for (i
= 0; i
< sc
->num_slices
; i
++) {
4334 err
= bus_setup_intr(sc
->dev
, sc
->msix_irq_res
[i
],
4336 mxge_intr
, &sc
->ss
[i
], &sc
->msix_ih
[i
],
4337 sc
->ifp
->if_serializer
);
4339 device_printf(sc
->dev
, "couldn't setup intr for "
4341 goto abort_with_intr
;
4346 device_printf(sc
->dev
, "using %d msix IRQs:",
4348 for (i
= 0; i
< sc
->num_slices
; i
++)
4349 kprintf(" %ld", rman_get_start(sc
->msix_irq_res
[i
]));
4355 for (i
= 0; i
< sc
->num_slices
; i
++) {
4356 if (sc
->msix_ih
[i
] != NULL
) {
4357 bus_teardown_intr(sc
->dev
, sc
->msix_irq_res
[i
],
4359 sc
->msix_ih
[i
] = NULL
;
4362 kfree(sc
->msix_ih
, M_DEVBUF
);
4366 for (i
= 0; i
< sc
->num_slices
; i
++) {
4368 if (sc
->msix_irq_res
[i
] != NULL
)
4369 bus_release_resource(sc
->dev
, SYS_RES_IRQ
, rid
,
4370 sc
->msix_irq_res
[i
]);
4371 sc
->msix_irq_res
[i
] = NULL
;
4373 kfree(sc
->msix_irq_res
, M_DEVBUF
);
4377 pci_release_msi(sc
->dev
);
4379 abort_with_msix_table
:
4380 bus_release_resource(sc
->dev
, SYS_RES_MEMORY
, PCIR_BAR(2),
4381 sc
->msix_table_res
);
4387 mxge_add_single_irq(mxge_softc_t
*sc
)
4389 int count
, err
, rid
;
4391 count
= pci_msi_count(sc
->dev
);
4392 if (count
== 1 && pci_alloc_msi(sc
->dev
, &count
) == 0) {
4398 sc
->irq_res
= bus_alloc_resource(sc
->dev
, SYS_RES_IRQ
, &rid
, 0, ~0,
4399 1, RF_SHAREABLE
| RF_ACTIVE
);
4400 if (sc
->irq_res
== NULL
) {
4401 device_printf(sc
->dev
, "could not alloc interrupt\n");
4405 device_printf(sc
->dev
, "using %s irq %ld\n",
4406 sc
->legacy_irq
? "INTx" : "MSI",
4407 rman_get_start(sc
->irq_res
));
4408 err
= bus_setup_intr(sc
->dev
, sc
->irq_res
,
4410 mxge_intr
, &sc
->ss
[0], &sc
->ih
,
4411 sc
->ifp
->if_serializer
);
4413 bus_release_resource(sc
->dev
, SYS_RES_IRQ
,
4414 sc
->legacy_irq
? 0 : 1, sc
->irq_res
);
4415 if (!sc
->legacy_irq
)
4416 pci_release_msi(sc
->dev
);
4422 mxge_rem_msix_irqs(mxge_softc_t
*sc
)
4426 for (i
= 0; i
< sc
->num_slices
; i
++) {
4427 if (sc
->msix_ih
[i
] != NULL
) {
4428 bus_teardown_intr(sc
->dev
, sc
->msix_irq_res
[i
],
4430 sc
->msix_ih
[i
] = NULL
;
4433 kfree(sc
->msix_ih
, M_DEVBUF
);
4435 for (i
= 0; i
< sc
->num_slices
; i
++) {
4437 if (sc
->msix_irq_res
[i
] != NULL
)
4438 bus_release_resource(sc
->dev
, SYS_RES_IRQ
, rid
,
4439 sc
->msix_irq_res
[i
]);
4440 sc
->msix_irq_res
[i
] = NULL
;
4442 kfree(sc
->msix_irq_res
, M_DEVBUF
);
4444 bus_release_resource(sc
->dev
, SYS_RES_MEMORY
, PCIR_BAR(2),
4445 sc
->msix_table_res
);
4447 pci_release_msi(sc
->dev
);
4452 mxge_rem_single_irq(mxge_softc_t
*sc
)
4454 bus_teardown_intr(sc
->dev
, sc
->irq_res
, sc
->ih
);
4455 bus_release_resource(sc
->dev
, SYS_RES_IRQ
,
4456 sc
->legacy_irq
? 0 : 1, sc
->irq_res
);
4457 if (!sc
->legacy_irq
)
4458 pci_release_msi(sc
->dev
);
4462 mxge_rem_irq(mxge_softc_t
*sc
)
4464 if (sc
->num_slices
> 1)
4465 mxge_rem_msix_irqs(sc
);
4467 mxge_rem_single_irq(sc
);
4471 mxge_add_irq(mxge_softc_t
*sc
)
4475 if (sc
->num_slices
> 1)
4476 err
= mxge_add_msix_irqs(sc
);
4478 err
= mxge_add_single_irq(sc
);
4480 if (0 && err
== 0 && sc
->num_slices
> 1) {
4481 mxge_rem_msix_irqs(sc
);
4482 err
= mxge_add_msix_irqs(sc
);
4489 mxge_attach(device_t dev
)
4491 mxge_softc_t
*sc
= device_get_softc(dev
);
4492 struct ifnet
*ifp
= &sc
->arpcom
.ac_if
;
4496 * avoid rewriting half the lines in this file to use
4497 * &sc->arpcom.ac_if instead
4501 mxge_fetch_tunables(sc
);
4503 err
= bus_dma_tag_create(NULL
, /* parent */
4506 BUS_SPACE_MAXADDR
, /* low */
4507 BUS_SPACE_MAXADDR
, /* high */
4508 NULL
, NULL
, /* filter */
4509 65536 + 256, /* maxsize */
4510 MXGE_MAX_SEND_DESC
, /* num segs */
4511 65536, /* maxsegsize */
4513 &sc
->parent_dmat
); /* tag */
4516 device_printf(sc
->dev
, "Err %d allocating parent dmat\n",
4518 goto abort_with_nothing
;
4522 if_initname(ifp
, device_get_name(dev
), device_get_unit(dev
));
4524 callout_init_mp(&sc
->co_hdl
);
4526 mxge_setup_cfg_space(sc
);
4528 /* Map the board into the kernel */
4530 sc
->mem_res
= bus_alloc_resource(dev
, SYS_RES_MEMORY
, &rid
, 0,
4532 if (sc
->mem_res
== NULL
) {
4533 device_printf(dev
, "could not map memory\n");
4535 goto abort_with_nothing
;
4537 sc
->sram
= rman_get_virtual(sc
->mem_res
);
4538 sc
->sram_size
= 2*1024*1024 - (2*(48*1024)+(32*1024)) - 0x100;
4539 if (sc
->sram_size
> rman_get_size(sc
->mem_res
)) {
4540 device_printf(dev
, "impossible memory region size %ld\n",
4541 rman_get_size(sc
->mem_res
));
4543 goto abort_with_mem_res
;
4546 /* make NULL terminated copy of the EEPROM strings section of
4548 bzero(sc
->eeprom_strings
, MXGE_EEPROM_STRINGS_SIZE
);
4549 bus_space_read_region_1(rman_get_bustag(sc
->mem_res
),
4550 rman_get_bushandle(sc
->mem_res
),
4551 sc
->sram_size
- MXGE_EEPROM_STRINGS_SIZE
,
4553 MXGE_EEPROM_STRINGS_SIZE
- 2);
4554 err
= mxge_parse_strings(sc
);
4556 goto abort_with_mem_res
;
4558 /* Enable write combining for efficient use of PCIe bus */
4561 /* Allocate the out of band dma memory */
4562 err
= mxge_dma_alloc(sc
, &sc
->cmd_dma
,
4563 sizeof (mxge_cmd_t
), 64);
4565 goto abort_with_mem_res
;
4566 sc
->cmd
= (mcp_cmd_response_t
*) sc
->cmd_dma
.addr
;
4567 err
= mxge_dma_alloc(sc
, &sc
->zeropad_dma
, 64, 64);
4569 goto abort_with_cmd_dma
;
4571 err
= mxge_dma_alloc(sc
, &sc
->dmabench_dma
, 4096, 4096);
4573 goto abort_with_zeropad_dma
;
4575 /* select & load the firmware */
4576 err
= mxge_select_firmware(sc
);
4578 goto abort_with_dmabench
;
4579 sc
->intr_coal_delay
= mxge_intr_coal_delay
;
4581 mxge_slice_probe(sc
);
4582 err
= mxge_alloc_slices(sc
);
4584 goto abort_with_dmabench
;
4586 err
= mxge_reset(sc
, 0);
4588 goto abort_with_slices
;
4590 err
= mxge_alloc_rings(sc
);
4592 device_printf(sc
->dev
, "failed to allocate rings\n");
4593 goto abort_with_dmabench
;
4596 ifp
->if_baudrate
= IF_Gbps(10UL);
4597 ifp
->if_capabilities
= IFCAP_RXCSUM
| IFCAP_TXCSUM
| IFCAP_TSO4
|
4600 ifp
->if_capabilities
|= IFCAP_LRO
;
4603 #ifdef MXGE_NEW_VLAN_API
4604 ifp
->if_capabilities
|= IFCAP_VLAN_HWTAGGING
| IFCAP_VLAN_HWCSUM
;
4607 sc
->max_mtu
= mxge_max_mtu(sc
);
4608 if (sc
->max_mtu
>= 9000)
4609 ifp
->if_capabilities
|= IFCAP_JUMBO_MTU
;
4611 device_printf(dev
, "MTU limited to %d. Install "
4612 "latest firmware for 9000 byte jumbo support\n",
4613 sc
->max_mtu
- ETHER_HDR_LEN
);
4614 ifp
->if_hwassist
= CSUM_TCP
| CSUM_UDP
| CSUM_TSO
;
4615 ifp
->if_capenable
= ifp
->if_capabilities
;
4616 if (sc
->lro_cnt
== 0)
4617 ifp
->if_capenable
&= ~IFCAP_LRO
;
4619 ifp
->if_init
= mxge_init
;
4621 ifp
->if_flags
= IFF_BROADCAST
| IFF_SIMPLEX
| IFF_MULTICAST
;
4622 ifp
->if_ioctl
= mxge_ioctl
;
4623 ifp
->if_start
= mxge_start
;
4624 /* Initialise the ifmedia structure */
4625 ifmedia_init(&sc
->media
, 0, mxge_media_change
,
4627 mxge_set_media(sc
, IFM_ETHER
| IFM_AUTO
);
4628 mxge_media_probe(sc
);
4630 ether_ifattach(ifp
, sc
->mac_addr
, NULL
);
4631 /* ether_ifattach sets mtu to ETHERMTU */
4632 if (mxge_initial_mtu
!= ETHERMTU
) {
4633 lwkt_serialize_enter(ifp
->if_serializer
);
4634 mxge_change_mtu(sc
, mxge_initial_mtu
);
4635 lwkt_serialize_exit(ifp
->if_serializer
);
4637 /* must come after ether_ifattach() */
4638 err
= mxge_add_irq(sc
);
4640 device_printf(sc
->dev
, "failed to add irq\n");
4641 goto abort_with_rings
;
4644 mxge_add_sysctls(sc
);
4645 #ifdef IFNET_BUF_RING
4646 ifp
->if_transmit
= mxge_transmit
;
4647 ifp
->if_qflush
= mxge_qflush
;
4652 mxge_free_rings(sc
);
4654 mxge_free_slices(sc
);
4655 abort_with_dmabench
:
4656 mxge_dma_free(&sc
->dmabench_dma
);
4657 abort_with_zeropad_dma
:
4658 mxge_dma_free(&sc
->zeropad_dma
);
4660 mxge_dma_free(&sc
->cmd_dma
);
4662 bus_release_resource(dev
, SYS_RES_MEMORY
, PCIR_BARS
, sc
->mem_res
);
4663 pci_disable_busmaster(dev
);
4664 bus_dma_tag_destroy(sc
->parent_dmat
);
4670 mxge_detach(device_t dev
)
4672 mxge_softc_t
*sc
= device_get_softc(dev
);
4674 lwkt_serialize_enter(sc
->ifp
->if_serializer
);
4676 if (sc
->ifp
->if_flags
& IFF_RUNNING
)
4679 * XXX: race: the callout callback could be spinning on
4680 * the serializer and run anyway
4682 callout_stop(&sc
->co_hdl
);
4683 lwkt_serialize_exit(sc
->ifp
->if_serializer
);
4685 ether_ifdetach(sc
->ifp
);
4686 ifmedia_removeall(&sc
->media
);
4687 mxge_dummy_rdma(sc
, 0);
4688 mxge_rem_sysctls(sc
);
4690 mxge_free_rings(sc
);
4691 mxge_free_slices(sc
);
4692 mxge_dma_free(&sc
->dmabench_dma
);
4693 mxge_dma_free(&sc
->zeropad_dma
);
4694 mxge_dma_free(&sc
->cmd_dma
);
4695 bus_release_resource(dev
, SYS_RES_MEMORY
, PCIR_BARS
, sc
->mem_res
);
4696 pci_disable_busmaster(dev
);
4697 bus_dma_tag_destroy(sc
->parent_dmat
);
4702 mxge_shutdown(device_t dev
)
4708 This file uses Myri10GE driver indentation.
4711 c-file-style:"linux"