2 * tg3.c: Broadcom Tigon3 ethernet driver.
4 * Copyright (C) 2001, 2002, 2003, 2004 David S. Miller (davem@redhat.com)
5 * Copyright (C) 2001, 2002, 2003 Jeff Garzik (jgarzik@pobox.com)
6 * Copyright (C) 2004 Sun Microsystems Inc.
7 * Copyright (C) 2005-2007 Broadcom Corporation.
10 * Derived from proprietary unpublished source code,
11 * Copyright (C) 2000-2003 Broadcom Corporation.
13 * Permission is hereby granted for the distribution of this firmware
14 * data in hexadecimal or equivalent format, provided this copyright
15 * notice is accompanying it.
19 #include <linux/module.h>
20 #include <linux/moduleparam.h>
21 #include <linux/kernel.h>
22 #include <linux/types.h>
23 #include <linux/compiler.h>
24 #include <linux/slab.h>
25 #include <linux/delay.h>
27 #include <linux/init.h>
28 #include <linux/ioport.h>
29 #include <linux/pci.h>
30 #include <linux/netdevice.h>
31 #include <linux/etherdevice.h>
32 #include <linux/skbuff.h>
33 #include <linux/ethtool.h>
34 #include <linux/mii.h>
35 #include <linux/if_vlan.h>
37 #include <linux/tcp.h>
38 #include <linux/workqueue.h>
39 #include <linux/prefetch.h>
40 #include <linux/dma-mapping.h>
42 #include <net/checksum.h>
45 #include <asm/system.h>
47 #include <asm/byteorder.h>
48 #include <asm/uaccess.h>
51 #include <asm/idprom.h>
55 #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE)
56 #define TG3_VLAN_TAG_USED 1
58 #define TG3_VLAN_TAG_USED 0
61 #define TG3_TSO_SUPPORT 1
65 #define DRV_MODULE_NAME "tg3"
66 #define PFX DRV_MODULE_NAME ": "
67 #define DRV_MODULE_VERSION "3.86"
68 #define DRV_MODULE_RELDATE "November 9, 2007"
70 #define TG3_DEF_MAC_MODE 0
71 #define TG3_DEF_RX_MODE 0
72 #define TG3_DEF_TX_MODE 0
73 #define TG3_DEF_MSG_ENABLE \
83 /* length of time before we decide the hardware is borked,
84 * and dev->tx_timeout() should be called to fix the problem
86 #define TG3_TX_TIMEOUT (5 * HZ)
88 /* hardware minimum and maximum for a single frame's data payload */
89 #define TG3_MIN_MTU 60
90 #define TG3_MAX_MTU(tp) \
91 ((tp->tg3_flags2 & TG3_FLG2_JUMBO_CAPABLE) ? 9000 : 1500)
93 /* These numbers seem to be hard coded in the NIC firmware somehow.
94 * You can't change the ring sizes, but you can change where you place
95 * them in the NIC onboard memory.
97 #define TG3_RX_RING_SIZE 512
98 #define TG3_DEF_RX_RING_PENDING 200
99 #define TG3_RX_JUMBO_RING_SIZE 256
100 #define TG3_DEF_RX_JUMBO_RING_PENDING 100
102 /* Do not place this n-ring entries value into the tp struct itself,
103 * we really want to expose these constants to GCC so that modulo et
104 * al. operations are done with shifts and masks instead of with
105 * hw multiply/modulo instructions. Another solution would be to
106 * replace things like '% foo' with '& (foo - 1)'.
108 #define TG3_RX_RCB_RING_SIZE(tp) \
109 ((tp->tg3_flags2 & TG3_FLG2_5705_PLUS) ? 512 : 1024)
111 #define TG3_TX_RING_SIZE 512
112 #define TG3_DEF_TX_RING_PENDING (TG3_TX_RING_SIZE - 1)
114 #define TG3_RX_RING_BYTES (sizeof(struct tg3_rx_buffer_desc) * \
116 #define TG3_RX_JUMBO_RING_BYTES (sizeof(struct tg3_rx_buffer_desc) * \
117 TG3_RX_JUMBO_RING_SIZE)
118 #define TG3_RX_RCB_RING_BYTES(tp) (sizeof(struct tg3_rx_buffer_desc) * \
119 TG3_RX_RCB_RING_SIZE(tp))
120 #define TG3_TX_RING_BYTES (sizeof(struct tg3_tx_buffer_desc) * \
122 #define NEXT_TX(N) (((N) + 1) & (TG3_TX_RING_SIZE - 1))
124 #define RX_PKT_BUF_SZ (1536 + tp->rx_offset + 64)
125 #define RX_JUMBO_PKT_BUF_SZ (9046 + tp->rx_offset + 64)
127 /* minimum number of free TX descriptors required to wake up TX process */
128 #define TG3_TX_WAKEUP_THRESH(tp) ((tp)->tx_pending / 4)
130 /* number of ETHTOOL_GSTATS u64's */
131 #define TG3_NUM_STATS (sizeof(struct tg3_ethtool_stats)/sizeof(u64))
133 #define TG3_NUM_TEST 6
135 static char version
[] __devinitdata
=
136 DRV_MODULE_NAME
".c:v" DRV_MODULE_VERSION
" (" DRV_MODULE_RELDATE
")\n";
138 MODULE_AUTHOR("David S. Miller (davem@redhat.com) and Jeff Garzik (jgarzik@pobox.com)");
139 MODULE_DESCRIPTION("Broadcom Tigon3 ethernet driver");
140 MODULE_LICENSE("GPL");
141 MODULE_VERSION(DRV_MODULE_VERSION
);
143 static int tg3_debug
= -1; /* -1 == use TG3_DEF_MSG_ENABLE as value */
144 module_param(tg3_debug
, int, 0);
145 MODULE_PARM_DESC(tg3_debug
, "Tigon3 bitmapped debugging message enable value");
147 static struct pci_device_id tg3_pci_tbl
[] = {
148 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5700
)},
149 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5701
)},
150 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5702
)},
151 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5703
)},
152 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5704
)},
153 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5702FE
)},
154 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5705
)},
155 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5705_2
)},
156 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5705M
)},
157 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5705M_2
)},
158 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5702X
)},
159 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5703X
)},
160 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5704S
)},
161 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5702A3
)},
162 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5703A3
)},
163 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5782
)},
164 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5788
)},
165 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5789
)},
166 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5901
)},
167 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5901_2
)},
168 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5704S_2
)},
169 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5705F
)},
170 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5720
)},
171 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5721
)},
172 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5722
)},
173 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5750
)},
174 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5751
)},
175 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5750M
)},
176 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5751M
)},
177 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5751F
)},
178 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5752
)},
179 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5752M
)},
180 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5753
)},
181 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5753M
)},
182 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5753F
)},
183 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5754
)},
184 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5754M
)},
185 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5755
)},
186 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5755M
)},
187 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5756
)},
188 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5786
)},
189 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5787
)},
190 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5787M
)},
191 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5787F
)},
192 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5714
)},
193 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5714S
)},
194 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5715
)},
195 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5715S
)},
196 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5780
)},
197 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5780S
)},
198 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5781
)},
199 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5906
)},
200 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5906M
)},
201 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5784
)},
202 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5764
)},
203 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5723
)},
204 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5761
)},
205 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM
, PCI_DEVICE_ID_TIGON3_5761E
)},
206 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT
, PCI_DEVICE_ID_SYSKONNECT_9DXX
)},
207 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT
, PCI_DEVICE_ID_SYSKONNECT_9MXX
)},
208 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA
, PCI_DEVICE_ID_ALTIMA_AC1000
)},
209 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA
, PCI_DEVICE_ID_ALTIMA_AC1001
)},
210 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA
, PCI_DEVICE_ID_ALTIMA_AC1003
)},
211 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA
, PCI_DEVICE_ID_ALTIMA_AC9100
)},
212 {PCI_DEVICE(PCI_VENDOR_ID_APPLE
, PCI_DEVICE_ID_APPLE_TIGON3
)},
216 MODULE_DEVICE_TABLE(pci
, tg3_pci_tbl
);
218 static const struct {
219 const char string
[ETH_GSTRING_LEN
];
220 } ethtool_stats_keys
[TG3_NUM_STATS
] = {
223 { "rx_ucast_packets" },
224 { "rx_mcast_packets" },
225 { "rx_bcast_packets" },
227 { "rx_align_errors" },
228 { "rx_xon_pause_rcvd" },
229 { "rx_xoff_pause_rcvd" },
230 { "rx_mac_ctrl_rcvd" },
231 { "rx_xoff_entered" },
232 { "rx_frame_too_long_errors" },
234 { "rx_undersize_packets" },
235 { "rx_in_length_errors" },
236 { "rx_out_length_errors" },
237 { "rx_64_or_less_octet_packets" },
238 { "rx_65_to_127_octet_packets" },
239 { "rx_128_to_255_octet_packets" },
240 { "rx_256_to_511_octet_packets" },
241 { "rx_512_to_1023_octet_packets" },
242 { "rx_1024_to_1522_octet_packets" },
243 { "rx_1523_to_2047_octet_packets" },
244 { "rx_2048_to_4095_octet_packets" },
245 { "rx_4096_to_8191_octet_packets" },
246 { "rx_8192_to_9022_octet_packets" },
253 { "tx_flow_control" },
255 { "tx_single_collisions" },
256 { "tx_mult_collisions" },
258 { "tx_excessive_collisions" },
259 { "tx_late_collisions" },
260 { "tx_collide_2times" },
261 { "tx_collide_3times" },
262 { "tx_collide_4times" },
263 { "tx_collide_5times" },
264 { "tx_collide_6times" },
265 { "tx_collide_7times" },
266 { "tx_collide_8times" },
267 { "tx_collide_9times" },
268 { "tx_collide_10times" },
269 { "tx_collide_11times" },
270 { "tx_collide_12times" },
271 { "tx_collide_13times" },
272 { "tx_collide_14times" },
273 { "tx_collide_15times" },
274 { "tx_ucast_packets" },
275 { "tx_mcast_packets" },
276 { "tx_bcast_packets" },
277 { "tx_carrier_sense_errors" },
281 { "dma_writeq_full" },
282 { "dma_write_prioq_full" },
286 { "rx_threshold_hit" },
288 { "dma_readq_full" },
289 { "dma_read_prioq_full" },
290 { "tx_comp_queue_full" },
292 { "ring_set_send_prod_index" },
293 { "ring_status_update" },
295 { "nic_avoided_irqs" },
296 { "nic_tx_threshold_hit" }
299 static const struct {
300 const char string
[ETH_GSTRING_LEN
];
301 } ethtool_test_keys
[TG3_NUM_TEST
] = {
302 { "nvram test (online) " },
303 { "link test (online) " },
304 { "register test (offline)" },
305 { "memory test (offline)" },
306 { "loopback test (offline)" },
307 { "interrupt test (offline)" },
310 static void tg3_write32(struct tg3
*tp
, u32 off
, u32 val
)
312 writel(val
, tp
->regs
+ off
);
315 static u32
tg3_read32(struct tg3
*tp
, u32 off
)
317 return (readl(tp
->regs
+ off
));
320 static void tg3_ape_write32(struct tg3
*tp
, u32 off
, u32 val
)
322 writel(val
, tp
->aperegs
+ off
);
325 static u32
tg3_ape_read32(struct tg3
*tp
, u32 off
)
327 return (readl(tp
->aperegs
+ off
));
330 static void tg3_write_indirect_reg32(struct tg3
*tp
, u32 off
, u32 val
)
334 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
335 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_BASE_ADDR
, off
);
336 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_DATA
, val
);
337 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
340 static void tg3_write_flush_reg32(struct tg3
*tp
, u32 off
, u32 val
)
342 writel(val
, tp
->regs
+ off
);
343 readl(tp
->regs
+ off
);
346 static u32
tg3_read_indirect_reg32(struct tg3
*tp
, u32 off
)
351 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
352 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_BASE_ADDR
, off
);
353 pci_read_config_dword(tp
->pdev
, TG3PCI_REG_DATA
, &val
);
354 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
358 static void tg3_write_indirect_mbox(struct tg3
*tp
, u32 off
, u32 val
)
362 if (off
== (MAILBOX_RCVRET_CON_IDX_0
+ TG3_64BIT_REG_LOW
)) {
363 pci_write_config_dword(tp
->pdev
, TG3PCI_RCV_RET_RING_CON_IDX
+
364 TG3_64BIT_REG_LOW
, val
);
367 if (off
== (MAILBOX_RCV_STD_PROD_IDX
+ TG3_64BIT_REG_LOW
)) {
368 pci_write_config_dword(tp
->pdev
, TG3PCI_STD_RING_PROD_IDX
+
369 TG3_64BIT_REG_LOW
, val
);
373 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
374 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_BASE_ADDR
, off
+ 0x5600);
375 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_DATA
, val
);
376 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
378 /* In indirect mode when disabling interrupts, we also need
379 * to clear the interrupt bit in the GRC local ctrl register.
381 if ((off
== (MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
)) &&
383 pci_write_config_dword(tp
->pdev
, TG3PCI_MISC_LOCAL_CTRL
,
384 tp
->grc_local_ctrl
|GRC_LCLCTRL_CLEARINT
);
388 static u32
tg3_read_indirect_mbox(struct tg3
*tp
, u32 off
)
393 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
394 pci_write_config_dword(tp
->pdev
, TG3PCI_REG_BASE_ADDR
, off
+ 0x5600);
395 pci_read_config_dword(tp
->pdev
, TG3PCI_REG_DATA
, &val
);
396 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
400 /* usec_wait specifies the wait time in usec when writing to certain registers
401 * where it is unsafe to read back the register without some delay.
402 * GRC_LOCAL_CTRL is one example if the GPIOs are toggled to switch power.
403 * TG3PCI_CLOCK_CTRL is another example if the clock frequencies are changed.
405 static void _tw32_flush(struct tg3
*tp
, u32 off
, u32 val
, u32 usec_wait
)
407 if ((tp
->tg3_flags
& TG3_FLAG_PCIX_TARGET_HWBUG
) ||
408 (tp
->tg3_flags2
& TG3_FLG2_ICH_WORKAROUND
))
409 /* Non-posted methods */
410 tp
->write32(tp
, off
, val
);
413 tg3_write32(tp
, off
, val
);
418 /* Wait again after the read for the posted method to guarantee that
419 * the wait time is met.
425 static inline void tw32_mailbox_flush(struct tg3
*tp
, u32 off
, u32 val
)
427 tp
->write32_mbox(tp
, off
, val
);
428 if (!(tp
->tg3_flags
& TG3_FLAG_MBOX_WRITE_REORDER
) &&
429 !(tp
->tg3_flags2
& TG3_FLG2_ICH_WORKAROUND
))
430 tp
->read32_mbox(tp
, off
);
433 static void tg3_write32_tx_mbox(struct tg3
*tp
, u32 off
, u32 val
)
435 void __iomem
*mbox
= tp
->regs
+ off
;
437 if (tp
->tg3_flags
& TG3_FLAG_TXD_MBOX_HWBUG
)
439 if (tp
->tg3_flags
& TG3_FLAG_MBOX_WRITE_REORDER
)
443 static u32
tg3_read32_mbox_5906(struct tg3
*tp
, u32 off
)
445 return (readl(tp
->regs
+ off
+ GRCMBOX_BASE
));
448 static void tg3_write32_mbox_5906(struct tg3
*tp
, u32 off
, u32 val
)
450 writel(val
, tp
->regs
+ off
+ GRCMBOX_BASE
);
453 #define tw32_mailbox(reg, val) tp->write32_mbox(tp, reg, val)
454 #define tw32_mailbox_f(reg, val) tw32_mailbox_flush(tp, (reg), (val))
455 #define tw32_rx_mbox(reg, val) tp->write32_rx_mbox(tp, reg, val)
456 #define tw32_tx_mbox(reg, val) tp->write32_tx_mbox(tp, reg, val)
457 #define tr32_mailbox(reg) tp->read32_mbox(tp, reg)
459 #define tw32(reg,val) tp->write32(tp, reg, val)
460 #define tw32_f(reg,val) _tw32_flush(tp,(reg),(val), 0)
461 #define tw32_wait_f(reg,val,us) _tw32_flush(tp,(reg),(val), (us))
462 #define tr32(reg) tp->read32(tp, reg)
464 static void tg3_write_mem(struct tg3
*tp
, u32 off
, u32 val
)
468 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) &&
469 (off
>= NIC_SRAM_STATS_BLK
) && (off
< NIC_SRAM_TX_BUFFER_DESC
))
472 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
473 if (tp
->tg3_flags
& TG3_FLAG_SRAM_USE_CONFIG
) {
474 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
, off
);
475 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_DATA
, val
);
477 /* Always leave this as zero. */
478 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
, 0);
480 tw32_f(TG3PCI_MEM_WIN_BASE_ADDR
, off
);
481 tw32_f(TG3PCI_MEM_WIN_DATA
, val
);
483 /* Always leave this as zero. */
484 tw32_f(TG3PCI_MEM_WIN_BASE_ADDR
, 0);
486 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
489 static void tg3_read_mem(struct tg3
*tp
, u32 off
, u32
*val
)
493 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) &&
494 (off
>= NIC_SRAM_STATS_BLK
) && (off
< NIC_SRAM_TX_BUFFER_DESC
)) {
499 spin_lock_irqsave(&tp
->indirect_lock
, flags
);
500 if (tp
->tg3_flags
& TG3_FLAG_SRAM_USE_CONFIG
) {
501 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
, off
);
502 pci_read_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_DATA
, val
);
504 /* Always leave this as zero. */
505 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
, 0);
507 tw32_f(TG3PCI_MEM_WIN_BASE_ADDR
, off
);
508 *val
= tr32(TG3PCI_MEM_WIN_DATA
);
510 /* Always leave this as zero. */
511 tw32_f(TG3PCI_MEM_WIN_BASE_ADDR
, 0);
513 spin_unlock_irqrestore(&tp
->indirect_lock
, flags
);
516 static void tg3_ape_lock_init(struct tg3
*tp
)
520 /* Make sure the driver hasn't any stale locks. */
521 for (i
= 0; i
< 8; i
++)
522 tg3_ape_write32(tp
, TG3_APE_LOCK_GRANT
+ 4 * i
,
523 APE_LOCK_GRANT_DRIVER
);
526 static int tg3_ape_lock(struct tg3
*tp
, int locknum
)
532 if (!(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
))
536 case TG3_APE_LOCK_MEM
:
544 tg3_ape_write32(tp
, TG3_APE_LOCK_REQ
+ off
, APE_LOCK_REQ_DRIVER
);
546 /* Wait for up to 1 millisecond to acquire lock. */
547 for (i
= 0; i
< 100; i
++) {
548 status
= tg3_ape_read32(tp
, TG3_APE_LOCK_GRANT
+ off
);
549 if (status
== APE_LOCK_GRANT_DRIVER
)
554 if (status
!= APE_LOCK_GRANT_DRIVER
) {
555 /* Revoke the lock request. */
556 tg3_ape_write32(tp
, TG3_APE_LOCK_GRANT
+ off
,
557 APE_LOCK_GRANT_DRIVER
);
565 static void tg3_ape_unlock(struct tg3
*tp
, int locknum
)
569 if (!(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
))
573 case TG3_APE_LOCK_MEM
:
580 tg3_ape_write32(tp
, TG3_APE_LOCK_GRANT
+ off
, APE_LOCK_GRANT_DRIVER
);
583 static void tg3_disable_ints(struct tg3
*tp
)
585 tw32(TG3PCI_MISC_HOST_CTRL
,
586 (tp
->misc_host_ctrl
| MISC_HOST_CTRL_MASK_PCI_INT
));
587 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
, 0x00000001);
590 static inline void tg3_cond_int(struct tg3
*tp
)
592 if (!(tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
) &&
593 (tp
->hw_status
->status
& SD_STATUS_UPDATED
))
594 tw32(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
| GRC_LCLCTRL_SETINT
);
596 tw32(HOSTCC_MODE
, tp
->coalesce_mode
|
597 (HOSTCC_MODE_ENABLE
| HOSTCC_MODE_NOW
));
600 static void tg3_enable_ints(struct tg3
*tp
)
605 tw32(TG3PCI_MISC_HOST_CTRL
,
606 (tp
->misc_host_ctrl
& ~MISC_HOST_CTRL_MASK_PCI_INT
));
607 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
,
608 (tp
->last_tag
<< 24));
609 if (tp
->tg3_flags2
& TG3_FLG2_1SHOT_MSI
)
610 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
,
611 (tp
->last_tag
<< 24));
615 static inline unsigned int tg3_has_work(struct tg3
*tp
)
617 struct tg3_hw_status
*sblk
= tp
->hw_status
;
618 unsigned int work_exists
= 0;
620 /* check for phy events */
621 if (!(tp
->tg3_flags
&
622 (TG3_FLAG_USE_LINKCHG_REG
|
623 TG3_FLAG_POLL_SERDES
))) {
624 if (sblk
->status
& SD_STATUS_LINK_CHG
)
627 /* check for RX/TX work to do */
628 if (sblk
->idx
[0].tx_consumer
!= tp
->tx_cons
||
629 sblk
->idx
[0].rx_producer
!= tp
->rx_rcb_ptr
)
636 * similar to tg3_enable_ints, but it accurately determines whether there
637 * is new work pending and can return without flushing the PIO write
638 * which reenables interrupts
640 static void tg3_restart_ints(struct tg3
*tp
)
642 tw32_mailbox(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
,
646 /* When doing tagged status, this work check is unnecessary.
647 * The last_tag we write above tells the chip which piece of
648 * work we've completed.
650 if (!(tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
) &&
652 tw32(HOSTCC_MODE
, tp
->coalesce_mode
|
653 (HOSTCC_MODE_ENABLE
| HOSTCC_MODE_NOW
));
656 static inline void tg3_netif_stop(struct tg3
*tp
)
658 tp
->dev
->trans_start
= jiffies
; /* prevent tx timeout */
659 napi_disable(&tp
->napi
);
660 netif_tx_disable(tp
->dev
);
663 static inline void tg3_netif_start(struct tg3
*tp
)
665 netif_wake_queue(tp
->dev
);
666 /* NOTE: unconditional netif_wake_queue is only appropriate
667 * so long as all callers are assured to have free tx slots
668 * (such as after tg3_init_hw)
670 napi_enable(&tp
->napi
);
671 tp
->hw_status
->status
|= SD_STATUS_UPDATED
;
675 static void tg3_switch_clocks(struct tg3
*tp
)
677 u32 clock_ctrl
= tr32(TG3PCI_CLOCK_CTRL
);
680 if ((tp
->tg3_flags
& TG3_FLAG_CPMU_PRESENT
) ||
681 (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
))
684 orig_clock_ctrl
= clock_ctrl
;
685 clock_ctrl
&= (CLOCK_CTRL_FORCE_CLKRUN
|
686 CLOCK_CTRL_CLKRUN_OENABLE
|
688 tp
->pci_clock_ctrl
= clock_ctrl
;
690 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
691 if (orig_clock_ctrl
& CLOCK_CTRL_625_CORE
) {
692 tw32_wait_f(TG3PCI_CLOCK_CTRL
,
693 clock_ctrl
| CLOCK_CTRL_625_CORE
, 40);
695 } else if ((orig_clock_ctrl
& CLOCK_CTRL_44MHZ_CORE
) != 0) {
696 tw32_wait_f(TG3PCI_CLOCK_CTRL
,
698 (CLOCK_CTRL_44MHZ_CORE
| CLOCK_CTRL_ALTCLK
),
700 tw32_wait_f(TG3PCI_CLOCK_CTRL
,
701 clock_ctrl
| (CLOCK_CTRL_ALTCLK
),
704 tw32_wait_f(TG3PCI_CLOCK_CTRL
, clock_ctrl
, 40);
707 #define PHY_BUSY_LOOPS 5000
709 static int tg3_readphy(struct tg3
*tp
, int reg
, u32
*val
)
715 if ((tp
->mi_mode
& MAC_MI_MODE_AUTO_POLL
) != 0) {
717 (tp
->mi_mode
& ~MAC_MI_MODE_AUTO_POLL
));
723 frame_val
= ((PHY_ADDR
<< MI_COM_PHY_ADDR_SHIFT
) &
724 MI_COM_PHY_ADDR_MASK
);
725 frame_val
|= ((reg
<< MI_COM_REG_ADDR_SHIFT
) &
726 MI_COM_REG_ADDR_MASK
);
727 frame_val
|= (MI_COM_CMD_READ
| MI_COM_START
);
729 tw32_f(MAC_MI_COM
, frame_val
);
731 loops
= PHY_BUSY_LOOPS
;
734 frame_val
= tr32(MAC_MI_COM
);
736 if ((frame_val
& MI_COM_BUSY
) == 0) {
738 frame_val
= tr32(MAC_MI_COM
);
746 *val
= frame_val
& MI_COM_DATA_MASK
;
750 if ((tp
->mi_mode
& MAC_MI_MODE_AUTO_POLL
) != 0) {
751 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
758 static int tg3_writephy(struct tg3
*tp
, int reg
, u32 val
)
764 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
&&
765 (reg
== MII_TG3_CTRL
|| reg
== MII_TG3_AUX_CTRL
))
768 if ((tp
->mi_mode
& MAC_MI_MODE_AUTO_POLL
) != 0) {
770 (tp
->mi_mode
& ~MAC_MI_MODE_AUTO_POLL
));
774 frame_val
= ((PHY_ADDR
<< MI_COM_PHY_ADDR_SHIFT
) &
775 MI_COM_PHY_ADDR_MASK
);
776 frame_val
|= ((reg
<< MI_COM_REG_ADDR_SHIFT
) &
777 MI_COM_REG_ADDR_MASK
);
778 frame_val
|= (val
& MI_COM_DATA_MASK
);
779 frame_val
|= (MI_COM_CMD_WRITE
| MI_COM_START
);
781 tw32_f(MAC_MI_COM
, frame_val
);
783 loops
= PHY_BUSY_LOOPS
;
786 frame_val
= tr32(MAC_MI_COM
);
787 if ((frame_val
& MI_COM_BUSY
) == 0) {
789 frame_val
= tr32(MAC_MI_COM
);
799 if ((tp
->mi_mode
& MAC_MI_MODE_AUTO_POLL
) != 0) {
800 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
807 static void tg3_phy_toggle_automdix(struct tg3
*tp
, int enable
)
811 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) ||
812 (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
))
815 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
818 if (!tg3_readphy(tp
, MII_TG3_EPHY_TEST
, &ephy
)) {
819 tg3_writephy(tp
, MII_TG3_EPHY_TEST
,
820 ephy
| MII_TG3_EPHY_SHADOW_EN
);
821 if (!tg3_readphy(tp
, MII_TG3_EPHYTST_MISCCTRL
, &phy
)) {
823 phy
|= MII_TG3_EPHYTST_MISCCTRL_MDIX
;
825 phy
&= ~MII_TG3_EPHYTST_MISCCTRL_MDIX
;
826 tg3_writephy(tp
, MII_TG3_EPHYTST_MISCCTRL
, phy
);
828 tg3_writephy(tp
, MII_TG3_EPHY_TEST
, ephy
);
831 phy
= MII_TG3_AUXCTL_MISC_RDSEL_MISC
|
832 MII_TG3_AUXCTL_SHDWSEL_MISC
;
833 if (!tg3_writephy(tp
, MII_TG3_AUX_CTRL
, phy
) &&
834 !tg3_readphy(tp
, MII_TG3_AUX_CTRL
, &phy
)) {
836 phy
|= MII_TG3_AUXCTL_MISC_FORCE_AMDIX
;
838 phy
&= ~MII_TG3_AUXCTL_MISC_FORCE_AMDIX
;
839 phy
|= MII_TG3_AUXCTL_MISC_WREN
;
840 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, phy
);
845 static void tg3_phy_set_wirespeed(struct tg3
*tp
)
849 if (tp
->tg3_flags2
& TG3_FLG2_NO_ETH_WIRE_SPEED
)
852 if (!tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x7007) &&
853 !tg3_readphy(tp
, MII_TG3_AUX_CTRL
, &val
))
854 tg3_writephy(tp
, MII_TG3_AUX_CTRL
,
855 (val
| (1 << 15) | (1 << 4)));
858 static int tg3_bmcr_reset(struct tg3
*tp
)
863 /* OK, reset it, and poll the BMCR_RESET bit until it
864 * clears or we time out.
866 phy_control
= BMCR_RESET
;
867 err
= tg3_writephy(tp
, MII_BMCR
, phy_control
);
873 err
= tg3_readphy(tp
, MII_BMCR
, &phy_control
);
877 if ((phy_control
& BMCR_RESET
) == 0) {
889 static int tg3_wait_macro_done(struct tg3
*tp
)
896 if (!tg3_readphy(tp
, 0x16, &tmp32
)) {
897 if ((tmp32
& 0x1000) == 0)
907 static int tg3_phy_write_and_check_testpat(struct tg3
*tp
, int *resetp
)
909 static const u32 test_pat
[4][6] = {
910 { 0x00005555, 0x00000005, 0x00002aaa, 0x0000000a, 0x00003456, 0x00000003 },
911 { 0x00002aaa, 0x0000000a, 0x00003333, 0x00000003, 0x0000789a, 0x00000005 },
912 { 0x00005a5a, 0x00000005, 0x00002a6a, 0x0000000a, 0x00001bcd, 0x00000003 },
913 { 0x00002a5a, 0x0000000a, 0x000033c3, 0x00000003, 0x00002ef1, 0x00000005 }
917 for (chan
= 0; chan
< 4; chan
++) {
920 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
,
921 (chan
* 0x2000) | 0x0200);
922 tg3_writephy(tp
, 0x16, 0x0002);
924 for (i
= 0; i
< 6; i
++)
925 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
,
928 tg3_writephy(tp
, 0x16, 0x0202);
929 if (tg3_wait_macro_done(tp
)) {
934 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
,
935 (chan
* 0x2000) | 0x0200);
936 tg3_writephy(tp
, 0x16, 0x0082);
937 if (tg3_wait_macro_done(tp
)) {
942 tg3_writephy(tp
, 0x16, 0x0802);
943 if (tg3_wait_macro_done(tp
)) {
948 for (i
= 0; i
< 6; i
+= 2) {
951 if (tg3_readphy(tp
, MII_TG3_DSP_RW_PORT
, &low
) ||
952 tg3_readphy(tp
, MII_TG3_DSP_RW_PORT
, &high
) ||
953 tg3_wait_macro_done(tp
)) {
959 if (low
!= test_pat
[chan
][i
] ||
960 high
!= test_pat
[chan
][i
+1]) {
961 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x000b);
962 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x4001);
963 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x4005);
973 static int tg3_phy_reset_chanpat(struct tg3
*tp
)
977 for (chan
= 0; chan
< 4; chan
++) {
980 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
,
981 (chan
* 0x2000) | 0x0200);
982 tg3_writephy(tp
, 0x16, 0x0002);
983 for (i
= 0; i
< 6; i
++)
984 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x000);
985 tg3_writephy(tp
, 0x16, 0x0202);
986 if (tg3_wait_macro_done(tp
))
993 static int tg3_phy_reset_5703_4_5(struct tg3
*tp
)
995 u32 reg32
, phy9_orig
;
996 int retries
, do_phy_reset
, err
;
1002 err
= tg3_bmcr_reset(tp
);
1008 /* Disable transmitter and interrupt. */
1009 if (tg3_readphy(tp
, MII_TG3_EXT_CTRL
, ®32
))
1013 tg3_writephy(tp
, MII_TG3_EXT_CTRL
, reg32
);
1015 /* Set full-duplex, 1000 mbps. */
1016 tg3_writephy(tp
, MII_BMCR
,
1017 BMCR_FULLDPLX
| TG3_BMCR_SPEED1000
);
1019 /* Set to master mode. */
1020 if (tg3_readphy(tp
, MII_TG3_CTRL
, &phy9_orig
))
1023 tg3_writephy(tp
, MII_TG3_CTRL
,
1024 (MII_TG3_CTRL_AS_MASTER
|
1025 MII_TG3_CTRL_ENABLE_AS_MASTER
));
1027 /* Enable SM_DSP_CLOCK and 6dB. */
1028 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0c00);
1030 /* Block the PHY control access. */
1031 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x8005);
1032 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0800);
1034 err
= tg3_phy_write_and_check_testpat(tp
, &do_phy_reset
);
1037 } while (--retries
);
1039 err
= tg3_phy_reset_chanpat(tp
);
1043 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x8005);
1044 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0000);
1046 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x8200);
1047 tg3_writephy(tp
, 0x16, 0x0000);
1049 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
1050 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) {
1051 /* Set Extended packet length bit for jumbo frames */
1052 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x4400);
1055 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0400);
1058 tg3_writephy(tp
, MII_TG3_CTRL
, phy9_orig
);
1060 if (!tg3_readphy(tp
, MII_TG3_EXT_CTRL
, ®32
)) {
1062 tg3_writephy(tp
, MII_TG3_EXT_CTRL
, reg32
);
1069 static void tg3_link_report(struct tg3
*);
1071 /* This will reset the tigon3 PHY if there is no valid
1072 * link unless the FORCE argument is non-zero.
1074 static int tg3_phy_reset(struct tg3
*tp
)
1079 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
1082 val
= tr32(GRC_MISC_CFG
);
1083 tw32_f(GRC_MISC_CFG
, val
& ~GRC_MISC_CFG_EPHY_IDDQ
);
1086 err
= tg3_readphy(tp
, MII_BMSR
, &phy_status
);
1087 err
|= tg3_readphy(tp
, MII_BMSR
, &phy_status
);
1091 if (netif_running(tp
->dev
) && netif_carrier_ok(tp
->dev
)) {
1092 netif_carrier_off(tp
->dev
);
1093 tg3_link_report(tp
);
1096 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
1097 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
||
1098 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) {
1099 err
= tg3_phy_reset_5703_4_5(tp
);
1105 err
= tg3_bmcr_reset(tp
);
1109 if (tp
->tg3_flags3
& TG3_FLG3_5761_5784_AX_FIXES
) {
1112 val
= tr32(TG3_CPMU_LSPD_1000MB_CLK
);
1113 if ((val
& CPMU_LSPD_1000MB_MACCLK_MASK
) ==
1114 CPMU_LSPD_1000MB_MACCLK_12_5
) {
1115 val
&= ~CPMU_LSPD_1000MB_MACCLK_MASK
;
1117 tw32_f(TG3_CPMU_LSPD_1000MB_CLK
, val
);
1120 /* Disable GPHY autopowerdown. */
1121 tg3_writephy(tp
, MII_TG3_MISC_SHDW
,
1122 MII_TG3_MISC_SHDW_WREN
|
1123 MII_TG3_MISC_SHDW_APD_SEL
|
1124 MII_TG3_MISC_SHDW_APD_WKTM_84MS
);
1128 if (tp
->tg3_flags2
& TG3_FLG2_PHY_ADC_BUG
) {
1129 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0c00);
1130 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x201f);
1131 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x2aaa);
1132 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x000a);
1133 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0323);
1134 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0400);
1136 if (tp
->tg3_flags2
& TG3_FLG2_PHY_5704_A0_BUG
) {
1137 tg3_writephy(tp
, 0x1c, 0x8d68);
1138 tg3_writephy(tp
, 0x1c, 0x8d68);
1140 if (tp
->tg3_flags2
& TG3_FLG2_PHY_BER_BUG
) {
1141 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0c00);
1142 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x000a);
1143 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x310b);
1144 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x201f);
1145 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x9506);
1146 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x401f);
1147 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x14e2);
1148 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0400);
1150 else if (tp
->tg3_flags2
& TG3_FLG2_PHY_JITTER_BUG
) {
1151 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0c00);
1152 tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x000a);
1153 if (tp
->tg3_flags2
& TG3_FLG2_PHY_ADJUST_TRIM
) {
1154 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x110b);
1155 tg3_writephy(tp
, MII_TG3_TEST1
,
1156 MII_TG3_TEST1_TRIM_EN
| 0x4);
1158 tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x010b);
1159 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0400);
1161 /* Set Extended packet length bit (bit 14) on all chips that */
1162 /* support jumbo frames */
1163 if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5401
) {
1164 /* Cannot do read-modify-write on 5401 */
1165 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x4c20);
1166 } else if (tp
->tg3_flags2
& TG3_FLG2_JUMBO_CAPABLE
) {
1169 /* Set bit 14 with read-modify-write to preserve other bits */
1170 if (!tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x0007) &&
1171 !tg3_readphy(tp
, MII_TG3_AUX_CTRL
, &phy_reg
))
1172 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, phy_reg
| 0x4000);
1175 /* Set phy register 0x10 bit 0 to high fifo elasticity to support
1176 * jumbo frames transmission.
1178 if (tp
->tg3_flags2
& TG3_FLG2_JUMBO_CAPABLE
) {
1181 if (!tg3_readphy(tp
, MII_TG3_EXT_CTRL
, &phy_reg
))
1182 tg3_writephy(tp
, MII_TG3_EXT_CTRL
,
1183 phy_reg
| MII_TG3_EXT_CTRL_FIFO_ELASTIC
);
1186 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
1187 /* adjust output voltage */
1188 tg3_writephy(tp
, MII_TG3_EPHY_PTEST
, 0x12);
1191 tg3_phy_toggle_automdix(tp
, 1);
1192 tg3_phy_set_wirespeed(tp
);
1196 static void tg3_frob_aux_power(struct tg3
*tp
)
1198 struct tg3
*tp_peer
= tp
;
1200 if ((tp
->tg3_flags2
& TG3_FLG2_IS_NIC
) == 0)
1203 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) ||
1204 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
)) {
1205 struct net_device
*dev_peer
;
1207 dev_peer
= pci_get_drvdata(tp
->pdev_peer
);
1208 /* remove_one() may have been run on the peer. */
1212 tp_peer
= netdev_priv(dev_peer
);
1215 if ((tp
->tg3_flags
& TG3_FLAG_WOL_ENABLE
) != 0 ||
1216 (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) != 0 ||
1217 (tp_peer
->tg3_flags
& TG3_FLAG_WOL_ENABLE
) != 0 ||
1218 (tp_peer
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) != 0) {
1219 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
1220 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) {
1221 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1222 (GRC_LCLCTRL_GPIO_OE0
|
1223 GRC_LCLCTRL_GPIO_OE1
|
1224 GRC_LCLCTRL_GPIO_OE2
|
1225 GRC_LCLCTRL_GPIO_OUTPUT0
|
1226 GRC_LCLCTRL_GPIO_OUTPUT1
),
1230 u32 grc_local_ctrl
= 0;
1232 if (tp_peer
!= tp
&&
1233 (tp_peer
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
) != 0)
1236 /* Workaround to prevent overdrawing Amps. */
1237 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
1239 grc_local_ctrl
|= GRC_LCLCTRL_GPIO_OE3
;
1240 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1241 grc_local_ctrl
, 100);
1244 /* On 5753 and variants, GPIO2 cannot be used. */
1245 no_gpio2
= tp
->nic_sram_data_cfg
&
1246 NIC_SRAM_DATA_CFG_NO_GPIO2
;
1248 grc_local_ctrl
|= GRC_LCLCTRL_GPIO_OE0
|
1249 GRC_LCLCTRL_GPIO_OE1
|
1250 GRC_LCLCTRL_GPIO_OE2
|
1251 GRC_LCLCTRL_GPIO_OUTPUT1
|
1252 GRC_LCLCTRL_GPIO_OUTPUT2
;
1254 grc_local_ctrl
&= ~(GRC_LCLCTRL_GPIO_OE2
|
1255 GRC_LCLCTRL_GPIO_OUTPUT2
);
1257 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1258 grc_local_ctrl
, 100);
1260 grc_local_ctrl
|= GRC_LCLCTRL_GPIO_OUTPUT0
;
1262 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1263 grc_local_ctrl
, 100);
1266 grc_local_ctrl
&= ~GRC_LCLCTRL_GPIO_OUTPUT2
;
1267 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1268 grc_local_ctrl
, 100);
1272 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
&&
1273 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5701
) {
1274 if (tp_peer
!= tp
&&
1275 (tp_peer
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
) != 0)
1278 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1279 (GRC_LCLCTRL_GPIO_OE1
|
1280 GRC_LCLCTRL_GPIO_OUTPUT1
), 100);
1282 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1283 GRC_LCLCTRL_GPIO_OE1
, 100);
1285 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
|
1286 (GRC_LCLCTRL_GPIO_OE1
|
1287 GRC_LCLCTRL_GPIO_OUTPUT1
), 100);
1292 static int tg3_5700_link_polarity(struct tg3
*tp
, u32 speed
)
1294 if (tp
->led_ctrl
== LED_CTRL_MODE_PHY_2
)
1296 else if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5411
) {
1297 if (speed
!= SPEED_10
)
1299 } else if (speed
== SPEED_10
)
1305 static int tg3_setup_phy(struct tg3
*, int);
1307 #define RESET_KIND_SHUTDOWN 0
1308 #define RESET_KIND_INIT 1
1309 #define RESET_KIND_SUSPEND 2
1311 static void tg3_write_sig_post_reset(struct tg3
*, int);
1312 static int tg3_halt_cpu(struct tg3
*, u32
);
1313 static int tg3_nvram_lock(struct tg3
*);
1314 static void tg3_nvram_unlock(struct tg3
*);
1316 static void tg3_power_down_phy(struct tg3
*tp
)
1320 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
1321 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) {
1322 u32 sg_dig_ctrl
= tr32(SG_DIG_CTRL
);
1323 u32 serdes_cfg
= tr32(MAC_SERDES_CFG
);
1326 SG_DIG_USING_HW_AUTONEG
| SG_DIG_SOFT_RESET
;
1327 tw32(SG_DIG_CTRL
, sg_dig_ctrl
);
1328 tw32(MAC_SERDES_CFG
, serdes_cfg
| (1 << 15));
1333 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
1335 val
= tr32(GRC_MISC_CFG
);
1336 tw32_f(GRC_MISC_CFG
, val
| GRC_MISC_CFG_EPHY_IDDQ
);
1340 tg3_writephy(tp
, MII_TG3_EXT_CTRL
,
1341 MII_TG3_EXT_CTRL_FORCE_LED_OFF
);
1342 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x01b2);
1345 /* The PHY should not be powered down on some chips because
1348 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
1349 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
||
1350 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5780
&&
1351 (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
)))
1354 if (tp
->tg3_flags3
& TG3_FLG3_5761_5784_AX_FIXES
) {
1355 val
= tr32(TG3_CPMU_LSPD_1000MB_CLK
);
1356 val
&= ~CPMU_LSPD_1000MB_MACCLK_MASK
;
1357 val
|= CPMU_LSPD_1000MB_MACCLK_12_5
;
1358 tw32_f(TG3_CPMU_LSPD_1000MB_CLK
, val
);
1361 tg3_writephy(tp
, MII_BMCR
, BMCR_PDOWN
);
1364 static int tg3_set_power_state(struct tg3
*tp
, pci_power_t state
)
1367 u16 power_control
, power_caps
;
1368 int pm
= tp
->pm_cap
;
1370 /* Make sure register accesses (indirect or otherwise)
1371 * will function correctly.
1373 pci_write_config_dword(tp
->pdev
,
1374 TG3PCI_MISC_HOST_CTRL
,
1375 tp
->misc_host_ctrl
);
1377 pci_read_config_word(tp
->pdev
,
1380 power_control
|= PCI_PM_CTRL_PME_STATUS
;
1381 power_control
&= ~(PCI_PM_CTRL_STATE_MASK
);
1385 pci_write_config_word(tp
->pdev
,
1388 udelay(100); /* Delay after power state change */
1390 /* Switch out of Vaux if it is a NIC */
1391 if (tp
->tg3_flags2
& TG3_FLG2_IS_NIC
)
1392 tw32_wait_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
, 100);
1409 printk(KERN_WARNING PFX
"%s: Invalid power state (%d) "
1411 tp
->dev
->name
, state
);
1415 power_control
|= PCI_PM_CTRL_PME_ENABLE
;
1417 misc_host_ctrl
= tr32(TG3PCI_MISC_HOST_CTRL
);
1418 tw32(TG3PCI_MISC_HOST_CTRL
,
1419 misc_host_ctrl
| MISC_HOST_CTRL_MASK_PCI_INT
);
1421 if (tp
->link_config
.phy_is_low_power
== 0) {
1422 tp
->link_config
.phy_is_low_power
= 1;
1423 tp
->link_config
.orig_speed
= tp
->link_config
.speed
;
1424 tp
->link_config
.orig_duplex
= tp
->link_config
.duplex
;
1425 tp
->link_config
.orig_autoneg
= tp
->link_config
.autoneg
;
1428 if (!(tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
)) {
1429 tp
->link_config
.speed
= SPEED_10
;
1430 tp
->link_config
.duplex
= DUPLEX_HALF
;
1431 tp
->link_config
.autoneg
= AUTONEG_ENABLE
;
1432 tg3_setup_phy(tp
, 0);
1435 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
1438 val
= tr32(GRC_VCPU_EXT_CTRL
);
1439 tw32(GRC_VCPU_EXT_CTRL
, val
| GRC_VCPU_EXT_CTRL_DISABLE_WOL
);
1440 } else if (!(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
)) {
1444 for (i
= 0; i
< 200; i
++) {
1445 tg3_read_mem(tp
, NIC_SRAM_FW_ASF_STATUS_MBOX
, &val
);
1446 if (val
== ~NIC_SRAM_FIRMWARE_MBOX_MAGIC1
)
1451 if (tp
->tg3_flags
& TG3_FLAG_WOL_CAP
)
1452 tg3_write_mem(tp
, NIC_SRAM_WOL_MBOX
, WOL_SIGNATURE
|
1453 WOL_DRV_STATE_SHUTDOWN
|
1457 pci_read_config_word(tp
->pdev
, pm
+ PCI_PM_PMC
, &power_caps
);
1459 if (tp
->tg3_flags
& TG3_FLAG_WOL_ENABLE
) {
1462 if (!(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)) {
1463 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x5a);
1466 if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
)
1467 mac_mode
= MAC_MODE_PORT_MODE_GMII
;
1469 mac_mode
= MAC_MODE_PORT_MODE_MII
;
1471 mac_mode
|= tp
->mac_mode
& MAC_MODE_LINK_POLARITY
;
1472 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
1474 u32 speed
= (tp
->tg3_flags
&
1475 TG3_FLAG_WOL_SPEED_100MB
) ?
1476 SPEED_100
: SPEED_10
;
1477 if (tg3_5700_link_polarity(tp
, speed
))
1478 mac_mode
|= MAC_MODE_LINK_POLARITY
;
1480 mac_mode
&= ~MAC_MODE_LINK_POLARITY
;
1483 mac_mode
= MAC_MODE_PORT_MODE_TBI
;
1486 if (!(tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
))
1487 tw32(MAC_LED_CTRL
, tp
->led_ctrl
);
1489 if (((power_caps
& PCI_PM_CAP_PME_D3cold
) &&
1490 (tp
->tg3_flags
& TG3_FLAG_WOL_ENABLE
)))
1491 mac_mode
|= MAC_MODE_MAGIC_PKT_ENABLE
;
1493 tw32_f(MAC_MODE
, mac_mode
);
1496 tw32_f(MAC_RX_MODE
, RX_MODE_ENABLE
);
1500 if (!(tp
->tg3_flags
& TG3_FLAG_WOL_SPEED_100MB
) &&
1501 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
1502 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
)) {
1505 base_val
= tp
->pci_clock_ctrl
;
1506 base_val
|= (CLOCK_CTRL_RXCLK_DISABLE
|
1507 CLOCK_CTRL_TXCLK_DISABLE
);
1509 tw32_wait_f(TG3PCI_CLOCK_CTRL
, base_val
| CLOCK_CTRL_ALTCLK
|
1510 CLOCK_CTRL_PWRDOWN_PLL133
, 40);
1511 } else if ((tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
) ||
1512 (tp
->tg3_flags
& TG3_FLAG_CPMU_PRESENT
) ||
1513 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)) {
1515 } else if (!((tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) &&
1516 (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
))) {
1517 u32 newbits1
, newbits2
;
1519 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
1520 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) {
1521 newbits1
= (CLOCK_CTRL_RXCLK_DISABLE
|
1522 CLOCK_CTRL_TXCLK_DISABLE
|
1524 newbits2
= newbits1
| CLOCK_CTRL_44MHZ_CORE
;
1525 } else if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
1526 newbits1
= CLOCK_CTRL_625_CORE
;
1527 newbits2
= newbits1
| CLOCK_CTRL_ALTCLK
;
1529 newbits1
= CLOCK_CTRL_ALTCLK
;
1530 newbits2
= newbits1
| CLOCK_CTRL_44MHZ_CORE
;
1533 tw32_wait_f(TG3PCI_CLOCK_CTRL
, tp
->pci_clock_ctrl
| newbits1
,
1536 tw32_wait_f(TG3PCI_CLOCK_CTRL
, tp
->pci_clock_ctrl
| newbits2
,
1539 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
1542 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
1543 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) {
1544 newbits3
= (CLOCK_CTRL_RXCLK_DISABLE
|
1545 CLOCK_CTRL_TXCLK_DISABLE
|
1546 CLOCK_CTRL_44MHZ_CORE
);
1548 newbits3
= CLOCK_CTRL_44MHZ_CORE
;
1551 tw32_wait_f(TG3PCI_CLOCK_CTRL
,
1552 tp
->pci_clock_ctrl
| newbits3
, 40);
1556 if (!(tp
->tg3_flags
& TG3_FLAG_WOL_ENABLE
) &&
1557 !(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) &&
1558 !(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
))
1559 tg3_power_down_phy(tp
);
1561 tg3_frob_aux_power(tp
);
1563 /* Workaround for unstable PLL clock */
1564 if ((GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5750_AX
) ||
1565 (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5750_BX
)) {
1566 u32 val
= tr32(0x7d00);
1568 val
&= ~((1 << 16) | (1 << 4) | (1 << 2) | (1 << 1) | 1);
1570 if (!(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
)) {
1573 err
= tg3_nvram_lock(tp
);
1574 tg3_halt_cpu(tp
, RX_CPU_BASE
);
1576 tg3_nvram_unlock(tp
);
1580 tg3_write_sig_post_reset(tp
, RESET_KIND_SHUTDOWN
);
1582 /* Finally, set the new power state. */
1583 pci_write_config_word(tp
->pdev
, pm
+ PCI_PM_CTRL
, power_control
);
1584 udelay(100); /* Delay after power state change */
1589 static void tg3_link_report(struct tg3
*tp
)
1591 if (!netif_carrier_ok(tp
->dev
)) {
1592 if (netif_msg_link(tp
))
1593 printk(KERN_INFO PFX
"%s: Link is down.\n",
1595 } else if (netif_msg_link(tp
)) {
1596 printk(KERN_INFO PFX
"%s: Link is up at %d Mbps, %s duplex.\n",
1598 (tp
->link_config
.active_speed
== SPEED_1000
?
1600 (tp
->link_config
.active_speed
== SPEED_100
?
1602 (tp
->link_config
.active_duplex
== DUPLEX_FULL
?
1605 printk(KERN_INFO PFX
"%s: Flow control is %s for TX and "
1608 (tp
->tg3_flags
& TG3_FLAG_TX_PAUSE
) ? "on" : "off",
1609 (tp
->tg3_flags
& TG3_FLAG_RX_PAUSE
) ? "on" : "off");
1613 static void tg3_setup_flow_control(struct tg3
*tp
, u32 local_adv
, u32 remote_adv
)
1615 u32 new_tg3_flags
= 0;
1616 u32 old_rx_mode
= tp
->rx_mode
;
1617 u32 old_tx_mode
= tp
->tx_mode
;
1619 if (tp
->tg3_flags
& TG3_FLAG_PAUSE_AUTONEG
) {
1621 /* Convert 1000BaseX flow control bits to 1000BaseT
1622 * bits before resolving flow control.
1624 if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) {
1625 local_adv
&= ~(ADVERTISE_PAUSE_CAP
|
1626 ADVERTISE_PAUSE_ASYM
);
1627 remote_adv
&= ~(LPA_PAUSE_CAP
| LPA_PAUSE_ASYM
);
1629 if (local_adv
& ADVERTISE_1000XPAUSE
)
1630 local_adv
|= ADVERTISE_PAUSE_CAP
;
1631 if (local_adv
& ADVERTISE_1000XPSE_ASYM
)
1632 local_adv
|= ADVERTISE_PAUSE_ASYM
;
1633 if (remote_adv
& LPA_1000XPAUSE
)
1634 remote_adv
|= LPA_PAUSE_CAP
;
1635 if (remote_adv
& LPA_1000XPAUSE_ASYM
)
1636 remote_adv
|= LPA_PAUSE_ASYM
;
1639 if (local_adv
& ADVERTISE_PAUSE_CAP
) {
1640 if (local_adv
& ADVERTISE_PAUSE_ASYM
) {
1641 if (remote_adv
& LPA_PAUSE_CAP
)
1643 (TG3_FLAG_RX_PAUSE
|
1645 else if (remote_adv
& LPA_PAUSE_ASYM
)
1647 (TG3_FLAG_RX_PAUSE
);
1649 if (remote_adv
& LPA_PAUSE_CAP
)
1651 (TG3_FLAG_RX_PAUSE
|
1654 } else if (local_adv
& ADVERTISE_PAUSE_ASYM
) {
1655 if ((remote_adv
& LPA_PAUSE_CAP
) &&
1656 (remote_adv
& LPA_PAUSE_ASYM
))
1657 new_tg3_flags
|= TG3_FLAG_TX_PAUSE
;
1660 tp
->tg3_flags
&= ~(TG3_FLAG_RX_PAUSE
| TG3_FLAG_TX_PAUSE
);
1661 tp
->tg3_flags
|= new_tg3_flags
;
1663 new_tg3_flags
= tp
->tg3_flags
;
1666 if (new_tg3_flags
& TG3_FLAG_RX_PAUSE
)
1667 tp
->rx_mode
|= RX_MODE_FLOW_CTRL_ENABLE
;
1669 tp
->rx_mode
&= ~RX_MODE_FLOW_CTRL_ENABLE
;
1671 if (old_rx_mode
!= tp
->rx_mode
) {
1672 tw32_f(MAC_RX_MODE
, tp
->rx_mode
);
1675 if (new_tg3_flags
& TG3_FLAG_TX_PAUSE
)
1676 tp
->tx_mode
|= TX_MODE_FLOW_CTRL_ENABLE
;
1678 tp
->tx_mode
&= ~TX_MODE_FLOW_CTRL_ENABLE
;
1680 if (old_tx_mode
!= tp
->tx_mode
) {
1681 tw32_f(MAC_TX_MODE
, tp
->tx_mode
);
1685 static void tg3_aux_stat_to_speed_duplex(struct tg3
*tp
, u32 val
, u16
*speed
, u8
*duplex
)
1687 switch (val
& MII_TG3_AUX_STAT_SPDMASK
) {
1688 case MII_TG3_AUX_STAT_10HALF
:
1690 *duplex
= DUPLEX_HALF
;
1693 case MII_TG3_AUX_STAT_10FULL
:
1695 *duplex
= DUPLEX_FULL
;
1698 case MII_TG3_AUX_STAT_100HALF
:
1700 *duplex
= DUPLEX_HALF
;
1703 case MII_TG3_AUX_STAT_100FULL
:
1705 *duplex
= DUPLEX_FULL
;
1708 case MII_TG3_AUX_STAT_1000HALF
:
1709 *speed
= SPEED_1000
;
1710 *duplex
= DUPLEX_HALF
;
1713 case MII_TG3_AUX_STAT_1000FULL
:
1714 *speed
= SPEED_1000
;
1715 *duplex
= DUPLEX_FULL
;
1719 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
1720 *speed
= (val
& MII_TG3_AUX_STAT_100
) ? SPEED_100
:
1722 *duplex
= (val
& MII_TG3_AUX_STAT_FULL
) ? DUPLEX_FULL
:
1726 *speed
= SPEED_INVALID
;
1727 *duplex
= DUPLEX_INVALID
;
1732 static void tg3_phy_copper_begin(struct tg3
*tp
)
1737 if (tp
->link_config
.phy_is_low_power
) {
1738 /* Entering low power mode. Disable gigabit and
1739 * 100baseT advertisements.
1741 tg3_writephy(tp
, MII_TG3_CTRL
, 0);
1743 new_adv
= (ADVERTISE_10HALF
| ADVERTISE_10FULL
|
1744 ADVERTISE_CSMA
| ADVERTISE_PAUSE_CAP
);
1745 if (tp
->tg3_flags
& TG3_FLAG_WOL_SPEED_100MB
)
1746 new_adv
|= (ADVERTISE_100HALF
| ADVERTISE_100FULL
);
1748 tg3_writephy(tp
, MII_ADVERTISE
, new_adv
);
1749 } else if (tp
->link_config
.speed
== SPEED_INVALID
) {
1750 if (tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
)
1751 tp
->link_config
.advertising
&=
1752 ~(ADVERTISED_1000baseT_Half
|
1753 ADVERTISED_1000baseT_Full
);
1755 new_adv
= (ADVERTISE_CSMA
| ADVERTISE_PAUSE_CAP
);
1756 if (tp
->link_config
.advertising
& ADVERTISED_10baseT_Half
)
1757 new_adv
|= ADVERTISE_10HALF
;
1758 if (tp
->link_config
.advertising
& ADVERTISED_10baseT_Full
)
1759 new_adv
|= ADVERTISE_10FULL
;
1760 if (tp
->link_config
.advertising
& ADVERTISED_100baseT_Half
)
1761 new_adv
|= ADVERTISE_100HALF
;
1762 if (tp
->link_config
.advertising
& ADVERTISED_100baseT_Full
)
1763 new_adv
|= ADVERTISE_100FULL
;
1764 tg3_writephy(tp
, MII_ADVERTISE
, new_adv
);
1766 if (tp
->link_config
.advertising
&
1767 (ADVERTISED_1000baseT_Half
| ADVERTISED_1000baseT_Full
)) {
1769 if (tp
->link_config
.advertising
& ADVERTISED_1000baseT_Half
)
1770 new_adv
|= MII_TG3_CTRL_ADV_1000_HALF
;
1771 if (tp
->link_config
.advertising
& ADVERTISED_1000baseT_Full
)
1772 new_adv
|= MII_TG3_CTRL_ADV_1000_FULL
;
1773 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
) &&
1774 (tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
||
1775 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B0
))
1776 new_adv
|= (MII_TG3_CTRL_AS_MASTER
|
1777 MII_TG3_CTRL_ENABLE_AS_MASTER
);
1778 tg3_writephy(tp
, MII_TG3_CTRL
, new_adv
);
1780 tg3_writephy(tp
, MII_TG3_CTRL
, 0);
1783 /* Asking for a specific link mode. */
1784 if (tp
->link_config
.speed
== SPEED_1000
) {
1785 new_adv
= ADVERTISE_CSMA
| ADVERTISE_PAUSE_CAP
;
1786 tg3_writephy(tp
, MII_ADVERTISE
, new_adv
);
1788 if (tp
->link_config
.duplex
== DUPLEX_FULL
)
1789 new_adv
= MII_TG3_CTRL_ADV_1000_FULL
;
1791 new_adv
= MII_TG3_CTRL_ADV_1000_HALF
;
1792 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
||
1793 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B0
)
1794 new_adv
|= (MII_TG3_CTRL_AS_MASTER
|
1795 MII_TG3_CTRL_ENABLE_AS_MASTER
);
1796 tg3_writephy(tp
, MII_TG3_CTRL
, new_adv
);
1798 tg3_writephy(tp
, MII_TG3_CTRL
, 0);
1800 new_adv
= ADVERTISE_CSMA
| ADVERTISE_PAUSE_CAP
;
1801 if (tp
->link_config
.speed
== SPEED_100
) {
1802 if (tp
->link_config
.duplex
== DUPLEX_FULL
)
1803 new_adv
|= ADVERTISE_100FULL
;
1805 new_adv
|= ADVERTISE_100HALF
;
1807 if (tp
->link_config
.duplex
== DUPLEX_FULL
)
1808 new_adv
|= ADVERTISE_10FULL
;
1810 new_adv
|= ADVERTISE_10HALF
;
1812 tg3_writephy(tp
, MII_ADVERTISE
, new_adv
);
1816 if (tp
->link_config
.autoneg
== AUTONEG_DISABLE
&&
1817 tp
->link_config
.speed
!= SPEED_INVALID
) {
1818 u32 bmcr
, orig_bmcr
;
1820 tp
->link_config
.active_speed
= tp
->link_config
.speed
;
1821 tp
->link_config
.active_duplex
= tp
->link_config
.duplex
;
1824 switch (tp
->link_config
.speed
) {
1830 bmcr
|= BMCR_SPEED100
;
1834 bmcr
|= TG3_BMCR_SPEED1000
;
1838 if (tp
->link_config
.duplex
== DUPLEX_FULL
)
1839 bmcr
|= BMCR_FULLDPLX
;
1841 if (!tg3_readphy(tp
, MII_BMCR
, &orig_bmcr
) &&
1842 (bmcr
!= orig_bmcr
)) {
1843 tg3_writephy(tp
, MII_BMCR
, BMCR_LOOPBACK
);
1844 for (i
= 0; i
< 1500; i
++) {
1848 if (tg3_readphy(tp
, MII_BMSR
, &tmp
) ||
1849 tg3_readphy(tp
, MII_BMSR
, &tmp
))
1851 if (!(tmp
& BMSR_LSTATUS
)) {
1856 tg3_writephy(tp
, MII_BMCR
, bmcr
);
1860 tg3_writephy(tp
, MII_BMCR
,
1861 BMCR_ANENABLE
| BMCR_ANRESTART
);
1865 static int tg3_init_5401phy_dsp(struct tg3
*tp
)
1869 /* Turn off tap power management. */
1870 /* Set Extended packet length bit */
1871 err
= tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x4c20);
1873 err
|= tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x0012);
1874 err
|= tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x1804);
1876 err
|= tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x0013);
1877 err
|= tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x1204);
1879 err
|= tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x8006);
1880 err
|= tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0132);
1882 err
|= tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x8006);
1883 err
|= tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0232);
1885 err
|= tg3_writephy(tp
, MII_TG3_DSP_ADDRESS
, 0x201f);
1886 err
|= tg3_writephy(tp
, MII_TG3_DSP_RW_PORT
, 0x0a20);
1893 static int tg3_copper_is_advertising_all(struct tg3
*tp
, u32 mask
)
1895 u32 adv_reg
, all_mask
= 0;
1897 if (mask
& ADVERTISED_10baseT_Half
)
1898 all_mask
|= ADVERTISE_10HALF
;
1899 if (mask
& ADVERTISED_10baseT_Full
)
1900 all_mask
|= ADVERTISE_10FULL
;
1901 if (mask
& ADVERTISED_100baseT_Half
)
1902 all_mask
|= ADVERTISE_100HALF
;
1903 if (mask
& ADVERTISED_100baseT_Full
)
1904 all_mask
|= ADVERTISE_100FULL
;
1906 if (tg3_readphy(tp
, MII_ADVERTISE
, &adv_reg
))
1909 if ((adv_reg
& all_mask
) != all_mask
)
1911 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
)) {
1915 if (mask
& ADVERTISED_1000baseT_Half
)
1916 all_mask
|= ADVERTISE_1000HALF
;
1917 if (mask
& ADVERTISED_1000baseT_Full
)
1918 all_mask
|= ADVERTISE_1000FULL
;
1920 if (tg3_readphy(tp
, MII_TG3_CTRL
, &tg3_ctrl
))
1923 if ((tg3_ctrl
& all_mask
) != all_mask
)
1929 static int tg3_setup_copper_phy(struct tg3
*tp
, int force_reset
)
1931 int current_link_up
;
1940 (MAC_STATUS_SYNC_CHANGED
|
1941 MAC_STATUS_CFG_CHANGED
|
1942 MAC_STATUS_MI_COMPLETION
|
1943 MAC_STATUS_LNKSTATE_CHANGED
));
1946 tp
->mi_mode
= MAC_MI_MODE_BASE
;
1947 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
1950 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x02);
1952 /* Some third-party PHYs need to be reset on link going
1955 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
1956 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
||
1957 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) &&
1958 netif_carrier_ok(tp
->dev
)) {
1959 tg3_readphy(tp
, MII_BMSR
, &bmsr
);
1960 if (!tg3_readphy(tp
, MII_BMSR
, &bmsr
) &&
1961 !(bmsr
& BMSR_LSTATUS
))
1967 if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5401
) {
1968 tg3_readphy(tp
, MII_BMSR
, &bmsr
);
1969 if (tg3_readphy(tp
, MII_BMSR
, &bmsr
) ||
1970 !(tp
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
))
1973 if (!(bmsr
& BMSR_LSTATUS
)) {
1974 err
= tg3_init_5401phy_dsp(tp
);
1978 tg3_readphy(tp
, MII_BMSR
, &bmsr
);
1979 for (i
= 0; i
< 1000; i
++) {
1981 if (!tg3_readphy(tp
, MII_BMSR
, &bmsr
) &&
1982 (bmsr
& BMSR_LSTATUS
)) {
1988 if ((tp
->phy_id
& PHY_ID_REV_MASK
) == PHY_REV_BCM5401_B0
&&
1989 !(bmsr
& BMSR_LSTATUS
) &&
1990 tp
->link_config
.active_speed
== SPEED_1000
) {
1991 err
= tg3_phy_reset(tp
);
1993 err
= tg3_init_5401phy_dsp(tp
);
1998 } else if (tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
||
1999 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B0
) {
2000 /* 5701 {A0,B0} CRC bug workaround */
2001 tg3_writephy(tp
, 0x15, 0x0a75);
2002 tg3_writephy(tp
, 0x1c, 0x8c68);
2003 tg3_writephy(tp
, 0x1c, 0x8d68);
2004 tg3_writephy(tp
, 0x1c, 0x8c68);
2007 /* Clear pending interrupts... */
2008 tg3_readphy(tp
, MII_TG3_ISTAT
, &dummy
);
2009 tg3_readphy(tp
, MII_TG3_ISTAT
, &dummy
);
2011 if (tp
->tg3_flags
& TG3_FLAG_USE_MI_INTERRUPT
)
2012 tg3_writephy(tp
, MII_TG3_IMASK
, ~MII_TG3_INT_LINKCHG
);
2013 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
)
2014 tg3_writephy(tp
, MII_TG3_IMASK
, ~0);
2016 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
2017 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) {
2018 if (tp
->led_ctrl
== LED_CTRL_MODE_PHY_1
)
2019 tg3_writephy(tp
, MII_TG3_EXT_CTRL
,
2020 MII_TG3_EXT_CTRL_LNK3_LED_MODE
);
2022 tg3_writephy(tp
, MII_TG3_EXT_CTRL
, 0);
2025 current_link_up
= 0;
2026 current_speed
= SPEED_INVALID
;
2027 current_duplex
= DUPLEX_INVALID
;
2029 if (tp
->tg3_flags2
& TG3_FLG2_CAPACITIVE_COUPLING
) {
2032 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, 0x4007);
2033 tg3_readphy(tp
, MII_TG3_AUX_CTRL
, &val
);
2034 if (!(val
& (1 << 10))) {
2036 tg3_writephy(tp
, MII_TG3_AUX_CTRL
, val
);
2042 for (i
= 0; i
< 100; i
++) {
2043 tg3_readphy(tp
, MII_BMSR
, &bmsr
);
2044 if (!tg3_readphy(tp
, MII_BMSR
, &bmsr
) &&
2045 (bmsr
& BMSR_LSTATUS
))
2050 if (bmsr
& BMSR_LSTATUS
) {
2053 tg3_readphy(tp
, MII_TG3_AUX_STAT
, &aux_stat
);
2054 for (i
= 0; i
< 2000; i
++) {
2056 if (!tg3_readphy(tp
, MII_TG3_AUX_STAT
, &aux_stat
) &&
2061 tg3_aux_stat_to_speed_duplex(tp
, aux_stat
,
2066 for (i
= 0; i
< 200; i
++) {
2067 tg3_readphy(tp
, MII_BMCR
, &bmcr
);
2068 if (tg3_readphy(tp
, MII_BMCR
, &bmcr
))
2070 if (bmcr
&& bmcr
!= 0x7fff)
2075 if (tp
->link_config
.autoneg
== AUTONEG_ENABLE
) {
2076 if (bmcr
& BMCR_ANENABLE
) {
2077 current_link_up
= 1;
2079 /* Force autoneg restart if we are exiting
2082 if (!tg3_copper_is_advertising_all(tp
,
2083 tp
->link_config
.advertising
))
2084 current_link_up
= 0;
2086 current_link_up
= 0;
2089 if (!(bmcr
& BMCR_ANENABLE
) &&
2090 tp
->link_config
.speed
== current_speed
&&
2091 tp
->link_config
.duplex
== current_duplex
) {
2092 current_link_up
= 1;
2094 current_link_up
= 0;
2098 tp
->link_config
.active_speed
= current_speed
;
2099 tp
->link_config
.active_duplex
= current_duplex
;
2102 if (current_link_up
== 1 &&
2103 (tp
->link_config
.active_duplex
== DUPLEX_FULL
) &&
2104 (tp
->link_config
.autoneg
== AUTONEG_ENABLE
)) {
2105 u32 local_adv
, remote_adv
;
2107 if (tg3_readphy(tp
, MII_ADVERTISE
, &local_adv
))
2109 local_adv
&= (ADVERTISE_PAUSE_CAP
| ADVERTISE_PAUSE_ASYM
);
2111 if (tg3_readphy(tp
, MII_LPA
, &remote_adv
))
2114 remote_adv
&= (LPA_PAUSE_CAP
| LPA_PAUSE_ASYM
);
2116 /* If we are not advertising full pause capability,
2117 * something is wrong. Bring the link down and reconfigure.
2119 if (local_adv
!= ADVERTISE_PAUSE_CAP
) {
2120 current_link_up
= 0;
2122 tg3_setup_flow_control(tp
, local_adv
, remote_adv
);
2126 if (current_link_up
== 0 || tp
->link_config
.phy_is_low_power
) {
2129 tg3_phy_copper_begin(tp
);
2131 tg3_readphy(tp
, MII_BMSR
, &tmp
);
2132 if (!tg3_readphy(tp
, MII_BMSR
, &tmp
) &&
2133 (tmp
& BMSR_LSTATUS
))
2134 current_link_up
= 1;
2137 tp
->mac_mode
&= ~MAC_MODE_PORT_MODE_MASK
;
2138 if (current_link_up
== 1) {
2139 if (tp
->link_config
.active_speed
== SPEED_100
||
2140 tp
->link_config
.active_speed
== SPEED_10
)
2141 tp
->mac_mode
|= MAC_MODE_PORT_MODE_MII
;
2143 tp
->mac_mode
|= MAC_MODE_PORT_MODE_GMII
;
2145 tp
->mac_mode
|= MAC_MODE_PORT_MODE_GMII
;
2147 tp
->mac_mode
&= ~MAC_MODE_HALF_DUPLEX
;
2148 if (tp
->link_config
.active_duplex
== DUPLEX_HALF
)
2149 tp
->mac_mode
|= MAC_MODE_HALF_DUPLEX
;
2151 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
) {
2152 if (current_link_up
== 1 &&
2153 tg3_5700_link_polarity(tp
, tp
->link_config
.active_speed
))
2154 tp
->mac_mode
|= MAC_MODE_LINK_POLARITY
;
2156 tp
->mac_mode
&= ~MAC_MODE_LINK_POLARITY
;
2159 /* ??? Without this setting Netgear GA302T PHY does not
2160 * ??? send/receive packets...
2162 if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5411
&&
2163 tp
->pci_chip_rev_id
== CHIPREV_ID_5700_ALTIMA
) {
2164 tp
->mi_mode
|= MAC_MI_MODE_AUTO_POLL
;
2165 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
2169 tw32_f(MAC_MODE
, tp
->mac_mode
);
2172 if (tp
->tg3_flags
& TG3_FLAG_USE_LINKCHG_REG
) {
2173 /* Polled via timer. */
2174 tw32_f(MAC_EVENT
, 0);
2176 tw32_f(MAC_EVENT
, MAC_EVENT_LNKSTATE_CHANGED
);
2180 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
&&
2181 current_link_up
== 1 &&
2182 tp
->link_config
.active_speed
== SPEED_1000
&&
2183 ((tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) ||
2184 (tp
->tg3_flags
& TG3_FLAG_PCI_HIGH_SPEED
))) {
2187 (MAC_STATUS_SYNC_CHANGED
|
2188 MAC_STATUS_CFG_CHANGED
));
2191 NIC_SRAM_FIRMWARE_MBOX
,
2192 NIC_SRAM_FIRMWARE_MBOX_MAGIC2
);
2195 if (current_link_up
!= netif_carrier_ok(tp
->dev
)) {
2196 if (current_link_up
)
2197 netif_carrier_on(tp
->dev
);
2199 netif_carrier_off(tp
->dev
);
2200 tg3_link_report(tp
);
2206 struct tg3_fiber_aneginfo
{
2208 #define ANEG_STATE_UNKNOWN 0
2209 #define ANEG_STATE_AN_ENABLE 1
2210 #define ANEG_STATE_RESTART_INIT 2
2211 #define ANEG_STATE_RESTART 3
2212 #define ANEG_STATE_DISABLE_LINK_OK 4
2213 #define ANEG_STATE_ABILITY_DETECT_INIT 5
2214 #define ANEG_STATE_ABILITY_DETECT 6
2215 #define ANEG_STATE_ACK_DETECT_INIT 7
2216 #define ANEG_STATE_ACK_DETECT 8
2217 #define ANEG_STATE_COMPLETE_ACK_INIT 9
2218 #define ANEG_STATE_COMPLETE_ACK 10
2219 #define ANEG_STATE_IDLE_DETECT_INIT 11
2220 #define ANEG_STATE_IDLE_DETECT 12
2221 #define ANEG_STATE_LINK_OK 13
2222 #define ANEG_STATE_NEXT_PAGE_WAIT_INIT 14
2223 #define ANEG_STATE_NEXT_PAGE_WAIT 15
2226 #define MR_AN_ENABLE 0x00000001
2227 #define MR_RESTART_AN 0x00000002
2228 #define MR_AN_COMPLETE 0x00000004
2229 #define MR_PAGE_RX 0x00000008
2230 #define MR_NP_LOADED 0x00000010
2231 #define MR_TOGGLE_TX 0x00000020
2232 #define MR_LP_ADV_FULL_DUPLEX 0x00000040
2233 #define MR_LP_ADV_HALF_DUPLEX 0x00000080
2234 #define MR_LP_ADV_SYM_PAUSE 0x00000100
2235 #define MR_LP_ADV_ASYM_PAUSE 0x00000200
2236 #define MR_LP_ADV_REMOTE_FAULT1 0x00000400
2237 #define MR_LP_ADV_REMOTE_FAULT2 0x00000800
2238 #define MR_LP_ADV_NEXT_PAGE 0x00001000
2239 #define MR_TOGGLE_RX 0x00002000
2240 #define MR_NP_RX 0x00004000
2242 #define MR_LINK_OK 0x80000000
2244 unsigned long link_time
, cur_time
;
2246 u32 ability_match_cfg
;
2247 int ability_match_count
;
2249 char ability_match
, idle_match
, ack_match
;
2251 u32 txconfig
, rxconfig
;
2252 #define ANEG_CFG_NP 0x00000080
2253 #define ANEG_CFG_ACK 0x00000040
2254 #define ANEG_CFG_RF2 0x00000020
2255 #define ANEG_CFG_RF1 0x00000010
2256 #define ANEG_CFG_PS2 0x00000001
2257 #define ANEG_CFG_PS1 0x00008000
2258 #define ANEG_CFG_HD 0x00004000
2259 #define ANEG_CFG_FD 0x00002000
2260 #define ANEG_CFG_INVAL 0x00001f06
2265 #define ANEG_TIMER_ENAB 2
2266 #define ANEG_FAILED -1
2268 #define ANEG_STATE_SETTLE_TIME 10000
2270 static int tg3_fiber_aneg_smachine(struct tg3
*tp
,
2271 struct tg3_fiber_aneginfo
*ap
)
2273 unsigned long delta
;
2277 if (ap
->state
== ANEG_STATE_UNKNOWN
) {
2281 ap
->ability_match_cfg
= 0;
2282 ap
->ability_match_count
= 0;
2283 ap
->ability_match
= 0;
2289 if (tr32(MAC_STATUS
) & MAC_STATUS_RCVD_CFG
) {
2290 rx_cfg_reg
= tr32(MAC_RX_AUTO_NEG
);
2292 if (rx_cfg_reg
!= ap
->ability_match_cfg
) {
2293 ap
->ability_match_cfg
= rx_cfg_reg
;
2294 ap
->ability_match
= 0;
2295 ap
->ability_match_count
= 0;
2297 if (++ap
->ability_match_count
> 1) {
2298 ap
->ability_match
= 1;
2299 ap
->ability_match_cfg
= rx_cfg_reg
;
2302 if (rx_cfg_reg
& ANEG_CFG_ACK
)
2310 ap
->ability_match_cfg
= 0;
2311 ap
->ability_match_count
= 0;
2312 ap
->ability_match
= 0;
2318 ap
->rxconfig
= rx_cfg_reg
;
2322 case ANEG_STATE_UNKNOWN
:
2323 if (ap
->flags
& (MR_AN_ENABLE
| MR_RESTART_AN
))
2324 ap
->state
= ANEG_STATE_AN_ENABLE
;
2327 case ANEG_STATE_AN_ENABLE
:
2328 ap
->flags
&= ~(MR_AN_COMPLETE
| MR_PAGE_RX
);
2329 if (ap
->flags
& MR_AN_ENABLE
) {
2332 ap
->ability_match_cfg
= 0;
2333 ap
->ability_match_count
= 0;
2334 ap
->ability_match
= 0;
2338 ap
->state
= ANEG_STATE_RESTART_INIT
;
2340 ap
->state
= ANEG_STATE_DISABLE_LINK_OK
;
2344 case ANEG_STATE_RESTART_INIT
:
2345 ap
->link_time
= ap
->cur_time
;
2346 ap
->flags
&= ~(MR_NP_LOADED
);
2348 tw32(MAC_TX_AUTO_NEG
, 0);
2349 tp
->mac_mode
|= MAC_MODE_SEND_CONFIGS
;
2350 tw32_f(MAC_MODE
, tp
->mac_mode
);
2353 ret
= ANEG_TIMER_ENAB
;
2354 ap
->state
= ANEG_STATE_RESTART
;
2357 case ANEG_STATE_RESTART
:
2358 delta
= ap
->cur_time
- ap
->link_time
;
2359 if (delta
> ANEG_STATE_SETTLE_TIME
) {
2360 ap
->state
= ANEG_STATE_ABILITY_DETECT_INIT
;
2362 ret
= ANEG_TIMER_ENAB
;
2366 case ANEG_STATE_DISABLE_LINK_OK
:
2370 case ANEG_STATE_ABILITY_DETECT_INIT
:
2371 ap
->flags
&= ~(MR_TOGGLE_TX
);
2372 ap
->txconfig
= (ANEG_CFG_FD
| ANEG_CFG_PS1
);
2373 tw32(MAC_TX_AUTO_NEG
, ap
->txconfig
);
2374 tp
->mac_mode
|= MAC_MODE_SEND_CONFIGS
;
2375 tw32_f(MAC_MODE
, tp
->mac_mode
);
2378 ap
->state
= ANEG_STATE_ABILITY_DETECT
;
2381 case ANEG_STATE_ABILITY_DETECT
:
2382 if (ap
->ability_match
!= 0 && ap
->rxconfig
!= 0) {
2383 ap
->state
= ANEG_STATE_ACK_DETECT_INIT
;
2387 case ANEG_STATE_ACK_DETECT_INIT
:
2388 ap
->txconfig
|= ANEG_CFG_ACK
;
2389 tw32(MAC_TX_AUTO_NEG
, ap
->txconfig
);
2390 tp
->mac_mode
|= MAC_MODE_SEND_CONFIGS
;
2391 tw32_f(MAC_MODE
, tp
->mac_mode
);
2394 ap
->state
= ANEG_STATE_ACK_DETECT
;
2397 case ANEG_STATE_ACK_DETECT
:
2398 if (ap
->ack_match
!= 0) {
2399 if ((ap
->rxconfig
& ~ANEG_CFG_ACK
) ==
2400 (ap
->ability_match_cfg
& ~ANEG_CFG_ACK
)) {
2401 ap
->state
= ANEG_STATE_COMPLETE_ACK_INIT
;
2403 ap
->state
= ANEG_STATE_AN_ENABLE
;
2405 } else if (ap
->ability_match
!= 0 &&
2406 ap
->rxconfig
== 0) {
2407 ap
->state
= ANEG_STATE_AN_ENABLE
;
2411 case ANEG_STATE_COMPLETE_ACK_INIT
:
2412 if (ap
->rxconfig
& ANEG_CFG_INVAL
) {
2416 ap
->flags
&= ~(MR_LP_ADV_FULL_DUPLEX
|
2417 MR_LP_ADV_HALF_DUPLEX
|
2418 MR_LP_ADV_SYM_PAUSE
|
2419 MR_LP_ADV_ASYM_PAUSE
|
2420 MR_LP_ADV_REMOTE_FAULT1
|
2421 MR_LP_ADV_REMOTE_FAULT2
|
2422 MR_LP_ADV_NEXT_PAGE
|
2425 if (ap
->rxconfig
& ANEG_CFG_FD
)
2426 ap
->flags
|= MR_LP_ADV_FULL_DUPLEX
;
2427 if (ap
->rxconfig
& ANEG_CFG_HD
)
2428 ap
->flags
|= MR_LP_ADV_HALF_DUPLEX
;
2429 if (ap
->rxconfig
& ANEG_CFG_PS1
)
2430 ap
->flags
|= MR_LP_ADV_SYM_PAUSE
;
2431 if (ap
->rxconfig
& ANEG_CFG_PS2
)
2432 ap
->flags
|= MR_LP_ADV_ASYM_PAUSE
;
2433 if (ap
->rxconfig
& ANEG_CFG_RF1
)
2434 ap
->flags
|= MR_LP_ADV_REMOTE_FAULT1
;
2435 if (ap
->rxconfig
& ANEG_CFG_RF2
)
2436 ap
->flags
|= MR_LP_ADV_REMOTE_FAULT2
;
2437 if (ap
->rxconfig
& ANEG_CFG_NP
)
2438 ap
->flags
|= MR_LP_ADV_NEXT_PAGE
;
2440 ap
->link_time
= ap
->cur_time
;
2442 ap
->flags
^= (MR_TOGGLE_TX
);
2443 if (ap
->rxconfig
& 0x0008)
2444 ap
->flags
|= MR_TOGGLE_RX
;
2445 if (ap
->rxconfig
& ANEG_CFG_NP
)
2446 ap
->flags
|= MR_NP_RX
;
2447 ap
->flags
|= MR_PAGE_RX
;
2449 ap
->state
= ANEG_STATE_COMPLETE_ACK
;
2450 ret
= ANEG_TIMER_ENAB
;
2453 case ANEG_STATE_COMPLETE_ACK
:
2454 if (ap
->ability_match
!= 0 &&
2455 ap
->rxconfig
== 0) {
2456 ap
->state
= ANEG_STATE_AN_ENABLE
;
2459 delta
= ap
->cur_time
- ap
->link_time
;
2460 if (delta
> ANEG_STATE_SETTLE_TIME
) {
2461 if (!(ap
->flags
& (MR_LP_ADV_NEXT_PAGE
))) {
2462 ap
->state
= ANEG_STATE_IDLE_DETECT_INIT
;
2464 if ((ap
->txconfig
& ANEG_CFG_NP
) == 0 &&
2465 !(ap
->flags
& MR_NP_RX
)) {
2466 ap
->state
= ANEG_STATE_IDLE_DETECT_INIT
;
2474 case ANEG_STATE_IDLE_DETECT_INIT
:
2475 ap
->link_time
= ap
->cur_time
;
2476 tp
->mac_mode
&= ~MAC_MODE_SEND_CONFIGS
;
2477 tw32_f(MAC_MODE
, tp
->mac_mode
);
2480 ap
->state
= ANEG_STATE_IDLE_DETECT
;
2481 ret
= ANEG_TIMER_ENAB
;
2484 case ANEG_STATE_IDLE_DETECT
:
2485 if (ap
->ability_match
!= 0 &&
2486 ap
->rxconfig
== 0) {
2487 ap
->state
= ANEG_STATE_AN_ENABLE
;
2490 delta
= ap
->cur_time
- ap
->link_time
;
2491 if (delta
> ANEG_STATE_SETTLE_TIME
) {
2492 /* XXX another gem from the Broadcom driver :( */
2493 ap
->state
= ANEG_STATE_LINK_OK
;
2497 case ANEG_STATE_LINK_OK
:
2498 ap
->flags
|= (MR_AN_COMPLETE
| MR_LINK_OK
);
2502 case ANEG_STATE_NEXT_PAGE_WAIT_INIT
:
2503 /* ??? unimplemented */
2506 case ANEG_STATE_NEXT_PAGE_WAIT
:
2507 /* ??? unimplemented */
2518 static int fiber_autoneg(struct tg3
*tp
, u32
*flags
)
2521 struct tg3_fiber_aneginfo aninfo
;
2522 int status
= ANEG_FAILED
;
2526 tw32_f(MAC_TX_AUTO_NEG
, 0);
2528 tmp
= tp
->mac_mode
& ~MAC_MODE_PORT_MODE_MASK
;
2529 tw32_f(MAC_MODE
, tmp
| MAC_MODE_PORT_MODE_GMII
);
2532 tw32_f(MAC_MODE
, tp
->mac_mode
| MAC_MODE_SEND_CONFIGS
);
2535 memset(&aninfo
, 0, sizeof(aninfo
));
2536 aninfo
.flags
|= MR_AN_ENABLE
;
2537 aninfo
.state
= ANEG_STATE_UNKNOWN
;
2538 aninfo
.cur_time
= 0;
2540 while (++tick
< 195000) {
2541 status
= tg3_fiber_aneg_smachine(tp
, &aninfo
);
2542 if (status
== ANEG_DONE
|| status
== ANEG_FAILED
)
2548 tp
->mac_mode
&= ~MAC_MODE_SEND_CONFIGS
;
2549 tw32_f(MAC_MODE
, tp
->mac_mode
);
2552 *flags
= aninfo
.flags
;
2554 if (status
== ANEG_DONE
&&
2555 (aninfo
.flags
& (MR_AN_COMPLETE
| MR_LINK_OK
|
2556 MR_LP_ADV_FULL_DUPLEX
)))
2562 static void tg3_init_bcm8002(struct tg3
*tp
)
2564 u32 mac_status
= tr32(MAC_STATUS
);
2567 /* Reset when initting first time or we have a link. */
2568 if ((tp
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
) &&
2569 !(mac_status
& MAC_STATUS_PCS_SYNCED
))
2572 /* Set PLL lock range. */
2573 tg3_writephy(tp
, 0x16, 0x8007);
2576 tg3_writephy(tp
, MII_BMCR
, BMCR_RESET
);
2578 /* Wait for reset to complete. */
2579 /* XXX schedule_timeout() ... */
2580 for (i
= 0; i
< 500; i
++)
2583 /* Config mode; select PMA/Ch 1 regs. */
2584 tg3_writephy(tp
, 0x10, 0x8411);
2586 /* Enable auto-lock and comdet, select txclk for tx. */
2587 tg3_writephy(tp
, 0x11, 0x0a10);
2589 tg3_writephy(tp
, 0x18, 0x00a0);
2590 tg3_writephy(tp
, 0x16, 0x41ff);
2592 /* Assert and deassert POR. */
2593 tg3_writephy(tp
, 0x13, 0x0400);
2595 tg3_writephy(tp
, 0x13, 0x0000);
2597 tg3_writephy(tp
, 0x11, 0x0a50);
2599 tg3_writephy(tp
, 0x11, 0x0a10);
2601 /* Wait for signal to stabilize */
2602 /* XXX schedule_timeout() ... */
2603 for (i
= 0; i
< 15000; i
++)
2606 /* Deselect the channel register so we can read the PHYID
2609 tg3_writephy(tp
, 0x10, 0x8011);
2612 static int tg3_setup_fiber_hw_autoneg(struct tg3
*tp
, u32 mac_status
)
2614 u32 sg_dig_ctrl
, sg_dig_status
;
2615 u32 serdes_cfg
, expected_sg_dig_ctrl
;
2616 int workaround
, port_a
;
2617 int current_link_up
;
2620 expected_sg_dig_ctrl
= 0;
2623 current_link_up
= 0;
2625 if (tp
->pci_chip_rev_id
!= CHIPREV_ID_5704_A0
&&
2626 tp
->pci_chip_rev_id
!= CHIPREV_ID_5704_A1
) {
2628 if (tr32(TG3PCI_DUAL_MAC_CTRL
) & DUAL_MAC_CTRL_ID
)
2631 /* preserve bits 0-11,13,14 for signal pre-emphasis */
2632 /* preserve bits 20-23 for voltage regulator */
2633 serdes_cfg
= tr32(MAC_SERDES_CFG
) & 0x00f06fff;
2636 sg_dig_ctrl
= tr32(SG_DIG_CTRL
);
2638 if (tp
->link_config
.autoneg
!= AUTONEG_ENABLE
) {
2639 if (sg_dig_ctrl
& (1 << 31)) {
2641 u32 val
= serdes_cfg
;
2647 tw32_f(MAC_SERDES_CFG
, val
);
2649 tw32_f(SG_DIG_CTRL
, 0x01388400);
2651 if (mac_status
& MAC_STATUS_PCS_SYNCED
) {
2652 tg3_setup_flow_control(tp
, 0, 0);
2653 current_link_up
= 1;
2658 /* Want auto-negotiation. */
2659 expected_sg_dig_ctrl
= 0x81388400;
2661 /* Pause capability */
2662 expected_sg_dig_ctrl
|= (1 << 11);
2664 /* Asymettric pause */
2665 expected_sg_dig_ctrl
|= (1 << 12);
2667 if (sg_dig_ctrl
!= expected_sg_dig_ctrl
) {
2668 if ((tp
->tg3_flags2
& TG3_FLG2_PARALLEL_DETECT
) &&
2669 tp
->serdes_counter
&&
2670 ((mac_status
& (MAC_STATUS_PCS_SYNCED
|
2671 MAC_STATUS_RCVD_CFG
)) ==
2672 MAC_STATUS_PCS_SYNCED
)) {
2673 tp
->serdes_counter
--;
2674 current_link_up
= 1;
2679 tw32_f(MAC_SERDES_CFG
, serdes_cfg
| 0xc011000);
2680 tw32_f(SG_DIG_CTRL
, expected_sg_dig_ctrl
| (1 << 30));
2682 tw32_f(SG_DIG_CTRL
, expected_sg_dig_ctrl
);
2684 tp
->serdes_counter
= SERDES_AN_TIMEOUT_5704S
;
2685 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
2686 } else if (mac_status
& (MAC_STATUS_PCS_SYNCED
|
2687 MAC_STATUS_SIGNAL_DET
)) {
2688 sg_dig_status
= tr32(SG_DIG_STATUS
);
2689 mac_status
= tr32(MAC_STATUS
);
2691 if ((sg_dig_status
& (1 << 1)) &&
2692 (mac_status
& MAC_STATUS_PCS_SYNCED
)) {
2693 u32 local_adv
, remote_adv
;
2695 local_adv
= ADVERTISE_PAUSE_CAP
;
2697 if (sg_dig_status
& (1 << 19))
2698 remote_adv
|= LPA_PAUSE_CAP
;
2699 if (sg_dig_status
& (1 << 20))
2700 remote_adv
|= LPA_PAUSE_ASYM
;
2702 tg3_setup_flow_control(tp
, local_adv
, remote_adv
);
2703 current_link_up
= 1;
2704 tp
->serdes_counter
= 0;
2705 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
2706 } else if (!(sg_dig_status
& (1 << 1))) {
2707 if (tp
->serdes_counter
)
2708 tp
->serdes_counter
--;
2711 u32 val
= serdes_cfg
;
2718 tw32_f(MAC_SERDES_CFG
, val
);
2721 tw32_f(SG_DIG_CTRL
, 0x01388400);
2724 /* Link parallel detection - link is up */
2725 /* only if we have PCS_SYNC and not */
2726 /* receiving config code words */
2727 mac_status
= tr32(MAC_STATUS
);
2728 if ((mac_status
& MAC_STATUS_PCS_SYNCED
) &&
2729 !(mac_status
& MAC_STATUS_RCVD_CFG
)) {
2730 tg3_setup_flow_control(tp
, 0, 0);
2731 current_link_up
= 1;
2733 TG3_FLG2_PARALLEL_DETECT
;
2734 tp
->serdes_counter
=
2735 SERDES_PARALLEL_DET_TIMEOUT
;
2737 goto restart_autoneg
;
2741 tp
->serdes_counter
= SERDES_AN_TIMEOUT_5704S
;
2742 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
2746 return current_link_up
;
2749 static int tg3_setup_fiber_by_hand(struct tg3
*tp
, u32 mac_status
)
2751 int current_link_up
= 0;
2753 if (!(mac_status
& MAC_STATUS_PCS_SYNCED
))
2756 if (tp
->link_config
.autoneg
== AUTONEG_ENABLE
) {
2760 if (fiber_autoneg(tp
, &flags
)) {
2761 u32 local_adv
, remote_adv
;
2763 local_adv
= ADVERTISE_PAUSE_CAP
;
2765 if (flags
& MR_LP_ADV_SYM_PAUSE
)
2766 remote_adv
|= LPA_PAUSE_CAP
;
2767 if (flags
& MR_LP_ADV_ASYM_PAUSE
)
2768 remote_adv
|= LPA_PAUSE_ASYM
;
2770 tg3_setup_flow_control(tp
, local_adv
, remote_adv
);
2772 current_link_up
= 1;
2774 for (i
= 0; i
< 30; i
++) {
2777 (MAC_STATUS_SYNC_CHANGED
|
2778 MAC_STATUS_CFG_CHANGED
));
2780 if ((tr32(MAC_STATUS
) &
2781 (MAC_STATUS_SYNC_CHANGED
|
2782 MAC_STATUS_CFG_CHANGED
)) == 0)
2786 mac_status
= tr32(MAC_STATUS
);
2787 if (current_link_up
== 0 &&
2788 (mac_status
& MAC_STATUS_PCS_SYNCED
) &&
2789 !(mac_status
& MAC_STATUS_RCVD_CFG
))
2790 current_link_up
= 1;
2792 /* Forcing 1000FD link up. */
2793 current_link_up
= 1;
2795 tw32_f(MAC_MODE
, (tp
->mac_mode
| MAC_MODE_SEND_CONFIGS
));
2798 tw32_f(MAC_MODE
, tp
->mac_mode
);
2803 return current_link_up
;
2806 static int tg3_setup_fiber_phy(struct tg3
*tp
, int force_reset
)
2809 u16 orig_active_speed
;
2810 u8 orig_active_duplex
;
2812 int current_link_up
;
2816 (tp
->tg3_flags
& (TG3_FLAG_RX_PAUSE
|
2817 TG3_FLAG_TX_PAUSE
));
2818 orig_active_speed
= tp
->link_config
.active_speed
;
2819 orig_active_duplex
= tp
->link_config
.active_duplex
;
2821 if (!(tp
->tg3_flags2
& TG3_FLG2_HW_AUTONEG
) &&
2822 netif_carrier_ok(tp
->dev
) &&
2823 (tp
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
)) {
2824 mac_status
= tr32(MAC_STATUS
);
2825 mac_status
&= (MAC_STATUS_PCS_SYNCED
|
2826 MAC_STATUS_SIGNAL_DET
|
2827 MAC_STATUS_CFG_CHANGED
|
2828 MAC_STATUS_RCVD_CFG
);
2829 if (mac_status
== (MAC_STATUS_PCS_SYNCED
|
2830 MAC_STATUS_SIGNAL_DET
)) {
2831 tw32_f(MAC_STATUS
, (MAC_STATUS_SYNC_CHANGED
|
2832 MAC_STATUS_CFG_CHANGED
));
2837 tw32_f(MAC_TX_AUTO_NEG
, 0);
2839 tp
->mac_mode
&= ~(MAC_MODE_PORT_MODE_MASK
| MAC_MODE_HALF_DUPLEX
);
2840 tp
->mac_mode
|= MAC_MODE_PORT_MODE_TBI
;
2841 tw32_f(MAC_MODE
, tp
->mac_mode
);
2844 if (tp
->phy_id
== PHY_ID_BCM8002
)
2845 tg3_init_bcm8002(tp
);
2847 /* Enable link change event even when serdes polling. */
2848 tw32_f(MAC_EVENT
, MAC_EVENT_LNKSTATE_CHANGED
);
2851 current_link_up
= 0;
2852 mac_status
= tr32(MAC_STATUS
);
2854 if (tp
->tg3_flags2
& TG3_FLG2_HW_AUTONEG
)
2855 current_link_up
= tg3_setup_fiber_hw_autoneg(tp
, mac_status
);
2857 current_link_up
= tg3_setup_fiber_by_hand(tp
, mac_status
);
2859 tp
->hw_status
->status
=
2860 (SD_STATUS_UPDATED
|
2861 (tp
->hw_status
->status
& ~SD_STATUS_LINK_CHG
));
2863 for (i
= 0; i
< 100; i
++) {
2864 tw32_f(MAC_STATUS
, (MAC_STATUS_SYNC_CHANGED
|
2865 MAC_STATUS_CFG_CHANGED
));
2867 if ((tr32(MAC_STATUS
) & (MAC_STATUS_SYNC_CHANGED
|
2868 MAC_STATUS_CFG_CHANGED
|
2869 MAC_STATUS_LNKSTATE_CHANGED
)) == 0)
2873 mac_status
= tr32(MAC_STATUS
);
2874 if ((mac_status
& MAC_STATUS_PCS_SYNCED
) == 0) {
2875 current_link_up
= 0;
2876 if (tp
->link_config
.autoneg
== AUTONEG_ENABLE
&&
2877 tp
->serdes_counter
== 0) {
2878 tw32_f(MAC_MODE
, (tp
->mac_mode
|
2879 MAC_MODE_SEND_CONFIGS
));
2881 tw32_f(MAC_MODE
, tp
->mac_mode
);
2885 if (current_link_up
== 1) {
2886 tp
->link_config
.active_speed
= SPEED_1000
;
2887 tp
->link_config
.active_duplex
= DUPLEX_FULL
;
2888 tw32(MAC_LED_CTRL
, (tp
->led_ctrl
|
2889 LED_CTRL_LNKLED_OVERRIDE
|
2890 LED_CTRL_1000MBPS_ON
));
2892 tp
->link_config
.active_speed
= SPEED_INVALID
;
2893 tp
->link_config
.active_duplex
= DUPLEX_INVALID
;
2894 tw32(MAC_LED_CTRL
, (tp
->led_ctrl
|
2895 LED_CTRL_LNKLED_OVERRIDE
|
2896 LED_CTRL_TRAFFIC_OVERRIDE
));
2899 if (current_link_up
!= netif_carrier_ok(tp
->dev
)) {
2900 if (current_link_up
)
2901 netif_carrier_on(tp
->dev
);
2903 netif_carrier_off(tp
->dev
);
2904 tg3_link_report(tp
);
2907 tp
->tg3_flags
& (TG3_FLAG_RX_PAUSE
|
2909 if (orig_pause_cfg
!= now_pause_cfg
||
2910 orig_active_speed
!= tp
->link_config
.active_speed
||
2911 orig_active_duplex
!= tp
->link_config
.active_duplex
)
2912 tg3_link_report(tp
);
2918 static int tg3_setup_fiber_mii_phy(struct tg3
*tp
, int force_reset
)
2920 int current_link_up
, err
= 0;
2925 tp
->mac_mode
|= MAC_MODE_PORT_MODE_GMII
;
2926 tw32_f(MAC_MODE
, tp
->mac_mode
);
2932 (MAC_STATUS_SYNC_CHANGED
|
2933 MAC_STATUS_CFG_CHANGED
|
2934 MAC_STATUS_MI_COMPLETION
|
2935 MAC_STATUS_LNKSTATE_CHANGED
));
2941 current_link_up
= 0;
2942 current_speed
= SPEED_INVALID
;
2943 current_duplex
= DUPLEX_INVALID
;
2945 err
|= tg3_readphy(tp
, MII_BMSR
, &bmsr
);
2946 err
|= tg3_readphy(tp
, MII_BMSR
, &bmsr
);
2947 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
) {
2948 if (tr32(MAC_TX_STATUS
) & TX_STATUS_LINK_UP
)
2949 bmsr
|= BMSR_LSTATUS
;
2951 bmsr
&= ~BMSR_LSTATUS
;
2954 err
|= tg3_readphy(tp
, MII_BMCR
, &bmcr
);
2956 if ((tp
->link_config
.autoneg
== AUTONEG_ENABLE
) && !force_reset
&&
2957 (tp
->tg3_flags2
& TG3_FLG2_PARALLEL_DETECT
)) {
2958 /* do nothing, just check for link up at the end */
2959 } else if (tp
->link_config
.autoneg
== AUTONEG_ENABLE
) {
2962 err
|= tg3_readphy(tp
, MII_ADVERTISE
, &adv
);
2963 new_adv
= adv
& ~(ADVERTISE_1000XFULL
| ADVERTISE_1000XHALF
|
2964 ADVERTISE_1000XPAUSE
|
2965 ADVERTISE_1000XPSE_ASYM
|
2968 /* Always advertise symmetric PAUSE just like copper */
2969 new_adv
|= ADVERTISE_1000XPAUSE
;
2971 if (tp
->link_config
.advertising
& ADVERTISED_1000baseT_Half
)
2972 new_adv
|= ADVERTISE_1000XHALF
;
2973 if (tp
->link_config
.advertising
& ADVERTISED_1000baseT_Full
)
2974 new_adv
|= ADVERTISE_1000XFULL
;
2976 if ((new_adv
!= adv
) || !(bmcr
& BMCR_ANENABLE
)) {
2977 tg3_writephy(tp
, MII_ADVERTISE
, new_adv
);
2978 bmcr
|= BMCR_ANENABLE
| BMCR_ANRESTART
;
2979 tg3_writephy(tp
, MII_BMCR
, bmcr
);
2981 tw32_f(MAC_EVENT
, MAC_EVENT_LNKSTATE_CHANGED
);
2982 tp
->serdes_counter
= SERDES_AN_TIMEOUT_5714S
;
2983 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
2990 bmcr
&= ~BMCR_SPEED1000
;
2991 new_bmcr
= bmcr
& ~(BMCR_ANENABLE
| BMCR_FULLDPLX
);
2993 if (tp
->link_config
.duplex
== DUPLEX_FULL
)
2994 new_bmcr
|= BMCR_FULLDPLX
;
2996 if (new_bmcr
!= bmcr
) {
2997 /* BMCR_SPEED1000 is a reserved bit that needs
2998 * to be set on write.
3000 new_bmcr
|= BMCR_SPEED1000
;
3002 /* Force a linkdown */
3003 if (netif_carrier_ok(tp
->dev
)) {
3006 err
|= tg3_readphy(tp
, MII_ADVERTISE
, &adv
);
3007 adv
&= ~(ADVERTISE_1000XFULL
|
3008 ADVERTISE_1000XHALF
|
3010 tg3_writephy(tp
, MII_ADVERTISE
, adv
);
3011 tg3_writephy(tp
, MII_BMCR
, bmcr
|
3015 netif_carrier_off(tp
->dev
);
3017 tg3_writephy(tp
, MII_BMCR
, new_bmcr
);
3019 err
|= tg3_readphy(tp
, MII_BMSR
, &bmsr
);
3020 err
|= tg3_readphy(tp
, MII_BMSR
, &bmsr
);
3021 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
3023 if (tr32(MAC_TX_STATUS
) & TX_STATUS_LINK_UP
)
3024 bmsr
|= BMSR_LSTATUS
;
3026 bmsr
&= ~BMSR_LSTATUS
;
3028 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
3032 if (bmsr
& BMSR_LSTATUS
) {
3033 current_speed
= SPEED_1000
;
3034 current_link_up
= 1;
3035 if (bmcr
& BMCR_FULLDPLX
)
3036 current_duplex
= DUPLEX_FULL
;
3038 current_duplex
= DUPLEX_HALF
;
3040 if (bmcr
& BMCR_ANENABLE
) {
3041 u32 local_adv
, remote_adv
, common
;
3043 err
|= tg3_readphy(tp
, MII_ADVERTISE
, &local_adv
);
3044 err
|= tg3_readphy(tp
, MII_LPA
, &remote_adv
);
3045 common
= local_adv
& remote_adv
;
3046 if (common
& (ADVERTISE_1000XHALF
|
3047 ADVERTISE_1000XFULL
)) {
3048 if (common
& ADVERTISE_1000XFULL
)
3049 current_duplex
= DUPLEX_FULL
;
3051 current_duplex
= DUPLEX_HALF
;
3053 tg3_setup_flow_control(tp
, local_adv
,
3057 current_link_up
= 0;
3061 tp
->mac_mode
&= ~MAC_MODE_HALF_DUPLEX
;
3062 if (tp
->link_config
.active_duplex
== DUPLEX_HALF
)
3063 tp
->mac_mode
|= MAC_MODE_HALF_DUPLEX
;
3065 tw32_f(MAC_MODE
, tp
->mac_mode
);
3068 tw32_f(MAC_EVENT
, MAC_EVENT_LNKSTATE_CHANGED
);
3070 tp
->link_config
.active_speed
= current_speed
;
3071 tp
->link_config
.active_duplex
= current_duplex
;
3073 if (current_link_up
!= netif_carrier_ok(tp
->dev
)) {
3074 if (current_link_up
)
3075 netif_carrier_on(tp
->dev
);
3077 netif_carrier_off(tp
->dev
);
3078 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
3080 tg3_link_report(tp
);
3085 static void tg3_serdes_parallel_detect(struct tg3
*tp
)
3087 if (tp
->serdes_counter
) {
3088 /* Give autoneg time to complete. */
3089 tp
->serdes_counter
--;
3092 if (!netif_carrier_ok(tp
->dev
) &&
3093 (tp
->link_config
.autoneg
== AUTONEG_ENABLE
)) {
3096 tg3_readphy(tp
, MII_BMCR
, &bmcr
);
3097 if (bmcr
& BMCR_ANENABLE
) {
3100 /* Select shadow register 0x1f */
3101 tg3_writephy(tp
, 0x1c, 0x7c00);
3102 tg3_readphy(tp
, 0x1c, &phy1
);
3104 /* Select expansion interrupt status register */
3105 tg3_writephy(tp
, 0x17, 0x0f01);
3106 tg3_readphy(tp
, 0x15, &phy2
);
3107 tg3_readphy(tp
, 0x15, &phy2
);
3109 if ((phy1
& 0x10) && !(phy2
& 0x20)) {
3110 /* We have signal detect and not receiving
3111 * config code words, link is up by parallel
3115 bmcr
&= ~BMCR_ANENABLE
;
3116 bmcr
|= BMCR_SPEED1000
| BMCR_FULLDPLX
;
3117 tg3_writephy(tp
, MII_BMCR
, bmcr
);
3118 tp
->tg3_flags2
|= TG3_FLG2_PARALLEL_DETECT
;
3122 else if (netif_carrier_ok(tp
->dev
) &&
3123 (tp
->link_config
.autoneg
== AUTONEG_ENABLE
) &&
3124 (tp
->tg3_flags2
& TG3_FLG2_PARALLEL_DETECT
)) {
3127 /* Select expansion interrupt status register */
3128 tg3_writephy(tp
, 0x17, 0x0f01);
3129 tg3_readphy(tp
, 0x15, &phy2
);
3133 /* Config code words received, turn on autoneg. */
3134 tg3_readphy(tp
, MII_BMCR
, &bmcr
);
3135 tg3_writephy(tp
, MII_BMCR
, bmcr
| BMCR_ANENABLE
);
3137 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
3143 static int tg3_setup_phy(struct tg3
*tp
, int force_reset
)
3147 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
3148 err
= tg3_setup_fiber_phy(tp
, force_reset
);
3149 } else if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) {
3150 err
= tg3_setup_fiber_mii_phy(tp
, force_reset
);
3152 err
= tg3_setup_copper_phy(tp
, force_reset
);
3155 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A0
||
3156 tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A1
) {
3159 val
= tr32(TG3_CPMU_CLCK_STAT
) & CPMU_CLCK_STAT_MAC_CLCK_MASK
;
3160 if (val
== CPMU_CLCK_STAT_MAC_CLCK_62_5
)
3162 else if (val
== CPMU_CLCK_STAT_MAC_CLCK_6_25
)
3167 val
= tr32(GRC_MISC_CFG
) & ~GRC_MISC_CFG_PRESCALAR_MASK
;
3168 val
|= (scale
<< GRC_MISC_CFG_PRESCALAR_SHIFT
);
3169 tw32(GRC_MISC_CFG
, val
);
3172 if (tp
->link_config
.active_speed
== SPEED_1000
&&
3173 tp
->link_config
.active_duplex
== DUPLEX_HALF
)
3174 tw32(MAC_TX_LENGTHS
,
3175 ((2 << TX_LENGTHS_IPG_CRS_SHIFT
) |
3176 (6 << TX_LENGTHS_IPG_SHIFT
) |
3177 (0xff << TX_LENGTHS_SLOT_TIME_SHIFT
)));
3179 tw32(MAC_TX_LENGTHS
,
3180 ((2 << TX_LENGTHS_IPG_CRS_SHIFT
) |
3181 (6 << TX_LENGTHS_IPG_SHIFT
) |
3182 (32 << TX_LENGTHS_SLOT_TIME_SHIFT
)));
3184 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
3185 if (netif_carrier_ok(tp
->dev
)) {
3186 tw32(HOSTCC_STAT_COAL_TICKS
,
3187 tp
->coal
.stats_block_coalesce_usecs
);
3189 tw32(HOSTCC_STAT_COAL_TICKS
, 0);
3193 if (tp
->tg3_flags
& TG3_FLAG_ASPM_WORKAROUND
) {
3194 u32 val
= tr32(PCIE_PWR_MGMT_THRESH
);
3195 if (!netif_carrier_ok(tp
->dev
))
3196 val
= (val
& ~PCIE_PWR_MGMT_L1_THRESH_MSK
) |
3199 val
|= PCIE_PWR_MGMT_L1_THRESH_MSK
;
3200 tw32(PCIE_PWR_MGMT_THRESH
, val
);
3206 /* This is called whenever we suspect that the system chipset is re-
3207 * ordering the sequence of MMIO to the tx send mailbox. The symptom
3208 * is bogus tx completions. We try to recover by setting the
3209 * TG3_FLAG_MBOX_WRITE_REORDER flag and resetting the chip later
3212 static void tg3_tx_recover(struct tg3
*tp
)
3214 BUG_ON((tp
->tg3_flags
& TG3_FLAG_MBOX_WRITE_REORDER
) ||
3215 tp
->write32_tx_mbox
== tg3_write_indirect_mbox
);
3217 printk(KERN_WARNING PFX
"%s: The system may be re-ordering memory-"
3218 "mapped I/O cycles to the network device, attempting to "
3219 "recover. Please report the problem to the driver maintainer "
3220 "and include system chipset information.\n", tp
->dev
->name
);
3222 spin_lock(&tp
->lock
);
3223 tp
->tg3_flags
|= TG3_FLAG_TX_RECOVERY_PENDING
;
3224 spin_unlock(&tp
->lock
);
3227 static inline u32
tg3_tx_avail(struct tg3
*tp
)
3230 return (tp
->tx_pending
-
3231 ((tp
->tx_prod
- tp
->tx_cons
) & (TG3_TX_RING_SIZE
- 1)));
3234 /* Tigon3 never reports partial packet sends. So we do not
3235 * need special logic to handle SKBs that have not had all
3236 * of their frags sent yet, like SunGEM does.
3238 static void tg3_tx(struct tg3
*tp
)
3240 u32 hw_idx
= tp
->hw_status
->idx
[0].tx_consumer
;
3241 u32 sw_idx
= tp
->tx_cons
;
3243 while (sw_idx
!= hw_idx
) {
3244 struct tx_ring_info
*ri
= &tp
->tx_buffers
[sw_idx
];
3245 struct sk_buff
*skb
= ri
->skb
;
3248 if (unlikely(skb
== NULL
)) {
3253 pci_unmap_single(tp
->pdev
,
3254 pci_unmap_addr(ri
, mapping
),
3260 sw_idx
= NEXT_TX(sw_idx
);
3262 for (i
= 0; i
< skb_shinfo(skb
)->nr_frags
; i
++) {
3263 ri
= &tp
->tx_buffers
[sw_idx
];
3264 if (unlikely(ri
->skb
!= NULL
|| sw_idx
== hw_idx
))
3267 pci_unmap_page(tp
->pdev
,
3268 pci_unmap_addr(ri
, mapping
),
3269 skb_shinfo(skb
)->frags
[i
].size
,
3272 sw_idx
= NEXT_TX(sw_idx
);
3277 if (unlikely(tx_bug
)) {
3283 tp
->tx_cons
= sw_idx
;
3285 /* Need to make the tx_cons update visible to tg3_start_xmit()
3286 * before checking for netif_queue_stopped(). Without the
3287 * memory barrier, there is a small possibility that tg3_start_xmit()
3288 * will miss it and cause the queue to be stopped forever.
3292 if (unlikely(netif_queue_stopped(tp
->dev
) &&
3293 (tg3_tx_avail(tp
) > TG3_TX_WAKEUP_THRESH(tp
)))) {
3294 netif_tx_lock(tp
->dev
);
3295 if (netif_queue_stopped(tp
->dev
) &&
3296 (tg3_tx_avail(tp
) > TG3_TX_WAKEUP_THRESH(tp
)))
3297 netif_wake_queue(tp
->dev
);
3298 netif_tx_unlock(tp
->dev
);
3302 /* Returns size of skb allocated or < 0 on error.
3304 * We only need to fill in the address because the other members
3305 * of the RX descriptor are invariant, see tg3_init_rings.
3307 * Note the purposeful assymetry of cpu vs. chip accesses. For
3308 * posting buffers we only dirty the first cache line of the RX
3309 * descriptor (containing the address). Whereas for the RX status
3310 * buffers the cpu only reads the last cacheline of the RX descriptor
3311 * (to fetch the error flags, vlan tag, checksum, and opaque cookie).
3313 static int tg3_alloc_rx_skb(struct tg3
*tp
, u32 opaque_key
,
3314 int src_idx
, u32 dest_idx_unmasked
)
3316 struct tg3_rx_buffer_desc
*desc
;
3317 struct ring_info
*map
, *src_map
;
3318 struct sk_buff
*skb
;
3320 int skb_size
, dest_idx
;
3323 switch (opaque_key
) {
3324 case RXD_OPAQUE_RING_STD
:
3325 dest_idx
= dest_idx_unmasked
% TG3_RX_RING_SIZE
;
3326 desc
= &tp
->rx_std
[dest_idx
];
3327 map
= &tp
->rx_std_buffers
[dest_idx
];
3329 src_map
= &tp
->rx_std_buffers
[src_idx
];
3330 skb_size
= tp
->rx_pkt_buf_sz
;
3333 case RXD_OPAQUE_RING_JUMBO
:
3334 dest_idx
= dest_idx_unmasked
% TG3_RX_JUMBO_RING_SIZE
;
3335 desc
= &tp
->rx_jumbo
[dest_idx
];
3336 map
= &tp
->rx_jumbo_buffers
[dest_idx
];
3338 src_map
= &tp
->rx_jumbo_buffers
[src_idx
];
3339 skb_size
= RX_JUMBO_PKT_BUF_SZ
;
3346 /* Do not overwrite any of the map or rp information
3347 * until we are sure we can commit to a new buffer.
3349 * Callers depend upon this behavior and assume that
3350 * we leave everything unchanged if we fail.
3352 skb
= netdev_alloc_skb(tp
->dev
, skb_size
);
3356 skb_reserve(skb
, tp
->rx_offset
);
3358 mapping
= pci_map_single(tp
->pdev
, skb
->data
,
3359 skb_size
- tp
->rx_offset
,
3360 PCI_DMA_FROMDEVICE
);
3363 pci_unmap_addr_set(map
, mapping
, mapping
);
3365 if (src_map
!= NULL
)
3366 src_map
->skb
= NULL
;
3368 desc
->addr_hi
= ((u64
)mapping
>> 32);
3369 desc
->addr_lo
= ((u64
)mapping
& 0xffffffff);
3374 /* We only need to move over in the address because the other
3375 * members of the RX descriptor are invariant. See notes above
3376 * tg3_alloc_rx_skb for full details.
3378 static void tg3_recycle_rx(struct tg3
*tp
, u32 opaque_key
,
3379 int src_idx
, u32 dest_idx_unmasked
)
3381 struct tg3_rx_buffer_desc
*src_desc
, *dest_desc
;
3382 struct ring_info
*src_map
, *dest_map
;
3385 switch (opaque_key
) {
3386 case RXD_OPAQUE_RING_STD
:
3387 dest_idx
= dest_idx_unmasked
% TG3_RX_RING_SIZE
;
3388 dest_desc
= &tp
->rx_std
[dest_idx
];
3389 dest_map
= &tp
->rx_std_buffers
[dest_idx
];
3390 src_desc
= &tp
->rx_std
[src_idx
];
3391 src_map
= &tp
->rx_std_buffers
[src_idx
];
3394 case RXD_OPAQUE_RING_JUMBO
:
3395 dest_idx
= dest_idx_unmasked
% TG3_RX_JUMBO_RING_SIZE
;
3396 dest_desc
= &tp
->rx_jumbo
[dest_idx
];
3397 dest_map
= &tp
->rx_jumbo_buffers
[dest_idx
];
3398 src_desc
= &tp
->rx_jumbo
[src_idx
];
3399 src_map
= &tp
->rx_jumbo_buffers
[src_idx
];
3406 dest_map
->skb
= src_map
->skb
;
3407 pci_unmap_addr_set(dest_map
, mapping
,
3408 pci_unmap_addr(src_map
, mapping
));
3409 dest_desc
->addr_hi
= src_desc
->addr_hi
;
3410 dest_desc
->addr_lo
= src_desc
->addr_lo
;
3412 src_map
->skb
= NULL
;
3415 #if TG3_VLAN_TAG_USED
3416 static int tg3_vlan_rx(struct tg3
*tp
, struct sk_buff
*skb
, u16 vlan_tag
)
3418 return vlan_hwaccel_receive_skb(skb
, tp
->vlgrp
, vlan_tag
);
3422 /* The RX ring scheme is composed of multiple rings which post fresh
3423 * buffers to the chip, and one special ring the chip uses to report
3424 * status back to the host.
3426 * The special ring reports the status of received packets to the
3427 * host. The chip does not write into the original descriptor the
3428 * RX buffer was obtained from. The chip simply takes the original
3429 * descriptor as provided by the host, updates the status and length
3430 * field, then writes this into the next status ring entry.
3432 * Each ring the host uses to post buffers to the chip is described
3433 * by a TG3_BDINFO entry in the chips SRAM area. When a packet arrives,
3434 * it is first placed into the on-chip ram. When the packet's length
3435 * is known, it walks down the TG3_BDINFO entries to select the ring.
3436 * Each TG3_BDINFO specifies a MAXLEN field and the first TG3_BDINFO
3437 * which is within the range of the new packet's length is chosen.
3439 * The "separate ring for rx status" scheme may sound queer, but it makes
3440 * sense from a cache coherency perspective. If only the host writes
3441 * to the buffer post rings, and only the chip writes to the rx status
3442 * rings, then cache lines never move beyond shared-modified state.
3443 * If both the host and chip were to write into the same ring, cache line
3444 * eviction could occur since both entities want it in an exclusive state.
3446 static int tg3_rx(struct tg3
*tp
, int budget
)
3448 u32 work_mask
, rx_std_posted
= 0;
3449 u32 sw_idx
= tp
->rx_rcb_ptr
;
3453 hw_idx
= tp
->hw_status
->idx
[0].rx_producer
;
3455 * We need to order the read of hw_idx and the read of
3456 * the opaque cookie.
3461 while (sw_idx
!= hw_idx
&& budget
> 0) {
3462 struct tg3_rx_buffer_desc
*desc
= &tp
->rx_rcb
[sw_idx
];
3464 struct sk_buff
*skb
;
3465 dma_addr_t dma_addr
;
3466 u32 opaque_key
, desc_idx
, *post_ptr
;
3468 desc_idx
= desc
->opaque
& RXD_OPAQUE_INDEX_MASK
;
3469 opaque_key
= desc
->opaque
& RXD_OPAQUE_RING_MASK
;
3470 if (opaque_key
== RXD_OPAQUE_RING_STD
) {
3471 dma_addr
= pci_unmap_addr(&tp
->rx_std_buffers
[desc_idx
],
3473 skb
= tp
->rx_std_buffers
[desc_idx
].skb
;
3474 post_ptr
= &tp
->rx_std_ptr
;
3476 } else if (opaque_key
== RXD_OPAQUE_RING_JUMBO
) {
3477 dma_addr
= pci_unmap_addr(&tp
->rx_jumbo_buffers
[desc_idx
],
3479 skb
= tp
->rx_jumbo_buffers
[desc_idx
].skb
;
3480 post_ptr
= &tp
->rx_jumbo_ptr
;
3483 goto next_pkt_nopost
;
3486 work_mask
|= opaque_key
;
3488 if ((desc
->err_vlan
& RXD_ERR_MASK
) != 0 &&
3489 (desc
->err_vlan
!= RXD_ERR_ODD_NIBBLE_RCVD_MII
)) {
3491 tg3_recycle_rx(tp
, opaque_key
,
3492 desc_idx
, *post_ptr
);
3494 /* Other statistics kept track of by card. */
3495 tp
->net_stats
.rx_dropped
++;
3499 len
= ((desc
->idx_len
& RXD_LEN_MASK
) >> RXD_LEN_SHIFT
) - 4; /* omit crc */
3501 if (len
> RX_COPY_THRESHOLD
3502 && tp
->rx_offset
== 2
3503 /* rx_offset != 2 iff this is a 5701 card running
3504 * in PCI-X mode [see tg3_get_invariants()] */
3508 skb_size
= tg3_alloc_rx_skb(tp
, opaque_key
,
3509 desc_idx
, *post_ptr
);
3513 pci_unmap_single(tp
->pdev
, dma_addr
,
3514 skb_size
- tp
->rx_offset
,
3515 PCI_DMA_FROMDEVICE
);
3519 struct sk_buff
*copy_skb
;
3521 tg3_recycle_rx(tp
, opaque_key
,
3522 desc_idx
, *post_ptr
);
3524 copy_skb
= netdev_alloc_skb(tp
->dev
, len
+ 2);
3525 if (copy_skb
== NULL
)
3526 goto drop_it_no_recycle
;
3528 skb_reserve(copy_skb
, 2);
3529 skb_put(copy_skb
, len
);
3530 pci_dma_sync_single_for_cpu(tp
->pdev
, dma_addr
, len
, PCI_DMA_FROMDEVICE
);
3531 skb_copy_from_linear_data(skb
, copy_skb
->data
, len
);
3532 pci_dma_sync_single_for_device(tp
->pdev
, dma_addr
, len
, PCI_DMA_FROMDEVICE
);
3534 /* We'll reuse the original ring buffer. */
3538 if ((tp
->tg3_flags
& TG3_FLAG_RX_CHECKSUMS
) &&
3539 (desc
->type_flags
& RXD_FLAG_TCPUDP_CSUM
) &&
3540 (((desc
->ip_tcp_csum
& RXD_TCPCSUM_MASK
)
3541 >> RXD_TCPCSUM_SHIFT
) == 0xffff))
3542 skb
->ip_summed
= CHECKSUM_UNNECESSARY
;
3544 skb
->ip_summed
= CHECKSUM_NONE
;
3546 skb
->protocol
= eth_type_trans(skb
, tp
->dev
);
3547 #if TG3_VLAN_TAG_USED
3548 if (tp
->vlgrp
!= NULL
&&
3549 desc
->type_flags
& RXD_FLAG_VLAN
) {
3550 tg3_vlan_rx(tp
, skb
,
3551 desc
->err_vlan
& RXD_VLAN_MASK
);
3554 netif_receive_skb(skb
);
3556 tp
->dev
->last_rx
= jiffies
;
3563 if (unlikely(rx_std_posted
>= tp
->rx_std_max_post
)) {
3564 u32 idx
= *post_ptr
% TG3_RX_RING_SIZE
;
3566 tw32_rx_mbox(MAILBOX_RCV_STD_PROD_IDX
+
3567 TG3_64BIT_REG_LOW
, idx
);
3568 work_mask
&= ~RXD_OPAQUE_RING_STD
;
3573 sw_idx
&= (TG3_RX_RCB_RING_SIZE(tp
) - 1);
3575 /* Refresh hw_idx to see if there is new work */
3576 if (sw_idx
== hw_idx
) {
3577 hw_idx
= tp
->hw_status
->idx
[0].rx_producer
;
3582 /* ACK the status ring. */
3583 tp
->rx_rcb_ptr
= sw_idx
;
3584 tw32_rx_mbox(MAILBOX_RCVRET_CON_IDX_0
+ TG3_64BIT_REG_LOW
, sw_idx
);
3586 /* Refill RX ring(s). */
3587 if (work_mask
& RXD_OPAQUE_RING_STD
) {
3588 sw_idx
= tp
->rx_std_ptr
% TG3_RX_RING_SIZE
;
3589 tw32_rx_mbox(MAILBOX_RCV_STD_PROD_IDX
+ TG3_64BIT_REG_LOW
,
3592 if (work_mask
& RXD_OPAQUE_RING_JUMBO
) {
3593 sw_idx
= tp
->rx_jumbo_ptr
% TG3_RX_JUMBO_RING_SIZE
;
3594 tw32_rx_mbox(MAILBOX_RCV_JUMBO_PROD_IDX
+ TG3_64BIT_REG_LOW
,
3602 static int tg3_poll_work(struct tg3
*tp
, int work_done
, int budget
)
3604 struct tg3_hw_status
*sblk
= tp
->hw_status
;
3606 /* handle link change and other phy events */
3607 if (!(tp
->tg3_flags
&
3608 (TG3_FLAG_USE_LINKCHG_REG
|
3609 TG3_FLAG_POLL_SERDES
))) {
3610 if (sblk
->status
& SD_STATUS_LINK_CHG
) {
3611 sblk
->status
= SD_STATUS_UPDATED
|
3612 (sblk
->status
& ~SD_STATUS_LINK_CHG
);
3613 spin_lock(&tp
->lock
);
3614 tg3_setup_phy(tp
, 0);
3615 spin_unlock(&tp
->lock
);
3619 /* run TX completion thread */
3620 if (sblk
->idx
[0].tx_consumer
!= tp
->tx_cons
) {
3622 if (unlikely(tp
->tg3_flags
& TG3_FLAG_TX_RECOVERY_PENDING
))
3626 /* run RX thread, within the bounds set by NAPI.
3627 * All RX "locking" is done by ensuring outside
3628 * code synchronizes with tg3->napi.poll()
3630 if (sblk
->idx
[0].rx_producer
!= tp
->rx_rcb_ptr
)
3631 work_done
+= tg3_rx(tp
, budget
- work_done
);
3636 static int tg3_poll(struct napi_struct
*napi
, int budget
)
3638 struct tg3
*tp
= container_of(napi
, struct tg3
, napi
);
3640 struct tg3_hw_status
*sblk
= tp
->hw_status
;
3643 work_done
= tg3_poll_work(tp
, work_done
, budget
);
3645 if (unlikely(tp
->tg3_flags
& TG3_FLAG_TX_RECOVERY_PENDING
))
3648 if (unlikely(work_done
>= budget
))
3651 if (tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
) {
3652 /* tp->last_tag is used in tg3_restart_ints() below
3653 * to tell the hw how much work has been processed,
3654 * so we must read it before checking for more work.
3656 tp
->last_tag
= sblk
->status_tag
;
3659 sblk
->status
&= ~SD_STATUS_UPDATED
;
3661 if (likely(!tg3_has_work(tp
))) {
3662 netif_rx_complete(tp
->dev
, napi
);
3663 tg3_restart_ints(tp
);
3671 /* work_done is guaranteed to be less than budget. */
3672 netif_rx_complete(tp
->dev
, napi
);
3673 schedule_work(&tp
->reset_task
);
3677 static void tg3_irq_quiesce(struct tg3
*tp
)
3679 BUG_ON(tp
->irq_sync
);
3684 synchronize_irq(tp
->pdev
->irq
);
3687 static inline int tg3_irq_sync(struct tg3
*tp
)
3689 return tp
->irq_sync
;
3692 /* Fully shutdown all tg3 driver activity elsewhere in the system.
3693 * If irq_sync is non-zero, then the IRQ handler must be synchronized
3694 * with as well. Most of the time, this is not necessary except when
3695 * shutting down the device.
3697 static inline void tg3_full_lock(struct tg3
*tp
, int irq_sync
)
3699 spin_lock_bh(&tp
->lock
);
3701 tg3_irq_quiesce(tp
);
3704 static inline void tg3_full_unlock(struct tg3
*tp
)
3706 spin_unlock_bh(&tp
->lock
);
3709 /* One-shot MSI handler - Chip automatically disables interrupt
3710 * after sending MSI so driver doesn't have to do it.
3712 static irqreturn_t
tg3_msi_1shot(int irq
, void *dev_id
)
3714 struct net_device
*dev
= dev_id
;
3715 struct tg3
*tp
= netdev_priv(dev
);
3717 prefetch(tp
->hw_status
);
3718 prefetch(&tp
->rx_rcb
[tp
->rx_rcb_ptr
]);
3720 if (likely(!tg3_irq_sync(tp
)))
3721 netif_rx_schedule(dev
, &tp
->napi
);
3726 /* MSI ISR - No need to check for interrupt sharing and no need to
3727 * flush status block and interrupt mailbox. PCI ordering rules
3728 * guarantee that MSI will arrive after the status block.
3730 static irqreturn_t
tg3_msi(int irq
, void *dev_id
)
3732 struct net_device
*dev
= dev_id
;
3733 struct tg3
*tp
= netdev_priv(dev
);
3735 prefetch(tp
->hw_status
);
3736 prefetch(&tp
->rx_rcb
[tp
->rx_rcb_ptr
]);
3738 * Writing any value to intr-mbox-0 clears PCI INTA# and
3739 * chip-internal interrupt pending events.
3740 * Writing non-zero to intr-mbox-0 additional tells the
3741 * NIC to stop sending us irqs, engaging "in-intr-handler"
3744 tw32_mailbox(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
, 0x00000001);
3745 if (likely(!tg3_irq_sync(tp
)))
3746 netif_rx_schedule(dev
, &tp
->napi
);
3748 return IRQ_RETVAL(1);
3751 static irqreturn_t
tg3_interrupt(int irq
, void *dev_id
)
3753 struct net_device
*dev
= dev_id
;
3754 struct tg3
*tp
= netdev_priv(dev
);
3755 struct tg3_hw_status
*sblk
= tp
->hw_status
;
3756 unsigned int handled
= 1;
3758 /* In INTx mode, it is possible for the interrupt to arrive at
3759 * the CPU before the status block posted prior to the interrupt.
3760 * Reading the PCI State register will confirm whether the
3761 * interrupt is ours and will flush the status block.
3763 if (unlikely(!(sblk
->status
& SD_STATUS_UPDATED
))) {
3764 if ((tp
->tg3_flags
& TG3_FLAG_CHIP_RESETTING
) ||
3765 (tr32(TG3PCI_PCISTATE
) & PCISTATE_INT_NOT_ACTIVE
)) {
3772 * Writing any value to intr-mbox-0 clears PCI INTA# and
3773 * chip-internal interrupt pending events.
3774 * Writing non-zero to intr-mbox-0 additional tells the
3775 * NIC to stop sending us irqs, engaging "in-intr-handler"
3778 * Flush the mailbox to de-assert the IRQ immediately to prevent
3779 * spurious interrupts. The flush impacts performance but
3780 * excessive spurious interrupts can be worse in some cases.
3782 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
, 0x00000001);
3783 if (tg3_irq_sync(tp
))
3785 sblk
->status
&= ~SD_STATUS_UPDATED
;
3786 if (likely(tg3_has_work(tp
))) {
3787 prefetch(&tp
->rx_rcb
[tp
->rx_rcb_ptr
]);
3788 netif_rx_schedule(dev
, &tp
->napi
);
3790 /* No work, shared interrupt perhaps? re-enable
3791 * interrupts, and flush that PCI write
3793 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
,
3797 return IRQ_RETVAL(handled
);
3800 static irqreturn_t
tg3_interrupt_tagged(int irq
, void *dev_id
)
3802 struct net_device
*dev
= dev_id
;
3803 struct tg3
*tp
= netdev_priv(dev
);
3804 struct tg3_hw_status
*sblk
= tp
->hw_status
;
3805 unsigned int handled
= 1;
3807 /* In INTx mode, it is possible for the interrupt to arrive at
3808 * the CPU before the status block posted prior to the interrupt.
3809 * Reading the PCI State register will confirm whether the
3810 * interrupt is ours and will flush the status block.
3812 if (unlikely(sblk
->status_tag
== tp
->last_tag
)) {
3813 if ((tp
->tg3_flags
& TG3_FLAG_CHIP_RESETTING
) ||
3814 (tr32(TG3PCI_PCISTATE
) & PCISTATE_INT_NOT_ACTIVE
)) {
3821 * writing any value to intr-mbox-0 clears PCI INTA# and
3822 * chip-internal interrupt pending events.
3823 * writing non-zero to intr-mbox-0 additional tells the
3824 * NIC to stop sending us irqs, engaging "in-intr-handler"
3827 * Flush the mailbox to de-assert the IRQ immediately to prevent
3828 * spurious interrupts. The flush impacts performance but
3829 * excessive spurious interrupts can be worse in some cases.
3831 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
, 0x00000001);
3832 if (tg3_irq_sync(tp
))
3834 if (netif_rx_schedule_prep(dev
, &tp
->napi
)) {
3835 prefetch(&tp
->rx_rcb
[tp
->rx_rcb_ptr
]);
3836 /* Update last_tag to mark that this status has been
3837 * seen. Because interrupt may be shared, we may be
3838 * racing with tg3_poll(), so only update last_tag
3839 * if tg3_poll() is not scheduled.
3841 tp
->last_tag
= sblk
->status_tag
;
3842 __netif_rx_schedule(dev
, &tp
->napi
);
3845 return IRQ_RETVAL(handled
);
3848 /* ISR for interrupt test */
3849 static irqreturn_t
tg3_test_isr(int irq
, void *dev_id
)
3851 struct net_device
*dev
= dev_id
;
3852 struct tg3
*tp
= netdev_priv(dev
);
3853 struct tg3_hw_status
*sblk
= tp
->hw_status
;
3855 if ((sblk
->status
& SD_STATUS_UPDATED
) ||
3856 !(tr32(TG3PCI_PCISTATE
) & PCISTATE_INT_NOT_ACTIVE
)) {
3857 tg3_disable_ints(tp
);
3858 return IRQ_RETVAL(1);
3860 return IRQ_RETVAL(0);
3863 static int tg3_init_hw(struct tg3
*, int);
3864 static int tg3_halt(struct tg3
*, int, int);
3866 /* Restart hardware after configuration changes, self-test, etc.
3867 * Invoked with tp->lock held.
3869 static int tg3_restart_hw(struct tg3
*tp
, int reset_phy
)
3873 err
= tg3_init_hw(tp
, reset_phy
);
3875 printk(KERN_ERR PFX
"%s: Failed to re-initialize device, "
3876 "aborting.\n", tp
->dev
->name
);
3877 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
3878 tg3_full_unlock(tp
);
3879 del_timer_sync(&tp
->timer
);
3881 napi_enable(&tp
->napi
);
3883 tg3_full_lock(tp
, 0);
3888 #ifdef CONFIG_NET_POLL_CONTROLLER
3889 static void tg3_poll_controller(struct net_device
*dev
)
3891 struct tg3
*tp
= netdev_priv(dev
);
3893 tg3_interrupt(tp
->pdev
->irq
, dev
);
3897 static void tg3_reset_task(struct work_struct
*work
)
3899 struct tg3
*tp
= container_of(work
, struct tg3
, reset_task
);
3900 unsigned int restart_timer
;
3902 tg3_full_lock(tp
, 0);
3904 if (!netif_running(tp
->dev
)) {
3905 tg3_full_unlock(tp
);
3909 tg3_full_unlock(tp
);
3913 tg3_full_lock(tp
, 1);
3915 restart_timer
= tp
->tg3_flags2
& TG3_FLG2_RESTART_TIMER
;
3916 tp
->tg3_flags2
&= ~TG3_FLG2_RESTART_TIMER
;
3918 if (tp
->tg3_flags
& TG3_FLAG_TX_RECOVERY_PENDING
) {
3919 tp
->write32_tx_mbox
= tg3_write32_tx_mbox
;
3920 tp
->write32_rx_mbox
= tg3_write_flush_reg32
;
3921 tp
->tg3_flags
|= TG3_FLAG_MBOX_WRITE_REORDER
;
3922 tp
->tg3_flags
&= ~TG3_FLAG_TX_RECOVERY_PENDING
;
3925 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 0);
3926 if (tg3_init_hw(tp
, 1))
3929 tg3_netif_start(tp
);
3932 mod_timer(&tp
->timer
, jiffies
+ 1);
3935 tg3_full_unlock(tp
);
3938 static void tg3_dump_short_state(struct tg3
*tp
)
3940 printk(KERN_ERR PFX
"DEBUG: MAC_TX_STATUS[%08x] MAC_RX_STATUS[%08x]\n",
3941 tr32(MAC_TX_STATUS
), tr32(MAC_RX_STATUS
));
3942 printk(KERN_ERR PFX
"DEBUG: RDMAC_STATUS[%08x] WDMAC_STATUS[%08x]\n",
3943 tr32(RDMAC_STATUS
), tr32(WDMAC_STATUS
));
3946 static void tg3_tx_timeout(struct net_device
*dev
)
3948 struct tg3
*tp
= netdev_priv(dev
);
3950 if (netif_msg_tx_err(tp
)) {
3951 printk(KERN_ERR PFX
"%s: transmit timed out, resetting\n",
3953 tg3_dump_short_state(tp
);
3956 schedule_work(&tp
->reset_task
);
3959 /* Test for DMA buffers crossing any 4GB boundaries: 4G, 8G, etc */
3960 static inline int tg3_4g_overflow_test(dma_addr_t mapping
, int len
)
3962 u32 base
= (u32
) mapping
& 0xffffffff;
3964 return ((base
> 0xffffdcc0) &&
3965 (base
+ len
+ 8 < base
));
3968 /* Test for DMA addresses > 40-bit */
3969 static inline int tg3_40bit_overflow_test(struct tg3
*tp
, dma_addr_t mapping
,
3972 #if defined(CONFIG_HIGHMEM) && (BITS_PER_LONG == 64)
3973 if (tp
->tg3_flags
& TG3_FLAG_40BIT_DMA_BUG
)
3974 return (((u64
) mapping
+ len
) > DMA_40BIT_MASK
);
3981 static void tg3_set_txd(struct tg3
*, int, dma_addr_t
, int, u32
, u32
);
3983 /* Workaround 4GB and 40-bit hardware DMA bugs. */
3984 static int tigon3_dma_hwbug_workaround(struct tg3
*tp
, struct sk_buff
*skb
,
3985 u32 last_plus_one
, u32
*start
,
3986 u32 base_flags
, u32 mss
)
3988 struct sk_buff
*new_skb
= skb_copy(skb
, GFP_ATOMIC
);
3989 dma_addr_t new_addr
= 0;
3996 /* New SKB is guaranteed to be linear. */
3998 new_addr
= pci_map_single(tp
->pdev
, new_skb
->data
, new_skb
->len
,
4000 /* Make sure new skb does not cross any 4G boundaries.
4001 * Drop the packet if it does.
4003 if (tg3_4g_overflow_test(new_addr
, new_skb
->len
)) {
4005 dev_kfree_skb(new_skb
);
4008 tg3_set_txd(tp
, entry
, new_addr
, new_skb
->len
,
4009 base_flags
, 1 | (mss
<< 1));
4010 *start
= NEXT_TX(entry
);
4014 /* Now clean up the sw ring entries. */
4016 while (entry
!= last_plus_one
) {
4020 len
= skb_headlen(skb
);
4022 len
= skb_shinfo(skb
)->frags
[i
-1].size
;
4023 pci_unmap_single(tp
->pdev
,
4024 pci_unmap_addr(&tp
->tx_buffers
[entry
], mapping
),
4025 len
, PCI_DMA_TODEVICE
);
4027 tp
->tx_buffers
[entry
].skb
= new_skb
;
4028 pci_unmap_addr_set(&tp
->tx_buffers
[entry
], mapping
, new_addr
);
4030 tp
->tx_buffers
[entry
].skb
= NULL
;
4032 entry
= NEXT_TX(entry
);
4041 static void tg3_set_txd(struct tg3
*tp
, int entry
,
4042 dma_addr_t mapping
, int len
, u32 flags
,
4045 struct tg3_tx_buffer_desc
*txd
= &tp
->tx_ring
[entry
];
4046 int is_end
= (mss_and_is_end
& 0x1);
4047 u32 mss
= (mss_and_is_end
>> 1);
4051 flags
|= TXD_FLAG_END
;
4052 if (flags
& TXD_FLAG_VLAN
) {
4053 vlan_tag
= flags
>> 16;
4056 vlan_tag
|= (mss
<< TXD_MSS_SHIFT
);
4058 txd
->addr_hi
= ((u64
) mapping
>> 32);
4059 txd
->addr_lo
= ((u64
) mapping
& 0xffffffff);
4060 txd
->len_flags
= (len
<< TXD_LEN_SHIFT
) | flags
;
4061 txd
->vlan_tag
= vlan_tag
<< TXD_VLAN_TAG_SHIFT
;
4064 /* hard_start_xmit for devices that don't have any bugs and
4065 * support TG3_FLG2_HW_TSO_2 only.
4067 static int tg3_start_xmit(struct sk_buff
*skb
, struct net_device
*dev
)
4069 struct tg3
*tp
= netdev_priv(dev
);
4071 u32 len
, entry
, base_flags
, mss
;
4073 len
= skb_headlen(skb
);
4075 /* We are running in BH disabled context with netif_tx_lock
4076 * and TX reclaim runs via tp->napi.poll inside of a software
4077 * interrupt. Furthermore, IRQ processing runs lockless so we have
4078 * no IRQ context deadlocks to worry about either. Rejoice!
4080 if (unlikely(tg3_tx_avail(tp
) <= (skb_shinfo(skb
)->nr_frags
+ 1))) {
4081 if (!netif_queue_stopped(dev
)) {
4082 netif_stop_queue(dev
);
4084 /* This is a hard error, log it. */
4085 printk(KERN_ERR PFX
"%s: BUG! Tx Ring full when "
4086 "queue awake!\n", dev
->name
);
4088 return NETDEV_TX_BUSY
;
4091 entry
= tp
->tx_prod
;
4094 if ((mss
= skb_shinfo(skb
)->gso_size
) != 0) {
4095 int tcp_opt_len
, ip_tcp_len
;
4097 if (skb_header_cloned(skb
) &&
4098 pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
)) {
4103 if (skb_shinfo(skb
)->gso_type
& SKB_GSO_TCPV6
)
4104 mss
|= (skb_headlen(skb
) - ETH_HLEN
) << 9;
4106 struct iphdr
*iph
= ip_hdr(skb
);
4108 tcp_opt_len
= tcp_optlen(skb
);
4109 ip_tcp_len
= ip_hdrlen(skb
) + sizeof(struct tcphdr
);
4112 iph
->tot_len
= htons(mss
+ ip_tcp_len
+ tcp_opt_len
);
4113 mss
|= (ip_tcp_len
+ tcp_opt_len
) << 9;
4116 base_flags
|= (TXD_FLAG_CPU_PRE_DMA
|
4117 TXD_FLAG_CPU_POST_DMA
);
4119 tcp_hdr(skb
)->check
= 0;
4122 else if (skb
->ip_summed
== CHECKSUM_PARTIAL
)
4123 base_flags
|= TXD_FLAG_TCPUDP_CSUM
;
4124 #if TG3_VLAN_TAG_USED
4125 if (tp
->vlgrp
!= NULL
&& vlan_tx_tag_present(skb
))
4126 base_flags
|= (TXD_FLAG_VLAN
|
4127 (vlan_tx_tag_get(skb
) << 16));
4130 /* Queue skb data, a.k.a. the main skb fragment. */
4131 mapping
= pci_map_single(tp
->pdev
, skb
->data
, len
, PCI_DMA_TODEVICE
);
4133 tp
->tx_buffers
[entry
].skb
= skb
;
4134 pci_unmap_addr_set(&tp
->tx_buffers
[entry
], mapping
, mapping
);
4136 tg3_set_txd(tp
, entry
, mapping
, len
, base_flags
,
4137 (skb_shinfo(skb
)->nr_frags
== 0) | (mss
<< 1));
4139 entry
= NEXT_TX(entry
);
4141 /* Now loop through additional data fragments, and queue them. */
4142 if (skb_shinfo(skb
)->nr_frags
> 0) {
4143 unsigned int i
, last
;
4145 last
= skb_shinfo(skb
)->nr_frags
- 1;
4146 for (i
= 0; i
<= last
; i
++) {
4147 skb_frag_t
*frag
= &skb_shinfo(skb
)->frags
[i
];
4150 mapping
= pci_map_page(tp
->pdev
,
4153 len
, PCI_DMA_TODEVICE
);
4155 tp
->tx_buffers
[entry
].skb
= NULL
;
4156 pci_unmap_addr_set(&tp
->tx_buffers
[entry
], mapping
, mapping
);
4158 tg3_set_txd(tp
, entry
, mapping
, len
,
4159 base_flags
, (i
== last
) | (mss
<< 1));
4161 entry
= NEXT_TX(entry
);
4165 /* Packets are ready, update Tx producer idx local and on card. */
4166 tw32_tx_mbox((MAILBOX_SNDHOST_PROD_IDX_0
+ TG3_64BIT_REG_LOW
), entry
);
4168 tp
->tx_prod
= entry
;
4169 if (unlikely(tg3_tx_avail(tp
) <= (MAX_SKB_FRAGS
+ 1))) {
4170 netif_stop_queue(dev
);
4171 if (tg3_tx_avail(tp
) > TG3_TX_WAKEUP_THRESH(tp
))
4172 netif_wake_queue(tp
->dev
);
4178 dev
->trans_start
= jiffies
;
4180 return NETDEV_TX_OK
;
4183 static int tg3_start_xmit_dma_bug(struct sk_buff
*, struct net_device
*);
4185 /* Use GSO to workaround a rare TSO bug that may be triggered when the
4186 * TSO header is greater than 80 bytes.
4188 static int tg3_tso_bug(struct tg3
*tp
, struct sk_buff
*skb
)
4190 struct sk_buff
*segs
, *nskb
;
4192 /* Estimate the number of fragments in the worst case */
4193 if (unlikely(tg3_tx_avail(tp
) <= (skb_shinfo(skb
)->gso_segs
* 3))) {
4194 netif_stop_queue(tp
->dev
);
4195 if (tg3_tx_avail(tp
) <= (skb_shinfo(skb
)->gso_segs
* 3))
4196 return NETDEV_TX_BUSY
;
4198 netif_wake_queue(tp
->dev
);
4201 segs
= skb_gso_segment(skb
, tp
->dev
->features
& ~NETIF_F_TSO
);
4202 if (unlikely(IS_ERR(segs
)))
4203 goto tg3_tso_bug_end
;
4209 tg3_start_xmit_dma_bug(nskb
, tp
->dev
);
4215 return NETDEV_TX_OK
;
4218 /* hard_start_xmit for devices that have the 4G bug and/or 40-bit bug and
4219 * support TG3_FLG2_HW_TSO_1 or firmware TSO only.
4221 static int tg3_start_xmit_dma_bug(struct sk_buff
*skb
, struct net_device
*dev
)
4223 struct tg3
*tp
= netdev_priv(dev
);
4225 u32 len
, entry
, base_flags
, mss
;
4226 int would_hit_hwbug
;
4228 len
= skb_headlen(skb
);
4230 /* We are running in BH disabled context with netif_tx_lock
4231 * and TX reclaim runs via tp->napi.poll inside of a software
4232 * interrupt. Furthermore, IRQ processing runs lockless so we have
4233 * no IRQ context deadlocks to worry about either. Rejoice!
4235 if (unlikely(tg3_tx_avail(tp
) <= (skb_shinfo(skb
)->nr_frags
+ 1))) {
4236 if (!netif_queue_stopped(dev
)) {
4237 netif_stop_queue(dev
);
4239 /* This is a hard error, log it. */
4240 printk(KERN_ERR PFX
"%s: BUG! Tx Ring full when "
4241 "queue awake!\n", dev
->name
);
4243 return NETDEV_TX_BUSY
;
4246 entry
= tp
->tx_prod
;
4248 if (skb
->ip_summed
== CHECKSUM_PARTIAL
)
4249 base_flags
|= TXD_FLAG_TCPUDP_CSUM
;
4251 if ((mss
= skb_shinfo(skb
)->gso_size
) != 0) {
4253 int tcp_opt_len
, ip_tcp_len
, hdr_len
;
4255 if (skb_header_cloned(skb
) &&
4256 pskb_expand_head(skb
, 0, 0, GFP_ATOMIC
)) {
4261 tcp_opt_len
= tcp_optlen(skb
);
4262 ip_tcp_len
= ip_hdrlen(skb
) + sizeof(struct tcphdr
);
4264 hdr_len
= ip_tcp_len
+ tcp_opt_len
;
4265 if (unlikely((ETH_HLEN
+ hdr_len
) > 80) &&
4266 (tp
->tg3_flags2
& TG3_FLG2_TSO_BUG
))
4267 return (tg3_tso_bug(tp
, skb
));
4269 base_flags
|= (TXD_FLAG_CPU_PRE_DMA
|
4270 TXD_FLAG_CPU_POST_DMA
);
4274 iph
->tot_len
= htons(mss
+ hdr_len
);
4275 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
) {
4276 tcp_hdr(skb
)->check
= 0;
4277 base_flags
&= ~TXD_FLAG_TCPUDP_CSUM
;
4279 tcp_hdr(skb
)->check
= ~csum_tcpudp_magic(iph
->saddr
,
4284 if ((tp
->tg3_flags2
& TG3_FLG2_HW_TSO
) ||
4285 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
)) {
4286 if (tcp_opt_len
|| iph
->ihl
> 5) {
4289 tsflags
= (iph
->ihl
- 5) + (tcp_opt_len
>> 2);
4290 mss
|= (tsflags
<< 11);
4293 if (tcp_opt_len
|| iph
->ihl
> 5) {
4296 tsflags
= (iph
->ihl
- 5) + (tcp_opt_len
>> 2);
4297 base_flags
|= tsflags
<< 12;
4301 #if TG3_VLAN_TAG_USED
4302 if (tp
->vlgrp
!= NULL
&& vlan_tx_tag_present(skb
))
4303 base_flags
|= (TXD_FLAG_VLAN
|
4304 (vlan_tx_tag_get(skb
) << 16));
4307 /* Queue skb data, a.k.a. the main skb fragment. */
4308 mapping
= pci_map_single(tp
->pdev
, skb
->data
, len
, PCI_DMA_TODEVICE
);
4310 tp
->tx_buffers
[entry
].skb
= skb
;
4311 pci_unmap_addr_set(&tp
->tx_buffers
[entry
], mapping
, mapping
);
4313 would_hit_hwbug
= 0;
4315 if (tg3_4g_overflow_test(mapping
, len
))
4316 would_hit_hwbug
= 1;
4318 tg3_set_txd(tp
, entry
, mapping
, len
, base_flags
,
4319 (skb_shinfo(skb
)->nr_frags
== 0) | (mss
<< 1));
4321 entry
= NEXT_TX(entry
);
4323 /* Now loop through additional data fragments, and queue them. */
4324 if (skb_shinfo(skb
)->nr_frags
> 0) {
4325 unsigned int i
, last
;
4327 last
= skb_shinfo(skb
)->nr_frags
- 1;
4328 for (i
= 0; i
<= last
; i
++) {
4329 skb_frag_t
*frag
= &skb_shinfo(skb
)->frags
[i
];
4332 mapping
= pci_map_page(tp
->pdev
,
4335 len
, PCI_DMA_TODEVICE
);
4337 tp
->tx_buffers
[entry
].skb
= NULL
;
4338 pci_unmap_addr_set(&tp
->tx_buffers
[entry
], mapping
, mapping
);
4340 if (tg3_4g_overflow_test(mapping
, len
))
4341 would_hit_hwbug
= 1;
4343 if (tg3_40bit_overflow_test(tp
, mapping
, len
))
4344 would_hit_hwbug
= 1;
4346 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
)
4347 tg3_set_txd(tp
, entry
, mapping
, len
,
4348 base_flags
, (i
== last
)|(mss
<< 1));
4350 tg3_set_txd(tp
, entry
, mapping
, len
,
4351 base_flags
, (i
== last
));
4353 entry
= NEXT_TX(entry
);
4357 if (would_hit_hwbug
) {
4358 u32 last_plus_one
= entry
;
4361 start
= entry
- 1 - skb_shinfo(skb
)->nr_frags
;
4362 start
&= (TG3_TX_RING_SIZE
- 1);
4364 /* If the workaround fails due to memory/mapping
4365 * failure, silently drop this packet.
4367 if (tigon3_dma_hwbug_workaround(tp
, skb
, last_plus_one
,
4368 &start
, base_flags
, mss
))
4374 /* Packets are ready, update Tx producer idx local and on card. */
4375 tw32_tx_mbox((MAILBOX_SNDHOST_PROD_IDX_0
+ TG3_64BIT_REG_LOW
), entry
);
4377 tp
->tx_prod
= entry
;
4378 if (unlikely(tg3_tx_avail(tp
) <= (MAX_SKB_FRAGS
+ 1))) {
4379 netif_stop_queue(dev
);
4380 if (tg3_tx_avail(tp
) > TG3_TX_WAKEUP_THRESH(tp
))
4381 netif_wake_queue(tp
->dev
);
4387 dev
->trans_start
= jiffies
;
4389 return NETDEV_TX_OK
;
4392 static inline void tg3_set_mtu(struct net_device
*dev
, struct tg3
*tp
,
4397 if (new_mtu
> ETH_DATA_LEN
) {
4398 if (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
) {
4399 tp
->tg3_flags2
&= ~TG3_FLG2_TSO_CAPABLE
;
4400 ethtool_op_set_tso(dev
, 0);
4403 tp
->tg3_flags
|= TG3_FLAG_JUMBO_RING_ENABLE
;
4405 if (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)
4406 tp
->tg3_flags2
|= TG3_FLG2_TSO_CAPABLE
;
4407 tp
->tg3_flags
&= ~TG3_FLAG_JUMBO_RING_ENABLE
;
4411 static int tg3_change_mtu(struct net_device
*dev
, int new_mtu
)
4413 struct tg3
*tp
= netdev_priv(dev
);
4416 if (new_mtu
< TG3_MIN_MTU
|| new_mtu
> TG3_MAX_MTU(tp
))
4419 if (!netif_running(dev
)) {
4420 /* We'll just catch it later when the
4423 tg3_set_mtu(dev
, tp
, new_mtu
);
4429 tg3_full_lock(tp
, 1);
4431 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
4433 tg3_set_mtu(dev
, tp
, new_mtu
);
4435 err
= tg3_restart_hw(tp
, 0);
4438 tg3_netif_start(tp
);
4440 tg3_full_unlock(tp
);
4445 /* Free up pending packets in all rx/tx rings.
4447 * The chip has been shut down and the driver detached from
4448 * the networking, so no interrupts or new tx packets will
4449 * end up in the driver. tp->{tx,}lock is not held and we are not
4450 * in an interrupt context and thus may sleep.
4452 static void tg3_free_rings(struct tg3
*tp
)
4454 struct ring_info
*rxp
;
4457 for (i
= 0; i
< TG3_RX_RING_SIZE
; i
++) {
4458 rxp
= &tp
->rx_std_buffers
[i
];
4460 if (rxp
->skb
== NULL
)
4462 pci_unmap_single(tp
->pdev
,
4463 pci_unmap_addr(rxp
, mapping
),
4464 tp
->rx_pkt_buf_sz
- tp
->rx_offset
,
4465 PCI_DMA_FROMDEVICE
);
4466 dev_kfree_skb_any(rxp
->skb
);
4470 for (i
= 0; i
< TG3_RX_JUMBO_RING_SIZE
; i
++) {
4471 rxp
= &tp
->rx_jumbo_buffers
[i
];
4473 if (rxp
->skb
== NULL
)
4475 pci_unmap_single(tp
->pdev
,
4476 pci_unmap_addr(rxp
, mapping
),
4477 RX_JUMBO_PKT_BUF_SZ
- tp
->rx_offset
,
4478 PCI_DMA_FROMDEVICE
);
4479 dev_kfree_skb_any(rxp
->skb
);
4483 for (i
= 0; i
< TG3_TX_RING_SIZE
; ) {
4484 struct tx_ring_info
*txp
;
4485 struct sk_buff
*skb
;
4488 txp
= &tp
->tx_buffers
[i
];
4496 pci_unmap_single(tp
->pdev
,
4497 pci_unmap_addr(txp
, mapping
),
4504 for (j
= 0; j
< skb_shinfo(skb
)->nr_frags
; j
++) {
4505 txp
= &tp
->tx_buffers
[i
& (TG3_TX_RING_SIZE
- 1)];
4506 pci_unmap_page(tp
->pdev
,
4507 pci_unmap_addr(txp
, mapping
),
4508 skb_shinfo(skb
)->frags
[j
].size
,
4513 dev_kfree_skb_any(skb
);
4517 /* Initialize tx/rx rings for packet processing.
4519 * The chip has been shut down and the driver detached from
4520 * the networking, so no interrupts or new tx packets will
4521 * end up in the driver. tp->{tx,}lock are held and thus
4524 static int tg3_init_rings(struct tg3
*tp
)
4528 /* Free up all the SKBs. */
4531 /* Zero out all descriptors. */
4532 memset(tp
->rx_std
, 0, TG3_RX_RING_BYTES
);
4533 memset(tp
->rx_jumbo
, 0, TG3_RX_JUMBO_RING_BYTES
);
4534 memset(tp
->rx_rcb
, 0, TG3_RX_RCB_RING_BYTES(tp
));
4535 memset(tp
->tx_ring
, 0, TG3_TX_RING_BYTES
);
4537 tp
->rx_pkt_buf_sz
= RX_PKT_BUF_SZ
;
4538 if ((tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
) &&
4539 (tp
->dev
->mtu
> ETH_DATA_LEN
))
4540 tp
->rx_pkt_buf_sz
= RX_JUMBO_PKT_BUF_SZ
;
4542 /* Initialize invariants of the rings, we only set this
4543 * stuff once. This works because the card does not
4544 * write into the rx buffer posting rings.
4546 for (i
= 0; i
< TG3_RX_RING_SIZE
; i
++) {
4547 struct tg3_rx_buffer_desc
*rxd
;
4549 rxd
= &tp
->rx_std
[i
];
4550 rxd
->idx_len
= (tp
->rx_pkt_buf_sz
- tp
->rx_offset
- 64)
4552 rxd
->type_flags
= (RXD_FLAG_END
<< RXD_FLAGS_SHIFT
);
4553 rxd
->opaque
= (RXD_OPAQUE_RING_STD
|
4554 (i
<< RXD_OPAQUE_INDEX_SHIFT
));
4557 if (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
) {
4558 for (i
= 0; i
< TG3_RX_JUMBO_RING_SIZE
; i
++) {
4559 struct tg3_rx_buffer_desc
*rxd
;
4561 rxd
= &tp
->rx_jumbo
[i
];
4562 rxd
->idx_len
= (RX_JUMBO_PKT_BUF_SZ
- tp
->rx_offset
- 64)
4564 rxd
->type_flags
= (RXD_FLAG_END
<< RXD_FLAGS_SHIFT
) |
4566 rxd
->opaque
= (RXD_OPAQUE_RING_JUMBO
|
4567 (i
<< RXD_OPAQUE_INDEX_SHIFT
));
4571 /* Now allocate fresh SKBs for each rx ring. */
4572 for (i
= 0; i
< tp
->rx_pending
; i
++) {
4573 if (tg3_alloc_rx_skb(tp
, RXD_OPAQUE_RING_STD
, -1, i
) < 0) {
4574 printk(KERN_WARNING PFX
4575 "%s: Using a smaller RX standard ring, "
4576 "only %d out of %d buffers were allocated "
4578 tp
->dev
->name
, i
, tp
->rx_pending
);
4586 if (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
) {
4587 for (i
= 0; i
< tp
->rx_jumbo_pending
; i
++) {
4588 if (tg3_alloc_rx_skb(tp
, RXD_OPAQUE_RING_JUMBO
,
4590 printk(KERN_WARNING PFX
4591 "%s: Using a smaller RX jumbo ring, "
4592 "only %d out of %d buffers were "
4593 "allocated successfully.\n",
4594 tp
->dev
->name
, i
, tp
->rx_jumbo_pending
);
4599 tp
->rx_jumbo_pending
= i
;
4608 * Must not be invoked with interrupt sources disabled and
4609 * the hardware shutdown down.
4611 static void tg3_free_consistent(struct tg3
*tp
)
4613 kfree(tp
->rx_std_buffers
);
4614 tp
->rx_std_buffers
= NULL
;
4616 pci_free_consistent(tp
->pdev
, TG3_RX_RING_BYTES
,
4617 tp
->rx_std
, tp
->rx_std_mapping
);
4621 pci_free_consistent(tp
->pdev
, TG3_RX_JUMBO_RING_BYTES
,
4622 tp
->rx_jumbo
, tp
->rx_jumbo_mapping
);
4623 tp
->rx_jumbo
= NULL
;
4626 pci_free_consistent(tp
->pdev
, TG3_RX_RCB_RING_BYTES(tp
),
4627 tp
->rx_rcb
, tp
->rx_rcb_mapping
);
4631 pci_free_consistent(tp
->pdev
, TG3_TX_RING_BYTES
,
4632 tp
->tx_ring
, tp
->tx_desc_mapping
);
4635 if (tp
->hw_status
) {
4636 pci_free_consistent(tp
->pdev
, TG3_HW_STATUS_SIZE
,
4637 tp
->hw_status
, tp
->status_mapping
);
4638 tp
->hw_status
= NULL
;
4641 pci_free_consistent(tp
->pdev
, sizeof(struct tg3_hw_stats
),
4642 tp
->hw_stats
, tp
->stats_mapping
);
4643 tp
->hw_stats
= NULL
;
4648 * Must not be invoked with interrupt sources disabled and
4649 * the hardware shutdown down. Can sleep.
4651 static int tg3_alloc_consistent(struct tg3
*tp
)
4653 tp
->rx_std_buffers
= kzalloc((sizeof(struct ring_info
) *
4655 TG3_RX_JUMBO_RING_SIZE
)) +
4656 (sizeof(struct tx_ring_info
) *
4659 if (!tp
->rx_std_buffers
)
4662 tp
->rx_jumbo_buffers
= &tp
->rx_std_buffers
[TG3_RX_RING_SIZE
];
4663 tp
->tx_buffers
= (struct tx_ring_info
*)
4664 &tp
->rx_jumbo_buffers
[TG3_RX_JUMBO_RING_SIZE
];
4666 tp
->rx_std
= pci_alloc_consistent(tp
->pdev
, TG3_RX_RING_BYTES
,
4667 &tp
->rx_std_mapping
);
4671 tp
->rx_jumbo
= pci_alloc_consistent(tp
->pdev
, TG3_RX_JUMBO_RING_BYTES
,
4672 &tp
->rx_jumbo_mapping
);
4677 tp
->rx_rcb
= pci_alloc_consistent(tp
->pdev
, TG3_RX_RCB_RING_BYTES(tp
),
4678 &tp
->rx_rcb_mapping
);
4682 tp
->tx_ring
= pci_alloc_consistent(tp
->pdev
, TG3_TX_RING_BYTES
,
4683 &tp
->tx_desc_mapping
);
4687 tp
->hw_status
= pci_alloc_consistent(tp
->pdev
,
4689 &tp
->status_mapping
);
4693 tp
->hw_stats
= pci_alloc_consistent(tp
->pdev
,
4694 sizeof(struct tg3_hw_stats
),
4695 &tp
->stats_mapping
);
4699 memset(tp
->hw_status
, 0, TG3_HW_STATUS_SIZE
);
4700 memset(tp
->hw_stats
, 0, sizeof(struct tg3_hw_stats
));
4705 tg3_free_consistent(tp
);
4709 #define MAX_WAIT_CNT 1000
4711 /* To stop a block, clear the enable bit and poll till it
4712 * clears. tp->lock is held.
4714 static int tg3_stop_block(struct tg3
*tp
, unsigned long ofs
, u32 enable_bit
, int silent
)
4719 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
4726 /* We can't enable/disable these bits of the
4727 * 5705/5750, just say success.
4740 for (i
= 0; i
< MAX_WAIT_CNT
; i
++) {
4743 if ((val
& enable_bit
) == 0)
4747 if (i
== MAX_WAIT_CNT
&& !silent
) {
4748 printk(KERN_ERR PFX
"tg3_stop_block timed out, "
4749 "ofs=%lx enable_bit=%x\n",
4757 /* tp->lock is held. */
4758 static int tg3_abort_hw(struct tg3
*tp
, int silent
)
4762 tg3_disable_ints(tp
);
4764 tp
->rx_mode
&= ~RX_MODE_ENABLE
;
4765 tw32_f(MAC_RX_MODE
, tp
->rx_mode
);
4768 err
= tg3_stop_block(tp
, RCVBDI_MODE
, RCVBDI_MODE_ENABLE
, silent
);
4769 err
|= tg3_stop_block(tp
, RCVLPC_MODE
, RCVLPC_MODE_ENABLE
, silent
);
4770 err
|= tg3_stop_block(tp
, RCVLSC_MODE
, RCVLSC_MODE_ENABLE
, silent
);
4771 err
|= tg3_stop_block(tp
, RCVDBDI_MODE
, RCVDBDI_MODE_ENABLE
, silent
);
4772 err
|= tg3_stop_block(tp
, RCVDCC_MODE
, RCVDCC_MODE_ENABLE
, silent
);
4773 err
|= tg3_stop_block(tp
, RCVCC_MODE
, RCVCC_MODE_ENABLE
, silent
);
4775 err
|= tg3_stop_block(tp
, SNDBDS_MODE
, SNDBDS_MODE_ENABLE
, silent
);
4776 err
|= tg3_stop_block(tp
, SNDBDI_MODE
, SNDBDI_MODE_ENABLE
, silent
);
4777 err
|= tg3_stop_block(tp
, SNDDATAI_MODE
, SNDDATAI_MODE_ENABLE
, silent
);
4778 err
|= tg3_stop_block(tp
, RDMAC_MODE
, RDMAC_MODE_ENABLE
, silent
);
4779 err
|= tg3_stop_block(tp
, SNDDATAC_MODE
, SNDDATAC_MODE_ENABLE
, silent
);
4780 err
|= tg3_stop_block(tp
, DMAC_MODE
, DMAC_MODE_ENABLE
, silent
);
4781 err
|= tg3_stop_block(tp
, SNDBDC_MODE
, SNDBDC_MODE_ENABLE
, silent
);
4783 tp
->mac_mode
&= ~MAC_MODE_TDE_ENABLE
;
4784 tw32_f(MAC_MODE
, tp
->mac_mode
);
4787 tp
->tx_mode
&= ~TX_MODE_ENABLE
;
4788 tw32_f(MAC_TX_MODE
, tp
->tx_mode
);
4790 for (i
= 0; i
< MAX_WAIT_CNT
; i
++) {
4792 if (!(tr32(MAC_TX_MODE
) & TX_MODE_ENABLE
))
4795 if (i
>= MAX_WAIT_CNT
) {
4796 printk(KERN_ERR PFX
"tg3_abort_hw timed out for %s, "
4797 "TX_MODE_ENABLE will not clear MAC_TX_MODE=%08x\n",
4798 tp
->dev
->name
, tr32(MAC_TX_MODE
));
4802 err
|= tg3_stop_block(tp
, HOSTCC_MODE
, HOSTCC_MODE_ENABLE
, silent
);
4803 err
|= tg3_stop_block(tp
, WDMAC_MODE
, WDMAC_MODE_ENABLE
, silent
);
4804 err
|= tg3_stop_block(tp
, MBFREE_MODE
, MBFREE_MODE_ENABLE
, silent
);
4806 tw32(FTQ_RESET
, 0xffffffff);
4807 tw32(FTQ_RESET
, 0x00000000);
4809 err
|= tg3_stop_block(tp
, BUFMGR_MODE
, BUFMGR_MODE_ENABLE
, silent
);
4810 err
|= tg3_stop_block(tp
, MEMARB_MODE
, MEMARB_MODE_ENABLE
, silent
);
4813 memset(tp
->hw_status
, 0, TG3_HW_STATUS_SIZE
);
4815 memset(tp
->hw_stats
, 0, sizeof(struct tg3_hw_stats
));
4820 /* tp->lock is held. */
4821 static int tg3_nvram_lock(struct tg3
*tp
)
4823 if (tp
->tg3_flags
& TG3_FLAG_NVRAM
) {
4826 if (tp
->nvram_lock_cnt
== 0) {
4827 tw32(NVRAM_SWARB
, SWARB_REQ_SET1
);
4828 for (i
= 0; i
< 8000; i
++) {
4829 if (tr32(NVRAM_SWARB
) & SWARB_GNT1
)
4834 tw32(NVRAM_SWARB
, SWARB_REQ_CLR1
);
4838 tp
->nvram_lock_cnt
++;
4843 /* tp->lock is held. */
4844 static void tg3_nvram_unlock(struct tg3
*tp
)
4846 if (tp
->tg3_flags
& TG3_FLAG_NVRAM
) {
4847 if (tp
->nvram_lock_cnt
> 0)
4848 tp
->nvram_lock_cnt
--;
4849 if (tp
->nvram_lock_cnt
== 0)
4850 tw32_f(NVRAM_SWARB
, SWARB_REQ_CLR1
);
4854 /* tp->lock is held. */
4855 static void tg3_enable_nvram_access(struct tg3
*tp
)
4857 if ((tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) &&
4858 !(tp
->tg3_flags2
& TG3_FLG2_PROTECTED_NVRAM
)) {
4859 u32 nvaccess
= tr32(NVRAM_ACCESS
);
4861 tw32(NVRAM_ACCESS
, nvaccess
| ACCESS_ENABLE
);
4865 /* tp->lock is held. */
4866 static void tg3_disable_nvram_access(struct tg3
*tp
)
4868 if ((tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) &&
4869 !(tp
->tg3_flags2
& TG3_FLG2_PROTECTED_NVRAM
)) {
4870 u32 nvaccess
= tr32(NVRAM_ACCESS
);
4872 tw32(NVRAM_ACCESS
, nvaccess
& ~ACCESS_ENABLE
);
4876 static void tg3_ape_send_event(struct tg3
*tp
, u32 event
)
4881 apedata
= tg3_ape_read32(tp
, TG3_APE_SEG_SIG
);
4882 if (apedata
!= APE_SEG_SIG_MAGIC
)
4885 apedata
= tg3_ape_read32(tp
, TG3_APE_FW_STATUS
);
4886 if (apedata
!= APE_FW_STATUS_READY
)
4889 /* Wait for up to 1 millisecond for APE to service previous event. */
4890 for (i
= 0; i
< 10; i
++) {
4891 if (tg3_ape_lock(tp
, TG3_APE_LOCK_MEM
))
4894 apedata
= tg3_ape_read32(tp
, TG3_APE_EVENT_STATUS
);
4896 if (!(apedata
& APE_EVENT_STATUS_EVENT_PENDING
))
4897 tg3_ape_write32(tp
, TG3_APE_EVENT_STATUS
,
4898 event
| APE_EVENT_STATUS_EVENT_PENDING
);
4900 tg3_ape_unlock(tp
, TG3_APE_LOCK_MEM
);
4902 if (!(apedata
& APE_EVENT_STATUS_EVENT_PENDING
))
4908 if (!(apedata
& APE_EVENT_STATUS_EVENT_PENDING
))
4909 tg3_ape_write32(tp
, TG3_APE_EVENT
, APE_EVENT_1
);
4912 static void tg3_ape_driver_state_change(struct tg3
*tp
, int kind
)
4917 if (!(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
))
4921 case RESET_KIND_INIT
:
4922 tg3_ape_write32(tp
, TG3_APE_HOST_SEG_SIG
,
4923 APE_HOST_SEG_SIG_MAGIC
);
4924 tg3_ape_write32(tp
, TG3_APE_HOST_SEG_LEN
,
4925 APE_HOST_SEG_LEN_MAGIC
);
4926 apedata
= tg3_ape_read32(tp
, TG3_APE_HOST_INIT_COUNT
);
4927 tg3_ape_write32(tp
, TG3_APE_HOST_INIT_COUNT
, ++apedata
);
4928 tg3_ape_write32(tp
, TG3_APE_HOST_DRIVER_ID
,
4929 APE_HOST_DRIVER_ID_MAGIC
);
4930 tg3_ape_write32(tp
, TG3_APE_HOST_BEHAVIOR
,
4931 APE_HOST_BEHAV_NO_PHYLOCK
);
4933 event
= APE_EVENT_STATUS_STATE_START
;
4935 case RESET_KIND_SHUTDOWN
:
4936 event
= APE_EVENT_STATUS_STATE_UNLOAD
;
4938 case RESET_KIND_SUSPEND
:
4939 event
= APE_EVENT_STATUS_STATE_SUSPEND
;
4945 event
|= APE_EVENT_STATUS_DRIVER_EVNT
| APE_EVENT_STATUS_STATE_CHNGE
;
4947 tg3_ape_send_event(tp
, event
);
4950 /* tp->lock is held. */
4951 static void tg3_write_sig_pre_reset(struct tg3
*tp
, int kind
)
4953 tg3_write_mem(tp
, NIC_SRAM_FIRMWARE_MBOX
,
4954 NIC_SRAM_FIRMWARE_MBOX_MAGIC1
);
4956 if (tp
->tg3_flags2
& TG3_FLG2_ASF_NEW_HANDSHAKE
) {
4958 case RESET_KIND_INIT
:
4959 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
4963 case RESET_KIND_SHUTDOWN
:
4964 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
4968 case RESET_KIND_SUSPEND
:
4969 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
4978 if (kind
== RESET_KIND_INIT
||
4979 kind
== RESET_KIND_SUSPEND
)
4980 tg3_ape_driver_state_change(tp
, kind
);
4983 /* tp->lock is held. */
4984 static void tg3_write_sig_post_reset(struct tg3
*tp
, int kind
)
4986 if (tp
->tg3_flags2
& TG3_FLG2_ASF_NEW_HANDSHAKE
) {
4988 case RESET_KIND_INIT
:
4989 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
4990 DRV_STATE_START_DONE
);
4993 case RESET_KIND_SHUTDOWN
:
4994 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
4995 DRV_STATE_UNLOAD_DONE
);
5003 if (kind
== RESET_KIND_SHUTDOWN
)
5004 tg3_ape_driver_state_change(tp
, kind
);
5007 /* tp->lock is held. */
5008 static void tg3_write_sig_legacy(struct tg3
*tp
, int kind
)
5010 if (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) {
5012 case RESET_KIND_INIT
:
5013 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
5017 case RESET_KIND_SHUTDOWN
:
5018 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
5022 case RESET_KIND_SUSPEND
:
5023 tg3_write_mem(tp
, NIC_SRAM_FW_DRV_STATE_MBOX
,
5033 static int tg3_poll_fw(struct tg3
*tp
)
5038 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
5039 /* Wait up to 20ms for init done. */
5040 for (i
= 0; i
< 200; i
++) {
5041 if (tr32(VCPU_STATUS
) & VCPU_STATUS_INIT_DONE
)
5048 /* Wait for firmware initialization to complete. */
5049 for (i
= 0; i
< 100000; i
++) {
5050 tg3_read_mem(tp
, NIC_SRAM_FIRMWARE_MBOX
, &val
);
5051 if (val
== ~NIC_SRAM_FIRMWARE_MBOX_MAGIC1
)
5056 /* Chip might not be fitted with firmware. Some Sun onboard
5057 * parts are configured like that. So don't signal the timeout
5058 * of the above loop as an error, but do report the lack of
5059 * running firmware once.
5062 !(tp
->tg3_flags2
& TG3_FLG2_NO_FWARE_REPORTED
)) {
5063 tp
->tg3_flags2
|= TG3_FLG2_NO_FWARE_REPORTED
;
5065 printk(KERN_INFO PFX
"%s: No firmware running.\n",
5072 /* Save PCI command register before chip reset */
5073 static void tg3_save_pci_state(struct tg3
*tp
)
5075 pci_read_config_word(tp
->pdev
, PCI_COMMAND
, &tp
->pci_cmd
);
5078 /* Restore PCI state after chip reset */
5079 static void tg3_restore_pci_state(struct tg3
*tp
)
5083 /* Re-enable indirect register accesses. */
5084 pci_write_config_dword(tp
->pdev
, TG3PCI_MISC_HOST_CTRL
,
5085 tp
->misc_host_ctrl
);
5087 /* Set MAX PCI retry to zero. */
5088 val
= (PCISTATE_ROM_ENABLE
| PCISTATE_ROM_RETRY_ENABLE
);
5089 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5704_A0
&&
5090 (tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
))
5091 val
|= PCISTATE_RETRY_SAME_DMA
;
5092 /* Allow reads and writes to the APE register and memory space. */
5093 if (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
)
5094 val
|= PCISTATE_ALLOW_APE_CTLSPC_WR
|
5095 PCISTATE_ALLOW_APE_SHMEM_WR
;
5096 pci_write_config_dword(tp
->pdev
, TG3PCI_PCISTATE
, val
);
5098 pci_write_config_word(tp
->pdev
, PCI_COMMAND
, tp
->pci_cmd
);
5100 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
)
5101 pcie_set_readrq(tp
->pdev
, 4096);
5103 pci_write_config_byte(tp
->pdev
, PCI_CACHE_LINE_SIZE
,
5104 tp
->pci_cacheline_sz
);
5105 pci_write_config_byte(tp
->pdev
, PCI_LATENCY_TIMER
,
5109 /* Make sure PCI-X relaxed ordering bit is clear. */
5113 pci_read_config_word(tp
->pdev
, tp
->pcix_cap
+ PCI_X_CMD
,
5115 pcix_cmd
&= ~PCI_X_CMD_ERO
;
5116 pci_write_config_word(tp
->pdev
, tp
->pcix_cap
+ PCI_X_CMD
,
5120 if (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
) {
5122 /* Chip reset on 5780 will reset MSI enable bit,
5123 * so need to restore it.
5125 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
5128 pci_read_config_word(tp
->pdev
,
5129 tp
->msi_cap
+ PCI_MSI_FLAGS
,
5131 pci_write_config_word(tp
->pdev
,
5132 tp
->msi_cap
+ PCI_MSI_FLAGS
,
5133 ctrl
| PCI_MSI_FLAGS_ENABLE
);
5134 val
= tr32(MSGINT_MODE
);
5135 tw32(MSGINT_MODE
, val
| MSGINT_MODE_ENABLE
);
5140 static void tg3_stop_fw(struct tg3
*);
5142 /* tp->lock is held. */
5143 static int tg3_chip_reset(struct tg3
*tp
)
5146 void (*write_op
)(struct tg3
*, u32
, u32
);
5151 /* No matching tg3_nvram_unlock() after this because
5152 * chip reset below will undo the nvram lock.
5154 tp
->nvram_lock_cnt
= 0;
5156 /* GRC_MISC_CFG core clock reset will clear the memory
5157 * enable bit in PCI register 4 and the MSI enable bit
5158 * on some chips, so we save relevant registers here.
5160 tg3_save_pci_state(tp
);
5162 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
||
5163 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
5164 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
5165 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
5166 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
5167 tw32(GRC_FASTBOOT_PC
, 0);
5170 * We must avoid the readl() that normally takes place.
5171 * It locks machines, causes machine checks, and other
5172 * fun things. So, temporarily disable the 5701
5173 * hardware workaround, while we do the reset.
5175 write_op
= tp
->write32
;
5176 if (write_op
== tg3_write_flush_reg32
)
5177 tp
->write32
= tg3_write32
;
5179 /* Prevent the irq handler from reading or writing PCI registers
5180 * during chip reset when the memory enable bit in the PCI command
5181 * register may be cleared. The chip does not generate interrupt
5182 * at this time, but the irq handler may still be called due to irq
5183 * sharing or irqpoll.
5185 tp
->tg3_flags
|= TG3_FLAG_CHIP_RESETTING
;
5186 if (tp
->hw_status
) {
5187 tp
->hw_status
->status
= 0;
5188 tp
->hw_status
->status_tag
= 0;
5192 synchronize_irq(tp
->pdev
->irq
);
5195 val
= GRC_MISC_CFG_CORECLK_RESET
;
5197 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
5198 if (tr32(0x7e2c) == 0x60) {
5201 if (tp
->pci_chip_rev_id
!= CHIPREV_ID_5750_A0
) {
5202 tw32(GRC_MISC_CFG
, (1 << 29));
5207 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
5208 tw32(VCPU_STATUS
, tr32(VCPU_STATUS
) | VCPU_STATUS_DRV_RESET
);
5209 tw32(GRC_VCPU_EXT_CTRL
,
5210 tr32(GRC_VCPU_EXT_CTRL
) & ~GRC_VCPU_EXT_CTRL_HALT_CPU
);
5213 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)
5214 val
|= GRC_MISC_CFG_KEEP_GPHY_POWER
;
5215 tw32(GRC_MISC_CFG
, val
);
5217 /* restore 5701 hardware bug workaround write method */
5218 tp
->write32
= write_op
;
5220 /* Unfortunately, we have to delay before the PCI read back.
5221 * Some 575X chips even will not respond to a PCI cfg access
5222 * when the reset command is given to the chip.
5224 * How do these hardware designers expect things to work
5225 * properly if the PCI write is posted for a long period
5226 * of time? It is always necessary to have some method by
5227 * which a register read back can occur to push the write
5228 * out which does the reset.
5230 * For most tg3 variants the trick below was working.
5235 /* Flush PCI posted writes. The normal MMIO registers
5236 * are inaccessible at this time so this is the only
5237 * way to make this reliably (actually, this is no longer
5238 * the case, see above). I tried to use indirect
5239 * register read/write but this upset some 5701 variants.
5241 pci_read_config_dword(tp
->pdev
, PCI_COMMAND
, &val
);
5245 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
5246 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5750_A0
) {
5250 /* Wait for link training to complete. */
5251 for (i
= 0; i
< 5000; i
++)
5254 pci_read_config_dword(tp
->pdev
, 0xc4, &cfg_val
);
5255 pci_write_config_dword(tp
->pdev
, 0xc4,
5256 cfg_val
| (1 << 15));
5258 /* Set PCIE max payload size and clear error status. */
5259 pci_write_config_dword(tp
->pdev
, 0xd8, 0xf5000);
5262 tg3_restore_pci_state(tp
);
5264 tp
->tg3_flags
&= ~TG3_FLAG_CHIP_RESETTING
;
5267 if (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)
5268 val
= tr32(MEMARB_MODE
);
5269 tw32(MEMARB_MODE
, val
| MEMARB_MODE_ENABLE
);
5271 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5750_A3
) {
5273 tw32(0x5000, 0x400);
5276 tw32(GRC_MODE
, tp
->grc_mode
);
5278 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A0
) {
5281 tw32(0xc4, val
| (1 << 15));
5284 if ((tp
->nic_sram_data_cfg
& NIC_SRAM_DATA_CFG_MINI_PCI
) != 0 &&
5285 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) {
5286 tp
->pci_clock_ctrl
|= CLOCK_CTRL_CLKRUN_OENABLE
;
5287 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A0
)
5288 tp
->pci_clock_ctrl
|= CLOCK_CTRL_FORCE_CLKRUN
;
5289 tw32(TG3PCI_CLOCK_CTRL
, tp
->pci_clock_ctrl
);
5292 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
5293 tp
->mac_mode
= MAC_MODE_PORT_MODE_TBI
;
5294 tw32_f(MAC_MODE
, tp
->mac_mode
);
5295 } else if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) {
5296 tp
->mac_mode
= MAC_MODE_PORT_MODE_GMII
;
5297 tw32_f(MAC_MODE
, tp
->mac_mode
);
5299 tw32_f(MAC_MODE
, 0);
5302 err
= tg3_poll_fw(tp
);
5306 if ((tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) &&
5307 tp
->pci_chip_rev_id
!= CHIPREV_ID_5750_A0
) {
5310 tw32(0x7c00, val
| (1 << 25));
5313 /* Reprobe ASF enable state. */
5314 tp
->tg3_flags
&= ~TG3_FLAG_ENABLE_ASF
;
5315 tp
->tg3_flags2
&= ~TG3_FLG2_ASF_NEW_HANDSHAKE
;
5316 tg3_read_mem(tp
, NIC_SRAM_DATA_SIG
, &val
);
5317 if (val
== NIC_SRAM_DATA_SIG_MAGIC
) {
5320 tg3_read_mem(tp
, NIC_SRAM_DATA_CFG
, &nic_cfg
);
5321 if (nic_cfg
& NIC_SRAM_DATA_CFG_ASF_ENABLE
) {
5322 tp
->tg3_flags
|= TG3_FLAG_ENABLE_ASF
;
5323 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
)
5324 tp
->tg3_flags2
|= TG3_FLG2_ASF_NEW_HANDSHAKE
;
5331 /* tp->lock is held. */
5332 static void tg3_stop_fw(struct tg3
*tp
)
5334 if ((tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) &&
5335 !(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
)) {
5339 tg3_write_mem(tp
, NIC_SRAM_FW_CMD_MBOX
, FWCMD_NICDRV_PAUSE_FW
);
5340 val
= tr32(GRC_RX_CPU_EVENT
);
5342 tw32(GRC_RX_CPU_EVENT
, val
);
5344 /* Wait for RX cpu to ACK the event. */
5345 for (i
= 0; i
< 100; i
++) {
5346 if (!(tr32(GRC_RX_CPU_EVENT
) & (1 << 14)))
5353 /* tp->lock is held. */
5354 static int tg3_halt(struct tg3
*tp
, int kind
, int silent
)
5360 tg3_write_sig_pre_reset(tp
, kind
);
5362 tg3_abort_hw(tp
, silent
);
5363 err
= tg3_chip_reset(tp
);
5365 tg3_write_sig_legacy(tp
, kind
);
5366 tg3_write_sig_post_reset(tp
, kind
);
5374 #define TG3_FW_RELEASE_MAJOR 0x0
5375 #define TG3_FW_RELASE_MINOR 0x0
5376 #define TG3_FW_RELEASE_FIX 0x0
5377 #define TG3_FW_START_ADDR 0x08000000
5378 #define TG3_FW_TEXT_ADDR 0x08000000
5379 #define TG3_FW_TEXT_LEN 0x9c0
5380 #define TG3_FW_RODATA_ADDR 0x080009c0
5381 #define TG3_FW_RODATA_LEN 0x60
5382 #define TG3_FW_DATA_ADDR 0x08000a40
5383 #define TG3_FW_DATA_LEN 0x20
5384 #define TG3_FW_SBSS_ADDR 0x08000a60
5385 #define TG3_FW_SBSS_LEN 0xc
5386 #define TG3_FW_BSS_ADDR 0x08000a70
5387 #define TG3_FW_BSS_LEN 0x10
5389 static const u32 tg3FwText
[(TG3_FW_TEXT_LEN
/ sizeof(u32
)) + 1] = {
5390 0x00000000, 0x10000003, 0x00000000, 0x0000000d, 0x0000000d, 0x3c1d0800,
5391 0x37bd3ffc, 0x03a0f021, 0x3c100800, 0x26100000, 0x0e000018, 0x00000000,
5392 0x0000000d, 0x3c1d0800, 0x37bd3ffc, 0x03a0f021, 0x3c100800, 0x26100034,
5393 0x0e00021c, 0x00000000, 0x0000000d, 0x00000000, 0x00000000, 0x00000000,
5394 0x27bdffe0, 0x3c1cc000, 0xafbf0018, 0xaf80680c, 0x0e00004c, 0x241b2105,
5395 0x97850000, 0x97870002, 0x9782002c, 0x9783002e, 0x3c040800, 0x248409c0,
5396 0xafa00014, 0x00021400, 0x00621825, 0x00052c00, 0xafa30010, 0x8f860010,
5397 0x00e52825, 0x0e000060, 0x24070102, 0x3c02ac00, 0x34420100, 0x3c03ac01,
5398 0x34630100, 0xaf820490, 0x3c02ffff, 0xaf820494, 0xaf830498, 0xaf82049c,
5399 0x24020001, 0xaf825ce0, 0x0e00003f, 0xaf825d00, 0x0e000140, 0x00000000,
5400 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x2402ffff, 0xaf825404, 0x8f835400,
5401 0x34630400, 0xaf835400, 0xaf825404, 0x3c020800, 0x24420034, 0xaf82541c,
5402 0x03e00008, 0xaf805400, 0x00000000, 0x00000000, 0x3c020800, 0x34423000,
5403 0x3c030800, 0x34633000, 0x3c040800, 0x348437ff, 0x3c010800, 0xac220a64,
5404 0x24020040, 0x3c010800, 0xac220a68, 0x3c010800, 0xac200a60, 0xac600000,
5405 0x24630004, 0x0083102b, 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000,
5406 0x00804821, 0x8faa0010, 0x3c020800, 0x8c420a60, 0x3c040800, 0x8c840a68,
5407 0x8fab0014, 0x24430001, 0x0044102b, 0x3c010800, 0xac230a60, 0x14400003,
5408 0x00004021, 0x3c010800, 0xac200a60, 0x3c020800, 0x8c420a60, 0x3c030800,
5409 0x8c630a64, 0x91240000, 0x00021140, 0x00431021, 0x00481021, 0x25080001,
5410 0xa0440000, 0x29020008, 0x1440fff4, 0x25290001, 0x3c020800, 0x8c420a60,
5411 0x3c030800, 0x8c630a64, 0x8f84680c, 0x00021140, 0x00431021, 0xac440008,
5412 0xac45000c, 0xac460010, 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c,
5413 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
5414 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
5416 0x02000008, 0x00000000, 0x0a0001e3, 0x3c0a0001, 0x0a0001e3, 0x3c0a0002,
5417 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
5418 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
5419 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
5420 0x0a0001e3, 0x3c0a0007, 0x0a0001e3, 0x3c0a0008, 0x0a0001e3, 0x3c0a0009,
5421 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a000b,
5422 0x0a0001e3, 0x3c0a000c, 0x0a0001e3, 0x3c0a000d, 0x0a0001e3, 0x00000000,
5423 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a000e, 0x0a0001e3, 0x00000000,
5424 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
5425 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x00000000,
5426 0x0a0001e3, 0x00000000, 0x0a0001e3, 0x3c0a0013, 0x0a0001e3, 0x3c0a0014,
5427 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
5428 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
5429 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
5430 0x27bdffe0, 0x00001821, 0x00001021, 0xafbf0018, 0xafb10014, 0xafb00010,
5431 0x3c010800, 0x00220821, 0xac200a70, 0x3c010800, 0x00220821, 0xac200a74,
5432 0x3c010800, 0x00220821, 0xac200a78, 0x24630001, 0x1860fff5, 0x2442000c,
5433 0x24110001, 0x8f906810, 0x32020004, 0x14400005, 0x24040001, 0x3c020800,
5434 0x8c420a78, 0x18400003, 0x00002021, 0x0e000182, 0x00000000, 0x32020001,
5435 0x10400003, 0x00000000, 0x0e000169, 0x00000000, 0x0a000153, 0xaf915028,
5436 0x8fbf0018, 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0020, 0x3c050800,
5437 0x8ca50a70, 0x3c060800, 0x8cc60a80, 0x3c070800, 0x8ce70a78, 0x27bdffe0,
5438 0x3c040800, 0x248409d0, 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014,
5439 0x0e00017b, 0x00002021, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x24020001,
5440 0x8f836810, 0x00821004, 0x00021027, 0x00621824, 0x03e00008, 0xaf836810,
5441 0x27bdffd8, 0xafbf0024, 0x1080002e, 0xafb00020, 0x8f825cec, 0xafa20018,
5442 0x8f825cec, 0x3c100800, 0x26100a78, 0xafa2001c, 0x34028000, 0xaf825cec,
5443 0x8e020000, 0x18400016, 0x00000000, 0x3c020800, 0x94420a74, 0x8fa3001c,
5444 0x000221c0, 0xac830004, 0x8fa2001c, 0x3c010800, 0x0e000201, 0xac220a74,
5445 0x10400005, 0x00000000, 0x8e020000, 0x24420001, 0x0a0001df, 0xae020000,
5446 0x3c020800, 0x8c420a70, 0x00021c02, 0x000321c0, 0x0a0001c5, 0xafa2001c,
5447 0x0e000201, 0x00000000, 0x1040001f, 0x00000000, 0x8e020000, 0x8fa3001c,
5448 0x24420001, 0x3c010800, 0xac230a70, 0x3c010800, 0xac230a74, 0x0a0001df,
5449 0xae020000, 0x3c100800, 0x26100a78, 0x8e020000, 0x18400028, 0x00000000,
5450 0x0e000201, 0x00000000, 0x14400024, 0x00000000, 0x8e020000, 0x3c030800,
5451 0x8c630a70, 0x2442ffff, 0xafa3001c, 0x18400006, 0xae020000, 0x00031402,
5452 0x000221c0, 0x8c820004, 0x3c010800, 0xac220a70, 0x97a2001e, 0x2442ff00,
5453 0x2c420300, 0x1440000b, 0x24024000, 0x3c040800, 0x248409dc, 0xafa00010,
5454 0xafa00014, 0x8fa6001c, 0x24050008, 0x0e000060, 0x00003821, 0x0a0001df,
5455 0x00000000, 0xaf825cf8, 0x3c020800, 0x8c420a40, 0x8fa3001c, 0x24420001,
5456 0xaf835cf8, 0x3c010800, 0xac220a40, 0x8fbf0024, 0x8fb00020, 0x03e00008,
5457 0x27bd0028, 0x27bdffe0, 0x3c040800, 0x248409e8, 0x00002821, 0x00003021,
5458 0x00003821, 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014, 0x8fbf0018,
5459 0x03e00008, 0x27bd0020, 0x8f82680c, 0x8f85680c, 0x00021827, 0x0003182b,
5460 0x00031823, 0x00431024, 0x00441021, 0x00a2282b, 0x10a00006, 0x00000000,
5461 0x00401821, 0x8f82680c, 0x0043102b, 0x1440fffd, 0x00000000, 0x03e00008,
5462 0x00000000, 0x3c040800, 0x8c840000, 0x3c030800, 0x8c630a40, 0x0064102b,
5463 0x54400002, 0x00831023, 0x00641023, 0x2c420008, 0x03e00008, 0x38420001,
5464 0x27bdffe0, 0x00802821, 0x3c040800, 0x24840a00, 0x00003021, 0x00003821,
5465 0xafbf0018, 0xafa00010, 0x0e000060, 0xafa00014, 0x0a000216, 0x00000000,
5466 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x00000000, 0x27bdffe0, 0x3c1cc000,
5467 0xafbf0018, 0x0e00004c, 0xaf80680c, 0x3c040800, 0x24840a10, 0x03802821,
5468 0x00003021, 0x00003821, 0xafa00010, 0x0e000060, 0xafa00014, 0x2402ffff,
5469 0xaf825404, 0x3c0200aa, 0x0e000234, 0xaf825434, 0x8fbf0018, 0x03e00008,
5470 0x27bd0020, 0x00000000, 0x00000000, 0x00000000, 0x27bdffe8, 0xafb00010,
5471 0x24100001, 0xafbf0014, 0x3c01c003, 0xac200000, 0x8f826810, 0x30422000,
5472 0x10400003, 0x00000000, 0x0e000246, 0x00000000, 0x0a00023a, 0xaf905428,
5473 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x27bdfff8, 0x8f845d0c,
5474 0x3c0200ff, 0x3c030800, 0x8c630a50, 0x3442fff8, 0x00821024, 0x1043001e,
5475 0x3c0500ff, 0x34a5fff8, 0x3c06c003, 0x3c074000, 0x00851824, 0x8c620010,
5476 0x3c010800, 0xac230a50, 0x30420008, 0x10400005, 0x00871025, 0x8cc20000,
5477 0x24420001, 0xacc20000, 0x00871025, 0xaf825d0c, 0x8fa20000, 0x24420001,
5478 0xafa20000, 0x8fa20000, 0x8fa20000, 0x24420001, 0xafa20000, 0x8fa20000,
5479 0x8f845d0c, 0x3c030800, 0x8c630a50, 0x00851024, 0x1443ffe8, 0x00851824,
5480 0x27bd0008, 0x03e00008, 0x00000000, 0x00000000, 0x00000000
5483 static const u32 tg3FwRodata
[(TG3_FW_RODATA_LEN
/ sizeof(u32
)) + 1] = {
5484 0x35373031, 0x726c7341, 0x00000000, 0x00000000, 0x53774576, 0x656e7430,
5485 0x00000000, 0x726c7045, 0x76656e74, 0x31000000, 0x556e6b6e, 0x45766e74,
5486 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x66617461, 0x6c457272,
5487 0x00000000, 0x00000000, 0x4d61696e, 0x43707542, 0x00000000, 0x00000000,
5491 #if 0 /* All zeros, don't eat up space with it. */
5492 u32 tg3FwData
[(TG3_FW_DATA_LEN
/ sizeof(u32
)) + 1] = {
5493 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000,
5494 0x00000000, 0x00000000, 0x00000000, 0x00000000
5498 #define RX_CPU_SCRATCH_BASE 0x30000
5499 #define RX_CPU_SCRATCH_SIZE 0x04000
5500 #define TX_CPU_SCRATCH_BASE 0x34000
5501 #define TX_CPU_SCRATCH_SIZE 0x04000
5503 /* tp->lock is held. */
5504 static int tg3_halt_cpu(struct tg3
*tp
, u32 offset
)
5508 BUG_ON(offset
== TX_CPU_BASE
&&
5509 (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
));
5511 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
5512 u32 val
= tr32(GRC_VCPU_EXT_CTRL
);
5514 tw32(GRC_VCPU_EXT_CTRL
, val
| GRC_VCPU_EXT_CTRL_HALT_CPU
);
5517 if (offset
== RX_CPU_BASE
) {
5518 for (i
= 0; i
< 10000; i
++) {
5519 tw32(offset
+ CPU_STATE
, 0xffffffff);
5520 tw32(offset
+ CPU_MODE
, CPU_MODE_HALT
);
5521 if (tr32(offset
+ CPU_MODE
) & CPU_MODE_HALT
)
5525 tw32(offset
+ CPU_STATE
, 0xffffffff);
5526 tw32_f(offset
+ CPU_MODE
, CPU_MODE_HALT
);
5529 for (i
= 0; i
< 10000; i
++) {
5530 tw32(offset
+ CPU_STATE
, 0xffffffff);
5531 tw32(offset
+ CPU_MODE
, CPU_MODE_HALT
);
5532 if (tr32(offset
+ CPU_MODE
) & CPU_MODE_HALT
)
5538 printk(KERN_ERR PFX
"tg3_reset_cpu timed out for %s, "
5541 (offset
== RX_CPU_BASE
? "RX" : "TX"));
5545 /* Clear firmware's nvram arbitration. */
5546 if (tp
->tg3_flags
& TG3_FLAG_NVRAM
)
5547 tw32(NVRAM_SWARB
, SWARB_REQ_CLR0
);
5552 unsigned int text_base
;
5553 unsigned int text_len
;
5554 const u32
*text_data
;
5555 unsigned int rodata_base
;
5556 unsigned int rodata_len
;
5557 const u32
*rodata_data
;
5558 unsigned int data_base
;
5559 unsigned int data_len
;
5560 const u32
*data_data
;
5563 /* tp->lock is held. */
5564 static int tg3_load_firmware_cpu(struct tg3
*tp
, u32 cpu_base
, u32 cpu_scratch_base
,
5565 int cpu_scratch_size
, struct fw_info
*info
)
5567 int err
, lock_err
, i
;
5568 void (*write_op
)(struct tg3
*, u32
, u32
);
5570 if (cpu_base
== TX_CPU_BASE
&&
5571 (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
5572 printk(KERN_ERR PFX
"tg3_load_firmware_cpu: Trying to load "
5573 "TX cpu firmware on %s which is 5705.\n",
5578 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)
5579 write_op
= tg3_write_mem
;
5581 write_op
= tg3_write_indirect_reg32
;
5583 /* It is possible that bootcode is still loading at this point.
5584 * Get the nvram lock first before halting the cpu.
5586 lock_err
= tg3_nvram_lock(tp
);
5587 err
= tg3_halt_cpu(tp
, cpu_base
);
5589 tg3_nvram_unlock(tp
);
5593 for (i
= 0; i
< cpu_scratch_size
; i
+= sizeof(u32
))
5594 write_op(tp
, cpu_scratch_base
+ i
, 0);
5595 tw32(cpu_base
+ CPU_STATE
, 0xffffffff);
5596 tw32(cpu_base
+ CPU_MODE
, tr32(cpu_base
+CPU_MODE
)|CPU_MODE_HALT
);
5597 for (i
= 0; i
< (info
->text_len
/ sizeof(u32
)); i
++)
5598 write_op(tp
, (cpu_scratch_base
+
5599 (info
->text_base
& 0xffff) +
5602 info
->text_data
[i
] : 0));
5603 for (i
= 0; i
< (info
->rodata_len
/ sizeof(u32
)); i
++)
5604 write_op(tp
, (cpu_scratch_base
+
5605 (info
->rodata_base
& 0xffff) +
5607 (info
->rodata_data
?
5608 info
->rodata_data
[i
] : 0));
5609 for (i
= 0; i
< (info
->data_len
/ sizeof(u32
)); i
++)
5610 write_op(tp
, (cpu_scratch_base
+
5611 (info
->data_base
& 0xffff) +
5614 info
->data_data
[i
] : 0));
5622 /* tp->lock is held. */
5623 static int tg3_load_5701_a0_firmware_fix(struct tg3
*tp
)
5625 struct fw_info info
;
5628 info
.text_base
= TG3_FW_TEXT_ADDR
;
5629 info
.text_len
= TG3_FW_TEXT_LEN
;
5630 info
.text_data
= &tg3FwText
[0];
5631 info
.rodata_base
= TG3_FW_RODATA_ADDR
;
5632 info
.rodata_len
= TG3_FW_RODATA_LEN
;
5633 info
.rodata_data
= &tg3FwRodata
[0];
5634 info
.data_base
= TG3_FW_DATA_ADDR
;
5635 info
.data_len
= TG3_FW_DATA_LEN
;
5636 info
.data_data
= NULL
;
5638 err
= tg3_load_firmware_cpu(tp
, RX_CPU_BASE
,
5639 RX_CPU_SCRATCH_BASE
, RX_CPU_SCRATCH_SIZE
,
5644 err
= tg3_load_firmware_cpu(tp
, TX_CPU_BASE
,
5645 TX_CPU_SCRATCH_BASE
, TX_CPU_SCRATCH_SIZE
,
5650 /* Now startup only the RX cpu. */
5651 tw32(RX_CPU_BASE
+ CPU_STATE
, 0xffffffff);
5652 tw32_f(RX_CPU_BASE
+ CPU_PC
, TG3_FW_TEXT_ADDR
);
5654 for (i
= 0; i
< 5; i
++) {
5655 if (tr32(RX_CPU_BASE
+ CPU_PC
) == TG3_FW_TEXT_ADDR
)
5657 tw32(RX_CPU_BASE
+ CPU_STATE
, 0xffffffff);
5658 tw32(RX_CPU_BASE
+ CPU_MODE
, CPU_MODE_HALT
);
5659 tw32_f(RX_CPU_BASE
+ CPU_PC
, TG3_FW_TEXT_ADDR
);
5663 printk(KERN_ERR PFX
"tg3_load_firmware fails for %s "
5664 "to set RX CPU PC, is %08x should be %08x\n",
5665 tp
->dev
->name
, tr32(RX_CPU_BASE
+ CPU_PC
),
5669 tw32(RX_CPU_BASE
+ CPU_STATE
, 0xffffffff);
5670 tw32_f(RX_CPU_BASE
+ CPU_MODE
, 0x00000000);
5676 #define TG3_TSO_FW_RELEASE_MAJOR 0x1
5677 #define TG3_TSO_FW_RELASE_MINOR 0x6
5678 #define TG3_TSO_FW_RELEASE_FIX 0x0
5679 #define TG3_TSO_FW_START_ADDR 0x08000000
5680 #define TG3_TSO_FW_TEXT_ADDR 0x08000000
5681 #define TG3_TSO_FW_TEXT_LEN 0x1aa0
5682 #define TG3_TSO_FW_RODATA_ADDR 0x08001aa0
5683 #define TG3_TSO_FW_RODATA_LEN 0x60
5684 #define TG3_TSO_FW_DATA_ADDR 0x08001b20
5685 #define TG3_TSO_FW_DATA_LEN 0x30
5686 #define TG3_TSO_FW_SBSS_ADDR 0x08001b50
5687 #define TG3_TSO_FW_SBSS_LEN 0x2c
5688 #define TG3_TSO_FW_BSS_ADDR 0x08001b80
5689 #define TG3_TSO_FW_BSS_LEN 0x894
5691 static const u32 tg3TsoFwText
[(TG3_TSO_FW_TEXT_LEN
/ 4) + 1] = {
5692 0x0e000003, 0x00000000, 0x08001b24, 0x00000000, 0x10000003, 0x00000000,
5693 0x0000000d, 0x0000000d, 0x3c1d0800, 0x37bd4000, 0x03a0f021, 0x3c100800,
5694 0x26100000, 0x0e000010, 0x00000000, 0x0000000d, 0x27bdffe0, 0x3c04fefe,
5695 0xafbf0018, 0x0e0005d8, 0x34840002, 0x0e000668, 0x00000000, 0x3c030800,
5696 0x90631b68, 0x24020002, 0x3c040800, 0x24841aac, 0x14620003, 0x24050001,
5697 0x3c040800, 0x24841aa0, 0x24060006, 0x00003821, 0xafa00010, 0x0e00067c,
5698 0xafa00014, 0x8f625c50, 0x34420001, 0xaf625c50, 0x8f625c90, 0x34420001,
5699 0xaf625c90, 0x2402ffff, 0x0e000034, 0xaf625404, 0x8fbf0018, 0x03e00008,
5700 0x27bd0020, 0x00000000, 0x00000000, 0x00000000, 0x27bdffe0, 0xafbf001c,
5701 0xafb20018, 0xafb10014, 0x0e00005b, 0xafb00010, 0x24120002, 0x24110001,
5702 0x8f706820, 0x32020100, 0x10400003, 0x00000000, 0x0e0000bb, 0x00000000,
5703 0x8f706820, 0x32022000, 0x10400004, 0x32020001, 0x0e0001f0, 0x24040001,
5704 0x32020001, 0x10400003, 0x00000000, 0x0e0000a3, 0x00000000, 0x3c020800,
5705 0x90421b98, 0x14520003, 0x00000000, 0x0e0004c0, 0x00000000, 0x0a00003c,
5706 0xaf715028, 0x8fbf001c, 0x8fb20018, 0x8fb10014, 0x8fb00010, 0x03e00008,
5707 0x27bd0020, 0x27bdffe0, 0x3c040800, 0x24841ac0, 0x00002821, 0x00003021,
5708 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014, 0x3c040800,
5709 0x248423d8, 0xa4800000, 0x3c010800, 0xa0201b98, 0x3c010800, 0xac201b9c,
5710 0x3c010800, 0xac201ba0, 0x3c010800, 0xac201ba4, 0x3c010800, 0xac201bac,
5711 0x3c010800, 0xac201bb8, 0x3c010800, 0xac201bbc, 0x8f624434, 0x3c010800,
5712 0xac221b88, 0x8f624438, 0x3c010800, 0xac221b8c, 0x8f624410, 0xac80f7a8,
5713 0x3c010800, 0xac201b84, 0x3c010800, 0xac2023e0, 0x3c010800, 0xac2023c8,
5714 0x3c010800, 0xac2023cc, 0x3c010800, 0xac202400, 0x3c010800, 0xac221b90,
5715 0x8f620068, 0x24030007, 0x00021702, 0x10430005, 0x00000000, 0x8f620068,
5716 0x00021702, 0x14400004, 0x24020001, 0x3c010800, 0x0a000097, 0xac20240c,
5717 0xac820034, 0x3c040800, 0x24841acc, 0x3c050800, 0x8ca5240c, 0x00003021,
5718 0x00003821, 0xafa00010, 0x0e00067c, 0xafa00014, 0x8fbf0018, 0x03e00008,
5719 0x27bd0020, 0x27bdffe0, 0x3c040800, 0x24841ad8, 0x00002821, 0x00003021,
5720 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014, 0x0e00005b,
5721 0x00000000, 0x0e0000b4, 0x00002021, 0x8fbf0018, 0x03e00008, 0x27bd0020,
5722 0x24020001, 0x8f636820, 0x00821004, 0x00021027, 0x00621824, 0x03e00008,
5723 0xaf636820, 0x27bdffd0, 0xafbf002c, 0xafb60028, 0xafb50024, 0xafb40020,
5724 0xafb3001c, 0xafb20018, 0xafb10014, 0xafb00010, 0x8f675c5c, 0x3c030800,
5725 0x24631bbc, 0x8c620000, 0x14470005, 0x3c0200ff, 0x3c020800, 0x90421b98,
5726 0x14400119, 0x3c0200ff, 0x3442fff8, 0x00e28824, 0xac670000, 0x00111902,
5727 0x306300ff, 0x30e20003, 0x000211c0, 0x00622825, 0x00a04021, 0x00071602,
5728 0x3c030800, 0x90631b98, 0x3044000f, 0x14600036, 0x00804821, 0x24020001,
5729 0x3c010800, 0xa0221b98, 0x00051100, 0x00821025, 0x3c010800, 0xac201b9c,
5730 0x3c010800, 0xac201ba0, 0x3c010800, 0xac201ba4, 0x3c010800, 0xac201bac,
5731 0x3c010800, 0xac201bb8, 0x3c010800, 0xac201bb0, 0x3c010800, 0xac201bb4,
5732 0x3c010800, 0xa42223d8, 0x9622000c, 0x30437fff, 0x3c010800, 0xa4222410,
5733 0x30428000, 0x3c010800, 0xa4231bc6, 0x10400005, 0x24020001, 0x3c010800,
5734 0xac2223f4, 0x0a000102, 0x2406003e, 0x24060036, 0x3c010800, 0xac2023f4,
5735 0x9622000a, 0x3c030800, 0x94631bc6, 0x3c010800, 0xac2023f0, 0x3c010800,
5736 0xac2023f8, 0x00021302, 0x00021080, 0x00c21021, 0x00621821, 0x3c010800,
5737 0xa42223d0, 0x3c010800, 0x0a000115, 0xa4231b96, 0x9622000c, 0x3c010800,
5738 0xa42223ec, 0x3c040800, 0x24841b9c, 0x8c820000, 0x00021100, 0x3c010800,
5739 0x00220821, 0xac311bc8, 0x8c820000, 0x00021100, 0x3c010800, 0x00220821,
5740 0xac271bcc, 0x8c820000, 0x25030001, 0x306601ff, 0x00021100, 0x3c010800,
5741 0x00220821, 0xac261bd0, 0x8c820000, 0x00021100, 0x3c010800, 0x00220821,
5742 0xac291bd4, 0x96230008, 0x3c020800, 0x8c421bac, 0x00432821, 0x3c010800,
5743 0xac251bac, 0x9622000a, 0x30420004, 0x14400018, 0x00061100, 0x8f630c14,
5744 0x3063000f, 0x2c620002, 0x1440000b, 0x3c02c000, 0x8f630c14, 0x3c020800,
5745 0x8c421b40, 0x3063000f, 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002,
5746 0x1040fff7, 0x3c02c000, 0x00e21825, 0xaf635c5c, 0x8f625c50, 0x30420002,
5747 0x10400014, 0x00000000, 0x0a000147, 0x00000000, 0x3c030800, 0x8c631b80,
5748 0x3c040800, 0x94841b94, 0x01221025, 0x3c010800, 0xa42223da, 0x24020001,
5749 0x3c010800, 0xac221bb8, 0x24630001, 0x0085202a, 0x3c010800, 0x10800003,
5750 0xac231b80, 0x3c010800, 0xa4251b94, 0x3c060800, 0x24c61b9c, 0x8cc20000,
5751 0x24420001, 0xacc20000, 0x28420080, 0x14400005, 0x00000000, 0x0e000656,
5752 0x24040002, 0x0a0001e6, 0x00000000, 0x3c020800, 0x8c421bb8, 0x10400078,
5753 0x24020001, 0x3c050800, 0x90a51b98, 0x14a20072, 0x00000000, 0x3c150800,
5754 0x96b51b96, 0x3c040800, 0x8c841bac, 0x32a3ffff, 0x0083102a, 0x1440006c,
5755 0x00000000, 0x14830003, 0x00000000, 0x3c010800, 0xac2523f0, 0x1060005c,
5756 0x00009021, 0x24d60004, 0x0060a021, 0x24d30014, 0x8ec20000, 0x00028100,
5757 0x3c110800, 0x02308821, 0x0e000625, 0x8e311bc8, 0x00402821, 0x10a00054,
5758 0x00000000, 0x9628000a, 0x31020040, 0x10400005, 0x2407180c, 0x8e22000c,
5759 0x2407188c, 0x00021400, 0xaca20018, 0x3c030800, 0x00701821, 0x8c631bd0,
5760 0x3c020800, 0x00501021, 0x8c421bd4, 0x00031d00, 0x00021400, 0x00621825,
5761 0xaca30014, 0x8ec30004, 0x96220008, 0x00432023, 0x3242ffff, 0x3083ffff,
5762 0x00431021, 0x0282102a, 0x14400002, 0x02b23023, 0x00803021, 0x8e620000,
5763 0x30c4ffff, 0x00441021, 0xae620000, 0x8e220000, 0xaca20000, 0x8e220004,
5764 0x8e63fff4, 0x00431021, 0xaca20004, 0xa4a6000e, 0x8e62fff4, 0x00441021,
5765 0xae62fff4, 0x96230008, 0x0043102a, 0x14400005, 0x02469021, 0x8e62fff0,
5766 0xae60fff4, 0x24420001, 0xae62fff0, 0xaca00008, 0x3242ffff, 0x14540008,
5767 0x24020305, 0x31020080, 0x54400001, 0x34e70010, 0x24020905, 0xa4a2000c,
5768 0x0a0001cb, 0x34e70020, 0xa4a2000c, 0x3c020800, 0x8c4223f0, 0x10400003,
5769 0x3c024b65, 0x0a0001d3, 0x34427654, 0x3c02b49a, 0x344289ab, 0xaca2001c,
5770 0x30e2ffff, 0xaca20010, 0x0e0005a2, 0x00a02021, 0x3242ffff, 0x0054102b,
5771 0x1440ffa9, 0x00000000, 0x24020002, 0x3c010800, 0x0a0001e6, 0xa0221b98,
5772 0x8ec2083c, 0x24420001, 0x0a0001e6, 0xaec2083c, 0x0e0004c0, 0x00000000,
5773 0x8fbf002c, 0x8fb60028, 0x8fb50024, 0x8fb40020, 0x8fb3001c, 0x8fb20018,
5774 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0030, 0x27bdffd0, 0xafbf0028,
5775 0xafb30024, 0xafb20020, 0xafb1001c, 0xafb00018, 0x8f725c9c, 0x3c0200ff,
5776 0x3442fff8, 0x3c070800, 0x24e71bb4, 0x02428824, 0x9623000e, 0x8ce20000,
5777 0x00431021, 0xace20000, 0x8e220010, 0x30420020, 0x14400011, 0x00809821,
5778 0x0e00063b, 0x02202021, 0x3c02c000, 0x02421825, 0xaf635c9c, 0x8f625c90,
5779 0x30420002, 0x1040011e, 0x00000000, 0xaf635c9c, 0x8f625c90, 0x30420002,
5780 0x10400119, 0x00000000, 0x0a00020d, 0x00000000, 0x8e240008, 0x8e230014,
5781 0x00041402, 0x000231c0, 0x00031502, 0x304201ff, 0x2442ffff, 0x3042007f,
5782 0x00031942, 0x30637800, 0x00021100, 0x24424000, 0x00624821, 0x9522000a,
5783 0x3084ffff, 0x30420008, 0x104000b0, 0x000429c0, 0x3c020800, 0x8c422400,
5784 0x14400024, 0x24c50008, 0x94c20014, 0x3c010800, 0xa42223d0, 0x8cc40010,
5785 0x00041402, 0x3c010800, 0xa42223d2, 0x3c010800, 0xa42423d4, 0x94c2000e,
5786 0x3083ffff, 0x00431023, 0x3c010800, 0xac222408, 0x94c2001a, 0x3c010800,
5787 0xac262400, 0x3c010800, 0xac322404, 0x3c010800, 0xac2223fc, 0x3c02c000,
5788 0x02421825, 0xaf635c9c, 0x8f625c90, 0x30420002, 0x104000e5, 0x00000000,
5789 0xaf635c9c, 0x8f625c90, 0x30420002, 0x104000e0, 0x00000000, 0x0a000246,
5790 0x00000000, 0x94c2000e, 0x3c030800, 0x946323d4, 0x00434023, 0x3103ffff,
5791 0x2c620008, 0x1040001c, 0x00000000, 0x94c20014, 0x24420028, 0x00a22821,
5792 0x00031042, 0x1840000b, 0x00002021, 0x24e60848, 0x00403821, 0x94a30000,
5793 0x8cc20000, 0x24840001, 0x00431021, 0xacc20000, 0x0087102a, 0x1440fff9,
5794 0x24a50002, 0x31020001, 0x1040001f, 0x3c024000, 0x3c040800, 0x248423fc,
5795 0xa0a00001, 0x94a30000, 0x8c820000, 0x00431021, 0x0a000285, 0xac820000,
5796 0x8f626800, 0x3c030010, 0x00431024, 0x10400009, 0x00000000, 0x94c2001a,
5797 0x3c030800, 0x8c6323fc, 0x00431021, 0x3c010800, 0xac2223fc, 0x0a000286,
5798 0x3c024000, 0x94c2001a, 0x94c4001c, 0x3c030800, 0x8c6323fc, 0x00441023,
5799 0x00621821, 0x3c010800, 0xac2323fc, 0x3c024000, 0x02421825, 0xaf635c9c,
5800 0x8f625c90, 0x30420002, 0x1440fffc, 0x00000000, 0x9522000a, 0x30420010,
5801 0x1040009b, 0x00000000, 0x3c030800, 0x946323d4, 0x3c070800, 0x24e72400,
5802 0x8ce40000, 0x8f626800, 0x24630030, 0x00832821, 0x3c030010, 0x00431024,
5803 0x1440000a, 0x00000000, 0x94a20004, 0x3c040800, 0x8c842408, 0x3c030800,
5804 0x8c6323fc, 0x00441023, 0x00621821, 0x3c010800, 0xac2323fc, 0x3c040800,
5805 0x8c8423fc, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402, 0x00822021,
5806 0x00041027, 0xa4a20006, 0x3c030800, 0x8c632404, 0x3c0200ff, 0x3442fff8,
5807 0x00628824, 0x96220008, 0x24050001, 0x24034000, 0x000231c0, 0x00801021,
5808 0xa4c2001a, 0xa4c0001c, 0xace00000, 0x3c010800, 0xac251b60, 0xaf635cb8,
5809 0x8f625cb0, 0x30420002, 0x10400003, 0x00000000, 0x3c010800, 0xac201b60,
5810 0x8e220008, 0xaf625cb8, 0x8f625cb0, 0x30420002, 0x10400003, 0x00000000,
5811 0x3c010800, 0xac201b60, 0x3c020800, 0x8c421b60, 0x1040ffec, 0x00000000,
5812 0x3c040800, 0x0e00063b, 0x8c842404, 0x0a00032a, 0x00000000, 0x3c030800,
5813 0x90631b98, 0x24020002, 0x14620003, 0x3c034b65, 0x0a0002e1, 0x00008021,
5814 0x8e22001c, 0x34637654, 0x10430002, 0x24100002, 0x24100001, 0x00c02021,
5815 0x0e000350, 0x02003021, 0x24020003, 0x3c010800, 0xa0221b98, 0x24020002,
5816 0x1202000a, 0x24020001, 0x3c030800, 0x8c6323f0, 0x10620006, 0x00000000,
5817 0x3c020800, 0x944223d8, 0x00021400, 0x0a00031f, 0xae220014, 0x3c040800,
5818 0x248423da, 0x94820000, 0x00021400, 0xae220014, 0x3c020800, 0x8c421bbc,
5819 0x3c03c000, 0x3c010800, 0xa0201b98, 0x00431025, 0xaf625c5c, 0x8f625c50,
5820 0x30420002, 0x10400009, 0x00000000, 0x2484f7e2, 0x8c820000, 0x00431025,
5821 0xaf625c5c, 0x8f625c50, 0x30420002, 0x1440fffa, 0x00000000, 0x3c020800,
5822 0x24421b84, 0x8c430000, 0x24630001, 0xac430000, 0x8f630c14, 0x3063000f,
5823 0x2c620002, 0x1440000c, 0x3c024000, 0x8f630c14, 0x3c020800, 0x8c421b40,
5824 0x3063000f, 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002, 0x1040fff7,
5825 0x00000000, 0x3c024000, 0x02421825, 0xaf635c9c, 0x8f625c90, 0x30420002,
5826 0x1440fffc, 0x00000000, 0x12600003, 0x00000000, 0x0e0004c0, 0x00000000,
5827 0x8fbf0028, 0x8fb30024, 0x8fb20020, 0x8fb1001c, 0x8fb00018, 0x03e00008,
5828 0x27bd0030, 0x8f634450, 0x3c040800, 0x24841b88, 0x8c820000, 0x00031c02,
5829 0x0043102b, 0x14400007, 0x3c038000, 0x8c840004, 0x8f624450, 0x00021c02,
5830 0x0083102b, 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024,
5831 0x1440fffd, 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3c024000,
5832 0x00822025, 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00000000,
5833 0x03e00008, 0x00000000, 0x27bdffe0, 0x00805821, 0x14c00011, 0x256e0008,
5834 0x3c020800, 0x8c4223f4, 0x10400007, 0x24020016, 0x3c010800, 0xa42223d2,
5835 0x2402002a, 0x3c010800, 0x0a000364, 0xa42223d4, 0x8d670010, 0x00071402,
5836 0x3c010800, 0xa42223d2, 0x3c010800, 0xa42723d4, 0x3c040800, 0x948423d4,
5837 0x3c030800, 0x946323d2, 0x95cf0006, 0x3c020800, 0x944223d0, 0x00832023,
5838 0x01e2c023, 0x3065ffff, 0x24a20028, 0x01c24821, 0x3082ffff, 0x14c0001a,
5839 0x01226021, 0x9582000c, 0x3042003f, 0x3c010800, 0xa42223d6, 0x95820004,
5840 0x95830006, 0x3c010800, 0xac2023e4, 0x3c010800, 0xac2023e8, 0x00021400,
5841 0x00431025, 0x3c010800, 0xac221bc0, 0x95220004, 0x3c010800, 0xa4221bc4,
5842 0x95230002, 0x01e51023, 0x0043102a, 0x10400010, 0x24020001, 0x3c010800,
5843 0x0a000398, 0xac2223f8, 0x3c030800, 0x8c6323e8, 0x3c020800, 0x94421bc4,
5844 0x00431021, 0xa5220004, 0x3c020800, 0x94421bc0, 0xa5820004, 0x3c020800,
5845 0x8c421bc0, 0xa5820006, 0x3c020800, 0x8c4223f0, 0x3c0d0800, 0x8dad23e4,
5846 0x3c0a0800, 0x144000e5, 0x8d4a23e8, 0x3c020800, 0x94421bc4, 0x004a1821,
5847 0x3063ffff, 0x0062182b, 0x24020002, 0x10c2000d, 0x01435023, 0x3c020800,
5848 0x944223d6, 0x30420009, 0x10400008, 0x00000000, 0x9582000c, 0x3042fff6,
5849 0xa582000c, 0x3c020800, 0x944223d6, 0x30420009, 0x01a26823, 0x3c020800,
5850 0x8c4223f8, 0x1040004a, 0x01203821, 0x3c020800, 0x944223d2, 0x00004021,
5851 0xa520000a, 0x01e21023, 0xa5220002, 0x3082ffff, 0x00021042, 0x18400008,
5852 0x00003021, 0x00401821, 0x94e20000, 0x25080001, 0x00c23021, 0x0103102a,
5853 0x1440fffb, 0x24e70002, 0x00061c02, 0x30c2ffff, 0x00623021, 0x00061402,
5854 0x00c23021, 0x00c02821, 0x00061027, 0xa522000a, 0x00003021, 0x2527000c,
5855 0x00004021, 0x94e20000, 0x25080001, 0x00c23021, 0x2d020004, 0x1440fffb,
5856 0x24e70002, 0x95220002, 0x00004021, 0x91230009, 0x00442023, 0x01803821,
5857 0x3082ffff, 0xa4e00010, 0x00621821, 0x00021042, 0x18400010, 0x00c33021,
5858 0x00404821, 0x94e20000, 0x24e70002, 0x00c23021, 0x30e2007f, 0x14400006,
5859 0x25080001, 0x8d630000, 0x3c02007f, 0x3442ff80, 0x00625824, 0x25670008,
5860 0x0109102a, 0x1440fff3, 0x00000000, 0x30820001, 0x10400005, 0x00061c02,
5861 0xa0e00001, 0x94e20000, 0x00c23021, 0x00061c02, 0x30c2ffff, 0x00623021,
5862 0x00061402, 0x00c23021, 0x0a00047d, 0x30c6ffff, 0x24020002, 0x14c20081,
5863 0x00000000, 0x3c020800, 0x8c42240c, 0x14400007, 0x00000000, 0x3c020800,
5864 0x944223d2, 0x95230002, 0x01e21023, 0x10620077, 0x00000000, 0x3c020800,
5865 0x944223d2, 0x01e21023, 0xa5220002, 0x3c020800, 0x8c42240c, 0x1040001a,
5866 0x31e3ffff, 0x8dc70010, 0x3c020800, 0x94421b96, 0x00e04021, 0x00072c02,
5867 0x00aa2021, 0x00431023, 0x00823823, 0x00072402, 0x30e2ffff, 0x00823821,
5868 0x00071027, 0xa522000a, 0x3102ffff, 0x3c040800, 0x948423d4, 0x00453023,
5869 0x00e02821, 0x00641823, 0x006d1821, 0x00c33021, 0x00061c02, 0x30c2ffff,
5870 0x0a00047d, 0x00623021, 0x01203821, 0x00004021, 0x3082ffff, 0x00021042,
5871 0x18400008, 0x00003021, 0x00401821, 0x94e20000, 0x25080001, 0x00c23021,
5872 0x0103102a, 0x1440fffb, 0x24e70002, 0x00061c02, 0x30c2ffff, 0x00623021,
5873 0x00061402, 0x00c23021, 0x00c02821, 0x00061027, 0xa522000a, 0x00003021,
5874 0x2527000c, 0x00004021, 0x94e20000, 0x25080001, 0x00c23021, 0x2d020004,
5875 0x1440fffb, 0x24e70002, 0x95220002, 0x00004021, 0x91230009, 0x00442023,
5876 0x01803821, 0x3082ffff, 0xa4e00010, 0x3c040800, 0x948423d4, 0x00621821,
5877 0x00c33021, 0x00061c02, 0x30c2ffff, 0x00623021, 0x00061c02, 0x3c020800,
5878 0x944223d0, 0x00c34821, 0x00441023, 0x00021fc2, 0x00431021, 0x00021043,
5879 0x18400010, 0x00003021, 0x00402021, 0x94e20000, 0x24e70002, 0x00c23021,
5880 0x30e2007f, 0x14400006, 0x25080001, 0x8d630000, 0x3c02007f, 0x3442ff80,
5881 0x00625824, 0x25670008, 0x0104102a, 0x1440fff3, 0x00000000, 0x3c020800,
5882 0x944223ec, 0x00c23021, 0x3122ffff, 0x00c23021, 0x00061c02, 0x30c2ffff,
5883 0x00623021, 0x00061402, 0x00c23021, 0x00c04021, 0x00061027, 0xa5820010,
5884 0xadc00014, 0x0a00049d, 0xadc00000, 0x8dc70010, 0x00e04021, 0x11400007,
5885 0x00072c02, 0x00aa3021, 0x00061402, 0x30c3ffff, 0x00433021, 0x00061402,
5886 0x00c22821, 0x00051027, 0xa522000a, 0x3c030800, 0x946323d4, 0x3102ffff,
5887 0x01e21021, 0x00433023, 0x00cd3021, 0x00061c02, 0x30c2ffff, 0x00623021,
5888 0x00061402, 0x00c23021, 0x00c04021, 0x00061027, 0xa5820010, 0x3102ffff,
5889 0x00051c00, 0x00431025, 0xadc20010, 0x3c020800, 0x8c4223f4, 0x10400005,
5890 0x2de205eb, 0x14400002, 0x25e2fff2, 0x34028870, 0xa5c20034, 0x3c030800,
5891 0x246323e8, 0x8c620000, 0x24420001, 0xac620000, 0x3c040800, 0x8c8423e4,
5892 0x3c020800, 0x8c421bc0, 0x3303ffff, 0x00832021, 0x00431821, 0x0062102b,
5893 0x3c010800, 0xac2423e4, 0x10400003, 0x2482ffff, 0x3c010800, 0xac2223e4,
5894 0x3c010800, 0xac231bc0, 0x03e00008, 0x27bd0020, 0x27bdffb8, 0x3c050800,
5895 0x24a51b96, 0xafbf0044, 0xafbe0040, 0xafb7003c, 0xafb60038, 0xafb50034,
5896 0xafb40030, 0xafb3002c, 0xafb20028, 0xafb10024, 0xafb00020, 0x94a90000,
5897 0x3c020800, 0x944223d0, 0x3c030800, 0x8c631bb0, 0x3c040800, 0x8c841bac,
5898 0x01221023, 0x0064182a, 0xa7a9001e, 0x106000be, 0xa7a20016, 0x24be0022,
5899 0x97b6001e, 0x24b3001a, 0x24b70016, 0x8fc20000, 0x14400008, 0x00000000,
5900 0x8fc2fff8, 0x97a30016, 0x8fc4fff4, 0x00431021, 0x0082202a, 0x148000b0,
5901 0x00000000, 0x97d50818, 0x32a2ffff, 0x104000a3, 0x00009021, 0x0040a021,
5902 0x00008821, 0x0e000625, 0x00000000, 0x00403021, 0x14c00007, 0x00000000,
5903 0x3c020800, 0x8c4223dc, 0x24420001, 0x3c010800, 0x0a000596, 0xac2223dc,
5904 0x3c100800, 0x02118021, 0x8e101bc8, 0x9608000a, 0x31020040, 0x10400005,
5905 0x2407180c, 0x8e02000c, 0x2407188c, 0x00021400, 0xacc20018, 0x31020080,
5906 0x54400001, 0x34e70010, 0x3c020800, 0x00511021, 0x8c421bd0, 0x3c030800,
5907 0x00711821, 0x8c631bd4, 0x00021500, 0x00031c00, 0x00431025, 0xacc20014,
5908 0x96040008, 0x3242ffff, 0x00821021, 0x0282102a, 0x14400002, 0x02b22823,
5909 0x00802821, 0x8e020000, 0x02459021, 0xacc20000, 0x8e020004, 0x00c02021,
5910 0x26310010, 0xac820004, 0x30e2ffff, 0xac800008, 0xa485000e, 0xac820010,
5911 0x24020305, 0x0e0005a2, 0xa482000c, 0x3242ffff, 0x0054102b, 0x1440ffc5,
5912 0x3242ffff, 0x0a00058e, 0x00000000, 0x8e620000, 0x8e63fffc, 0x0043102a,
5913 0x10400067, 0x00000000, 0x8e62fff0, 0x00028900, 0x3c100800, 0x02118021,
5914 0x0e000625, 0x8e101bc8, 0x00403021, 0x14c00005, 0x00000000, 0x8e62082c,
5915 0x24420001, 0x0a000596, 0xae62082c, 0x9608000a, 0x31020040, 0x10400005,
5916 0x2407180c, 0x8e02000c, 0x2407188c, 0x00021400, 0xacc20018, 0x3c020800,
5917 0x00511021, 0x8c421bd0, 0x3c030800, 0x00711821, 0x8c631bd4, 0x00021500,
5918 0x00031c00, 0x00431025, 0xacc20014, 0x8e63fff4, 0x96020008, 0x00432023,
5919 0x3242ffff, 0x3083ffff, 0x00431021, 0x02c2102a, 0x10400003, 0x00802821,
5920 0x97a9001e, 0x01322823, 0x8e620000, 0x30a4ffff, 0x00441021, 0xae620000,
5921 0xa4c5000e, 0x8e020000, 0xacc20000, 0x8e020004, 0x8e63fff4, 0x00431021,
5922 0xacc20004, 0x8e63fff4, 0x96020008, 0x00641821, 0x0062102a, 0x14400006,
5923 0x02459021, 0x8e62fff0, 0xae60fff4, 0x24420001, 0x0a000571, 0xae62fff0,
5924 0xae63fff4, 0xacc00008, 0x3242ffff, 0x10560003, 0x31020004, 0x10400006,
5925 0x24020305, 0x31020080, 0x54400001, 0x34e70010, 0x34e70020, 0x24020905,
5926 0xa4c2000c, 0x8ee30000, 0x8ee20004, 0x14620007, 0x3c02b49a, 0x8ee20860,
5927 0x54400001, 0x34e70400, 0x3c024b65, 0x0a000588, 0x34427654, 0x344289ab,
5928 0xacc2001c, 0x30e2ffff, 0xacc20010, 0x0e0005a2, 0x00c02021, 0x3242ffff,
5929 0x0056102b, 0x1440ff9b, 0x00000000, 0x8e620000, 0x8e63fffc, 0x0043102a,
5930 0x1440ff48, 0x00000000, 0x8fbf0044, 0x8fbe0040, 0x8fb7003c, 0x8fb60038,
5931 0x8fb50034, 0x8fb40030, 0x8fb3002c, 0x8fb20028, 0x8fb10024, 0x8fb00020,
5932 0x03e00008, 0x27bd0048, 0x27bdffe8, 0xafbf0014, 0xafb00010, 0x8f624450,
5933 0x8f634410, 0x0a0005b1, 0x00808021, 0x8f626820, 0x30422000, 0x10400003,
5934 0x00000000, 0x0e0001f0, 0x00002021, 0x8f624450, 0x8f634410, 0x3042ffff,
5935 0x0043102b, 0x1440fff5, 0x00000000, 0x8f630c14, 0x3063000f, 0x2c620002,
5936 0x1440000b, 0x00000000, 0x8f630c14, 0x3c020800, 0x8c421b40, 0x3063000f,
5937 0x24420001, 0x3c010800, 0xac221b40, 0x2c620002, 0x1040fff7, 0x00000000,
5938 0xaf705c18, 0x8f625c10, 0x30420002, 0x10400009, 0x00000000, 0x8f626820,
5939 0x30422000, 0x1040fff8, 0x00000000, 0x0e0001f0, 0x00002021, 0x0a0005c4,
5940 0x00000000, 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x00000000,
5941 0x00000000, 0x00000000, 0x27bdffe8, 0x3c1bc000, 0xafbf0014, 0xafb00010,
5942 0xaf60680c, 0x8f626804, 0x34420082, 0xaf626804, 0x8f634000, 0x24020b50,
5943 0x3c010800, 0xac221b54, 0x24020b78, 0x3c010800, 0xac221b64, 0x34630002,
5944 0xaf634000, 0x0e000605, 0x00808021, 0x3c010800, 0xa0221b68, 0x304200ff,
5945 0x24030002, 0x14430005, 0x00000000, 0x3c020800, 0x8c421b54, 0x0a0005f8,
5946 0xac5000c0, 0x3c020800, 0x8c421b54, 0xac5000bc, 0x8f624434, 0x8f634438,
5947 0x8f644410, 0x3c010800, 0xac221b5c, 0x3c010800, 0xac231b6c, 0x3c010800,
5948 0xac241b58, 0x8fbf0014, 0x8fb00010, 0x03e00008, 0x27bd0018, 0x3c040800,
5949 0x8c870000, 0x3c03aa55, 0x3463aa55, 0x3c06c003, 0xac830000, 0x8cc20000,
5950 0x14430007, 0x24050002, 0x3c0355aa, 0x346355aa, 0xac830000, 0x8cc20000,
5951 0x50430001, 0x24050001, 0x3c020800, 0xac470000, 0x03e00008, 0x00a01021,
5952 0x27bdfff8, 0x18800009, 0x00002821, 0x8f63680c, 0x8f62680c, 0x1043fffe,
5953 0x00000000, 0x24a50001, 0x00a4102a, 0x1440fff9, 0x00000000, 0x03e00008,
5954 0x27bd0008, 0x8f634450, 0x3c020800, 0x8c421b5c, 0x00031c02, 0x0043102b,
5955 0x14400008, 0x3c038000, 0x3c040800, 0x8c841b6c, 0x8f624450, 0x00021c02,
5956 0x0083102b, 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024,
5957 0x1440fffd, 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3082ffff,
5958 0x2442e000, 0x2c422001, 0x14400003, 0x3c024000, 0x0a000648, 0x2402ffff,
5959 0x00822025, 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00001021,
5960 0x03e00008, 0x00000000, 0x8f624450, 0x3c030800, 0x8c631b58, 0x0a000651,
5961 0x3042ffff, 0x8f624450, 0x3042ffff, 0x0043102b, 0x1440fffc, 0x00000000,
5962 0x03e00008, 0x00000000, 0x27bdffe0, 0x00802821, 0x3c040800, 0x24841af0,
5963 0x00003021, 0x00003821, 0xafbf0018, 0xafa00010, 0x0e00067c, 0xafa00014,
5964 0x0a000660, 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x00000000,
5965 0x00000000, 0x00000000, 0x3c020800, 0x34423000, 0x3c030800, 0x34633000,
5966 0x3c040800, 0x348437ff, 0x3c010800, 0xac221b74, 0x24020040, 0x3c010800,
5967 0xac221b78, 0x3c010800, 0xac201b70, 0xac600000, 0x24630004, 0x0083102b,
5968 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000, 0x00804821, 0x8faa0010,
5969 0x3c020800, 0x8c421b70, 0x3c040800, 0x8c841b78, 0x8fab0014, 0x24430001,
5970 0x0044102b, 0x3c010800, 0xac231b70, 0x14400003, 0x00004021, 0x3c010800,
5971 0xac201b70, 0x3c020800, 0x8c421b70, 0x3c030800, 0x8c631b74, 0x91240000,
5972 0x00021140, 0x00431021, 0x00481021, 0x25080001, 0xa0440000, 0x29020008,
5973 0x1440fff4, 0x25290001, 0x3c020800, 0x8c421b70, 0x3c030800, 0x8c631b74,
5974 0x8f64680c, 0x00021140, 0x00431021, 0xac440008, 0xac45000c, 0xac460010,
5975 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c, 0x00000000, 0x00000000,
5978 static const u32 tg3TsoFwRodata
[] = {
5979 0x4d61696e, 0x43707542, 0x00000000, 0x4d61696e, 0x43707541, 0x00000000,
5980 0x00000000, 0x00000000, 0x73746b6f, 0x66666c64, 0x496e0000, 0x73746b6f,
5981 0x66662a2a, 0x00000000, 0x53774576, 0x656e7430, 0x00000000, 0x00000000,
5982 0x00000000, 0x00000000, 0x66617461, 0x6c457272, 0x00000000, 0x00000000,
5986 static const u32 tg3TsoFwData
[] = {
5987 0x00000000, 0x73746b6f, 0x66666c64, 0x5f76312e, 0x362e3000, 0x00000000,
5988 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000, 0x00000000,
5992 /* 5705 needs a special version of the TSO firmware. */
5993 #define TG3_TSO5_FW_RELEASE_MAJOR 0x1
5994 #define TG3_TSO5_FW_RELASE_MINOR 0x2
5995 #define TG3_TSO5_FW_RELEASE_FIX 0x0
5996 #define TG3_TSO5_FW_START_ADDR 0x00010000
5997 #define TG3_TSO5_FW_TEXT_ADDR 0x00010000
5998 #define TG3_TSO5_FW_TEXT_LEN 0xe90
5999 #define TG3_TSO5_FW_RODATA_ADDR 0x00010e90
6000 #define TG3_TSO5_FW_RODATA_LEN 0x50
6001 #define TG3_TSO5_FW_DATA_ADDR 0x00010f00
6002 #define TG3_TSO5_FW_DATA_LEN 0x20
6003 #define TG3_TSO5_FW_SBSS_ADDR 0x00010f20
6004 #define TG3_TSO5_FW_SBSS_LEN 0x28
6005 #define TG3_TSO5_FW_BSS_ADDR 0x00010f50
6006 #define TG3_TSO5_FW_BSS_LEN 0x88
6008 static const u32 tg3Tso5FwText
[(TG3_TSO5_FW_TEXT_LEN
/ 4) + 1] = {
6009 0x0c004003, 0x00000000, 0x00010f04, 0x00000000, 0x10000003, 0x00000000,
6010 0x0000000d, 0x0000000d, 0x3c1d0001, 0x37bde000, 0x03a0f021, 0x3c100001,
6011 0x26100000, 0x0c004010, 0x00000000, 0x0000000d, 0x27bdffe0, 0x3c04fefe,
6012 0xafbf0018, 0x0c0042e8, 0x34840002, 0x0c004364, 0x00000000, 0x3c030001,
6013 0x90630f34, 0x24020002, 0x3c040001, 0x24840e9c, 0x14620003, 0x24050001,
6014 0x3c040001, 0x24840e90, 0x24060002, 0x00003821, 0xafa00010, 0x0c004378,
6015 0xafa00014, 0x0c00402c, 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020,
6016 0x00000000, 0x00000000, 0x27bdffe0, 0xafbf001c, 0xafb20018, 0xafb10014,
6017 0x0c0042d4, 0xafb00010, 0x3c128000, 0x24110001, 0x8f706810, 0x32020400,
6018 0x10400007, 0x00000000, 0x8f641008, 0x00921024, 0x14400003, 0x00000000,
6019 0x0c004064, 0x00000000, 0x3c020001, 0x90420f56, 0x10510003, 0x32020200,
6020 0x1040fff1, 0x00000000, 0x0c0041b4, 0x00000000, 0x08004034, 0x00000000,
6021 0x8fbf001c, 0x8fb20018, 0x8fb10014, 0x8fb00010, 0x03e00008, 0x27bd0020,
6022 0x27bdffe0, 0x3c040001, 0x24840eb0, 0x00002821, 0x00003021, 0x00003821,
6023 0xafbf0018, 0xafa00010, 0x0c004378, 0xafa00014, 0x0000d021, 0x24020130,
6024 0xaf625000, 0x3c010001, 0xa4200f50, 0x3c010001, 0xa0200f57, 0x8fbf0018,
6025 0x03e00008, 0x27bd0020, 0x00000000, 0x00000000, 0x3c030001, 0x24630f60,
6026 0x90620000, 0x27bdfff0, 0x14400003, 0x0080c021, 0x08004073, 0x00004821,
6027 0x3c022000, 0x03021024, 0x10400003, 0x24090002, 0x08004073, 0xa0600000,
6028 0x24090001, 0x00181040, 0x30431f80, 0x346f8008, 0x1520004b, 0x25eb0028,
6029 0x3c040001, 0x00832021, 0x8c848010, 0x3c050001, 0x24a50f7a, 0x00041402,
6030 0xa0a20000, 0x3c010001, 0xa0240f7b, 0x3c020001, 0x00431021, 0x94428014,
6031 0x3c010001, 0xa0220f7c, 0x3c0c0001, 0x01836021, 0x8d8c8018, 0x304200ff,
6032 0x24420008, 0x000220c3, 0x24020001, 0x3c010001, 0xa0220f60, 0x0124102b,
6033 0x1040000c, 0x00003821, 0x24a6000e, 0x01602821, 0x8ca20000, 0x8ca30004,
6034 0x24a50008, 0x24e70001, 0xacc20000, 0xacc30004, 0x00e4102b, 0x1440fff8,
6035 0x24c60008, 0x00003821, 0x3c080001, 0x25080f7b, 0x91060000, 0x3c020001,
6036 0x90420f7c, 0x2503000d, 0x00c32821, 0x00461023, 0x00021fc2, 0x00431021,
6037 0x00021043, 0x1840000c, 0x00002021, 0x91020001, 0x00461023, 0x00021fc2,
6038 0x00431021, 0x00021843, 0x94a20000, 0x24e70001, 0x00822021, 0x00e3102a,
6039 0x1440fffb, 0x24a50002, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402,
6040 0x00822021, 0x3c02ffff, 0x01821024, 0x3083ffff, 0x00431025, 0x3c010001,
6041 0x080040fa, 0xac220f80, 0x3c050001, 0x24a50f7c, 0x90a20000, 0x3c0c0001,
6042 0x01836021, 0x8d8c8018, 0x000220c2, 0x1080000e, 0x00003821, 0x01603021,
6043 0x24a5000c, 0x8ca20000, 0x8ca30004, 0x24a50008, 0x24e70001, 0xacc20000,
6044 0xacc30004, 0x00e4102b, 0x1440fff8, 0x24c60008, 0x3c050001, 0x24a50f7c,
6045 0x90a20000, 0x30430007, 0x24020004, 0x10620011, 0x28620005, 0x10400005,
6046 0x24020002, 0x10620008, 0x000710c0, 0x080040fa, 0x00000000, 0x24020006,
6047 0x1062000e, 0x000710c0, 0x080040fa, 0x00000000, 0x00a21821, 0x9463000c,
6048 0x004b1021, 0x080040fa, 0xa4430000, 0x000710c0, 0x00a21821, 0x8c63000c,
6049 0x004b1021, 0x080040fa, 0xac430000, 0x00a21821, 0x8c63000c, 0x004b2021,
6050 0x00a21021, 0xac830000, 0x94420010, 0xa4820004, 0x95e70006, 0x3c020001,
6051 0x90420f7c, 0x3c030001, 0x90630f7a, 0x00e2c823, 0x3c020001, 0x90420f7b,
6052 0x24630028, 0x01e34021, 0x24420028, 0x15200012, 0x01e23021, 0x94c2000c,
6053 0x3c010001, 0xa4220f78, 0x94c20004, 0x94c30006, 0x3c010001, 0xa4200f76,
6054 0x3c010001, 0xa4200f72, 0x00021400, 0x00431025, 0x3c010001, 0xac220f6c,
6055 0x95020004, 0x3c010001, 0x08004124, 0xa4220f70, 0x3c020001, 0x94420f70,
6056 0x3c030001, 0x94630f72, 0x00431021, 0xa5020004, 0x3c020001, 0x94420f6c,
6057 0xa4c20004, 0x3c020001, 0x8c420f6c, 0xa4c20006, 0x3c040001, 0x94840f72,
6058 0x3c020001, 0x94420f70, 0x3c0a0001, 0x954a0f76, 0x00441821, 0x3063ffff,
6059 0x0062182a, 0x24020002, 0x1122000b, 0x00832023, 0x3c030001, 0x94630f78,
6060 0x30620009, 0x10400006, 0x3062fff6, 0xa4c2000c, 0x3c020001, 0x94420f78,
6061 0x30420009, 0x01425023, 0x24020001, 0x1122001b, 0x29220002, 0x50400005,
6062 0x24020002, 0x11200007, 0x31a2ffff, 0x08004197, 0x00000000, 0x1122001d,
6063 0x24020016, 0x08004197, 0x31a2ffff, 0x3c0e0001, 0x95ce0f80, 0x10800005,
6064 0x01806821, 0x01c42021, 0x00041c02, 0x3082ffff, 0x00627021, 0x000e1027,
6065 0xa502000a, 0x3c030001, 0x90630f7b, 0x31a2ffff, 0x00e21021, 0x0800418d,
6066 0x00432023, 0x3c020001, 0x94420f80, 0x00442021, 0x00041c02, 0x3082ffff,
6067 0x00622021, 0x00807021, 0x00041027, 0x08004185, 0xa502000a, 0x3c050001,
6068 0x24a50f7a, 0x90a30000, 0x14620002, 0x24e2fff2, 0xa5e20034, 0x90a20000,
6069 0x00e21023, 0xa5020002, 0x3c030001, 0x94630f80, 0x3c020001, 0x94420f5a,
6070 0x30e5ffff, 0x00641821, 0x00451023, 0x00622023, 0x00041c02, 0x3082ffff,
6071 0x00622021, 0x00041027, 0xa502000a, 0x3c030001, 0x90630f7c, 0x24620001,
6072 0x14a20005, 0x00807021, 0x01631021, 0x90420000, 0x08004185, 0x00026200,
6073 0x24620002, 0x14a20003, 0x306200fe, 0x004b1021, 0x944c0000, 0x3c020001,
6074 0x94420f82, 0x3183ffff, 0x3c040001, 0x90840f7b, 0x00431021, 0x00e21021,
6075 0x00442023, 0x008a2021, 0x00041c02, 0x3082ffff, 0x00622021, 0x00041402,
6076 0x00822021, 0x00806821, 0x00041027, 0xa4c20010, 0x31a2ffff, 0x000e1c00,
6077 0x00431025, 0x3c040001, 0x24840f72, 0xade20010, 0x94820000, 0x3c050001,
6078 0x94a50f76, 0x3c030001, 0x8c630f6c, 0x24420001, 0x00b92821, 0xa4820000,
6079 0x3322ffff, 0x00622021, 0x0083182b, 0x3c010001, 0xa4250f76, 0x10600003,
6080 0x24a2ffff, 0x3c010001, 0xa4220f76, 0x3c024000, 0x03021025, 0x3c010001,
6081 0xac240f6c, 0xaf621008, 0x03e00008, 0x27bd0010, 0x3c030001, 0x90630f56,
6082 0x27bdffe8, 0x24020001, 0xafbf0014, 0x10620026, 0xafb00010, 0x8f620cf4,
6083 0x2442ffff, 0x3042007f, 0x00021100, 0x8c434000, 0x3c010001, 0xac230f64,
6084 0x8c434008, 0x24444000, 0x8c5c4004, 0x30620040, 0x14400002, 0x24020088,
6085 0x24020008, 0x3c010001, 0xa4220f68, 0x30620004, 0x10400005, 0x24020001,
6086 0x3c010001, 0xa0220f57, 0x080041d5, 0x00031402, 0x3c010001, 0xa0200f57,
6087 0x00031402, 0x3c010001, 0xa4220f54, 0x9483000c, 0x24020001, 0x3c010001,
6088 0xa4200f50, 0x3c010001, 0xa0220f56, 0x3c010001, 0xa4230f62, 0x24020001,
6089 0x1342001e, 0x00000000, 0x13400005, 0x24020003, 0x13420067, 0x00000000,
6090 0x080042cf, 0x00000000, 0x3c020001, 0x94420f62, 0x241a0001, 0x3c010001,
6091 0xa4200f5e, 0x3c010001, 0xa4200f52, 0x304407ff, 0x00021bc2, 0x00031823,
6092 0x3063003e, 0x34630036, 0x00021242, 0x3042003c, 0x00621821, 0x3c010001,
6093 0xa4240f58, 0x00832021, 0x24630030, 0x3c010001, 0xa4240f5a, 0x3c010001,
6094 0xa4230f5c, 0x3c060001, 0x24c60f52, 0x94c50000, 0x94c30002, 0x3c040001,
6095 0x94840f5a, 0x00651021, 0x0044102a, 0x10400013, 0x3c108000, 0x00a31021,
6096 0xa4c20000, 0x3c02a000, 0xaf620cf4, 0x3c010001, 0xa0200f56, 0x8f641008,
6097 0x00901024, 0x14400003, 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4,
6098 0x00501024, 0x104000b7, 0x00000000, 0x0800420f, 0x00000000, 0x3c030001,
6099 0x94630f50, 0x00851023, 0xa4c40000, 0x00621821, 0x3042ffff, 0x3c010001,
6100 0xa4230f50, 0xaf620ce8, 0x3c020001, 0x94420f68, 0x34420024, 0xaf620cec,
6101 0x94c30002, 0x3c020001, 0x94420f50, 0x14620012, 0x3c028000, 0x3c108000,
6102 0x3c02a000, 0xaf620cf4, 0x3c010001, 0xa0200f56, 0x8f641008, 0x00901024,
6103 0x14400003, 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4, 0x00501024,
6104 0x1440fff7, 0x00000000, 0x080042cf, 0x241a0003, 0xaf620cf4, 0x3c108000,
6105 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064, 0x00000000,
6106 0x8f620cf4, 0x00501024, 0x1440fff7, 0x00000000, 0x080042cf, 0x241a0003,
6107 0x3c070001, 0x24e70f50, 0x94e20000, 0x03821021, 0xaf620ce0, 0x3c020001,
6108 0x8c420f64, 0xaf620ce4, 0x3c050001, 0x94a50f54, 0x94e30000, 0x3c040001,
6109 0x94840f58, 0x3c020001, 0x94420f5e, 0x00a32823, 0x00822023, 0x30a6ffff,
6110 0x3083ffff, 0x00c3102b, 0x14400043, 0x00000000, 0x3c020001, 0x94420f5c,
6111 0x00021400, 0x00621025, 0xaf620ce8, 0x94e20000, 0x3c030001, 0x94630f54,
6112 0x00441021, 0xa4e20000, 0x3042ffff, 0x14430021, 0x3c020008, 0x3c020001,
6113 0x90420f57, 0x10400006, 0x3c03000c, 0x3c020001, 0x94420f68, 0x34630624,
6114 0x0800427c, 0x0000d021, 0x3c020001, 0x94420f68, 0x3c030008, 0x34630624,
6115 0x00431025, 0xaf620cec, 0x3c108000, 0x3c02a000, 0xaf620cf4, 0x3c010001,
6116 0xa0200f56, 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064,
6117 0x00000000, 0x8f620cf4, 0x00501024, 0x10400015, 0x00000000, 0x08004283,
6118 0x00000000, 0x3c030001, 0x94630f68, 0x34420624, 0x3c108000, 0x00621825,
6119 0x3c028000, 0xaf630cec, 0xaf620cf4, 0x8f641008, 0x00901024, 0x14400003,
6120 0x00000000, 0x0c004064, 0x00000000, 0x8f620cf4, 0x00501024, 0x1440fff7,
6121 0x00000000, 0x3c010001, 0x080042cf, 0xa4200f5e, 0x3c020001, 0x94420f5c,
6122 0x00021400, 0x00c21025, 0xaf620ce8, 0x3c020001, 0x90420f57, 0x10400009,
6123 0x3c03000c, 0x3c020001, 0x94420f68, 0x34630624, 0x0000d021, 0x00431025,
6124 0xaf620cec, 0x080042c1, 0x3c108000, 0x3c020001, 0x94420f68, 0x3c030008,
6125 0x34630604, 0x00431025, 0xaf620cec, 0x3c020001, 0x94420f5e, 0x00451021,
6126 0x3c010001, 0xa4220f5e, 0x3c108000, 0x3c02a000, 0xaf620cf4, 0x3c010001,
6127 0xa0200f56, 0x8f641008, 0x00901024, 0x14400003, 0x00000000, 0x0c004064,
6128 0x00000000, 0x8f620cf4, 0x00501024, 0x1440fff7, 0x00000000, 0x8fbf0014,
6129 0x8fb00010, 0x03e00008, 0x27bd0018, 0x00000000, 0x27bdffe0, 0x3c040001,
6130 0x24840ec0, 0x00002821, 0x00003021, 0x00003821, 0xafbf0018, 0xafa00010,
6131 0x0c004378, 0xafa00014, 0x0000d021, 0x24020130, 0xaf625000, 0x3c010001,
6132 0xa4200f50, 0x3c010001, 0xa0200f57, 0x8fbf0018, 0x03e00008, 0x27bd0020,
6133 0x27bdffe8, 0x3c1bc000, 0xafbf0014, 0xafb00010, 0xaf60680c, 0x8f626804,
6134 0x34420082, 0xaf626804, 0x8f634000, 0x24020b50, 0x3c010001, 0xac220f20,
6135 0x24020b78, 0x3c010001, 0xac220f30, 0x34630002, 0xaf634000, 0x0c004315,
6136 0x00808021, 0x3c010001, 0xa0220f34, 0x304200ff, 0x24030002, 0x14430005,
6137 0x00000000, 0x3c020001, 0x8c420f20, 0x08004308, 0xac5000c0, 0x3c020001,
6138 0x8c420f20, 0xac5000bc, 0x8f624434, 0x8f634438, 0x8f644410, 0x3c010001,
6139 0xac220f28, 0x3c010001, 0xac230f38, 0x3c010001, 0xac240f24, 0x8fbf0014,
6140 0x8fb00010, 0x03e00008, 0x27bd0018, 0x03e00008, 0x24020001, 0x27bdfff8,
6141 0x18800009, 0x00002821, 0x8f63680c, 0x8f62680c, 0x1043fffe, 0x00000000,
6142 0x24a50001, 0x00a4102a, 0x1440fff9, 0x00000000, 0x03e00008, 0x27bd0008,
6143 0x8f634450, 0x3c020001, 0x8c420f28, 0x00031c02, 0x0043102b, 0x14400008,
6144 0x3c038000, 0x3c040001, 0x8c840f38, 0x8f624450, 0x00021c02, 0x0083102b,
6145 0x1040fffc, 0x3c038000, 0xaf634444, 0x8f624444, 0x00431024, 0x1440fffd,
6146 0x00000000, 0x8f624448, 0x03e00008, 0x3042ffff, 0x3082ffff, 0x2442e000,
6147 0x2c422001, 0x14400003, 0x3c024000, 0x08004347, 0x2402ffff, 0x00822025,
6148 0xaf645c38, 0x8f625c30, 0x30420002, 0x1440fffc, 0x00001021, 0x03e00008,
6149 0x00000000, 0x8f624450, 0x3c030001, 0x8c630f24, 0x08004350, 0x3042ffff,
6150 0x8f624450, 0x3042ffff, 0x0043102b, 0x1440fffc, 0x00000000, 0x03e00008,
6151 0x00000000, 0x27bdffe0, 0x00802821, 0x3c040001, 0x24840ed0, 0x00003021,
6152 0x00003821, 0xafbf0018, 0xafa00010, 0x0c004378, 0xafa00014, 0x0800435f,
6153 0x00000000, 0x8fbf0018, 0x03e00008, 0x27bd0020, 0x3c020001, 0x3442d600,
6154 0x3c030001, 0x3463d600, 0x3c040001, 0x3484ddff, 0x3c010001, 0xac220f40,
6155 0x24020040, 0x3c010001, 0xac220f44, 0x3c010001, 0xac200f3c, 0xac600000,
6156 0x24630004, 0x0083102b, 0x5040fffd, 0xac600000, 0x03e00008, 0x00000000,
6157 0x00804821, 0x8faa0010, 0x3c020001, 0x8c420f3c, 0x3c040001, 0x8c840f44,
6158 0x8fab0014, 0x24430001, 0x0044102b, 0x3c010001, 0xac230f3c, 0x14400003,
6159 0x00004021, 0x3c010001, 0xac200f3c, 0x3c020001, 0x8c420f3c, 0x3c030001,
6160 0x8c630f40, 0x91240000, 0x00021140, 0x00431021, 0x00481021, 0x25080001,
6161 0xa0440000, 0x29020008, 0x1440fff4, 0x25290001, 0x3c020001, 0x8c420f3c,
6162 0x3c030001, 0x8c630f40, 0x8f64680c, 0x00021140, 0x00431021, 0xac440008,
6163 0xac45000c, 0xac460010, 0xac470014, 0xac4a0018, 0x03e00008, 0xac4b001c,
6164 0x00000000, 0x00000000, 0x00000000,
6167 static const u32 tg3Tso5FwRodata
[(TG3_TSO5_FW_RODATA_LEN
/ 4) + 1] = {
6168 0x4d61696e, 0x43707542, 0x00000000, 0x4d61696e, 0x43707541, 0x00000000,
6169 0x00000000, 0x00000000, 0x73746b6f, 0x66666c64, 0x00000000, 0x00000000,
6170 0x73746b6f, 0x66666c64, 0x00000000, 0x00000000, 0x66617461, 0x6c457272,
6171 0x00000000, 0x00000000, 0x00000000,
6174 static const u32 tg3Tso5FwData
[(TG3_TSO5_FW_DATA_LEN
/ 4) + 1] = {
6175 0x00000000, 0x73746b6f, 0x66666c64, 0x5f76312e, 0x322e3000, 0x00000000,
6176 0x00000000, 0x00000000, 0x00000000,
6179 /* tp->lock is held. */
6180 static int tg3_load_tso_firmware(struct tg3
*tp
)
6182 struct fw_info info
;
6183 unsigned long cpu_base
, cpu_scratch_base
, cpu_scratch_size
;
6186 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
)
6189 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) {
6190 info
.text_base
= TG3_TSO5_FW_TEXT_ADDR
;
6191 info
.text_len
= TG3_TSO5_FW_TEXT_LEN
;
6192 info
.text_data
= &tg3Tso5FwText
[0];
6193 info
.rodata_base
= TG3_TSO5_FW_RODATA_ADDR
;
6194 info
.rodata_len
= TG3_TSO5_FW_RODATA_LEN
;
6195 info
.rodata_data
= &tg3Tso5FwRodata
[0];
6196 info
.data_base
= TG3_TSO5_FW_DATA_ADDR
;
6197 info
.data_len
= TG3_TSO5_FW_DATA_LEN
;
6198 info
.data_data
= &tg3Tso5FwData
[0];
6199 cpu_base
= RX_CPU_BASE
;
6200 cpu_scratch_base
= NIC_SRAM_MBUF_POOL_BASE5705
;
6201 cpu_scratch_size
= (info
.text_len
+
6204 TG3_TSO5_FW_SBSS_LEN
+
6205 TG3_TSO5_FW_BSS_LEN
);
6207 info
.text_base
= TG3_TSO_FW_TEXT_ADDR
;
6208 info
.text_len
= TG3_TSO_FW_TEXT_LEN
;
6209 info
.text_data
= &tg3TsoFwText
[0];
6210 info
.rodata_base
= TG3_TSO_FW_RODATA_ADDR
;
6211 info
.rodata_len
= TG3_TSO_FW_RODATA_LEN
;
6212 info
.rodata_data
= &tg3TsoFwRodata
[0];
6213 info
.data_base
= TG3_TSO_FW_DATA_ADDR
;
6214 info
.data_len
= TG3_TSO_FW_DATA_LEN
;
6215 info
.data_data
= &tg3TsoFwData
[0];
6216 cpu_base
= TX_CPU_BASE
;
6217 cpu_scratch_base
= TX_CPU_SCRATCH_BASE
;
6218 cpu_scratch_size
= TX_CPU_SCRATCH_SIZE
;
6221 err
= tg3_load_firmware_cpu(tp
, cpu_base
,
6222 cpu_scratch_base
, cpu_scratch_size
,
6227 /* Now startup the cpu. */
6228 tw32(cpu_base
+ CPU_STATE
, 0xffffffff);
6229 tw32_f(cpu_base
+ CPU_PC
, info
.text_base
);
6231 for (i
= 0; i
< 5; i
++) {
6232 if (tr32(cpu_base
+ CPU_PC
) == info
.text_base
)
6234 tw32(cpu_base
+ CPU_STATE
, 0xffffffff);
6235 tw32(cpu_base
+ CPU_MODE
, CPU_MODE_HALT
);
6236 tw32_f(cpu_base
+ CPU_PC
, info
.text_base
);
6240 printk(KERN_ERR PFX
"tg3_load_tso_firmware fails for %s "
6241 "to set CPU PC, is %08x should be %08x\n",
6242 tp
->dev
->name
, tr32(cpu_base
+ CPU_PC
),
6246 tw32(cpu_base
+ CPU_STATE
, 0xffffffff);
6247 tw32_f(cpu_base
+ CPU_MODE
, 0x00000000);
6252 /* tp->lock is held. */
6253 static void __tg3_set_mac_addr(struct tg3
*tp
, int skip_mac_1
)
6255 u32 addr_high
, addr_low
;
6258 addr_high
= ((tp
->dev
->dev_addr
[0] << 8) |
6259 tp
->dev
->dev_addr
[1]);
6260 addr_low
= ((tp
->dev
->dev_addr
[2] << 24) |
6261 (tp
->dev
->dev_addr
[3] << 16) |
6262 (tp
->dev
->dev_addr
[4] << 8) |
6263 (tp
->dev
->dev_addr
[5] << 0));
6264 for (i
= 0; i
< 4; i
++) {
6265 if (i
== 1 && skip_mac_1
)
6267 tw32(MAC_ADDR_0_HIGH
+ (i
* 8), addr_high
);
6268 tw32(MAC_ADDR_0_LOW
+ (i
* 8), addr_low
);
6271 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
6272 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) {
6273 for (i
= 0; i
< 12; i
++) {
6274 tw32(MAC_EXTADDR_0_HIGH
+ (i
* 8), addr_high
);
6275 tw32(MAC_EXTADDR_0_LOW
+ (i
* 8), addr_low
);
6279 addr_high
= (tp
->dev
->dev_addr
[0] +
6280 tp
->dev
->dev_addr
[1] +
6281 tp
->dev
->dev_addr
[2] +
6282 tp
->dev
->dev_addr
[3] +
6283 tp
->dev
->dev_addr
[4] +
6284 tp
->dev
->dev_addr
[5]) &
6285 TX_BACKOFF_SEED_MASK
;
6286 tw32(MAC_TX_BACKOFF_SEED
, addr_high
);
6289 static int tg3_set_mac_addr(struct net_device
*dev
, void *p
)
6291 struct tg3
*tp
= netdev_priv(dev
);
6292 struct sockaddr
*addr
= p
;
6293 int err
= 0, skip_mac_1
= 0;
6295 if (!is_valid_ether_addr(addr
->sa_data
))
6298 memcpy(dev
->dev_addr
, addr
->sa_data
, dev
->addr_len
);
6300 if (!netif_running(dev
))
6303 if (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) {
6304 u32 addr0_high
, addr0_low
, addr1_high
, addr1_low
;
6306 addr0_high
= tr32(MAC_ADDR_0_HIGH
);
6307 addr0_low
= tr32(MAC_ADDR_0_LOW
);
6308 addr1_high
= tr32(MAC_ADDR_1_HIGH
);
6309 addr1_low
= tr32(MAC_ADDR_1_LOW
);
6311 /* Skip MAC addr 1 if ASF is using it. */
6312 if ((addr0_high
!= addr1_high
|| addr0_low
!= addr1_low
) &&
6313 !(addr1_high
== 0 && addr1_low
== 0))
6316 spin_lock_bh(&tp
->lock
);
6317 __tg3_set_mac_addr(tp
, skip_mac_1
);
6318 spin_unlock_bh(&tp
->lock
);
6323 /* tp->lock is held. */
6324 static void tg3_set_bdinfo(struct tg3
*tp
, u32 bdinfo_addr
,
6325 dma_addr_t mapping
, u32 maxlen_flags
,
6329 (bdinfo_addr
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_HIGH
),
6330 ((u64
) mapping
>> 32));
6332 (bdinfo_addr
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_LOW
),
6333 ((u64
) mapping
& 0xffffffff));
6335 (bdinfo_addr
+ TG3_BDINFO_MAXLEN_FLAGS
),
6338 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
6340 (bdinfo_addr
+ TG3_BDINFO_NIC_ADDR
),
6344 static void __tg3_set_rx_mode(struct net_device
*);
6345 static void __tg3_set_coalesce(struct tg3
*tp
, struct ethtool_coalesce
*ec
)
6347 tw32(HOSTCC_RXCOL_TICKS
, ec
->rx_coalesce_usecs
);
6348 tw32(HOSTCC_TXCOL_TICKS
, ec
->tx_coalesce_usecs
);
6349 tw32(HOSTCC_RXMAX_FRAMES
, ec
->rx_max_coalesced_frames
);
6350 tw32(HOSTCC_TXMAX_FRAMES
, ec
->tx_max_coalesced_frames
);
6351 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6352 tw32(HOSTCC_RXCOAL_TICK_INT
, ec
->rx_coalesce_usecs_irq
);
6353 tw32(HOSTCC_TXCOAL_TICK_INT
, ec
->tx_coalesce_usecs_irq
);
6355 tw32(HOSTCC_RXCOAL_MAXF_INT
, ec
->rx_max_coalesced_frames_irq
);
6356 tw32(HOSTCC_TXCOAL_MAXF_INT
, ec
->tx_max_coalesced_frames_irq
);
6357 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6358 u32 val
= ec
->stats_block_coalesce_usecs
;
6360 if (!netif_carrier_ok(tp
->dev
))
6363 tw32(HOSTCC_STAT_COAL_TICKS
, val
);
6367 /* tp->lock is held. */
6368 static int tg3_reset_hw(struct tg3
*tp
, int reset_phy
)
6370 u32 val
, rdmac_mode
;
6373 tg3_disable_ints(tp
);
6377 tg3_write_sig_pre_reset(tp
, RESET_KIND_INIT
);
6379 if (tp
->tg3_flags
& TG3_FLAG_INIT_COMPLETE
) {
6380 tg3_abort_hw(tp
, 1);
6386 err
= tg3_chip_reset(tp
);
6390 tg3_write_sig_legacy(tp
, RESET_KIND_INIT
);
6392 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A0
||
6393 tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A1
) {
6394 val
= tr32(TG3_CPMU_CTRL
);
6395 val
&= ~(CPMU_CTRL_LINK_AWARE_MODE
| CPMU_CTRL_LINK_IDLE_MODE
);
6396 tw32(TG3_CPMU_CTRL
, val
);
6398 val
= tr32(TG3_CPMU_LSPD_10MB_CLK
);
6399 val
&= ~CPMU_LSPD_10MB_MACCLK_MASK
;
6400 val
|= CPMU_LSPD_10MB_MACCLK_6_25
;
6401 tw32(TG3_CPMU_LSPD_10MB_CLK
, val
);
6403 val
= tr32(TG3_CPMU_LNK_AWARE_PWRMD
);
6404 val
&= ~CPMU_LNK_AWARE_MACCLK_MASK
;
6405 val
|= CPMU_LNK_AWARE_MACCLK_6_25
;
6406 tw32(TG3_CPMU_LNK_AWARE_PWRMD
, val
);
6408 val
= tr32(TG3_CPMU_HST_ACC
);
6409 val
&= ~CPMU_HST_ACC_MACCLK_MASK
;
6410 val
|= CPMU_HST_ACC_MACCLK_6_25
;
6411 tw32(TG3_CPMU_HST_ACC
, val
);
6414 /* This works around an issue with Athlon chipsets on
6415 * B3 tigon3 silicon. This bit has no effect on any
6416 * other revision. But do not set this on PCI Express
6417 * chips and don't even touch the clocks if the CPMU is present.
6419 if (!(tp
->tg3_flags
& TG3_FLAG_CPMU_PRESENT
)) {
6420 if (!(tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
))
6421 tp
->pci_clock_ctrl
|= CLOCK_CTRL_DELAY_PCI_GRANT
;
6422 tw32_f(TG3PCI_CLOCK_CTRL
, tp
->pci_clock_ctrl
);
6425 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5704_A0
&&
6426 (tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
)) {
6427 val
= tr32(TG3PCI_PCISTATE
);
6428 val
|= PCISTATE_RETRY_SAME_DMA
;
6429 tw32(TG3PCI_PCISTATE
, val
);
6432 if (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
) {
6433 /* Allow reads and writes to the
6434 * APE register and memory space.
6436 val
= tr32(TG3PCI_PCISTATE
);
6437 val
|= PCISTATE_ALLOW_APE_CTLSPC_WR
|
6438 PCISTATE_ALLOW_APE_SHMEM_WR
;
6439 tw32(TG3PCI_PCISTATE
, val
);
6442 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5704_BX
) {
6443 /* Enable some hw fixes. */
6444 val
= tr32(TG3PCI_MSI_DATA
);
6445 val
|= (1 << 26) | (1 << 28) | (1 << 29);
6446 tw32(TG3PCI_MSI_DATA
, val
);
6449 /* Descriptor ring init may make accesses to the
6450 * NIC SRAM area to setup the TX descriptors, so we
6451 * can only do this after the hardware has been
6452 * successfully reset.
6454 err
= tg3_init_rings(tp
);
6458 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5784
&&
6459 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5761
) {
6460 /* This value is determined during the probe time DMA
6461 * engine test, tg3_test_dma.
6463 tw32(TG3PCI_DMA_RW_CTRL
, tp
->dma_rwctrl
);
6466 tp
->grc_mode
&= ~(GRC_MODE_HOST_SENDBDS
|
6467 GRC_MODE_4X_NIC_SEND_RINGS
|
6468 GRC_MODE_NO_TX_PHDR_CSUM
|
6469 GRC_MODE_NO_RX_PHDR_CSUM
);
6470 tp
->grc_mode
|= GRC_MODE_HOST_SENDBDS
;
6472 /* Pseudo-header checksum is done by hardware logic and not
6473 * the offload processers, so make the chip do the pseudo-
6474 * header checksums on receive. For transmit it is more
6475 * convenient to do the pseudo-header checksum in software
6476 * as Linux does that on transmit for us in all cases.
6478 tp
->grc_mode
|= GRC_MODE_NO_TX_PHDR_CSUM
;
6482 (GRC_MODE_IRQ_ON_MAC_ATTN
| GRC_MODE_HOST_STACKUP
));
6484 /* Setup the timer prescalar register. Clock is always 66Mhz. */
6485 val
= tr32(GRC_MISC_CFG
);
6487 val
|= (65 << GRC_MISC_CFG_PRESCALAR_SHIFT
);
6488 tw32(GRC_MISC_CFG
, val
);
6490 /* Initialize MBUF/DESC pool. */
6491 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) {
6493 } else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5705
) {
6494 tw32(BUFMGR_MB_POOL_ADDR
, NIC_SRAM_MBUF_POOL_BASE
);
6495 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
)
6496 tw32(BUFMGR_MB_POOL_SIZE
, NIC_SRAM_MBUF_POOL_SIZE64
);
6498 tw32(BUFMGR_MB_POOL_SIZE
, NIC_SRAM_MBUF_POOL_SIZE96
);
6499 tw32(BUFMGR_DMA_DESC_POOL_ADDR
, NIC_SRAM_DMA_DESC_POOL_BASE
);
6500 tw32(BUFMGR_DMA_DESC_POOL_SIZE
, NIC_SRAM_DMA_DESC_POOL_SIZE
);
6502 else if (tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
) {
6505 fw_len
= (TG3_TSO5_FW_TEXT_LEN
+
6506 TG3_TSO5_FW_RODATA_LEN
+
6507 TG3_TSO5_FW_DATA_LEN
+
6508 TG3_TSO5_FW_SBSS_LEN
+
6509 TG3_TSO5_FW_BSS_LEN
);
6510 fw_len
= (fw_len
+ (0x80 - 1)) & ~(0x80 - 1);
6511 tw32(BUFMGR_MB_POOL_ADDR
,
6512 NIC_SRAM_MBUF_POOL_BASE5705
+ fw_len
);
6513 tw32(BUFMGR_MB_POOL_SIZE
,
6514 NIC_SRAM_MBUF_POOL_SIZE5705
- fw_len
- 0xa00);
6517 if (tp
->dev
->mtu
<= ETH_DATA_LEN
) {
6518 tw32(BUFMGR_MB_RDMA_LOW_WATER
,
6519 tp
->bufmgr_config
.mbuf_read_dma_low_water
);
6520 tw32(BUFMGR_MB_MACRX_LOW_WATER
,
6521 tp
->bufmgr_config
.mbuf_mac_rx_low_water
);
6522 tw32(BUFMGR_MB_HIGH_WATER
,
6523 tp
->bufmgr_config
.mbuf_high_water
);
6525 tw32(BUFMGR_MB_RDMA_LOW_WATER
,
6526 tp
->bufmgr_config
.mbuf_read_dma_low_water_jumbo
);
6527 tw32(BUFMGR_MB_MACRX_LOW_WATER
,
6528 tp
->bufmgr_config
.mbuf_mac_rx_low_water_jumbo
);
6529 tw32(BUFMGR_MB_HIGH_WATER
,
6530 tp
->bufmgr_config
.mbuf_high_water_jumbo
);
6532 tw32(BUFMGR_DMA_LOW_WATER
,
6533 tp
->bufmgr_config
.dma_low_water
);
6534 tw32(BUFMGR_DMA_HIGH_WATER
,
6535 tp
->bufmgr_config
.dma_high_water
);
6537 tw32(BUFMGR_MODE
, BUFMGR_MODE_ENABLE
| BUFMGR_MODE_ATTN_ENABLE
);
6538 for (i
= 0; i
< 2000; i
++) {
6539 if (tr32(BUFMGR_MODE
) & BUFMGR_MODE_ENABLE
)
6544 printk(KERN_ERR PFX
"tg3_reset_hw cannot enable BUFMGR for %s.\n",
6549 /* Setup replenish threshold. */
6550 val
= tp
->rx_pending
/ 8;
6553 else if (val
> tp
->rx_std_max_post
)
6554 val
= tp
->rx_std_max_post
;
6555 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
6556 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5906_A1
)
6557 tw32(ISO_PKT_TX
, (tr32(ISO_PKT_TX
) & ~0x3) | 0x2);
6559 if (val
> (TG3_RX_INTERNAL_RING_SZ_5906
/ 2))
6560 val
= TG3_RX_INTERNAL_RING_SZ_5906
/ 2;
6563 tw32(RCVBDI_STD_THRESH
, val
);
6565 /* Initialize TG3_BDINFO's at:
6566 * RCVDBDI_STD_BD: standard eth size rx ring
6567 * RCVDBDI_JUMBO_BD: jumbo frame rx ring
6568 * RCVDBDI_MINI_BD: small frame rx ring (??? does not work)
6571 * TG3_BDINFO_HOST_ADDR: high/low parts of DMA address of ring
6572 * TG3_BDINFO_MAXLEN_FLAGS: (rx max buffer size << 16) |
6573 * ring attribute flags
6574 * TG3_BDINFO_NIC_ADDR: location of descriptors in nic SRAM
6576 * Standard receive ring @ NIC_SRAM_RX_BUFFER_DESC, 512 entries.
6577 * Jumbo receive ring @ NIC_SRAM_RX_JUMBO_BUFFER_DESC, 256 entries.
6579 * The size of each ring is fixed in the firmware, but the location is
6582 tw32(RCVDBDI_STD_BD
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_HIGH
,
6583 ((u64
) tp
->rx_std_mapping
>> 32));
6584 tw32(RCVDBDI_STD_BD
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_LOW
,
6585 ((u64
) tp
->rx_std_mapping
& 0xffffffff));
6586 tw32(RCVDBDI_STD_BD
+ TG3_BDINFO_NIC_ADDR
,
6587 NIC_SRAM_RX_BUFFER_DESC
);
6589 /* Don't even try to program the JUMBO/MINI buffer descriptor
6592 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
6593 tw32(RCVDBDI_STD_BD
+ TG3_BDINFO_MAXLEN_FLAGS
,
6594 RX_STD_MAX_SIZE_5705
<< BDINFO_FLAGS_MAXLEN_SHIFT
);
6596 tw32(RCVDBDI_STD_BD
+ TG3_BDINFO_MAXLEN_FLAGS
,
6597 RX_STD_MAX_SIZE
<< BDINFO_FLAGS_MAXLEN_SHIFT
);
6599 tw32(RCVDBDI_MINI_BD
+ TG3_BDINFO_MAXLEN_FLAGS
,
6600 BDINFO_FLAGS_DISABLED
);
6602 /* Setup replenish threshold. */
6603 tw32(RCVBDI_JUMBO_THRESH
, tp
->rx_jumbo_pending
/ 8);
6605 if (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
) {
6606 tw32(RCVDBDI_JUMBO_BD
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_HIGH
,
6607 ((u64
) tp
->rx_jumbo_mapping
>> 32));
6608 tw32(RCVDBDI_JUMBO_BD
+ TG3_BDINFO_HOST_ADDR
+ TG3_64BIT_REG_LOW
,
6609 ((u64
) tp
->rx_jumbo_mapping
& 0xffffffff));
6610 tw32(RCVDBDI_JUMBO_BD
+ TG3_BDINFO_MAXLEN_FLAGS
,
6611 RX_JUMBO_MAX_SIZE
<< BDINFO_FLAGS_MAXLEN_SHIFT
);
6612 tw32(RCVDBDI_JUMBO_BD
+ TG3_BDINFO_NIC_ADDR
,
6613 NIC_SRAM_RX_JUMBO_BUFFER_DESC
);
6615 tw32(RCVDBDI_JUMBO_BD
+ TG3_BDINFO_MAXLEN_FLAGS
,
6616 BDINFO_FLAGS_DISABLED
);
6621 /* There is only one send ring on 5705/5750, no need to explicitly
6622 * disable the others.
6624 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6625 /* Clear out send RCB ring in SRAM. */
6626 for (i
= NIC_SRAM_SEND_RCB
; i
< NIC_SRAM_RCV_RET_RCB
; i
+= TG3_BDINFO_SIZE
)
6627 tg3_write_mem(tp
, i
+ TG3_BDINFO_MAXLEN_FLAGS
,
6628 BDINFO_FLAGS_DISABLED
);
6633 tw32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0
+ TG3_64BIT_REG_LOW
, 0);
6634 tw32_tx_mbox(MAILBOX_SNDNIC_PROD_IDX_0
+ TG3_64BIT_REG_LOW
, 0);
6636 tg3_set_bdinfo(tp
, NIC_SRAM_SEND_RCB
,
6637 tp
->tx_desc_mapping
,
6638 (TG3_TX_RING_SIZE
<<
6639 BDINFO_FLAGS_MAXLEN_SHIFT
),
6640 NIC_SRAM_TX_BUFFER_DESC
);
6642 /* There is only one receive return ring on 5705/5750, no need
6643 * to explicitly disable the others.
6645 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6646 for (i
= NIC_SRAM_RCV_RET_RCB
; i
< NIC_SRAM_STATS_BLK
;
6647 i
+= TG3_BDINFO_SIZE
) {
6648 tg3_write_mem(tp
, i
+ TG3_BDINFO_MAXLEN_FLAGS
,
6649 BDINFO_FLAGS_DISABLED
);
6654 tw32_rx_mbox(MAILBOX_RCVRET_CON_IDX_0
+ TG3_64BIT_REG_LOW
, 0);
6656 tg3_set_bdinfo(tp
, NIC_SRAM_RCV_RET_RCB
,
6658 (TG3_RX_RCB_RING_SIZE(tp
) <<
6659 BDINFO_FLAGS_MAXLEN_SHIFT
),
6662 tp
->rx_std_ptr
= tp
->rx_pending
;
6663 tw32_rx_mbox(MAILBOX_RCV_STD_PROD_IDX
+ TG3_64BIT_REG_LOW
,
6666 tp
->rx_jumbo_ptr
= (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
) ?
6667 tp
->rx_jumbo_pending
: 0;
6668 tw32_rx_mbox(MAILBOX_RCV_JUMBO_PROD_IDX
+ TG3_64BIT_REG_LOW
,
6671 /* Initialize MAC address and backoff seed. */
6672 __tg3_set_mac_addr(tp
, 0);
6674 /* MTU + ethernet header + FCS + optional VLAN tag */
6675 tw32(MAC_RX_MTU_SIZE
, tp
->dev
->mtu
+ ETH_HLEN
+ 8);
6677 /* The slot time is changed by tg3_setup_phy if we
6678 * run at gigabit with half duplex.
6680 tw32(MAC_TX_LENGTHS
,
6681 (2 << TX_LENGTHS_IPG_CRS_SHIFT
) |
6682 (6 << TX_LENGTHS_IPG_SHIFT
) |
6683 (32 << TX_LENGTHS_SLOT_TIME_SHIFT
));
6685 /* Receive rules. */
6686 tw32(MAC_RCV_RULE_CFG
, RCV_RULE_CFG_DEFAULT_CLASS
);
6687 tw32(RCVLPC_CONFIG
, 0x0181);
6689 /* Calculate RDMAC_MODE setting early, we need it to determine
6690 * the RCVLPC_STATE_ENABLE mask.
6692 rdmac_mode
= (RDMAC_MODE_ENABLE
| RDMAC_MODE_TGTABORT_ENAB
|
6693 RDMAC_MODE_MSTABORT_ENAB
| RDMAC_MODE_PARITYERR_ENAB
|
6694 RDMAC_MODE_ADDROFLOW_ENAB
| RDMAC_MODE_FIFOOFLOW_ENAB
|
6695 RDMAC_MODE_FIFOURUN_ENAB
| RDMAC_MODE_FIFOOREAD_ENAB
|
6696 RDMAC_MODE_LNGREAD_ENAB
);
6698 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
)
6699 rdmac_mode
|= RDMAC_MODE_BD_SBD_CRPT_ENAB
|
6700 RDMAC_MODE_MBUF_RBD_CRPT_ENAB
|
6701 RDMAC_MODE_MBUF_SBD_CRPT_ENAB
;
6703 /* If statement applies to 5705 and 5750 PCI devices only */
6704 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
&&
6705 tp
->pci_chip_rev_id
!= CHIPREV_ID_5705_A0
) ||
6706 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
)) {
6707 if (tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
&&
6708 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) {
6709 rdmac_mode
|= RDMAC_MODE_FIFO_SIZE_128
;
6710 } else if (!(tr32(TG3PCI_PCISTATE
) & PCISTATE_BUS_SPEED_HIGH
) &&
6711 !(tp
->tg3_flags2
& TG3_FLG2_IS_5788
)) {
6712 rdmac_mode
|= RDMAC_MODE_FIFO_LONG_BURST
;
6716 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
)
6717 rdmac_mode
|= RDMAC_MODE_FIFO_LONG_BURST
;
6719 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
)
6720 rdmac_mode
|= (1 << 27);
6722 /* Receive/send statistics. */
6723 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) {
6724 val
= tr32(RCVLPC_STATS_ENABLE
);
6725 val
&= ~RCVLPC_STATSENAB_DACK_FIX
;
6726 tw32(RCVLPC_STATS_ENABLE
, val
);
6727 } else if ((rdmac_mode
& RDMAC_MODE_FIFO_SIZE_128
) &&
6728 (tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
)) {
6729 val
= tr32(RCVLPC_STATS_ENABLE
);
6730 val
&= ~RCVLPC_STATSENAB_LNGBRST_RFIX
;
6731 tw32(RCVLPC_STATS_ENABLE
, val
);
6733 tw32(RCVLPC_STATS_ENABLE
, 0xffffff);
6735 tw32(RCVLPC_STATSCTRL
, RCVLPC_STATSCTRL_ENABLE
);
6736 tw32(SNDDATAI_STATSENAB
, 0xffffff);
6737 tw32(SNDDATAI_STATSCTRL
,
6738 (SNDDATAI_SCTRL_ENABLE
|
6739 SNDDATAI_SCTRL_FASTUPD
));
6741 /* Setup host coalescing engine. */
6742 tw32(HOSTCC_MODE
, 0);
6743 for (i
= 0; i
< 2000; i
++) {
6744 if (!(tr32(HOSTCC_MODE
) & HOSTCC_MODE_ENABLE
))
6749 __tg3_set_coalesce(tp
, &tp
->coal
);
6751 /* set status block DMA address */
6752 tw32(HOSTCC_STATUS_BLK_HOST_ADDR
+ TG3_64BIT_REG_HIGH
,
6753 ((u64
) tp
->status_mapping
>> 32));
6754 tw32(HOSTCC_STATUS_BLK_HOST_ADDR
+ TG3_64BIT_REG_LOW
,
6755 ((u64
) tp
->status_mapping
& 0xffffffff));
6757 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6758 /* Status/statistics block address. See tg3_timer,
6759 * the tg3_periodic_fetch_stats call there, and
6760 * tg3_get_stats to see how this works for 5705/5750 chips.
6762 tw32(HOSTCC_STATS_BLK_HOST_ADDR
+ TG3_64BIT_REG_HIGH
,
6763 ((u64
) tp
->stats_mapping
>> 32));
6764 tw32(HOSTCC_STATS_BLK_HOST_ADDR
+ TG3_64BIT_REG_LOW
,
6765 ((u64
) tp
->stats_mapping
& 0xffffffff));
6766 tw32(HOSTCC_STATS_BLK_NIC_ADDR
, NIC_SRAM_STATS_BLK
);
6767 tw32(HOSTCC_STATUS_BLK_NIC_ADDR
, NIC_SRAM_STATUS_BLK
);
6770 tw32(HOSTCC_MODE
, HOSTCC_MODE_ENABLE
| tp
->coalesce_mode
);
6772 tw32(RCVCC_MODE
, RCVCC_MODE_ENABLE
| RCVCC_MODE_ATTN_ENABLE
);
6773 tw32(RCVLPC_MODE
, RCVLPC_MODE_ENABLE
);
6774 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
6775 tw32(RCVLSC_MODE
, RCVLSC_MODE_ENABLE
| RCVLSC_MODE_ATTN_ENABLE
);
6777 /* Clear statistics/status block in chip, and status block in ram. */
6778 for (i
= NIC_SRAM_STATS_BLK
;
6779 i
< NIC_SRAM_STATUS_BLK
+ TG3_HW_STATUS_SIZE
;
6781 tg3_write_mem(tp
, i
, 0);
6784 memset(tp
->hw_status
, 0, TG3_HW_STATUS_SIZE
);
6786 if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) {
6787 tp
->tg3_flags2
&= ~TG3_FLG2_PARALLEL_DETECT
;
6788 /* reset to prevent losing 1st rx packet intermittently */
6789 tw32_f(MAC_RX_MODE
, RX_MODE_RESET
);
6793 tp
->mac_mode
= MAC_MODE_TXSTAT_ENABLE
| MAC_MODE_RXSTAT_ENABLE
|
6794 MAC_MODE_TDE_ENABLE
| MAC_MODE_RDE_ENABLE
| MAC_MODE_FHDE_ENABLE
;
6795 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) &&
6796 !(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) &&
6797 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
)
6798 tp
->mac_mode
|= MAC_MODE_LINK_POLARITY
;
6799 tw32_f(MAC_MODE
, tp
->mac_mode
| MAC_MODE_RXSTAT_CLEAR
| MAC_MODE_TXSTAT_CLEAR
);
6802 /* tp->grc_local_ctrl is partially set up during tg3_get_invariants().
6803 * If TG3_FLG2_IS_NIC is zero, we should read the
6804 * register to preserve the GPIO settings for LOMs. The GPIOs,
6805 * whether used as inputs or outputs, are set by boot code after
6808 if (!(tp
->tg3_flags2
& TG3_FLG2_IS_NIC
)) {
6811 gpio_mask
= GRC_LCLCTRL_GPIO_OE0
| GRC_LCLCTRL_GPIO_OE1
|
6812 GRC_LCLCTRL_GPIO_OE2
| GRC_LCLCTRL_GPIO_OUTPUT0
|
6813 GRC_LCLCTRL_GPIO_OUTPUT1
| GRC_LCLCTRL_GPIO_OUTPUT2
;
6815 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
)
6816 gpio_mask
|= GRC_LCLCTRL_GPIO_OE3
|
6817 GRC_LCLCTRL_GPIO_OUTPUT3
;
6819 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
)
6820 gpio_mask
|= GRC_LCLCTRL_GPIO_UART_SEL
;
6822 tp
->grc_local_ctrl
&= ~gpio_mask
;
6823 tp
->grc_local_ctrl
|= tr32(GRC_LOCAL_CTRL
) & gpio_mask
;
6825 /* GPIO1 must be driven high for eeprom write protect */
6826 if (tp
->tg3_flags
& TG3_FLAG_EEPROM_WRITE_PROT
)
6827 tp
->grc_local_ctrl
|= (GRC_LCLCTRL_GPIO_OE1
|
6828 GRC_LCLCTRL_GPIO_OUTPUT1
);
6830 tw32_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
);
6833 tw32_mailbox_f(MAILBOX_INTERRUPT_0
+ TG3_64BIT_REG_LOW
, 0);
6836 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
6837 tw32_f(DMAC_MODE
, DMAC_MODE_ENABLE
);
6841 val
= (WDMAC_MODE_ENABLE
| WDMAC_MODE_TGTABORT_ENAB
|
6842 WDMAC_MODE_MSTABORT_ENAB
| WDMAC_MODE_PARITYERR_ENAB
|
6843 WDMAC_MODE_ADDROFLOW_ENAB
| WDMAC_MODE_FIFOOFLOW_ENAB
|
6844 WDMAC_MODE_FIFOURUN_ENAB
| WDMAC_MODE_FIFOOREAD_ENAB
|
6845 WDMAC_MODE_LNGREAD_ENAB
);
6847 /* If statement applies to 5705 and 5750 PCI devices only */
6848 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
&&
6849 tp
->pci_chip_rev_id
!= CHIPREV_ID_5705_A0
) ||
6850 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
) {
6851 if ((tp
->tg3_flags
& TG3_FLG2_TSO_CAPABLE
) &&
6852 (tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A1
||
6853 tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A2
)) {
6855 } else if (!(tr32(TG3PCI_PCISTATE
) & PCISTATE_BUS_SPEED_HIGH
) &&
6856 !(tp
->tg3_flags2
& TG3_FLG2_IS_5788
) &&
6857 !(tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
)) {
6858 val
|= WDMAC_MODE_RX_ACCEL
;
6862 /* Enable host coalescing bug fix */
6863 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
) ||
6864 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
) ||
6865 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
) ||
6866 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
))
6869 tw32_f(WDMAC_MODE
, val
);
6872 if (tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) {
6875 pci_read_config_word(tp
->pdev
, tp
->pcix_cap
+ PCI_X_CMD
,
6877 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
) {
6878 pcix_cmd
&= ~PCI_X_CMD_MAX_READ
;
6879 pcix_cmd
|= PCI_X_CMD_READ_2K
;
6880 } else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) {
6881 pcix_cmd
&= ~(PCI_X_CMD_MAX_SPLIT
| PCI_X_CMD_MAX_READ
);
6882 pcix_cmd
|= PCI_X_CMD_READ_2K
;
6884 pci_write_config_word(tp
->pdev
, tp
->pcix_cap
+ PCI_X_CMD
,
6888 tw32_f(RDMAC_MODE
, rdmac_mode
);
6891 tw32(RCVDCC_MODE
, RCVDCC_MODE_ENABLE
| RCVDCC_MODE_ATTN_ENABLE
);
6892 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
6893 tw32(MBFREE_MODE
, MBFREE_MODE_ENABLE
);
6895 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
6897 SNDDATAC_MODE_ENABLE
| SNDDATAC_MODE_CDELAY
);
6899 tw32(SNDDATAC_MODE
, SNDDATAC_MODE_ENABLE
);
6901 tw32(SNDBDC_MODE
, SNDBDC_MODE_ENABLE
| SNDBDC_MODE_ATTN_ENABLE
);
6902 tw32(RCVBDI_MODE
, RCVBDI_MODE_ENABLE
| RCVBDI_MODE_RCB_ATTN_ENAB
);
6903 tw32(RCVDBDI_MODE
, RCVDBDI_MODE_ENABLE
| RCVDBDI_MODE_INV_RING_SZ
);
6904 tw32(SNDDATAI_MODE
, SNDDATAI_MODE_ENABLE
);
6905 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
)
6906 tw32(SNDDATAI_MODE
, SNDDATAI_MODE_ENABLE
| 0x8);
6907 tw32(SNDBDI_MODE
, SNDBDI_MODE_ENABLE
| SNDBDI_MODE_ATTN_ENABLE
);
6908 tw32(SNDBDS_MODE
, SNDBDS_MODE_ENABLE
| SNDBDS_MODE_ATTN_ENABLE
);
6910 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
) {
6911 err
= tg3_load_5701_a0_firmware_fix(tp
);
6916 if (tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
) {
6917 err
= tg3_load_tso_firmware(tp
);
6922 tp
->tx_mode
= TX_MODE_ENABLE
;
6923 tw32_f(MAC_TX_MODE
, tp
->tx_mode
);
6926 tp
->rx_mode
= RX_MODE_ENABLE
;
6927 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
6928 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
6929 tp
->rx_mode
|= RX_MODE_IPV6_CSUM_ENABLE
;
6931 tw32_f(MAC_RX_MODE
, tp
->rx_mode
);
6934 if (tp
->link_config
.phy_is_low_power
) {
6935 tp
->link_config
.phy_is_low_power
= 0;
6936 tp
->link_config
.speed
= tp
->link_config
.orig_speed
;
6937 tp
->link_config
.duplex
= tp
->link_config
.orig_duplex
;
6938 tp
->link_config
.autoneg
= tp
->link_config
.orig_autoneg
;
6941 tp
->mi_mode
= MAC_MI_MODE_BASE
;
6942 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
6945 tw32(MAC_LED_CTRL
, tp
->led_ctrl
);
6947 tw32(MAC_MI_STAT
, MAC_MI_STAT_LNKSTAT_ATTN_ENAB
);
6948 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
6949 tw32_f(MAC_RX_MODE
, RX_MODE_RESET
);
6952 tw32_f(MAC_RX_MODE
, tp
->rx_mode
);
6955 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
6956 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) &&
6957 !(tp
->tg3_flags2
& TG3_FLG2_SERDES_PREEMPHASIS
)) {
6958 /* Set drive transmission level to 1.2V */
6959 /* only if the signal pre-emphasis bit is not set */
6960 val
= tr32(MAC_SERDES_CFG
);
6963 tw32(MAC_SERDES_CFG
, val
);
6965 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5703_A1
)
6966 tw32(MAC_SERDES_CFG
, 0x616000);
6969 /* Prevent chip from dropping frames when flow control
6972 tw32_f(MAC_LOW_WMARK_MAX_RX_FRAME
, 2);
6974 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
&&
6975 (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)) {
6976 /* Use hardware link auto-negotiation */
6977 tp
->tg3_flags2
|= TG3_FLG2_HW_AUTONEG
;
6980 if ((tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) &&
6981 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
)) {
6984 tmp
= tr32(SERDES_RX_CTRL
);
6985 tw32(SERDES_RX_CTRL
, tmp
| SERDES_RX_SIG_DETECT
);
6986 tp
->grc_local_ctrl
&= ~GRC_LCLCTRL_USE_EXT_SIG_DETECT
;
6987 tp
->grc_local_ctrl
|= GRC_LCLCTRL_USE_SIG_DETECT
;
6988 tw32(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
);
6991 err
= tg3_setup_phy(tp
, 0);
6995 if (!(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) &&
6996 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
) {
6999 /* Clear CRC stats. */
7000 if (!tg3_readphy(tp
, MII_TG3_TEST1
, &tmp
)) {
7001 tg3_writephy(tp
, MII_TG3_TEST1
,
7002 tmp
| MII_TG3_TEST1_CRC_EN
);
7003 tg3_readphy(tp
, 0x14, &tmp
);
7007 __tg3_set_rx_mode(tp
->dev
);
7009 /* Initialize receive rules. */
7010 tw32(MAC_RCV_RULE_0
, 0xc2000000 & RCV_RULE_DISABLE_MASK
);
7011 tw32(MAC_RCV_VALUE_0
, 0xffffffff & RCV_RULE_DISABLE_MASK
);
7012 tw32(MAC_RCV_RULE_1
, 0x86000004 & RCV_RULE_DISABLE_MASK
);
7013 tw32(MAC_RCV_VALUE_1
, 0xffffffff & RCV_RULE_DISABLE_MASK
);
7015 if ((tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) &&
7016 !(tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
))
7020 if (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
)
7024 tw32(MAC_RCV_RULE_15
, 0); tw32(MAC_RCV_VALUE_15
, 0);
7026 tw32(MAC_RCV_RULE_14
, 0); tw32(MAC_RCV_VALUE_14
, 0);
7028 tw32(MAC_RCV_RULE_13
, 0); tw32(MAC_RCV_VALUE_13
, 0);
7030 tw32(MAC_RCV_RULE_12
, 0); tw32(MAC_RCV_VALUE_12
, 0);
7032 tw32(MAC_RCV_RULE_11
, 0); tw32(MAC_RCV_VALUE_11
, 0);
7034 tw32(MAC_RCV_RULE_10
, 0); tw32(MAC_RCV_VALUE_10
, 0);
7036 tw32(MAC_RCV_RULE_9
, 0); tw32(MAC_RCV_VALUE_9
, 0);
7038 tw32(MAC_RCV_RULE_8
, 0); tw32(MAC_RCV_VALUE_8
, 0);
7040 tw32(MAC_RCV_RULE_7
, 0); tw32(MAC_RCV_VALUE_7
, 0);
7042 tw32(MAC_RCV_RULE_6
, 0); tw32(MAC_RCV_VALUE_6
, 0);
7044 tw32(MAC_RCV_RULE_5
, 0); tw32(MAC_RCV_VALUE_5
, 0);
7046 tw32(MAC_RCV_RULE_4
, 0); tw32(MAC_RCV_VALUE_4
, 0);
7048 /* tw32(MAC_RCV_RULE_3, 0); tw32(MAC_RCV_VALUE_3, 0); */
7050 /* tw32(MAC_RCV_RULE_2, 0); tw32(MAC_RCV_VALUE_2, 0); */
7058 if (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
)
7059 /* Write our heartbeat update interval to APE. */
7060 tg3_ape_write32(tp
, TG3_APE_HOST_HEARTBEAT_INT_MS
,
7061 APE_HOST_HEARTBEAT_INT_DISABLE
);
7063 tg3_write_sig_post_reset(tp
, RESET_KIND_INIT
);
7068 /* Called at device open time to get the chip ready for
7069 * packet processing. Invoked with tp->lock held.
7071 static int tg3_init_hw(struct tg3
*tp
, int reset_phy
)
7075 /* Force the chip into D0. */
7076 err
= tg3_set_power_state(tp
, PCI_D0
);
7080 tg3_switch_clocks(tp
);
7082 tw32(TG3PCI_MEM_WIN_BASE_ADDR
, 0);
7084 err
= tg3_reset_hw(tp
, reset_phy
);
7090 #define TG3_STAT_ADD32(PSTAT, REG) \
7091 do { u32 __val = tr32(REG); \
7092 (PSTAT)->low += __val; \
7093 if ((PSTAT)->low < __val) \
7094 (PSTAT)->high += 1; \
7097 static void tg3_periodic_fetch_stats(struct tg3
*tp
)
7099 struct tg3_hw_stats
*sp
= tp
->hw_stats
;
7101 if (!netif_carrier_ok(tp
->dev
))
7104 TG3_STAT_ADD32(&sp
->tx_octets
, MAC_TX_STATS_OCTETS
);
7105 TG3_STAT_ADD32(&sp
->tx_collisions
, MAC_TX_STATS_COLLISIONS
);
7106 TG3_STAT_ADD32(&sp
->tx_xon_sent
, MAC_TX_STATS_XON_SENT
);
7107 TG3_STAT_ADD32(&sp
->tx_xoff_sent
, MAC_TX_STATS_XOFF_SENT
);
7108 TG3_STAT_ADD32(&sp
->tx_mac_errors
, MAC_TX_STATS_MAC_ERRORS
);
7109 TG3_STAT_ADD32(&sp
->tx_single_collisions
, MAC_TX_STATS_SINGLE_COLLISIONS
);
7110 TG3_STAT_ADD32(&sp
->tx_mult_collisions
, MAC_TX_STATS_MULT_COLLISIONS
);
7111 TG3_STAT_ADD32(&sp
->tx_deferred
, MAC_TX_STATS_DEFERRED
);
7112 TG3_STAT_ADD32(&sp
->tx_excessive_collisions
, MAC_TX_STATS_EXCESSIVE_COL
);
7113 TG3_STAT_ADD32(&sp
->tx_late_collisions
, MAC_TX_STATS_LATE_COL
);
7114 TG3_STAT_ADD32(&sp
->tx_ucast_packets
, MAC_TX_STATS_UCAST
);
7115 TG3_STAT_ADD32(&sp
->tx_mcast_packets
, MAC_TX_STATS_MCAST
);
7116 TG3_STAT_ADD32(&sp
->tx_bcast_packets
, MAC_TX_STATS_BCAST
);
7118 TG3_STAT_ADD32(&sp
->rx_octets
, MAC_RX_STATS_OCTETS
);
7119 TG3_STAT_ADD32(&sp
->rx_fragments
, MAC_RX_STATS_FRAGMENTS
);
7120 TG3_STAT_ADD32(&sp
->rx_ucast_packets
, MAC_RX_STATS_UCAST
);
7121 TG3_STAT_ADD32(&sp
->rx_mcast_packets
, MAC_RX_STATS_MCAST
);
7122 TG3_STAT_ADD32(&sp
->rx_bcast_packets
, MAC_RX_STATS_BCAST
);
7123 TG3_STAT_ADD32(&sp
->rx_fcs_errors
, MAC_RX_STATS_FCS_ERRORS
);
7124 TG3_STAT_ADD32(&sp
->rx_align_errors
, MAC_RX_STATS_ALIGN_ERRORS
);
7125 TG3_STAT_ADD32(&sp
->rx_xon_pause_rcvd
, MAC_RX_STATS_XON_PAUSE_RECVD
);
7126 TG3_STAT_ADD32(&sp
->rx_xoff_pause_rcvd
, MAC_RX_STATS_XOFF_PAUSE_RECVD
);
7127 TG3_STAT_ADD32(&sp
->rx_mac_ctrl_rcvd
, MAC_RX_STATS_MAC_CTRL_RECVD
);
7128 TG3_STAT_ADD32(&sp
->rx_xoff_entered
, MAC_RX_STATS_XOFF_ENTERED
);
7129 TG3_STAT_ADD32(&sp
->rx_frame_too_long_errors
, MAC_RX_STATS_FRAME_TOO_LONG
);
7130 TG3_STAT_ADD32(&sp
->rx_jabbers
, MAC_RX_STATS_JABBERS
);
7131 TG3_STAT_ADD32(&sp
->rx_undersize_packets
, MAC_RX_STATS_UNDERSIZE
);
7133 TG3_STAT_ADD32(&sp
->rxbds_empty
, RCVLPC_NO_RCV_BD_CNT
);
7134 TG3_STAT_ADD32(&sp
->rx_discards
, RCVLPC_IN_DISCARDS_CNT
);
7135 TG3_STAT_ADD32(&sp
->rx_errors
, RCVLPC_IN_ERRORS_CNT
);
7138 static void tg3_timer(unsigned long __opaque
)
7140 struct tg3
*tp
= (struct tg3
*) __opaque
;
7145 spin_lock(&tp
->lock
);
7147 if (!(tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
)) {
7148 /* All of this garbage is because when using non-tagged
7149 * IRQ status the mailbox/status_block protocol the chip
7150 * uses with the cpu is race prone.
7152 if (tp
->hw_status
->status
& SD_STATUS_UPDATED
) {
7153 tw32(GRC_LOCAL_CTRL
,
7154 tp
->grc_local_ctrl
| GRC_LCLCTRL_SETINT
);
7156 tw32(HOSTCC_MODE
, tp
->coalesce_mode
|
7157 (HOSTCC_MODE_ENABLE
| HOSTCC_MODE_NOW
));
7160 if (!(tr32(WDMAC_MODE
) & WDMAC_MODE_ENABLE
)) {
7161 tp
->tg3_flags2
|= TG3_FLG2_RESTART_TIMER
;
7162 spin_unlock(&tp
->lock
);
7163 schedule_work(&tp
->reset_task
);
7168 /* This part only runs once per second. */
7169 if (!--tp
->timer_counter
) {
7170 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)
7171 tg3_periodic_fetch_stats(tp
);
7173 if (tp
->tg3_flags
& TG3_FLAG_USE_LINKCHG_REG
) {
7177 mac_stat
= tr32(MAC_STATUS
);
7180 if (tp
->tg3_flags
& TG3_FLAG_USE_MI_INTERRUPT
) {
7181 if (mac_stat
& MAC_STATUS_MI_INTERRUPT
)
7183 } else if (mac_stat
& MAC_STATUS_LNKSTATE_CHANGED
)
7187 tg3_setup_phy(tp
, 0);
7188 } else if (tp
->tg3_flags
& TG3_FLAG_POLL_SERDES
) {
7189 u32 mac_stat
= tr32(MAC_STATUS
);
7192 if (netif_carrier_ok(tp
->dev
) &&
7193 (mac_stat
& MAC_STATUS_LNKSTATE_CHANGED
)) {
7196 if (! netif_carrier_ok(tp
->dev
) &&
7197 (mac_stat
& (MAC_STATUS_PCS_SYNCED
|
7198 MAC_STATUS_SIGNAL_DET
))) {
7202 if (!tp
->serdes_counter
) {
7205 ~MAC_MODE_PORT_MODE_MASK
));
7207 tw32_f(MAC_MODE
, tp
->mac_mode
);
7210 tg3_setup_phy(tp
, 0);
7212 } else if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
)
7213 tg3_serdes_parallel_detect(tp
);
7215 tp
->timer_counter
= tp
->timer_multiplier
;
7218 /* Heartbeat is only sent once every 2 seconds.
7220 * The heartbeat is to tell the ASF firmware that the host
7221 * driver is still alive. In the event that the OS crashes,
7222 * ASF needs to reset the hardware to free up the FIFO space
7223 * that may be filled with rx packets destined for the host.
7224 * If the FIFO is full, ASF will no longer function properly.
7226 * Unintended resets have been reported on real time kernels
7227 * where the timer doesn't run on time. Netpoll will also have
7230 * The new FWCMD_NICDRV_ALIVE3 command tells the ASF firmware
7231 * to check the ring condition when the heartbeat is expiring
7232 * before doing the reset. This will prevent most unintended
7235 if (!--tp
->asf_counter
) {
7236 if (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) {
7239 tg3_write_mem(tp
, NIC_SRAM_FW_CMD_MBOX
,
7240 FWCMD_NICDRV_ALIVE3
);
7241 tg3_write_mem(tp
, NIC_SRAM_FW_CMD_LEN_MBOX
, 4);
7242 /* 5 seconds timeout */
7243 tg3_write_mem(tp
, NIC_SRAM_FW_CMD_DATA_MBOX
, 5);
7244 val
= tr32(GRC_RX_CPU_EVENT
);
7246 tw32(GRC_RX_CPU_EVENT
, val
);
7248 tp
->asf_counter
= tp
->asf_multiplier
;
7251 spin_unlock(&tp
->lock
);
7254 tp
->timer
.expires
= jiffies
+ tp
->timer_offset
;
7255 add_timer(&tp
->timer
);
7258 static int tg3_request_irq(struct tg3
*tp
)
7261 unsigned long flags
;
7262 struct net_device
*dev
= tp
->dev
;
7264 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7266 if (tp
->tg3_flags2
& TG3_FLG2_1SHOT_MSI
)
7268 flags
= IRQF_SAMPLE_RANDOM
;
7271 if (tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
)
7272 fn
= tg3_interrupt_tagged
;
7273 flags
= IRQF_SHARED
| IRQF_SAMPLE_RANDOM
;
7275 return (request_irq(tp
->pdev
->irq
, fn
, flags
, dev
->name
, dev
));
7278 static int tg3_test_interrupt(struct tg3
*tp
)
7280 struct net_device
*dev
= tp
->dev
;
7281 int err
, i
, intr_ok
= 0;
7283 if (!netif_running(dev
))
7286 tg3_disable_ints(tp
);
7288 free_irq(tp
->pdev
->irq
, dev
);
7290 err
= request_irq(tp
->pdev
->irq
, tg3_test_isr
,
7291 IRQF_SHARED
| IRQF_SAMPLE_RANDOM
, dev
->name
, dev
);
7295 tp
->hw_status
->status
&= ~SD_STATUS_UPDATED
;
7296 tg3_enable_ints(tp
);
7298 tw32_f(HOSTCC_MODE
, tp
->coalesce_mode
| HOSTCC_MODE_ENABLE
|
7301 for (i
= 0; i
< 5; i
++) {
7302 u32 int_mbox
, misc_host_ctrl
;
7304 int_mbox
= tr32_mailbox(MAILBOX_INTERRUPT_0
+
7306 misc_host_ctrl
= tr32(TG3PCI_MISC_HOST_CTRL
);
7308 if ((int_mbox
!= 0) ||
7309 (misc_host_ctrl
& MISC_HOST_CTRL_MASK_PCI_INT
)) {
7317 tg3_disable_ints(tp
);
7319 free_irq(tp
->pdev
->irq
, dev
);
7321 err
= tg3_request_irq(tp
);
7332 /* Returns 0 if MSI test succeeds or MSI test fails and INTx mode is
7333 * successfully restored
7335 static int tg3_test_msi(struct tg3
*tp
)
7337 struct net_device
*dev
= tp
->dev
;
7341 if (!(tp
->tg3_flags2
& TG3_FLG2_USING_MSI
))
7344 /* Turn off SERR reporting in case MSI terminates with Master
7347 pci_read_config_word(tp
->pdev
, PCI_COMMAND
, &pci_cmd
);
7348 pci_write_config_word(tp
->pdev
, PCI_COMMAND
,
7349 pci_cmd
& ~PCI_COMMAND_SERR
);
7351 err
= tg3_test_interrupt(tp
);
7353 pci_write_config_word(tp
->pdev
, PCI_COMMAND
, pci_cmd
);
7358 /* other failures */
7362 /* MSI test failed, go back to INTx mode */
7363 printk(KERN_WARNING PFX
"%s: No interrupt was generated using MSI, "
7364 "switching to INTx mode. Please report this failure to "
7365 "the PCI maintainer and include system chipset information.\n",
7368 free_irq(tp
->pdev
->irq
, dev
);
7369 pci_disable_msi(tp
->pdev
);
7371 tp
->tg3_flags2
&= ~TG3_FLG2_USING_MSI
;
7373 err
= tg3_request_irq(tp
);
7377 /* Need to reset the chip because the MSI cycle may have terminated
7378 * with Master Abort.
7380 tg3_full_lock(tp
, 1);
7382 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
7383 err
= tg3_init_hw(tp
, 1);
7385 tg3_full_unlock(tp
);
7388 free_irq(tp
->pdev
->irq
, dev
);
7393 static int tg3_open(struct net_device
*dev
)
7395 struct tg3
*tp
= netdev_priv(dev
);
7398 netif_carrier_off(tp
->dev
);
7400 tg3_full_lock(tp
, 0);
7402 err
= tg3_set_power_state(tp
, PCI_D0
);
7404 tg3_full_unlock(tp
);
7408 tg3_disable_ints(tp
);
7409 tp
->tg3_flags
&= ~TG3_FLAG_INIT_COMPLETE
;
7411 tg3_full_unlock(tp
);
7413 /* The placement of this call is tied
7414 * to the setup and use of Host TX descriptors.
7416 err
= tg3_alloc_consistent(tp
);
7420 if (tp
->tg3_flags
& TG3_FLAG_SUPPORT_MSI
) {
7421 /* All MSI supporting chips should support tagged
7422 * status. Assert that this is the case.
7424 if (!(tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
)) {
7425 printk(KERN_WARNING PFX
"%s: MSI without TAGGED? "
7426 "Not using MSI.\n", tp
->dev
->name
);
7427 } else if (pci_enable_msi(tp
->pdev
) == 0) {
7430 msi_mode
= tr32(MSGINT_MODE
);
7431 tw32(MSGINT_MODE
, msi_mode
| MSGINT_MODE_ENABLE
);
7432 tp
->tg3_flags2
|= TG3_FLG2_USING_MSI
;
7435 err
= tg3_request_irq(tp
);
7438 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7439 pci_disable_msi(tp
->pdev
);
7440 tp
->tg3_flags2
&= ~TG3_FLG2_USING_MSI
;
7442 tg3_free_consistent(tp
);
7446 napi_enable(&tp
->napi
);
7448 tg3_full_lock(tp
, 0);
7450 err
= tg3_init_hw(tp
, 1);
7452 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
7455 if (tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
)
7456 tp
->timer_offset
= HZ
;
7458 tp
->timer_offset
= HZ
/ 10;
7460 BUG_ON(tp
->timer_offset
> HZ
);
7461 tp
->timer_counter
= tp
->timer_multiplier
=
7462 (HZ
/ tp
->timer_offset
);
7463 tp
->asf_counter
= tp
->asf_multiplier
=
7464 ((HZ
/ tp
->timer_offset
) * 2);
7466 init_timer(&tp
->timer
);
7467 tp
->timer
.expires
= jiffies
+ tp
->timer_offset
;
7468 tp
->timer
.data
= (unsigned long) tp
;
7469 tp
->timer
.function
= tg3_timer
;
7472 tg3_full_unlock(tp
);
7475 napi_disable(&tp
->napi
);
7476 free_irq(tp
->pdev
->irq
, dev
);
7477 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7478 pci_disable_msi(tp
->pdev
);
7479 tp
->tg3_flags2
&= ~TG3_FLG2_USING_MSI
;
7481 tg3_free_consistent(tp
);
7485 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7486 err
= tg3_test_msi(tp
);
7489 tg3_full_lock(tp
, 0);
7491 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7492 pci_disable_msi(tp
->pdev
);
7493 tp
->tg3_flags2
&= ~TG3_FLG2_USING_MSI
;
7495 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
7497 tg3_free_consistent(tp
);
7499 tg3_full_unlock(tp
);
7501 napi_disable(&tp
->napi
);
7506 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7507 if (tp
->tg3_flags2
& TG3_FLG2_1SHOT_MSI
) {
7508 u32 val
= tr32(PCIE_TRANSACTION_CFG
);
7510 tw32(PCIE_TRANSACTION_CFG
,
7511 val
| PCIE_TRANS_CFG_1SHOT_MSI
);
7516 tg3_full_lock(tp
, 0);
7518 add_timer(&tp
->timer
);
7519 tp
->tg3_flags
|= TG3_FLAG_INIT_COMPLETE
;
7520 tg3_enable_ints(tp
);
7522 tg3_full_unlock(tp
);
7524 netif_start_queue(dev
);
7530 /*static*/ void tg3_dump_state(struct tg3
*tp
)
7532 u32 val32
, val32_2
, val32_3
, val32_4
, val32_5
;
7536 pci_read_config_word(tp
->pdev
, PCI_STATUS
, &val16
);
7537 pci_read_config_dword(tp
->pdev
, TG3PCI_PCISTATE
, &val32
);
7538 printk("DEBUG: PCI status [%04x] TG3PCI state[%08x]\n",
7542 printk("DEBUG: MAC_MODE[%08x] MAC_STATUS[%08x]\n",
7543 tr32(MAC_MODE
), tr32(MAC_STATUS
));
7544 printk(" MAC_EVENT[%08x] MAC_LED_CTRL[%08x]\n",
7545 tr32(MAC_EVENT
), tr32(MAC_LED_CTRL
));
7546 printk("DEBUG: MAC_TX_MODE[%08x] MAC_TX_STATUS[%08x]\n",
7547 tr32(MAC_TX_MODE
), tr32(MAC_TX_STATUS
));
7548 printk(" MAC_RX_MODE[%08x] MAC_RX_STATUS[%08x]\n",
7549 tr32(MAC_RX_MODE
), tr32(MAC_RX_STATUS
));
7551 /* Send data initiator control block */
7552 printk("DEBUG: SNDDATAI_MODE[%08x] SNDDATAI_STATUS[%08x]\n",
7553 tr32(SNDDATAI_MODE
), tr32(SNDDATAI_STATUS
));
7554 printk(" SNDDATAI_STATSCTRL[%08x]\n",
7555 tr32(SNDDATAI_STATSCTRL
));
7557 /* Send data completion control block */
7558 printk("DEBUG: SNDDATAC_MODE[%08x]\n", tr32(SNDDATAC_MODE
));
7560 /* Send BD ring selector block */
7561 printk("DEBUG: SNDBDS_MODE[%08x] SNDBDS_STATUS[%08x]\n",
7562 tr32(SNDBDS_MODE
), tr32(SNDBDS_STATUS
));
7564 /* Send BD initiator control block */
7565 printk("DEBUG: SNDBDI_MODE[%08x] SNDBDI_STATUS[%08x]\n",
7566 tr32(SNDBDI_MODE
), tr32(SNDBDI_STATUS
));
7568 /* Send BD completion control block */
7569 printk("DEBUG: SNDBDC_MODE[%08x]\n", tr32(SNDBDC_MODE
));
7571 /* Receive list placement control block */
7572 printk("DEBUG: RCVLPC_MODE[%08x] RCVLPC_STATUS[%08x]\n",
7573 tr32(RCVLPC_MODE
), tr32(RCVLPC_STATUS
));
7574 printk(" RCVLPC_STATSCTRL[%08x]\n",
7575 tr32(RCVLPC_STATSCTRL
));
7577 /* Receive data and receive BD initiator control block */
7578 printk("DEBUG: RCVDBDI_MODE[%08x] RCVDBDI_STATUS[%08x]\n",
7579 tr32(RCVDBDI_MODE
), tr32(RCVDBDI_STATUS
));
7581 /* Receive data completion control block */
7582 printk("DEBUG: RCVDCC_MODE[%08x]\n",
7585 /* Receive BD initiator control block */
7586 printk("DEBUG: RCVBDI_MODE[%08x] RCVBDI_STATUS[%08x]\n",
7587 tr32(RCVBDI_MODE
), tr32(RCVBDI_STATUS
));
7589 /* Receive BD completion control block */
7590 printk("DEBUG: RCVCC_MODE[%08x] RCVCC_STATUS[%08x]\n",
7591 tr32(RCVCC_MODE
), tr32(RCVCC_STATUS
));
7593 /* Receive list selector control block */
7594 printk("DEBUG: RCVLSC_MODE[%08x] RCVLSC_STATUS[%08x]\n",
7595 tr32(RCVLSC_MODE
), tr32(RCVLSC_STATUS
));
7597 /* Mbuf cluster free block */
7598 printk("DEBUG: MBFREE_MODE[%08x] MBFREE_STATUS[%08x]\n",
7599 tr32(MBFREE_MODE
), tr32(MBFREE_STATUS
));
7601 /* Host coalescing control block */
7602 printk("DEBUG: HOSTCC_MODE[%08x] HOSTCC_STATUS[%08x]\n",
7603 tr32(HOSTCC_MODE
), tr32(HOSTCC_STATUS
));
7604 printk("DEBUG: HOSTCC_STATS_BLK_HOST_ADDR[%08x%08x]\n",
7605 tr32(HOSTCC_STATS_BLK_HOST_ADDR
+ TG3_64BIT_REG_HIGH
),
7606 tr32(HOSTCC_STATS_BLK_HOST_ADDR
+ TG3_64BIT_REG_LOW
));
7607 printk("DEBUG: HOSTCC_STATUS_BLK_HOST_ADDR[%08x%08x]\n",
7608 tr32(HOSTCC_STATUS_BLK_HOST_ADDR
+ TG3_64BIT_REG_HIGH
),
7609 tr32(HOSTCC_STATUS_BLK_HOST_ADDR
+ TG3_64BIT_REG_LOW
));
7610 printk("DEBUG: HOSTCC_STATS_BLK_NIC_ADDR[%08x]\n",
7611 tr32(HOSTCC_STATS_BLK_NIC_ADDR
));
7612 printk("DEBUG: HOSTCC_STATUS_BLK_NIC_ADDR[%08x]\n",
7613 tr32(HOSTCC_STATUS_BLK_NIC_ADDR
));
7615 /* Memory arbiter control block */
7616 printk("DEBUG: MEMARB_MODE[%08x] MEMARB_STATUS[%08x]\n",
7617 tr32(MEMARB_MODE
), tr32(MEMARB_STATUS
));
7619 /* Buffer manager control block */
7620 printk("DEBUG: BUFMGR_MODE[%08x] BUFMGR_STATUS[%08x]\n",
7621 tr32(BUFMGR_MODE
), tr32(BUFMGR_STATUS
));
7622 printk("DEBUG: BUFMGR_MB_POOL_ADDR[%08x] BUFMGR_MB_POOL_SIZE[%08x]\n",
7623 tr32(BUFMGR_MB_POOL_ADDR
), tr32(BUFMGR_MB_POOL_SIZE
));
7624 printk("DEBUG: BUFMGR_DMA_DESC_POOL_ADDR[%08x] "
7625 "BUFMGR_DMA_DESC_POOL_SIZE[%08x]\n",
7626 tr32(BUFMGR_DMA_DESC_POOL_ADDR
),
7627 tr32(BUFMGR_DMA_DESC_POOL_SIZE
));
7629 /* Read DMA control block */
7630 printk("DEBUG: RDMAC_MODE[%08x] RDMAC_STATUS[%08x]\n",
7631 tr32(RDMAC_MODE
), tr32(RDMAC_STATUS
));
7633 /* Write DMA control block */
7634 printk("DEBUG: WDMAC_MODE[%08x] WDMAC_STATUS[%08x]\n",
7635 tr32(WDMAC_MODE
), tr32(WDMAC_STATUS
));
7637 /* DMA completion block */
7638 printk("DEBUG: DMAC_MODE[%08x]\n",
7642 printk("DEBUG: GRC_MODE[%08x] GRC_MISC_CFG[%08x]\n",
7643 tr32(GRC_MODE
), tr32(GRC_MISC_CFG
));
7644 printk("DEBUG: GRC_LOCAL_CTRL[%08x]\n",
7645 tr32(GRC_LOCAL_CTRL
));
7648 printk("DEBUG: RCVDBDI_JUMBO_BD[%08x%08x:%08x:%08x]\n",
7649 tr32(RCVDBDI_JUMBO_BD
+ 0x0),
7650 tr32(RCVDBDI_JUMBO_BD
+ 0x4),
7651 tr32(RCVDBDI_JUMBO_BD
+ 0x8),
7652 tr32(RCVDBDI_JUMBO_BD
+ 0xc));
7653 printk("DEBUG: RCVDBDI_STD_BD[%08x%08x:%08x:%08x]\n",
7654 tr32(RCVDBDI_STD_BD
+ 0x0),
7655 tr32(RCVDBDI_STD_BD
+ 0x4),
7656 tr32(RCVDBDI_STD_BD
+ 0x8),
7657 tr32(RCVDBDI_STD_BD
+ 0xc));
7658 printk("DEBUG: RCVDBDI_MINI_BD[%08x%08x:%08x:%08x]\n",
7659 tr32(RCVDBDI_MINI_BD
+ 0x0),
7660 tr32(RCVDBDI_MINI_BD
+ 0x4),
7661 tr32(RCVDBDI_MINI_BD
+ 0x8),
7662 tr32(RCVDBDI_MINI_BD
+ 0xc));
7664 tg3_read_mem(tp
, NIC_SRAM_SEND_RCB
+ 0x0, &val32
);
7665 tg3_read_mem(tp
, NIC_SRAM_SEND_RCB
+ 0x4, &val32_2
);
7666 tg3_read_mem(tp
, NIC_SRAM_SEND_RCB
+ 0x8, &val32_3
);
7667 tg3_read_mem(tp
, NIC_SRAM_SEND_RCB
+ 0xc, &val32_4
);
7668 printk("DEBUG: SRAM_SEND_RCB_0[%08x%08x:%08x:%08x]\n",
7669 val32
, val32_2
, val32_3
, val32_4
);
7671 tg3_read_mem(tp
, NIC_SRAM_RCV_RET_RCB
+ 0x0, &val32
);
7672 tg3_read_mem(tp
, NIC_SRAM_RCV_RET_RCB
+ 0x4, &val32_2
);
7673 tg3_read_mem(tp
, NIC_SRAM_RCV_RET_RCB
+ 0x8, &val32_3
);
7674 tg3_read_mem(tp
, NIC_SRAM_RCV_RET_RCB
+ 0xc, &val32_4
);
7675 printk("DEBUG: SRAM_RCV_RET_RCB_0[%08x%08x:%08x:%08x]\n",
7676 val32
, val32_2
, val32_3
, val32_4
);
7678 tg3_read_mem(tp
, NIC_SRAM_STATUS_BLK
+ 0x0, &val32
);
7679 tg3_read_mem(tp
, NIC_SRAM_STATUS_BLK
+ 0x4, &val32_2
);
7680 tg3_read_mem(tp
, NIC_SRAM_STATUS_BLK
+ 0x8, &val32_3
);
7681 tg3_read_mem(tp
, NIC_SRAM_STATUS_BLK
+ 0xc, &val32_4
);
7682 tg3_read_mem(tp
, NIC_SRAM_STATUS_BLK
+ 0x10, &val32_5
);
7683 printk("DEBUG: SRAM_STATUS_BLK[%08x:%08x:%08x:%08x:%08x]\n",
7684 val32
, val32_2
, val32_3
, val32_4
, val32_5
);
7686 /* SW status block */
7687 printk("DEBUG: Host status block [%08x:%08x:(%04x:%04x:%04x):(%04x:%04x)]\n",
7688 tp
->hw_status
->status
,
7689 tp
->hw_status
->status_tag
,
7690 tp
->hw_status
->rx_jumbo_consumer
,
7691 tp
->hw_status
->rx_consumer
,
7692 tp
->hw_status
->rx_mini_consumer
,
7693 tp
->hw_status
->idx
[0].rx_producer
,
7694 tp
->hw_status
->idx
[0].tx_consumer
);
7696 /* SW statistics block */
7697 printk("DEBUG: Host statistics block [%08x:%08x:%08x:%08x]\n",
7698 ((u32
*)tp
->hw_stats
)[0],
7699 ((u32
*)tp
->hw_stats
)[1],
7700 ((u32
*)tp
->hw_stats
)[2],
7701 ((u32
*)tp
->hw_stats
)[3]);
7704 printk("DEBUG: SNDHOST_PROD[%08x%08x] SNDNIC_PROD[%08x%08x]\n",
7705 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0
+ 0x0),
7706 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0
+ 0x4),
7707 tr32_mailbox(MAILBOX_SNDNIC_PROD_IDX_0
+ 0x0),
7708 tr32_mailbox(MAILBOX_SNDNIC_PROD_IDX_0
+ 0x4));
7710 /* NIC side send descriptors. */
7711 for (i
= 0; i
< 6; i
++) {
7714 txd
= tp
->regs
+ NIC_SRAM_WIN_BASE
+ NIC_SRAM_TX_BUFFER_DESC
7715 + (i
* sizeof(struct tg3_tx_buffer_desc
));
7716 printk("DEBUG: NIC TXD(%d)[%08x:%08x:%08x:%08x]\n",
7718 readl(txd
+ 0x0), readl(txd
+ 0x4),
7719 readl(txd
+ 0x8), readl(txd
+ 0xc));
7722 /* NIC side RX descriptors. */
7723 for (i
= 0; i
< 6; i
++) {
7726 rxd
= tp
->regs
+ NIC_SRAM_WIN_BASE
+ NIC_SRAM_RX_BUFFER_DESC
7727 + (i
* sizeof(struct tg3_rx_buffer_desc
));
7728 printk("DEBUG: NIC RXD_STD(%d)[0][%08x:%08x:%08x:%08x]\n",
7730 readl(rxd
+ 0x0), readl(rxd
+ 0x4),
7731 readl(rxd
+ 0x8), readl(rxd
+ 0xc));
7732 rxd
+= (4 * sizeof(u32
));
7733 printk("DEBUG: NIC RXD_STD(%d)[1][%08x:%08x:%08x:%08x]\n",
7735 readl(rxd
+ 0x0), readl(rxd
+ 0x4),
7736 readl(rxd
+ 0x8), readl(rxd
+ 0xc));
7739 for (i
= 0; i
< 6; i
++) {
7742 rxd
= tp
->regs
+ NIC_SRAM_WIN_BASE
+ NIC_SRAM_RX_JUMBO_BUFFER_DESC
7743 + (i
* sizeof(struct tg3_rx_buffer_desc
));
7744 printk("DEBUG: NIC RXD_JUMBO(%d)[0][%08x:%08x:%08x:%08x]\n",
7746 readl(rxd
+ 0x0), readl(rxd
+ 0x4),
7747 readl(rxd
+ 0x8), readl(rxd
+ 0xc));
7748 rxd
+= (4 * sizeof(u32
));
7749 printk("DEBUG: NIC RXD_JUMBO(%d)[1][%08x:%08x:%08x:%08x]\n",
7751 readl(rxd
+ 0x0), readl(rxd
+ 0x4),
7752 readl(rxd
+ 0x8), readl(rxd
+ 0xc));
7757 static struct net_device_stats
*tg3_get_stats(struct net_device
*);
7758 static struct tg3_ethtool_stats
*tg3_get_estats(struct tg3
*);
7760 static int tg3_close(struct net_device
*dev
)
7762 struct tg3
*tp
= netdev_priv(dev
);
7764 napi_disable(&tp
->napi
);
7765 cancel_work_sync(&tp
->reset_task
);
7767 netif_stop_queue(dev
);
7769 del_timer_sync(&tp
->timer
);
7771 tg3_full_lock(tp
, 1);
7776 tg3_disable_ints(tp
);
7778 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
7780 tp
->tg3_flags
&= ~TG3_FLAG_INIT_COMPLETE
;
7782 tg3_full_unlock(tp
);
7784 free_irq(tp
->pdev
->irq
, dev
);
7785 if (tp
->tg3_flags2
& TG3_FLG2_USING_MSI
) {
7786 pci_disable_msi(tp
->pdev
);
7787 tp
->tg3_flags2
&= ~TG3_FLG2_USING_MSI
;
7790 memcpy(&tp
->net_stats_prev
, tg3_get_stats(tp
->dev
),
7791 sizeof(tp
->net_stats_prev
));
7792 memcpy(&tp
->estats_prev
, tg3_get_estats(tp
),
7793 sizeof(tp
->estats_prev
));
7795 tg3_free_consistent(tp
);
7797 tg3_set_power_state(tp
, PCI_D3hot
);
7799 netif_carrier_off(tp
->dev
);
7804 static inline unsigned long get_stat64(tg3_stat64_t
*val
)
7808 #if (BITS_PER_LONG == 32)
7811 ret
= ((u64
)val
->high
<< 32) | ((u64
)val
->low
);
7816 static unsigned long calc_crc_errors(struct tg3
*tp
)
7818 struct tg3_hw_stats
*hw_stats
= tp
->hw_stats
;
7820 if (!(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) &&
7821 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
7822 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
)) {
7825 spin_lock_bh(&tp
->lock
);
7826 if (!tg3_readphy(tp
, MII_TG3_TEST1
, &val
)) {
7827 tg3_writephy(tp
, MII_TG3_TEST1
,
7828 val
| MII_TG3_TEST1_CRC_EN
);
7829 tg3_readphy(tp
, 0x14, &val
);
7832 spin_unlock_bh(&tp
->lock
);
7834 tp
->phy_crc_errors
+= val
;
7836 return tp
->phy_crc_errors
;
7839 return get_stat64(&hw_stats
->rx_fcs_errors
);
7842 #define ESTAT_ADD(member) \
7843 estats->member = old_estats->member + \
7844 get_stat64(&hw_stats->member)
7846 static struct tg3_ethtool_stats
*tg3_get_estats(struct tg3
*tp
)
7848 struct tg3_ethtool_stats
*estats
= &tp
->estats
;
7849 struct tg3_ethtool_stats
*old_estats
= &tp
->estats_prev
;
7850 struct tg3_hw_stats
*hw_stats
= tp
->hw_stats
;
7855 ESTAT_ADD(rx_octets
);
7856 ESTAT_ADD(rx_fragments
);
7857 ESTAT_ADD(rx_ucast_packets
);
7858 ESTAT_ADD(rx_mcast_packets
);
7859 ESTAT_ADD(rx_bcast_packets
);
7860 ESTAT_ADD(rx_fcs_errors
);
7861 ESTAT_ADD(rx_align_errors
);
7862 ESTAT_ADD(rx_xon_pause_rcvd
);
7863 ESTAT_ADD(rx_xoff_pause_rcvd
);
7864 ESTAT_ADD(rx_mac_ctrl_rcvd
);
7865 ESTAT_ADD(rx_xoff_entered
);
7866 ESTAT_ADD(rx_frame_too_long_errors
);
7867 ESTAT_ADD(rx_jabbers
);
7868 ESTAT_ADD(rx_undersize_packets
);
7869 ESTAT_ADD(rx_in_length_errors
);
7870 ESTAT_ADD(rx_out_length_errors
);
7871 ESTAT_ADD(rx_64_or_less_octet_packets
);
7872 ESTAT_ADD(rx_65_to_127_octet_packets
);
7873 ESTAT_ADD(rx_128_to_255_octet_packets
);
7874 ESTAT_ADD(rx_256_to_511_octet_packets
);
7875 ESTAT_ADD(rx_512_to_1023_octet_packets
);
7876 ESTAT_ADD(rx_1024_to_1522_octet_packets
);
7877 ESTAT_ADD(rx_1523_to_2047_octet_packets
);
7878 ESTAT_ADD(rx_2048_to_4095_octet_packets
);
7879 ESTAT_ADD(rx_4096_to_8191_octet_packets
);
7880 ESTAT_ADD(rx_8192_to_9022_octet_packets
);
7882 ESTAT_ADD(tx_octets
);
7883 ESTAT_ADD(tx_collisions
);
7884 ESTAT_ADD(tx_xon_sent
);
7885 ESTAT_ADD(tx_xoff_sent
);
7886 ESTAT_ADD(tx_flow_control
);
7887 ESTAT_ADD(tx_mac_errors
);
7888 ESTAT_ADD(tx_single_collisions
);
7889 ESTAT_ADD(tx_mult_collisions
);
7890 ESTAT_ADD(tx_deferred
);
7891 ESTAT_ADD(tx_excessive_collisions
);
7892 ESTAT_ADD(tx_late_collisions
);
7893 ESTAT_ADD(tx_collide_2times
);
7894 ESTAT_ADD(tx_collide_3times
);
7895 ESTAT_ADD(tx_collide_4times
);
7896 ESTAT_ADD(tx_collide_5times
);
7897 ESTAT_ADD(tx_collide_6times
);
7898 ESTAT_ADD(tx_collide_7times
);
7899 ESTAT_ADD(tx_collide_8times
);
7900 ESTAT_ADD(tx_collide_9times
);
7901 ESTAT_ADD(tx_collide_10times
);
7902 ESTAT_ADD(tx_collide_11times
);
7903 ESTAT_ADD(tx_collide_12times
);
7904 ESTAT_ADD(tx_collide_13times
);
7905 ESTAT_ADD(tx_collide_14times
);
7906 ESTAT_ADD(tx_collide_15times
);
7907 ESTAT_ADD(tx_ucast_packets
);
7908 ESTAT_ADD(tx_mcast_packets
);
7909 ESTAT_ADD(tx_bcast_packets
);
7910 ESTAT_ADD(tx_carrier_sense_errors
);
7911 ESTAT_ADD(tx_discards
);
7912 ESTAT_ADD(tx_errors
);
7914 ESTAT_ADD(dma_writeq_full
);
7915 ESTAT_ADD(dma_write_prioq_full
);
7916 ESTAT_ADD(rxbds_empty
);
7917 ESTAT_ADD(rx_discards
);
7918 ESTAT_ADD(rx_errors
);
7919 ESTAT_ADD(rx_threshold_hit
);
7921 ESTAT_ADD(dma_readq_full
);
7922 ESTAT_ADD(dma_read_prioq_full
);
7923 ESTAT_ADD(tx_comp_queue_full
);
7925 ESTAT_ADD(ring_set_send_prod_index
);
7926 ESTAT_ADD(ring_status_update
);
7927 ESTAT_ADD(nic_irqs
);
7928 ESTAT_ADD(nic_avoided_irqs
);
7929 ESTAT_ADD(nic_tx_threshold_hit
);
7934 static struct net_device_stats
*tg3_get_stats(struct net_device
*dev
)
7936 struct tg3
*tp
= netdev_priv(dev
);
7937 struct net_device_stats
*stats
= &tp
->net_stats
;
7938 struct net_device_stats
*old_stats
= &tp
->net_stats_prev
;
7939 struct tg3_hw_stats
*hw_stats
= tp
->hw_stats
;
7944 stats
->rx_packets
= old_stats
->rx_packets
+
7945 get_stat64(&hw_stats
->rx_ucast_packets
) +
7946 get_stat64(&hw_stats
->rx_mcast_packets
) +
7947 get_stat64(&hw_stats
->rx_bcast_packets
);
7949 stats
->tx_packets
= old_stats
->tx_packets
+
7950 get_stat64(&hw_stats
->tx_ucast_packets
) +
7951 get_stat64(&hw_stats
->tx_mcast_packets
) +
7952 get_stat64(&hw_stats
->tx_bcast_packets
);
7954 stats
->rx_bytes
= old_stats
->rx_bytes
+
7955 get_stat64(&hw_stats
->rx_octets
);
7956 stats
->tx_bytes
= old_stats
->tx_bytes
+
7957 get_stat64(&hw_stats
->tx_octets
);
7959 stats
->rx_errors
= old_stats
->rx_errors
+
7960 get_stat64(&hw_stats
->rx_errors
);
7961 stats
->tx_errors
= old_stats
->tx_errors
+
7962 get_stat64(&hw_stats
->tx_errors
) +
7963 get_stat64(&hw_stats
->tx_mac_errors
) +
7964 get_stat64(&hw_stats
->tx_carrier_sense_errors
) +
7965 get_stat64(&hw_stats
->tx_discards
);
7967 stats
->multicast
= old_stats
->multicast
+
7968 get_stat64(&hw_stats
->rx_mcast_packets
);
7969 stats
->collisions
= old_stats
->collisions
+
7970 get_stat64(&hw_stats
->tx_collisions
);
7972 stats
->rx_length_errors
= old_stats
->rx_length_errors
+
7973 get_stat64(&hw_stats
->rx_frame_too_long_errors
) +
7974 get_stat64(&hw_stats
->rx_undersize_packets
);
7976 stats
->rx_over_errors
= old_stats
->rx_over_errors
+
7977 get_stat64(&hw_stats
->rxbds_empty
);
7978 stats
->rx_frame_errors
= old_stats
->rx_frame_errors
+
7979 get_stat64(&hw_stats
->rx_align_errors
);
7980 stats
->tx_aborted_errors
= old_stats
->tx_aborted_errors
+
7981 get_stat64(&hw_stats
->tx_discards
);
7982 stats
->tx_carrier_errors
= old_stats
->tx_carrier_errors
+
7983 get_stat64(&hw_stats
->tx_carrier_sense_errors
);
7985 stats
->rx_crc_errors
= old_stats
->rx_crc_errors
+
7986 calc_crc_errors(tp
);
7988 stats
->rx_missed_errors
= old_stats
->rx_missed_errors
+
7989 get_stat64(&hw_stats
->rx_discards
);
7994 static inline u32
calc_crc(unsigned char *buf
, int len
)
8002 for (j
= 0; j
< len
; j
++) {
8005 for (k
= 0; k
< 8; k
++) {
8019 static void tg3_set_multi(struct tg3
*tp
, unsigned int accept_all
)
8021 /* accept or reject all multicast frames */
8022 tw32(MAC_HASH_REG_0
, accept_all
? 0xffffffff : 0);
8023 tw32(MAC_HASH_REG_1
, accept_all
? 0xffffffff : 0);
8024 tw32(MAC_HASH_REG_2
, accept_all
? 0xffffffff : 0);
8025 tw32(MAC_HASH_REG_3
, accept_all
? 0xffffffff : 0);
8028 static void __tg3_set_rx_mode(struct net_device
*dev
)
8030 struct tg3
*tp
= netdev_priv(dev
);
8033 rx_mode
= tp
->rx_mode
& ~(RX_MODE_PROMISC
|
8034 RX_MODE_KEEP_VLAN_TAG
);
8036 /* When ASF is in use, we always keep the RX_MODE_KEEP_VLAN_TAG
8039 #if TG3_VLAN_TAG_USED
8041 !(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
))
8042 rx_mode
|= RX_MODE_KEEP_VLAN_TAG
;
8044 /* By definition, VLAN is disabled always in this
8047 if (!(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
))
8048 rx_mode
|= RX_MODE_KEEP_VLAN_TAG
;
8051 if (dev
->flags
& IFF_PROMISC
) {
8052 /* Promiscuous mode. */
8053 rx_mode
|= RX_MODE_PROMISC
;
8054 } else if (dev
->flags
& IFF_ALLMULTI
) {
8055 /* Accept all multicast. */
8056 tg3_set_multi (tp
, 1);
8057 } else if (dev
->mc_count
< 1) {
8058 /* Reject all multicast. */
8059 tg3_set_multi (tp
, 0);
8061 /* Accept one or more multicast(s). */
8062 struct dev_mc_list
*mclist
;
8064 u32 mc_filter
[4] = { 0, };
8069 for (i
= 0, mclist
= dev
->mc_list
; mclist
&& i
< dev
->mc_count
;
8070 i
++, mclist
= mclist
->next
) {
8072 crc
= calc_crc (mclist
->dmi_addr
, ETH_ALEN
);
8074 regidx
= (bit
& 0x60) >> 5;
8076 mc_filter
[regidx
] |= (1 << bit
);
8079 tw32(MAC_HASH_REG_0
, mc_filter
[0]);
8080 tw32(MAC_HASH_REG_1
, mc_filter
[1]);
8081 tw32(MAC_HASH_REG_2
, mc_filter
[2]);
8082 tw32(MAC_HASH_REG_3
, mc_filter
[3]);
8085 if (rx_mode
!= tp
->rx_mode
) {
8086 tp
->rx_mode
= rx_mode
;
8087 tw32_f(MAC_RX_MODE
, rx_mode
);
8092 static void tg3_set_rx_mode(struct net_device
*dev
)
8094 struct tg3
*tp
= netdev_priv(dev
);
8096 if (!netif_running(dev
))
8099 tg3_full_lock(tp
, 0);
8100 __tg3_set_rx_mode(dev
);
8101 tg3_full_unlock(tp
);
8104 #define TG3_REGDUMP_LEN (32 * 1024)
8106 static int tg3_get_regs_len(struct net_device
*dev
)
8108 return TG3_REGDUMP_LEN
;
8111 static void tg3_get_regs(struct net_device
*dev
,
8112 struct ethtool_regs
*regs
, void *_p
)
8115 struct tg3
*tp
= netdev_priv(dev
);
8121 memset(p
, 0, TG3_REGDUMP_LEN
);
8123 if (tp
->link_config
.phy_is_low_power
)
8126 tg3_full_lock(tp
, 0);
8128 #define __GET_REG32(reg) (*(p)++ = tr32(reg))
8129 #define GET_REG32_LOOP(base,len) \
8130 do { p = (u32 *)(orig_p + (base)); \
8131 for (i = 0; i < len; i += 4) \
8132 __GET_REG32((base) + i); \
8134 #define GET_REG32_1(reg) \
8135 do { p = (u32 *)(orig_p + (reg)); \
8136 __GET_REG32((reg)); \
8139 GET_REG32_LOOP(TG3PCI_VENDOR
, 0xb0);
8140 GET_REG32_LOOP(MAILBOX_INTERRUPT_0
, 0x200);
8141 GET_REG32_LOOP(MAC_MODE
, 0x4f0);
8142 GET_REG32_LOOP(SNDDATAI_MODE
, 0xe0);
8143 GET_REG32_1(SNDDATAC_MODE
);
8144 GET_REG32_LOOP(SNDBDS_MODE
, 0x80);
8145 GET_REG32_LOOP(SNDBDI_MODE
, 0x48);
8146 GET_REG32_1(SNDBDC_MODE
);
8147 GET_REG32_LOOP(RCVLPC_MODE
, 0x20);
8148 GET_REG32_LOOP(RCVLPC_SELLST_BASE
, 0x15c);
8149 GET_REG32_LOOP(RCVDBDI_MODE
, 0x0c);
8150 GET_REG32_LOOP(RCVDBDI_JUMBO_BD
, 0x3c);
8151 GET_REG32_LOOP(RCVDBDI_BD_PROD_IDX_0
, 0x44);
8152 GET_REG32_1(RCVDCC_MODE
);
8153 GET_REG32_LOOP(RCVBDI_MODE
, 0x20);
8154 GET_REG32_LOOP(RCVCC_MODE
, 0x14);
8155 GET_REG32_LOOP(RCVLSC_MODE
, 0x08);
8156 GET_REG32_1(MBFREE_MODE
);
8157 GET_REG32_LOOP(HOSTCC_MODE
, 0x100);
8158 GET_REG32_LOOP(MEMARB_MODE
, 0x10);
8159 GET_REG32_LOOP(BUFMGR_MODE
, 0x58);
8160 GET_REG32_LOOP(RDMAC_MODE
, 0x08);
8161 GET_REG32_LOOP(WDMAC_MODE
, 0x08);
8162 GET_REG32_1(RX_CPU_MODE
);
8163 GET_REG32_1(RX_CPU_STATE
);
8164 GET_REG32_1(RX_CPU_PGMCTR
);
8165 GET_REG32_1(RX_CPU_HWBKPT
);
8166 GET_REG32_1(TX_CPU_MODE
);
8167 GET_REG32_1(TX_CPU_STATE
);
8168 GET_REG32_1(TX_CPU_PGMCTR
);
8169 GET_REG32_LOOP(GRCMBOX_INTERRUPT_0
, 0x110);
8170 GET_REG32_LOOP(FTQ_RESET
, 0x120);
8171 GET_REG32_LOOP(MSGINT_MODE
, 0x0c);
8172 GET_REG32_1(DMAC_MODE
);
8173 GET_REG32_LOOP(GRC_MODE
, 0x4c);
8174 if (tp
->tg3_flags
& TG3_FLAG_NVRAM
)
8175 GET_REG32_LOOP(NVRAM_CMD
, 0x24);
8178 #undef GET_REG32_LOOP
8181 tg3_full_unlock(tp
);
8184 static int tg3_get_eeprom_len(struct net_device
*dev
)
8186 struct tg3
*tp
= netdev_priv(dev
);
8188 return tp
->nvram_size
;
8191 static int tg3_nvram_read(struct tg3
*tp
, u32 offset
, u32
*val
);
8192 static int tg3_nvram_read_le(struct tg3
*tp
, u32 offset
, __le32
*val
);
8193 static int tg3_nvram_read_swab(struct tg3
*tp
, u32 offset
, u32
*val
);
8195 static int tg3_get_eeprom(struct net_device
*dev
, struct ethtool_eeprom
*eeprom
, u8
*data
)
8197 struct tg3
*tp
= netdev_priv(dev
);
8200 u32 i
, offset
, len
, b_offset
, b_count
;
8203 if (tp
->link_config
.phy_is_low_power
)
8206 offset
= eeprom
->offset
;
8210 eeprom
->magic
= TG3_EEPROM_MAGIC
;
8213 /* adjustments to start on required 4 byte boundary */
8214 b_offset
= offset
& 3;
8215 b_count
= 4 - b_offset
;
8216 if (b_count
> len
) {
8217 /* i.e. offset=1 len=2 */
8220 ret
= tg3_nvram_read_le(tp
, offset
-b_offset
, &val
);
8223 memcpy(data
, ((char*)&val
) + b_offset
, b_count
);
8226 eeprom
->len
+= b_count
;
8229 /* read bytes upto the last 4 byte boundary */
8230 pd
= &data
[eeprom
->len
];
8231 for (i
= 0; i
< (len
- (len
& 3)); i
+= 4) {
8232 ret
= tg3_nvram_read_le(tp
, offset
+ i
, &val
);
8237 memcpy(pd
+ i
, &val
, 4);
8242 /* read last bytes not ending on 4 byte boundary */
8243 pd
= &data
[eeprom
->len
];
8245 b_offset
= offset
+ len
- b_count
;
8246 ret
= tg3_nvram_read_le(tp
, b_offset
, &val
);
8249 memcpy(pd
, &val
, b_count
);
8250 eeprom
->len
+= b_count
;
8255 static int tg3_nvram_write_block(struct tg3
*tp
, u32 offset
, u32 len
, u8
*buf
);
8257 static int tg3_set_eeprom(struct net_device
*dev
, struct ethtool_eeprom
*eeprom
, u8
*data
)
8259 struct tg3
*tp
= netdev_priv(dev
);
8261 u32 offset
, len
, b_offset
, odd_len
;
8265 if (tp
->link_config
.phy_is_low_power
)
8268 if (eeprom
->magic
!= TG3_EEPROM_MAGIC
)
8271 offset
= eeprom
->offset
;
8274 if ((b_offset
= (offset
& 3))) {
8275 /* adjustments to start on required 4 byte boundary */
8276 ret
= tg3_nvram_read_le(tp
, offset
-b_offset
, &start
);
8287 /* adjustments to end on required 4 byte boundary */
8289 len
= (len
+ 3) & ~3;
8290 ret
= tg3_nvram_read_le(tp
, offset
+len
-4, &end
);
8296 if (b_offset
|| odd_len
) {
8297 buf
= kmalloc(len
, GFP_KERNEL
);
8301 memcpy(buf
, &start
, 4);
8303 memcpy(buf
+len
-4, &end
, 4);
8304 memcpy(buf
+ b_offset
, data
, eeprom
->len
);
8307 ret
= tg3_nvram_write_block(tp
, offset
, len
, buf
);
8315 static int tg3_get_settings(struct net_device
*dev
, struct ethtool_cmd
*cmd
)
8317 struct tg3
*tp
= netdev_priv(dev
);
8319 cmd
->supported
= (SUPPORTED_Autoneg
);
8321 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
))
8322 cmd
->supported
|= (SUPPORTED_1000baseT_Half
|
8323 SUPPORTED_1000baseT_Full
);
8325 if (!(tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
)) {
8326 cmd
->supported
|= (SUPPORTED_100baseT_Half
|
8327 SUPPORTED_100baseT_Full
|
8328 SUPPORTED_10baseT_Half
|
8329 SUPPORTED_10baseT_Full
|
8331 cmd
->port
= PORT_TP
;
8333 cmd
->supported
|= SUPPORTED_FIBRE
;
8334 cmd
->port
= PORT_FIBRE
;
8337 cmd
->advertising
= tp
->link_config
.advertising
;
8338 if (netif_running(dev
)) {
8339 cmd
->speed
= tp
->link_config
.active_speed
;
8340 cmd
->duplex
= tp
->link_config
.active_duplex
;
8342 cmd
->phy_address
= PHY_ADDR
;
8343 cmd
->transceiver
= 0;
8344 cmd
->autoneg
= tp
->link_config
.autoneg
;
8350 static int tg3_set_settings(struct net_device
*dev
, struct ethtool_cmd
*cmd
)
8352 struct tg3
*tp
= netdev_priv(dev
);
8354 if (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
) {
8355 /* These are the only valid advertisement bits allowed. */
8356 if (cmd
->autoneg
== AUTONEG_ENABLE
&&
8357 (cmd
->advertising
& ~(ADVERTISED_1000baseT_Half
|
8358 ADVERTISED_1000baseT_Full
|
8359 ADVERTISED_Autoneg
|
8362 /* Fiber can only do SPEED_1000. */
8363 else if ((cmd
->autoneg
!= AUTONEG_ENABLE
) &&
8364 (cmd
->speed
!= SPEED_1000
))
8366 /* Copper cannot force SPEED_1000. */
8367 } else if ((cmd
->autoneg
!= AUTONEG_ENABLE
) &&
8368 (cmd
->speed
== SPEED_1000
))
8370 else if ((cmd
->speed
== SPEED_1000
) &&
8371 (tp
->tg3_flags2
& TG3_FLAG_10_100_ONLY
))
8374 tg3_full_lock(tp
, 0);
8376 tp
->link_config
.autoneg
= cmd
->autoneg
;
8377 if (cmd
->autoneg
== AUTONEG_ENABLE
) {
8378 tp
->link_config
.advertising
= (cmd
->advertising
|
8379 ADVERTISED_Autoneg
);
8380 tp
->link_config
.speed
= SPEED_INVALID
;
8381 tp
->link_config
.duplex
= DUPLEX_INVALID
;
8383 tp
->link_config
.advertising
= 0;
8384 tp
->link_config
.speed
= cmd
->speed
;
8385 tp
->link_config
.duplex
= cmd
->duplex
;
8388 tp
->link_config
.orig_speed
= tp
->link_config
.speed
;
8389 tp
->link_config
.orig_duplex
= tp
->link_config
.duplex
;
8390 tp
->link_config
.orig_autoneg
= tp
->link_config
.autoneg
;
8392 if (netif_running(dev
))
8393 tg3_setup_phy(tp
, 1);
8395 tg3_full_unlock(tp
);
8400 static void tg3_get_drvinfo(struct net_device
*dev
, struct ethtool_drvinfo
*info
)
8402 struct tg3
*tp
= netdev_priv(dev
);
8404 strcpy(info
->driver
, DRV_MODULE_NAME
);
8405 strcpy(info
->version
, DRV_MODULE_VERSION
);
8406 strcpy(info
->fw_version
, tp
->fw_ver
);
8407 strcpy(info
->bus_info
, pci_name(tp
->pdev
));
8410 static void tg3_get_wol(struct net_device
*dev
, struct ethtool_wolinfo
*wol
)
8412 struct tg3
*tp
= netdev_priv(dev
);
8414 if (tp
->tg3_flags
& TG3_FLAG_WOL_CAP
)
8415 wol
->supported
= WAKE_MAGIC
;
8419 if (tp
->tg3_flags
& TG3_FLAG_WOL_ENABLE
)
8420 wol
->wolopts
= WAKE_MAGIC
;
8421 memset(&wol
->sopass
, 0, sizeof(wol
->sopass
));
8424 static int tg3_set_wol(struct net_device
*dev
, struct ethtool_wolinfo
*wol
)
8426 struct tg3
*tp
= netdev_priv(dev
);
8428 if (wol
->wolopts
& ~WAKE_MAGIC
)
8430 if ((wol
->wolopts
& WAKE_MAGIC
) &&
8431 !(tp
->tg3_flags
& TG3_FLAG_WOL_CAP
))
8434 spin_lock_bh(&tp
->lock
);
8435 if (wol
->wolopts
& WAKE_MAGIC
)
8436 tp
->tg3_flags
|= TG3_FLAG_WOL_ENABLE
;
8438 tp
->tg3_flags
&= ~TG3_FLAG_WOL_ENABLE
;
8439 spin_unlock_bh(&tp
->lock
);
8444 static u32
tg3_get_msglevel(struct net_device
*dev
)
8446 struct tg3
*tp
= netdev_priv(dev
);
8447 return tp
->msg_enable
;
8450 static void tg3_set_msglevel(struct net_device
*dev
, u32 value
)
8452 struct tg3
*tp
= netdev_priv(dev
);
8453 tp
->msg_enable
= value
;
8456 static int tg3_set_tso(struct net_device
*dev
, u32 value
)
8458 struct tg3
*tp
= netdev_priv(dev
);
8460 if (!(tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
)) {
8465 if ((tp
->tg3_flags2
& TG3_FLG2_HW_TSO_2
) &&
8466 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
)) {
8468 dev
->features
|= NETIF_F_TSO6
;
8469 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
8470 dev
->features
|= NETIF_F_TSO_ECN
;
8472 dev
->features
&= ~(NETIF_F_TSO6
| NETIF_F_TSO_ECN
);
8474 return ethtool_op_set_tso(dev
, value
);
8477 static int tg3_nway_reset(struct net_device
*dev
)
8479 struct tg3
*tp
= netdev_priv(dev
);
8483 if (!netif_running(dev
))
8486 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)
8489 spin_lock_bh(&tp
->lock
);
8491 tg3_readphy(tp
, MII_BMCR
, &bmcr
);
8492 if (!tg3_readphy(tp
, MII_BMCR
, &bmcr
) &&
8493 ((bmcr
& BMCR_ANENABLE
) ||
8494 (tp
->tg3_flags2
& TG3_FLG2_PARALLEL_DETECT
))) {
8495 tg3_writephy(tp
, MII_BMCR
, bmcr
| BMCR_ANRESTART
|
8499 spin_unlock_bh(&tp
->lock
);
8504 static void tg3_get_ringparam(struct net_device
*dev
, struct ethtool_ringparam
*ering
)
8506 struct tg3
*tp
= netdev_priv(dev
);
8508 ering
->rx_max_pending
= TG3_RX_RING_SIZE
- 1;
8509 ering
->rx_mini_max_pending
= 0;
8510 if (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
)
8511 ering
->rx_jumbo_max_pending
= TG3_RX_JUMBO_RING_SIZE
- 1;
8513 ering
->rx_jumbo_max_pending
= 0;
8515 ering
->tx_max_pending
= TG3_TX_RING_SIZE
- 1;
8517 ering
->rx_pending
= tp
->rx_pending
;
8518 ering
->rx_mini_pending
= 0;
8519 if (tp
->tg3_flags
& TG3_FLAG_JUMBO_RING_ENABLE
)
8520 ering
->rx_jumbo_pending
= tp
->rx_jumbo_pending
;
8522 ering
->rx_jumbo_pending
= 0;
8524 ering
->tx_pending
= tp
->tx_pending
;
8527 static int tg3_set_ringparam(struct net_device
*dev
, struct ethtool_ringparam
*ering
)
8529 struct tg3
*tp
= netdev_priv(dev
);
8530 int irq_sync
= 0, err
= 0;
8532 if ((ering
->rx_pending
> TG3_RX_RING_SIZE
- 1) ||
8533 (ering
->rx_jumbo_pending
> TG3_RX_JUMBO_RING_SIZE
- 1) ||
8534 (ering
->tx_pending
> TG3_TX_RING_SIZE
- 1) ||
8535 (ering
->tx_pending
<= MAX_SKB_FRAGS
) ||
8536 ((tp
->tg3_flags2
& TG3_FLG2_TSO_BUG
) &&
8537 (ering
->tx_pending
<= (MAX_SKB_FRAGS
* 3))))
8540 if (netif_running(dev
)) {
8545 tg3_full_lock(tp
, irq_sync
);
8547 tp
->rx_pending
= ering
->rx_pending
;
8549 if ((tp
->tg3_flags2
& TG3_FLG2_MAX_RXPEND_64
) &&
8550 tp
->rx_pending
> 63)
8551 tp
->rx_pending
= 63;
8552 tp
->rx_jumbo_pending
= ering
->rx_jumbo_pending
;
8553 tp
->tx_pending
= ering
->tx_pending
;
8555 if (netif_running(dev
)) {
8556 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
8557 err
= tg3_restart_hw(tp
, 1);
8559 tg3_netif_start(tp
);
8562 tg3_full_unlock(tp
);
8567 static void tg3_get_pauseparam(struct net_device
*dev
, struct ethtool_pauseparam
*epause
)
8569 struct tg3
*tp
= netdev_priv(dev
);
8571 epause
->autoneg
= (tp
->tg3_flags
& TG3_FLAG_PAUSE_AUTONEG
) != 0;
8572 epause
->rx_pause
= (tp
->tg3_flags
& TG3_FLAG_RX_PAUSE
) != 0;
8573 epause
->tx_pause
= (tp
->tg3_flags
& TG3_FLAG_TX_PAUSE
) != 0;
8576 static int tg3_set_pauseparam(struct net_device
*dev
, struct ethtool_pauseparam
*epause
)
8578 struct tg3
*tp
= netdev_priv(dev
);
8579 int irq_sync
= 0, err
= 0;
8581 if (netif_running(dev
)) {
8586 tg3_full_lock(tp
, irq_sync
);
8588 if (epause
->autoneg
)
8589 tp
->tg3_flags
|= TG3_FLAG_PAUSE_AUTONEG
;
8591 tp
->tg3_flags
&= ~TG3_FLAG_PAUSE_AUTONEG
;
8592 if (epause
->rx_pause
)
8593 tp
->tg3_flags
|= TG3_FLAG_RX_PAUSE
;
8595 tp
->tg3_flags
&= ~TG3_FLAG_RX_PAUSE
;
8596 if (epause
->tx_pause
)
8597 tp
->tg3_flags
|= TG3_FLAG_TX_PAUSE
;
8599 tp
->tg3_flags
&= ~TG3_FLAG_TX_PAUSE
;
8601 if (netif_running(dev
)) {
8602 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
8603 err
= tg3_restart_hw(tp
, 1);
8605 tg3_netif_start(tp
);
8608 tg3_full_unlock(tp
);
8613 static u32
tg3_get_rx_csum(struct net_device
*dev
)
8615 struct tg3
*tp
= netdev_priv(dev
);
8616 return (tp
->tg3_flags
& TG3_FLAG_RX_CHECKSUMS
) != 0;
8619 static int tg3_set_rx_csum(struct net_device
*dev
, u32 data
)
8621 struct tg3
*tp
= netdev_priv(dev
);
8623 if (tp
->tg3_flags
& TG3_FLAG_BROKEN_CHECKSUMS
) {
8629 spin_lock_bh(&tp
->lock
);
8631 tp
->tg3_flags
|= TG3_FLAG_RX_CHECKSUMS
;
8633 tp
->tg3_flags
&= ~TG3_FLAG_RX_CHECKSUMS
;
8634 spin_unlock_bh(&tp
->lock
);
8639 static int tg3_set_tx_csum(struct net_device
*dev
, u32 data
)
8641 struct tg3
*tp
= netdev_priv(dev
);
8643 if (tp
->tg3_flags
& TG3_FLAG_BROKEN_CHECKSUMS
) {
8649 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
8650 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
8651 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
8652 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
8653 ethtool_op_set_tx_ipv6_csum(dev
, data
);
8655 ethtool_op_set_tx_csum(dev
, data
);
8660 static int tg3_get_sset_count (struct net_device
*dev
, int sset
)
8664 return TG3_NUM_TEST
;
8666 return TG3_NUM_STATS
;
8672 static void tg3_get_strings (struct net_device
*dev
, u32 stringset
, u8
*buf
)
8674 switch (stringset
) {
8676 memcpy(buf
, ðtool_stats_keys
, sizeof(ethtool_stats_keys
));
8679 memcpy(buf
, ðtool_test_keys
, sizeof(ethtool_test_keys
));
8682 WARN_ON(1); /* we need a WARN() */
8687 static int tg3_phys_id(struct net_device
*dev
, u32 data
)
8689 struct tg3
*tp
= netdev_priv(dev
);
8692 if (!netif_running(tp
->dev
))
8698 for (i
= 0; i
< (data
* 2); i
++) {
8700 tw32(MAC_LED_CTRL
, LED_CTRL_LNKLED_OVERRIDE
|
8701 LED_CTRL_1000MBPS_ON
|
8702 LED_CTRL_100MBPS_ON
|
8703 LED_CTRL_10MBPS_ON
|
8704 LED_CTRL_TRAFFIC_OVERRIDE
|
8705 LED_CTRL_TRAFFIC_BLINK
|
8706 LED_CTRL_TRAFFIC_LED
);
8709 tw32(MAC_LED_CTRL
, LED_CTRL_LNKLED_OVERRIDE
|
8710 LED_CTRL_TRAFFIC_OVERRIDE
);
8712 if (msleep_interruptible(500))
8715 tw32(MAC_LED_CTRL
, tp
->led_ctrl
);
8719 static void tg3_get_ethtool_stats (struct net_device
*dev
,
8720 struct ethtool_stats
*estats
, u64
*tmp_stats
)
8722 struct tg3
*tp
= netdev_priv(dev
);
8723 memcpy(tmp_stats
, tg3_get_estats(tp
), sizeof(tp
->estats
));
8726 #define NVRAM_TEST_SIZE 0x100
8727 #define NVRAM_SELFBOOT_FORMAT1_0_SIZE 0x14
8728 #define NVRAM_SELFBOOT_FORMAT1_2_SIZE 0x18
8729 #define NVRAM_SELFBOOT_FORMAT1_3_SIZE 0x1c
8730 #define NVRAM_SELFBOOT_HW_SIZE 0x20
8731 #define NVRAM_SELFBOOT_DATA_SIZE 0x1c
8733 static int tg3_test_nvram(struct tg3
*tp
)
8737 int i
, j
, k
, err
= 0, size
;
8739 if (tg3_nvram_read_swab(tp
, 0, &magic
) != 0)
8742 if (magic
== TG3_EEPROM_MAGIC
)
8743 size
= NVRAM_TEST_SIZE
;
8744 else if ((magic
& TG3_EEPROM_MAGIC_FW_MSK
) == TG3_EEPROM_MAGIC_FW
) {
8745 if ((magic
& TG3_EEPROM_SB_FORMAT_MASK
) ==
8746 TG3_EEPROM_SB_FORMAT_1
) {
8747 switch (magic
& TG3_EEPROM_SB_REVISION_MASK
) {
8748 case TG3_EEPROM_SB_REVISION_0
:
8749 size
= NVRAM_SELFBOOT_FORMAT1_0_SIZE
;
8751 case TG3_EEPROM_SB_REVISION_2
:
8752 size
= NVRAM_SELFBOOT_FORMAT1_2_SIZE
;
8754 case TG3_EEPROM_SB_REVISION_3
:
8755 size
= NVRAM_SELFBOOT_FORMAT1_3_SIZE
;
8762 } else if ((magic
& TG3_EEPROM_MAGIC_HW_MSK
) == TG3_EEPROM_MAGIC_HW
)
8763 size
= NVRAM_SELFBOOT_HW_SIZE
;
8767 buf
= kmalloc(size
, GFP_KERNEL
);
8772 for (i
= 0, j
= 0; i
< size
; i
+= 4, j
++) {
8773 if ((err
= tg3_nvram_read_le(tp
, i
, &buf
[j
])) != 0)
8779 /* Selfboot format */
8780 magic
= swab32(le32_to_cpu(buf
[0]));
8781 if ((magic
& TG3_EEPROM_MAGIC_FW_MSK
) ==
8782 TG3_EEPROM_MAGIC_FW
) {
8783 u8
*buf8
= (u8
*) buf
, csum8
= 0;
8785 if ((magic
& TG3_EEPROM_SB_REVISION_MASK
) ==
8786 TG3_EEPROM_SB_REVISION_2
) {
8787 /* For rev 2, the csum doesn't include the MBA. */
8788 for (i
= 0; i
< TG3_EEPROM_SB_F1R2_MBA_OFF
; i
++)
8790 for (i
= TG3_EEPROM_SB_F1R2_MBA_OFF
+ 4; i
< size
; i
++)
8793 for (i
= 0; i
< size
; i
++)
8806 if ((magic
& TG3_EEPROM_MAGIC_HW_MSK
) ==
8807 TG3_EEPROM_MAGIC_HW
) {
8808 u8 data
[NVRAM_SELFBOOT_DATA_SIZE
];
8809 u8 parity
[NVRAM_SELFBOOT_DATA_SIZE
];
8810 u8
*buf8
= (u8
*) buf
;
8812 /* Separate the parity bits and the data bytes. */
8813 for (i
= 0, j
= 0, k
= 0; i
< NVRAM_SELFBOOT_HW_SIZE
; i
++) {
8814 if ((i
== 0) || (i
== 8)) {
8818 for (l
= 0, msk
= 0x80; l
< 7; l
++, msk
>>= 1)
8819 parity
[k
++] = buf8
[i
] & msk
;
8826 for (l
= 0, msk
= 0x20; l
< 6; l
++, msk
>>= 1)
8827 parity
[k
++] = buf8
[i
] & msk
;
8830 for (l
= 0, msk
= 0x80; l
< 8; l
++, msk
>>= 1)
8831 parity
[k
++] = buf8
[i
] & msk
;
8834 data
[j
++] = buf8
[i
];
8838 for (i
= 0; i
< NVRAM_SELFBOOT_DATA_SIZE
; i
++) {
8839 u8 hw8
= hweight8(data
[i
]);
8841 if ((hw8
& 0x1) && parity
[i
])
8843 else if (!(hw8
& 0x1) && !parity
[i
])
8850 /* Bootstrap checksum at offset 0x10 */
8851 csum
= calc_crc((unsigned char *) buf
, 0x10);
8852 if(csum
!= le32_to_cpu(buf
[0x10/4]))
8855 /* Manufacturing block starts at offset 0x74, checksum at 0xfc */
8856 csum
= calc_crc((unsigned char *) &buf
[0x74/4], 0x88);
8857 if (csum
!= le32_to_cpu(buf
[0xfc/4]))
8867 #define TG3_SERDES_TIMEOUT_SEC 2
8868 #define TG3_COPPER_TIMEOUT_SEC 6
8870 static int tg3_test_link(struct tg3
*tp
)
8874 if (!netif_running(tp
->dev
))
8877 if (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
)
8878 max
= TG3_SERDES_TIMEOUT_SEC
;
8880 max
= TG3_COPPER_TIMEOUT_SEC
;
8882 for (i
= 0; i
< max
; i
++) {
8883 if (netif_carrier_ok(tp
->dev
))
8886 if (msleep_interruptible(1000))
8893 /* Only test the commonly used registers */
8894 static int tg3_test_registers(struct tg3
*tp
)
8896 int i
, is_5705
, is_5750
;
8897 u32 offset
, read_mask
, write_mask
, val
, save_val
, read_val
;
8901 #define TG3_FL_5705 0x1
8902 #define TG3_FL_NOT_5705 0x2
8903 #define TG3_FL_NOT_5788 0x4
8904 #define TG3_FL_NOT_5750 0x8
8908 /* MAC Control Registers */
8909 { MAC_MODE
, TG3_FL_NOT_5705
,
8910 0x00000000, 0x00ef6f8c },
8911 { MAC_MODE
, TG3_FL_5705
,
8912 0x00000000, 0x01ef6b8c },
8913 { MAC_STATUS
, TG3_FL_NOT_5705
,
8914 0x03800107, 0x00000000 },
8915 { MAC_STATUS
, TG3_FL_5705
,
8916 0x03800100, 0x00000000 },
8917 { MAC_ADDR_0_HIGH
, 0x0000,
8918 0x00000000, 0x0000ffff },
8919 { MAC_ADDR_0_LOW
, 0x0000,
8920 0x00000000, 0xffffffff },
8921 { MAC_RX_MTU_SIZE
, 0x0000,
8922 0x00000000, 0x0000ffff },
8923 { MAC_TX_MODE
, 0x0000,
8924 0x00000000, 0x00000070 },
8925 { MAC_TX_LENGTHS
, 0x0000,
8926 0x00000000, 0x00003fff },
8927 { MAC_RX_MODE
, TG3_FL_NOT_5705
,
8928 0x00000000, 0x000007fc },
8929 { MAC_RX_MODE
, TG3_FL_5705
,
8930 0x00000000, 0x000007dc },
8931 { MAC_HASH_REG_0
, 0x0000,
8932 0x00000000, 0xffffffff },
8933 { MAC_HASH_REG_1
, 0x0000,
8934 0x00000000, 0xffffffff },
8935 { MAC_HASH_REG_2
, 0x0000,
8936 0x00000000, 0xffffffff },
8937 { MAC_HASH_REG_3
, 0x0000,
8938 0x00000000, 0xffffffff },
8940 /* Receive Data and Receive BD Initiator Control Registers. */
8941 { RCVDBDI_JUMBO_BD
+0, TG3_FL_NOT_5705
,
8942 0x00000000, 0xffffffff },
8943 { RCVDBDI_JUMBO_BD
+4, TG3_FL_NOT_5705
,
8944 0x00000000, 0xffffffff },
8945 { RCVDBDI_JUMBO_BD
+8, TG3_FL_NOT_5705
,
8946 0x00000000, 0x00000003 },
8947 { RCVDBDI_JUMBO_BD
+0xc, TG3_FL_NOT_5705
,
8948 0x00000000, 0xffffffff },
8949 { RCVDBDI_STD_BD
+0, 0x0000,
8950 0x00000000, 0xffffffff },
8951 { RCVDBDI_STD_BD
+4, 0x0000,
8952 0x00000000, 0xffffffff },
8953 { RCVDBDI_STD_BD
+8, 0x0000,
8954 0x00000000, 0xffff0002 },
8955 { RCVDBDI_STD_BD
+0xc, 0x0000,
8956 0x00000000, 0xffffffff },
8958 /* Receive BD Initiator Control Registers. */
8959 { RCVBDI_STD_THRESH
, TG3_FL_NOT_5705
,
8960 0x00000000, 0xffffffff },
8961 { RCVBDI_STD_THRESH
, TG3_FL_5705
,
8962 0x00000000, 0x000003ff },
8963 { RCVBDI_JUMBO_THRESH
, TG3_FL_NOT_5705
,
8964 0x00000000, 0xffffffff },
8966 /* Host Coalescing Control Registers. */
8967 { HOSTCC_MODE
, TG3_FL_NOT_5705
,
8968 0x00000000, 0x00000004 },
8969 { HOSTCC_MODE
, TG3_FL_5705
,
8970 0x00000000, 0x000000f6 },
8971 { HOSTCC_RXCOL_TICKS
, TG3_FL_NOT_5705
,
8972 0x00000000, 0xffffffff },
8973 { HOSTCC_RXCOL_TICKS
, TG3_FL_5705
,
8974 0x00000000, 0x000003ff },
8975 { HOSTCC_TXCOL_TICKS
, TG3_FL_NOT_5705
,
8976 0x00000000, 0xffffffff },
8977 { HOSTCC_TXCOL_TICKS
, TG3_FL_5705
,
8978 0x00000000, 0x000003ff },
8979 { HOSTCC_RXMAX_FRAMES
, TG3_FL_NOT_5705
,
8980 0x00000000, 0xffffffff },
8981 { HOSTCC_RXMAX_FRAMES
, TG3_FL_5705
| TG3_FL_NOT_5788
,
8982 0x00000000, 0x000000ff },
8983 { HOSTCC_TXMAX_FRAMES
, TG3_FL_NOT_5705
,
8984 0x00000000, 0xffffffff },
8985 { HOSTCC_TXMAX_FRAMES
, TG3_FL_5705
| TG3_FL_NOT_5788
,
8986 0x00000000, 0x000000ff },
8987 { HOSTCC_RXCOAL_TICK_INT
, TG3_FL_NOT_5705
,
8988 0x00000000, 0xffffffff },
8989 { HOSTCC_TXCOAL_TICK_INT
, TG3_FL_NOT_5705
,
8990 0x00000000, 0xffffffff },
8991 { HOSTCC_RXCOAL_MAXF_INT
, TG3_FL_NOT_5705
,
8992 0x00000000, 0xffffffff },
8993 { HOSTCC_RXCOAL_MAXF_INT
, TG3_FL_5705
| TG3_FL_NOT_5788
,
8994 0x00000000, 0x000000ff },
8995 { HOSTCC_TXCOAL_MAXF_INT
, TG3_FL_NOT_5705
,
8996 0x00000000, 0xffffffff },
8997 { HOSTCC_TXCOAL_MAXF_INT
, TG3_FL_5705
| TG3_FL_NOT_5788
,
8998 0x00000000, 0x000000ff },
8999 { HOSTCC_STAT_COAL_TICKS
, TG3_FL_NOT_5705
,
9000 0x00000000, 0xffffffff },
9001 { HOSTCC_STATS_BLK_HOST_ADDR
, TG3_FL_NOT_5705
,
9002 0x00000000, 0xffffffff },
9003 { HOSTCC_STATS_BLK_HOST_ADDR
+4, TG3_FL_NOT_5705
,
9004 0x00000000, 0xffffffff },
9005 { HOSTCC_STATUS_BLK_HOST_ADDR
, 0x0000,
9006 0x00000000, 0xffffffff },
9007 { HOSTCC_STATUS_BLK_HOST_ADDR
+4, 0x0000,
9008 0x00000000, 0xffffffff },
9009 { HOSTCC_STATS_BLK_NIC_ADDR
, 0x0000,
9010 0xffffffff, 0x00000000 },
9011 { HOSTCC_STATUS_BLK_NIC_ADDR
, 0x0000,
9012 0xffffffff, 0x00000000 },
9014 /* Buffer Manager Control Registers. */
9015 { BUFMGR_MB_POOL_ADDR
, TG3_FL_NOT_5750
,
9016 0x00000000, 0x007fff80 },
9017 { BUFMGR_MB_POOL_SIZE
, TG3_FL_NOT_5750
,
9018 0x00000000, 0x007fffff },
9019 { BUFMGR_MB_RDMA_LOW_WATER
, 0x0000,
9020 0x00000000, 0x0000003f },
9021 { BUFMGR_MB_MACRX_LOW_WATER
, 0x0000,
9022 0x00000000, 0x000001ff },
9023 { BUFMGR_MB_HIGH_WATER
, 0x0000,
9024 0x00000000, 0x000001ff },
9025 { BUFMGR_DMA_DESC_POOL_ADDR
, TG3_FL_NOT_5705
,
9026 0xffffffff, 0x00000000 },
9027 { BUFMGR_DMA_DESC_POOL_SIZE
, TG3_FL_NOT_5705
,
9028 0xffffffff, 0x00000000 },
9030 /* Mailbox Registers */
9031 { GRCMBOX_RCVSTD_PROD_IDX
+4, 0x0000,
9032 0x00000000, 0x000001ff },
9033 { GRCMBOX_RCVJUMBO_PROD_IDX
+4, TG3_FL_NOT_5705
,
9034 0x00000000, 0x000001ff },
9035 { GRCMBOX_RCVRET_CON_IDX_0
+4, 0x0000,
9036 0x00000000, 0x000007ff },
9037 { GRCMBOX_SNDHOST_PROD_IDX_0
+4, 0x0000,
9038 0x00000000, 0x000001ff },
9040 { 0xffff, 0x0000, 0x00000000, 0x00000000 },
9043 is_5705
= is_5750
= 0;
9044 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
9046 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
)
9050 for (i
= 0; reg_tbl
[i
].offset
!= 0xffff; i
++) {
9051 if (is_5705
&& (reg_tbl
[i
].flags
& TG3_FL_NOT_5705
))
9054 if (!is_5705
&& (reg_tbl
[i
].flags
& TG3_FL_5705
))
9057 if ((tp
->tg3_flags2
& TG3_FLG2_IS_5788
) &&
9058 (reg_tbl
[i
].flags
& TG3_FL_NOT_5788
))
9061 if (is_5750
&& (reg_tbl
[i
].flags
& TG3_FL_NOT_5750
))
9064 offset
= (u32
) reg_tbl
[i
].offset
;
9065 read_mask
= reg_tbl
[i
].read_mask
;
9066 write_mask
= reg_tbl
[i
].write_mask
;
9068 /* Save the original register content */
9069 save_val
= tr32(offset
);
9071 /* Determine the read-only value. */
9072 read_val
= save_val
& read_mask
;
9074 /* Write zero to the register, then make sure the read-only bits
9075 * are not changed and the read/write bits are all zeros.
9081 /* Test the read-only and read/write bits. */
9082 if (((val
& read_mask
) != read_val
) || (val
& write_mask
))
9085 /* Write ones to all the bits defined by RdMask and WrMask, then
9086 * make sure the read-only bits are not changed and the
9087 * read/write bits are all ones.
9089 tw32(offset
, read_mask
| write_mask
);
9093 /* Test the read-only bits. */
9094 if ((val
& read_mask
) != read_val
)
9097 /* Test the read/write bits. */
9098 if ((val
& write_mask
) != write_mask
)
9101 tw32(offset
, save_val
);
9107 if (netif_msg_hw(tp
))
9108 printk(KERN_ERR PFX
"Register test failed at offset %x\n",
9110 tw32(offset
, save_val
);
9114 static int tg3_do_mem_test(struct tg3
*tp
, u32 offset
, u32 len
)
9116 static const u32 test_pattern
[] = { 0x00000000, 0xffffffff, 0xaa55a55a };
9120 for (i
= 0; i
< ARRAY_SIZE(test_pattern
); i
++) {
9121 for (j
= 0; j
< len
; j
+= 4) {
9124 tg3_write_mem(tp
, offset
+ j
, test_pattern
[i
]);
9125 tg3_read_mem(tp
, offset
+ j
, &val
);
9126 if (val
!= test_pattern
[i
])
9133 static int tg3_test_memory(struct tg3
*tp
)
9135 static struct mem_entry
{
9138 } mem_tbl_570x
[] = {
9139 { 0x00000000, 0x00b50},
9140 { 0x00002000, 0x1c000},
9141 { 0xffffffff, 0x00000}
9142 }, mem_tbl_5705
[] = {
9143 { 0x00000100, 0x0000c},
9144 { 0x00000200, 0x00008},
9145 { 0x00004000, 0x00800},
9146 { 0x00006000, 0x01000},
9147 { 0x00008000, 0x02000},
9148 { 0x00010000, 0x0e000},
9149 { 0xffffffff, 0x00000}
9150 }, mem_tbl_5755
[] = {
9151 { 0x00000200, 0x00008},
9152 { 0x00004000, 0x00800},
9153 { 0x00006000, 0x00800},
9154 { 0x00008000, 0x02000},
9155 { 0x00010000, 0x0c000},
9156 { 0xffffffff, 0x00000}
9157 }, mem_tbl_5906
[] = {
9158 { 0x00000200, 0x00008},
9159 { 0x00004000, 0x00400},
9160 { 0x00006000, 0x00400},
9161 { 0x00008000, 0x01000},
9162 { 0x00010000, 0x01000},
9163 { 0xffffffff, 0x00000}
9165 struct mem_entry
*mem_tbl
;
9169 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
9170 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
9171 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
9172 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
9173 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
9174 mem_tbl
= mem_tbl_5755
;
9175 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
9176 mem_tbl
= mem_tbl_5906
;
9178 mem_tbl
= mem_tbl_5705
;
9180 mem_tbl
= mem_tbl_570x
;
9182 for (i
= 0; mem_tbl
[i
].offset
!= 0xffffffff; i
++) {
9183 if ((err
= tg3_do_mem_test(tp
, mem_tbl
[i
].offset
,
9184 mem_tbl
[i
].len
)) != 0)
9191 #define TG3_MAC_LOOPBACK 0
9192 #define TG3_PHY_LOOPBACK 1
9194 static int tg3_run_loopback(struct tg3
*tp
, int loopback_mode
)
9196 u32 mac_mode
, rx_start_idx
, rx_idx
, tx_idx
, opaque_key
;
9198 struct sk_buff
*skb
, *rx_skb
;
9201 int num_pkts
, tx_len
, rx_len
, i
, err
;
9202 struct tg3_rx_buffer_desc
*desc
;
9204 if (loopback_mode
== TG3_MAC_LOOPBACK
) {
9205 /* HW errata - mac loopback fails in some cases on 5780.
9206 * Normal traffic and PHY loopback are not affected by
9209 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5780
)
9212 mac_mode
= (tp
->mac_mode
& ~MAC_MODE_PORT_MODE_MASK
) |
9213 MAC_MODE_PORT_INT_LPBACK
;
9214 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
9215 mac_mode
|= MAC_MODE_LINK_POLARITY
;
9216 if (tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
)
9217 mac_mode
|= MAC_MODE_PORT_MODE_MII
;
9219 mac_mode
|= MAC_MODE_PORT_MODE_GMII
;
9220 tw32(MAC_MODE
, mac_mode
);
9221 } else if (loopback_mode
== TG3_PHY_LOOPBACK
) {
9224 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
9227 if (!tg3_readphy(tp
, MII_TG3_EPHY_TEST
, &phytest
)) {
9230 tg3_writephy(tp
, MII_TG3_EPHY_TEST
,
9231 phytest
| MII_TG3_EPHY_SHADOW_EN
);
9232 if (!tg3_readphy(tp
, 0x1b, &phy
))
9233 tg3_writephy(tp
, 0x1b, phy
& ~0x20);
9234 tg3_writephy(tp
, MII_TG3_EPHY_TEST
, phytest
);
9236 val
= BMCR_LOOPBACK
| BMCR_FULLDPLX
| BMCR_SPEED100
;
9238 val
= BMCR_LOOPBACK
| BMCR_FULLDPLX
| BMCR_SPEED1000
;
9240 tg3_phy_toggle_automdix(tp
, 0);
9242 tg3_writephy(tp
, MII_BMCR
, val
);
9245 mac_mode
= tp
->mac_mode
& ~MAC_MODE_PORT_MODE_MASK
;
9246 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
9247 tg3_writephy(tp
, MII_TG3_EPHY_PTEST
, 0x1800);
9248 mac_mode
|= MAC_MODE_PORT_MODE_MII
;
9250 mac_mode
|= MAC_MODE_PORT_MODE_GMII
;
9252 /* reset to prevent losing 1st rx packet intermittently */
9253 if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
) {
9254 tw32_f(MAC_RX_MODE
, RX_MODE_RESET
);
9256 tw32_f(MAC_RX_MODE
, tp
->rx_mode
);
9258 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
) {
9259 if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5401
)
9260 mac_mode
&= ~MAC_MODE_LINK_POLARITY
;
9261 else if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5411
)
9262 mac_mode
|= MAC_MODE_LINK_POLARITY
;
9263 tg3_writephy(tp
, MII_TG3_EXT_CTRL
,
9264 MII_TG3_EXT_CTRL_LNK3_LED_MODE
);
9266 tw32(MAC_MODE
, mac_mode
);
9274 skb
= netdev_alloc_skb(tp
->dev
, tx_len
);
9278 tx_data
= skb_put(skb
, tx_len
);
9279 memcpy(tx_data
, tp
->dev
->dev_addr
, 6);
9280 memset(tx_data
+ 6, 0x0, 8);
9282 tw32(MAC_RX_MTU_SIZE
, tx_len
+ 4);
9284 for (i
= 14; i
< tx_len
; i
++)
9285 tx_data
[i
] = (u8
) (i
& 0xff);
9287 map
= pci_map_single(tp
->pdev
, skb
->data
, tx_len
, PCI_DMA_TODEVICE
);
9289 tw32_f(HOSTCC_MODE
, tp
->coalesce_mode
| HOSTCC_MODE_ENABLE
|
9294 rx_start_idx
= tp
->hw_status
->idx
[0].rx_producer
;
9298 tg3_set_txd(tp
, tp
->tx_prod
, map
, tx_len
, 0, 1);
9303 tw32_tx_mbox(MAILBOX_SNDHOST_PROD_IDX_0
+ TG3_64BIT_REG_LOW
,
9305 tr32_mailbox(MAILBOX_SNDHOST_PROD_IDX_0
+ TG3_64BIT_REG_LOW
);
9309 /* 250 usec to allow enough time on some 10/100 Mbps devices. */
9310 for (i
= 0; i
< 25; i
++) {
9311 tw32_f(HOSTCC_MODE
, tp
->coalesce_mode
| HOSTCC_MODE_ENABLE
|
9316 tx_idx
= tp
->hw_status
->idx
[0].tx_consumer
;
9317 rx_idx
= tp
->hw_status
->idx
[0].rx_producer
;
9318 if ((tx_idx
== tp
->tx_prod
) &&
9319 (rx_idx
== (rx_start_idx
+ num_pkts
)))
9323 pci_unmap_single(tp
->pdev
, map
, tx_len
, PCI_DMA_TODEVICE
);
9326 if (tx_idx
!= tp
->tx_prod
)
9329 if (rx_idx
!= rx_start_idx
+ num_pkts
)
9332 desc
= &tp
->rx_rcb
[rx_start_idx
];
9333 desc_idx
= desc
->opaque
& RXD_OPAQUE_INDEX_MASK
;
9334 opaque_key
= desc
->opaque
& RXD_OPAQUE_RING_MASK
;
9335 if (opaque_key
!= RXD_OPAQUE_RING_STD
)
9338 if ((desc
->err_vlan
& RXD_ERR_MASK
) != 0 &&
9339 (desc
->err_vlan
!= RXD_ERR_ODD_NIBBLE_RCVD_MII
))
9342 rx_len
= ((desc
->idx_len
& RXD_LEN_MASK
) >> RXD_LEN_SHIFT
) - 4;
9343 if (rx_len
!= tx_len
)
9346 rx_skb
= tp
->rx_std_buffers
[desc_idx
].skb
;
9348 map
= pci_unmap_addr(&tp
->rx_std_buffers
[desc_idx
], mapping
);
9349 pci_dma_sync_single_for_cpu(tp
->pdev
, map
, rx_len
, PCI_DMA_FROMDEVICE
);
9351 for (i
= 14; i
< tx_len
; i
++) {
9352 if (*(rx_skb
->data
+ i
) != (u8
) (i
& 0xff))
9357 /* tg3_free_rings will unmap and free the rx_skb */
9362 #define TG3_MAC_LOOPBACK_FAILED 1
9363 #define TG3_PHY_LOOPBACK_FAILED 2
9364 #define TG3_LOOPBACK_FAILED (TG3_MAC_LOOPBACK_FAILED | \
9365 TG3_PHY_LOOPBACK_FAILED)
9367 static int tg3_test_loopback(struct tg3
*tp
)
9372 if (!netif_running(tp
->dev
))
9373 return TG3_LOOPBACK_FAILED
;
9375 err
= tg3_reset_hw(tp
, 1);
9377 return TG3_LOOPBACK_FAILED
;
9379 if (tp
->tg3_flags3
& TG3_FLG3_5761_5784_AX_FIXES
) {
9383 tw32(TG3_CPMU_MUTEX_REQ
, CPMU_MUTEX_REQ_DRIVER
);
9385 /* Wait for up to 40 microseconds to acquire lock. */
9386 for (i
= 0; i
< 4; i
++) {
9387 status
= tr32(TG3_CPMU_MUTEX_GNT
);
9388 if (status
== CPMU_MUTEX_GNT_DRIVER
)
9393 if (status
!= CPMU_MUTEX_GNT_DRIVER
)
9394 return TG3_LOOPBACK_FAILED
;
9396 /* Turn off power management based on link speed. */
9397 cpmuctrl
= tr32(TG3_CPMU_CTRL
);
9399 cpmuctrl
& ~(CPMU_CTRL_LINK_SPEED_MODE
|
9400 CPMU_CTRL_LINK_AWARE_MODE
));
9403 if (tg3_run_loopback(tp
, TG3_MAC_LOOPBACK
))
9404 err
|= TG3_MAC_LOOPBACK_FAILED
;
9406 if (tp
->tg3_flags3
& TG3_FLG3_5761_5784_AX_FIXES
) {
9407 tw32(TG3_CPMU_CTRL
, cpmuctrl
);
9409 /* Release the mutex */
9410 tw32(TG3_CPMU_MUTEX_GNT
, CPMU_MUTEX_GNT_DRIVER
);
9413 if (!(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)) {
9414 if (tg3_run_loopback(tp
, TG3_PHY_LOOPBACK
))
9415 err
|= TG3_PHY_LOOPBACK_FAILED
;
9421 static void tg3_self_test(struct net_device
*dev
, struct ethtool_test
*etest
,
9424 struct tg3
*tp
= netdev_priv(dev
);
9426 if (tp
->link_config
.phy_is_low_power
)
9427 tg3_set_power_state(tp
, PCI_D0
);
9429 memset(data
, 0, sizeof(u64
) * TG3_NUM_TEST
);
9431 if (tg3_test_nvram(tp
) != 0) {
9432 etest
->flags
|= ETH_TEST_FL_FAILED
;
9435 if (tg3_test_link(tp
) != 0) {
9436 etest
->flags
|= ETH_TEST_FL_FAILED
;
9439 if (etest
->flags
& ETH_TEST_FL_OFFLINE
) {
9440 int err
, irq_sync
= 0;
9442 if (netif_running(dev
)) {
9447 tg3_full_lock(tp
, irq_sync
);
9449 tg3_halt(tp
, RESET_KIND_SUSPEND
, 1);
9450 err
= tg3_nvram_lock(tp
);
9451 tg3_halt_cpu(tp
, RX_CPU_BASE
);
9452 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
9453 tg3_halt_cpu(tp
, TX_CPU_BASE
);
9455 tg3_nvram_unlock(tp
);
9457 if (tp
->tg3_flags2
& TG3_FLG2_MII_SERDES
)
9460 if (tg3_test_registers(tp
) != 0) {
9461 etest
->flags
|= ETH_TEST_FL_FAILED
;
9464 if (tg3_test_memory(tp
) != 0) {
9465 etest
->flags
|= ETH_TEST_FL_FAILED
;
9468 if ((data
[4] = tg3_test_loopback(tp
)) != 0)
9469 etest
->flags
|= ETH_TEST_FL_FAILED
;
9471 tg3_full_unlock(tp
);
9473 if (tg3_test_interrupt(tp
) != 0) {
9474 etest
->flags
|= ETH_TEST_FL_FAILED
;
9478 tg3_full_lock(tp
, 0);
9480 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
9481 if (netif_running(dev
)) {
9482 tp
->tg3_flags
|= TG3_FLAG_INIT_COMPLETE
;
9483 if (!tg3_restart_hw(tp
, 1))
9484 tg3_netif_start(tp
);
9487 tg3_full_unlock(tp
);
9489 if (tp
->link_config
.phy_is_low_power
)
9490 tg3_set_power_state(tp
, PCI_D3hot
);
9494 static int tg3_ioctl(struct net_device
*dev
, struct ifreq
*ifr
, int cmd
)
9496 struct mii_ioctl_data
*data
= if_mii(ifr
);
9497 struct tg3
*tp
= netdev_priv(dev
);
9502 data
->phy_id
= PHY_ADDR
;
9508 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)
9509 break; /* We have no PHY */
9511 if (tp
->link_config
.phy_is_low_power
)
9514 spin_lock_bh(&tp
->lock
);
9515 err
= tg3_readphy(tp
, data
->reg_num
& 0x1f, &mii_regval
);
9516 spin_unlock_bh(&tp
->lock
);
9518 data
->val_out
= mii_regval
;
9524 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)
9525 break; /* We have no PHY */
9527 if (!capable(CAP_NET_ADMIN
))
9530 if (tp
->link_config
.phy_is_low_power
)
9533 spin_lock_bh(&tp
->lock
);
9534 err
= tg3_writephy(tp
, data
->reg_num
& 0x1f, data
->val_in
);
9535 spin_unlock_bh(&tp
->lock
);
9546 #if TG3_VLAN_TAG_USED
9547 static void tg3_vlan_rx_register(struct net_device
*dev
, struct vlan_group
*grp
)
9549 struct tg3
*tp
= netdev_priv(dev
);
9551 if (netif_running(dev
))
9554 tg3_full_lock(tp
, 0);
9558 /* Update RX_MODE_KEEP_VLAN_TAG bit in RX_MODE register. */
9559 __tg3_set_rx_mode(dev
);
9561 if (netif_running(dev
))
9562 tg3_netif_start(tp
);
9564 tg3_full_unlock(tp
);
9568 static int tg3_get_coalesce(struct net_device
*dev
, struct ethtool_coalesce
*ec
)
9570 struct tg3
*tp
= netdev_priv(dev
);
9572 memcpy(ec
, &tp
->coal
, sizeof(*ec
));
9576 static int tg3_set_coalesce(struct net_device
*dev
, struct ethtool_coalesce
*ec
)
9578 struct tg3
*tp
= netdev_priv(dev
);
9579 u32 max_rxcoal_tick_int
= 0, max_txcoal_tick_int
= 0;
9580 u32 max_stat_coal_ticks
= 0, min_stat_coal_ticks
= 0;
9582 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
)) {
9583 max_rxcoal_tick_int
= MAX_RXCOAL_TICK_INT
;
9584 max_txcoal_tick_int
= MAX_TXCOAL_TICK_INT
;
9585 max_stat_coal_ticks
= MAX_STAT_COAL_TICKS
;
9586 min_stat_coal_ticks
= MIN_STAT_COAL_TICKS
;
9589 if ((ec
->rx_coalesce_usecs
> MAX_RXCOL_TICKS
) ||
9590 (ec
->tx_coalesce_usecs
> MAX_TXCOL_TICKS
) ||
9591 (ec
->rx_max_coalesced_frames
> MAX_RXMAX_FRAMES
) ||
9592 (ec
->tx_max_coalesced_frames
> MAX_TXMAX_FRAMES
) ||
9593 (ec
->rx_coalesce_usecs_irq
> max_rxcoal_tick_int
) ||
9594 (ec
->tx_coalesce_usecs_irq
> max_txcoal_tick_int
) ||
9595 (ec
->rx_max_coalesced_frames_irq
> MAX_RXCOAL_MAXF_INT
) ||
9596 (ec
->tx_max_coalesced_frames_irq
> MAX_TXCOAL_MAXF_INT
) ||
9597 (ec
->stats_block_coalesce_usecs
> max_stat_coal_ticks
) ||
9598 (ec
->stats_block_coalesce_usecs
< min_stat_coal_ticks
))
9601 /* No rx interrupts will be generated if both are zero */
9602 if ((ec
->rx_coalesce_usecs
== 0) &&
9603 (ec
->rx_max_coalesced_frames
== 0))
9606 /* No tx interrupts will be generated if both are zero */
9607 if ((ec
->tx_coalesce_usecs
== 0) &&
9608 (ec
->tx_max_coalesced_frames
== 0))
9611 /* Only copy relevant parameters, ignore all others. */
9612 tp
->coal
.rx_coalesce_usecs
= ec
->rx_coalesce_usecs
;
9613 tp
->coal
.tx_coalesce_usecs
= ec
->tx_coalesce_usecs
;
9614 tp
->coal
.rx_max_coalesced_frames
= ec
->rx_max_coalesced_frames
;
9615 tp
->coal
.tx_max_coalesced_frames
= ec
->tx_max_coalesced_frames
;
9616 tp
->coal
.rx_coalesce_usecs_irq
= ec
->rx_coalesce_usecs_irq
;
9617 tp
->coal
.tx_coalesce_usecs_irq
= ec
->tx_coalesce_usecs_irq
;
9618 tp
->coal
.rx_max_coalesced_frames_irq
= ec
->rx_max_coalesced_frames_irq
;
9619 tp
->coal
.tx_max_coalesced_frames_irq
= ec
->tx_max_coalesced_frames_irq
;
9620 tp
->coal
.stats_block_coalesce_usecs
= ec
->stats_block_coalesce_usecs
;
9622 if (netif_running(dev
)) {
9623 tg3_full_lock(tp
, 0);
9624 __tg3_set_coalesce(tp
, &tp
->coal
);
9625 tg3_full_unlock(tp
);
9630 static const struct ethtool_ops tg3_ethtool_ops
= {
9631 .get_settings
= tg3_get_settings
,
9632 .set_settings
= tg3_set_settings
,
9633 .get_drvinfo
= tg3_get_drvinfo
,
9634 .get_regs_len
= tg3_get_regs_len
,
9635 .get_regs
= tg3_get_regs
,
9636 .get_wol
= tg3_get_wol
,
9637 .set_wol
= tg3_set_wol
,
9638 .get_msglevel
= tg3_get_msglevel
,
9639 .set_msglevel
= tg3_set_msglevel
,
9640 .nway_reset
= tg3_nway_reset
,
9641 .get_link
= ethtool_op_get_link
,
9642 .get_eeprom_len
= tg3_get_eeprom_len
,
9643 .get_eeprom
= tg3_get_eeprom
,
9644 .set_eeprom
= tg3_set_eeprom
,
9645 .get_ringparam
= tg3_get_ringparam
,
9646 .set_ringparam
= tg3_set_ringparam
,
9647 .get_pauseparam
= tg3_get_pauseparam
,
9648 .set_pauseparam
= tg3_set_pauseparam
,
9649 .get_rx_csum
= tg3_get_rx_csum
,
9650 .set_rx_csum
= tg3_set_rx_csum
,
9651 .set_tx_csum
= tg3_set_tx_csum
,
9652 .set_sg
= ethtool_op_set_sg
,
9653 .set_tso
= tg3_set_tso
,
9654 .self_test
= tg3_self_test
,
9655 .get_strings
= tg3_get_strings
,
9656 .phys_id
= tg3_phys_id
,
9657 .get_ethtool_stats
= tg3_get_ethtool_stats
,
9658 .get_coalesce
= tg3_get_coalesce
,
9659 .set_coalesce
= tg3_set_coalesce
,
9660 .get_sset_count
= tg3_get_sset_count
,
9663 static void __devinit
tg3_get_eeprom_size(struct tg3
*tp
)
9665 u32 cursize
, val
, magic
;
9667 tp
->nvram_size
= EEPROM_CHIP_SIZE
;
9669 if (tg3_nvram_read_swab(tp
, 0, &magic
) != 0)
9672 if ((magic
!= TG3_EEPROM_MAGIC
) &&
9673 ((magic
& TG3_EEPROM_MAGIC_FW_MSK
) != TG3_EEPROM_MAGIC_FW
) &&
9674 ((magic
& TG3_EEPROM_MAGIC_HW_MSK
) != TG3_EEPROM_MAGIC_HW
))
9678 * Size the chip by reading offsets at increasing powers of two.
9679 * When we encounter our validation signature, we know the addressing
9680 * has wrapped around, and thus have our chip size.
9684 while (cursize
< tp
->nvram_size
) {
9685 if (tg3_nvram_read_swab(tp
, cursize
, &val
) != 0)
9694 tp
->nvram_size
= cursize
;
9697 static void __devinit
tg3_get_nvram_size(struct tg3
*tp
)
9701 if (tg3_nvram_read_swab(tp
, 0, &val
) != 0)
9704 /* Selfboot format */
9705 if (val
!= TG3_EEPROM_MAGIC
) {
9706 tg3_get_eeprom_size(tp
);
9710 if (tg3_nvram_read(tp
, 0xf0, &val
) == 0) {
9712 tp
->nvram_size
= (val
>> 16) * 1024;
9716 tp
->nvram_size
= 0x80000;
9719 static void __devinit
tg3_get_nvram_info(struct tg3
*tp
)
9723 nvcfg1
= tr32(NVRAM_CFG1
);
9724 if (nvcfg1
& NVRAM_CFG1_FLASHIF_ENAB
) {
9725 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9728 nvcfg1
&= ~NVRAM_CFG1_COMPAT_BYPASS
;
9729 tw32(NVRAM_CFG1
, nvcfg1
);
9732 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
) ||
9733 (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)) {
9734 switch (nvcfg1
& NVRAM_CFG1_VENDOR_MASK
) {
9735 case FLASH_VENDOR_ATMEL_FLASH_BUFFERED
:
9736 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9737 tp
->nvram_pagesize
= ATMEL_AT45DB0X1B_PAGE_SIZE
;
9738 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9740 case FLASH_VENDOR_ATMEL_FLASH_UNBUFFERED
:
9741 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9742 tp
->nvram_pagesize
= ATMEL_AT25F512_PAGE_SIZE
;
9744 case FLASH_VENDOR_ATMEL_EEPROM
:
9745 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9746 tp
->nvram_pagesize
= ATMEL_AT24C512_CHIP_SIZE
;
9747 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9749 case FLASH_VENDOR_ST
:
9750 tp
->nvram_jedecnum
= JEDEC_ST
;
9751 tp
->nvram_pagesize
= ST_M45PEX0_PAGE_SIZE
;
9752 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9754 case FLASH_VENDOR_SAIFUN
:
9755 tp
->nvram_jedecnum
= JEDEC_SAIFUN
;
9756 tp
->nvram_pagesize
= SAIFUN_SA25F0XX_PAGE_SIZE
;
9758 case FLASH_VENDOR_SST_SMALL
:
9759 case FLASH_VENDOR_SST_LARGE
:
9760 tp
->nvram_jedecnum
= JEDEC_SST
;
9761 tp
->nvram_pagesize
= SST_25VF0X0_PAGE_SIZE
;
9766 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9767 tp
->nvram_pagesize
= ATMEL_AT45DB0X1B_PAGE_SIZE
;
9768 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9772 static void __devinit
tg3_get_5752_nvram_info(struct tg3
*tp
)
9776 nvcfg1
= tr32(NVRAM_CFG1
);
9778 /* NVRAM protection for TPM */
9779 if (nvcfg1
& (1 << 27))
9780 tp
->tg3_flags2
|= TG3_FLG2_PROTECTED_NVRAM
;
9782 switch (nvcfg1
& NVRAM_CFG1_5752VENDOR_MASK
) {
9783 case FLASH_5752VENDOR_ATMEL_EEPROM_64KHZ
:
9784 case FLASH_5752VENDOR_ATMEL_EEPROM_376KHZ
:
9785 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9786 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9788 case FLASH_5752VENDOR_ATMEL_FLASH_BUFFERED
:
9789 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9790 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9791 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9793 case FLASH_5752VENDOR_ST_M45PE10
:
9794 case FLASH_5752VENDOR_ST_M45PE20
:
9795 case FLASH_5752VENDOR_ST_M45PE40
:
9796 tp
->nvram_jedecnum
= JEDEC_ST
;
9797 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9798 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9802 if (tp
->tg3_flags2
& TG3_FLG2_FLASH
) {
9803 switch (nvcfg1
& NVRAM_CFG1_5752PAGE_SIZE_MASK
) {
9804 case FLASH_5752PAGE_SIZE_256
:
9805 tp
->nvram_pagesize
= 256;
9807 case FLASH_5752PAGE_SIZE_512
:
9808 tp
->nvram_pagesize
= 512;
9810 case FLASH_5752PAGE_SIZE_1K
:
9811 tp
->nvram_pagesize
= 1024;
9813 case FLASH_5752PAGE_SIZE_2K
:
9814 tp
->nvram_pagesize
= 2048;
9816 case FLASH_5752PAGE_SIZE_4K
:
9817 tp
->nvram_pagesize
= 4096;
9819 case FLASH_5752PAGE_SIZE_264
:
9820 tp
->nvram_pagesize
= 264;
9825 /* For eeprom, set pagesize to maximum eeprom size */
9826 tp
->nvram_pagesize
= ATMEL_AT24C512_CHIP_SIZE
;
9828 nvcfg1
&= ~NVRAM_CFG1_COMPAT_BYPASS
;
9829 tw32(NVRAM_CFG1
, nvcfg1
);
9833 static void __devinit
tg3_get_5755_nvram_info(struct tg3
*tp
)
9835 u32 nvcfg1
, protect
= 0;
9837 nvcfg1
= tr32(NVRAM_CFG1
);
9839 /* NVRAM protection for TPM */
9840 if (nvcfg1
& (1 << 27)) {
9841 tp
->tg3_flags2
|= TG3_FLG2_PROTECTED_NVRAM
;
9845 nvcfg1
&= NVRAM_CFG1_5752VENDOR_MASK
;
9847 case FLASH_5755VENDOR_ATMEL_FLASH_1
:
9848 case FLASH_5755VENDOR_ATMEL_FLASH_2
:
9849 case FLASH_5755VENDOR_ATMEL_FLASH_3
:
9850 case FLASH_5755VENDOR_ATMEL_FLASH_5
:
9851 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9852 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9853 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9854 tp
->nvram_pagesize
= 264;
9855 if (nvcfg1
== FLASH_5755VENDOR_ATMEL_FLASH_1
||
9856 nvcfg1
== FLASH_5755VENDOR_ATMEL_FLASH_5
)
9857 tp
->nvram_size
= (protect
? 0x3e200 : 0x80000);
9858 else if (nvcfg1
== FLASH_5755VENDOR_ATMEL_FLASH_2
)
9859 tp
->nvram_size
= (protect
? 0x1f200 : 0x40000);
9861 tp
->nvram_size
= (protect
? 0x1f200 : 0x20000);
9863 case FLASH_5752VENDOR_ST_M45PE10
:
9864 case FLASH_5752VENDOR_ST_M45PE20
:
9865 case FLASH_5752VENDOR_ST_M45PE40
:
9866 tp
->nvram_jedecnum
= JEDEC_ST
;
9867 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9868 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9869 tp
->nvram_pagesize
= 256;
9870 if (nvcfg1
== FLASH_5752VENDOR_ST_M45PE10
)
9871 tp
->nvram_size
= (protect
? 0x10000 : 0x20000);
9872 else if (nvcfg1
== FLASH_5752VENDOR_ST_M45PE20
)
9873 tp
->nvram_size
= (protect
? 0x10000 : 0x40000);
9875 tp
->nvram_size
= (protect
? 0x20000 : 0x80000);
9880 static void __devinit
tg3_get_5787_nvram_info(struct tg3
*tp
)
9884 nvcfg1
= tr32(NVRAM_CFG1
);
9886 switch (nvcfg1
& NVRAM_CFG1_5752VENDOR_MASK
) {
9887 case FLASH_5787VENDOR_ATMEL_EEPROM_64KHZ
:
9888 case FLASH_5787VENDOR_ATMEL_EEPROM_376KHZ
:
9889 case FLASH_5787VENDOR_MICRO_EEPROM_64KHZ
:
9890 case FLASH_5787VENDOR_MICRO_EEPROM_376KHZ
:
9891 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9892 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9893 tp
->nvram_pagesize
= ATMEL_AT24C512_CHIP_SIZE
;
9895 nvcfg1
&= ~NVRAM_CFG1_COMPAT_BYPASS
;
9896 tw32(NVRAM_CFG1
, nvcfg1
);
9898 case FLASH_5752VENDOR_ATMEL_FLASH_BUFFERED
:
9899 case FLASH_5755VENDOR_ATMEL_FLASH_1
:
9900 case FLASH_5755VENDOR_ATMEL_FLASH_2
:
9901 case FLASH_5755VENDOR_ATMEL_FLASH_3
:
9902 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9903 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9904 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9905 tp
->nvram_pagesize
= 264;
9907 case FLASH_5752VENDOR_ST_M45PE10
:
9908 case FLASH_5752VENDOR_ST_M45PE20
:
9909 case FLASH_5752VENDOR_ST_M45PE40
:
9910 tp
->nvram_jedecnum
= JEDEC_ST
;
9911 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9912 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9913 tp
->nvram_pagesize
= 256;
9918 static void __devinit
tg3_get_5761_nvram_info(struct tg3
*tp
)
9920 u32 nvcfg1
, protect
= 0;
9922 nvcfg1
= tr32(NVRAM_CFG1
);
9924 /* NVRAM protection for TPM */
9925 if (nvcfg1
& (1 << 27)) {
9926 tp
->tg3_flags2
|= TG3_FLG2_PROTECTED_NVRAM
;
9930 nvcfg1
&= NVRAM_CFG1_5752VENDOR_MASK
;
9932 case FLASH_5761VENDOR_ATMEL_ADB021D
:
9933 case FLASH_5761VENDOR_ATMEL_ADB041D
:
9934 case FLASH_5761VENDOR_ATMEL_ADB081D
:
9935 case FLASH_5761VENDOR_ATMEL_ADB161D
:
9936 case FLASH_5761VENDOR_ATMEL_MDB021D
:
9937 case FLASH_5761VENDOR_ATMEL_MDB041D
:
9938 case FLASH_5761VENDOR_ATMEL_MDB081D
:
9939 case FLASH_5761VENDOR_ATMEL_MDB161D
:
9940 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9941 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9942 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9943 tp
->tg3_flags3
|= TG3_FLG3_NO_NVRAM_ADDR_TRANS
;
9944 tp
->nvram_pagesize
= 256;
9946 case FLASH_5761VENDOR_ST_A_M45PE20
:
9947 case FLASH_5761VENDOR_ST_A_M45PE40
:
9948 case FLASH_5761VENDOR_ST_A_M45PE80
:
9949 case FLASH_5761VENDOR_ST_A_M45PE16
:
9950 case FLASH_5761VENDOR_ST_M_M45PE20
:
9951 case FLASH_5761VENDOR_ST_M_M45PE40
:
9952 case FLASH_5761VENDOR_ST_M_M45PE80
:
9953 case FLASH_5761VENDOR_ST_M_M45PE16
:
9954 tp
->nvram_jedecnum
= JEDEC_ST
;
9955 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9956 tp
->tg3_flags2
|= TG3_FLG2_FLASH
;
9957 tp
->nvram_pagesize
= 256;
9962 tp
->nvram_size
= tr32(NVRAM_ADDR_LOCKOUT
);
9965 case FLASH_5761VENDOR_ATMEL_ADB161D
:
9966 case FLASH_5761VENDOR_ATMEL_MDB161D
:
9967 case FLASH_5761VENDOR_ST_A_M45PE16
:
9968 case FLASH_5761VENDOR_ST_M_M45PE16
:
9969 tp
->nvram_size
= 0x100000;
9971 case FLASH_5761VENDOR_ATMEL_ADB081D
:
9972 case FLASH_5761VENDOR_ATMEL_MDB081D
:
9973 case FLASH_5761VENDOR_ST_A_M45PE80
:
9974 case FLASH_5761VENDOR_ST_M_M45PE80
:
9975 tp
->nvram_size
= 0x80000;
9977 case FLASH_5761VENDOR_ATMEL_ADB041D
:
9978 case FLASH_5761VENDOR_ATMEL_MDB041D
:
9979 case FLASH_5761VENDOR_ST_A_M45PE40
:
9980 case FLASH_5761VENDOR_ST_M_M45PE40
:
9981 tp
->nvram_size
= 0x40000;
9983 case FLASH_5761VENDOR_ATMEL_ADB021D
:
9984 case FLASH_5761VENDOR_ATMEL_MDB021D
:
9985 case FLASH_5761VENDOR_ST_A_M45PE20
:
9986 case FLASH_5761VENDOR_ST_M_M45PE20
:
9987 tp
->nvram_size
= 0x20000;
9993 static void __devinit
tg3_get_5906_nvram_info(struct tg3
*tp
)
9995 tp
->nvram_jedecnum
= JEDEC_ATMEL
;
9996 tp
->tg3_flags
|= TG3_FLAG_NVRAM_BUFFERED
;
9997 tp
->nvram_pagesize
= ATMEL_AT24C512_CHIP_SIZE
;
10000 /* Chips other than 5700/5701 use the NVRAM for fetching info. */
10001 static void __devinit
tg3_nvram_init(struct tg3
*tp
)
10003 tw32_f(GRC_EEPROM_ADDR
,
10004 (EEPROM_ADDR_FSM_RESET
|
10005 (EEPROM_DEFAULT_CLOCK_PERIOD
<<
10006 EEPROM_ADDR_CLKPERD_SHIFT
)));
10010 /* Enable seeprom accesses. */
10011 tw32_f(GRC_LOCAL_CTRL
,
10012 tr32(GRC_LOCAL_CTRL
) | GRC_LCLCTRL_AUTO_SEEPROM
);
10015 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
&&
10016 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5701
) {
10017 tp
->tg3_flags
|= TG3_FLAG_NVRAM
;
10019 if (tg3_nvram_lock(tp
)) {
10020 printk(KERN_WARNING PFX
"%s: Cannot get nvarm lock, "
10021 "tg3_nvram_init failed.\n", tp
->dev
->name
);
10024 tg3_enable_nvram_access(tp
);
10026 tp
->nvram_size
= 0;
10028 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
)
10029 tg3_get_5752_nvram_info(tp
);
10030 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
)
10031 tg3_get_5755_nvram_info(tp
);
10032 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
10033 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
)
10034 tg3_get_5787_nvram_info(tp
);
10035 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
10036 tg3_get_5761_nvram_info(tp
);
10037 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
10038 tg3_get_5906_nvram_info(tp
);
10040 tg3_get_nvram_info(tp
);
10042 if (tp
->nvram_size
== 0)
10043 tg3_get_nvram_size(tp
);
10045 tg3_disable_nvram_access(tp
);
10046 tg3_nvram_unlock(tp
);
10049 tp
->tg3_flags
&= ~(TG3_FLAG_NVRAM
| TG3_FLAG_NVRAM_BUFFERED
);
10051 tg3_get_eeprom_size(tp
);
10055 static int tg3_nvram_read_using_eeprom(struct tg3
*tp
,
10056 u32 offset
, u32
*val
)
10061 if (offset
> EEPROM_ADDR_ADDR_MASK
||
10065 tmp
= tr32(GRC_EEPROM_ADDR
) & ~(EEPROM_ADDR_ADDR_MASK
|
10066 EEPROM_ADDR_DEVID_MASK
|
10068 tw32(GRC_EEPROM_ADDR
,
10070 (0 << EEPROM_ADDR_DEVID_SHIFT
) |
10071 ((offset
<< EEPROM_ADDR_ADDR_SHIFT
) &
10072 EEPROM_ADDR_ADDR_MASK
) |
10073 EEPROM_ADDR_READ
| EEPROM_ADDR_START
);
10075 for (i
= 0; i
< 1000; i
++) {
10076 tmp
= tr32(GRC_EEPROM_ADDR
);
10078 if (tmp
& EEPROM_ADDR_COMPLETE
)
10082 if (!(tmp
& EEPROM_ADDR_COMPLETE
))
10085 *val
= tr32(GRC_EEPROM_DATA
);
10089 #define NVRAM_CMD_TIMEOUT 10000
10091 static int tg3_nvram_exec_cmd(struct tg3
*tp
, u32 nvram_cmd
)
10095 tw32(NVRAM_CMD
, nvram_cmd
);
10096 for (i
= 0; i
< NVRAM_CMD_TIMEOUT
; i
++) {
10098 if (tr32(NVRAM_CMD
) & NVRAM_CMD_DONE
) {
10103 if (i
== NVRAM_CMD_TIMEOUT
) {
10109 static u32
tg3_nvram_phys_addr(struct tg3
*tp
, u32 addr
)
10111 if ((tp
->tg3_flags
& TG3_FLAG_NVRAM
) &&
10112 (tp
->tg3_flags
& TG3_FLAG_NVRAM_BUFFERED
) &&
10113 (tp
->tg3_flags2
& TG3_FLG2_FLASH
) &&
10114 !(tp
->tg3_flags3
& TG3_FLG3_NO_NVRAM_ADDR_TRANS
) &&
10115 (tp
->nvram_jedecnum
== JEDEC_ATMEL
))
10117 addr
= ((addr
/ tp
->nvram_pagesize
) <<
10118 ATMEL_AT45DB0X1B_PAGE_POS
) +
10119 (addr
% tp
->nvram_pagesize
);
10124 static u32
tg3_nvram_logical_addr(struct tg3
*tp
, u32 addr
)
10126 if ((tp
->tg3_flags
& TG3_FLAG_NVRAM
) &&
10127 (tp
->tg3_flags
& TG3_FLAG_NVRAM_BUFFERED
) &&
10128 (tp
->tg3_flags2
& TG3_FLG2_FLASH
) &&
10129 !(tp
->tg3_flags3
& TG3_FLG3_NO_NVRAM_ADDR_TRANS
) &&
10130 (tp
->nvram_jedecnum
== JEDEC_ATMEL
))
10132 addr
= ((addr
>> ATMEL_AT45DB0X1B_PAGE_POS
) *
10133 tp
->nvram_pagesize
) +
10134 (addr
& ((1 << ATMEL_AT45DB0X1B_PAGE_POS
) - 1));
10139 static int tg3_nvram_read(struct tg3
*tp
, u32 offset
, u32
*val
)
10143 if (!(tp
->tg3_flags
& TG3_FLAG_NVRAM
))
10144 return tg3_nvram_read_using_eeprom(tp
, offset
, val
);
10146 offset
= tg3_nvram_phys_addr(tp
, offset
);
10148 if (offset
> NVRAM_ADDR_MSK
)
10151 ret
= tg3_nvram_lock(tp
);
10155 tg3_enable_nvram_access(tp
);
10157 tw32(NVRAM_ADDR
, offset
);
10158 ret
= tg3_nvram_exec_cmd(tp
, NVRAM_CMD_RD
| NVRAM_CMD_GO
|
10159 NVRAM_CMD_FIRST
| NVRAM_CMD_LAST
| NVRAM_CMD_DONE
);
10162 *val
= swab32(tr32(NVRAM_RDDATA
));
10164 tg3_disable_nvram_access(tp
);
10166 tg3_nvram_unlock(tp
);
10171 static int tg3_nvram_read_le(struct tg3
*tp
, u32 offset
, __le32
*val
)
10174 int res
= tg3_nvram_read(tp
, offset
, &v
);
10176 *val
= cpu_to_le32(v
);
10180 static int tg3_nvram_read_swab(struct tg3
*tp
, u32 offset
, u32
*val
)
10185 err
= tg3_nvram_read(tp
, offset
, &tmp
);
10186 *val
= swab32(tmp
);
10190 static int tg3_nvram_write_block_using_eeprom(struct tg3
*tp
,
10191 u32 offset
, u32 len
, u8
*buf
)
10196 for (i
= 0; i
< len
; i
+= 4) {
10202 memcpy(&data
, buf
+ i
, 4);
10204 tw32(GRC_EEPROM_DATA
, le32_to_cpu(data
));
10206 val
= tr32(GRC_EEPROM_ADDR
);
10207 tw32(GRC_EEPROM_ADDR
, val
| EEPROM_ADDR_COMPLETE
);
10209 val
&= ~(EEPROM_ADDR_ADDR_MASK
| EEPROM_ADDR_DEVID_MASK
|
10211 tw32(GRC_EEPROM_ADDR
, val
|
10212 (0 << EEPROM_ADDR_DEVID_SHIFT
) |
10213 (addr
& EEPROM_ADDR_ADDR_MASK
) |
10214 EEPROM_ADDR_START
|
10215 EEPROM_ADDR_WRITE
);
10217 for (j
= 0; j
< 1000; j
++) {
10218 val
= tr32(GRC_EEPROM_ADDR
);
10220 if (val
& EEPROM_ADDR_COMPLETE
)
10224 if (!(val
& EEPROM_ADDR_COMPLETE
)) {
10233 /* offset and length are dword aligned */
10234 static int tg3_nvram_write_block_unbuffered(struct tg3
*tp
, u32 offset
, u32 len
,
10238 u32 pagesize
= tp
->nvram_pagesize
;
10239 u32 pagemask
= pagesize
- 1;
10243 tmp
= kmalloc(pagesize
, GFP_KERNEL
);
10249 u32 phy_addr
, page_off
, size
;
10251 phy_addr
= offset
& ~pagemask
;
10253 for (j
= 0; j
< pagesize
; j
+= 4) {
10254 if ((ret
= tg3_nvram_read_le(tp
, phy_addr
+ j
,
10255 (__le32
*) (tmp
+ j
))))
10261 page_off
= offset
& pagemask
;
10268 memcpy(tmp
+ page_off
, buf
, size
);
10270 offset
= offset
+ (pagesize
- page_off
);
10272 tg3_enable_nvram_access(tp
);
10275 * Before we can erase the flash page, we need
10276 * to issue a special "write enable" command.
10278 nvram_cmd
= NVRAM_CMD_WREN
| NVRAM_CMD_GO
| NVRAM_CMD_DONE
;
10280 if (tg3_nvram_exec_cmd(tp
, nvram_cmd
))
10283 /* Erase the target page */
10284 tw32(NVRAM_ADDR
, phy_addr
);
10286 nvram_cmd
= NVRAM_CMD_GO
| NVRAM_CMD_DONE
| NVRAM_CMD_WR
|
10287 NVRAM_CMD_FIRST
| NVRAM_CMD_LAST
| NVRAM_CMD_ERASE
;
10289 if (tg3_nvram_exec_cmd(tp
, nvram_cmd
))
10292 /* Issue another write enable to start the write. */
10293 nvram_cmd
= NVRAM_CMD_WREN
| NVRAM_CMD_GO
| NVRAM_CMD_DONE
;
10295 if (tg3_nvram_exec_cmd(tp
, nvram_cmd
))
10298 for (j
= 0; j
< pagesize
; j
+= 4) {
10301 data
= *((__be32
*) (tmp
+ j
));
10302 /* swab32(le32_to_cpu(data)), actually */
10303 tw32(NVRAM_WRDATA
, be32_to_cpu(data
));
10305 tw32(NVRAM_ADDR
, phy_addr
+ j
);
10307 nvram_cmd
= NVRAM_CMD_GO
| NVRAM_CMD_DONE
|
10311 nvram_cmd
|= NVRAM_CMD_FIRST
;
10312 else if (j
== (pagesize
- 4))
10313 nvram_cmd
|= NVRAM_CMD_LAST
;
10315 if ((ret
= tg3_nvram_exec_cmd(tp
, nvram_cmd
)))
10322 nvram_cmd
= NVRAM_CMD_WRDI
| NVRAM_CMD_GO
| NVRAM_CMD_DONE
;
10323 tg3_nvram_exec_cmd(tp
, nvram_cmd
);
10330 /* offset and length are dword aligned */
10331 static int tg3_nvram_write_block_buffered(struct tg3
*tp
, u32 offset
, u32 len
,
10336 for (i
= 0; i
< len
; i
+= 4, offset
+= 4) {
10337 u32 page_off
, phy_addr
, nvram_cmd
;
10340 memcpy(&data
, buf
+ i
, 4);
10341 tw32(NVRAM_WRDATA
, be32_to_cpu(data
));
10343 page_off
= offset
% tp
->nvram_pagesize
;
10345 phy_addr
= tg3_nvram_phys_addr(tp
, offset
);
10347 tw32(NVRAM_ADDR
, phy_addr
);
10349 nvram_cmd
= NVRAM_CMD_GO
| NVRAM_CMD_DONE
| NVRAM_CMD_WR
;
10351 if ((page_off
== 0) || (i
== 0))
10352 nvram_cmd
|= NVRAM_CMD_FIRST
;
10353 if (page_off
== (tp
->nvram_pagesize
- 4))
10354 nvram_cmd
|= NVRAM_CMD_LAST
;
10356 if (i
== (len
- 4))
10357 nvram_cmd
|= NVRAM_CMD_LAST
;
10359 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5752
) &&
10360 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5755
) &&
10361 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5787
) &&
10362 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5784
) &&
10363 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5761
) &&
10364 (tp
->nvram_jedecnum
== JEDEC_ST
) &&
10365 (nvram_cmd
& NVRAM_CMD_FIRST
)) {
10367 if ((ret
= tg3_nvram_exec_cmd(tp
,
10368 NVRAM_CMD_WREN
| NVRAM_CMD_GO
|
10373 if (!(tp
->tg3_flags2
& TG3_FLG2_FLASH
)) {
10374 /* We always do complete word writes to eeprom. */
10375 nvram_cmd
|= (NVRAM_CMD_FIRST
| NVRAM_CMD_LAST
);
10378 if ((ret
= tg3_nvram_exec_cmd(tp
, nvram_cmd
)))
10384 /* offset and length are dword aligned */
10385 static int tg3_nvram_write_block(struct tg3
*tp
, u32 offset
, u32 len
, u8
*buf
)
10389 if (tp
->tg3_flags
& TG3_FLAG_EEPROM_WRITE_PROT
) {
10390 tw32_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
&
10391 ~GRC_LCLCTRL_GPIO_OUTPUT1
);
10395 if (!(tp
->tg3_flags
& TG3_FLAG_NVRAM
)) {
10396 ret
= tg3_nvram_write_block_using_eeprom(tp
, offset
, len
, buf
);
10401 ret
= tg3_nvram_lock(tp
);
10405 tg3_enable_nvram_access(tp
);
10406 if ((tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) &&
10407 !(tp
->tg3_flags2
& TG3_FLG2_PROTECTED_NVRAM
))
10408 tw32(NVRAM_WRITE1
, 0x406);
10410 grc_mode
= tr32(GRC_MODE
);
10411 tw32(GRC_MODE
, grc_mode
| GRC_MODE_NVRAM_WR_ENABLE
);
10413 if ((tp
->tg3_flags
& TG3_FLAG_NVRAM_BUFFERED
) ||
10414 !(tp
->tg3_flags2
& TG3_FLG2_FLASH
)) {
10416 ret
= tg3_nvram_write_block_buffered(tp
, offset
, len
,
10420 ret
= tg3_nvram_write_block_unbuffered(tp
, offset
, len
,
10424 grc_mode
= tr32(GRC_MODE
);
10425 tw32(GRC_MODE
, grc_mode
& ~GRC_MODE_NVRAM_WR_ENABLE
);
10427 tg3_disable_nvram_access(tp
);
10428 tg3_nvram_unlock(tp
);
10431 if (tp
->tg3_flags
& TG3_FLAG_EEPROM_WRITE_PROT
) {
10432 tw32_f(GRC_LOCAL_CTRL
, tp
->grc_local_ctrl
);
10439 struct subsys_tbl_ent
{
10440 u16 subsys_vendor
, subsys_devid
;
10444 static struct subsys_tbl_ent subsys_id_to_phy_id
[] = {
10445 /* Broadcom boards. */
10446 { PCI_VENDOR_ID_BROADCOM
, 0x1644, PHY_ID_BCM5401
}, /* BCM95700A6 */
10447 { PCI_VENDOR_ID_BROADCOM
, 0x0001, PHY_ID_BCM5701
}, /* BCM95701A5 */
10448 { PCI_VENDOR_ID_BROADCOM
, 0x0002, PHY_ID_BCM8002
}, /* BCM95700T6 */
10449 { PCI_VENDOR_ID_BROADCOM
, 0x0003, 0 }, /* BCM95700A9 */
10450 { PCI_VENDOR_ID_BROADCOM
, 0x0005, PHY_ID_BCM5701
}, /* BCM95701T1 */
10451 { PCI_VENDOR_ID_BROADCOM
, 0x0006, PHY_ID_BCM5701
}, /* BCM95701T8 */
10452 { PCI_VENDOR_ID_BROADCOM
, 0x0007, 0 }, /* BCM95701A7 */
10453 { PCI_VENDOR_ID_BROADCOM
, 0x0008, PHY_ID_BCM5701
}, /* BCM95701A10 */
10454 { PCI_VENDOR_ID_BROADCOM
, 0x8008, PHY_ID_BCM5701
}, /* BCM95701A12 */
10455 { PCI_VENDOR_ID_BROADCOM
, 0x0009, PHY_ID_BCM5703
}, /* BCM95703Ax1 */
10456 { PCI_VENDOR_ID_BROADCOM
, 0x8009, PHY_ID_BCM5703
}, /* BCM95703Ax2 */
10459 { PCI_VENDOR_ID_3COM
, 0x1000, PHY_ID_BCM5401
}, /* 3C996T */
10460 { PCI_VENDOR_ID_3COM
, 0x1006, PHY_ID_BCM5701
}, /* 3C996BT */
10461 { PCI_VENDOR_ID_3COM
, 0x1004, 0 }, /* 3C996SX */
10462 { PCI_VENDOR_ID_3COM
, 0x1007, PHY_ID_BCM5701
}, /* 3C1000T */
10463 { PCI_VENDOR_ID_3COM
, 0x1008, PHY_ID_BCM5701
}, /* 3C940BR01 */
10466 { PCI_VENDOR_ID_DELL
, 0x00d1, PHY_ID_BCM5401
}, /* VIPER */
10467 { PCI_VENDOR_ID_DELL
, 0x0106, PHY_ID_BCM5401
}, /* JAGUAR */
10468 { PCI_VENDOR_ID_DELL
, 0x0109, PHY_ID_BCM5411
}, /* MERLOT */
10469 { PCI_VENDOR_ID_DELL
, 0x010a, PHY_ID_BCM5411
}, /* SLIM_MERLOT */
10471 /* Compaq boards. */
10472 { PCI_VENDOR_ID_COMPAQ
, 0x007c, PHY_ID_BCM5701
}, /* BANSHEE */
10473 { PCI_VENDOR_ID_COMPAQ
, 0x009a, PHY_ID_BCM5701
}, /* BANSHEE_2 */
10474 { PCI_VENDOR_ID_COMPAQ
, 0x007d, 0 }, /* CHANGELING */
10475 { PCI_VENDOR_ID_COMPAQ
, 0x0085, PHY_ID_BCM5701
}, /* NC7780 */
10476 { PCI_VENDOR_ID_COMPAQ
, 0x0099, PHY_ID_BCM5701
}, /* NC7780_2 */
10479 { PCI_VENDOR_ID_IBM
, 0x0281, 0 } /* IBM??? */
10482 static inline struct subsys_tbl_ent
*lookup_by_subsys(struct tg3
*tp
)
10486 for (i
= 0; i
< ARRAY_SIZE(subsys_id_to_phy_id
); i
++) {
10487 if ((subsys_id_to_phy_id
[i
].subsys_vendor
==
10488 tp
->pdev
->subsystem_vendor
) &&
10489 (subsys_id_to_phy_id
[i
].subsys_devid
==
10490 tp
->pdev
->subsystem_device
))
10491 return &subsys_id_to_phy_id
[i
];
10496 static void __devinit
tg3_get_eeprom_hw_cfg(struct tg3
*tp
)
10501 /* On some early chips the SRAM cannot be accessed in D3hot state,
10502 * so need make sure we're in D0.
10504 pci_read_config_word(tp
->pdev
, tp
->pm_cap
+ PCI_PM_CTRL
, &pmcsr
);
10505 pmcsr
&= ~PCI_PM_CTRL_STATE_MASK
;
10506 pci_write_config_word(tp
->pdev
, tp
->pm_cap
+ PCI_PM_CTRL
, pmcsr
);
10509 /* Make sure register accesses (indirect or otherwise)
10510 * will function correctly.
10512 pci_write_config_dword(tp
->pdev
, TG3PCI_MISC_HOST_CTRL
,
10513 tp
->misc_host_ctrl
);
10515 /* The memory arbiter has to be enabled in order for SRAM accesses
10516 * to succeed. Normally on powerup the tg3 chip firmware will make
10517 * sure it is enabled, but other entities such as system netboot
10518 * code might disable it.
10520 val
= tr32(MEMARB_MODE
);
10521 tw32(MEMARB_MODE
, val
| MEMARB_MODE_ENABLE
);
10523 tp
->phy_id
= PHY_ID_INVALID
;
10524 tp
->led_ctrl
= LED_CTRL_MODE_PHY_1
;
10526 /* Assume an onboard device and WOL capable by default. */
10527 tp
->tg3_flags
|= TG3_FLAG_EEPROM_WRITE_PROT
| TG3_FLAG_WOL_CAP
;
10529 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
10530 if (!(tr32(PCIE_TRANSACTION_CFG
) & PCIE_TRANS_CFG_LOM
)) {
10531 tp
->tg3_flags
&= ~TG3_FLAG_EEPROM_WRITE_PROT
;
10532 tp
->tg3_flags2
|= TG3_FLG2_IS_NIC
;
10534 val
= tr32(VCPU_CFGSHDW
);
10535 if (val
& VCPU_CFGSHDW_ASPM_DBNC
)
10536 tp
->tg3_flags
|= TG3_FLAG_ASPM_WORKAROUND
;
10537 if ((val
& VCPU_CFGSHDW_WOL_ENABLE
) &&
10538 (val
& VCPU_CFGSHDW_WOL_MAGPKT
))
10539 tp
->tg3_flags
|= TG3_FLAG_WOL_ENABLE
;
10543 tg3_read_mem(tp
, NIC_SRAM_DATA_SIG
, &val
);
10544 if (val
== NIC_SRAM_DATA_SIG_MAGIC
) {
10545 u32 nic_cfg
, led_cfg
;
10546 u32 nic_phy_id
, ver
, cfg2
= 0, eeprom_phy_id
;
10547 int eeprom_phy_serdes
= 0;
10549 tg3_read_mem(tp
, NIC_SRAM_DATA_CFG
, &nic_cfg
);
10550 tp
->nic_sram_data_cfg
= nic_cfg
;
10552 tg3_read_mem(tp
, NIC_SRAM_DATA_VER
, &ver
);
10553 ver
>>= NIC_SRAM_DATA_VER_SHIFT
;
10554 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
) &&
10555 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5701
) &&
10556 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5703
) &&
10557 (ver
> 0) && (ver
< 0x100))
10558 tg3_read_mem(tp
, NIC_SRAM_DATA_CFG_2
, &cfg2
);
10560 if ((nic_cfg
& NIC_SRAM_DATA_CFG_PHY_TYPE_MASK
) ==
10561 NIC_SRAM_DATA_CFG_PHY_TYPE_FIBER
)
10562 eeprom_phy_serdes
= 1;
10564 tg3_read_mem(tp
, NIC_SRAM_DATA_PHY_ID
, &nic_phy_id
);
10565 if (nic_phy_id
!= 0) {
10566 u32 id1
= nic_phy_id
& NIC_SRAM_DATA_PHY_ID1_MASK
;
10567 u32 id2
= nic_phy_id
& NIC_SRAM_DATA_PHY_ID2_MASK
;
10569 eeprom_phy_id
= (id1
>> 16) << 10;
10570 eeprom_phy_id
|= (id2
& 0xfc00) << 16;
10571 eeprom_phy_id
|= (id2
& 0x03ff) << 0;
10575 tp
->phy_id
= eeprom_phy_id
;
10576 if (eeprom_phy_serdes
) {
10577 if (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)
10578 tp
->tg3_flags2
|= TG3_FLG2_MII_SERDES
;
10580 tp
->tg3_flags2
|= TG3_FLG2_PHY_SERDES
;
10583 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
)
10584 led_cfg
= cfg2
& (NIC_SRAM_DATA_CFG_LED_MODE_MASK
|
10585 SHASTA_EXT_LED_MODE_MASK
);
10587 led_cfg
= nic_cfg
& NIC_SRAM_DATA_CFG_LED_MODE_MASK
;
10591 case NIC_SRAM_DATA_CFG_LED_MODE_PHY_1
:
10592 tp
->led_ctrl
= LED_CTRL_MODE_PHY_1
;
10595 case NIC_SRAM_DATA_CFG_LED_MODE_PHY_2
:
10596 tp
->led_ctrl
= LED_CTRL_MODE_PHY_2
;
10599 case NIC_SRAM_DATA_CFG_LED_MODE_MAC
:
10600 tp
->led_ctrl
= LED_CTRL_MODE_MAC
;
10602 /* Default to PHY_1_MODE if 0 (MAC_MODE) is
10603 * read on some older 5700/5701 bootcode.
10605 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
10607 GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
10609 tp
->led_ctrl
= LED_CTRL_MODE_PHY_1
;
10613 case SHASTA_EXT_LED_SHARED
:
10614 tp
->led_ctrl
= LED_CTRL_MODE_SHARED
;
10615 if (tp
->pci_chip_rev_id
!= CHIPREV_ID_5750_A0
&&
10616 tp
->pci_chip_rev_id
!= CHIPREV_ID_5750_A1
)
10617 tp
->led_ctrl
|= (LED_CTRL_MODE_PHY_1
|
10618 LED_CTRL_MODE_PHY_2
);
10621 case SHASTA_EXT_LED_MAC
:
10622 tp
->led_ctrl
= LED_CTRL_MODE_SHASTA_MAC
;
10625 case SHASTA_EXT_LED_COMBO
:
10626 tp
->led_ctrl
= LED_CTRL_MODE_COMBO
;
10627 if (tp
->pci_chip_rev_id
!= CHIPREV_ID_5750_A0
)
10628 tp
->led_ctrl
|= (LED_CTRL_MODE_PHY_1
|
10629 LED_CTRL_MODE_PHY_2
);
10634 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
10635 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) &&
10636 tp
->pdev
->subsystem_vendor
== PCI_VENDOR_ID_DELL
)
10637 tp
->led_ctrl
= LED_CTRL_MODE_PHY_2
;
10639 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A0
||
10640 tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A1
)
10641 tp
->led_ctrl
= LED_CTRL_MODE_MAC
;
10643 if (nic_cfg
& NIC_SRAM_DATA_CFG_EEPROM_WP
) {
10644 tp
->tg3_flags
|= TG3_FLAG_EEPROM_WRITE_PROT
;
10645 if ((tp
->pdev
->subsystem_vendor
==
10646 PCI_VENDOR_ID_ARIMA
) &&
10647 (tp
->pdev
->subsystem_device
== 0x205a ||
10648 tp
->pdev
->subsystem_device
== 0x2063))
10649 tp
->tg3_flags
&= ~TG3_FLAG_EEPROM_WRITE_PROT
;
10651 tp
->tg3_flags
&= ~TG3_FLAG_EEPROM_WRITE_PROT
;
10652 tp
->tg3_flags2
|= TG3_FLG2_IS_NIC
;
10655 if (nic_cfg
& NIC_SRAM_DATA_CFG_ASF_ENABLE
) {
10656 tp
->tg3_flags
|= TG3_FLAG_ENABLE_ASF
;
10657 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
)
10658 tp
->tg3_flags2
|= TG3_FLG2_ASF_NEW_HANDSHAKE
;
10660 if (nic_cfg
& NIC_SRAM_DATA_CFG_APE_ENABLE
)
10661 tp
->tg3_flags3
|= TG3_FLG3_ENABLE_APE
;
10662 if (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
&&
10663 !(nic_cfg
& NIC_SRAM_DATA_CFG_FIBER_WOL
))
10664 tp
->tg3_flags
&= ~TG3_FLAG_WOL_CAP
;
10666 if (tp
->tg3_flags
& TG3_FLAG_WOL_CAP
&&
10667 nic_cfg
& NIC_SRAM_DATA_CFG_WOL_ENABLE
)
10668 tp
->tg3_flags
|= TG3_FLAG_WOL_ENABLE
;
10670 if (cfg2
& (1 << 17))
10671 tp
->tg3_flags2
|= TG3_FLG2_CAPACITIVE_COUPLING
;
10673 /* serdes signal pre-emphasis in register 0x590 set by */
10674 /* bootcode if bit 18 is set */
10675 if (cfg2
& (1 << 18))
10676 tp
->tg3_flags2
|= TG3_FLG2_SERDES_PREEMPHASIS
;
10678 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
10681 tg3_read_mem(tp
, NIC_SRAM_DATA_CFG_3
, &cfg3
);
10682 if (cfg3
& NIC_SRAM_ASPM_DEBOUNCE
)
10683 tp
->tg3_flags
|= TG3_FLAG_ASPM_WORKAROUND
;
10688 static int __devinit
tg3_phy_probe(struct tg3
*tp
)
10690 u32 hw_phy_id_1
, hw_phy_id_2
;
10691 u32 hw_phy_id
, hw_phy_id_masked
;
10694 /* Reading the PHY ID register can conflict with ASF
10695 * firwmare access to the PHY hardware.
10698 if ((tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) ||
10699 (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
)) {
10700 hw_phy_id
= hw_phy_id_masked
= PHY_ID_INVALID
;
10702 /* Now read the physical PHY_ID from the chip and verify
10703 * that it is sane. If it doesn't look good, we fall back
10704 * to either the hard-coded table based PHY_ID and failing
10705 * that the value found in the eeprom area.
10707 err
|= tg3_readphy(tp
, MII_PHYSID1
, &hw_phy_id_1
);
10708 err
|= tg3_readphy(tp
, MII_PHYSID2
, &hw_phy_id_2
);
10710 hw_phy_id
= (hw_phy_id_1
& 0xffff) << 10;
10711 hw_phy_id
|= (hw_phy_id_2
& 0xfc00) << 16;
10712 hw_phy_id
|= (hw_phy_id_2
& 0x03ff) << 0;
10714 hw_phy_id_masked
= hw_phy_id
& PHY_ID_MASK
;
10717 if (!err
&& KNOWN_PHY_ID(hw_phy_id_masked
)) {
10718 tp
->phy_id
= hw_phy_id
;
10719 if (hw_phy_id_masked
== PHY_ID_BCM8002
)
10720 tp
->tg3_flags2
|= TG3_FLG2_PHY_SERDES
;
10722 tp
->tg3_flags2
&= ~TG3_FLG2_PHY_SERDES
;
10724 if (tp
->phy_id
!= PHY_ID_INVALID
) {
10725 /* Do nothing, phy ID already set up in
10726 * tg3_get_eeprom_hw_cfg().
10729 struct subsys_tbl_ent
*p
;
10731 /* No eeprom signature? Try the hardcoded
10732 * subsys device table.
10734 p
= lookup_by_subsys(tp
);
10738 tp
->phy_id
= p
->phy_id
;
10740 tp
->phy_id
== PHY_ID_BCM8002
)
10741 tp
->tg3_flags2
|= TG3_FLG2_PHY_SERDES
;
10745 if (!(tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
) &&
10746 !(tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
) &&
10747 !(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
)) {
10748 u32 bmsr
, adv_reg
, tg3_ctrl
, mask
;
10750 tg3_readphy(tp
, MII_BMSR
, &bmsr
);
10751 if (!tg3_readphy(tp
, MII_BMSR
, &bmsr
) &&
10752 (bmsr
& BMSR_LSTATUS
))
10753 goto skip_phy_reset
;
10755 err
= tg3_phy_reset(tp
);
10759 adv_reg
= (ADVERTISE_10HALF
| ADVERTISE_10FULL
|
10760 ADVERTISE_100HALF
| ADVERTISE_100FULL
|
10761 ADVERTISE_CSMA
| ADVERTISE_PAUSE_CAP
);
10763 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
)) {
10764 tg3_ctrl
= (MII_TG3_CTRL_ADV_1000_HALF
|
10765 MII_TG3_CTRL_ADV_1000_FULL
);
10766 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
||
10767 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B0
)
10768 tg3_ctrl
|= (MII_TG3_CTRL_AS_MASTER
|
10769 MII_TG3_CTRL_ENABLE_AS_MASTER
);
10772 mask
= (ADVERTISED_10baseT_Half
| ADVERTISED_10baseT_Full
|
10773 ADVERTISED_100baseT_Half
| ADVERTISED_100baseT_Full
|
10774 ADVERTISED_1000baseT_Half
| ADVERTISED_1000baseT_Full
);
10775 if (!tg3_copper_is_advertising_all(tp
, mask
)) {
10776 tg3_writephy(tp
, MII_ADVERTISE
, adv_reg
);
10778 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
))
10779 tg3_writephy(tp
, MII_TG3_CTRL
, tg3_ctrl
);
10781 tg3_writephy(tp
, MII_BMCR
,
10782 BMCR_ANENABLE
| BMCR_ANRESTART
);
10784 tg3_phy_set_wirespeed(tp
);
10786 tg3_writephy(tp
, MII_ADVERTISE
, adv_reg
);
10787 if (!(tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
))
10788 tg3_writephy(tp
, MII_TG3_CTRL
, tg3_ctrl
);
10792 if ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5401
) {
10793 err
= tg3_init_5401phy_dsp(tp
);
10798 if (!err
&& ((tp
->phy_id
& PHY_ID_MASK
) == PHY_ID_BCM5401
)) {
10799 err
= tg3_init_5401phy_dsp(tp
);
10802 if (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
)
10803 tp
->link_config
.advertising
=
10804 (ADVERTISED_1000baseT_Half
|
10805 ADVERTISED_1000baseT_Full
|
10806 ADVERTISED_Autoneg
|
10808 if (tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
)
10809 tp
->link_config
.advertising
&=
10810 ~(ADVERTISED_1000baseT_Half
|
10811 ADVERTISED_1000baseT_Full
);
10816 static void __devinit
tg3_read_partno(struct tg3
*tp
)
10818 unsigned char vpd_data
[256];
10822 if (tg3_nvram_read_swab(tp
, 0x0, &magic
))
10823 goto out_not_found
;
10825 if (magic
== TG3_EEPROM_MAGIC
) {
10826 for (i
= 0; i
< 256; i
+= 4) {
10829 if (tg3_nvram_read(tp
, 0x100 + i
, &tmp
))
10830 goto out_not_found
;
10832 vpd_data
[i
+ 0] = ((tmp
>> 0) & 0xff);
10833 vpd_data
[i
+ 1] = ((tmp
>> 8) & 0xff);
10834 vpd_data
[i
+ 2] = ((tmp
>> 16) & 0xff);
10835 vpd_data
[i
+ 3] = ((tmp
>> 24) & 0xff);
10840 vpd_cap
= pci_find_capability(tp
->pdev
, PCI_CAP_ID_VPD
);
10841 for (i
= 0; i
< 256; i
+= 4) {
10846 pci_write_config_word(tp
->pdev
, vpd_cap
+ PCI_VPD_ADDR
,
10848 while (j
++ < 100) {
10849 pci_read_config_word(tp
->pdev
, vpd_cap
+
10850 PCI_VPD_ADDR
, &tmp16
);
10851 if (tmp16
& 0x8000)
10855 if (!(tmp16
& 0x8000))
10856 goto out_not_found
;
10858 pci_read_config_dword(tp
->pdev
, vpd_cap
+ PCI_VPD_DATA
,
10860 v
= cpu_to_le32(tmp
);
10861 memcpy(&vpd_data
[i
], &v
, 4);
10865 /* Now parse and find the part number. */
10866 for (i
= 0; i
< 254; ) {
10867 unsigned char val
= vpd_data
[i
];
10868 unsigned int block_end
;
10870 if (val
== 0x82 || val
== 0x91) {
10873 (vpd_data
[i
+ 2] << 8)));
10878 goto out_not_found
;
10880 block_end
= (i
+ 3 +
10882 (vpd_data
[i
+ 2] << 8)));
10885 if (block_end
> 256)
10886 goto out_not_found
;
10888 while (i
< (block_end
- 2)) {
10889 if (vpd_data
[i
+ 0] == 'P' &&
10890 vpd_data
[i
+ 1] == 'N') {
10891 int partno_len
= vpd_data
[i
+ 2];
10894 if (partno_len
> 24 || (partno_len
+ i
) > 256)
10895 goto out_not_found
;
10897 memcpy(tp
->board_part_number
,
10898 &vpd_data
[i
], partno_len
);
10903 i
+= 3 + vpd_data
[i
+ 2];
10906 /* Part number not found. */
10907 goto out_not_found
;
10911 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
10912 strcpy(tp
->board_part_number
, "BCM95906");
10914 strcpy(tp
->board_part_number
, "none");
10917 static int __devinit
tg3_fw_img_is_valid(struct tg3
*tp
, u32 offset
)
10921 if (tg3_nvram_read_swab(tp
, offset
, &val
) ||
10922 (val
& 0xfc000000) != 0x0c000000 ||
10923 tg3_nvram_read_swab(tp
, offset
+ 4, &val
) ||
10930 static void __devinit
tg3_read_fw_ver(struct tg3
*tp
)
10932 u32 val
, offset
, start
;
10936 if (tg3_nvram_read_swab(tp
, 0, &val
))
10939 if (val
!= TG3_EEPROM_MAGIC
)
10942 if (tg3_nvram_read_swab(tp
, 0xc, &offset
) ||
10943 tg3_nvram_read_swab(tp
, 0x4, &start
))
10946 offset
= tg3_nvram_logical_addr(tp
, offset
);
10948 if (!tg3_fw_img_is_valid(tp
, offset
) ||
10949 tg3_nvram_read_swab(tp
, offset
+ 8, &ver_offset
))
10952 offset
= offset
+ ver_offset
- start
;
10953 for (i
= 0; i
< 16; i
+= 4) {
10955 if (tg3_nvram_read_le(tp
, offset
+ i
, &v
))
10958 memcpy(tp
->fw_ver
+ i
, &v
, 4);
10961 if (!(tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) ||
10962 (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
))
10965 for (offset
= TG3_NVM_DIR_START
;
10966 offset
< TG3_NVM_DIR_END
;
10967 offset
+= TG3_NVM_DIRENT_SIZE
) {
10968 if (tg3_nvram_read_swab(tp
, offset
, &val
))
10971 if ((val
>> TG3_NVM_DIRTYPE_SHIFT
) == TG3_NVM_DIRTYPE_ASFINI
)
10975 if (offset
== TG3_NVM_DIR_END
)
10978 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
))
10979 start
= 0x08000000;
10980 else if (tg3_nvram_read_swab(tp
, offset
- 4, &start
))
10983 if (tg3_nvram_read_swab(tp
, offset
+ 4, &offset
) ||
10984 !tg3_fw_img_is_valid(tp
, offset
) ||
10985 tg3_nvram_read_swab(tp
, offset
+ 8, &val
))
10988 offset
+= val
- start
;
10990 bcnt
= strlen(tp
->fw_ver
);
10992 tp
->fw_ver
[bcnt
++] = ',';
10993 tp
->fw_ver
[bcnt
++] = ' ';
10995 for (i
= 0; i
< 4; i
++) {
10997 if (tg3_nvram_read_le(tp
, offset
, &v
))
11000 offset
+= sizeof(v
);
11002 if (bcnt
> TG3_VER_SIZE
- sizeof(v
)) {
11003 memcpy(&tp
->fw_ver
[bcnt
], &v
, TG3_VER_SIZE
- bcnt
);
11007 memcpy(&tp
->fw_ver
[bcnt
], &v
, sizeof(v
));
11011 tp
->fw_ver
[TG3_VER_SIZE
- 1] = 0;
11014 static struct pci_dev
* __devinit
tg3_find_peer(struct tg3
*);
11016 static int __devinit
tg3_get_invariants(struct tg3
*tp
)
11018 static struct pci_device_id write_reorder_chipsets
[] = {
11019 { PCI_DEVICE(PCI_VENDOR_ID_AMD
,
11020 PCI_DEVICE_ID_AMD_FE_GATE_700C
) },
11021 { PCI_DEVICE(PCI_VENDOR_ID_AMD
,
11022 PCI_DEVICE_ID_AMD_8131_BRIDGE
) },
11023 { PCI_DEVICE(PCI_VENDOR_ID_VIA
,
11024 PCI_DEVICE_ID_VIA_8385_0
) },
11028 u32 cacheline_sz_reg
;
11029 u32 pci_state_reg
, grc_misc_cfg
;
11034 /* Force memory write invalidate off. If we leave it on,
11035 * then on 5700_BX chips we have to enable a workaround.
11036 * The workaround is to set the TG3PCI_DMA_RW_CTRL boundary
11037 * to match the cacheline size. The Broadcom driver have this
11038 * workaround but turns MWI off all the times so never uses
11039 * it. This seems to suggest that the workaround is insufficient.
11041 pci_read_config_word(tp
->pdev
, PCI_COMMAND
, &pci_cmd
);
11042 pci_cmd
&= ~PCI_COMMAND_INVALIDATE
;
11043 pci_write_config_word(tp
->pdev
, PCI_COMMAND
, pci_cmd
);
11045 /* It is absolutely critical that TG3PCI_MISC_HOST_CTRL
11046 * has the register indirect write enable bit set before
11047 * we try to access any of the MMIO registers. It is also
11048 * critical that the PCI-X hw workaround situation is decided
11049 * before that as well.
11051 pci_read_config_dword(tp
->pdev
, TG3PCI_MISC_HOST_CTRL
,
11054 tp
->pci_chip_rev_id
= (misc_ctrl_reg
>>
11055 MISC_HOST_CTRL_CHIPREV_SHIFT
);
11056 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_USE_PROD_ID_REG
) {
11057 u32 prod_id_asic_rev
;
11059 pci_read_config_dword(tp
->pdev
, TG3PCI_PRODID_ASICREV
,
11060 &prod_id_asic_rev
);
11061 tp
->pci_chip_rev_id
= prod_id_asic_rev
& PROD_ID_ASIC_REV_MASK
;
11064 /* Wrong chip ID in 5752 A0. This code can be removed later
11065 * as A0 is not in production.
11067 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5752_A0_HW
)
11068 tp
->pci_chip_rev_id
= CHIPREV_ID_5752_A0
;
11070 /* If we have 5702/03 A1 or A2 on certain ICH chipsets,
11071 * we need to disable memory and use config. cycles
11072 * only to access all registers. The 5702/03 chips
11073 * can mistakenly decode the special cycles from the
11074 * ICH chipsets as memory write cycles, causing corruption
11075 * of register and memory space. Only certain ICH bridges
11076 * will drive special cycles with non-zero data during the
11077 * address phase which can fall within the 5703's address
11078 * range. This is not an ICH bug as the PCI spec allows
11079 * non-zero address during special cycles. However, only
11080 * these ICH bridges are known to drive non-zero addresses
11081 * during special cycles.
11083 * Since special cycles do not cross PCI bridges, we only
11084 * enable this workaround if the 5703 is on the secondary
11085 * bus of these ICH bridges.
11087 if ((tp
->pci_chip_rev_id
== CHIPREV_ID_5703_A1
) ||
11088 (tp
->pci_chip_rev_id
== CHIPREV_ID_5703_A2
)) {
11089 static struct tg3_dev_id
{
11093 } ich_chipsets
[] = {
11094 { PCI_VENDOR_ID_INTEL
, PCI_DEVICE_ID_INTEL_82801AA_8
,
11096 { PCI_VENDOR_ID_INTEL
, PCI_DEVICE_ID_INTEL_82801AB_8
,
11098 { PCI_VENDOR_ID_INTEL
, PCI_DEVICE_ID_INTEL_82801BA_11
,
11100 { PCI_VENDOR_ID_INTEL
, PCI_DEVICE_ID_INTEL_82801BA_6
,
11104 struct tg3_dev_id
*pci_id
= &ich_chipsets
[0];
11105 struct pci_dev
*bridge
= NULL
;
11107 while (pci_id
->vendor
!= 0) {
11108 bridge
= pci_get_device(pci_id
->vendor
, pci_id
->device
,
11114 if (pci_id
->rev
!= PCI_ANY_ID
) {
11115 if (bridge
->revision
> pci_id
->rev
)
11118 if (bridge
->subordinate
&&
11119 (bridge
->subordinate
->number
==
11120 tp
->pdev
->bus
->number
)) {
11122 tp
->tg3_flags2
|= TG3_FLG2_ICH_WORKAROUND
;
11123 pci_dev_put(bridge
);
11129 /* The EPB bridge inside 5714, 5715, and 5780 cannot support
11130 * DMA addresses > 40-bit. This bridge may have other additional
11131 * 57xx devices behind it in some 4-port NIC designs for example.
11132 * Any tg3 device found behind the bridge will also need the 40-bit
11135 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5780
||
11136 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
) {
11137 tp
->tg3_flags2
|= TG3_FLG2_5780_CLASS
;
11138 tp
->tg3_flags
|= TG3_FLAG_40BIT_DMA_BUG
;
11139 tp
->msi_cap
= pci_find_capability(tp
->pdev
, PCI_CAP_ID_MSI
);
11142 struct pci_dev
*bridge
= NULL
;
11145 bridge
= pci_get_device(PCI_VENDOR_ID_SERVERWORKS
,
11146 PCI_DEVICE_ID_SERVERWORKS_EPB
,
11148 if (bridge
&& bridge
->subordinate
&&
11149 (bridge
->subordinate
->number
<=
11150 tp
->pdev
->bus
->number
) &&
11151 (bridge
->subordinate
->subordinate
>=
11152 tp
->pdev
->bus
->number
)) {
11153 tp
->tg3_flags
|= TG3_FLAG_40BIT_DMA_BUG
;
11154 pci_dev_put(bridge
);
11160 /* Initialize misc host control in PCI block. */
11161 tp
->misc_host_ctrl
|= (misc_ctrl_reg
&
11162 MISC_HOST_CTRL_CHIPREV
);
11163 pci_write_config_dword(tp
->pdev
, TG3PCI_MISC_HOST_CTRL
,
11164 tp
->misc_host_ctrl
);
11166 pci_read_config_dword(tp
->pdev
, TG3PCI_CACHELINESZ
,
11167 &cacheline_sz_reg
);
11169 tp
->pci_cacheline_sz
= (cacheline_sz_reg
>> 0) & 0xff;
11170 tp
->pci_lat_timer
= (cacheline_sz_reg
>> 8) & 0xff;
11171 tp
->pci_hdr_type
= (cacheline_sz_reg
>> 16) & 0xff;
11172 tp
->pci_bist
= (cacheline_sz_reg
>> 24) & 0xff;
11174 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) ||
11175 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
))
11176 tp
->pdev_peer
= tg3_find_peer(tp
);
11178 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
||
11179 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
||
11180 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
11181 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
11182 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
11183 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
||
11184 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
||
11185 (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
))
11186 tp
->tg3_flags2
|= TG3_FLG2_5750_PLUS
;
11188 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) ||
11189 (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
))
11190 tp
->tg3_flags2
|= TG3_FLG2_5705_PLUS
;
11192 if (tp
->tg3_flags2
& TG3_FLG2_5750_PLUS
) {
11193 tp
->tg3_flags
|= TG3_FLAG_SUPPORT_MSI
;
11194 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5750_AX
||
11195 GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5750_BX
||
11196 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
&&
11197 tp
->pci_chip_rev_id
<= CHIPREV_ID_5714_A2
&&
11198 tp
->pdev_peer
== tp
->pdev
))
11199 tp
->tg3_flags
&= ~TG3_FLAG_SUPPORT_MSI
;
11201 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
11202 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
11203 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
11204 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
||
11205 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
11206 tp
->tg3_flags2
|= TG3_FLG2_HW_TSO_2
;
11207 tp
->tg3_flags2
|= TG3_FLG2_1SHOT_MSI
;
11209 tp
->tg3_flags2
|= TG3_FLG2_HW_TSO_1
| TG3_FLG2_TSO_BUG
;
11210 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) ==
11212 tp
->pci_chip_rev_id
>= CHIPREV_ID_5750_C2
)
11213 tp
->tg3_flags2
&= ~TG3_FLG2_TSO_BUG
;
11217 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5705
&&
11218 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5750
&&
11219 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5752
&&
11220 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5755
&&
11221 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5787
&&
11222 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5784
&&
11223 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5761
&&
11224 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
)
11225 tp
->tg3_flags2
|= TG3_FLG2_JUMBO_CAPABLE
;
11227 pcie_cap
= pci_find_capability(tp
->pdev
, PCI_CAP_ID_EXP
);
11228 if (pcie_cap
!= 0) {
11229 tp
->tg3_flags2
|= TG3_FLG2_PCI_EXPRESS
;
11231 pcie_set_readrq(tp
->pdev
, 4096);
11233 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
11236 pci_read_config_word(tp
->pdev
,
11237 pcie_cap
+ PCI_EXP_LNKCTL
,
11239 if (lnkctl
& PCI_EXP_LNKCTL_CLKREQ_EN
)
11240 tp
->tg3_flags2
&= ~TG3_FLG2_HW_TSO_2
;
11244 /* If we have an AMD 762 or VIA K8T800 chipset, write
11245 * reordering to the mailbox registers done by the host
11246 * controller can cause major troubles. We read back from
11247 * every mailbox register write to force the writes to be
11248 * posted to the chip in order.
11250 if (pci_dev_present(write_reorder_chipsets
) &&
11251 !(tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
))
11252 tp
->tg3_flags
|= TG3_FLAG_MBOX_WRITE_REORDER
;
11254 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
&&
11255 tp
->pci_lat_timer
< 64) {
11256 tp
->pci_lat_timer
= 64;
11258 cacheline_sz_reg
= ((tp
->pci_cacheline_sz
& 0xff) << 0);
11259 cacheline_sz_reg
|= ((tp
->pci_lat_timer
& 0xff) << 8);
11260 cacheline_sz_reg
|= ((tp
->pci_hdr_type
& 0xff) << 16);
11261 cacheline_sz_reg
|= ((tp
->pci_bist
& 0xff) << 24);
11263 pci_write_config_dword(tp
->pdev
, TG3PCI_CACHELINESZ
,
11267 if (!(tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) ||
11268 (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)) {
11269 tp
->pcix_cap
= pci_find_capability(tp
->pdev
, PCI_CAP_ID_PCIX
);
11270 if (!tp
->pcix_cap
) {
11271 printk(KERN_ERR PFX
"Cannot find PCI-X "
11272 "capability, aborting.\n");
11277 pci_read_config_dword(tp
->pdev
, TG3PCI_PCISTATE
,
11280 if (tp
->pcix_cap
&& (pci_state_reg
& PCISTATE_CONV_PCI_MODE
) == 0) {
11281 tp
->tg3_flags
|= TG3_FLAG_PCIX_MODE
;
11283 /* If this is a 5700 BX chipset, and we are in PCI-X
11284 * mode, enable register write workaround.
11286 * The workaround is to use indirect register accesses
11287 * for all chip writes not to mailbox registers.
11289 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5700_BX
) {
11292 tp
->tg3_flags
|= TG3_FLAG_PCIX_TARGET_HWBUG
;
11294 /* The chip can have it's power management PCI config
11295 * space registers clobbered due to this bug.
11296 * So explicitly force the chip into D0 here.
11298 pci_read_config_dword(tp
->pdev
,
11299 tp
->pm_cap
+ PCI_PM_CTRL
,
11301 pm_reg
&= ~PCI_PM_CTRL_STATE_MASK
;
11302 pm_reg
|= PCI_PM_CTRL_PME_ENABLE
| 0 /* D0 */;
11303 pci_write_config_dword(tp
->pdev
,
11304 tp
->pm_cap
+ PCI_PM_CTRL
,
11307 /* Also, force SERR#/PERR# in PCI command. */
11308 pci_read_config_word(tp
->pdev
, PCI_COMMAND
, &pci_cmd
);
11309 pci_cmd
|= PCI_COMMAND_PARITY
| PCI_COMMAND_SERR
;
11310 pci_write_config_word(tp
->pdev
, PCI_COMMAND
, pci_cmd
);
11314 /* 5700 BX chips need to have their TX producer index mailboxes
11315 * written twice to workaround a bug.
11317 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5700_BX
)
11318 tp
->tg3_flags
|= TG3_FLAG_TXD_MBOX_HWBUG
;
11320 if ((pci_state_reg
& PCISTATE_BUS_SPEED_HIGH
) != 0)
11321 tp
->tg3_flags
|= TG3_FLAG_PCI_HIGH_SPEED
;
11322 if ((pci_state_reg
& PCISTATE_BUS_32BIT
) != 0)
11323 tp
->tg3_flags
|= TG3_FLAG_PCI_32BIT
;
11325 /* Chip-specific fixup from Broadcom driver */
11326 if ((tp
->pci_chip_rev_id
== CHIPREV_ID_5704_A0
) &&
11327 (!(pci_state_reg
& PCISTATE_RETRY_SAME_DMA
))) {
11328 pci_state_reg
|= PCISTATE_RETRY_SAME_DMA
;
11329 pci_write_config_dword(tp
->pdev
, TG3PCI_PCISTATE
, pci_state_reg
);
11332 /* Default fast path register access methods */
11333 tp
->read32
= tg3_read32
;
11334 tp
->write32
= tg3_write32
;
11335 tp
->read32_mbox
= tg3_read32
;
11336 tp
->write32_mbox
= tg3_write32
;
11337 tp
->write32_tx_mbox
= tg3_write32
;
11338 tp
->write32_rx_mbox
= tg3_write32
;
11340 /* Various workaround register access methods */
11341 if (tp
->tg3_flags
& TG3_FLAG_PCIX_TARGET_HWBUG
)
11342 tp
->write32
= tg3_write_indirect_reg32
;
11343 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
||
11344 ((tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) &&
11345 tp
->pci_chip_rev_id
== CHIPREV_ID_5750_A0
)) {
11347 * Back to back register writes can cause problems on these
11348 * chips, the workaround is to read back all reg writes
11349 * except those to mailbox regs.
11351 * See tg3_write_indirect_reg32().
11353 tp
->write32
= tg3_write_flush_reg32
;
11357 if ((tp
->tg3_flags
& TG3_FLAG_TXD_MBOX_HWBUG
) ||
11358 (tp
->tg3_flags
& TG3_FLAG_MBOX_WRITE_REORDER
)) {
11359 tp
->write32_tx_mbox
= tg3_write32_tx_mbox
;
11360 if (tp
->tg3_flags
& TG3_FLAG_MBOX_WRITE_REORDER
)
11361 tp
->write32_rx_mbox
= tg3_write_flush_reg32
;
11364 if (tp
->tg3_flags2
& TG3_FLG2_ICH_WORKAROUND
) {
11365 tp
->read32
= tg3_read_indirect_reg32
;
11366 tp
->write32
= tg3_write_indirect_reg32
;
11367 tp
->read32_mbox
= tg3_read_indirect_mbox
;
11368 tp
->write32_mbox
= tg3_write_indirect_mbox
;
11369 tp
->write32_tx_mbox
= tg3_write_indirect_mbox
;
11370 tp
->write32_rx_mbox
= tg3_write_indirect_mbox
;
11375 pci_read_config_word(tp
->pdev
, PCI_COMMAND
, &pci_cmd
);
11376 pci_cmd
&= ~PCI_COMMAND_MEMORY
;
11377 pci_write_config_word(tp
->pdev
, PCI_COMMAND
, pci_cmd
);
11379 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
11380 tp
->read32_mbox
= tg3_read32_mbox_5906
;
11381 tp
->write32_mbox
= tg3_write32_mbox_5906
;
11382 tp
->write32_tx_mbox
= tg3_write32_mbox_5906
;
11383 tp
->write32_rx_mbox
= tg3_write32_mbox_5906
;
11386 if (tp
->write32
== tg3_write_indirect_reg32
||
11387 ((tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) &&
11388 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
11389 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
)))
11390 tp
->tg3_flags
|= TG3_FLAG_SRAM_USE_CONFIG
;
11392 /* Get eeprom hw config before calling tg3_set_power_state().
11393 * In particular, the TG3_FLG2_IS_NIC flag must be
11394 * determined before calling tg3_set_power_state() so that
11395 * we know whether or not to switch out of Vaux power.
11396 * When the flag is set, it means that GPIO1 is used for eeprom
11397 * write protect and also implies that it is a LOM where GPIOs
11398 * are not used to switch power.
11400 tg3_get_eeprom_hw_cfg(tp
);
11402 if (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
) {
11403 /* Allow reads and writes to the
11404 * APE register and memory space.
11406 pci_state_reg
|= PCISTATE_ALLOW_APE_CTLSPC_WR
|
11407 PCISTATE_ALLOW_APE_SHMEM_WR
;
11408 pci_write_config_dword(tp
->pdev
, TG3PCI_PCISTATE
,
11412 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
11413 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
) {
11414 tp
->tg3_flags
|= TG3_FLAG_CPMU_PRESENT
;
11416 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A0
||
11417 tp
->pci_chip_rev_id
== CHIPREV_ID_5784_A1
||
11418 tp
->pci_chip_rev_id
== CHIPREV_ID_5761_A0
||
11419 tp
->pci_chip_rev_id
== CHIPREV_ID_5761_A1
)
11420 tp
->tg3_flags3
|= TG3_FLG3_5761_5784_AX_FIXES
;
11423 /* Set up tp->grc_local_ctrl before calling tg3_set_power_state().
11424 * GPIO1 driven high will bring 5700's external PHY out of reset.
11425 * It is also used as eeprom write protect on LOMs.
11427 tp
->grc_local_ctrl
= GRC_LCLCTRL_INT_ON_ATTN
| GRC_LCLCTRL_AUTO_SEEPROM
;
11428 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
) ||
11429 (tp
->tg3_flags
& TG3_FLAG_EEPROM_WRITE_PROT
))
11430 tp
->grc_local_ctrl
|= (GRC_LCLCTRL_GPIO_OE1
|
11431 GRC_LCLCTRL_GPIO_OUTPUT1
);
11432 /* Unused GPIO3 must be driven as output on 5752 because there
11433 * are no pull-up resistors on unused GPIO pins.
11435 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
)
11436 tp
->grc_local_ctrl
|= GRC_LCLCTRL_GPIO_OE3
;
11438 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
)
11439 tp
->grc_local_ctrl
|= GRC_LCLCTRL_GPIO_UART_SEL
;
11441 /* Force the chip into D0. */
11442 err
= tg3_set_power_state(tp
, PCI_D0
);
11444 printk(KERN_ERR PFX
"(%s) transition to D0 failed\n",
11445 pci_name(tp
->pdev
));
11449 /* 5700 B0 chips do not support checksumming correctly due
11450 * to hardware bugs.
11452 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5700_B0
)
11453 tp
->tg3_flags
|= TG3_FLAG_BROKEN_CHECKSUMS
;
11455 /* Derive initial jumbo mode from MTU assigned in
11456 * ether_setup() via the alloc_etherdev() call
11458 if (tp
->dev
->mtu
> ETH_DATA_LEN
&&
11459 !(tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
))
11460 tp
->tg3_flags
|= TG3_FLAG_JUMBO_RING_ENABLE
;
11462 /* Determine WakeOnLan speed to use. */
11463 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
11464 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_A0
||
11465 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B0
||
11466 tp
->pci_chip_rev_id
== CHIPREV_ID_5701_B2
) {
11467 tp
->tg3_flags
&= ~(TG3_FLAG_WOL_SPEED_100MB
);
11469 tp
->tg3_flags
|= TG3_FLAG_WOL_SPEED_100MB
;
11472 /* A few boards don't want Ethernet@WireSpeed phy feature */
11473 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
) ||
11474 ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
) &&
11475 (tp
->pci_chip_rev_id
!= CHIPREV_ID_5705_A0
) &&
11476 (tp
->pci_chip_rev_id
!= CHIPREV_ID_5705_A1
)) ||
11477 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) ||
11478 (tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
))
11479 tp
->tg3_flags2
|= TG3_FLG2_NO_ETH_WIRE_SPEED
;
11481 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5703_AX
||
11482 GET_CHIP_REV(tp
->pci_chip_rev_id
) == CHIPREV_5704_AX
)
11483 tp
->tg3_flags2
|= TG3_FLG2_PHY_ADC_BUG
;
11484 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5704_A0
)
11485 tp
->tg3_flags2
|= TG3_FLG2_PHY_5704_A0_BUG
;
11487 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
11488 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
11489 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
11490 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
11491 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
) {
11492 if (tp
->pdev
->device
!= PCI_DEVICE_ID_TIGON3_5756
&&
11493 tp
->pdev
->device
!= PCI_DEVICE_ID_TIGON3_5722
)
11494 tp
->tg3_flags2
|= TG3_FLG2_PHY_JITTER_BUG
;
11495 if (tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5755M
)
11496 tp
->tg3_flags2
|= TG3_FLG2_PHY_ADJUST_TRIM
;
11497 } else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
)
11498 tp
->tg3_flags2
|= TG3_FLG2_PHY_BER_BUG
;
11501 tp
->coalesce_mode
= 0;
11502 if (GET_CHIP_REV(tp
->pci_chip_rev_id
) != CHIPREV_5700_AX
&&
11503 GET_CHIP_REV(tp
->pci_chip_rev_id
) != CHIPREV_5700_BX
)
11504 tp
->coalesce_mode
|= HOSTCC_MODE_32BYTE
;
11506 /* Initialize MAC MI mode, polling disabled. */
11507 tw32_f(MAC_MI_MODE
, tp
->mi_mode
);
11510 /* Initialize data/descriptor byte/word swapping. */
11511 val
= tr32(GRC_MODE
);
11512 val
&= GRC_MODE_HOST_STACKUP
;
11513 tw32(GRC_MODE
, val
| tp
->grc_mode
);
11515 tg3_switch_clocks(tp
);
11517 /* Clear this out for sanity. */
11518 tw32(TG3PCI_MEM_WIN_BASE_ADDR
, 0);
11520 pci_read_config_dword(tp
->pdev
, TG3PCI_PCISTATE
,
11522 if ((pci_state_reg
& PCISTATE_CONV_PCI_MODE
) == 0 &&
11523 (tp
->tg3_flags
& TG3_FLAG_PCIX_TARGET_HWBUG
) == 0) {
11524 u32 chiprevid
= GET_CHIP_REV_ID(tp
->misc_host_ctrl
);
11526 if (chiprevid
== CHIPREV_ID_5701_A0
||
11527 chiprevid
== CHIPREV_ID_5701_B0
||
11528 chiprevid
== CHIPREV_ID_5701_B2
||
11529 chiprevid
== CHIPREV_ID_5701_B5
) {
11530 void __iomem
*sram_base
;
11532 /* Write some dummy words into the SRAM status block
11533 * area, see if it reads back correctly. If the return
11534 * value is bad, force enable the PCIX workaround.
11536 sram_base
= tp
->regs
+ NIC_SRAM_WIN_BASE
+ NIC_SRAM_STATS_BLK
;
11538 writel(0x00000000, sram_base
);
11539 writel(0x00000000, sram_base
+ 4);
11540 writel(0xffffffff, sram_base
+ 4);
11541 if (readl(sram_base
) != 0x00000000)
11542 tp
->tg3_flags
|= TG3_FLAG_PCIX_TARGET_HWBUG
;
11547 tg3_nvram_init(tp
);
11549 grc_misc_cfg
= tr32(GRC_MISC_CFG
);
11550 grc_misc_cfg
&= GRC_MISC_CFG_BOARD_ID_MASK
;
11552 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
&&
11553 (grc_misc_cfg
== GRC_MISC_CFG_BOARD_ID_5788
||
11554 grc_misc_cfg
== GRC_MISC_CFG_BOARD_ID_5788M
))
11555 tp
->tg3_flags2
|= TG3_FLG2_IS_5788
;
11557 if (!(tp
->tg3_flags2
& TG3_FLG2_IS_5788
) &&
11558 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
))
11559 tp
->tg3_flags
|= TG3_FLAG_TAGGED_STATUS
;
11560 if (tp
->tg3_flags
& TG3_FLAG_TAGGED_STATUS
) {
11561 tp
->coalesce_mode
|= (HOSTCC_MODE_CLRTICK_RXBD
|
11562 HOSTCC_MODE_CLRTICK_TXBD
);
11564 tp
->misc_host_ctrl
|= MISC_HOST_CTRL_TAGGED_STATUS
;
11565 pci_write_config_dword(tp
->pdev
, TG3PCI_MISC_HOST_CTRL
,
11566 tp
->misc_host_ctrl
);
11569 /* these are limited to 10/100 only */
11570 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
&&
11571 (grc_misc_cfg
== 0x8000 || grc_misc_cfg
== 0x4000)) ||
11572 (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
&&
11573 tp
->pdev
->vendor
== PCI_VENDOR_ID_BROADCOM
&&
11574 (tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5901
||
11575 tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5901_2
||
11576 tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5705F
)) ||
11577 (tp
->pdev
->vendor
== PCI_VENDOR_ID_BROADCOM
&&
11578 (tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5751F
||
11579 tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5753F
||
11580 tp
->pdev
->device
== PCI_DEVICE_ID_TIGON3_5787F
)) ||
11581 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
11582 tp
->tg3_flags
|= TG3_FLAG_10_100_ONLY
;
11584 err
= tg3_phy_probe(tp
);
11586 printk(KERN_ERR PFX
"(%s) phy probe failed, err %d\n",
11587 pci_name(tp
->pdev
), err
);
11588 /* ... but do not return immediately ... */
11591 tg3_read_partno(tp
);
11592 tg3_read_fw_ver(tp
);
11594 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
) {
11595 tp
->tg3_flags
&= ~TG3_FLAG_USE_MI_INTERRUPT
;
11597 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
)
11598 tp
->tg3_flags
|= TG3_FLAG_USE_MI_INTERRUPT
;
11600 tp
->tg3_flags
&= ~TG3_FLAG_USE_MI_INTERRUPT
;
11603 /* 5700 {AX,BX} chips have a broken status block link
11604 * change bit implementation, so we must use the
11605 * status register in those cases.
11607 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
)
11608 tp
->tg3_flags
|= TG3_FLAG_USE_LINKCHG_REG
;
11610 tp
->tg3_flags
&= ~TG3_FLAG_USE_LINKCHG_REG
;
11612 /* The led_ctrl is set during tg3_phy_probe, here we might
11613 * have to force the link status polling mechanism based
11614 * upon subsystem IDs.
11616 if (tp
->pdev
->subsystem_vendor
== PCI_VENDOR_ID_DELL
&&
11617 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
&&
11618 !(tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)) {
11619 tp
->tg3_flags
|= (TG3_FLAG_USE_MI_INTERRUPT
|
11620 TG3_FLAG_USE_LINKCHG_REG
);
11623 /* For all SERDES we poll the MAC status register. */
11624 if (tp
->tg3_flags2
& TG3_FLG2_PHY_SERDES
)
11625 tp
->tg3_flags
|= TG3_FLAG_POLL_SERDES
;
11627 tp
->tg3_flags
&= ~TG3_FLAG_POLL_SERDES
;
11629 /* All chips before 5787 can get confused if TX buffers
11630 * straddle the 4GB address boundary in some cases.
11632 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
11633 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
11634 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
11635 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
||
11636 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
11637 tp
->dev
->hard_start_xmit
= tg3_start_xmit
;
11639 tp
->dev
->hard_start_xmit
= tg3_start_xmit_dma_bug
;
11642 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
&&
11643 (tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) != 0)
11646 tp
->rx_std_max_post
= TG3_RX_RING_SIZE
;
11648 /* Increment the rx prod index on the rx std ring by at most
11649 * 8 for these chips to workaround hw errata.
11651 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
||
11652 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5752
||
11653 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
)
11654 tp
->rx_std_max_post
= 8;
11656 if (tp
->tg3_flags
& TG3_FLAG_ASPM_WORKAROUND
)
11657 tp
->pwrmgmt_thresh
= tr32(PCIE_PWR_MGMT_THRESH
) &
11658 PCIE_PWR_MGMT_L1_THRESH_MSK
;
11663 #ifdef CONFIG_SPARC
11664 static int __devinit
tg3_get_macaddr_sparc(struct tg3
*tp
)
11666 struct net_device
*dev
= tp
->dev
;
11667 struct pci_dev
*pdev
= tp
->pdev
;
11668 struct device_node
*dp
= pci_device_to_OF_node(pdev
);
11669 const unsigned char *addr
;
11672 addr
= of_get_property(dp
, "local-mac-address", &len
);
11673 if (addr
&& len
== 6) {
11674 memcpy(dev
->dev_addr
, addr
, 6);
11675 memcpy(dev
->perm_addr
, dev
->dev_addr
, 6);
11681 static int __devinit
tg3_get_default_macaddr_sparc(struct tg3
*tp
)
11683 struct net_device
*dev
= tp
->dev
;
11685 memcpy(dev
->dev_addr
, idprom
->id_ethaddr
, 6);
11686 memcpy(dev
->perm_addr
, idprom
->id_ethaddr
, 6);
11691 static int __devinit
tg3_get_device_address(struct tg3
*tp
)
11693 struct net_device
*dev
= tp
->dev
;
11694 u32 hi
, lo
, mac_offset
;
11697 #ifdef CONFIG_SPARC
11698 if (!tg3_get_macaddr_sparc(tp
))
11703 if ((GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) ||
11704 (tp
->tg3_flags2
& TG3_FLG2_5780_CLASS
)) {
11705 if (tr32(TG3PCI_DUAL_MAC_CTRL
) & DUAL_MAC_CTRL_ID
)
11707 if (tg3_nvram_lock(tp
))
11708 tw32_f(NVRAM_CMD
, NVRAM_CMD_RESET
);
11710 tg3_nvram_unlock(tp
);
11712 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
)
11715 /* First try to get it from MAC address mailbox. */
11716 tg3_read_mem(tp
, NIC_SRAM_MAC_ADDR_HIGH_MBOX
, &hi
);
11717 if ((hi
>> 16) == 0x484b) {
11718 dev
->dev_addr
[0] = (hi
>> 8) & 0xff;
11719 dev
->dev_addr
[1] = (hi
>> 0) & 0xff;
11721 tg3_read_mem(tp
, NIC_SRAM_MAC_ADDR_LOW_MBOX
, &lo
);
11722 dev
->dev_addr
[2] = (lo
>> 24) & 0xff;
11723 dev
->dev_addr
[3] = (lo
>> 16) & 0xff;
11724 dev
->dev_addr
[4] = (lo
>> 8) & 0xff;
11725 dev
->dev_addr
[5] = (lo
>> 0) & 0xff;
11727 /* Some old bootcode may report a 0 MAC address in SRAM */
11728 addr_ok
= is_valid_ether_addr(&dev
->dev_addr
[0]);
11731 /* Next, try NVRAM. */
11732 if (!tg3_nvram_read(tp
, mac_offset
+ 0, &hi
) &&
11733 !tg3_nvram_read(tp
, mac_offset
+ 4, &lo
)) {
11734 dev
->dev_addr
[0] = ((hi
>> 16) & 0xff);
11735 dev
->dev_addr
[1] = ((hi
>> 24) & 0xff);
11736 dev
->dev_addr
[2] = ((lo
>> 0) & 0xff);
11737 dev
->dev_addr
[3] = ((lo
>> 8) & 0xff);
11738 dev
->dev_addr
[4] = ((lo
>> 16) & 0xff);
11739 dev
->dev_addr
[5] = ((lo
>> 24) & 0xff);
11741 /* Finally just fetch it out of the MAC control regs. */
11743 hi
= tr32(MAC_ADDR_0_HIGH
);
11744 lo
= tr32(MAC_ADDR_0_LOW
);
11746 dev
->dev_addr
[5] = lo
& 0xff;
11747 dev
->dev_addr
[4] = (lo
>> 8) & 0xff;
11748 dev
->dev_addr
[3] = (lo
>> 16) & 0xff;
11749 dev
->dev_addr
[2] = (lo
>> 24) & 0xff;
11750 dev
->dev_addr
[1] = hi
& 0xff;
11751 dev
->dev_addr
[0] = (hi
>> 8) & 0xff;
11755 if (!is_valid_ether_addr(&dev
->dev_addr
[0])) {
11756 #ifdef CONFIG_SPARC64
11757 if (!tg3_get_default_macaddr_sparc(tp
))
11762 memcpy(dev
->perm_addr
, dev
->dev_addr
, dev
->addr_len
);
11766 #define BOUNDARY_SINGLE_CACHELINE 1
11767 #define BOUNDARY_MULTI_CACHELINE 2
11769 static u32 __devinit
tg3_calc_dma_bndry(struct tg3
*tp
, u32 val
)
11771 int cacheline_size
;
11775 pci_read_config_byte(tp
->pdev
, PCI_CACHE_LINE_SIZE
, &byte
);
11777 cacheline_size
= 1024;
11779 cacheline_size
= (int) byte
* 4;
11781 /* On 5703 and later chips, the boundary bits have no
11784 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
&&
11785 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5701
&&
11786 !(tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
))
11789 #if defined(CONFIG_PPC64) || defined(CONFIG_IA64) || defined(CONFIG_PARISC)
11790 goal
= BOUNDARY_MULTI_CACHELINE
;
11792 #if defined(CONFIG_SPARC64) || defined(CONFIG_ALPHA)
11793 goal
= BOUNDARY_SINGLE_CACHELINE
;
11802 /* PCI controllers on most RISC systems tend to disconnect
11803 * when a device tries to burst across a cache-line boundary.
11804 * Therefore, letting tg3 do so just wastes PCI bandwidth.
11806 * Unfortunately, for PCI-E there are only limited
11807 * write-side controls for this, and thus for reads
11808 * we will still get the disconnects. We'll also waste
11809 * these PCI cycles for both read and write for chips
11810 * other than 5700 and 5701 which do not implement the
11813 if ((tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) &&
11814 !(tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
)) {
11815 switch (cacheline_size
) {
11820 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11821 val
|= (DMA_RWCTRL_READ_BNDRY_128_PCIX
|
11822 DMA_RWCTRL_WRITE_BNDRY_128_PCIX
);
11824 val
|= (DMA_RWCTRL_READ_BNDRY_384_PCIX
|
11825 DMA_RWCTRL_WRITE_BNDRY_384_PCIX
);
11830 val
|= (DMA_RWCTRL_READ_BNDRY_256_PCIX
|
11831 DMA_RWCTRL_WRITE_BNDRY_256_PCIX
);
11835 val
|= (DMA_RWCTRL_READ_BNDRY_384_PCIX
|
11836 DMA_RWCTRL_WRITE_BNDRY_384_PCIX
);
11839 } else if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
11840 switch (cacheline_size
) {
11844 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11845 val
&= ~DMA_RWCTRL_WRITE_BNDRY_DISAB_PCIE
;
11846 val
|= DMA_RWCTRL_WRITE_BNDRY_64_PCIE
;
11852 val
&= ~DMA_RWCTRL_WRITE_BNDRY_DISAB_PCIE
;
11853 val
|= DMA_RWCTRL_WRITE_BNDRY_128_PCIE
;
11857 switch (cacheline_size
) {
11859 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11860 val
|= (DMA_RWCTRL_READ_BNDRY_16
|
11861 DMA_RWCTRL_WRITE_BNDRY_16
);
11866 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11867 val
|= (DMA_RWCTRL_READ_BNDRY_32
|
11868 DMA_RWCTRL_WRITE_BNDRY_32
);
11873 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11874 val
|= (DMA_RWCTRL_READ_BNDRY_64
|
11875 DMA_RWCTRL_WRITE_BNDRY_64
);
11880 if (goal
== BOUNDARY_SINGLE_CACHELINE
) {
11881 val
|= (DMA_RWCTRL_READ_BNDRY_128
|
11882 DMA_RWCTRL_WRITE_BNDRY_128
);
11887 val
|= (DMA_RWCTRL_READ_BNDRY_256
|
11888 DMA_RWCTRL_WRITE_BNDRY_256
);
11891 val
|= (DMA_RWCTRL_READ_BNDRY_512
|
11892 DMA_RWCTRL_WRITE_BNDRY_512
);
11896 val
|= (DMA_RWCTRL_READ_BNDRY_1024
|
11897 DMA_RWCTRL_WRITE_BNDRY_1024
);
11906 static int __devinit
tg3_do_test_dma(struct tg3
*tp
, u32
*buf
, dma_addr_t buf_dma
, int size
, int to_device
)
11908 struct tg3_internal_buffer_desc test_desc
;
11909 u32 sram_dma_descs
;
11912 sram_dma_descs
= NIC_SRAM_DMA_DESC_POOL_BASE
;
11914 tw32(FTQ_RCVBD_COMP_FIFO_ENQDEQ
, 0);
11915 tw32(FTQ_RCVDATA_COMP_FIFO_ENQDEQ
, 0);
11916 tw32(RDMAC_STATUS
, 0);
11917 tw32(WDMAC_STATUS
, 0);
11919 tw32(BUFMGR_MODE
, 0);
11920 tw32(FTQ_RESET
, 0);
11922 test_desc
.addr_hi
= ((u64
) buf_dma
) >> 32;
11923 test_desc
.addr_lo
= buf_dma
& 0xffffffff;
11924 test_desc
.nic_mbuf
= 0x00002100;
11925 test_desc
.len
= size
;
11928 * HP ZX1 was seeing test failures for 5701 cards running at 33Mhz
11929 * the *second* time the tg3 driver was getting loaded after an
11932 * Broadcom tells me:
11933 * ...the DMA engine is connected to the GRC block and a DMA
11934 * reset may affect the GRC block in some unpredictable way...
11935 * The behavior of resets to individual blocks has not been tested.
11937 * Broadcom noted the GRC reset will also reset all sub-components.
11940 test_desc
.cqid_sqid
= (13 << 8) | 2;
11942 tw32_f(RDMAC_MODE
, RDMAC_MODE_ENABLE
);
11945 test_desc
.cqid_sqid
= (16 << 8) | 7;
11947 tw32_f(WDMAC_MODE
, WDMAC_MODE_ENABLE
);
11950 test_desc
.flags
= 0x00000005;
11952 for (i
= 0; i
< (sizeof(test_desc
) / sizeof(u32
)); i
++) {
11955 val
= *(((u32
*)&test_desc
) + i
);
11956 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
,
11957 sram_dma_descs
+ (i
* sizeof(u32
)));
11958 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_DATA
, val
);
11960 pci_write_config_dword(tp
->pdev
, TG3PCI_MEM_WIN_BASE_ADDR
, 0);
11963 tw32(FTQ_DMA_HIGH_READ_FIFO_ENQDEQ
, sram_dma_descs
);
11965 tw32(FTQ_DMA_HIGH_WRITE_FIFO_ENQDEQ
, sram_dma_descs
);
11969 for (i
= 0; i
< 40; i
++) {
11973 val
= tr32(FTQ_RCVBD_COMP_FIFO_ENQDEQ
);
11975 val
= tr32(FTQ_RCVDATA_COMP_FIFO_ENQDEQ
);
11976 if ((val
& 0xffff) == sram_dma_descs
) {
11987 #define TEST_BUFFER_SIZE 0x2000
11989 static int __devinit
tg3_test_dma(struct tg3
*tp
)
11991 dma_addr_t buf_dma
;
11992 u32
*buf
, saved_dma_rwctrl
;
11995 buf
= pci_alloc_consistent(tp
->pdev
, TEST_BUFFER_SIZE
, &buf_dma
);
12001 tp
->dma_rwctrl
= ((0x7 << DMA_RWCTRL_PCI_WRITE_CMD_SHIFT
) |
12002 (0x6 << DMA_RWCTRL_PCI_READ_CMD_SHIFT
));
12004 tp
->dma_rwctrl
= tg3_calc_dma_bndry(tp
, tp
->dma_rwctrl
);
12006 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
12007 /* DMA read watermark not used on PCIE */
12008 tp
->dma_rwctrl
|= 0x00180000;
12009 } else if (!(tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
)) {
12010 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5705
||
12011 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5750
)
12012 tp
->dma_rwctrl
|= 0x003f0000;
12014 tp
->dma_rwctrl
|= 0x003f000f;
12016 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
12017 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
) {
12018 u32 ccval
= (tr32(TG3PCI_CLOCK_CTRL
) & 0x1f);
12019 u32 read_water
= 0x7;
12021 /* If the 5704 is behind the EPB bridge, we can
12022 * do the less restrictive ONE_DMA workaround for
12023 * better performance.
12025 if ((tp
->tg3_flags
& TG3_FLAG_40BIT_DMA_BUG
) &&
12026 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
)
12027 tp
->dma_rwctrl
|= 0x8000;
12028 else if (ccval
== 0x6 || ccval
== 0x7)
12029 tp
->dma_rwctrl
|= DMA_RWCTRL_ONE_DMA
;
12031 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
)
12033 /* Set bit 23 to enable PCIX hw bug fix */
12035 (read_water
<< DMA_RWCTRL_READ_WATER_SHIFT
) |
12036 (0x3 << DMA_RWCTRL_WRITE_WATER_SHIFT
) |
12038 } else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5780
) {
12039 /* 5780 always in PCIX mode */
12040 tp
->dma_rwctrl
|= 0x00144000;
12041 } else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5714
) {
12042 /* 5714 always in PCIX mode */
12043 tp
->dma_rwctrl
|= 0x00148000;
12045 tp
->dma_rwctrl
|= 0x001b000f;
12049 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5703
||
12050 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5704
)
12051 tp
->dma_rwctrl
&= 0xfffffff0;
12053 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
12054 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
) {
12055 /* Remove this if it causes problems for some boards. */
12056 tp
->dma_rwctrl
|= DMA_RWCTRL_USE_MEM_READ_MULT
;
12058 /* On 5700/5701 chips, we need to set this bit.
12059 * Otherwise the chip will issue cacheline transactions
12060 * to streamable DMA memory with not all the byte
12061 * enables turned on. This is an error on several
12062 * RISC PCI controllers, in particular sparc64.
12064 * On 5703/5704 chips, this bit has been reassigned
12065 * a different meaning. In particular, it is used
12066 * on those chips to enable a PCI-X workaround.
12068 tp
->dma_rwctrl
|= DMA_RWCTRL_ASSERT_ALL_BE
;
12071 tw32(TG3PCI_DMA_RW_CTRL
, tp
->dma_rwctrl
);
12074 /* Unneeded, already done by tg3_get_invariants. */
12075 tg3_switch_clocks(tp
);
12079 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5700
&&
12080 GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5701
)
12083 /* It is best to perform DMA test with maximum write burst size
12084 * to expose the 5700/5701 write DMA bug.
12086 saved_dma_rwctrl
= tp
->dma_rwctrl
;
12087 tp
->dma_rwctrl
&= ~DMA_RWCTRL_WRITE_BNDRY_MASK
;
12088 tw32(TG3PCI_DMA_RW_CTRL
, tp
->dma_rwctrl
);
12093 for (i
= 0; i
< TEST_BUFFER_SIZE
/ sizeof(u32
); i
++)
12096 /* Send the buffer to the chip. */
12097 ret
= tg3_do_test_dma(tp
, buf
, buf_dma
, TEST_BUFFER_SIZE
, 1);
12099 printk(KERN_ERR
"tg3_test_dma() Write the buffer failed %d\n", ret
);
12104 /* validate data reached card RAM correctly. */
12105 for (i
= 0; i
< TEST_BUFFER_SIZE
/ sizeof(u32
); i
++) {
12107 tg3_read_mem(tp
, 0x2100 + (i
*4), &val
);
12108 if (le32_to_cpu(val
) != p
[i
]) {
12109 printk(KERN_ERR
" tg3_test_dma() Card buffer corrupted on write! (%d != %d)\n", val
, i
);
12110 /* ret = -ENODEV here? */
12115 /* Now read it back. */
12116 ret
= tg3_do_test_dma(tp
, buf
, buf_dma
, TEST_BUFFER_SIZE
, 0);
12118 printk(KERN_ERR
"tg3_test_dma() Read the buffer failed %d\n", ret
);
12124 for (i
= 0; i
< TEST_BUFFER_SIZE
/ sizeof(u32
); i
++) {
12128 if ((tp
->dma_rwctrl
& DMA_RWCTRL_WRITE_BNDRY_MASK
) !=
12129 DMA_RWCTRL_WRITE_BNDRY_16
) {
12130 tp
->dma_rwctrl
&= ~DMA_RWCTRL_WRITE_BNDRY_MASK
;
12131 tp
->dma_rwctrl
|= DMA_RWCTRL_WRITE_BNDRY_16
;
12132 tw32(TG3PCI_DMA_RW_CTRL
, tp
->dma_rwctrl
);
12135 printk(KERN_ERR
"tg3_test_dma() buffer corrupted on read back! (%d != %d)\n", p
[i
], i
);
12141 if (i
== (TEST_BUFFER_SIZE
/ sizeof(u32
))) {
12147 if ((tp
->dma_rwctrl
& DMA_RWCTRL_WRITE_BNDRY_MASK
) !=
12148 DMA_RWCTRL_WRITE_BNDRY_16
) {
12149 static struct pci_device_id dma_wait_state_chipsets
[] = {
12150 { PCI_DEVICE(PCI_VENDOR_ID_APPLE
,
12151 PCI_DEVICE_ID_APPLE_UNI_N_PCI15
) },
12155 /* DMA test passed without adjusting DMA boundary,
12156 * now look for chipsets that are known to expose the
12157 * DMA bug without failing the test.
12159 if (pci_dev_present(dma_wait_state_chipsets
)) {
12160 tp
->dma_rwctrl
&= ~DMA_RWCTRL_WRITE_BNDRY_MASK
;
12161 tp
->dma_rwctrl
|= DMA_RWCTRL_WRITE_BNDRY_16
;
12164 /* Safe to use the calculated DMA boundary. */
12165 tp
->dma_rwctrl
= saved_dma_rwctrl
;
12167 tw32(TG3PCI_DMA_RW_CTRL
, tp
->dma_rwctrl
);
12171 pci_free_consistent(tp
->pdev
, TEST_BUFFER_SIZE
, buf
, buf_dma
);
12176 static void __devinit
tg3_init_link_config(struct tg3
*tp
)
12178 tp
->link_config
.advertising
=
12179 (ADVERTISED_10baseT_Half
| ADVERTISED_10baseT_Full
|
12180 ADVERTISED_100baseT_Half
| ADVERTISED_100baseT_Full
|
12181 ADVERTISED_1000baseT_Half
| ADVERTISED_1000baseT_Full
|
12182 ADVERTISED_Autoneg
| ADVERTISED_MII
);
12183 tp
->link_config
.speed
= SPEED_INVALID
;
12184 tp
->link_config
.duplex
= DUPLEX_INVALID
;
12185 tp
->link_config
.autoneg
= AUTONEG_ENABLE
;
12186 tp
->link_config
.active_speed
= SPEED_INVALID
;
12187 tp
->link_config
.active_duplex
= DUPLEX_INVALID
;
12188 tp
->link_config
.phy_is_low_power
= 0;
12189 tp
->link_config
.orig_speed
= SPEED_INVALID
;
12190 tp
->link_config
.orig_duplex
= DUPLEX_INVALID
;
12191 tp
->link_config
.orig_autoneg
= AUTONEG_INVALID
;
12194 static void __devinit
tg3_init_bufmgr_config(struct tg3
*tp
)
12196 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
12197 tp
->bufmgr_config
.mbuf_read_dma_low_water
=
12198 DEFAULT_MB_RDMA_LOW_WATER_5705
;
12199 tp
->bufmgr_config
.mbuf_mac_rx_low_water
=
12200 DEFAULT_MB_MACRX_LOW_WATER_5705
;
12201 tp
->bufmgr_config
.mbuf_high_water
=
12202 DEFAULT_MB_HIGH_WATER_5705
;
12203 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
) {
12204 tp
->bufmgr_config
.mbuf_mac_rx_low_water
=
12205 DEFAULT_MB_MACRX_LOW_WATER_5906
;
12206 tp
->bufmgr_config
.mbuf_high_water
=
12207 DEFAULT_MB_HIGH_WATER_5906
;
12210 tp
->bufmgr_config
.mbuf_read_dma_low_water_jumbo
=
12211 DEFAULT_MB_RDMA_LOW_WATER_JUMBO_5780
;
12212 tp
->bufmgr_config
.mbuf_mac_rx_low_water_jumbo
=
12213 DEFAULT_MB_MACRX_LOW_WATER_JUMBO_5780
;
12214 tp
->bufmgr_config
.mbuf_high_water_jumbo
=
12215 DEFAULT_MB_HIGH_WATER_JUMBO_5780
;
12217 tp
->bufmgr_config
.mbuf_read_dma_low_water
=
12218 DEFAULT_MB_RDMA_LOW_WATER
;
12219 tp
->bufmgr_config
.mbuf_mac_rx_low_water
=
12220 DEFAULT_MB_MACRX_LOW_WATER
;
12221 tp
->bufmgr_config
.mbuf_high_water
=
12222 DEFAULT_MB_HIGH_WATER
;
12224 tp
->bufmgr_config
.mbuf_read_dma_low_water_jumbo
=
12225 DEFAULT_MB_RDMA_LOW_WATER_JUMBO
;
12226 tp
->bufmgr_config
.mbuf_mac_rx_low_water_jumbo
=
12227 DEFAULT_MB_MACRX_LOW_WATER_JUMBO
;
12228 tp
->bufmgr_config
.mbuf_high_water_jumbo
=
12229 DEFAULT_MB_HIGH_WATER_JUMBO
;
12232 tp
->bufmgr_config
.dma_low_water
= DEFAULT_DMA_LOW_WATER
;
12233 tp
->bufmgr_config
.dma_high_water
= DEFAULT_DMA_HIGH_WATER
;
12236 static char * __devinit
tg3_phy_string(struct tg3
*tp
)
12238 switch (tp
->phy_id
& PHY_ID_MASK
) {
12239 case PHY_ID_BCM5400
: return "5400";
12240 case PHY_ID_BCM5401
: return "5401";
12241 case PHY_ID_BCM5411
: return "5411";
12242 case PHY_ID_BCM5701
: return "5701";
12243 case PHY_ID_BCM5703
: return "5703";
12244 case PHY_ID_BCM5704
: return "5704";
12245 case PHY_ID_BCM5705
: return "5705";
12246 case PHY_ID_BCM5750
: return "5750";
12247 case PHY_ID_BCM5752
: return "5752";
12248 case PHY_ID_BCM5714
: return "5714";
12249 case PHY_ID_BCM5780
: return "5780";
12250 case PHY_ID_BCM5755
: return "5755";
12251 case PHY_ID_BCM5787
: return "5787";
12252 case PHY_ID_BCM5784
: return "5784";
12253 case PHY_ID_BCM5756
: return "5722/5756";
12254 case PHY_ID_BCM5906
: return "5906";
12255 case PHY_ID_BCM5761
: return "5761";
12256 case PHY_ID_BCM8002
: return "8002/serdes";
12257 case 0: return "serdes";
12258 default: return "unknown";
12262 static char * __devinit
tg3_bus_string(struct tg3
*tp
, char *str
)
12264 if (tp
->tg3_flags2
& TG3_FLG2_PCI_EXPRESS
) {
12265 strcpy(str
, "PCI Express");
12267 } else if (tp
->tg3_flags
& TG3_FLAG_PCIX_MODE
) {
12268 u32 clock_ctrl
= tr32(TG3PCI_CLOCK_CTRL
) & 0x1f;
12270 strcpy(str
, "PCIX:");
12272 if ((clock_ctrl
== 7) ||
12273 ((tr32(GRC_MISC_CFG
) & GRC_MISC_CFG_BOARD_ID_MASK
) ==
12274 GRC_MISC_CFG_BOARD_ID_5704CIOBE
))
12275 strcat(str
, "133MHz");
12276 else if (clock_ctrl
== 0)
12277 strcat(str
, "33MHz");
12278 else if (clock_ctrl
== 2)
12279 strcat(str
, "50MHz");
12280 else if (clock_ctrl
== 4)
12281 strcat(str
, "66MHz");
12282 else if (clock_ctrl
== 6)
12283 strcat(str
, "100MHz");
12285 strcpy(str
, "PCI:");
12286 if (tp
->tg3_flags
& TG3_FLAG_PCI_HIGH_SPEED
)
12287 strcat(str
, "66MHz");
12289 strcat(str
, "33MHz");
12291 if (tp
->tg3_flags
& TG3_FLAG_PCI_32BIT
)
12292 strcat(str
, ":32-bit");
12294 strcat(str
, ":64-bit");
12298 static struct pci_dev
* __devinit
tg3_find_peer(struct tg3
*tp
)
12300 struct pci_dev
*peer
;
12301 unsigned int func
, devnr
= tp
->pdev
->devfn
& ~7;
12303 for (func
= 0; func
< 8; func
++) {
12304 peer
= pci_get_slot(tp
->pdev
->bus
, devnr
| func
);
12305 if (peer
&& peer
!= tp
->pdev
)
12309 /* 5704 can be configured in single-port mode, set peer to
12310 * tp->pdev in that case.
12318 * We don't need to keep the refcount elevated; there's no way
12319 * to remove one half of this device without removing the other
12326 static void __devinit
tg3_init_coal(struct tg3
*tp
)
12328 struct ethtool_coalesce
*ec
= &tp
->coal
;
12330 memset(ec
, 0, sizeof(*ec
));
12331 ec
->cmd
= ETHTOOL_GCOALESCE
;
12332 ec
->rx_coalesce_usecs
= LOW_RXCOL_TICKS
;
12333 ec
->tx_coalesce_usecs
= LOW_TXCOL_TICKS
;
12334 ec
->rx_max_coalesced_frames
= LOW_RXMAX_FRAMES
;
12335 ec
->tx_max_coalesced_frames
= LOW_TXMAX_FRAMES
;
12336 ec
->rx_coalesce_usecs_irq
= DEFAULT_RXCOAL_TICK_INT
;
12337 ec
->tx_coalesce_usecs_irq
= DEFAULT_TXCOAL_TICK_INT
;
12338 ec
->rx_max_coalesced_frames_irq
= DEFAULT_RXCOAL_MAXF_INT
;
12339 ec
->tx_max_coalesced_frames_irq
= DEFAULT_TXCOAL_MAXF_INT
;
12340 ec
->stats_block_coalesce_usecs
= DEFAULT_STAT_COAL_TICKS
;
12342 if (tp
->coalesce_mode
& (HOSTCC_MODE_CLRTICK_RXBD
|
12343 HOSTCC_MODE_CLRTICK_TXBD
)) {
12344 ec
->rx_coalesce_usecs
= LOW_RXCOL_TICKS_CLRTCKS
;
12345 ec
->rx_coalesce_usecs_irq
= DEFAULT_RXCOAL_TICK_INT_CLRTCKS
;
12346 ec
->tx_coalesce_usecs
= LOW_TXCOL_TICKS_CLRTCKS
;
12347 ec
->tx_coalesce_usecs_irq
= DEFAULT_TXCOAL_TICK_INT_CLRTCKS
;
12350 if (tp
->tg3_flags2
& TG3_FLG2_5705_PLUS
) {
12351 ec
->rx_coalesce_usecs_irq
= 0;
12352 ec
->tx_coalesce_usecs_irq
= 0;
12353 ec
->stats_block_coalesce_usecs
= 0;
12357 static int __devinit
tg3_init_one(struct pci_dev
*pdev
,
12358 const struct pci_device_id
*ent
)
12360 static int tg3_version_printed
= 0;
12361 unsigned long tg3reg_base
, tg3reg_len
;
12362 struct net_device
*dev
;
12364 int i
, err
, pm_cap
;
12366 u64 dma_mask
, persist_dma_mask
;
12368 if (tg3_version_printed
++ == 0)
12369 printk(KERN_INFO
"%s", version
);
12371 err
= pci_enable_device(pdev
);
12373 printk(KERN_ERR PFX
"Cannot enable PCI device, "
12378 if (!(pci_resource_flags(pdev
, 0) & IORESOURCE_MEM
)) {
12379 printk(KERN_ERR PFX
"Cannot find proper PCI device "
12380 "base address, aborting.\n");
12382 goto err_out_disable_pdev
;
12385 err
= pci_request_regions(pdev
, DRV_MODULE_NAME
);
12387 printk(KERN_ERR PFX
"Cannot obtain PCI resources, "
12389 goto err_out_disable_pdev
;
12392 pci_set_master(pdev
);
12394 /* Find power-management capability. */
12395 pm_cap
= pci_find_capability(pdev
, PCI_CAP_ID_PM
);
12397 printk(KERN_ERR PFX
"Cannot find PowerManagement capability, "
12400 goto err_out_free_res
;
12403 tg3reg_base
= pci_resource_start(pdev
, 0);
12404 tg3reg_len
= pci_resource_len(pdev
, 0);
12406 dev
= alloc_etherdev(sizeof(*tp
));
12408 printk(KERN_ERR PFX
"Etherdev alloc failed, aborting.\n");
12410 goto err_out_free_res
;
12413 SET_NETDEV_DEV(dev
, &pdev
->dev
);
12415 #if TG3_VLAN_TAG_USED
12416 dev
->features
|= NETIF_F_HW_VLAN_TX
| NETIF_F_HW_VLAN_RX
;
12417 dev
->vlan_rx_register
= tg3_vlan_rx_register
;
12420 tp
= netdev_priv(dev
);
12423 tp
->pm_cap
= pm_cap
;
12424 tp
->mac_mode
= TG3_DEF_MAC_MODE
;
12425 tp
->rx_mode
= TG3_DEF_RX_MODE
;
12426 tp
->tx_mode
= TG3_DEF_TX_MODE
;
12427 tp
->mi_mode
= MAC_MI_MODE_BASE
;
12429 tp
->msg_enable
= tg3_debug
;
12431 tp
->msg_enable
= TG3_DEF_MSG_ENABLE
;
12433 /* The word/byte swap controls here control register access byte
12434 * swapping. DMA data byte swapping is controlled in the GRC_MODE
12437 tp
->misc_host_ctrl
=
12438 MISC_HOST_CTRL_MASK_PCI_INT
|
12439 MISC_HOST_CTRL_WORD_SWAP
|
12440 MISC_HOST_CTRL_INDIR_ACCESS
|
12441 MISC_HOST_CTRL_PCISTATE_RW
;
12443 /* The NONFRM (non-frame) byte/word swap controls take effect
12444 * on descriptor entries, anything which isn't packet data.
12446 * The StrongARM chips on the board (one for tx, one for rx)
12447 * are running in big-endian mode.
12449 tp
->grc_mode
= (GRC_MODE_WSWAP_DATA
| GRC_MODE_BSWAP_DATA
|
12450 GRC_MODE_WSWAP_NONFRM_DATA
);
12451 #ifdef __BIG_ENDIAN
12452 tp
->grc_mode
|= GRC_MODE_BSWAP_NONFRM_DATA
;
12454 spin_lock_init(&tp
->lock
);
12455 spin_lock_init(&tp
->indirect_lock
);
12456 INIT_WORK(&tp
->reset_task
, tg3_reset_task
);
12458 tp
->regs
= ioremap_nocache(tg3reg_base
, tg3reg_len
);
12460 printk(KERN_ERR PFX
"Cannot map device registers, "
12463 goto err_out_free_dev
;
12466 tg3_init_link_config(tp
);
12468 tp
->rx_pending
= TG3_DEF_RX_RING_PENDING
;
12469 tp
->rx_jumbo_pending
= TG3_DEF_RX_JUMBO_RING_PENDING
;
12470 tp
->tx_pending
= TG3_DEF_TX_RING_PENDING
;
12472 dev
->open
= tg3_open
;
12473 dev
->stop
= tg3_close
;
12474 dev
->get_stats
= tg3_get_stats
;
12475 dev
->set_multicast_list
= tg3_set_rx_mode
;
12476 dev
->set_mac_address
= tg3_set_mac_addr
;
12477 dev
->do_ioctl
= tg3_ioctl
;
12478 dev
->tx_timeout
= tg3_tx_timeout
;
12479 netif_napi_add(dev
, &tp
->napi
, tg3_poll
, 64);
12480 dev
->ethtool_ops
= &tg3_ethtool_ops
;
12481 dev
->watchdog_timeo
= TG3_TX_TIMEOUT
;
12482 dev
->change_mtu
= tg3_change_mtu
;
12483 dev
->irq
= pdev
->irq
;
12484 #ifdef CONFIG_NET_POLL_CONTROLLER
12485 dev
->poll_controller
= tg3_poll_controller
;
12488 err
= tg3_get_invariants(tp
);
12490 printk(KERN_ERR PFX
"Problem fetching invariants of chip, "
12492 goto err_out_iounmap
;
12495 /* The EPB bridge inside 5714, 5715, and 5780 and any
12496 * device behind the EPB cannot support DMA addresses > 40-bit.
12497 * On 64-bit systems with IOMMU, use 40-bit dma_mask.
12498 * On 64-bit systems without IOMMU, use 64-bit dma_mask and
12499 * do DMA address check in tg3_start_xmit().
12501 if (tp
->tg3_flags2
& TG3_FLG2_IS_5788
)
12502 persist_dma_mask
= dma_mask
= DMA_32BIT_MASK
;
12503 else if (tp
->tg3_flags
& TG3_FLAG_40BIT_DMA_BUG
) {
12504 persist_dma_mask
= dma_mask
= DMA_40BIT_MASK
;
12505 #ifdef CONFIG_HIGHMEM
12506 dma_mask
= DMA_64BIT_MASK
;
12509 persist_dma_mask
= dma_mask
= DMA_64BIT_MASK
;
12511 /* Configure DMA attributes. */
12512 if (dma_mask
> DMA_32BIT_MASK
) {
12513 err
= pci_set_dma_mask(pdev
, dma_mask
);
12515 dev
->features
|= NETIF_F_HIGHDMA
;
12516 err
= pci_set_consistent_dma_mask(pdev
,
12519 printk(KERN_ERR PFX
"Unable to obtain 64 bit "
12520 "DMA for consistent allocations\n");
12521 goto err_out_iounmap
;
12525 if (err
|| dma_mask
== DMA_32BIT_MASK
) {
12526 err
= pci_set_dma_mask(pdev
, DMA_32BIT_MASK
);
12528 printk(KERN_ERR PFX
"No usable DMA configuration, "
12530 goto err_out_iounmap
;
12534 tg3_init_bufmgr_config(tp
);
12536 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
) {
12537 tp
->tg3_flags2
|= TG3_FLG2_TSO_CAPABLE
;
12539 else if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5700
||
12540 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5701
||
12541 tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A0
||
12542 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5906
||
12543 (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) != 0) {
12544 tp
->tg3_flags2
&= ~TG3_FLG2_TSO_CAPABLE
;
12546 tp
->tg3_flags2
|= TG3_FLG2_TSO_CAPABLE
| TG3_FLG2_TSO_BUG
;
12549 /* TSO is on by default on chips that support hardware TSO.
12550 * Firmware TSO on older chips gives lower performance, so it
12551 * is off by default, but can be enabled using ethtool.
12553 if (tp
->tg3_flags2
& TG3_FLG2_HW_TSO
) {
12554 dev
->features
|= NETIF_F_TSO
;
12555 if ((tp
->tg3_flags2
& TG3_FLG2_HW_TSO_2
) &&
12556 (GET_ASIC_REV(tp
->pci_chip_rev_id
) != ASIC_REV_5906
))
12557 dev
->features
|= NETIF_F_TSO6
;
12558 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
12559 dev
->features
|= NETIF_F_TSO_ECN
;
12563 if (tp
->pci_chip_rev_id
== CHIPREV_ID_5705_A1
&&
12564 !(tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
) &&
12565 !(tr32(TG3PCI_PCISTATE
) & PCISTATE_BUS_SPEED_HIGH
)) {
12566 tp
->tg3_flags2
|= TG3_FLG2_MAX_RXPEND_64
;
12567 tp
->rx_pending
= 63;
12570 err
= tg3_get_device_address(tp
);
12572 printk(KERN_ERR PFX
"Could not obtain valid ethernet address, "
12574 goto err_out_iounmap
;
12577 if (tp
->tg3_flags3
& TG3_FLG3_ENABLE_APE
) {
12578 if (!(pci_resource_flags(pdev
, 2) & IORESOURCE_MEM
)) {
12579 printk(KERN_ERR PFX
"Cannot find proper PCI device "
12580 "base address for APE, aborting.\n");
12582 goto err_out_iounmap
;
12585 tg3reg_base
= pci_resource_start(pdev
, 2);
12586 tg3reg_len
= pci_resource_len(pdev
, 2);
12588 tp
->aperegs
= ioremap_nocache(tg3reg_base
, tg3reg_len
);
12589 if (tp
->aperegs
== 0UL) {
12590 printk(KERN_ERR PFX
"Cannot map APE registers, "
12593 goto err_out_iounmap
;
12596 tg3_ape_lock_init(tp
);
12600 * Reset chip in case UNDI or EFI driver did not shutdown
12601 * DMA self test will enable WDMAC and we'll see (spurious)
12602 * pending DMA on the PCI bus at that point.
12604 if ((tr32(HOSTCC_MODE
) & HOSTCC_MODE_ENABLE
) ||
12605 (tr32(WDMAC_MODE
) & WDMAC_MODE_ENABLE
)) {
12606 tw32(MEMARB_MODE
, MEMARB_MODE_ENABLE
);
12607 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
12610 err
= tg3_test_dma(tp
);
12612 printk(KERN_ERR PFX
"DMA engine test failed, aborting.\n");
12613 goto err_out_apeunmap
;
12616 /* Tigon3 can do ipv4 only... and some chips have buggy
12619 if ((tp
->tg3_flags
& TG3_FLAG_BROKEN_CHECKSUMS
) == 0) {
12620 dev
->features
|= NETIF_F_IP_CSUM
| NETIF_F_SG
;
12621 if (GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5755
||
12622 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5787
||
12623 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5784
||
12624 GET_ASIC_REV(tp
->pci_chip_rev_id
) == ASIC_REV_5761
)
12625 dev
->features
|= NETIF_F_IPV6_CSUM
;
12627 tp
->tg3_flags
|= TG3_FLAG_RX_CHECKSUMS
;
12629 tp
->tg3_flags
&= ~TG3_FLAG_RX_CHECKSUMS
;
12631 /* flow control autonegotiation is default behavior */
12632 tp
->tg3_flags
|= TG3_FLAG_PAUSE_AUTONEG
;
12636 pci_set_drvdata(pdev
, dev
);
12638 err
= register_netdev(dev
);
12640 printk(KERN_ERR PFX
"Cannot register net device, "
12642 goto err_out_apeunmap
;
12645 printk(KERN_INFO
"%s: Tigon3 [partno(%s) rev %04x PHY(%s)] (%s) %s Ethernet ",
12647 tp
->board_part_number
,
12648 tp
->pci_chip_rev_id
,
12649 tg3_phy_string(tp
),
12650 tg3_bus_string(tp
, str
),
12651 ((tp
->tg3_flags
& TG3_FLAG_10_100_ONLY
) ? "10/100Base-TX" :
12652 ((tp
->tg3_flags2
& TG3_FLG2_ANY_SERDES
) ? "1000Base-SX" :
12653 "10/100/1000Base-T")));
12655 for (i
= 0; i
< 6; i
++)
12656 printk("%2.2x%c", dev
->dev_addr
[i
],
12657 i
== 5 ? '\n' : ':');
12659 printk(KERN_INFO
"%s: RXcsums[%d] LinkChgREG[%d] "
12660 "MIirq[%d] ASF[%d] WireSpeed[%d] TSOcap[%d]\n",
12662 (tp
->tg3_flags
& TG3_FLAG_RX_CHECKSUMS
) != 0,
12663 (tp
->tg3_flags
& TG3_FLAG_USE_LINKCHG_REG
) != 0,
12664 (tp
->tg3_flags
& TG3_FLAG_USE_MI_INTERRUPT
) != 0,
12665 (tp
->tg3_flags
& TG3_FLAG_ENABLE_ASF
) != 0,
12666 (tp
->tg3_flags2
& TG3_FLG2_NO_ETH_WIRE_SPEED
) == 0,
12667 (tp
->tg3_flags2
& TG3_FLG2_TSO_CAPABLE
) != 0);
12668 printk(KERN_INFO
"%s: dma_rwctrl[%08x] dma_mask[%d-bit]\n",
12669 dev
->name
, tp
->dma_rwctrl
,
12670 (pdev
->dma_mask
== DMA_32BIT_MASK
) ? 32 :
12671 (((u64
) pdev
->dma_mask
== DMA_40BIT_MASK
) ? 40 : 64));
12677 iounmap(tp
->aperegs
);
12678 tp
->aperegs
= NULL
;
12691 pci_release_regions(pdev
);
12693 err_out_disable_pdev
:
12694 pci_disable_device(pdev
);
12695 pci_set_drvdata(pdev
, NULL
);
12699 static void __devexit
tg3_remove_one(struct pci_dev
*pdev
)
12701 struct net_device
*dev
= pci_get_drvdata(pdev
);
12704 struct tg3
*tp
= netdev_priv(dev
);
12706 flush_scheduled_work();
12707 unregister_netdev(dev
);
12709 iounmap(tp
->aperegs
);
12710 tp
->aperegs
= NULL
;
12717 pci_release_regions(pdev
);
12718 pci_disable_device(pdev
);
12719 pci_set_drvdata(pdev
, NULL
);
12723 static int tg3_suspend(struct pci_dev
*pdev
, pm_message_t state
)
12725 struct net_device
*dev
= pci_get_drvdata(pdev
);
12726 struct tg3
*tp
= netdev_priv(dev
);
12729 /* PCI register 4 needs to be saved whether netif_running() or not.
12730 * MSI address and data need to be saved if using MSI and
12733 pci_save_state(pdev
);
12735 if (!netif_running(dev
))
12738 flush_scheduled_work();
12739 tg3_netif_stop(tp
);
12741 del_timer_sync(&tp
->timer
);
12743 tg3_full_lock(tp
, 1);
12744 tg3_disable_ints(tp
);
12745 tg3_full_unlock(tp
);
12747 netif_device_detach(dev
);
12749 tg3_full_lock(tp
, 0);
12750 tg3_halt(tp
, RESET_KIND_SHUTDOWN
, 1);
12751 tp
->tg3_flags
&= ~TG3_FLAG_INIT_COMPLETE
;
12752 tg3_full_unlock(tp
);
12754 err
= tg3_set_power_state(tp
, pci_choose_state(pdev
, state
));
12756 tg3_full_lock(tp
, 0);
12758 tp
->tg3_flags
|= TG3_FLAG_INIT_COMPLETE
;
12759 if (tg3_restart_hw(tp
, 1))
12762 tp
->timer
.expires
= jiffies
+ tp
->timer_offset
;
12763 add_timer(&tp
->timer
);
12765 netif_device_attach(dev
);
12766 tg3_netif_start(tp
);
12769 tg3_full_unlock(tp
);
12775 static int tg3_resume(struct pci_dev
*pdev
)
12777 struct net_device
*dev
= pci_get_drvdata(pdev
);
12778 struct tg3
*tp
= netdev_priv(dev
);
12781 pci_restore_state(tp
->pdev
);
12783 if (!netif_running(dev
))
12786 err
= tg3_set_power_state(tp
, PCI_D0
);
12790 netif_device_attach(dev
);
12792 tg3_full_lock(tp
, 0);
12794 tp
->tg3_flags
|= TG3_FLAG_INIT_COMPLETE
;
12795 err
= tg3_restart_hw(tp
, 1);
12799 tp
->timer
.expires
= jiffies
+ tp
->timer_offset
;
12800 add_timer(&tp
->timer
);
12802 tg3_netif_start(tp
);
12805 tg3_full_unlock(tp
);
12810 static struct pci_driver tg3_driver
= {
12811 .name
= DRV_MODULE_NAME
,
12812 .id_table
= tg3_pci_tbl
,
12813 .probe
= tg3_init_one
,
12814 .remove
= __devexit_p(tg3_remove_one
),
12815 .suspend
= tg3_suspend
,
12816 .resume
= tg3_resume
12819 static int __init
tg3_init(void)
12821 return pci_register_driver(&tg3_driver
);
12824 static void __exit
tg3_cleanup(void)
12826 pci_unregister_driver(&tg3_driver
);
12829 module_init(tg3_init
);
12830 module_exit(tg3_cleanup
);