MOXA linux-2.6.x / linux-2.6.9-uc0 from sdlinux-moxaart.tgz
[linux-2.6.9-moxart.git] / drivers / net / acenic.h
blob6eb134e7b57430e16de55f1aa276599673a81ea5
1 #ifndef _ACENIC_H_
2 #define _ACENIC_H_
4 #include <linux/config.h>
6 /*
7 * Generate TX index update each time, when TX ring is closed.
8 * Normally, this is not useful, because results in more dma (and irqs
9 * without TX_COAL_INTS_ONLY).
11 #define USE_TX_COAL_NOW 0
13 #ifndef MAX_SKB_FRAGS
14 #define MAX_SKB_FRAGS 0
15 #endif
19 * Addressing:
21 * The Tigon uses 64-bit host addresses, regardless of their actual
22 * length, and it expects a big-endian format. For 32 bit systems the
23 * upper 32 bits of the address are simply ignored (zero), however for
24 * little endian 64 bit systems (Alpha) this looks strange with the
25 * two parts of the address word being swapped.
27 * The addresses are split in two 32 bit words for all architectures
28 * as some of them are in PCI shared memory and it is necessary to use
29 * readl/writel to access them.
31 * The addressing code is derived from Pete Wyckoff's work, but
32 * modified to deal properly with readl/writel usage.
35 struct ace_regs {
36 u32 pad0[16]; /* PCI control registers */
38 u32 HostCtrl; /* 0x40 */
39 u32 LocalCtrl;
41 u32 pad1[2];
43 u32 MiscCfg; /* 0x50 */
45 u32 pad2[2];
47 u32 PciState;
49 u32 pad3[2]; /* 0x60 */
51 u32 WinBase;
52 u32 WinData;
54 u32 pad4[12]; /* 0x70 */
56 u32 DmaWriteState; /* 0xa0 */
57 u32 pad5[3];
58 u32 DmaReadState; /* 0xb0 */
60 u32 pad6[26];
62 u32 AssistState;
64 u32 pad7[8]; /* 0x120 */
66 u32 CpuCtrl; /* 0x140 */
67 u32 Pc;
69 u32 pad8[3];
71 u32 SramAddr; /* 0x154 */
72 u32 SramData;
74 u32 pad9[49];
76 u32 MacRxState; /* 0x220 */
78 u32 pad10[7];
80 u32 CpuBCtrl; /* 0x240 */
81 u32 PcB;
83 u32 pad11[3];
85 u32 SramBAddr; /* 0x254 */
86 u32 SramBData;
88 u32 pad12[105];
90 u32 pad13[32]; /* 0x400 */
91 u32 Stats[32];
93 u32 Mb0Hi; /* 0x500 */
94 u32 Mb0Lo;
95 u32 Mb1Hi;
96 u32 CmdPrd;
97 u32 Mb2Hi;
98 u32 TxPrd;
99 u32 Mb3Hi;
100 u32 RxStdPrd;
101 u32 Mb4Hi;
102 u32 RxJumboPrd;
103 u32 Mb5Hi;
104 u32 RxMiniPrd;
105 u32 Mb6Hi;
106 u32 Mb6Lo;
107 u32 Mb7Hi;
108 u32 Mb7Lo;
109 u32 Mb8Hi;
110 u32 Mb8Lo;
111 u32 Mb9Hi;
112 u32 Mb9Lo;
113 u32 MbAHi;
114 u32 MbALo;
115 u32 MbBHi;
116 u32 MbBLo;
117 u32 MbCHi;
118 u32 MbCLo;
119 u32 MbDHi;
120 u32 MbDLo;
121 u32 MbEHi;
122 u32 MbELo;
123 u32 MbFHi;
124 u32 MbFLo;
126 u32 pad14[32];
128 u32 MacAddrHi; /* 0x600 */
129 u32 MacAddrLo;
130 u32 InfoPtrHi;
131 u32 InfoPtrLo;
132 u32 MultiCastHi; /* 0x610 */
133 u32 MultiCastLo;
134 u32 ModeStat;
135 u32 DmaReadCfg;
136 u32 DmaWriteCfg; /* 0x620 */
137 u32 TxBufRat;
138 u32 EvtCsm;
139 u32 CmdCsm;
140 u32 TuneRxCoalTicks;/* 0x630 */
141 u32 TuneTxCoalTicks;
142 u32 TuneStatTicks;
143 u32 TuneMaxTxDesc;
144 u32 TuneMaxRxDesc; /* 0x640 */
145 u32 TuneTrace;
146 u32 TuneLink;
147 u32 TuneFastLink;
148 u32 TracePtr; /* 0x650 */
149 u32 TraceStrt;
150 u32 TraceLen;
151 u32 IfIdx;
152 u32 IfMtu; /* 0x660 */
153 u32 MaskInt;
154 u32 GigLnkState;
155 u32 FastLnkState;
156 u32 pad16[4]; /* 0x670 */
157 u32 RxRetCsm; /* 0x680 */
159 u32 pad17[31];
161 u32 CmdRng[64]; /* 0x700 */
162 u32 Window[0x200];
166 typedef struct {
167 u32 addrhi;
168 u32 addrlo;
169 } aceaddr;
172 #define ACE_WINDOW_SIZE 0x800
174 #define ACE_JUMBO_MTU 9000
175 #define ACE_STD_MTU 1500
177 #define ACE_TRACE_SIZE 0x8000
180 * Host control register bits.
183 #define IN_INT 0x01
184 #define CLR_INT 0x02
185 #define HW_RESET 0x08
186 #define BYTE_SWAP 0x10
187 #define WORD_SWAP 0x20
188 #define MASK_INTS 0x40
191 * Local control register bits.
194 #define EEPROM_DATA_IN 0x800000
195 #define EEPROM_DATA_OUT 0x400000
196 #define EEPROM_WRITE_ENABLE 0x200000
197 #define EEPROM_CLK_OUT 0x100000
199 #define EEPROM_BASE 0xa0000000
201 #define EEPROM_WRITE_SELECT 0xa0
202 #define EEPROM_READ_SELECT 0xa1
204 #define SRAM_BANK_512K 0x200
208 * udelay() values for when clocking the eeprom
210 #define ACE_SHORT_DELAY 2
211 #define ACE_LONG_DELAY 4
215 * Misc Config bits
218 #define SYNC_SRAM_TIMING 0x100000
222 * CPU state bits.
225 #define CPU_RESET 0x01
226 #define CPU_TRACE 0x02
227 #define CPU_PROM_FAILED 0x10
228 #define CPU_HALT 0x00010000
229 #define CPU_HALTED 0xffff0000
233 * PCI State bits.
236 #define DMA_READ_MAX_4 0x04
237 #define DMA_READ_MAX_16 0x08
238 #define DMA_READ_MAX_32 0x0c
239 #define DMA_READ_MAX_64 0x10
240 #define DMA_READ_MAX_128 0x14
241 #define DMA_READ_MAX_256 0x18
242 #define DMA_READ_MAX_1K 0x1c
243 #define DMA_WRITE_MAX_4 0x20
244 #define DMA_WRITE_MAX_16 0x40
245 #define DMA_WRITE_MAX_32 0x60
246 #define DMA_WRITE_MAX_64 0x80
247 #define DMA_WRITE_MAX_128 0xa0
248 #define DMA_WRITE_MAX_256 0xc0
249 #define DMA_WRITE_MAX_1K 0xe0
250 #define DMA_READ_WRITE_MASK 0xfc
251 #define MEM_READ_MULTIPLE 0x00020000
252 #define PCI_66MHZ 0x00080000
253 #define PCI_32BIT 0x00100000
254 #define DMA_WRITE_ALL_ALIGN 0x00800000
255 #define READ_CMD_MEM 0x06000000
256 #define WRITE_CMD_MEM 0x70000000
260 * Mode status
263 #define ACE_BYTE_SWAP_BD 0x02
264 #define ACE_WORD_SWAP_BD 0x04 /* not actually used */
265 #define ACE_WARN 0x08
266 #define ACE_BYTE_SWAP_DMA 0x10
267 #define ACE_NO_JUMBO_FRAG 0x200
268 #define ACE_FATAL 0x40000000
272 * DMA config
275 #define DMA_THRESH_1W 0x10
276 #define DMA_THRESH_2W 0x20
277 #define DMA_THRESH_4W 0x40
278 #define DMA_THRESH_8W 0x80
279 #define DMA_THRESH_16W 0x100
280 #define DMA_THRESH_32W 0x0 /* not described in doc, but exists. */
284 * Tuning parameters
287 #define TICKS_PER_SEC 1000000
291 * Link bits
294 #define LNK_PREF 0x00008000
295 #define LNK_10MB 0x00010000
296 #define LNK_100MB 0x00020000
297 #define LNK_1000MB 0x00040000
298 #define LNK_FULL_DUPLEX 0x00080000
299 #define LNK_HALF_DUPLEX 0x00100000
300 #define LNK_TX_FLOW_CTL_Y 0x00200000
301 #define LNK_NEG_ADVANCED 0x00400000
302 #define LNK_RX_FLOW_CTL_Y 0x00800000
303 #define LNK_NIC 0x01000000
304 #define LNK_JAM 0x02000000
305 #define LNK_JUMBO 0x04000000
306 #define LNK_ALTEON 0x08000000
307 #define LNK_NEG_FCTL 0x10000000
308 #define LNK_NEGOTIATE 0x20000000
309 #define LNK_ENABLE 0x40000000
310 #define LNK_UP 0x80000000
314 * Event definitions
317 #define EVT_RING_ENTRIES 256
318 #define EVT_RING_SIZE (EVT_RING_ENTRIES * sizeof(struct event))
320 struct event {
321 #ifdef __LITTLE_ENDIAN_BITFIELD
322 u32 idx:12;
323 u32 code:12;
324 u32 evt:8;
325 #else
326 u32 evt:8;
327 u32 code:12;
328 u32 idx:12;
329 #endif
330 u32 pad;
335 * Events
338 #define E_FW_RUNNING 0x01
339 #define E_STATS_UPDATED 0x04
341 #define E_STATS_UPDATE 0x04
343 #define E_LNK_STATE 0x06
344 #define E_C_LINK_UP 0x01
345 #define E_C_LINK_DOWN 0x02
346 #define E_C_LINK_10_100 0x03
348 #define E_ERROR 0x07
349 #define E_C_ERR_INVAL_CMD 0x01
350 #define E_C_ERR_UNIMP_CMD 0x02
351 #define E_C_ERR_BAD_CFG 0x03
353 #define E_MCAST_LIST 0x08
354 #define E_C_MCAST_ADDR_ADD 0x01
355 #define E_C_MCAST_ADDR_DEL 0x02
357 #define E_RESET_JUMBO_RNG 0x09
361 * Commands
364 #define CMD_RING_ENTRIES 64
366 struct cmd {
367 #ifdef __LITTLE_ENDIAN_BITFIELD
368 u32 idx:12;
369 u32 code:12;
370 u32 evt:8;
371 #else
372 u32 evt:8;
373 u32 code:12;
374 u32 idx:12;
375 #endif
379 #define C_HOST_STATE 0x01
380 #define C_C_STACK_UP 0x01
381 #define C_C_STACK_DOWN 0x02
383 #define C_FDR_FILTERING 0x02
384 #define C_C_FDR_FILT_ENABLE 0x01
385 #define C_C_FDR_FILT_DISABLE 0x02
387 #define C_SET_RX_PRD_IDX 0x03
388 #define C_UPDATE_STATS 0x04
389 #define C_RESET_JUMBO_RNG 0x05
390 #define C_ADD_MULTICAST_ADDR 0x08
391 #define C_DEL_MULTICAST_ADDR 0x09
393 #define C_SET_PROMISC_MODE 0x0a
394 #define C_C_PROMISC_ENABLE 0x01
395 #define C_C_PROMISC_DISABLE 0x02
397 #define C_LNK_NEGOTIATION 0x0b
398 #define C_C_NEGOTIATE_BOTH 0x00
399 #define C_C_NEGOTIATE_GIG 0x01
400 #define C_C_NEGOTIATE_10_100 0x02
402 #define C_SET_MAC_ADDR 0x0c
403 #define C_CLEAR_PROFILE 0x0d
405 #define C_SET_MULTICAST_MODE 0x0e
406 #define C_C_MCAST_ENABLE 0x01
407 #define C_C_MCAST_DISABLE 0x02
409 #define C_CLEAR_STATS 0x0f
410 #define C_SET_RX_JUMBO_PRD_IDX 0x10
411 #define C_REFRESH_STATS 0x11
415 * Descriptor flags
417 #define BD_FLG_TCP_UDP_SUM 0x01
418 #define BD_FLG_IP_SUM 0x02
419 #define BD_FLG_END 0x04
420 #define BD_FLG_MORE 0x08
421 #define BD_FLG_JUMBO 0x10
422 #define BD_FLG_UCAST 0x20
423 #define BD_FLG_MCAST 0x40
424 #define BD_FLG_BCAST 0x60
425 #define BD_FLG_TYP_MASK 0x60
426 #define BD_FLG_IP_FRAG 0x80
427 #define BD_FLG_IP_FRAG_END 0x100
428 #define BD_FLG_VLAN_TAG 0x200
429 #define BD_FLG_FRAME_ERROR 0x400
430 #define BD_FLG_COAL_NOW 0x800
431 #define BD_FLG_MINI 0x1000
435 * Ring Control block flags
437 #define RCB_FLG_TCP_UDP_SUM 0x01
438 #define RCB_FLG_IP_SUM 0x02
439 #define RCB_FLG_NO_PSEUDO_HDR 0x08
440 #define RCB_FLG_VLAN_ASSIST 0x10
441 #define RCB_FLG_COAL_INT_ONLY 0x20
442 #define RCB_FLG_TX_HOST_RING 0x40
443 #define RCB_FLG_IEEE_SNAP_SUM 0x80
444 #define RCB_FLG_EXT_RX_BD 0x100
445 #define RCB_FLG_RNG_DISABLE 0x200
449 * TX ring - maximum TX ring entries for Tigon I's is 128
451 #define MAX_TX_RING_ENTRIES 256
452 #define TIGON_I_TX_RING_ENTRIES 128
453 #define TX_RING_SIZE (MAX_TX_RING_ENTRIES * sizeof(struct tx_desc))
454 #define TX_RING_BASE 0x3800
456 struct tx_desc{
457 aceaddr addr;
458 u32 flagsize;
459 #if 0
461 * This is in PCI shared mem and must be accessed with readl/writel
462 * real layout is:
464 #if __LITTLE_ENDIAN
465 u16 flags;
466 u16 size;
467 u16 vlan;
468 u16 reserved;
469 #else
470 u16 size;
471 u16 flags;
472 u16 reserved;
473 u16 vlan;
474 #endif
475 #endif
476 u32 vlanres;
480 #define RX_STD_RING_ENTRIES 512
481 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc))
483 #define RX_JUMBO_RING_ENTRIES 256
484 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc))
486 #define RX_MINI_RING_ENTRIES 1024
487 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc))
489 #define RX_RETURN_RING_ENTRIES 2048
490 #define RX_RETURN_RING_SIZE (RX_MAX_RETURN_RING_ENTRIES * \
491 sizeof(struct rx_desc))
493 struct rx_desc{
494 aceaddr addr;
495 #ifdef __LITTLE_ENDIAN
496 u16 size;
497 u16 idx;
498 #else
499 u16 idx;
500 u16 size;
501 #endif
502 #ifdef __LITTLE_ENDIAN
503 u16 flags;
504 u16 type;
505 #else
506 u16 type;
507 u16 flags;
508 #endif
509 #ifdef __LITTLE_ENDIAN
510 u16 tcp_udp_csum;
511 u16 ip_csum;
512 #else
513 u16 ip_csum;
514 u16 tcp_udp_csum;
515 #endif
516 #ifdef __LITTLE_ENDIAN
517 u16 vlan;
518 u16 err_flags;
519 #else
520 u16 err_flags;
521 u16 vlan;
522 #endif
523 u32 reserved;
524 u32 opague;
529 * This struct is shared with the NIC firmware.
531 struct ring_ctrl {
532 aceaddr rngptr;
533 #ifdef __LITTLE_ENDIAN
534 u16 flags;
535 u16 max_len;
536 #else
537 u16 max_len;
538 u16 flags;
539 #endif
540 u32 pad;
544 struct ace_mac_stats {
545 u32 excess_colls;
546 u32 coll_1;
547 u32 coll_2;
548 u32 coll_3;
549 u32 coll_4;
550 u32 coll_5;
551 u32 coll_6;
552 u32 coll_7;
553 u32 coll_8;
554 u32 coll_9;
555 u32 coll_10;
556 u32 coll_11;
557 u32 coll_12;
558 u32 coll_13;
559 u32 coll_14;
560 u32 coll_15;
561 u32 late_coll;
562 u32 defers;
563 u32 crc_err;
564 u32 underrun;
565 u32 crs_err;
566 u32 pad[3];
567 u32 drop_ula;
568 u32 drop_mc;
569 u32 drop_fc;
570 u32 drop_space;
571 u32 coll;
572 u32 kept_bc;
573 u32 kept_mc;
574 u32 kept_uc;
578 struct ace_info {
579 union {
580 u32 stats[256];
581 } s;
582 struct ring_ctrl evt_ctrl;
583 struct ring_ctrl cmd_ctrl;
584 struct ring_ctrl tx_ctrl;
585 struct ring_ctrl rx_std_ctrl;
586 struct ring_ctrl rx_jumbo_ctrl;
587 struct ring_ctrl rx_mini_ctrl;
588 struct ring_ctrl rx_return_ctrl;
589 aceaddr evt_prd_ptr;
590 aceaddr rx_ret_prd_ptr;
591 aceaddr tx_csm_ptr;
592 aceaddr stats2_ptr;
596 struct ring_info {
597 struct sk_buff *skb;
598 DECLARE_PCI_UNMAP_ADDR(mapping)
603 * Funny... As soon as we add maplen on alpha, it starts to work
604 * much slower. Hmm... is it because struct does not fit to one cacheline?
605 * So, split tx_ring_info.
607 struct tx_ring_info {
608 struct sk_buff *skb;
609 DECLARE_PCI_UNMAP_ADDR(mapping)
610 DECLARE_PCI_UNMAP_LEN(maplen)
615 * struct ace_skb holding the rings of skb's. This is an awful lot of
616 * pointers, but I don't see any other smart mode to do this in an
617 * efficient manner ;-(
619 struct ace_skb
621 struct tx_ring_info tx_skbuff[MAX_TX_RING_ENTRIES];
622 struct ring_info rx_std_skbuff[RX_STD_RING_ENTRIES];
623 struct ring_info rx_mini_skbuff[RX_MINI_RING_ENTRIES];
624 struct ring_info rx_jumbo_skbuff[RX_JUMBO_RING_ENTRIES];
629 * Struct private for the AceNIC.
631 * Elements are grouped so variables used by the tx handling goes
632 * together, and will go into the same cache lines etc. in order to
633 * avoid cache line contention between the rx and tx handling on SMP.
635 * Frequently accessed variables are put at the beginning of the
636 * struct to help the compiler generate better/shorter code.
638 struct ace_private
640 struct ace_info *info;
641 struct ace_regs *regs; /* register base */
642 struct ace_skb *skb;
643 dma_addr_t info_dma; /* 32/64 bit */
645 int version, link;
646 int promisc, mcast_all;
649 * TX elements
651 struct tx_desc *tx_ring;
652 u32 tx_prd;
653 volatile u32 tx_ret_csm;
654 int tx_ring_entries;
657 * RX elements
659 unsigned long std_refill_busy
660 __attribute__ ((aligned (SMP_CACHE_BYTES)));
661 unsigned long mini_refill_busy, jumbo_refill_busy;
662 atomic_t cur_rx_bufs;
663 atomic_t cur_mini_bufs;
664 atomic_t cur_jumbo_bufs;
665 u32 rx_std_skbprd, rx_mini_skbprd, rx_jumbo_skbprd;
666 u32 cur_rx;
668 struct rx_desc *rx_std_ring;
669 struct rx_desc *rx_jumbo_ring;
670 struct rx_desc *rx_mini_ring;
671 struct rx_desc *rx_return_ring;
673 #if ACENIC_DO_VLAN
674 struct vlan_group *vlgrp;
675 #endif
677 int tasklet_pending, jumbo;
678 struct tasklet_struct ace_tasklet;
680 struct event *evt_ring;
682 volatile u32 *evt_prd, *rx_ret_prd, *tx_csm;
684 dma_addr_t tx_ring_dma; /* 32/64 bit */
685 dma_addr_t rx_ring_base_dma;
686 dma_addr_t evt_ring_dma;
687 dma_addr_t evt_prd_dma, rx_ret_prd_dma, tx_csm_dma;
689 unsigned char *trace_buf;
690 struct pci_dev *pdev;
691 struct net_device *next;
692 volatile int fw_running;
693 int board_idx;
694 u16 pci_command;
695 u8 pci_latency;
696 const char *name;
697 #ifdef INDEX_DEBUG
698 spinlock_t debug_lock
699 __attribute__ ((aligned (SMP_CACHE_BYTES)));
700 u32 last_tx, last_std_rx, last_mini_rx;
701 #endif
702 struct net_device_stats stats;
703 int pci_using_dac;
707 #define TX_RESERVED MAX_SKB_FRAGS
709 static inline int tx_space (struct ace_private *ap, u32 csm, u32 prd)
711 return (csm - prd - 1) & (ACE_TX_RING_ENTRIES(ap) - 1);
714 #define tx_free(ap) tx_space((ap)->tx_ret_csm, (ap)->tx_prd, ap)
716 #if MAX_SKB_FRAGS
717 #define tx_ring_full(ap, csm, prd) (tx_space(ap, csm, prd) <= TX_RESERVED)
718 #else
719 #define tx_ring_full 0
720 #endif
723 static inline void set_aceaddr(aceaddr *aa, dma_addr_t addr)
725 u64 baddr = (u64) addr;
726 aa->addrlo = baddr & 0xffffffff;
727 aa->addrhi = baddr >> 32;
728 wmb();
732 static inline void ace_set_txprd(struct ace_regs *regs,
733 struct ace_private *ap, u32 value)
735 #ifdef INDEX_DEBUG
736 unsigned long flags;
737 spin_lock_irqsave(&ap->debug_lock, flags);
738 writel(value, &regs->TxPrd);
739 if (value == ap->last_tx)
740 printk(KERN_ERR "AceNIC RACE ALERT! writing identical value "
741 "to tx producer (%i)\n", value);
742 ap->last_tx = value;
743 spin_unlock_irqrestore(&ap->debug_lock, flags);
744 #else
745 writel(value, &regs->TxPrd);
746 #endif
747 wmb();
751 static inline void ace_mask_irq(struct net_device *dev)
753 struct ace_private *ap = dev->priv;
754 struct ace_regs *regs = ap->regs;
756 if (ACE_IS_TIGON_I(ap))
757 writel(1, &regs->MaskInt);
758 else
759 writel(readl(&regs->HostCtrl) | MASK_INTS, &regs->HostCtrl);
761 ace_sync_irq(dev->irq);
765 static inline void ace_unmask_irq(struct net_device *dev)
767 struct ace_private *ap = dev->priv;
768 struct ace_regs *regs = ap->regs;
770 if (ACE_IS_TIGON_I(ap))
771 writel(0, &regs->MaskInt);
772 else
773 writel(readl(&regs->HostCtrl) & ~MASK_INTS, &regs->HostCtrl);
778 * Prototypes
780 static int ace_init(struct net_device *dev);
781 static void ace_load_std_rx_ring(struct ace_private *ap, int nr_bufs);
782 static void ace_load_mini_rx_ring(struct ace_private *ap, int nr_bufs);
783 static void ace_load_jumbo_rx_ring(struct ace_private *ap, int nr_bufs);
784 static irqreturn_t ace_interrupt(int irq, void *dev_id, struct pt_regs *regs);
785 static int ace_load_firmware(struct net_device *dev);
786 static int ace_open(struct net_device *dev);
787 static int ace_start_xmit(struct sk_buff *skb, struct net_device *dev);
788 static int ace_close(struct net_device *dev);
789 static void ace_tasklet(unsigned long dev);
790 static void ace_dump_trace(struct ace_private *ap);
791 static void ace_set_multicast_list(struct net_device *dev);
792 static int ace_change_mtu(struct net_device *dev, int new_mtu);
793 static int ace_set_mac_addr(struct net_device *dev, void *p);
794 static void ace_set_rxtx_parms(struct net_device *dev, int jumbo);
795 static int ace_allocate_descriptors(struct net_device *dev);
796 static void ace_free_descriptors(struct net_device *dev);
797 static void ace_init_cleanup(struct net_device *dev);
798 static struct net_device_stats *ace_get_stats(struct net_device *dev);
799 static int read_eeprom_byte(struct net_device *dev, unsigned long offset);
800 #if ACENIC_DO_VLAN
801 static void ace_vlan_rx_register(struct net_device *dev, struct vlan_group *grp);
802 static void ace_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid);
803 #endif
805 #endif /* _ACENIC_H_ */