1 #ifndef _ACENIC_H_
2 #define _ACENIC_H_
3
4
5 /*
6 * Generate TX index update each time, when TX ring is closed.
7 * Normally, this is not useful, because results in more dma (and irqs
8 * without TX_COAL_INTS_ONLY).
9 */
10 #define USE_TX_COAL_NOW 0
11
12 /*
13 * Addressing:
14 *
15 * The Tigon uses 64-bit host addresses, regardless of their actual
16 * length, and it expects a big-endian format. For 32 bit systems the
17 * upper 32 bits of the address are simply ignored (zero), however for
18 * little endian 64 bit systems (Alpha) this looks strange with the
19 * two parts of the address word being swapped.
20 *
21 * The addresses are split in two 32 bit words for all architectures
22 * as some of them are in PCI shared memory and it is necessary to use
23 * readl/writel to access them.
24 *
25 * The addressing code is derived from Pete Wyckoff's work, but
26 * modified to deal properly with readl/writel usage.
27 */
28
29 struct ace_regs {
30 u32 pad0[16]; /* PCI control registers */
31
32 u32 HostCtrl; /* 0x40 */
33 u32 LocalCtrl;
34
35 u32 pad1[2];
36
37 u32 MiscCfg; /* 0x50 */
38
39 u32 pad2[2];
40
41 u32 PciState;
42
43 u32 pad3[2]; /* 0x60 */
44
45 u32 WinBase;
46 u32 WinData;
47
48 u32 pad4[12]; /* 0x70 */
49
50 u32 DmaWriteState; /* 0xa0 */
51 u32 pad5[3];
52 u32 DmaReadState; /* 0xb0 */
53
54 u32 pad6[26];
55
56 u32 AssistState;
57
58 u32 pad7[8]; /* 0x120 */
59
60 u32 CpuCtrl; /* 0x140 */
61 u32 Pc;
62
63 u32 pad8[3];
64
65 u32 SramAddr; /* 0x154 */
66 u32 SramData;
67
68 u32 pad9[49];
69
70 u32 MacRxState; /* 0x220 */
71
72 u32 pad10[7];
73
74 u32 CpuBCtrl; /* 0x240 */
75 u32 PcB;
76
77 u32 pad11[3];
78
79 u32 SramBAddr; /* 0x254 */
80 u32 SramBData;
81
82 u32 pad12[105];
83
84 u32 pad13[32]; /* 0x400 */
85 u32 Stats[32];
86
87 u32 Mb0Hi; /* 0x500 */
88 u32 Mb0Lo;
89 u32 Mb1Hi;
90 u32 CmdPrd;
91 u32 Mb2Hi;
92 u32 TxPrd;
93 u32 Mb3Hi;
94 u32 RxStdPrd;
95 u32 Mb4Hi;
96 u32 RxJumboPrd;
97 u32 Mb5Hi;
98 u32 RxMiniPrd;
99 u32 Mb6Hi;
100 u32 Mb6Lo;
101 u32 Mb7Hi;
102 u32 Mb7Lo;
103 u32 Mb8Hi;
104 u32 Mb8Lo;
105 u32 Mb9Hi;
106 u32 Mb9Lo;
107 u32 MbAHi;
108 u32 MbALo;
109 u32 MbBHi;
110 u32 MbBLo;
111 u32 MbCHi;
112 u32 MbCLo;
113 u32 MbDHi;
114 u32 MbDLo;
115 u32 MbEHi;
116 u32 MbELo;
117 u32 MbFHi;
118 u32 MbFLo;
119
120 u32 pad14[32];
121
122 u32 MacAddrHi; /* 0x600 */
123 u32 MacAddrLo;
124 u32 InfoPtrHi;
125 u32 InfoPtrLo;
126 u32 MultiCastHi; /* 0x610 */
127 u32 MultiCastLo;
128 u32 ModeStat;
129 u32 DmaReadCfg;
130 u32 DmaWriteCfg; /* 0x620 */
131 u32 TxBufRat;
132 u32 EvtCsm;
133 u32 CmdCsm;
134 u32 TuneRxCoalTicks;/* 0x630 */
135 u32 TuneTxCoalTicks;
136 u32 TuneStatTicks;
137 u32 TuneMaxTxDesc;
138 u32 TuneMaxRxDesc; /* 0x640 */
139 u32 TuneTrace;
140 u32 TuneLink;
141 u32 TuneFastLink;
142 u32 TracePtr; /* 0x650 */
143 u32 TraceStrt;
144 u32 TraceLen;
145 u32 IfIdx;
146 u32 IfMtu; /* 0x660 */
147 u32 MaskInt;
148 u32 GigLnkState;
149 u32 FastLnkState;
150 u32 pad16[4]; /* 0x670 */
151 u32 RxRetCsm; /* 0x680 */
152
153 u32 pad17[31];
154
155 u32 CmdRng[64]; /* 0x700 */
156 u32 Window[0x200];
157 };
158
159
160 typedef struct {
161 u32 addrhi;
162 u32 addrlo;
163 } aceaddr;
164
165
166 #define ACE_WINDOW_SIZE 0x800
167
168 #define ACE_JUMBO_MTU 9000
169 #define ACE_STD_MTU 1500
170
171 #define ACE_TRACE_SIZE 0x8000
172
173 /*
174 * Host control register bits.
175 */
176
177 #define IN_INT 0x01
178 #define CLR_INT 0x02
179 #define HW_RESET 0x08
180 #define BYTE_SWAP 0x10
181 #define WORD_SWAP 0x20
182 #define MASK_INTS 0x40
183
184 /*
185 * Local control register bits.
186 */
187
188 #define EEPROM_DATA_IN 0x800000
189 #define EEPROM_DATA_OUT 0x400000
190 #define EEPROM_WRITE_ENABLE 0x200000
191 #define EEPROM_CLK_OUT 0x100000
192
193 #define EEPROM_BASE 0xa0000000
194
195 #define EEPROM_WRITE_SELECT 0xa0
196 #define EEPROM_READ_SELECT 0xa1
197
198 #define SRAM_BANK_512K 0x200
199
200
201 /*
202 * udelay() values for when clocking the eeprom
203 */
204 #define ACE_SHORT_DELAY 2
205 #define ACE_LONG_DELAY 4
206
207
208 /*
209 * Misc Config bits
210 */
211
212 #define SYNC_SRAM_TIMING 0x100000
213
214
215 /*
216 * CPU state bits.
217 */
218
219 #define CPU_RESET 0x01
220 #define CPU_TRACE 0x02
221 #define CPU_PROM_FAILED 0x10
222 #define CPU_HALT 0x00010000
223 #define CPU_HALTED 0xffff0000
224
225
226 /*
227 * PCI State bits.
228 */
229
230 #define DMA_READ_MAX_4 0x04
231 #define DMA_READ_MAX_16 0x08
232 #define DMA_READ_MAX_32 0x0c
233 #define DMA_READ_MAX_64 0x10
234 #define DMA_READ_MAX_128 0x14
235 #define DMA_READ_MAX_256 0x18
236 #define DMA_READ_MAX_1K 0x1c
237 #define DMA_WRITE_MAX_4 0x20
238 #define DMA_WRITE_MAX_16 0x40
239 #define DMA_WRITE_MAX_32 0x60
240 #define DMA_WRITE_MAX_64 0x80
241 #define DMA_WRITE_MAX_128 0xa0
242 #define DMA_WRITE_MAX_256 0xc0
243 #define DMA_WRITE_MAX_1K 0xe0
244 #define DMA_READ_WRITE_MASK 0xfc
245 #define MEM_READ_MULTIPLE 0x00020000
246 #define PCI_66MHZ 0x00080000
247 #define PCI_32BIT 0x00100000
248 #define DMA_WRITE_ALL_ALIGN 0x00800000
249 #define READ_CMD_MEM 0x06000000
250 #define WRITE_CMD_MEM 0x70000000
251
252
253 /*
254 * Mode status
255 */
256
257 #define ACE_BYTE_SWAP_BD 0x02
258 #define ACE_WORD_SWAP_BD 0x04 /* not actually used */
259 #define ACE_WARN 0x08
260 #define ACE_BYTE_SWAP_DMA 0x10
261 #define ACE_NO_JUMBO_FRAG 0x200
262 #define ACE_FATAL 0x40000000
263
264
265 /*
266 * DMA config
267 */
268
269 #define DMA_THRESH_1W 0x10
270 #define DMA_THRESH_2W 0x20
271 #define DMA_THRESH_4W 0x40
272 #define DMA_THRESH_8W 0x80
273 #define DMA_THRESH_16W 0x100
274 #define DMA_THRESH_32W 0x0 /* not described in doc, but exists. */
275
276
277 /*
278 * Tuning parameters
279 */
280
281 #define TICKS_PER_SEC 1000000
282
283
284 /*
285 * Link bits
286 */
287
288 #define LNK_PREF 0x00008000
289 #define LNK_10MB 0x00010000
290 #define LNK_100MB 0x00020000
291 #define LNK_1000MB 0x00040000
292 #define LNK_FULL_DUPLEX 0x00080000
293 #define LNK_HALF_DUPLEX 0x00100000
294 #define LNK_TX_FLOW_CTL_Y 0x00200000
295 #define LNK_NEG_ADVANCED 0x00400000
296 #define LNK_RX_FLOW_CTL_Y 0x00800000
297 #define LNK_NIC 0x01000000
298 #define LNK_JAM 0x02000000
299 #define LNK_JUMBO 0x04000000
300 #define LNK_ALTEON 0x08000000
301 #define LNK_NEG_FCTL 0x10000000
302 #define LNK_NEGOTIATE 0x20000000
303 #define LNK_ENABLE 0x40000000
304 #define LNK_UP 0x80000000
305
306
307 /*
308 * Event definitions
309 */
310
311 #define EVT_RING_ENTRIES 256
312 #define EVT_RING_SIZE (EVT_RING_ENTRIES * sizeof(struct event))
313
314 struct event {
315 #ifdef __LITTLE_ENDIAN_BITFIELD
316 u32 idx:12;
317 u32 code:12;
318 u32 evt:8;
319 #else
320 u32 evt:8;
321 u32 code:12;
322 u32 idx:12;
323 #endif
324 u32 pad;
325 };
326
327
328 /*
329 * Events
330 */
331
332 #define E_FW_RUNNING 0x01
333 #define E_STATS_UPDATED 0x04
334
335 #define E_STATS_UPDATE 0x04
336
337 #define E_LNK_STATE 0x06
338 #define E_C_LINK_UP 0x01
339 #define E_C_LINK_DOWN 0x02
340 #define E_C_LINK_10_100 0x03
341
342 #define E_ERROR 0x07
343 #define E_C_ERR_INVAL_CMD 0x01
344 #define E_C_ERR_UNIMP_CMD 0x02
345 #define E_C_ERR_BAD_CFG 0x03
346
347 #define E_MCAST_LIST 0x08
348 #define E_C_MCAST_ADDR_ADD 0x01
349 #define E_C_MCAST_ADDR_DEL 0x02
350
351 #define E_RESET_JUMBO_RNG 0x09
352
353
354 /*
355 * Commands
356 */
357
358 #define CMD_RING_ENTRIES 64
359
360 struct cmd {
361 #ifdef __LITTLE_ENDIAN_BITFIELD
362 u32 idx:12;
363 u32 code:12;
364 u32 evt:8;
365 #else
366 u32 evt:8;
367 u32 code:12;
368 u32 idx:12;
369 #endif
370 };
371
372
373 #define C_HOST_STATE 0x01
374 #define C_C_STACK_UP 0x01
375 #define C_C_STACK_DOWN 0x02
376
377 #define C_FDR_FILTERING 0x02
378 #define C_C_FDR_FILT_ENABLE 0x01
379 #define C_C_FDR_FILT_DISABLE 0x02
380
381 #define C_SET_RX_PRD_IDX 0x03
382 #define C_UPDATE_STATS 0x04
383 #define C_RESET_JUMBO_RNG 0x05
384 #define C_ADD_MULTICAST_ADDR 0x08
385 #define C_DEL_MULTICAST_ADDR 0x09
386
387 #define C_SET_PROMISC_MODE 0x0a
388 #define C_C_PROMISC_ENABLE 0x01
389 #define C_C_PROMISC_DISABLE 0x02
390
391 #define C_LNK_NEGOTIATION 0x0b
392 #define C_C_NEGOTIATE_BOTH 0x00
393 #define C_C_NEGOTIATE_GIG 0x01
394 #define C_C_NEGOTIATE_10_100 0x02
395
396 #define C_SET_MAC_ADDR 0x0c
397 #define C_CLEAR_PROFILE 0x0d
398
399 #define C_SET_MULTICAST_MODE 0x0e
400 #define C_C_MCAST_ENABLE 0x01
401 #define C_C_MCAST_DISABLE 0x02
402
403 #define C_CLEAR_STATS 0x0f
404 #define C_SET_RX_JUMBO_PRD_IDX 0x10
405 #define C_REFRESH_STATS 0x11
406
407
408 /*
409 * Descriptor flags
410 */
411 #define BD_FLG_TCP_UDP_SUM 0x01
412 #define BD_FLG_IP_SUM 0x02
413 #define BD_FLG_END 0x04
414 #define BD_FLG_MORE 0x08
415 #define BD_FLG_JUMBO 0x10
416 #define BD_FLG_UCAST 0x20
417 #define BD_FLG_MCAST 0x40
418 #define BD_FLG_BCAST 0x60
419 #define BD_FLG_TYP_MASK 0x60
420 #define BD_FLG_IP_FRAG 0x80
421 #define BD_FLG_IP_FRAG_END 0x100
422 #define BD_FLG_VLAN_TAG 0x200
423 #define BD_FLG_FRAME_ERROR 0x400
424 #define BD_FLG_COAL_NOW 0x800
425 #define BD_FLG_MINI 0x1000
426
427
428 /*
429 * Ring Control block flags
430 */
431 #define RCB_FLG_TCP_UDP_SUM 0x01
432 #define RCB_FLG_IP_SUM 0x02
433 #define RCB_FLG_NO_PSEUDO_HDR 0x08
434 #define RCB_FLG_VLAN_ASSIST 0x10
435 #define RCB_FLG_COAL_INT_ONLY 0x20
436 #define RCB_FLG_TX_HOST_RING 0x40
437 #define RCB_FLG_IEEE_SNAP_SUM 0x80
438 #define RCB_FLG_EXT_RX_BD 0x100
439 #define RCB_FLG_RNG_DISABLE 0x200
440
441
442 /*
443 * TX ring - maximum TX ring entries for Tigon I's is 128
444 */
445 #define MAX_TX_RING_ENTRIES 256
446 #define TIGON_I_TX_RING_ENTRIES 128
447 #define TX_RING_SIZE (MAX_TX_RING_ENTRIES * sizeof(struct tx_desc))
448 #define TX_RING_BASE 0x3800
449
450 struct tx_desc{
451 aceaddr addr;
452 u32 flagsize;
453 #if 0
454 /*
455 * This is in PCI shared mem and must be accessed with readl/writel
456 * real layout is:
457 */
458 #if __LITTLE_ENDIAN
459 u16 flags;
460 u16 size;
461 u16 vlan;
462 u16 reserved;
463 #else
464 u16 size;
465 u16 flags;
466 u16 reserved;
467 u16 vlan;
468 #endif
469 #endif
470 u32 vlanres;
471 };
472
473
474 #define RX_STD_RING_ENTRIES 512
475 #define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc))
476
477 #define RX_JUMBO_RING_ENTRIES 256
478 #define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc))
479
480 #define RX_MINI_RING_ENTRIES 1024
481 #define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc))
482
483 #define RX_RETURN_RING_ENTRIES 2048
484 #define RX_RETURN_RING_SIZE (RX_MAX_RETURN_RING_ENTRIES * \
485 sizeof(struct rx_desc))
486
487 struct rx_desc{
488 aceaddr addr;
489 #ifdef __LITTLE_ENDIAN
490 u16 size;
491 u16 idx;
492 #else
493 u16 idx;
494 u16 size;
495 #endif
496 #ifdef __LITTLE_ENDIAN
497 u16 flags;
498 u16 type;
499 #else
500 u16 type;
501 u16 flags;
502 #endif
503 #ifdef __LITTLE_ENDIAN
504 u16 tcp_udp_csum;
505 u16 ip_csum;
506 #else
507 u16 ip_csum;
508 u16 tcp_udp_csum;
509 #endif
510 #ifdef __LITTLE_ENDIAN
511 u16 vlan;
512 u16 err_flags;
513 #else
514 u16 err_flags;
515 u16 vlan;
516 #endif
517 u32 reserved;
518 u32 opague;
519 };
520
521
522 /*
523 * This struct is shared with the NIC firmware.
524 */
525 struct ring_ctrl {
526 aceaddr rngptr;
527 #ifdef __LITTLE_ENDIAN
528 u16 flags;
529 u16 max_len;
530 #else
531 u16 max_len;
532 u16 flags;
533 #endif
534 u32 pad;
535 };
536
537
538 struct ace_mac_stats {
539 u32 excess_colls;
540 u32 coll_1;
541 u32 coll_2;
542 u32 coll_3;
543 u32 coll_4;
544 u32 coll_5;
545 u32 coll_6;
546 u32 coll_7;
547 u32 coll_8;
548 u32 coll_9;
549 u32 coll_10;
550 u32 coll_11;
551 u32 coll_12;
552 u32 coll_13;
553 u32 coll_14;
554 u32 coll_15;
555 u32 late_coll;
556 u32 defers;
557 u32 crc_err;
558 u32 underrun;
559 u32 crs_err;
560 u32 pad[3];
561 u32 drop_ula;
562 u32 drop_mc;
563 u32 drop_fc;
564 u32 drop_space;
565 u32 coll;
566 u32 kept_bc;
567 u32 kept_mc;
568 u32 kept_uc;
569 };
570
571
572 struct ace_info {
573 union {
574 u32 stats[256];
575 } s;
576 struct ring_ctrl evt_ctrl;
577 struct ring_ctrl cmd_ctrl;
578 struct ring_ctrl tx_ctrl;
579 struct ring_ctrl rx_std_ctrl;
580 struct ring_ctrl rx_jumbo_ctrl;
581 struct ring_ctrl rx_mini_ctrl;
582 struct ring_ctrl rx_return_ctrl;
583 aceaddr evt_prd_ptr;
584 aceaddr rx_ret_prd_ptr;
585 aceaddr tx_csm_ptr;
586 aceaddr stats2_ptr;
587 };
588
589
590 struct ring_info {
591 struct sk_buff *skb;
592 DECLARE_PCI_UNMAP_ADDR(mapping)
593 };
594
595
596 /*
597 * Funny... As soon as we add maplen on alpha, it starts to work
598 * much slower. Hmm... is it because struct does not fit to one cacheline?
599 * So, split tx_ring_info.
600 */
601 struct tx_ring_info {
602 struct sk_buff *skb;
603 DECLARE_PCI_UNMAP_ADDR(mapping)
604 DECLARE_PCI_UNMAP_LEN(maplen)
605 };
606
607
608 /*
609 * struct ace_skb holding the rings of skb's. This is an awful lot of
610 * pointers, but I don't see any other smart mode to do this in an
611 * efficient manner ;-(
612 */
613 struct ace_skb
614 {
615 struct tx_ring_info tx_skbuff[MAX_TX_RING_ENTRIES];
616 struct ring_info rx_std_skbuff[RX_STD_RING_ENTRIES];
617 struct ring_info rx_mini_skbuff[RX_MINI_RING_ENTRIES];
618 struct ring_info rx_jumbo_skbuff[RX_JUMBO_RING_ENTRIES];
619 };
620
621
622 /*
623 * Struct private for the AceNIC.
624 *
625 * Elements are grouped so variables used by the tx handling goes
626 * together, and will go into the same cache lines etc. in order to
627 * avoid cache line contention between the rx and tx handling on SMP.
628 *
629 * Frequently accessed variables are put at the beginning of the
630 * struct to help the compiler generate better/shorter code.
631 */
632 struct ace_private
633 {
634 struct ace_info *info;
635 struct ace_regs __iomem *regs; /* register base */
636 struct ace_skb *skb;
637 dma_addr_t info_dma; /* 32/64 bit */
638
639 int version, link;
640 int promisc, mcast_all;
641
642 /*
643 * TX elements
644 */
645 struct tx_desc *tx_ring;
646 u32 tx_prd;
647 volatile u32 tx_ret_csm;
648 int tx_ring_entries;
649
650 /*
651 * RX elements
652 */
653 unsigned long std_refill_busy
654 __attribute__ ((aligned (SMP_CACHE_BYTES)));
655 unsigned long mini_refill_busy, jumbo_refill_busy;
656 atomic_t cur_rx_bufs;
657 atomic_t cur_mini_bufs;
658 atomic_t cur_jumbo_bufs;
659 u32 rx_std_skbprd, rx_mini_skbprd, rx_jumbo_skbprd;
660 u32 cur_rx;
661
662 struct rx_desc *rx_std_ring;
663 struct rx_desc *rx_jumbo_ring;
664 struct rx_desc *rx_mini_ring;
665 struct rx_desc *rx_return_ring;
666
667 #if ACENIC_DO_VLAN
668 struct vlan_group *vlgrp;
669 #endif
670
671 int tasklet_pending, jumbo;
672 struct tasklet_struct ace_tasklet;
673
674 struct event *evt_ring;
675
676 volatile u32 *evt_prd, *rx_ret_prd, *tx_csm;
677
678 dma_addr_t tx_ring_dma; /* 32/64 bit */
679 dma_addr_t rx_ring_base_dma;
680 dma_addr_t evt_ring_dma;
681 dma_addr_t evt_prd_dma, rx_ret_prd_dma, tx_csm_dma;
682
683 unsigned char *trace_buf;
684 struct pci_dev *pdev;
685 struct net_device *next;
686 volatile int fw_running;
687 int board_idx;
688 u16 pci_command;
689 u8 pci_latency;
690 const char *name;
691 #ifdef INDEX_DEBUG
692 spinlock_t debug_lock
693 __attribute__ ((aligned (SMP_CACHE_BYTES)));
694 u32 last_tx, last_std_rx, last_mini_rx;
695 #endif
696 int pci_using_dac;
697 u8 firmware_major;
698 u8 firmware_minor;
699 u8 firmware_fix;
700 u32 firmware_start;
701 };
702
703
704 #define TX_RESERVED MAX_SKB_FRAGS
705
tx_space(struct ace_private * ap,u32 csm,u32 prd)706 static inline int tx_space (struct ace_private *ap, u32 csm, u32 prd)
707 {
708 return (csm - prd - 1) & (ACE_TX_RING_ENTRIES(ap) - 1);
709 }
710
711 #define tx_free(ap) tx_space((ap)->tx_ret_csm, (ap)->tx_prd, ap)
712 #define tx_ring_full(ap, csm, prd) (tx_space(ap, csm, prd) <= TX_RESERVED)
713
set_aceaddr(aceaddr * aa,dma_addr_t addr)714 static inline void set_aceaddr(aceaddr *aa, dma_addr_t addr)
715 {
716 u64 baddr = (u64) addr;
717 aa->addrlo = baddr & 0xffffffff;
718 aa->addrhi = baddr >> 32;
719 wmb();
720 }
721
722
ace_set_txprd(struct ace_regs __iomem * regs,struct ace_private * ap,u32 value)723 static inline void ace_set_txprd(struct ace_regs __iomem *regs,
724 struct ace_private *ap, u32 value)
725 {
726 #ifdef INDEX_DEBUG
727 unsigned long flags;
728 spin_lock_irqsave(&ap->debug_lock, flags);
729 writel(value, ®s->TxPrd);
730 if (value == ap->last_tx)
731 printk(KERN_ERR "AceNIC RACE ALERT! writing identical value "
732 "to tx producer (%i)\n", value);
733 ap->last_tx = value;
734 spin_unlock_irqrestore(&ap->debug_lock, flags);
735 #else
736 writel(value, ®s->TxPrd);
737 #endif
738 wmb();
739 }
740
741
ace_mask_irq(struct net_device * dev)742 static inline void ace_mask_irq(struct net_device *dev)
743 {
744 struct ace_private *ap = netdev_priv(dev);
745 struct ace_regs __iomem *regs = ap->regs;
746
747 if (ACE_IS_TIGON_I(ap))
748 writel(1, ®s->MaskInt);
749 else
750 writel(readl(®s->HostCtrl) | MASK_INTS, ®s->HostCtrl);
751
752 ace_sync_irq(dev->irq);
753 }
754
755
ace_unmask_irq(struct net_device * dev)756 static inline void ace_unmask_irq(struct net_device *dev)
757 {
758 struct ace_private *ap = netdev_priv(dev);
759 struct ace_regs __iomem *regs = ap->regs;
760
761 if (ACE_IS_TIGON_I(ap))
762 writel(0, ®s->MaskInt);
763 else
764 writel(readl(®s->HostCtrl) & ~MASK_INTS, ®s->HostCtrl);
765 }
766
767
768 /*
769 * Prototypes
770 */
771 static int ace_init(struct net_device *dev);
772 static void ace_load_std_rx_ring(struct ace_private *ap, int nr_bufs);
773 static void ace_load_mini_rx_ring(struct ace_private *ap, int nr_bufs);
774 static void ace_load_jumbo_rx_ring(struct ace_private *ap, int nr_bufs);
775 static irqreturn_t ace_interrupt(int irq, void *dev_id);
776 static int ace_load_firmware(struct net_device *dev);
777 static int ace_open(struct net_device *dev);
778 static int ace_start_xmit(struct sk_buff *skb, struct net_device *dev);
779 static int ace_close(struct net_device *dev);
780 static void ace_tasklet(unsigned long dev);
781 static void ace_dump_trace(struct ace_private *ap);
782 static void ace_set_multicast_list(struct net_device *dev);
783 static int ace_change_mtu(struct net_device *dev, int new_mtu);
784 static int ace_set_mac_addr(struct net_device *dev, void *p);
785 static void ace_set_rxtx_parms(struct net_device *dev, int jumbo);
786 static int ace_allocate_descriptors(struct net_device *dev);
787 static void ace_free_descriptors(struct net_device *dev);
788 static void ace_init_cleanup(struct net_device *dev);
789 static struct net_device_stats *ace_get_stats(struct net_device *dev);
790 static int read_eeprom_byte(struct net_device *dev, unsigned long offset);
791 #if ACENIC_DO_VLAN
792 static void ace_vlan_rx_register(struct net_device *dev, struct vlan_group *grp);
793 #endif
794
795 #endif /* _ACENIC_H_ */
796