792 lines
16 KiB
C
792 lines
16 KiB
C
#ifndef _ACENIC_H_
|
|
#define _ACENIC_H_
|
|
|
|
|
|
/*
|
|
* Generate TX index update each time, when TX ring is closed.
|
|
* Normally, this is not useful, because results in more dma (and irqs
|
|
* without TX_COAL_INTS_ONLY).
|
|
*/
|
|
#define USE_TX_COAL_NOW 0
|
|
|
|
/*
|
|
* Addressing:
|
|
*
|
|
* The Tigon uses 64-bit host addresses, regardless of their actual
|
|
* length, and it expects a big-endian format. For 32 bit systems the
|
|
* upper 32 bits of the address are simply ignored (zero), however for
|
|
* little endian 64 bit systems (Alpha) this looks strange with the
|
|
* two parts of the address word being swapped.
|
|
*
|
|
* The addresses are split in two 32 bit words for all architectures
|
|
* as some of them are in PCI shared memory and it is necessary to use
|
|
* readl/writel to access them.
|
|
*
|
|
* The addressing code is derived from Pete Wyckoff's work, but
|
|
* modified to deal properly with readl/writel usage.
|
|
*/
|
|
|
|
struct ace_regs {
|
|
u32 pad0[16]; /* PCI control registers */
|
|
|
|
u32 HostCtrl; /* 0x40 */
|
|
u32 LocalCtrl;
|
|
|
|
u32 pad1[2];
|
|
|
|
u32 MiscCfg; /* 0x50 */
|
|
|
|
u32 pad2[2];
|
|
|
|
u32 PciState;
|
|
|
|
u32 pad3[2]; /* 0x60 */
|
|
|
|
u32 WinBase;
|
|
u32 WinData;
|
|
|
|
u32 pad4[12]; /* 0x70 */
|
|
|
|
u32 DmaWriteState; /* 0xa0 */
|
|
u32 pad5[3];
|
|
u32 DmaReadState; /* 0xb0 */
|
|
|
|
u32 pad6[26];
|
|
|
|
u32 AssistState;
|
|
|
|
u32 pad7[8]; /* 0x120 */
|
|
|
|
u32 CpuCtrl; /* 0x140 */
|
|
u32 Pc;
|
|
|
|
u32 pad8[3];
|
|
|
|
u32 SramAddr; /* 0x154 */
|
|
u32 SramData;
|
|
|
|
u32 pad9[49];
|
|
|
|
u32 MacRxState; /* 0x220 */
|
|
|
|
u32 pad10[7];
|
|
|
|
u32 CpuBCtrl; /* 0x240 */
|
|
u32 PcB;
|
|
|
|
u32 pad11[3];
|
|
|
|
u32 SramBAddr; /* 0x254 */
|
|
u32 SramBData;
|
|
|
|
u32 pad12[105];
|
|
|
|
u32 pad13[32]; /* 0x400 */
|
|
u32 Stats[32];
|
|
|
|
u32 Mb0Hi; /* 0x500 */
|
|
u32 Mb0Lo;
|
|
u32 Mb1Hi;
|
|
u32 CmdPrd;
|
|
u32 Mb2Hi;
|
|
u32 TxPrd;
|
|
u32 Mb3Hi;
|
|
u32 RxStdPrd;
|
|
u32 Mb4Hi;
|
|
u32 RxJumboPrd;
|
|
u32 Mb5Hi;
|
|
u32 RxMiniPrd;
|
|
u32 Mb6Hi;
|
|
u32 Mb6Lo;
|
|
u32 Mb7Hi;
|
|
u32 Mb7Lo;
|
|
u32 Mb8Hi;
|
|
u32 Mb8Lo;
|
|
u32 Mb9Hi;
|
|
u32 Mb9Lo;
|
|
u32 MbAHi;
|
|
u32 MbALo;
|
|
u32 MbBHi;
|
|
u32 MbBLo;
|
|
u32 MbCHi;
|
|
u32 MbCLo;
|
|
u32 MbDHi;
|
|
u32 MbDLo;
|
|
u32 MbEHi;
|
|
u32 MbELo;
|
|
u32 MbFHi;
|
|
u32 MbFLo;
|
|
|
|
u32 pad14[32];
|
|
|
|
u32 MacAddrHi; /* 0x600 */
|
|
u32 MacAddrLo;
|
|
u32 InfoPtrHi;
|
|
u32 InfoPtrLo;
|
|
u32 MultiCastHi; /* 0x610 */
|
|
u32 MultiCastLo;
|
|
u32 ModeStat;
|
|
u32 DmaReadCfg;
|
|
u32 DmaWriteCfg; /* 0x620 */
|
|
u32 TxBufRat;
|
|
u32 EvtCsm;
|
|
u32 CmdCsm;
|
|
u32 TuneRxCoalTicks;/* 0x630 */
|
|
u32 TuneTxCoalTicks;
|
|
u32 TuneStatTicks;
|
|
u32 TuneMaxTxDesc;
|
|
u32 TuneMaxRxDesc; /* 0x640 */
|
|
u32 TuneTrace;
|
|
u32 TuneLink;
|
|
u32 TuneFastLink;
|
|
u32 TracePtr; /* 0x650 */
|
|
u32 TraceStrt;
|
|
u32 TraceLen;
|
|
u32 IfIdx;
|
|
u32 IfMtu; /* 0x660 */
|
|
u32 MaskInt;
|
|
u32 GigLnkState;
|
|
u32 FastLnkState;
|
|
u32 pad16[4]; /* 0x670 */
|
|
u32 RxRetCsm; /* 0x680 */
|
|
|
|
u32 pad17[31];
|
|
|
|
u32 CmdRng[64]; /* 0x700 */
|
|
u32 Window[0x200];
|
|
};
|
|
|
|
|
|
typedef struct {
|
|
u32 addrhi;
|
|
u32 addrlo;
|
|
} aceaddr;
|
|
|
|
|
|
#define ACE_WINDOW_SIZE 0x800
|
|
|
|
#define ACE_JUMBO_MTU 9000
|
|
#define ACE_STD_MTU 1500
|
|
|
|
#define ACE_TRACE_SIZE 0x8000
|
|
|
|
/*
|
|
* Host control register bits.
|
|
*/
|
|
|
|
#define IN_INT 0x01
|
|
#define CLR_INT 0x02
|
|
#define HW_RESET 0x08
|
|
#define BYTE_SWAP 0x10
|
|
#define WORD_SWAP 0x20
|
|
#define MASK_INTS 0x40
|
|
|
|
/*
|
|
* Local control register bits.
|
|
*/
|
|
|
|
#define EEPROM_DATA_IN 0x800000
|
|
#define EEPROM_DATA_OUT 0x400000
|
|
#define EEPROM_WRITE_ENABLE 0x200000
|
|
#define EEPROM_CLK_OUT 0x100000
|
|
|
|
#define EEPROM_BASE 0xa0000000
|
|
|
|
#define EEPROM_WRITE_SELECT 0xa0
|
|
#define EEPROM_READ_SELECT 0xa1
|
|
|
|
#define SRAM_BANK_512K 0x200
|
|
|
|
|
|
/*
|
|
* udelay() values for when clocking the eeprom
|
|
*/
|
|
#define ACE_SHORT_DELAY 2
|
|
#define ACE_LONG_DELAY 4
|
|
|
|
|
|
/*
|
|
* Misc Config bits
|
|
*/
|
|
|
|
#define SYNC_SRAM_TIMING 0x100000
|
|
|
|
|
|
/*
|
|
* CPU state bits.
|
|
*/
|
|
|
|
#define CPU_RESET 0x01
|
|
#define CPU_TRACE 0x02
|
|
#define CPU_PROM_FAILED 0x10
|
|
#define CPU_HALT 0x00010000
|
|
#define CPU_HALTED 0xffff0000
|
|
|
|
|
|
/*
|
|
* PCI State bits.
|
|
*/
|
|
|
|
#define DMA_READ_MAX_4 0x04
|
|
#define DMA_READ_MAX_16 0x08
|
|
#define DMA_READ_MAX_32 0x0c
|
|
#define DMA_READ_MAX_64 0x10
|
|
#define DMA_READ_MAX_128 0x14
|
|
#define DMA_READ_MAX_256 0x18
|
|
#define DMA_READ_MAX_1K 0x1c
|
|
#define DMA_WRITE_MAX_4 0x20
|
|
#define DMA_WRITE_MAX_16 0x40
|
|
#define DMA_WRITE_MAX_32 0x60
|
|
#define DMA_WRITE_MAX_64 0x80
|
|
#define DMA_WRITE_MAX_128 0xa0
|
|
#define DMA_WRITE_MAX_256 0xc0
|
|
#define DMA_WRITE_MAX_1K 0xe0
|
|
#define DMA_READ_WRITE_MASK 0xfc
|
|
#define MEM_READ_MULTIPLE 0x00020000
|
|
#define PCI_66MHZ 0x00080000
|
|
#define PCI_32BIT 0x00100000
|
|
#define DMA_WRITE_ALL_ALIGN 0x00800000
|
|
#define READ_CMD_MEM 0x06000000
|
|
#define WRITE_CMD_MEM 0x70000000
|
|
|
|
|
|
/*
|
|
* Mode status
|
|
*/
|
|
|
|
#define ACE_BYTE_SWAP_BD 0x02
|
|
#define ACE_WORD_SWAP_BD 0x04 /* not actually used */
|
|
#define ACE_WARN 0x08
|
|
#define ACE_BYTE_SWAP_DMA 0x10
|
|
#define ACE_NO_JUMBO_FRAG 0x200
|
|
#define ACE_FATAL 0x40000000
|
|
|
|
|
|
/*
|
|
* DMA config
|
|
*/
|
|
|
|
#define DMA_THRESH_1W 0x10
|
|
#define DMA_THRESH_2W 0x20
|
|
#define DMA_THRESH_4W 0x40
|
|
#define DMA_THRESH_8W 0x80
|
|
#define DMA_THRESH_16W 0x100
|
|
#define DMA_THRESH_32W 0x0 /* not described in doc, but exists. */
|
|
|
|
|
|
/*
|
|
* Tuning parameters
|
|
*/
|
|
|
|
#define TICKS_PER_SEC 1000000
|
|
|
|
|
|
/*
|
|
* Link bits
|
|
*/
|
|
|
|
#define LNK_PREF 0x00008000
|
|
#define LNK_10MB 0x00010000
|
|
#define LNK_100MB 0x00020000
|
|
#define LNK_1000MB 0x00040000
|
|
#define LNK_FULL_DUPLEX 0x00080000
|
|
#define LNK_HALF_DUPLEX 0x00100000
|
|
#define LNK_TX_FLOW_CTL_Y 0x00200000
|
|
#define LNK_NEG_ADVANCED 0x00400000
|
|
#define LNK_RX_FLOW_CTL_Y 0x00800000
|
|
#define LNK_NIC 0x01000000
|
|
#define LNK_JAM 0x02000000
|
|
#define LNK_JUMBO 0x04000000
|
|
#define LNK_ALTEON 0x08000000
|
|
#define LNK_NEG_FCTL 0x10000000
|
|
#define LNK_NEGOTIATE 0x20000000
|
|
#define LNK_ENABLE 0x40000000
|
|
#define LNK_UP 0x80000000
|
|
|
|
|
|
/*
|
|
* Event definitions
|
|
*/
|
|
|
|
#define EVT_RING_ENTRIES 256
|
|
#define EVT_RING_SIZE (EVT_RING_ENTRIES * sizeof(struct event))
|
|
|
|
struct event {
|
|
#ifdef __LITTLE_ENDIAN_BITFIELD
|
|
u32 idx:12;
|
|
u32 code:12;
|
|
u32 evt:8;
|
|
#else
|
|
u32 evt:8;
|
|
u32 code:12;
|
|
u32 idx:12;
|
|
#endif
|
|
u32 pad;
|
|
};
|
|
|
|
|
|
/*
|
|
* Events
|
|
*/
|
|
|
|
#define E_FW_RUNNING 0x01
|
|
#define E_STATS_UPDATED 0x04
|
|
|
|
#define E_STATS_UPDATE 0x04
|
|
|
|
#define E_LNK_STATE 0x06
|
|
#define E_C_LINK_UP 0x01
|
|
#define E_C_LINK_DOWN 0x02
|
|
#define E_C_LINK_10_100 0x03
|
|
|
|
#define E_ERROR 0x07
|
|
#define E_C_ERR_INVAL_CMD 0x01
|
|
#define E_C_ERR_UNIMP_CMD 0x02
|
|
#define E_C_ERR_BAD_CFG 0x03
|
|
|
|
#define E_MCAST_LIST 0x08
|
|
#define E_C_MCAST_ADDR_ADD 0x01
|
|
#define E_C_MCAST_ADDR_DEL 0x02
|
|
|
|
#define E_RESET_JUMBO_RNG 0x09
|
|
|
|
|
|
/*
|
|
* Commands
|
|
*/
|
|
|
|
#define CMD_RING_ENTRIES 64
|
|
|
|
struct cmd {
|
|
#ifdef __LITTLE_ENDIAN_BITFIELD
|
|
u32 idx:12;
|
|
u32 code:12;
|
|
u32 evt:8;
|
|
#else
|
|
u32 evt:8;
|
|
u32 code:12;
|
|
u32 idx:12;
|
|
#endif
|
|
};
|
|
|
|
|
|
#define C_HOST_STATE 0x01
|
|
#define C_C_STACK_UP 0x01
|
|
#define C_C_STACK_DOWN 0x02
|
|
|
|
#define C_FDR_FILTERING 0x02
|
|
#define C_C_FDR_FILT_ENABLE 0x01
|
|
#define C_C_FDR_FILT_DISABLE 0x02
|
|
|
|
#define C_SET_RX_PRD_IDX 0x03
|
|
#define C_UPDATE_STATS 0x04
|
|
#define C_RESET_JUMBO_RNG 0x05
|
|
#define C_ADD_MULTICAST_ADDR 0x08
|
|
#define C_DEL_MULTICAST_ADDR 0x09
|
|
|
|
#define C_SET_PROMISC_MODE 0x0a
|
|
#define C_C_PROMISC_ENABLE 0x01
|
|
#define C_C_PROMISC_DISABLE 0x02
|
|
|
|
#define C_LNK_NEGOTIATION 0x0b
|
|
#define C_C_NEGOTIATE_BOTH 0x00
|
|
#define C_C_NEGOTIATE_GIG 0x01
|
|
#define C_C_NEGOTIATE_10_100 0x02
|
|
|
|
#define C_SET_MAC_ADDR 0x0c
|
|
#define C_CLEAR_PROFILE 0x0d
|
|
|
|
#define C_SET_MULTICAST_MODE 0x0e
|
|
#define C_C_MCAST_ENABLE 0x01
|
|
#define C_C_MCAST_DISABLE 0x02
|
|
|
|
#define C_CLEAR_STATS 0x0f
|
|
#define C_SET_RX_JUMBO_PRD_IDX 0x10
|
|
#define C_REFRESH_STATS 0x11
|
|
|
|
|
|
/*
|
|
* Descriptor flags
|
|
*/
|
|
#define BD_FLG_TCP_UDP_SUM 0x01
|
|
#define BD_FLG_IP_SUM 0x02
|
|
#define BD_FLG_END 0x04
|
|
#define BD_FLG_MORE 0x08
|
|
#define BD_FLG_JUMBO 0x10
|
|
#define BD_FLG_UCAST 0x20
|
|
#define BD_FLG_MCAST 0x40
|
|
#define BD_FLG_BCAST 0x60
|
|
#define BD_FLG_TYP_MASK 0x60
|
|
#define BD_FLG_IP_FRAG 0x80
|
|
#define BD_FLG_IP_FRAG_END 0x100
|
|
#define BD_FLG_VLAN_TAG 0x200
|
|
#define BD_FLG_FRAME_ERROR 0x400
|
|
#define BD_FLG_COAL_NOW 0x800
|
|
#define BD_FLG_MINI 0x1000
|
|
|
|
|
|
/*
|
|
* Ring Control block flags
|
|
*/
|
|
#define RCB_FLG_TCP_UDP_SUM 0x01
|
|
#define RCB_FLG_IP_SUM 0x02
|
|
#define RCB_FLG_NO_PSEUDO_HDR 0x08
|
|
#define RCB_FLG_VLAN_ASSIST 0x10
|
|
#define RCB_FLG_COAL_INT_ONLY 0x20
|
|
#define RCB_FLG_TX_HOST_RING 0x40
|
|
#define RCB_FLG_IEEE_SNAP_SUM 0x80
|
|
#define RCB_FLG_EXT_RX_BD 0x100
|
|
#define RCB_FLG_RNG_DISABLE 0x200
|
|
|
|
|
|
/*
|
|
* TX ring - maximum TX ring entries for Tigon I's is 128
|
|
*/
|
|
#define MAX_TX_RING_ENTRIES 256
|
|
#define TIGON_I_TX_RING_ENTRIES 128
|
|
#define TX_RING_SIZE (MAX_TX_RING_ENTRIES * sizeof(struct tx_desc))
|
|
#define TX_RING_BASE 0x3800
|
|
|
|
struct tx_desc{
|
|
aceaddr addr;
|
|
u32 flagsize;
|
|
#if 0
|
|
/*
|
|
* This is in PCI shared mem and must be accessed with readl/writel
|
|
* real layout is:
|
|
*/
|
|
#if __LITTLE_ENDIAN
|
|
u16 flags;
|
|
u16 size;
|
|
u16 vlan;
|
|
u16 reserved;
|
|
#else
|
|
u16 size;
|
|
u16 flags;
|
|
u16 reserved;
|
|
u16 vlan;
|
|
#endif
|
|
#endif
|
|
u32 vlanres;
|
|
};
|
|
|
|
|
|
#define RX_STD_RING_ENTRIES 512
|
|
#define RX_STD_RING_SIZE (RX_STD_RING_ENTRIES * sizeof(struct rx_desc))
|
|
|
|
#define RX_JUMBO_RING_ENTRIES 256
|
|
#define RX_JUMBO_RING_SIZE (RX_JUMBO_RING_ENTRIES *sizeof(struct rx_desc))
|
|
|
|
#define RX_MINI_RING_ENTRIES 1024
|
|
#define RX_MINI_RING_SIZE (RX_MINI_RING_ENTRIES *sizeof(struct rx_desc))
|
|
|
|
#define RX_RETURN_RING_ENTRIES 2048
|
|
#define RX_RETURN_RING_SIZE (RX_MAX_RETURN_RING_ENTRIES * \
|
|
sizeof(struct rx_desc))
|
|
|
|
struct rx_desc{
|
|
aceaddr addr;
|
|
#ifdef __LITTLE_ENDIAN
|
|
u16 size;
|
|
u16 idx;
|
|
#else
|
|
u16 idx;
|
|
u16 size;
|
|
#endif
|
|
#ifdef __LITTLE_ENDIAN
|
|
u16 flags;
|
|
u16 type;
|
|
#else
|
|
u16 type;
|
|
u16 flags;
|
|
#endif
|
|
#ifdef __LITTLE_ENDIAN
|
|
u16 tcp_udp_csum;
|
|
u16 ip_csum;
|
|
#else
|
|
u16 ip_csum;
|
|
u16 tcp_udp_csum;
|
|
#endif
|
|
#ifdef __LITTLE_ENDIAN
|
|
u16 vlan;
|
|
u16 err_flags;
|
|
#else
|
|
u16 err_flags;
|
|
u16 vlan;
|
|
#endif
|
|
u32 reserved;
|
|
u32 opague;
|
|
};
|
|
|
|
|
|
/*
|
|
* This struct is shared with the NIC firmware.
|
|
*/
|
|
struct ring_ctrl {
|
|
aceaddr rngptr;
|
|
#ifdef __LITTLE_ENDIAN
|
|
u16 flags;
|
|
u16 max_len;
|
|
#else
|
|
u16 max_len;
|
|
u16 flags;
|
|
#endif
|
|
u32 pad;
|
|
};
|
|
|
|
|
|
struct ace_mac_stats {
|
|
u32 excess_colls;
|
|
u32 coll_1;
|
|
u32 coll_2;
|
|
u32 coll_3;
|
|
u32 coll_4;
|
|
u32 coll_5;
|
|
u32 coll_6;
|
|
u32 coll_7;
|
|
u32 coll_8;
|
|
u32 coll_9;
|
|
u32 coll_10;
|
|
u32 coll_11;
|
|
u32 coll_12;
|
|
u32 coll_13;
|
|
u32 coll_14;
|
|
u32 coll_15;
|
|
u32 late_coll;
|
|
u32 defers;
|
|
u32 crc_err;
|
|
u32 underrun;
|
|
u32 crs_err;
|
|
u32 pad[3];
|
|
u32 drop_ula;
|
|
u32 drop_mc;
|
|
u32 drop_fc;
|
|
u32 drop_space;
|
|
u32 coll;
|
|
u32 kept_bc;
|
|
u32 kept_mc;
|
|
u32 kept_uc;
|
|
};
|
|
|
|
|
|
struct ace_info {
|
|
union {
|
|
u32 stats[256];
|
|
} s;
|
|
struct ring_ctrl evt_ctrl;
|
|
struct ring_ctrl cmd_ctrl;
|
|
struct ring_ctrl tx_ctrl;
|
|
struct ring_ctrl rx_std_ctrl;
|
|
struct ring_ctrl rx_jumbo_ctrl;
|
|
struct ring_ctrl rx_mini_ctrl;
|
|
struct ring_ctrl rx_return_ctrl;
|
|
aceaddr evt_prd_ptr;
|
|
aceaddr rx_ret_prd_ptr;
|
|
aceaddr tx_csm_ptr;
|
|
aceaddr stats2_ptr;
|
|
};
|
|
|
|
|
|
struct ring_info {
|
|
struct sk_buff *skb;
|
|
DECLARE_PCI_UNMAP_ADDR(mapping)
|
|
};
|
|
|
|
|
|
/*
|
|
* Funny... As soon as we add maplen on alpha, it starts to work
|
|
* much slower. Hmm... is it because struct does not fit to one cacheline?
|
|
* So, split tx_ring_info.
|
|
*/
|
|
struct tx_ring_info {
|
|
struct sk_buff *skb;
|
|
DECLARE_PCI_UNMAP_ADDR(mapping)
|
|
DECLARE_PCI_UNMAP_LEN(maplen)
|
|
};
|
|
|
|
|
|
/*
|
|
* struct ace_skb holding the rings of skb's. This is an awful lot of
|
|
* pointers, but I don't see any other smart mode to do this in an
|
|
* efficient manner ;-(
|
|
*/
|
|
struct ace_skb
|
|
{
|
|
struct tx_ring_info tx_skbuff[MAX_TX_RING_ENTRIES];
|
|
struct ring_info rx_std_skbuff[RX_STD_RING_ENTRIES];
|
|
struct ring_info rx_mini_skbuff[RX_MINI_RING_ENTRIES];
|
|
struct ring_info rx_jumbo_skbuff[RX_JUMBO_RING_ENTRIES];
|
|
};
|
|
|
|
|
|
/*
|
|
* Struct private for the AceNIC.
|
|
*
|
|
* Elements are grouped so variables used by the tx handling goes
|
|
* together, and will go into the same cache lines etc. in order to
|
|
* avoid cache line contention between the rx and tx handling on SMP.
|
|
*
|
|
* Frequently accessed variables are put at the beginning of the
|
|
* struct to help the compiler generate better/shorter code.
|
|
*/
|
|
struct ace_private
|
|
{
|
|
struct ace_info *info;
|
|
struct ace_regs __iomem *regs; /* register base */
|
|
struct ace_skb *skb;
|
|
dma_addr_t info_dma; /* 32/64 bit */
|
|
|
|
int version, link;
|
|
int promisc, mcast_all;
|
|
|
|
/*
|
|
* TX elements
|
|
*/
|
|
struct tx_desc *tx_ring;
|
|
u32 tx_prd;
|
|
volatile u32 tx_ret_csm;
|
|
int tx_ring_entries;
|
|
|
|
/*
|
|
* RX elements
|
|
*/
|
|
unsigned long std_refill_busy
|
|
__attribute__ ((aligned (SMP_CACHE_BYTES)));
|
|
unsigned long mini_refill_busy, jumbo_refill_busy;
|
|
atomic_t cur_rx_bufs;
|
|
atomic_t cur_mini_bufs;
|
|
atomic_t cur_jumbo_bufs;
|
|
u32 rx_std_skbprd, rx_mini_skbprd, rx_jumbo_skbprd;
|
|
u32 cur_rx;
|
|
|
|
struct rx_desc *rx_std_ring;
|
|
struct rx_desc *rx_jumbo_ring;
|
|
struct rx_desc *rx_mini_ring;
|
|
struct rx_desc *rx_return_ring;
|
|
|
|
#if ACENIC_DO_VLAN
|
|
struct vlan_group *vlgrp;
|
|
#endif
|
|
|
|
int tasklet_pending, jumbo;
|
|
struct tasklet_struct ace_tasklet;
|
|
|
|
struct event *evt_ring;
|
|
|
|
volatile u32 *evt_prd, *rx_ret_prd, *tx_csm;
|
|
|
|
dma_addr_t tx_ring_dma; /* 32/64 bit */
|
|
dma_addr_t rx_ring_base_dma;
|
|
dma_addr_t evt_ring_dma;
|
|
dma_addr_t evt_prd_dma, rx_ret_prd_dma, tx_csm_dma;
|
|
|
|
unsigned char *trace_buf;
|
|
struct pci_dev *pdev;
|
|
struct net_device *next;
|
|
volatile int fw_running;
|
|
int board_idx;
|
|
u16 pci_command;
|
|
u8 pci_latency;
|
|
const char *name;
|
|
#ifdef INDEX_DEBUG
|
|
spinlock_t debug_lock
|
|
__attribute__ ((aligned (SMP_CACHE_BYTES)));
|
|
u32 last_tx, last_std_rx, last_mini_rx;
|
|
#endif
|
|
int pci_using_dac;
|
|
};
|
|
|
|
|
|
#define TX_RESERVED MAX_SKB_FRAGS
|
|
|
|
static inline int tx_space (struct ace_private *ap, u32 csm, u32 prd)
|
|
{
|
|
return (csm - prd - 1) & (ACE_TX_RING_ENTRIES(ap) - 1);
|
|
}
|
|
|
|
#define tx_free(ap) tx_space((ap)->tx_ret_csm, (ap)->tx_prd, ap)
|
|
#define tx_ring_full(ap, csm, prd) (tx_space(ap, csm, prd) <= TX_RESERVED)
|
|
|
|
static inline void set_aceaddr(aceaddr *aa, dma_addr_t addr)
|
|
{
|
|
u64 baddr = (u64) addr;
|
|
aa->addrlo = baddr & 0xffffffff;
|
|
aa->addrhi = baddr >> 32;
|
|
wmb();
|
|
}
|
|
|
|
|
|
static inline void ace_set_txprd(struct ace_regs __iomem *regs,
|
|
struct ace_private *ap, u32 value)
|
|
{
|
|
#ifdef INDEX_DEBUG
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&ap->debug_lock, flags);
|
|
writel(value, ®s->TxPrd);
|
|
if (value == ap->last_tx)
|
|
printk(KERN_ERR "AceNIC RACE ALERT! writing identical value "
|
|
"to tx producer (%i)\n", value);
|
|
ap->last_tx = value;
|
|
spin_unlock_irqrestore(&ap->debug_lock, flags);
|
|
#else
|
|
writel(value, ®s->TxPrd);
|
|
#endif
|
|
wmb();
|
|
}
|
|
|
|
|
|
static inline void ace_mask_irq(struct net_device *dev)
|
|
{
|
|
struct ace_private *ap = netdev_priv(dev);
|
|
struct ace_regs __iomem *regs = ap->regs;
|
|
|
|
if (ACE_IS_TIGON_I(ap))
|
|
writel(1, ®s->MaskInt);
|
|
else
|
|
writel(readl(®s->HostCtrl) | MASK_INTS, ®s->HostCtrl);
|
|
|
|
ace_sync_irq(dev->irq);
|
|
}
|
|
|
|
|
|
static inline void ace_unmask_irq(struct net_device *dev)
|
|
{
|
|
struct ace_private *ap = netdev_priv(dev);
|
|
struct ace_regs __iomem *regs = ap->regs;
|
|
|
|
if (ACE_IS_TIGON_I(ap))
|
|
writel(0, ®s->MaskInt);
|
|
else
|
|
writel(readl(®s->HostCtrl) & ~MASK_INTS, ®s->HostCtrl);
|
|
}
|
|
|
|
|
|
/*
|
|
* Prototypes
|
|
*/
|
|
static int ace_init(struct net_device *dev);
|
|
static void ace_load_std_rx_ring(struct ace_private *ap, int nr_bufs);
|
|
static void ace_load_mini_rx_ring(struct ace_private *ap, int nr_bufs);
|
|
static void ace_load_jumbo_rx_ring(struct ace_private *ap, int nr_bufs);
|
|
static irqreturn_t ace_interrupt(int irq, void *dev_id);
|
|
static int ace_load_firmware(struct net_device *dev);
|
|
static int ace_open(struct net_device *dev);
|
|
static int ace_start_xmit(struct sk_buff *skb, struct net_device *dev);
|
|
static int ace_close(struct net_device *dev);
|
|
static void ace_tasklet(unsigned long dev);
|
|
static void ace_dump_trace(struct ace_private *ap);
|
|
static void ace_set_multicast_list(struct net_device *dev);
|
|
static int ace_change_mtu(struct net_device *dev, int new_mtu);
|
|
static int ace_set_mac_addr(struct net_device *dev, void *p);
|
|
static void ace_set_rxtx_parms(struct net_device *dev, int jumbo);
|
|
static int ace_allocate_descriptors(struct net_device *dev);
|
|
static void ace_free_descriptors(struct net_device *dev);
|
|
static void ace_init_cleanup(struct net_device *dev);
|
|
static struct net_device_stats *ace_get_stats(struct net_device *dev);
|
|
static int read_eeprom_byte(struct net_device *dev, unsigned long offset);
|
|
#if ACENIC_DO_VLAN
|
|
static void ace_vlan_rx_register(struct net_device *dev, struct vlan_group *grp);
|
|
#endif
|
|
|
|
#endif /* _ACENIC_H_ */
|