1
0
mirror of git://projects.qi-hardware.com/openwrt-xburst.git synced 2024-12-12 06:35:19 +02:00
openwrt-xburst/target/linux/storm/patches/002-gmac.patch
kaloz d57e4d97e7 rename and renumber storm patches
git-svn-id: svn://svn.openwrt.org/openwrt/trunk@11368 3c298f89-4303-0410-b956-a3cf2f4a3e73
2008-06-05 17:14:02 +00:00

18616 lines
615 KiB
Diff
Raw Blame History

--- /dev/null
+++ b/drivers/net/sl2312_emac.c
@@ -0,0 +1,4604 @@
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch-sl2312/irqs.h>
+#include <asm/arch/it8712.h>
+#include <asm/arch/sl2312.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/sysctl_storlink.h>
+
+#define BIG_ENDIAN 0
+
+#define GMAC_DEBUG 0
+
+#define GMAC_PHY_IF 2
+
+/* define PHY address */
+#define HPHY_ADDR 0x01
+#define GPHY_ADDR 0x02
+
+#define CONFIG_ADM_6999 1
+/* define chip information */
+#define DRV_NAME "SL2312"
+#define DRV_VERSION "0.1.1"
+#define SL2312_DRIVER_NAME DRV_NAME " Fast Ethernet driver " DRV_VERSION
+
+/* define TX/RX descriptor parameter */
+#define MAX_ETH_FRAME_SIZE 1920
+#define TX_BUF_SIZE MAX_ETH_FRAME_SIZE
+#define TX_DESC_NUM 128
+#define TX_BUF_TOT_LEN (TX_BUF_SIZE * TX_DESC_NUM)
+#define RX_BUF_SIZE MAX_ETH_FRAME_SIZE
+#define RX_DESC_NUM 256
+#define RX_BUF_TOT_LEN (RX_BUF_SIZE * RX_DESC_NUM)
+#define MAX_ISR_WORK 20
+
+unsigned int int_status = 0;
+
+/* define GMAC base address */
+#define GMAC_PHYSICAL_BASE_ADDR (SL2312_GMAC_BASE)
+#define GMAC_BASE_ADDR (IO_ADDRESS(GMAC_PHYSICAL_BASE_ADDR))
+#define GMAC_GLOBAL_BASE_ADDR (IO_ADDRESS(SL2312_GLOBAL_BASE))
+
+#define GMAC0_BASE (IO_ADDRESS(SL2312_GMAC0_BASE))
+#define GMAC1_BASE (IO_ADDRESS(SL2312_GMAC1_BASE))
+
+/* memory management utility */
+#define DMA_MALLOC(size,handle) pci_alloc_consistent(NULL,size,handle)
+#define DMA_MFREE(mem,size,handle) pci_free_consistent(NULL,size,mem,handle)
+
+//#define gmac_read_reg(offset) (readl(GMAC_BASE_ADDR + offset))
+//#define gmac_write_reg(offset,data,mask) writel( (gmac_read_reg(offset)&~mask) |(data&mask),(GMAC_BASE_ADDR+offset))
+
+/* define owner bit */
+#define CPU 0
+#define DMA 1
+
+#define ACTIVE 1
+#define NONACTIVE 0
+
+#define CONFIG_SL_NAPI
+
+#ifndef CONFIG_SL2312_MPAGE
+#define CONFIG_SL2312_MPAGE
+#endif
+
+#ifdef CONFIG_SL2312_MPAGE
+#include <linux/skbuff.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#endif
+
+#ifndef CONFIG_TXINT_DISABLE
+//#define CONFIG_TXINT_DISABLE
+#endif
+
+enum phy_state
+{
+ LINK_DOWN = 0,
+ LINK_UP = 1
+};
+
+
+/* transmit timeout value */
+#define TX_TIMEOUT (6*HZ)
+
+/***************************************/
+/* the offset address of GMAC register */
+/***************************************/
+enum GMAC_REGISTER {
+ GMAC_STA_ADD0 = 0x0000,
+ GMAC_STA_ADD1 = 0x0004,
+ GMAC_STA_ADD2 = 0x0008,
+ GMAC_RX_FLTR = 0x000c,
+ GMAC_MCAST_FIL0 = 0x0010,
+ GMAC_MCAST_FIL1 = 0x0014,
+ GMAC_CONFIG0 = 0x0018,
+ GMAC_CONFIG1 = 0x001c,
+ GMAC_CONFIG2 = 0x0020,
+ GMAC_BNCR = 0x0024,
+ GMAC_RBNR = 0x0028,
+ GMAC_STATUS = 0x002c,
+ GMAC_IN_DISCARDS= 0x0030,
+ GMAC_IN_ERRORS = 0x0034,
+ GMAC_IN_MCAST = 0x0038,
+ GMAC_IN_BCAST = 0x003c,
+ GMAC_IN_MAC1 = 0x0040,
+ GMAC_IN_MAC2 = 0x0044
+};
+
+/*******************************************/
+/* the offset address of GMAC DMA register */
+/*******************************************/
+enum GMAC_DMA_REGISTER {
+ GMAC_DMA_DEVICE_ID = 0xff00,
+ GMAC_DMA_STATUS = 0xff04,
+ GMAC_TXDMA_CTRL = 0xff08,
+ GMAC_TXDMA_FIRST_DESC = 0xff0c,
+ GMAC_TXDMA_CURR_DESC = 0xff10,
+ GMAC_RXDMA_CTRL = 0xff14,
+ GMAC_RXDMA_FIRST_DESC = 0xff18,
+ GMAC_RXDMA_CURR_DESC = 0xff1c,
+};
+
+/*******************************************/
+/* the register structure of GMAC */
+/*******************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0004
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int sta_add2_l16 : 16; /* station MAC address2 bits 15 to 0 */
+ unsigned int sta_add1_h16 : 16; /* station MAC address1 bits 47 to 32 */
+#else
+ unsigned int sta_add1_h16 : 16; /* station MAC address1 bits 47 to 32 */
+ unsigned int sta_add2_l16 : 16; /* station MAC address2 bits 15 to 0 */
+#endif
+ } bits;
+} GMAC_STA_ADD1_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_000c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 27;
+ unsigned int error : 1; /* enable receive of all error frames */
+ unsigned int promiscuous : 1; /* enable receive of all frames */
+ unsigned int broadcast : 1; /* enable receive of broadcast frames */
+ unsigned int multicast : 1; /* enable receive of multicast frames that pass multicast filter */
+ unsigned int unicast : 1; /* enable receive of unicast frames that are sent to STA address */
+#else
+ unsigned int unicast : 1; /* enable receive of unicast frames that are sent to STA address */
+ unsigned int multicast : 1; /* enable receive of multicast frames that pass multicast filter */
+ unsigned int broadcast : 1; /* enable receive of broadcast frames */
+ unsigned int promiscuous : 1; /* enable receive of all frames */
+ unsigned int error : 1; /* enable receive of all error frames */
+ unsigned int : 27;
+#endif
+ } bits;
+} GMAC_RX_FLTR_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0018
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 10;
+ unsigned int inv_rx_clk : 1; /* Inverse RX Clock */
+ unsigned int rising_latch : 1;
+ unsigned int rx_tag_remove : 1; /* Remove Rx VLAN tag */
+ unsigned int ipv6_tss_rx_en : 1; /* IPv6 TSS RX enable */
+ unsigned int ipv4_tss_rx_en : 1; /* IPv4 TSS RX enable */
+ unsigned int rgmii_en : 1; /* RGMII in-band status enable */
+ unsigned int tx_fc_en : 1; /* TX flow control enable */
+ unsigned int rx_fc_en : 1; /* RX flow control enable */
+ unsigned int sim_test : 1; /* speed up timers in simulation */
+ unsigned int dis_col : 1; /* disable 16 collisions abort function */
+ unsigned int dis_bkoff : 1; /* disable back-off function */
+ unsigned int max_len : 3; /* maximum receive frame length allowed */
+ unsigned int adj_ifg : 4; /* adjust IFG from 96+/-56 */
+ unsigned int : 1; /* reserved */
+ unsigned int loop_back : 1; /* transmit data loopback enable */
+ unsigned int dis_rx : 1; /* disable receive */
+ unsigned int dis_tx : 1; /* disable transmit */
+#else
+ unsigned int dis_tx : 1; /* disable transmit */
+ unsigned int dis_rx : 1; /* disable receive */
+ unsigned int loop_back : 1; /* transmit data loopback enable */
+ unsigned int : 1; /* reserved */
+ unsigned int adj_ifg : 4; /* adjust IFG from 96+/-56 */
+ unsigned int max_len : 3; /* maximum receive frame length allowed */
+ unsigned int dis_bkoff : 1; /* disable back-off function */
+ unsigned int dis_col : 1; /* disable 16 collisions abort function */
+ unsigned int sim_test : 1; /* speed up timers in simulation */
+ unsigned int rx_fc_en : 1; /* RX flow control enable */
+ unsigned int tx_fc_en : 1; /* TX flow control enable */
+ unsigned int rgmii_en : 1; /* RGMII in-band status enable */
+ unsigned int ipv4_tss_rx_en : 1; /* IPv4 TSS RX enable */
+ unsigned int ipv6_tss_rx_en : 1; /* IPv6 TSS RX enable */
+ unsigned int rx_tag_remove : 1; /* Remove Rx VLAN tag */
+ unsigned int rising_latch : 1;
+ unsigned int inv_rx_clk : 1; /* Inverse RX Clock */
+ unsigned int : 10;
+#endif
+ } bits;
+} GMAC_CONFIG0_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_001c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 28;
+ unsigned int buf_size : 4; /* per packet buffer size */
+#else
+ unsigned int buf_size : 4; /* per packet buffer size */
+ unsigned int : 28;
+#endif
+ } bits;
+} GMAC_CONFIG1_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0020
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+ unsigned int set_threshold : 16; /* flow control set threshold */
+#else
+ unsigned int set_threshold : 16; /* flow control set threshold */
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+#endif
+ } bits;
+} GMAC_CONFIG2_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0024
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 16;
+ unsigned int buf_num : 16; /* return buffer number from software */
+#else
+ unsigned int buf_num : 16; /* return buffer number from software */
+ unsigned int : 16;
+#endif
+ } bits;
+} GMAC_BNCR_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0028
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 16;
+ unsigned int buf_remain : 16; /* remaining buffer number */
+#else
+ unsigned int buf_remain : 16; /* remaining buffer number */
+ unsigned int : 16;
+#endif
+ } bits;
+} GMAC_RBNR_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_002c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 25;
+ unsigned int mii_rmii : 2; /* PHY interface type */
+ unsigned int phy_mode : 1; /* PHY interface mode in 10M-bps */
+ unsigned int duplex : 1; /* duplex mode */
+ unsigned int speed : 2; /* link speed(00->2.5M 01->25M 10->125M) */
+ unsigned int link : 1; /* link status */
+#else
+ unsigned int link : 1; /* link status */
+ unsigned int speed : 2; /* link speed(00->2.5M 01->25M 10->125M) */
+ unsigned int duplex : 1; /* duplex mode */
+ unsigned int phy_mode : 1; /* PHY interface mode in 10M-bps */
+ unsigned int mii_rmii : 2; /* PHY interface type */
+ unsigned int : 25;
+#endif
+ } bits;
+} GMAC_STATUS_T;
+
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_009
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 10;
+ unsigned int tx_fail : 1; /* Tx fail interrupt */
+ unsigned int cnt_full : 1; /* MIB counters half full interrupt */
+ unsigned int rx_pause_on : 1; /* received pause on frame interrupt */
+ unsigned int tx_pause_on : 1; /* transmit pause on frame interrupt */
+ unsigned int rx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int tx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt */
+ unsigned int tx_underrun : 1; /* GMAC Tx FIFO underrun interrupt */
+ unsigned int : 6;
+ unsigned int m_tx_fail : 1; /* Tx fail interrupt mask */
+ unsigned int m_cnt_full : 1; /* MIB counters half full interrupt mask */
+ unsigned int m_rx_pause_on : 1; /* received pause on frame interrupt mask */
+ unsigned int m_tx_pause_on : 1; /* transmit pause on frame interrupt mask */
+ unsigned int m_rx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_tx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt mask */
+ unsigned int m_tx_underrun : 1; /* GMAC Tx FIFO underrun interrupt mask */
+#else
+ unsigned int m_tx_underrun : 1; /* GMAC Tx FIFO underrun interrupt mask */
+ unsigned int m_rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt mask */
+ unsigned int m_tx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_rx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_tx_pause_on : 1; /* transmit pause on frame interrupt mask */
+ unsigned int m_rx_pause_on : 1; /* received pause on frame interrupt mask */
+ unsigned int m_cnt_full : 1; /* MIB counters half full interrupt mask */
+ unsigned int m_tx_fail : 1; /* Tx fail interrupt mask */
+ unsigned int : 6;
+ unsigned int tx_underrun : 1; /* GMAC Tx FIFO underrun interrupt */
+ unsigned int rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt */
+ unsigned int tx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int rx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int tx_pause_on : 1; /* transmit pause on frame interrupt */
+ unsigned int rx_pause_on : 1; /* received pause on frame interrupt */
+ unsigned int cnt_full : 1; /* MIB counters half full interrupt */
+ unsigned int tx_fail : 1; /* Tx fail interrupt */
+ unsigned int : 10;
+#endif
+ } bits;
+} GMAC_INT_MASK_T;
+
+
+/*******************************************/
+/* the register structure of GMAC DMA */
+/*******************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff00
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 7; /* reserved */
+ unsigned int s_ahb_err : 1; /* Slave AHB bus error */
+ unsigned int tx_err_code : 4; /* TxDMA error code */
+ unsigned int rx_err_code : 4; /* RxDMA error code */
+ unsigned int device_id : 12;
+ unsigned int revision_id : 4;
+#else
+ unsigned int revision_id : 4;
+ unsigned int device_id : 12;
+ unsigned int rx_err_code : 4; /* RxDMA error code */
+ unsigned int tx_err_code : 4; /* TxDMA error code */
+ unsigned int s_ahb_err : 1; /* Slave AHB bus error */
+ unsigned int : 7; /* reserved */
+#endif
+ } bits;
+} GMAC_DMA_DEVICE_ID_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff04
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int ts_finish : 1; /* finished tx interrupt */
+ unsigned int ts_derr : 1; /* AHB Bus Error while tx */
+ unsigned int ts_perr : 1; /* Tx Descriptor protocol error */
+ unsigned int ts_eodi : 1; /* TxDMA end of descriptor interrupt */
+ unsigned int ts_eofi : 1; /* TxDMA end of frame interrupt */
+ unsigned int rs_finish : 1; /* finished rx interrupt */
+ unsigned int rs_derr : 1; /* AHB Bus Error while rx */
+ unsigned int rs_perr : 1; /* Rx Descriptor protocol error */
+ unsigned int rs_eodi : 1; /* RxDMA end of descriptor interrupt */
+ unsigned int rs_eofi : 1; /* RxDMA end of frame interrupt */
+ unsigned int : 1; /* Tx fail interrupt */
+ unsigned int cnt_full : 1; /* MIB counters half full interrupt */
+ unsigned int rx_pause_on : 1; /* received pause on frame interrupt */
+ unsigned int tx_pause_on : 1; /* transmit pause on frame interrupt */
+ unsigned int rx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int tx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt */
+ unsigned int link_change : 1; /* GMAC link changed Interrupt for RGMII mode */
+ unsigned int : 1;
+ unsigned int : 1;
+ unsigned int : 3;
+ unsigned int loop_back : 1; /* loopback TxDMA to RxDMA */
+ unsigned int : 1; /* Tx fail interrupt mask */
+ unsigned int m_cnt_full : 1; /* MIB counters half full interrupt mask */
+ unsigned int m_rx_pause_on : 1; /* received pause on frame interrupt mask */
+ unsigned int m_tx_pause_on : 1; /* transmit pause on frame interrupt mask */
+ unsigned int m_rx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_tx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt mask */
+ unsigned int m_link_change : 1; /* GMAC link changed Interrupt mask for RGMII mode */
+#else
+ unsigned int m_link_change : 1; /* GMAC link changed Interrupt mask for RGMII mode */
+ unsigned int m_rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt mask */
+ unsigned int m_tx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_rx_pause_off : 1; /* received pause off frame interrupt mask */
+ unsigned int m_tx_pause_on : 1; /* transmit pause on frame interrupt mask */
+ unsigned int m_rx_pause_on : 1; /* received pause on frame interrupt mask */
+ unsigned int m_cnt_full : 1; /* MIB counters half full interrupt mask */
+ unsigned int : 1; /* Tx fail interrupt mask */
+ unsigned int loop_back : 1; /* loopback TxDMA to RxDMA */
+ unsigned int : 3;
+ unsigned int : 1;
+ unsigned int : 1;
+ unsigned int link_change : 1; /* GMAC link changed Interrupt for RGMII mode */
+ unsigned int rx_overrun : 1; /* GMAC Rx FIFO overrun interrupt */
+ unsigned int tx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int rx_pause_off : 1; /* received pause off frame interrupt */
+ unsigned int tx_pause_on : 1; /* transmit pause on frame interrupt */
+ unsigned int rx_pause_on : 1; /* received pause on frame interrupt */
+ unsigned int cnt_full : 1; /* MIB counters half full interrupt */
+ unsigned int : 1; /* Tx fail interrupt */
+ unsigned int rs_eofi : 1; /* RxDMA end of frame interrupt */
+ unsigned int rs_eodi : 1; /* RxDMA end of descriptor interrupt */
+ unsigned int rs_perr : 1; /* Rx Descriptor protocol error */
+ unsigned int rs_derr : 1; /* AHB Bus Error while rx */
+ unsigned int rs_finish : 1; /* finished rx interrupt */
+ unsigned int ts_eofi : 1; /* TxDMA end of frame interrupt */
+ unsigned int ts_eodi : 1; /* TxDMA end of descriptor interrupt */
+ unsigned int ts_perr : 1; /* Tx Descriptor protocol error */
+ unsigned int ts_derr : 1; /* AHB Bus Error while tx */
+ unsigned int ts_finish : 1; /* finished tx interrupt */
+#endif
+ } bits;
+} GMAC_DMA_STATUS_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff08
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int td_start : 1; /* Start DMA transfer */
+ unsigned int td_continue : 1; /* Continue DMA operation */
+ unsigned int td_chain_mode : 1; /* Descriptor Chain Mode;1-Descriptor Chain mode, 0-Direct DMA mode*/
+ unsigned int : 1;
+ unsigned int td_prot : 4; /* TxDMA protection control */
+ unsigned int td_burst_size : 2; /* TxDMA max burst size for every AHB request */
+ unsigned int td_bus : 2; /* peripheral bus width;0x->8 bits,10->16 bits,11->32 bits */
+ unsigned int td_endian : 1; /* AHB Endian. 0-little endian; 1-big endian */
+ unsigned int td_finish_en : 1; /* DMA Finish Event Interrupt Enable;1-enable;0-mask */
+ unsigned int td_fail_en : 1; /* DMA Fail Interrupt Enable;1-enable;0-mask */
+ unsigned int td_perr_en : 1; /* Protocol Failure Interrupt Enable;1-enable;0-mask */
+ unsigned int td_eod_en : 1; /* End of Descriptor interrupt Enable;1-enable;0-mask */
+ unsigned int td_eof_en : 1; /* End of frame interrupt Enable;1-enable;0-mask */
+ unsigned int : 14;
+#else
+ unsigned int : 14;
+ unsigned int td_eof_en : 1; /* End of frame interrupt Enable;1-enable;0-mask */
+ unsigned int td_eod_en : 1; /* End of Descriptor interrupt Enable;1-enable;0-mask */
+ unsigned int td_perr_en : 1; /* Protocol Failure Interrupt Enable;1-enable;0-mask */
+ unsigned int td_fail_en : 1; /* DMA Fail Interrupt Enable;1-enable;0-mask */
+ unsigned int td_finish_en : 1; /* DMA Finish Event Interrupt Enable;1-enable;0-mask */
+ unsigned int td_endian : 1; /* AHB Endian. 0-little endian; 1-big endian */
+ unsigned int td_bus : 2; /* peripheral bus width;0x->8 bits,10->16 bits,11->32 bits */
+ unsigned int td_burst_size : 2; /* TxDMA max burst size for every AHB request */
+ unsigned int td_prot : 4; /* TxDMA protection control */
+ unsigned int : 1;
+ unsigned int td_chain_mode : 1; /* Descriptor Chain Mode;1-Descriptor Chain mode, 0-Direct DMA mode*/
+ unsigned int td_continue : 1; /* Continue DMA operation */
+ unsigned int td_start : 1; /* Start DMA transfer */
+#endif
+ } bits;
+} GMAC_TXDMA_CTRL_T;
+
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff0c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int td_first_des_ptr : 28;/* first descriptor address */
+ unsigned int td_busy : 1;/* 1-TxDMA busy; 0-TxDMA idle */
+ unsigned int : 3;
+#else
+ unsigned int : 3;
+ unsigned int td_busy : 1;/* 1-TxDMA busy; 0-TxDMA idle */
+ unsigned int td_first_des_ptr : 28;/* first descriptor address */
+#endif
+ } bits;
+} GMAC_TXDMA_FIRST_DESC_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff10
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int ndar : 28; /* next descriptor address */
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int : 1;
+ unsigned int sof_eof : 2;
+#else
+ unsigned int sof_eof : 2;
+ unsigned int : 1;
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int ndar : 28; /* next descriptor address */
+#endif
+ } bits;
+} GMAC_TXDMA_CURR_DESC_T;
+
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff14
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rd_start : 1; /* Start DMA transfer */
+ unsigned int rd_continue : 1; /* Continue DMA operation */
+ unsigned int rd_chain_mode : 1; /* Descriptor Chain Mode;1-Descriptor Chain mode, 0-Direct DMA mode*/
+ unsigned int : 1;
+ unsigned int rd_prot : 4; /* DMA protection control */
+ unsigned int rd_burst_size : 2; /* DMA max burst size for every AHB request */
+ unsigned int rd_bus : 2; /* peripheral bus width;0x->8 bits,10->16 bits,11->32 bits */
+ unsigned int rd_endian : 1; /* AHB Endian. 0-little endian; 1-big endian */
+ unsigned int rd_finish_en : 1; /* DMA Finish Event Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_fail_en : 1; /* DMA Fail Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_perr_en : 1; /* Protocol Failure Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_eod_en : 1; /* End of Descriptor interrupt Enable;1-enable;0-mask */
+ unsigned int rd_eof_en : 1; /* End of frame interrupt Enable;1-enable;0-mask */
+ unsigned int : 14;
+#else
+ unsigned int : 14;
+ unsigned int rd_eof_en : 1; /* End of frame interrupt Enable;1-enable;0-mask */
+ unsigned int rd_eod_en : 1; /* End of Descriptor interrupt Enable;1-enable;0-mask */
+ unsigned int rd_perr_en : 1; /* Protocol Failure Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_fail_en : 1; /* DMA Fail Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_finish_en : 1; /* DMA Finish Event Interrupt Enable;1-enable;0-mask */
+ unsigned int rd_endian : 1; /* AHB Endian. 0-little endian; 1-big endian */
+ unsigned int rd_bus : 2; /* peripheral bus width;0x->8 bits,10->16 bits,11->32 bits */
+ unsigned int rd_burst_size : 2; /* DMA max burst size for every AHB request */
+ unsigned int rd_prot : 4; /* DMA protection control */
+ unsigned int : 1;
+ unsigned int rd_chain_mode : 1; /* Descriptor Chain Mode;1-Descriptor Chain mode, 0-Direct DMA mode*/
+ unsigned int rd_continue : 1; /* Continue DMA operation */
+ unsigned int rd_start : 1; /* Start DMA transfer */
+#endif
+ } bits;
+} GMAC_RXDMA_CTRL_T;
+
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff18
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rd_first_des_ptr : 28;/* first descriptor address */
+ unsigned int rd_busy : 1;/* 1-RxDMA busy; 0-RxDMA idle */
+ unsigned int : 3;
+#else
+ unsigned int : 3;
+ unsigned int rd_busy : 1;/* 1-RxDMA busy; 0-RxDMA idle */
+ unsigned int rd_first_des_ptr : 28;/* first descriptor address */
+#endif
+ } bits;
+} GMAC_RXDMA_FIRST_DESC_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit2_ff1c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int ndar : 28; /* next descriptor address */
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int : 1;
+ unsigned int sof_eof : 2;
+#else
+ unsigned int sof_eof : 2;
+ unsigned int : 1;
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int ndar : 28; /* next descriptor address */
+#endif
+ } bits;
+} GMAC_RXDMA_CURR_DESC_T;
+
+
+/********************************************/
+/* Descriptor Format */
+/********************************************/
+
+typedef struct descriptor_t
+{
+ union frame_control_t
+ {
+ unsigned int bits32;
+ struct bits_0000
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int csum_state : 3; /* checksum error status */
+ unsigned int vlan_tag : 1; /* 802.1q vlan tag packet */
+ unsigned int frame_state: 3; /* reference Rx Status1 */
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+#else
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int frame_state: 3; /* reference Rx Status1 */
+ unsigned int vlan_tag : 1; /* 802.1q vlan tag packet */
+ unsigned int csum_state : 3; /* checksum error status */
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+#endif
+ } bits_rx;
+
+ struct bits_0001
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int : 6;
+ unsigned int success_tx : 1; /* successful transmitted */
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+#else
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int success_tx : 1; /* successful transmitted */
+ unsigned int : 6;
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+#endif
+ } bits_tx_in;
+
+ struct bits_0002
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int : 2;
+ unsigned int udp_csum_en: 1; /* TSS UDP checksum enable */
+ unsigned int tcp_csum_en: 1; /* TSS TCP checksum enable */
+ unsigned int ipv6_tx_en : 1; /* TSS IPv6 TX enable */
+ unsigned int ip_csum_en : 1; /* TSS IPv4 IP Header checksum enable */
+ unsigned int vlan_enable: 1; /* VLAN TIC insertion enable */
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+#else
+ unsigned int buffer_size:16; /* transfer buffer size associated with current description*/
+ unsigned int desc_count : 6; /* number of descriptors used for the current frame */
+ unsigned int vlan_enable: 1; /* VLAN TIC insertion enable */
+ unsigned int ip_csum_en : 1; /* TSS IPv4 IP Header checksum enable */
+ unsigned int ipv6_tx_en : 1; /* TSS IPv6 TX enable */
+ unsigned int tcp_csum_en: 1; /* TSS TCP checksum enable */
+ unsigned int udp_csum_en: 1; /* TSS UDP checksum enable */
+ unsigned int : 2;
+ unsigned int perr : 1; /* protocol error during processing this descriptor */
+ unsigned int derr : 1; /* data error during processing this descriptor */
+ unsigned int own : 1; /* owner bit. 0-CPU, 1-DMA */
+#endif
+ } bits_tx_out;
+
+ } frame_ctrl;
+
+ union flag_status_t
+ {
+ unsigned int bits32;
+ struct bits_0004
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int priority : 3; /* user priority extracted from receiving frame*/
+ unsigned int cfi : 1; /* cfi extracted from receiving frame*/
+ unsigned int vlan_id :12; /* VLAN ID extracted from receiving frame */
+ unsigned int frame_count:16; /* received frame byte count,include CRC,not include VLAN TIC */
+#else
+ unsigned int frame_count:16; /* received frame byte count,include CRC,not include VLAN TIC */
+ unsigned int vlan_id :12; /* VLAN ID extracted from receiving frame */
+ unsigned int cfi : 1; /* cfi extracted from receiving frame*/
+ unsigned int priority : 3; /* user priority extracted from receiving frame*/
+#endif
+ } bits_rx_status;
+
+ struct bits_0005
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int priority : 3; /* user priority to transmit*/
+ unsigned int cfi : 1; /* cfi to transmit*/
+ unsigned int vlan_id :12; /* VLAN ID to transmit */
+ unsigned int frame_count:16; /* total tx frame byte count */
+#else
+ unsigned int frame_count:16; /* total tx frame byte count */
+ unsigned int vlan_id :12; /* VLAN ID to transmit */
+ unsigned int cfi : 1; /* cfi to transmit*/
+ unsigned int priority : 3; /* user priority to transmit*/
+#endif
+ } bits_tx_flag;
+ } flag_status;
+
+ unsigned int buf_adr; /* data buffer address */
+
+ union next_desc_t
+ {
+ unsigned int next_descriptor;
+ struct bits_000c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int ndar :28; /* next descriptor address */
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int : 1;
+ unsigned int sof_eof : 2; /* 00-the linking descriptor 01-the last descriptor of a frame*/
+ /* 10-the first descriptor of a frame 11-only one descriptor for a frame*/
+#else
+ unsigned int sof_eof : 2; /* 00-the linking descriptor 01-the last descriptor of a frame*/
+ /* 10-the first descriptor of a frame 11-only one descriptor for a frame*/
+ unsigned int : 1;
+ unsigned int eofie : 1; /* end of frame interrupt enable */
+ unsigned int ndar :28; /* next descriptor address */
+#endif
+ } bits;
+ } next_desc;
+} GMAC_DESCRIPTOR_T;
+
+typedef struct gmac_conf {
+ struct net_device *dev;
+ int portmap;
+ int vid;
+ int flag; /* 1: active 0: non-active */
+} sys_gmac_conf;
+
+struct gmac_private {
+ unsigned char *tx_bufs; /* Tx bounce buffer region. */
+ unsigned char *rx_bufs;
+ GMAC_DESCRIPTOR_T *tx_desc; /* point to virtual TX descriptor address*/
+ GMAC_DESCRIPTOR_T *rx_desc; /* point to virtual RX descriptor address*/
+ GMAC_DESCRIPTOR_T *tx_cur_desc; /* point to current TX descriptor */
+ GMAC_DESCRIPTOR_T *rx_cur_desc; /* point to current RX descriptor */
+ GMAC_DESCRIPTOR_T *tx_finished_desc;
+ GMAC_DESCRIPTOR_T *rx_finished_desc;
+ unsigned long cur_tx;
+ unsigned int cur_rx; /* Index into the Rx buffer of next Rx pkt. */
+ unsigned int tx_flag;
+ unsigned long dirty_tx;
+ unsigned char *tx_buf[TX_DESC_NUM]; /* Tx bounce buffers */
+ dma_addr_t tx_desc_dma; /* physical TX descriptor address */
+ dma_addr_t rx_desc_dma; /* physical RX descriptor address */
+ dma_addr_t tx_bufs_dma; /* physical TX descriptor address */
+ dma_addr_t rx_bufs_dma; /* physical RX descriptor address */
+ struct net_device_stats stats;
+ pid_t thr_pid;
+ wait_queue_head_t thr_wait;
+ struct completion thr_exited;
+ spinlock_t lock;
+ int time_to_die;
+ unsigned int tx_desc_hdr[GMAC_PHY_IF]; /* the descriptor which sw can fill */
+ unsigned int tx_desc_tail[GMAC_PHY_IF]; /* the descriptor which is not cleaned yet */
+};
+
+
+struct reg_ioctl_data {
+ unsigned int reg_addr; /* the register address */
+ unsigned int val_in; /* data write to the register */
+ unsigned int val_out; /* data read from the register */
+};
+
+#ifdef CONFIG_SL2312_MPAGE
+typedef struct tx_data_t {
+ int freeable; // 1 when it's skb. it can be freed in tx interrupt handler
+ struct sk_buff* skb; // skb
+ int desc_in_use; // 1 when the desc is in use. 0 when desc is available.
+ long end_seq; // to find out packets are in seq.
+ // so this value is the seq of next packet.
+} tx_data;
+#endif
+
+/*************************************************************
+ * Global Variable
+ *************************************************************/
+struct semaphore sem_gmac; /* semaphore for share pins issue */
+
+/*************************************************************
+ * Static Global Variable
+ *************************************************************/
+// static unsigned int MAC_BASE_ADDR = GMAC0_BASE;
+static unsigned int gmac_base_addr[GMAC_PHY_IF] = {GMAC0_BASE,GMAC1_BASE};
+static unsigned int gmac_irq[GMAC_PHY_IF] = {IRQ_GMAC0,IRQ_GMAC1};
+static struct net_device *gmac_dev[GMAC_PHY_IF];
+
+static unsigned int FLAG_SWITCH=0; /* if 1-->switch chip presented. if 0-->switch chip unpresented */
+static unsigned int flow_control_enable[GMAC_PHY_IF] = {1,1};
+static unsigned int pre_phy_status[GMAC_PHY_IF] = {LINK_DOWN,LINK_DOWN};
+static unsigned int tx_desc_virtual_base[GMAC_PHY_IF];
+static unsigned int rx_desc_virtual_base[GMAC_PHY_IF];
+static unsigned int full_duplex = 1;
+static unsigned int speed = 1;
+#ifdef CONFIG_SL2312_MPAGE
+static tx_data tx_skb[GMAC_PHY_IF][TX_DESC_NUM];
+#else
+static struct sk_buff *tx_skb[GMAC_PHY_IF][TX_DESC_NUM];
+#endif
+static struct sk_buff *rx_skb[GMAC_PHY_IF][RX_DESC_NUM];
+static unsigned int tx_desc_start_adr[GMAC_PHY_IF];
+static unsigned int rx_desc_start_adr[GMAC_PHY_IF];
+static unsigned char eth0_mac[6]= {0x00,0x50,0xc2,0x2b,0xd3,0x25};
+static unsigned char eth1_mac[6]= {0x00,0x50,0xc2,0x2b,0xdf,0xfe};
+static unsigned int next_tick = 3 * HZ;
+
+static unsigned int phy_addr[GMAC_PHY_IF] = {0x01,0x02}; /* define PHY address */
+
+DECLARE_WAIT_QUEUE_HEAD(gmac_queue);
+//static wait_queue_t wait;
+
+struct gmac_conf VLAN_conf[] = {
+#ifdef CONFIG_ADM_6999
+ { (struct net_device *)0,0x7F,1 },
+ { (struct net_device *)0,0x80,2 }
+#endif
+#ifdef CONFIG_ADM_6996
+ { (struct net_device *)0,0x0F,1 },
+ { (struct net_device *)0,0x10,2 }
+#endif
+};
+
+#define NUM_VLAN_IF (sizeof(VLAN_conf)/sizeof(struct gmac_conf))
+
+
+/************************************************/
+/* GMAC function declare */
+/************************************************/
+
+unsigned int mii_read(unsigned char phyad,unsigned char regad);
+void mii_write(unsigned char phyad,unsigned char regad,unsigned int value);
+static void gmac_set_phy_status(struct net_device *dev);
+static void gmac_get_phy_status(struct net_device *dev);
+static int gmac_phy_thread (void *data);
+static int gmac_set_mac_address(struct net_device *dev, void *addr);
+static void gmac_tx_timeout(struct net_device *dev);
+static void gmac_tx_packet_complete(struct net_device *dev);
+static int gmac_start_xmit(struct sk_buff *skb, struct net_device *dev);
+static void gmac_set_rx_mode(struct net_device *dev);
+static void gmac_rx_packet(struct net_device *dev);
+static int gmac_open (struct net_device *dev);
+static int gmac_netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
+
+static unsigned int gmac_get_dev_index(struct net_device *dev);
+static unsigned int gmac_select_interface(struct net_device *dev);
+
+#ifdef CONFIG_SL2312_MPAGE
+int printk_all(int dev_index, struct gmac_private* tp);
+#endif
+
+/****************************************/
+/* SPI Function Declare */
+/****************************************/
+void SPI_write(unsigned char addr,unsigned int value);
+unsigned int SPI_read(unsigned char table,unsigned char addr);
+void SPI_write_bit(char bit_EEDO);
+unsigned int SPI_read_bit(void);
+void SPI_default(void);
+void SPI_reset(unsigned char rstype,unsigned char port_cnt);
+void SPI_pre_st(void);
+void SPI_CS_enable(unsigned char enable);
+void SPI_Set_VLAN(unsigned char LAN,unsigned int port_mask);
+void SPI_Set_tag(unsigned int port,unsigned tag);
+void SPI_Set_PVID(unsigned int PVID,unsigned int port_mask);
+unsigned int SPI_Get_PVID(unsigned int port);
+void SPI_mac_lock(unsigned int port, unsigned char lock);
+void SPI_get_port_state(unsigned int port);
+void SPI_port_enable(unsigned int port,unsigned char enable);
+unsigned int SPI_get_identifier(void);
+void SPI_get_status(unsigned int port);
+
+/****************************************/
+/* VLAN Function Declare */
+/****************************************/
+int getVLANfromdev (struct net_device *dev );
+struct net_device * getdevfromVLAN( int VID);
+
+
+
+/************************************************/
+/* function body */
+/************************************************/
+#if 0
+void hw_memcpy(void *to,const void *from,unsigned long n)
+{
+ writel(from,SL2312_DRAM_CTRL_BASE+0x20); /* set source address */
+ writel(to,SL2312_DRAM_CTRL_BASE+0x24); /* set destination address */
+ writel(n,SL2312_DRAM_CTRL_BASE+0x28); /* set byte count */
+ writel(0x00000001,SL2312_DRAM_CTRL_BASE+0x2c);
+ while (readl(SL2312_DRAM_CTRL_BASE+0x2c));
+}
+#endif
+
+static unsigned int gmac_read_reg(unsigned int addr)
+{
+ unsigned int reg_val;
+// unsigned int flags;
+// spinlock_t lock;
+
+// spin_lock_irqsave(&lock, flags);
+ reg_val = readl(addr); // Gary Chen
+// spin_unlock_irqrestore(&lock, flags);
+ return (reg_val);
+}
+
+static void gmac_write_reg(unsigned int addr,unsigned int data,unsigned int bit_mask)
+{
+ unsigned int reg_val;
+ //unsigned int *addr;
+// unsigned int flags;
+// spinlock_t lock;
+
+// spin_lock_irqsave(&lock, flags);
+ reg_val = ( gmac_read_reg(addr) & (~bit_mask) ) | (data & bit_mask);
+ writel(reg_val,addr);
+// spin_unlock_irqrestore(&lock, flags);
+ return;
+}
+
+
+static void gmac_sw_reset(struct net_device *dev)
+{
+ unsigned int index;
+ unsigned int reg_val;
+
+ index = gmac_get_dev_index(dev);
+ if (index==0)
+ reg_val = readl(GMAC_GLOBAL_BASE_ADDR+0x0c) | 0x00000020; /* GMAC0 S/W reset */
+ else
+ reg_val = readl(GMAC_GLOBAL_BASE_ADDR+0x0c) | 0x00000040; /* GMAC1 S/W reset */
+
+ writel(reg_val,GMAC_GLOBAL_BASE_ADDR+0x0c);
+ return;
+}
+
+static void gmac_get_mac_address(void)
+{
+#ifdef CONFIG_MTD
+ extern int get_vlaninfo(vlaninfo* vlan);
+ static vlaninfo vlan[2];
+
+ if (get_vlaninfo(&vlan[0]))
+ {
+ memcpy(eth0_mac,vlan[0].mac,6);
+ VLAN_conf[0].vid = vlan[0].vlanid;
+ VLAN_conf[0].portmap = vlan[0].vlanmap;
+ memcpy(eth1_mac,vlan[1].mac,6);
+ VLAN_conf[1].vid = vlan[1].vlanid;
+ VLAN_conf[1].portmap = vlan[1].vlanmap;
+ }
+#else
+ unsigned int reg_val;
+
+ reg_val = readl(IO_ADDRESS(SL2312_SECURITY_BASE)+0xac);
+ eth0_mac[4] = (reg_val & 0xff00) >> 8;
+ eth0_mac[5] = reg_val & 0x00ff;
+ reg_val = readl(IO_ADDRESS(SL2312_SECURITY_BASE)+0xac);
+ eth1_mac[4] = (reg_val & 0xff00) >> 8;
+ eth1_mac[5] = reg_val & 0x00ff;
+#endif
+ return;
+}
+
+static unsigned int gmac_get_dev_index(struct net_device *dev)
+{
+ unsigned int i;
+
+ /* get device index number */
+ for (i=0;i<GMAC_PHY_IF;i++)
+ {
+ if (gmac_dev[i]==dev)
+ {
+ return(i);
+ }
+ }
+ return (0xff);
+}
+
+static unsigned int gmac_select_interface(struct net_device *dev)
+{
+ unsigned int index;
+
+ index = gmac_get_dev_index(dev);
+ // MAC_BASE_ADDR = gmac_base_addr[index]; // Gary Chen
+ return (index);
+}
+
+
+static void gmac_dump_register(struct net_device *dev)
+{
+#if 0
+ unsigned int i,val,index;
+
+ index = gmac_select_interface(dev);
+
+ printk("========== GMAC%d ==========\n",index);
+ for (i=0;i<=0x7c;i=i+4)
+ {
+ val = gmac_read_reg(gmac_base_addr[index] + i);
+ printk("offset = %08x value = %08x\n",i,val);
+ }
+ for (i=0xff00;i<=0xff7c;i=i+4)
+ {
+ val = gmac_read_reg(gmac_base_addr[index] + i);
+ printk("offset = %08x value = %08x\n",i,val);
+ }
+#endif
+}
+
+static int gmac_init_chip(struct net_device *dev)
+{
+ GMAC_RBNR_T rbnr_val,rbnr_mask;
+ GMAC_CONFIG2_T config2_val;
+ GMAC_CONFIG0_T config0,config0_mask;
+ GMAC_CONFIG1_T config1;
+ struct sockaddr sock;
+ unsigned int status;
+ unsigned int phy_mode;
+ unsigned int index;
+
+ index = gmac_get_dev_index(dev);
+
+ /* set GMAC RMII mode */
+ if (index==0)
+ phy_mode = 0; /* 0->MII 1->GMII 2->RGMII(10/100) 3->RGMII(1000) */
+ else
+ phy_mode = 2; /* 0->MII 1->GMII 2->RGMII(10/100) 3->RGMII(1000) */
+
+ /* set PHY operation mode */
+ status = (phy_mode<<5) | 0x11 | (full_duplex<<3) | (speed<<1);
+ gmac_write_reg(gmac_base_addr[index] + GMAC_STATUS,status ,0x0000007f);
+
+ /* set station MAC address1 and address2 */
+ if (index==0)
+ memcpy(&sock.sa_data[0],&eth0_mac[0],6);
+ else
+ memcpy(&sock.sa_data[0],&eth1_mac[0],6);
+ gmac_set_mac_address(dev,(void *)&sock);
+
+ /* set RX_FLTR register to receive all multicast packet */
+ gmac_write_reg(gmac_base_addr[index] + GMAC_RX_FLTR,0x0000001F,0x0000001f);
+ //gmac_write_reg(gmac_base_addr[index] + GMAC_RX_FLTR,0x00000007,0x0000001f);
+
+ /* set per packet buffer size */
+ config1.bits32 = 0;
+ config1.bits.buf_size = 11; /* buffer size = 2048-byte */
+ gmac_write_reg(gmac_base_addr[index] + GMAC_CONFIG1,config1.bits32,0x0000000f);
+
+ /* set flow control threshold */
+ config2_val.bits32 = 0;
+ config2_val.bits.set_threshold = RX_DESC_NUM/4;
+ config2_val.bits.rel_threshold = RX_DESC_NUM*3/4;
+ gmac_write_reg(gmac_base_addr[index] + GMAC_CONFIG2,config2_val.bits32,0xffffffff);
+
+ /* init remaining buffer number register */
+ rbnr_val.bits32 = 0;
+ rbnr_val.bits.buf_remain = RX_DESC_NUM;
+ rbnr_mask.bits32 = 0;
+ rbnr_mask.bits.buf_remain = 0xffff;
+ gmac_write_reg(gmac_base_addr[index] + GMAC_RBNR,rbnr_val.bits32,rbnr_mask.bits32);
+
+ /* disable TX/RX and disable internal loop back */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.max_len = 2;
+ if (flow_control_enable[index]==1)
+ {
+ config0.bits.tx_fc_en = 1; /* enable tx flow control */
+ config0.bits.rx_fc_en = 1; /* enable rx flow control */
+ printk("Enable MAC Flow Control...\n");
+ }
+ else
+ {
+ config0.bits.tx_fc_en = 0; /* disable tx flow control */
+ config0.bits.rx_fc_en = 0; /* disable rx flow control */
+ printk("Disable MAC Flow Control...\n");
+ }
+ config0.bits.dis_rx = 1; /* disable rx */
+ config0.bits.dis_tx = 1; /* disable tx */
+ config0.bits.loop_back = 0; /* enable/disable GMAC loopback */
+ config0.bits.inv_rx_clk = 0;
+ config0.bits.rising_latch = 1;
+ config0.bits.ipv4_tss_rx_en = 1; /* enable H/W to check ip checksum */
+ config0.bits.ipv6_tss_rx_en = 1; /* enable H/W to check ip checksum */
+
+ config0_mask.bits.max_len = 7;
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ config0_mask.bits.loop_back = 1;
+ config0_mask.bits.inv_rx_clk = 1;
+ config0_mask.bits.rising_latch = 1;
+ config0_mask.bits.ipv4_tss_rx_en = 1;
+ config0_mask.bits.ipv6_tss_rx_en = 1;
+ gmac_write_reg(gmac_base_addr[index] + GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+
+ return (0);
+}
+
+static void gmac_enable_tx_rx(struct net_device *dev)
+{
+ GMAC_CONFIG0_T config0,config0_mask;
+ int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 0; /* enable rx */
+ config0.bits.dis_tx = 0; /* enable tx */
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+}
+
+static void gmac_disable_tx_rx(struct net_device *dev)
+{
+ GMAC_CONFIG0_T config0,config0_mask;
+ int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 1; /* disable rx */
+ config0.bits.dis_tx = 1; /* disable tx */
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+}
+
+#ifdef CONFIG_SL_NAPI
+static int gmac_rx_poll_ga(struct net_device *dev, int *budget)
+{
+ struct gmac_private *tp = dev->priv;
+ struct sk_buff *skb;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ GMAC_RXDMA_FIRST_DESC_T rxdma_busy;
+ GMAC_DESCRIPTOR_T *rx_desc;
+ unsigned int pkt_size;
+ unsigned int desc_count;
+ unsigned int vid;
+// unsigned int priority;
+ unsigned int own;
+ unsigned int good_frame = 0;
+ unsigned int index;
+ unsigned int dev_index;
+ int work = 0;
+ int work_done = 0;
+ int quota = min(dev->quota, *budget);
+
+ dev_index = gmac_select_interface(dev);
+
+ for (;;)
+ {
+ own = tp->rx_cur_desc->frame_ctrl.bits32 >> 31;
+ if (own == CPU) /* check owner bit */
+ {
+ rx_desc = tp->rx_cur_desc;
+#if (GMAC_DEBUG==1)
+ /* check error interrupt */
+ if ( (rx_desc->frame_ctrl.bits_rx.derr==1)||(rx_desc->frame_ctrl.bits_rx.perr==1) )
+ {
+ printk("%s::Rx Descriptor Processing Error !!!\n",__func__);
+ }
+#endif
+ /* get frame information from the first descriptor of the frame */
+ pkt_size = rx_desc->flag_status.bits_rx_status.frame_count - 4; /*total byte count in a frame*/
+#if (GMAC_DEBUG==1)
+ priority = rx_desc->flag_status.bits_rx_status.priority; /* 802.1p priority */
+#endif
+ vid = rx_desc->flag_status.bits_rx_status.vlan_id; /* 802.1q vlan id */
+ if (vid == 0)
+ {
+ vid = 1; /* default vlan */
+ }
+ desc_count = rx_desc->frame_ctrl.bits_rx.desc_count; /* get descriptor count per frame */
+
+ if (rx_desc->frame_ctrl.bits_rx.frame_state == 0x000) /* good frame */
+ {
+ tp->stats.rx_bytes += pkt_size;
+ tp->stats.rx_packets++;
+ good_frame = 1;
+ }
+ else
+ {
+ tp->stats.rx_errors++;
+ good_frame = 0;
+ printk("RX status: 0x%x\n",rx_desc->frame_ctrl.bits_rx.frame_state);
+ }
+ }
+ else
+ {
+ work_done = 1;
+ break; /* Rx process is completed */
+ }
+
+ if (good_frame == 1)
+ {
+ /* get rx skb buffer index */
+ index = ((unsigned int)tp->rx_cur_desc - rx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ if (rx_skb[dev_index][index])
+ {
+ skb_reserve (rx_skb[dev_index][index], 2); /* 16 byte align the IP fields. */
+ rx_skb[dev_index][index]->dev = dev;
+ rx_skb[dev_index][index]->ip_summed = CHECKSUM_UNNECESSARY;
+ skb_put(rx_skb[dev_index][index],pkt_size);
+ rx_skb[dev_index][index]->protocol = eth_type_trans(rx_skb[dev_index][index],dev); /* set skb protocol */
+ netif_rx(rx_skb[dev_index][index]); /* socket rx */
+ dev->last_rx = jiffies;
+
+ /* allocate rx skb buffer */
+ if ( (skb = dev_alloc_skb(RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__);
+ }
+ rx_skb[dev_index][index] = skb;
+ tp->rx_cur_desc->buf_adr = (unsigned int)__pa(skb->data) | 0x02; /* insert two bytes in the beginning of rx data */
+ }
+ else
+ {
+ printk("%s::rx skb index error !\n",__func__);
+ }
+ }
+
+ tp->rx_cur_desc->frame_ctrl.bits_rx.own = DMA; /* release rx descriptor to DMA */
+ /* point to next rx descriptor */
+ tp->rx_cur_desc = (GMAC_DESCRIPTOR_T *)((tp->rx_cur_desc->next_desc.next_descriptor & 0xfffffff0)+rx_desc_virtual_base[dev_index]);
+
+ /* release buffer to Remaining Buffer Number Register */
+ if (flow_control_enable[dev_index] ==1)
+ {
+// gmac_write_reg(gmac_base_addr[dev_index] + GMAC_BNCR,desc_count,0x0000ffff);
+ writel(desc_count,(unsigned int *)(gmac_base_addr[dev_index] + GMAC_BNCR));
+ }
+
+ if (work++ >= quota )
+ {
+ break;
+ }
+ }
+
+ /* if RX DMA process is stoped , restart it */
+ rxdma_busy.bits.rd_first_des_ptr = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_FIRST_DESC);
+ if (rxdma_busy.bits.rd_busy == 0)
+ {
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ }
+
+ dev->quota -= work;
+ *budget -= work;
+ if (work_done==1)
+ {
+ /* Receive descriptor is empty now */
+ netif_rx_complete(dev);
+ /* enable receive interrupt */
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,0x0007c000,0x0007c000); /* enable rx interrupt */
+ return 0;
+ }
+ else
+ {
+ return -1;
+ }
+}
+
+static int gmac_rx_poll_gb(struct net_device *dev, int *budget)
+{
+ struct gmac_private *tp = dev->priv;
+ struct sk_buff *skb;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ GMAC_RXDMA_FIRST_DESC_T rxdma_busy;
+ GMAC_DESCRIPTOR_T *rx_desc;
+ unsigned int pkt_size;
+ unsigned int desc_count;
+ unsigned int vid;
+// unsigned int priority;
+ unsigned int own;
+ unsigned int good_frame = 0;
+ unsigned int index;
+ unsigned int dev_index;
+ int work = 0;
+ int work_done = 0;
+ int quota = min(dev->quota, *budget);
+
+ dev_index = gmac_select_interface(dev);
+
+ for (;;)
+ {
+ own = tp->rx_cur_desc->frame_ctrl.bits32 >> 31;
+ if (own == CPU) /* check owner bit */
+ {
+ rx_desc = tp->rx_cur_desc;
+#if (GMAC_DEBUG==1)
+ /* check error interrupt */
+ if ( (rx_desc->frame_ctrl.bits_rx.derr==1)||(rx_desc->frame_ctrl.bits_rx.perr==1) )
+ {
+ printk("%s::Rx Descriptor Processing Error !!!\n",__func__);
+ }
+#endif
+ /* get frame information from the first descriptor of the frame */
+ pkt_size = rx_desc->flag_status.bits_rx_status.frame_count - 4; /*total byte count in a frame*/
+#if (GMAC_DEBUG==1)
+ priority = rx_desc->flag_status.bits_rx_status.priority; /* 802.1p priority */
+#endif
+ vid = rx_desc->flag_status.bits_rx_status.vlan_id; /* 802.1q vlan id */
+ if (vid == 0)
+ {
+ vid = 1; /* default vlan */
+ }
+ desc_count = rx_desc->frame_ctrl.bits_rx.desc_count; /* get descriptor count per frame */
+
+ if (rx_desc->frame_ctrl.bits_rx.frame_state == 0x000) /* good frame */
+ {
+ tp->stats.rx_bytes += pkt_size;
+ tp->stats.rx_packets++;
+ good_frame = 1;
+ }
+ else
+ {
+ tp->stats.rx_errors++;
+ good_frame = 0;
+ printk("RX status: 0x%x\n",rx_desc->frame_ctrl.bits_rx.frame_state);
+ }
+ }
+ else
+ {
+ work_done = 1;
+ break; /* Rx process is completed */
+ }
+
+ if (good_frame == 1)
+ {
+ /* get rx skb buffer index */
+ index = ((unsigned int)tp->rx_cur_desc - rx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ if (rx_skb[dev_index][index])
+ {
+ skb_reserve (rx_skb[dev_index][index], 2); /* 16 byte align the IP fields. */
+ rx_skb[dev_index][index]->dev = dev;
+ rx_skb[dev_index][index]->ip_summed = CHECKSUM_UNNECESSARY;
+ skb_put(rx_skb[dev_index][index],pkt_size);
+ rx_skb[dev_index][index]->protocol = eth_type_trans(rx_skb[dev_index][index],dev); /* set skb protocol */
+ netif_rx(rx_skb[dev_index][index]); /* socket rx */
+ dev->last_rx = jiffies;
+
+ /* allocate rx skb buffer */
+ if ( (skb = dev_alloc_skb(RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__);
+ }
+ rx_skb[dev_index][index] = skb;
+ tp->rx_cur_desc->buf_adr = (unsigned int)__pa(skb->data) | 0x02; /* insert two bytes in the beginning of rx data */
+ }
+ else
+ {
+ printk("%s::rx skb index error !\n",__func__);
+ }
+ }
+
+ tp->rx_cur_desc->frame_ctrl.bits_rx.own = DMA; /* release rx descriptor to DMA */
+ /* point to next rx descriptor */
+ tp->rx_cur_desc = (GMAC_DESCRIPTOR_T *)((tp->rx_cur_desc->next_desc.next_descriptor & 0xfffffff0)+rx_desc_virtual_base[dev_index]);
+
+ /* release buffer to Remaining Buffer Number Register */
+ if (flow_control_enable[dev_index] ==1)
+ {
+// gmac_write_reg(gmac_base_addr[dev_index] + GMAC_BNCR,desc_count,0x0000ffff);
+ writel(desc_count,(unsigned int *)(gmac_base_addr[dev_index] + GMAC_BNCR));
+ }
+
+ if (work++ >= quota )
+ {
+ break;
+ }
+ }
+
+ /* if RX DMA process is stoped , restart it */
+ rxdma_busy.bits.rd_first_des_ptr = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_FIRST_DESC);
+ if (rxdma_busy.bits.rd_busy == 0)
+ {
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ }
+
+ dev->quota -= work;
+ *budget -= work;
+ if (work_done==1)
+ {
+ /* Receive descriptor is empty now */
+ netif_rx_complete(dev);
+ /* enable receive interrupt */
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,0x0007c000,0x0007c000); /* enable rx interrupt */
+ return 0;
+ }
+ else
+ {
+ return -1;
+ }
+}
+
+#endif
+
+static void gmac_rx_packet(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ struct sk_buff *skb;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ GMAC_RXDMA_FIRST_DESC_T rxdma_busy;
+ GMAC_DESCRIPTOR_T *rx_desc;
+ unsigned int pkt_size;
+ unsigned int desc_count;
+ unsigned int vid;
+// unsigned int priority;
+ unsigned int own;
+ unsigned int good_frame = 0;
+ unsigned int i,index;
+ unsigned int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ for (i=0;i<256;i++)
+ {
+ own = tp->rx_cur_desc->frame_ctrl.bits32 >> 31;
+ if (own == CPU) /* check owner bit */
+ {
+ rx_desc = tp->rx_cur_desc;
+#if (GMAC_DEBUG==1)
+ /* check error interrupt */
+ if ( (rx_desc->frame_ctrl.bits_rx.derr==1)||(rx_desc->frame_ctrl.bits_rx.perr==1) )
+ {
+ printk("%s::Rx Descriptor Processing Error !!!\n",__func__);
+ }
+#endif
+ /* get frame information from the first descriptor of the frame */
+ pkt_size = rx_desc->flag_status.bits_rx_status.frame_count - 4; /*total byte count in a frame*/
+#if (GMAC_DEBUG==1)
+ priority = rx_desc->flag_status.bits_rx_status.priority; /* 802.1p priority */
+#endif
+ vid = rx_desc->flag_status.bits_rx_status.vlan_id; /* 802.1q vlan id */
+ if (vid == 0)
+ {
+ vid = 1; /* default vlan */
+ }
+ desc_count = rx_desc->frame_ctrl.bits_rx.desc_count; /* get descriptor count per frame */
+
+ if (rx_desc->frame_ctrl.bits_rx.frame_state == 0x000) /* good frame */
+ {
+ tp->stats.rx_bytes += pkt_size;
+ tp->stats.rx_packets++;
+ good_frame = 1;
+ }
+ else
+ {
+ tp->stats.rx_errors++;
+ good_frame = 0;
+ printk("RX status: 0x%x\n",rx_desc->frame_ctrl.bits_rx.frame_state);
+ }
+ }
+ else
+ {
+ break; /* Rx process is completed */
+ }
+
+ if (good_frame == 1)
+ {
+ /* get rx skb buffer index */
+ index = ((unsigned int)tp->rx_cur_desc - rx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ if (rx_skb[dev_index][index])
+ {
+ skb_reserve (rx_skb[dev_index][index], 2); /* 16 byte align the IP fields. */
+ rx_skb[dev_index][index]->dev = dev;
+ rx_skb[dev_index][index]->ip_summed = CHECKSUM_UNNECESSARY;
+ skb_put(rx_skb[dev_index][index],pkt_size);
+ rx_skb[dev_index][index]->protocol = eth_type_trans(rx_skb[dev_index][index],dev); /* set skb protocol */
+ netif_rx(rx_skb[dev_index][index]); /* socket rx */
+ dev->last_rx = jiffies;
+
+ /* allocate rx skb buffer */
+ if ( (skb = dev_alloc_skb(RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__);
+ }
+ rx_skb[dev_index][index] = skb;
+ tp->rx_cur_desc->buf_adr = (unsigned int)__pa(skb->data) | 0x02; /* insert two bytes in the beginning of rx data */
+ }
+ else
+ {
+ printk("%s::rx skb index error !\n",__func__);
+ }
+ }
+
+ tp->rx_cur_desc->frame_ctrl.bits_rx.own = DMA; /* release rx descriptor to DMA */
+ /* point to next rx descriptor */
+ tp->rx_cur_desc = (GMAC_DESCRIPTOR_T *)((tp->rx_cur_desc->next_desc.next_descriptor & 0xfffffff0)+rx_desc_virtual_base[dev_index]);
+
+ /* release buffer to Remaining Buffer Number Register */
+ if (flow_control_enable[dev_index] ==1)
+ {
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_BNCR,desc_count,0x0000ffff);
+ }
+ }
+
+ /* if RX DMA process is stoped , restart it */
+ rxdma_busy.bits.rd_first_des_ptr = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_FIRST_DESC);
+ if (rxdma_busy.bits.rd_busy == 0)
+ {
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ }
+}
+
+#ifdef CONFIG_SL2312_MPAGE
+static inline void free_tx_buf(int dev_index, int desc_index)
+{
+ if (tx_skb[dev_index][desc_index].freeable &&
+ tx_skb[dev_index][desc_index].skb) {
+ struct sk_buff* skb = tx_skb[dev_index][desc_index].skb;
+ //printk("free_skb %x, len %d\n", skb, skb->len);
+#ifdef CONFIG_TXINT_DISABLE
+ dev_kfree_skb(skb);
+#else
+ dev_kfree_skb_irq(skb);
+#endif
+ tx_skb[dev_index][desc_index].skb = 0;
+ }
+}
+
+#ifdef CONFIG_TXINT_DISABLE
+static void gmac_tx_packet_complete(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ GMAC_DESCRIPTOR_T *tx_hw_complete_desc, *next_desc;
+ unsigned int desc_cnt=0;
+ unsigned int i,index,dev_index;
+ unsigned int tx_current_descriptor = 0;
+ // int own_dma = 0;
+
+ dev_index = gmac_select_interface(dev);
+
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ if (tx_skb[dev_index][index].desc_in_use && tp->tx_finished_desc->frame_ctrl.bits_tx_in.own == CPU) {
+ free_tx_buf(dev_index, index);
+ tx_skb[dev_index][index].desc_in_use = 0;
+ }
+ next_desc = (GMAC_DESCRIPTOR_T*)((tp->tx_finished_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+
+ for (;;) {
+ tx_hw_complete_desc = (GMAC_DESCRIPTOR_T *)((gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC) & 0xfffffff0)+ tx_desc_virtual_base[dev_index]);
+ if (next_desc == tx_hw_complete_desc)
+ break;
+ if (next_desc->frame_ctrl.bits_tx_in.own == CPU) {
+ if (next_desc->frame_ctrl.bits_tx_in.success_tx == 1) {
+ tp->stats.tx_bytes += next_desc->flag_status.bits_tx_flag.frame_count;
+ tp->stats.tx_packets ++;
+ } else {
+ tp->stats.tx_errors++;
+ }
+ desc_cnt = next_desc->frame_ctrl.bits_tx_in.desc_count;
+ for (i=1; i<desc_cnt; i++) {
+ /* get tx skb buffer index */
+ index = ((unsigned int)next_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ next_desc->frame_ctrl.bits_tx_in.own = CPU;
+ free_tx_buf(dev_index, index);
+ tx_skb[dev_index][index].desc_in_use = 0;
+ tp->tx_desc_tail[dev_index] = (tp->tx_desc_tail[dev_index] +1) & (TX_DESC_NUM-1);
+ /* release Tx descriptor to CPU */
+ next_desc = (GMAC_DESCRIPTOR_T *)((next_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ }
+ /* get tx skb buffer index */
+ index = ((unsigned int)next_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ /* free skb buffer */
+ next_desc->frame_ctrl.bits_tx_in.own = CPU;
+ free_tx_buf(dev_index, index);
+ tx_skb[dev_index][index].desc_in_use = 0;
+ tp->tx_desc_tail[dev_index] = (tp->tx_desc_tail[dev_index] +1) & (TX_DESC_NUM-1);
+ tp->tx_finished_desc = next_desc;
+// printk("finish tx_desc index %d\n", index);
+ next_desc = (GMAC_DESCRIPTOR_T *)((next_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ }
+ else
+ break;
+ }
+ if (netif_queue_stopped(dev))
+ {
+ netif_wake_queue(dev);
+ }
+
+}
+#else
+static void gmac_tx_packet_complete(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ GMAC_DESCRIPTOR_T *tx_hw_complete_desc;
+ unsigned int desc_cnt=0;
+ unsigned int i,index,dev_index;
+ unsigned int tx_current_descriptor = 0;
+ // int own_dma = 0;
+
+ dev_index = gmac_select_interface(dev);
+
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+
+ /* check tx status and accumulate tx statistics */
+ for (;;)
+ {
+
+ for (i=0;i<1000;i++)
+ {
+ tx_current_descriptor = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC);
+ if ( ((tx_current_descriptor & 0x00000003)==0x00000003) || /* only one descriptor */
+ ((tx_current_descriptor & 0x00000003)==0x00000001) ) /* the last descriptor */
+ {
+ break;
+ }
+ udelay(1);
+ }
+ if (i==1000)
+ {
+// gmac_dump_register(dev);
+// printk("%s: tx current descriptor = %x \n",__func__,tx_current_descriptor);
+// printk_all(dev_index, tp);
+ continue;
+ }
+
+ /* get tx H/W completed descriptor virtual address */
+ tx_hw_complete_desc = (GMAC_DESCRIPTOR_T *)((tx_current_descriptor & 0xfffffff0)+ tx_desc_virtual_base[dev_index]);
+// tx_hw_complete_desc = (GMAC_DESCRIPTOR_T *)((gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC) & 0xfffffff0)+ tx_desc_virtual_base[dev_index]);
+ if (tp->tx_finished_desc == tx_hw_complete_desc ) // ||
+ //tx_skb[dev_index][index].desc_in_use ) /* complete tx processing */
+ {
+ break;
+ }
+
+ for (;;)
+ {
+ if (tp->tx_finished_desc->frame_ctrl.bits_tx_in.own == CPU)
+ {
+ #if (GMAC_DEBUG==1)
+ if ( (tp->tx_finished_desc->frame_ctrl.bits_tx_in.derr) ||
+ (tp->tx_finished_desc->frame_ctrl.bits_tx_in.perr) )
+ {
+ printk("%s::Descriptor Processing Error !!!\n",__func__);
+ }
+ #endif
+ if (tp->tx_finished_desc->frame_ctrl.bits_tx_in.success_tx == 1)
+ {
+ tp->stats.tx_bytes += tp->tx_finished_desc->flag_status.bits_tx_flag.frame_count;
+ tp->stats.tx_packets ++;
+ }
+ else
+ {
+ tp->stats.tx_errors++;
+ }
+ desc_cnt = tp->tx_finished_desc->frame_ctrl.bits_tx_in.desc_count;
+ for (i=1; i<desc_cnt; i++) /* multi-descriptor in one packet */
+ {
+ /* get tx skb buffer index */
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ tp->tx_finished_desc->frame_ctrl.bits_tx_in.own = CPU;
+ free_tx_buf(dev_index, index);
+ tx_skb[dev_index][index].desc_in_use = 0;
+ /* release Tx descriptor to CPU */
+ tp->tx_finished_desc = (GMAC_DESCRIPTOR_T *)((tp->tx_finished_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ }
+ /* get tx skb buffer index */
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ /* free skb buffer */
+ tp->tx_finished_desc->frame_ctrl.bits_tx_in.own = CPU;
+ free_tx_buf(dev_index, index);
+ tx_skb[dev_index][index].desc_in_use = 0;
+ tp->tx_finished_desc = (GMAC_DESCRIPTOR_T *)((tp->tx_finished_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+
+ if (tp->tx_finished_desc == tx_hw_complete_desc )
+ {
+ break;
+ }
+ }
+ else
+ {
+ break;
+ }
+ }
+ }
+
+ if (netif_queue_stopped(dev))
+ {
+ netif_wake_queue(dev);
+ }
+
+}
+#endif
+#else
+
+static void gmac_tx_packet_complete(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ GMAC_DESCRIPTOR_T *tx_hw_complete_desc;
+ unsigned int desc_cnt=0;
+ unsigned int i,index,dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* get tx H/W completed descriptor virtual address */
+ tx_hw_complete_desc = (GMAC_DESCRIPTOR_T *)((gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC) & 0xfffffff0)+ tx_desc_virtual_base[dev_index]);
+ /* check tx status and accumulate tx statistics */
+ for (;;)
+ {
+ if (tp->tx_finished_desc == tx_hw_complete_desc) /* complete tx processing */
+ {
+ break;
+ }
+ if (tp->tx_finished_desc->frame_ctrl.bits_tx_in.own == CPU)
+ {
+#if (GMAC_DEBUG==1)
+ if ( (tp->tx_finished_desc->frame_ctrl.bits_tx_in.derr) ||
+ (tp->tx_finished_desc->frame_ctrl.bits_tx_in.perr) )
+ {
+ printk("%s::Descriptor Processing Error !!!\n",__func__);
+ }
+#endif
+ if (tp->tx_finished_desc->frame_ctrl.bits_tx_in.success_tx == 1)
+ {
+ tp->stats.tx_bytes += tp->tx_finished_desc->flag_status.bits_tx_flag.frame_count;
+ tp->stats.tx_packets ++;
+ }
+ else
+ {
+ tp->stats.tx_errors++;
+ }
+ desc_cnt = tp->tx_finished_desc->frame_ctrl.bits_tx_in.desc_count;
+ for (i=1; i<desc_cnt; i++) /* multi-descriptor in one packet */
+ {
+ /* get tx skb buffer index */
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ /* free skb buffer */
+ if (tx_skb[dev_index][index])
+ {
+ dev_kfree_skb_irq(tx_skb[dev_index][index]);
+ }
+ /* release Tx descriptor to CPU */
+ tp->tx_finished_desc = (GMAC_DESCRIPTOR_T *)((tp->tx_finished_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ tp->tx_finished_desc->frame_ctrl.bits_tx_in.own = CPU;
+ }
+ /* get tx skb buffer index */
+ index = ((unsigned int)tp->tx_finished_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ /* free skb buffer */
+ if (tx_skb[dev_index][index])
+ {
+ dev_kfree_skb_irq(tx_skb[dev_index][index]);
+ }
+ tp->tx_finished_desc = (GMAC_DESCRIPTOR_T *)((tp->tx_finished_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ }
+ }
+
+ if (netif_queue_stopped(dev))
+ {
+ netif_wake_queue(dev);
+ }
+
+}
+
+
+#endif
+
+#if 0
+static void gmac_weird_interrupt(struct net_device *dev)
+{
+ gmac_dump_register(dev);
+}
+#endif
+
+/* The interrupt handler does all of the Rx thread work and cleans up
+ after the Tx thread. */
+static irqreturn_t gmac_interrupt (int irq, void *dev_instance, struct pt_regs *regs)
+{
+ struct net_device *dev = (struct net_device *)dev_instance;
+ GMAC_RXDMA_FIRST_DESC_T rxdma_busy;
+// GMAC_TXDMA_FIRST_DESC_T txdma_busy;
+// GMAC_TXDMA_CTRL_T txdma_ctrl,txdma_ctrl_mask;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ GMAC_DMA_STATUS_T status;
+ unsigned int i,dev_index;
+ int handled = 0;
+
+ dev_index = gmac_select_interface(dev);
+
+ handled = 1;
+
+#ifdef CONFIG_SL_NAPI
+ disable_irq(gmac_irq[dev_index]); /* disable GMAC interrupt */
+
+ status.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_DMA_STATUS); /* read DMA status */
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_DMA_STATUS,status.bits32,status.bits32); /* clear DMA status */
+
+ if (status.bits.rx_overrun == 1)
+ {
+ printk("%s::RX Overrun !!!%d\n",__func__,gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RBNR));
+ gmac_dump_register(dev);
+ /* if RX DMA process is stoped , restart it */
+ rxdma_busy.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_FIRST_DESC) ;
+ if (rxdma_busy.bits.rd_busy == 0)
+ {
+ /* restart Rx DMA process */
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ }
+ }
+
+ /* process rx packet */
+ if (netif_running(dev) && ((status.bits.rs_eofi==1)||(status.bits.rs_finish==1)))
+ {
+ if (likely(netif_rx_schedule_prep(dev)))
+ {
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,0,0x0007c000); /* disable rx interrupt */
+ __netif_rx_schedule(dev);
+ }
+ }
+#ifndef CONFIG_TXINT_DISABLE
+ /* process tx packet */
+ if (netif_running(dev) && ((status.bits.ts_eofi==1)||(status.bits.ts_finish==1)))
+ {
+ gmac_tx_packet_complete(dev);
+ }
+#endif
+
+ enable_irq(gmac_irq[dev_index]); /* enable GMAC interrupt */
+ return IRQ_RETVAL(handled);
+#endif
+
+ /* disable GMAC interrupt */
+ disable_irq(gmac_irq[dev_index]);
+ for (i=0;i<MAX_ISR_WORK;i++)
+ {
+ /* read DMA status */
+ status.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_DMA_STATUS);
+int_status = status.bits32;
+ /* clear DMA status */
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_DMA_STATUS,status.bits32,status.bits32);
+
+ if ((status.bits32 & 0xffffc000)==0)
+ {
+ break;
+ }
+
+ if (status.bits.rx_overrun == 1)
+ {
+ printk("%s::RX Overrun !!!%d\n",__func__,gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RBNR));
+ gmac_dump_register(dev);
+ /* if RX DMA process is stoped , restart it */
+ rxdma_busy.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_FIRST_DESC) ;
+ if (rxdma_busy.bits.rd_busy == 0)
+ {
+ /* restart Rx DMA process */
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ }
+ }
+
+ /* receive rx interrupt */
+ if (netif_running(dev) && ((status.bits.rs_eofi==1)||(status.bits.rs_finish==1)))
+ {
+ gmac_rx_packet(dev);
+// gmac_tx_packet_complete(dev);
+ }
+
+ /* receive tx interrupt */
+ // if (netif_running(dev) && (status.bits.ts_finish==1))
+#ifndef CONFIG_TXINT_DISABLE
+ if (netif_running(dev) && ((status.bits.ts_eofi==1)||
+ (status.bits.ts_finish==1)))
+ {
+ gmac_tx_packet_complete(dev);
+ }
+#endif
+ /* check uncommon events */
+/* if ((status.bits32 & 0x632fc000)!=0)
+ {
+ printk("%s::DMA Status = %08x \n",__func__,status.bits32);
+ gmac_weird_interrupt(dev);
+ }
+*/
+ }
+
+ /* enable GMAC interrupt */
+ enable_irq(gmac_irq[dev_index]);
+ //printk("gmac_interrupt complete!\n\n");
+ return IRQ_RETVAL(handled);
+}
+
+static void gmac_hw_start(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ GMAC_TXDMA_CURR_DESC_T tx_desc;
+ GMAC_RXDMA_CURR_DESC_T rx_desc;
+ GMAC_TXDMA_CTRL_T txdma_ctrl,txdma_ctrl_mask;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ GMAC_DMA_STATUS_T dma_status,dma_status_mask;
+ int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* program TxDMA Current Descriptor Address register for first descriptor */
+ tx_desc.bits32 = (unsigned int)(tp->tx_desc_dma);
+ tx_desc.bits.eofie = 1;
+ tx_desc.bits.sof_eof = 0x03;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC,tx_desc.bits32,0xffffffff);
+ gmac_write_reg(gmac_base_addr[dev_index] + 0xff2c,tx_desc.bits32,0xffffffff); /* tx next descriptor address */
+
+ /* program RxDMA Current Descriptor Address register for first descriptor */
+ rx_desc.bits32 = (unsigned int)(tp->rx_desc_dma);
+ rx_desc.bits.eofie = 1;
+ rx_desc.bits.sof_eof = 0x03;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CURR_DESC,rx_desc.bits32,0xffffffff);
+ gmac_write_reg(gmac_base_addr[dev_index] + 0xff3c,rx_desc.bits32,0xffffffff); /* rx next descriptor address */
+
+ /* enable GMAC interrupt & disable loopback */
+ dma_status.bits32 = 0;
+ dma_status.bits.loop_back = 0; /* disable DMA loop-back mode */
+// dma_status.bits.m_tx_fail = 1;
+ dma_status.bits.m_cnt_full = 1;
+ dma_status.bits.m_rx_pause_on = 1;
+ dma_status.bits.m_tx_pause_on = 1;
+ dma_status.bits.m_rx_pause_off = 1;
+ dma_status.bits.m_tx_pause_off = 1;
+ dma_status.bits.m_rx_overrun = 1;
+ dma_status.bits.m_link_change = 1;
+ dma_status_mask.bits32 = 0;
+ dma_status_mask.bits.loop_back = 1;
+// dma_status_mask.bits.m_tx_fail = 1;
+ dma_status_mask.bits.m_cnt_full = 1;
+ dma_status_mask.bits.m_rx_pause_on = 1;
+ dma_status_mask.bits.m_tx_pause_on = 1;
+ dma_status_mask.bits.m_rx_pause_off = 1;
+ dma_status_mask.bits.m_tx_pause_off = 1;
+ dma_status_mask.bits.m_rx_overrun = 1;
+ dma_status_mask.bits.m_link_change = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_DMA_STATUS,dma_status.bits32,dma_status_mask.bits32);
+
+ /* program tx dma control register */
+ txdma_ctrl.bits32 = 0;
+ txdma_ctrl.bits.td_start = 0; /* start TX DMA transfer */
+ txdma_ctrl.bits.td_continue = 0; /* continue Tx DMA operation */
+ txdma_ctrl.bits.td_chain_mode = 1; /* chain mode */
+ txdma_ctrl.bits.td_prot = 0;
+ txdma_ctrl.bits.td_burst_size = 2; /* DMA burst size for every AHB request */
+ txdma_ctrl.bits.td_bus = 2; /* peripheral bus width */
+ txdma_ctrl.bits.td_endian = 0; /* little endian */
+#ifdef CONFIG_TXINT_DISABLE
+ txdma_ctrl.bits.td_finish_en = 0; /* DMA finish event interrupt disable */
+#else
+ txdma_ctrl.bits.td_finish_en = 1; /* DMA finish event interrupt enable */
+#endif
+ txdma_ctrl.bits.td_fail_en = 1; /* DMA fail interrupt enable */
+ txdma_ctrl.bits.td_perr_en = 1; /* protocol failure interrupt enable */
+ txdma_ctrl.bits.td_eod_en = 0; /* disable Tx End of Descriptor Interrupt */
+ //txdma_ctrl.bits.td_eod_en = 0; /* disable Tx End of Descriptor Interrupt */
+#ifdef CONFIG_TXINT_DISABLE
+ txdma_ctrl.bits.td_eof_en = 0; /* end of frame interrupt disable */
+#else
+ txdma_ctrl.bits.td_eof_en = 1; /* end of frame interrupt enable */
+#endif
+ txdma_ctrl_mask.bits32 = 0;
+ txdma_ctrl_mask.bits.td_start = 1;
+ txdma_ctrl_mask.bits.td_continue = 1;
+ txdma_ctrl_mask.bits.td_chain_mode = 1;
+ txdma_ctrl_mask.bits.td_prot = 15;
+ txdma_ctrl_mask.bits.td_burst_size = 3;
+ txdma_ctrl_mask.bits.td_bus = 3;
+ txdma_ctrl_mask.bits.td_endian = 1;
+ txdma_ctrl_mask.bits.td_finish_en = 1;
+ txdma_ctrl_mask.bits.td_fail_en = 1;
+ txdma_ctrl_mask.bits.td_perr_en = 1;
+ txdma_ctrl_mask.bits.td_eod_en = 1;
+ //txdma_ctrl_mask.bits.td_eod_en = 1;
+ txdma_ctrl_mask.bits.td_eof_en = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CTRL,txdma_ctrl.bits32,txdma_ctrl_mask.bits32);
+
+ /* program rx dma control register */
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 1; /* start RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 1; /* continue RX DMA operation */
+ rxdma_ctrl.bits.rd_chain_mode = 1; /* chain mode */
+ rxdma_ctrl.bits.rd_prot = 0;
+ rxdma_ctrl.bits.rd_burst_size = 2; /* DMA burst size for every AHB request */
+ rxdma_ctrl.bits.rd_bus = 2; /* peripheral bus width */
+ rxdma_ctrl.bits.rd_endian = 0; /* little endian */
+ rxdma_ctrl.bits.rd_finish_en = 1; /* DMA finish event interrupt enable */
+ rxdma_ctrl.bits.rd_fail_en = 1; /* DMA fail interrupt enable */
+ rxdma_ctrl.bits.rd_perr_en = 1; /* protocol failure interrupt enable */
+ rxdma_ctrl.bits.rd_eod_en = 0; /* disable Rx End of Descriptor Interrupt */
+ rxdma_ctrl.bits.rd_eof_en = 1; /* end of frame interrupt enable */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ rxdma_ctrl_mask.bits.rd_chain_mode = 1;
+ rxdma_ctrl_mask.bits.rd_prot = 15;
+ rxdma_ctrl_mask.bits.rd_burst_size = 3;
+ rxdma_ctrl_mask.bits.rd_bus = 3;
+ rxdma_ctrl_mask.bits.rd_endian = 1;
+ rxdma_ctrl_mask.bits.rd_finish_en = 1;
+ rxdma_ctrl_mask.bits.rd_fail_en = 1;
+ rxdma_ctrl_mask.bits.rd_perr_en = 1;
+ rxdma_ctrl_mask.bits.rd_eod_en = 1;
+ rxdma_ctrl_mask.bits.rd_eof_en = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+ return;
+}
+
+static void gmac_hw_stop(struct net_device *dev)
+{
+ GMAC_TXDMA_CTRL_T txdma_ctrl,txdma_ctrl_mask;
+ GMAC_RXDMA_CTRL_T rxdma_ctrl,rxdma_ctrl_mask;
+ int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* program tx dma control register */
+ txdma_ctrl.bits32 = 0;
+ txdma_ctrl.bits.td_start = 0;
+ txdma_ctrl.bits.td_continue = 0;
+ txdma_ctrl_mask.bits32 = 0;
+ txdma_ctrl_mask.bits.td_start = 1;
+ txdma_ctrl_mask.bits.td_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CTRL,txdma_ctrl.bits32,txdma_ctrl_mask.bits32);
+ /* program rx dma control register */
+ rxdma_ctrl.bits32 = 0;
+ rxdma_ctrl.bits.rd_start = 0; /* stop RX DMA transfer */
+ rxdma_ctrl.bits.rd_continue = 0; /* stop continue RX DMA operation */
+ rxdma_ctrl_mask.bits32 = 0;
+ rxdma_ctrl_mask.bits.rd_start = 1;
+ rxdma_ctrl_mask.bits.rd_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RXDMA_CTRL,rxdma_ctrl.bits32,rxdma_ctrl_mask.bits32);
+}
+
+static int gmac_init_desc_buf(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ struct sk_buff *skb;
+ dma_addr_t tx_first_desc_dma=0;
+ dma_addr_t rx_first_desc_dma=0;
+ dma_addr_t rx_first_buf_dma=0;
+ unsigned int i,index;
+
+ printk("Descriptor buffer init......\n");
+
+ /* get device index number */
+ index = gmac_get_dev_index(dev);
+#ifdef CONFIG_SL2312_MPAGE
+ for (i=0; i<TX_DESC_NUM; i++) {
+ tx_skb[index][i].freeable = 0;
+ tx_skb[index][i].skb = 0;
+ tx_skb[index][i].desc_in_use = 0;
+ tx_skb[index][i].end_seq = 0;
+ }
+#else
+ for (i=0;i<TX_DESC_NUM;i++)
+ {
+ tx_skb[index][i] = NULL;
+ }
+#endif
+ for (i=0;i<RX_DESC_NUM;i++)
+ {
+ rx_skb[index][i] = NULL;
+ }
+
+ /* allocates TX/RX descriptors */
+ tp->tx_desc = DMA_MALLOC(TX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),(dma_addr_t *)&tp->tx_desc_dma);
+ tx_desc_virtual_base[index] = (unsigned int)tp->tx_desc - (unsigned int)tp->tx_desc_dma;
+ memset(tp->tx_desc,0x00,TX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T));
+ tp->rx_desc = DMA_MALLOC(RX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),(dma_addr_t *)&tp->rx_desc_dma);
+ rx_desc_virtual_base[index] = (unsigned int)tp->rx_desc - (unsigned int)tp->rx_desc_dma;
+ memset(tp->rx_desc,0x00,RX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T));
+ tx_desc_start_adr[index] = (unsigned int)tp->tx_desc; /* for tx skb index calculation */
+ rx_desc_start_adr[index] = (unsigned int)tp->rx_desc; /* for rx skb index calculation */
+ printk("tx_desc = %08x\n",(unsigned int)tp->tx_desc);
+ printk("rx_desc = %08x\n",(unsigned int)tp->rx_desc);
+ printk("tx_desc_dma = %08x\n",tp->tx_desc_dma);
+ printk("rx_desc_dma = %08x\n",tp->rx_desc_dma);
+
+ if (tp->tx_desc==0x00 || tp->rx_desc==0x00)
+ {
+ free_irq(dev->irq, dev);
+
+ if (tp->tx_desc)
+ DMA_MFREE(tp->tx_desc, TX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),tp->tx_desc_dma);
+ if (tp->rx_desc)
+ DMA_MFREE(tp->rx_desc, RX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),tp->rx_desc_dma);
+ return -ENOMEM;
+ }
+
+ /* TX descriptors initial */
+ tp->tx_cur_desc = tp->tx_desc; /* virtual address */
+ tp->tx_finished_desc = tp->tx_desc; /* virtual address */
+ tx_first_desc_dma = tp->tx_desc_dma; /* physical address */
+ for (i = 1; i < TX_DESC_NUM; i++)
+ {
+ tp->tx_desc->frame_ctrl.bits_tx_out.own = CPU; /* set owner to CPU */
+ tp->tx_desc->frame_ctrl.bits_tx_out.buffer_size = TX_BUF_SIZE; /* set tx buffer size for descriptor */
+ tp->tx_desc_dma = tp->tx_desc_dma + sizeof(GMAC_DESCRIPTOR_T); /* next tx descriptor DMA address */
+ tp->tx_desc->next_desc.next_descriptor = tp->tx_desc_dma | 0x0000000b;
+ tp->tx_desc = &tp->tx_desc[1] ; /* next tx descriptor virtual address */
+ }
+ /* the last descriptor will point back to first descriptor */
+ tp->tx_desc->frame_ctrl.bits_tx_out.own = CPU;
+ tp->tx_desc->frame_ctrl.bits_tx_out.buffer_size = TX_BUF_SIZE;
+ tp->tx_desc->next_desc.next_descriptor = tx_first_desc_dma | 0x0000000b;
+ tp->tx_desc = tp->tx_cur_desc;
+ tp->tx_desc_dma = tx_first_desc_dma;
+
+ /* RX descriptors initial */
+ tp->rx_cur_desc = tp->rx_desc; /* virtual address */
+ rx_first_desc_dma = tp->rx_desc_dma; /* physical address */
+ for (i = 1; i < RX_DESC_NUM; i++)
+ {
+ if ( (skb = dev_alloc_skb(RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__);
+ }
+ rx_skb[index][i-1] = skb;
+ tp->rx_desc->buf_adr = (unsigned int)__pa(skb->data) | 0x02; /* insert two bytes in the beginning of rx data */
+ tp->rx_desc->frame_ctrl.bits_rx.own = DMA; /* set owner bit to DMA */
+ tp->rx_desc->frame_ctrl.bits_rx.buffer_size = RX_BUF_SIZE; /* set rx buffer size for descriptor */
+ tp->rx_bufs_dma = tp->rx_bufs_dma + RX_BUF_SIZE; /* point to next buffer address */
+ tp->rx_desc_dma = tp->rx_desc_dma + sizeof(GMAC_DESCRIPTOR_T); /* next rx descriptor DMA address */
+ tp->rx_desc->next_desc.next_descriptor = tp->rx_desc_dma | 0x0000000b;
+ tp->rx_desc = &tp->rx_desc[1]; /* next rx descriptor virtual address */
+ }
+ /* the last descriptor will point back to first descriptor */
+ if ( (skb = dev_alloc_skb(RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__);
+ }
+ rx_skb[index][i-1] = skb;
+ tp->rx_desc->buf_adr = (unsigned int)__pa(skb->data) | 0x02; /* insert two bytes in the beginning of rx data */
+ tp->rx_desc->frame_ctrl.bits_rx.own = DMA;
+ tp->rx_desc->frame_ctrl.bits_rx.buffer_size = RX_BUF_SIZE;
+ tp->rx_desc->next_desc.next_descriptor = rx_first_desc_dma | 0x0000000b;
+ tp->rx_desc = tp->rx_cur_desc;
+ tp->rx_desc_dma = rx_first_desc_dma;
+ tp->rx_bufs_dma = rx_first_buf_dma;
+
+ for (i=0; i<GMAC_PHY_IF; i++) {
+ tp->tx_desc_hdr[i] = 0;
+ tp->tx_desc_tail[i] = 0;
+ }
+ return (0);
+}
+
+static int gmac_clear_counter (struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ unsigned int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+// tp = gmac_dev[index]->priv;
+ /* clear counter */
+ gmac_read_reg(gmac_base_addr[dev_index] + GMAC_IN_DISCARDS);
+ gmac_read_reg(gmac_base_addr[dev_index] + GMAC_IN_ERRORS);
+ tp->stats.tx_bytes = 0;
+ tp->stats.tx_packets = 0;
+ tp->stats.tx_errors = 0;
+ tp->stats.rx_bytes = 0;
+ tp->stats.rx_packets = 0;
+ tp->stats.rx_errors = 0;
+ tp->stats.rx_dropped = 0;
+ return (0);
+}
+
+static int gmac_open (struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ int retval;
+
+ gmac_select_interface(dev);
+
+ /* chip reset */
+ gmac_sw_reset(dev);
+
+ /* allocates tx/rx descriptor and buffer memory */
+ gmac_init_desc_buf(dev);
+
+ /* get mac address from FLASH */
+ gmac_get_mac_address();
+
+ /* set PHY register to start autonegition process */
+ gmac_set_phy_status(dev);
+
+ /* GMAC initialization */
+ if (gmac_init_chip(dev))
+ {
+ printk (KERN_ERR "GMAC init fail\n");
+ }
+
+ /* start DMA process */
+ gmac_hw_start(dev);
+
+ /* enable tx/rx register */
+ gmac_enable_tx_rx(dev);
+
+ /* clear statistic counter */
+ gmac_clear_counter(dev);
+
+ netif_start_queue (dev);
+
+ /* hook ISR */
+ retval = request_irq (dev->irq, gmac_interrupt, SA_INTERRUPT, dev->name, dev);
+ if (retval)
+ return retval;
+
+ if(!FLAG_SWITCH)
+ {
+ init_waitqueue_head (&tp->thr_wait);
+ init_completion(&tp->thr_exited);
+
+ tp->time_to_die = 0;
+ tp->thr_pid = kernel_thread (gmac_phy_thread, dev, CLONE_FS | CLONE_FILES);
+ if (tp->thr_pid < 0)
+ {
+ printk (KERN_WARNING "%s: unable to start kernel thread\n",dev->name);
+ }
+ }
+ return (0);
+}
+
+static int gmac_close(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ unsigned int i,dev_index;
+ unsigned int ret;
+
+ dev_index = gmac_get_dev_index(dev);
+
+ /* stop tx/rx packet */
+ gmac_disable_tx_rx(dev);
+
+ /* stop the chip's Tx and Rx DMA processes */
+ gmac_hw_stop(dev);
+
+ netif_stop_queue(dev);
+
+ /* disable interrupts by clearing the interrupt mask */
+ synchronize_irq();
+ free_irq(dev->irq,dev);
+
+ DMA_MFREE(tp->tx_desc, TX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),(unsigned int)tp->tx_desc_dma);
+ DMA_MFREE(tp->rx_desc, RX_DESC_NUM*sizeof(GMAC_DESCRIPTOR_T),(unsigned int)tp->rx_desc_dma);
+
+#ifdef CONFIG_SL2312_MPAGE
+// kfree(tx_skb);
+#endif
+
+ for (i=0;i<RX_DESC_NUM;i++)
+ {
+ if (rx_skb[dev_index][i])
+ {
+ dev_kfree_skb(rx_skb[dev_index][i]);
+ }
+ }
+ if(!FLAG_SWITCH)
+ {
+ if (tp->thr_pid >= 0)
+ {
+ tp->time_to_die = 1;
+ wmb();
+ ret = kill_proc (tp->thr_pid, SIGTERM, 1);
+ if (ret)
+ {
+ printk (KERN_ERR "%s: unable to signal thread\n", dev->name);
+ return ret;
+ }
+// wait_for_completion (&tp->thr_exited);
+ }
+ }
+
+ return (0);
+}
+
+#ifdef CONFIG_SL2312_MPAGE
+int printk_all(int dev_index, struct gmac_private* tp)
+{
+ int i=0;
+ unsigned int tx_current_descriptor = 0;
+ int hw_index;
+ int fi;
+ GMAC_DESCRIPTOR_T* tmp_desc;
+
+ GMAC_DESCRIPTOR_T* cur_desc=tp->tx_cur_desc;
+ fi = ((unsigned int)cur_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ printk("tmp_desc %x, id %d\n", (int)cur_desc, fi);
+
+ tmp_desc = (GMAC_DESCRIPTOR_T*)((gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC) & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ hw_index = ((unsigned int)tmp_desc - tx_desc_start_adr[dev_index])/ sizeof(GMAC_DESCRIPTOR_T);
+ printk("hd_desc %x, ind %d, fin desc %x\n",(int)tmp_desc, hw_index, (int)tp->tx_finished_desc);
+
+ for (i=0; i<TX_DESC_NUM; i++) {
+ printk("**id %4d, hw_index %4d ==> ", fi, hw_index);
+ printk("fc %8x ", tmp_desc->frame_ctrl.bits32);
+ printk("fs %8x ", tmp_desc->flag_status.bits32);
+ printk("fb %8x ", tmp_desc->buf_adr);
+ printk("fd %8x\n", tmp_desc->next_desc.next_descriptor);
+ tmp_desc = (GMAC_DESCRIPTOR_T*)((tmp_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ fi = ((unsigned int)tmp_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ }
+ tx_current_descriptor = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CURR_DESC);
+ printk("%s: tx current descriptor = %x \n",__func__,tx_current_descriptor);
+ printk("%s: interrupt status = %x \n",__func__,int_status);
+ return 0;
+}
+
+int cleanup_desc(int dev_index, struct gmac_private* tp)
+{
+ int i=0;
+ int index = ((unsigned int)tp->tx_cur_desc - tx_desc_start_adr[dev_index])/sizeof(GMAC_DESCRIPTOR_T);
+ GMAC_DESCRIPTOR_T* fill_desc = tp->tx_cur_desc;
+
+ for (i=0; i< TX_DESC_NUM; i++)
+ {
+ fill_desc->frame_ctrl.bits_tx_out.own = CPU;
+ fill_desc->frame_ctrl.bits_tx_out.buffer_size = TX_BUF_SIZE;
+ tx_skb[dev_index][index].desc_in_use = 0;
+ free_tx_buf(dev_index, index);
+ printk("cleanup di %d\n", index);
+ fill_desc = (GMAC_DESCRIPTOR_T*)((fill_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ index++;
+ if (index > TX_DESC_NUM)
+ index = 0;
+ }
+ return 1;
+}
+
+size_t get_available_tx_desc(struct net_device* dev, int dev_index)
+{
+ struct gmac_private *tp = dev->priv;
+ unsigned int desc_hdr = tp->tx_desc_hdr[dev_index];
+ unsigned int desc_tail = tp->tx_desc_tail[dev_index];
+ int available_desc_num = (TX_DESC_NUM - desc_hdr + desc_tail) & (TX_DESC_NUM-1);
+ if (!available_desc_num) {
+ if (tx_skb[dev_index][desc_hdr].desc_in_use)
+ return 0;
+ else
+ return TX_DESC_NUM;
+ }
+ return available_desc_num;
+}
+
+int check_free_tx_desc(int dev_index, int n, GMAC_DESCRIPTOR_T* desc)
+{
+ int i,index;
+ GMAC_DESCRIPTOR_T* tmp_desc = desc;
+
+ if (n > TX_DESC_NUM)
+ return 0;
+
+ index = ((unsigned int)tmp_desc - tx_desc_start_adr[dev_index])/sizeof(GMAC_DESCRIPTOR_T);
+ for (i=0; i<n; i++)
+ {
+ if (tx_skb[dev_index][index].desc_in_use)
+ {
+ printk("sw desc %d is in use\n", index);
+ /* cleanup all the descriptors to check if DMA still running */
+ return 0;
+ }
+ index++;
+ if (index == TX_DESC_NUM)
+ index = 0;
+ }
+ return 1;
+}
+
+#define TCPHDRLEN(tcp_hdr) ((ntohs(*((__u16 *)tcp_hdr + 6)) >> 12) & 0x000F)
+
+inline int fill_in_desc(int dev_index, GMAC_DESCRIPTOR_T *desc, char* data, int len, int total_len, int sof, int freeable, int ownership, struct sk_buff* skb)
+{
+ int index = ((unsigned int)desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+
+ if (desc->frame_ctrl.bits_tx_in.own == CPU)
+ {
+ tx_skb[dev_index][index].freeable = freeable;
+ if ((sof & 0x01) && skb) {
+ tx_skb[dev_index][index].skb = skb;
+ }
+ else
+ tx_skb[dev_index][index].skb = 0;
+
+ if (sof != 2)
+ tx_skb[dev_index][index].desc_in_use = 1;
+ else
+ tx_skb[dev_index][index].desc_in_use = 0;
+
+ consistent_sync(data, len, PCI_DMA_TODEVICE);
+ desc->buf_adr = (unsigned int)__pa(data);
+ desc->frame_ctrl.bits_tx_out.buffer_size = len;
+ desc->flag_status.bits_tx_flag.frame_count = total_len;
+ desc->next_desc.bits.eofie = 1;
+ desc->next_desc.bits.sof_eof = sof;
+ desc->frame_ctrl.bits_tx_out.vlan_enable = 0;
+ desc->frame_ctrl.bits_tx_out.ip_csum_en = 1; /* TSS IPv4 IP header checksum enable */
+ desc->frame_ctrl.bits_tx_out.ipv6_tx_en = 1; /* TSS IPv6 tx enable */
+ desc->frame_ctrl.bits_tx_out.tcp_csum_en = 1; /* TSS TCP checksum enable */
+ desc->frame_ctrl.bits_tx_out.udp_csum_en = 1; /* TSS UDP checksum enable */
+ wmb();
+ desc->frame_ctrl.bits_tx_out.own = ownership;
+// consistent_sync(desc, sizeof(GMAC_DESCRIPTOR_T), PCI_DMA_TODEVICE);
+ }
+ return 0;
+}
+#endif
+
+static int gmac_start_xmit(struct sk_buff *skb, struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ GMAC_TXDMA_CTRL_T tx_ctrl,tx_ctrl_mask;
+ GMAC_TXDMA_FIRST_DESC_T txdma_busy;
+ unsigned int len = skb->len;
+ unsigned int dev_index;
+ static unsigned int pcount = 0;
+#ifdef CONFIG_SL2312_MPAGE
+ GMAC_DESCRIPTOR_T *fill_desc;
+ int snd_pages = skb_shinfo(skb)->nr_frags; /* get number of descriptor */
+ int desc_needed = 1; // for jumbo packet, one descriptor is enough.
+ int header_len = skb->len;
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ int tcp_hdr_len;
+ int data_len;
+ int prv_index;
+ long seq_num;
+ int first_desc_index;
+ int ownership, freeable;
+ int eof;
+ int i=0;
+#endif
+#ifdef CONFIG_TXINT_DISABLE
+ int available_desc_cnt = 0;
+#endif
+
+ dev_index = gmac_select_interface(dev);
+
+#ifdef CONFIG_TXINT_DISABLE
+ available_desc_cnt = get_available_tx_desc(dev, dev_index);
+
+ if (available_desc_cnt < (TX_DESC_NUM >> 2)) {
+ gmac_tx_packet_complete(dev);
+ }
+#endif
+
+#ifdef CONFIG_SL2312_MPAGE
+
+ fill_desc = tp->tx_cur_desc;
+ if(!fill_desc) {
+ printk("cur_desc is NULL!\n");
+ return -1;
+ }
+
+ if (storlink_ctl.recvfile==2)
+ {
+ printk("snd_pages=%d skb->len=%d\n",snd_pages,skb->len);
+ }
+
+ if (snd_pages)
+ desc_needed += snd_pages; /* decriptors needed for this large packet */
+
+ if (!check_free_tx_desc(dev_index, desc_needed, fill_desc)) {
+ printk("no available desc!\n");
+ gmac_dump_register(dev);
+ printk_all(dev_index, tp);
+ tp->stats.tx_dropped++;
+ if (pcount++ > 10)
+ {
+ for (;;);
+ }
+ return -1;
+ }
+
+ first_desc_index = ((unsigned int)fill_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+
+ /* check if the tcp packet is in order*/
+ ip_hdr = (struct iphdr*) &(skb->data[14]);
+ tcp_hdr = (struct tcphdr*) &(skb->data[14+ip_hdr->ihl * 4]);
+ tcp_hdr_len = TCPHDRLEN(tcp_hdr) * 4;
+ data_len = skb->len - 14 - ip_hdr->ihl *4 - tcp_hdr_len;
+
+ prv_index = first_desc_index-1;
+ if (prv_index <0)
+ prv_index += TX_DESC_NUM;
+ seq_num = ntohl(tcp_hdr->seq);
+
+ if (snd_pages)
+ {
+ // calculate header length
+ // check fragment total length and header len = skb len - frag len
+ // or parse the header.
+ for (i=0; i<snd_pages; i++) {
+ skb_frag_t* frag = &skb_shinfo(skb)->frags[i];
+ header_len -= frag->size;
+ }
+ ownership = CPU;
+ freeable = 0;
+ /* fill header into first descriptor */
+ fill_in_desc(dev_index, fill_desc, skb->data, header_len, len, 2, freeable, ownership, 0);
+ fill_desc = (GMAC_DESCRIPTOR_T*)((fill_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ tx_skb[dev_index][first_desc_index].end_seq = seq_num + data_len;
+
+ eof = 0;
+ ownership = DMA;
+ for (i=0; i<snd_pages; i++)
+ {
+ skb_frag_t* frag = &skb_shinfo(skb)->frags[i];
+ int start_pos = frag->page_offset;
+ char* data_buf = page_address(frag->page);
+ int data_size = frag->size;
+ int cur_index;
+
+ if (i == snd_pages-1)
+ {
+ eof=1;
+ freeable = 1;
+ }
+ fill_in_desc(dev_index, fill_desc, data_buf+(start_pos), data_size,
+ len, eof, freeable, ownership, skb);
+ cur_index = ((unsigned int)fill_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+
+ fill_desc = (GMAC_DESCRIPTOR_T*)((fill_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ }
+ /* pass the ownership of the first descriptor to hardware */
+// disable_irq(gmac_irq[dev_index]);
+ tx_skb[dev_index][first_desc_index].desc_in_use = 1;
+ wmb();
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.own = DMA;
+// consistent_sync(tp->tx_cur_desc, sizeof(GMAC_DESCRIPTOR_T), PCI_DMA_TODEVICE);
+ tp->tx_cur_desc = fill_desc;
+ dev->trans_start = jiffies;
+// enable_irq(gmac_irq[dev_index]);
+ }
+ else if ( tp->tx_cur_desc->frame_ctrl.bits_tx_out.own == CPU )
+ {
+// tx_skb[dev_index][first_desc_index].end_seq = seq_num + data_len;
+// disable_irq(gmac_irq[dev_index]);
+ fill_in_desc(dev_index, tp->tx_cur_desc, skb->data, skb->len, skb->len, 3, 1, DMA, skb);
+// enable_irq(gmac_irq[dev_index]);
+ //consistent_sync(tp->tx_cur_desc, sizeof(GMAC_DESCRIPTOR_T), PCI_DMA_TODEVICE);
+ tp->tx_cur_desc = (GMAC_DESCRIPTOR_T*)((tp->tx_cur_desc->next_desc.next_descriptor & 0xfffffff0) + tx_desc_virtual_base[dev_index]);
+ dev->trans_start = jiffies;
+ }
+ else
+ {
+ printk("gmac tx drop!\n");
+ tp->stats.tx_dropped++;
+ return -1;
+ }
+
+#ifdef CONFIG_TXINT_DISABLE
+ tp->tx_desc_hdr[dev_index] = (tp->tx_desc_hdr[dev_index] + desc_needed) & (TX_DESC_NUM-1);
+#endif
+
+#else
+ if ((tp->tx_cur_desc->frame_ctrl.bits_tx_out.own == CPU) && (len < TX_BUF_SIZE))
+ {
+ index = ((unsigned int)tp->tx_cur_desc - tx_desc_start_adr[dev_index]) / sizeof(GMAC_DESCRIPTOR_T);
+ tx_skb[dev_index][index] = skb;
+ consistent_sync(skb->data,skb->len,PCI_DMA_TODEVICE);
+ tp->tx_cur_desc->buf_adr = (unsigned int)__pa(skb->data);
+ tp->tx_cur_desc->flag_status.bits_tx_flag.frame_count = len; /* total frame byte count */
+ tp->tx_cur_desc->next_desc.bits.sof_eof = 0x03; /*only one descriptor*/
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.buffer_size = len; /* descriptor byte count */
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.vlan_enable = 0;
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.ip_csum_en = 0; /* TSS IPv4 IP header checksum enable */
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.ipv6_tx_en = 0 ; /* TSS IPv6 tx enable */
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.tcp_csum_en = 0; /* TSS TCP checksum enable */
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.udp_csum_en = 0; /* TSS UDP checksum enable */
+ wmb();
+ tp->tx_cur_desc->frame_ctrl.bits_tx_out.own = DMA; /* set owner bit */
+ tp->tx_cur_desc = (GMAC_DESCRIPTOR_T *)((tp->tx_cur_desc->next_desc.next_descriptor & 0xfffffff0)+tx_desc_virtual_base[dev_index]);
+ dev->trans_start = jiffies;
+ }
+ else
+ {
+ /* no free tx descriptor */
+ dev_kfree_skb(skb);
+ netif_stop_queue(dev);
+ tp->stats.tx_dropped++;
+ return (-1);
+ }
+#endif
+ /* if TX DMA process is stoped , restart it */
+ txdma_busy.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_FIRST_DESC);
+ if (txdma_busy.bits.td_busy == 0)
+ {
+ /* restart DMA process */
+ tx_ctrl.bits32 = 0;
+ tx_ctrl.bits.td_start = 1;
+ tx_ctrl.bits.td_continue = 1;
+ tx_ctrl_mask.bits32 = 0;
+ tx_ctrl_mask.bits.td_start = 1;
+ tx_ctrl_mask.bits.td_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CTRL,tx_ctrl.bits32,tx_ctrl_mask.bits32);
+ }
+ return (0);
+}
+
+
+struct net_device_stats * gmac_get_stats(struct net_device *dev)
+{
+ struct gmac_private *tp = dev->priv;
+ unsigned long flags;
+ unsigned int pkt_drop;
+ unsigned int pkt_error;
+ unsigned int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+// if (storlink_ctl.recvfile==3)
+// {
+// printk("GMAC_GLOBAL_BASE_ADDR=%x\n", readl(GMAC_GLOBAL_BASE_ADDR+0x30));
+// gmac_dump_register(dev);
+// printk_all(0, dev);
+// }
+
+ if (netif_running(dev))
+ {
+ /* read H/W counter */
+ spin_lock_irqsave(&tp->lock,flags);
+ pkt_drop = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_IN_DISCARDS);
+ pkt_error = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_IN_ERRORS);
+ tp->stats.rx_dropped = tp->stats.rx_dropped + pkt_drop;
+ tp->stats.rx_errors = tp->stats.rx_errors + pkt_error;
+ spin_unlock_irqrestore(&tp->lock,flags);
+ }
+ return &tp->stats;
+}
+
+static unsigned const ethernet_polynomial = 0x04c11db7U;
+static inline u32 ether_crc (int length, unsigned char *data)
+{
+ int crc = -1;
+ unsigned int i;
+ unsigned int crc_val=0;
+
+ while (--length >= 0) {
+ unsigned char current_octet = *data++;
+ int bit;
+ for (bit = 0; bit < 8; bit++, current_octet >>= 1)
+ crc = (crc << 1) ^ ((crc < 0) ^ (current_octet & 1) ?
+ ethernet_polynomial : 0);
+ }
+ crc = ~crc;
+ for (i=0;i<32;i++)
+ {
+ crc_val = crc_val + (((crc << i) & 0x80000000) >> (31-i));
+ }
+ return crc_val;
+}
+
+static void gmac_set_rx_mode(struct net_device *dev)
+{
+ GMAC_RX_FLTR_T filter;
+ unsigned int mc_filter[2]; /* Multicast hash filter */
+ int bit_nr;
+ unsigned int i, dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+// printk("%s : dev->flags = %x \n",__func__,dev->flags);
+// dev->flags |= IFF_ALLMULTI; /* temp */
+ filter.bits32 = 0;
+ filter.bits.error = 0;
+ if (dev->flags & IFF_PROMISC)
+ {
+ filter.bits.error = 1;
+ filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0xffffffff;
+ }
+ else if (dev->flags & IFF_ALLMULTI)
+ {
+ filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0xffffffff;
+ }
+ else
+ {
+ struct dev_mc_list *mclist;
+
+ filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0;
+ for (i = 0, mclist = dev->mc_list; mclist && i < dev->mc_count;i++, mclist = mclist->next)
+ {
+ bit_nr = ether_crc(ETH_ALEN,mclist->dmi_addr) & 0x0000003f;
+ if (bit_nr < 32)
+ {
+ mc_filter[0] = mc_filter[0] | (1<<bit_nr);
+ }
+ else
+ {
+ mc_filter[1] = mc_filter[1] | (1<<(bit_nr-32));
+ }
+ }
+ }
+ filter.bits32 = 0x1f;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_RX_FLTR,filter.bits32,0xffffffff);
+
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_MCAST_FIL0,mc_filter[0],0xffffffff);
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_MCAST_FIL1,mc_filter[1],0xffffffff);
+ return;
+}
+
+static int gmac_set_mac_address(struct net_device *dev, void *addr)
+{
+ struct sockaddr *sock;
+ unsigned int reg_val;
+ unsigned int dev_index;
+ unsigned int i;
+
+ dev_index = gmac_select_interface(dev);
+
+ sock = (struct sockaddr *) addr;
+ for (i = 0; i < 6; i++)
+ {
+ dev->dev_addr[i] = sock->sa_data[i];
+ }
+
+ reg_val = dev->dev_addr[0] + (dev->dev_addr[1]<<8) + (dev->dev_addr[2]<<16) + (dev->dev_addr[3]<<24);
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_STA_ADD0,reg_val,0xffffffff);
+ reg_val = dev->dev_addr[4] + (dev->dev_addr[5]<<8) ;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_STA_ADD1,reg_val,0x0000ffff);
+ memcpy(&eth0_mac[0],&dev->dev_addr[0],6);
+ printk("Storlink %s address = ",dev->name);
+ printk("%02x",dev->dev_addr[0]);
+ printk("%02x",dev->dev_addr[1]);
+ printk("%02x",dev->dev_addr[2]);
+ printk("%02x",dev->dev_addr[3]);
+ printk("%02x",dev->dev_addr[4]);
+ printk("%02x\n",dev->dev_addr[5]);
+
+ return (0);
+}
+
+static void gmac_tx_timeout(struct net_device *dev)
+{
+ GMAC_TXDMA_CTRL_T tx_ctrl,tx_ctrl_mask;
+ GMAC_TXDMA_FIRST_DESC_T txdma_busy;
+ int dev_index;
+
+ dev_index = gmac_select_interface(dev);
+
+ /* if TX DMA process is stoped , restart it */
+ txdma_busy.bits32 = gmac_read_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_FIRST_DESC);
+ if (txdma_busy.bits.td_busy == 0)
+ {
+ /* restart DMA process */
+ tx_ctrl.bits32 = 0;
+ tx_ctrl.bits.td_start = 1;
+ tx_ctrl.bits.td_continue = 1;
+ tx_ctrl_mask.bits32 = 0;
+ tx_ctrl_mask.bits.td_start = 1;
+ tx_ctrl_mask.bits.td_continue = 1;
+ gmac_write_reg(gmac_base_addr[dev_index] + GMAC_TXDMA_CTRL,tx_ctrl.bits32,tx_ctrl_mask.bits32);
+ }
+ netif_wake_queue(dev);
+ return;
+}
+
+static int gmac_netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
+{
+ int rc = 0;
+ unsigned char *hwa = rq->ifr_ifru.ifru_hwaddr.sa_data;
+
+ if (!netif_running(dev))
+ {
+ printk("Before changing the H/W address,please down the device.\n");
+ return -EINVAL;
+ }
+
+ switch (cmd) {
+ case SIOCETHTOOL:
+ break;
+
+ case SIOCSIFHWADDR:
+ gmac_set_mac_address(dev,hwa);
+ break;
+
+ case SIOCGMIIPHY: /* Get the address of the PHY in use. */
+ case SIOCDEVPRIVATE: /* binary compat, remove in 2.5 */
+ break;
+
+ case SIOCGMIIREG: /* Read the specified MII register. */
+ case SIOCDEVPRIVATE+1:
+ break;
+
+ case SIOCSMIIREG: /* Write the specified MII register */
+ case SIOCDEVPRIVATE+2:
+ break;
+
+ default:
+ rc = -EOPNOTSUPP;
+ break;
+ }
+
+ return rc;
+}
+
+static void gmac_cleanup_module(void)
+{
+ int i;
+
+ for (i=0;i<GMAC_PHY_IF;i++)
+ {
+ unregister_netdev(gmac_dev[i]);
+ }
+ return ;
+}
+
+static int __init gmac_init_module(void)
+{
+ struct gmac_private *tp;
+ struct net_device *dev[GMAC_PHY_IF];
+ unsigned int i;
+
+#ifdef MODULE
+ printk (KERN_INFO RTL8139_DRIVER_NAME "\n");
+#endif
+// init_waitqueue_entry(&wait, current);
+
+ printk("GMAC Init......\n");
+ for(i = 0; i<GMAC_PHY_IF; i++)
+ {
+ dev[i] = alloc_etherdev(sizeof(struct gmac_private));
+ if (dev[i] == NULL)
+ {
+ printk (KERN_ERR "Can't allocate ethernet device #%d .\n",i);
+ return -ENOMEM;
+ }
+ gmac_dev[i] = dev[i];
+
+ SET_MODULE_OWNER(dev[i]);
+
+ tp = dev[i]->priv;
+
+ dev[i]->base_addr = gmac_base_addr[i];
+ dev[i]->irq = gmac_irq[i];
+ dev[i]->open = gmac_open;
+ dev[i]->stop = gmac_close;
+ dev[i]->hard_start_xmit = gmac_start_xmit;
+ dev[i]->get_stats = gmac_get_stats;
+ dev[i]->set_multicast_list = gmac_set_rx_mode;
+ dev[i]->set_mac_address = gmac_set_mac_address;
+ dev[i]->do_ioctl = gmac_netdev_ioctl;
+ dev[i]->tx_timeout = gmac_tx_timeout;
+ dev[i]->watchdog_timeo = TX_TIMEOUT;
+ dev[i]->features |= NETIF_F_SG|NETIF_F_HW_CSUM|NETIF_F_TSO;
+#ifdef CONFIG_SL_NAPI
+ printk("NAPI driver is enabled.\n");
+ if (i==0)
+ {
+ dev[i]->poll = gmac_rx_poll_ga;
+ dev[i]->weight = 64;
+ }
+ else
+ {
+ dev[i]->poll = gmac_rx_poll_gb;
+ dev[i]->weight = 64;
+ }
+#endif
+
+ if (register_netdev(dev[i]))
+ {
+ gmac_cleanup_module();
+ return(-1);
+ }
+ }
+
+#ifdef CONFIG_SL3516_ASIC
+{
+ unsigned int val;
+
+ /* set GMAC global register */
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x10);
+ val = val | 0x005a0000;
+ writel(val,GMAC_GLOBAL_BASE_ADDR+0x10);
+ writel(0x07f007f0,GMAC_GLOBAL_BASE_ADDR+0x1c);
+ writel(0x77770000,GMAC_GLOBAL_BASE_ADDR+0x20);
+ writel(0x77770000,GMAC_GLOBAL_BASE_ADDR+0x24);
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x04);
+ if((val&(1<<20))==0){ // GMAC1 enable
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x30);
+ val = (val & 0xe7ffffff) | 0x08000000;
+ writel(val,GMAC_GLOBAL_BASE_ADDR+0x30);
+ }
+
+}
+#endif
+
+// printk("%s: dev0=%x dev1=%x \n",__func__,dev[0],dev[1]);
+// FLAG_SWITCH = 0 ;
+// FLAG_SWITCH = SPI_get_identifier();
+// if(FLAG_SWITCH)
+// {
+// printk("Configure ADM699X...\n");
+// SPI_default(); //Add by jason for ADM699X configuration
+// }
+ return (0);
+}
+
+
+module_init(gmac_init_module);
+module_exit(gmac_cleanup_module);
+
+static int gmac_phy_thread (void *data)
+{
+ struct net_device *dev = data;
+ struct gmac_private *tp = dev->priv;
+ unsigned long timeout;
+
+ daemonize("%s", dev->name);
+ allow_signal(SIGTERM);
+// reparent_to_init();
+// spin_lock_irq(&current->sigmask_lock);
+// sigemptyset(&current->blocked);
+// recalc_sigpending(current);
+// spin_unlock_irq(&current->sigmask_lock);
+// strncpy (current->comm, dev->name, sizeof(current->comm) - 1);
+// current->comm[sizeof(current->comm) - 1] = '\0';
+
+ while (1)
+ {
+ timeout = next_tick;
+ do
+ {
+ timeout = interruptible_sleep_on_timeout (&tp->thr_wait, timeout);
+ } while (!signal_pending (current) && (timeout > 0));
+
+ if (signal_pending (current))
+ {
+// spin_lock_irq(&current->sigmask_lock);
+ flush_signals(current);
+// spin_unlock_irq(&current->sigmask_lock);
+ }
+
+ if (tp->time_to_die)
+ break;
+
+// printk("%s : Polling PHY Status...%x\n",__func__,dev);
+ rtnl_lock ();
+ gmac_get_phy_status(dev);
+ rtnl_unlock ();
+ }
+ complete_and_exit (&tp->thr_exited, 0);
+}
+
+static void gmac_set_phy_status(struct net_device *dev)
+{
+ GMAC_STATUS_T status;
+ unsigned int reg_val;
+ unsigned int i = 0;
+ unsigned int index;
+
+ if (FLAG_SWITCH==1)
+ {
+ return; /* GMAC connects to a switch chip, not PHY */
+ }
+
+ index = gmac_get_dev_index(dev);
+
+ if (index == 0)
+ {
+// mii_write(phy_addr[index],0x04,0x0461); /* advertisement 10M full duplex, pause capable on */
+// mii_write(phy_addr[index],0x04,0x0421); /* advertisement 10M half duplex, pause capable on */
+ mii_write(phy_addr[index],0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+// mii_write(phy_addr[index],0x04,0x04a1); /* advertisement 100M half duplex, pause capable on */
+#ifdef CONFIG_SL3516_ASIC
+ mii_write(phy_addr[index],0x09,0x0300); /* advertisement 1000M full duplex, pause capable on */
+// mii_write(phy_addr[index],0x09,0x0000); /* advertisement 1000M full duplex, pause capable on */
+#endif
+ }
+ else
+ {
+// mii_write(phy_addr[index],0x04,0x0461); /* advertisement 10M full duplex, pause capable on */
+// mii_write(phy_addr[index],0x04,0x0421); /* advertisement 10M half duplex, pause capable on */
+ mii_write(phy_addr[index],0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+// mii_write(phy_addr[index],0x04,0x04a1); /* advertisement 100M half duplex, pause capable on */
+#ifdef CONFIG_SL3516_ASIC
+// mii_write(phy_addr[index],0x09,0x0000); /* advertisement no 1000M */
+ mii_write(phy_addr[index],0x09,0x0300); /* advertisement 1000M full duplex, pause capable on */
+#endif
+ }
+
+ mii_write(phy_addr[index],0x00,0x1200); /* Enable and Restart Auto-Negotiation */
+ mii_write(phy_addr[index],0x18,0x0041); /* Enable Active led */
+ while (((reg_val=mii_read(phy_addr[index],0x01)) & 0x00000004)!=0x04)
+ {
+ i++;
+ if (i > 30)
+ {
+ break;
+ }
+ msleep(100);
+ }
+ if (i>30)
+ {
+ pre_phy_status[index] = LINK_DOWN;
+ clear_bit(__LINK_STATE_START, &dev->state);
+ netif_stop_queue(dev);
+ storlink_ctl.link = 0;
+ printk("Link Down (%04x) ",reg_val);
+ }
+ else
+ {
+ pre_phy_status[index] = LINK_UP;
+ set_bit(__LINK_STATE_START, &dev->state);
+ netif_wake_queue(dev);
+ storlink_ctl.link = 1;
+ printk("Link Up (%04x) ",reg_val);
+ }
+
+ status.bits32 = 0;
+ reg_val = mii_read(phy_addr[index],10);
+ printk("reg_val0 = %x \n",reg_val);
+ if ((reg_val & 0x0800) == 0x0800)
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 2;
+ printk(" 1000M/Full \n");
+ }
+ else if ((reg_val & 0x0400) == 0x0400)
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 2;
+ printk(" 1000M/Half \n");
+ }
+ else
+ {
+ reg_val = (mii_read(phy_addr[index],0x05) & 0x05E0) >> 5;
+ printk("reg_val1 = %x \n",reg_val);
+ if ((reg_val & 0x08)==0x08) /* 100M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 1;
+ printk(" 100M/Full \n");
+ }
+ else if ((reg_val & 0x04)==0x04) /* 100M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 1;
+ printk(" 100M/Half \n");
+ }
+ else if ((reg_val & 0x02)==0x02) /* 10M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 0;
+ printk(" 10M/Full \n");
+ }
+ else if ((reg_val & 0x01)==0x01) /* 10M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 0;
+ printk(" 100M/Half \n");
+ }
+ }
+
+ reg_val = (mii_read(phy_addr[index],0x05) & 0x05E0) >> 5;
+ if ((reg_val & 0x20)==0x20)
+ {
+ flow_control_enable[index] = 1;
+ printk("Flow Control Enable. \n");
+ }
+ else
+ {
+ flow_control_enable[index] = 0;
+ printk("Flow Control Disable. \n");
+ }
+ full_duplex = status.bits.duplex;
+ speed = status.bits.speed;
+}
+
+static void gmac_get_phy_status(struct net_device *dev)
+{
+ GMAC_CONFIG0_T config0,config0_mask;
+ GMAC_STATUS_T status;
+ unsigned int reg_val;
+ unsigned int index;
+
+ index = gmac_select_interface(dev);
+
+ status.bits32 = 0;
+ status.bits.phy_mode = 1;
+
+#ifdef CONFIG_SL3516_ASIC
+ status.bits.mii_rmii = 2; /* default value for ASIC version */
+// status.bits.speed = 1;
+#else
+ if (index==0)
+ status.bits.mii_rmii = 0;
+ else
+ status.bits.mii_rmii = 2;
+#endif
+
+ /* read PHY status register */
+ reg_val = mii_read(phy_addr[index],0x01);
+ if ((reg_val & 0x0024) == 0x0024) /* link is established and auto_negotiate process completed */
+ {
+ /* read PHY Auto-Negotiation Link Partner Ability Register */
+ reg_val = mii_read(phy_addr[index],10);
+ if ((reg_val & 0x0800) == 0x0800)
+ {
+ status.bits.mii_rmii = 3; /* RGMII 1000Mbps mode */
+ status.bits.duplex = 1;
+ status.bits.speed = 2;
+ }
+ else if ((reg_val & 0x0400) == 0x0400)
+ {
+ status.bits.mii_rmii = 3; /* RGMII 1000Mbps mode */
+ status.bits.duplex = 0;
+ status.bits.speed = 2;
+ }
+ else
+ {
+ reg_val = (mii_read(phy_addr[index],0x05) & 0x05E0) >> 5;
+ if ((reg_val & 0x08)==0x08) /* 100M full duplex */
+ {
+ status.bits.mii_rmii = 2; /* RGMII 10/100Mbps mode */
+ status.bits.duplex = 1;
+ status.bits.speed = 1;
+ }
+ else if ((reg_val & 0x04)==0x04) /* 100M half duplex */
+ {
+ status.bits.mii_rmii = 2; /* RGMII 10/100Mbps mode */
+ status.bits.duplex = 0;
+ status.bits.speed = 1;
+ }
+ else if ((reg_val & 0x02)==0x02) /* 10M full duplex */
+ {
+ status.bits.mii_rmii = 2; /* RGMII 10/100Mbps mode */
+ status.bits.duplex = 1;
+ status.bits.speed = 0;
+ }
+ else if ((reg_val & 0x01)==0x01) /* 10M half duplex */
+ {
+ status.bits.mii_rmii = 2; /* RGMII 10/100Mbps mode */
+ status.bits.duplex = 0;
+ status.bits.speed = 0;
+ }
+ }
+ status.bits.link = LINK_UP; /* link up */
+ netif_wake_queue(dev);
+
+ reg_val = (mii_read(phy_addr[index],0x05) & 0x05E0) >> 5;
+ if ((reg_val & 0x20)==0x20)
+ {
+ if (flow_control_enable[index] == 0)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 1; /* enable tx flow control */
+ config0.bits.rx_fc_en = 1; /* enable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(gmac_base_addr[index] + GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+// printk("eth%d Flow Control Enable. \n",index);
+ }
+ flow_control_enable[index] = 1;
+ }
+ else
+ {
+ if (flow_control_enable[index] == 1)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 0; /* disable tx flow control */
+ config0.bits.rx_fc_en = 0; /* disable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(gmac_base_addr[index] + GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+// printk("eth%d Flow Control Disable. \n",index);
+ }
+ flow_control_enable[index] = 0;
+ }
+
+ if (pre_phy_status[index] == LINK_DOWN)
+ {
+ gmac_enable_tx_rx(dev);
+ pre_phy_status[index] = LINK_UP;
+ set_bit(__LINK_STATE_START, &dev->state);
+ storlink_ctl.link = 1;
+// printk("eth%d Link Up ...\n",index);
+ }
+ }
+ else
+ {
+ status.bits.link = LINK_DOWN; /* link down */
+ netif_stop_queue(dev);
+ flow_control_enable[index] = 0;
+ storlink_ctl.link = 0;
+ if (pre_phy_status[index] == LINK_UP)
+ {
+ gmac_disable_tx_rx(dev);
+ pre_phy_status[index] = LINK_DOWN;
+ clear_bit(__LINK_STATE_START, &dev->state);
+// printk("eth%d Link Down ...\n",index);
+ }
+
+ }
+
+ reg_val = gmac_read_reg(gmac_base_addr[index] + GMAC_STATUS);
+ if (reg_val != status.bits32)
+ {
+ gmac_write_reg(gmac_base_addr[index] + GMAC_STATUS,status.bits32,0x0000007f);
+ }
+}
+
+/***************************************/
+/* define GPIO module base address */
+/***************************************/
+#define GPIO_BASE_ADDR (IO_ADDRESS(SL2312_GPIO_BASE))
+
+/* define GPIO pin for MDC/MDIO */
+
+// for gemini ASIC
+#ifdef CONFIG_SL3516_ASIC
+#define H_MDC_PIN 22
+#define H_MDIO_PIN 21
+#define G_MDC_PIN 22
+#define G_MDIO_PIN 21
+#else
+#define H_MDC_PIN 3
+#define H_MDIO_PIN 2
+#define G_MDC_PIN 0
+#define G_MDIO_PIN 1
+#endif
+
+//#define GPIO_MDC 0x80000000
+//#define GPIO_MDIO 0x00400000
+
+static unsigned int GPIO_MDC = 0;
+static unsigned int GPIO_MDIO = 0;
+static unsigned int GPIO_MDC_PIN = 0;
+static unsigned int GPIO_MDIO_PIN = 0;
+
+// For PHY test definition!!
+#define LPC_EECK 0x02
+#define LPC_EDIO 0x04
+#define LPC_GPIO_SET 3
+#define LPC_BASE_ADDR IO_ADDRESS(IT8712_IO_BASE)
+#define inb_gpio(x) inb(LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+#define outb_gpio(x, y) outb(y, LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+
+enum GPIO_REG
+{
+ GPIO_DATA_OUT = 0x00,
+ GPIO_DATA_IN = 0x04,
+ GPIO_PIN_DIR = 0x08,
+ GPIO_BY_PASS = 0x0c,
+ GPIO_DATA_SET = 0x10,
+ GPIO_DATA_CLEAR = 0x14,
+};
+/***********************/
+/* MDC : GPIO[31] */
+/* MDIO: GPIO[22] */
+/***********************/
+
+/***************************************************
+* All the commands should have the frame structure:
+*<PRE><ST><OP><PHYAD><REGAD><TA><DATA><IDLE>
+****************************************************/
+
+/*****************************************************************
+* Inject a bit to NWay register through CSR9_MDC,MDIO
+*******************************************************************/
+void mii_serial_write(char bit_MDO) // write data into mii PHY
+{
+#if 0 //def CONFIG_SL2312_LPC_IT8712
+ unsigned char iomode,status;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode |= (LPC_EECK|LPC_EDIO) ; // Set EECK,EDIO,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ if(bit_MDO)
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= LPC_EDIO ; //EDIO high
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+ else
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status &= ~(LPC_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+
+ status |= LPC_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(LPC_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+#else
+ unsigned int addr;
+ unsigned int value;
+
+ addr = GPIO_BASE_ADDR + GPIO_PIN_DIR;
+ value = readl(addr) | GPIO_MDC | GPIO_MDIO; /* set MDC/MDIO Pin to output */
+ writel(value,addr);
+ if(bit_MDO)
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDIO,addr); /* set MDIO to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDIO,addr); /* set MDIO to 0 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+ }
+
+#endif
+}
+
+/**********************************************************************
+* read a bit from NWay register through CSR9_MDC,MDIO
+***********************************************************************/
+unsigned int mii_serial_read(void) // read data from mii PHY
+{
+#if 0 //def CONFIG_SL2312_LPC_IT8712
+ unsigned char iomode,status;
+ unsigned int value ;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode &= ~(LPC_EDIO) ; // Set EDIO input
+ iomode |= (LPC_EECK) ; // Set EECK,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= LPC_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(LPC_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+ value = inb_gpio( LPC_GPIO_SET);
+
+ value = value>>2 ;
+ value &= 0x01;
+
+ return value ;
+
+#else
+ unsigned int *addr;
+ unsigned int value;
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_PIN_DIR);
+ value = readl(addr) & ~GPIO_MDIO; //0xffbfffff; /* set MDC to output and MDIO to input */
+ writel(value,addr);
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_IN);
+ value = readl(addr);
+ value = (value & (1<<GPIO_MDIO_PIN)) >> GPIO_MDIO_PIN;
+ return(value);
+
+#endif
+}
+
+/***************************************
+* preamble + ST
+***************************************/
+void mii_pre_st(void)
+{
+ unsigned char i;
+
+ for(i=0;i<32;i++) // PREAMBLE
+ mii_serial_write(1);
+ mii_serial_write(0); // ST
+ mii_serial_write(1);
+}
+
+
+/******************************************
+* Read MII register
+* phyad -> physical address
+* regad -> register address
+***************************************** */
+unsigned int mii_read(unsigned char phyad,unsigned char regad)
+{
+ unsigned int i,value;
+ unsigned int bit;
+
+ if (phyad == GPHY_ADDR)
+ {
+ GPIO_MDC_PIN = G_MDC_PIN; /* assigned MDC pin for giga PHY */
+ GPIO_MDIO_PIN = G_MDIO_PIN; /* assigned MDIO pin for giga PHY */
+ }
+ else
+ {
+ GPIO_MDC_PIN = H_MDC_PIN; /* assigned MDC pin for 10/100 PHY */
+ GPIO_MDIO_PIN = H_MDIO_PIN; /* assigned MDIO pin for 10/100 PHY */
+ }
+ GPIO_MDC = (1<<GPIO_MDC_PIN);
+ GPIO_MDIO = (1<<GPIO_MDIO_PIN);
+
+ mii_pre_st(); // PRE+ST
+ mii_serial_write(1); // OP
+ mii_serial_write(0);
+
+ for (i=0;i<5;i++) { // PHYAD
+ bit= ((phyad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ for (i=0;i<5;i++) { // REGAD
+ bit= ((regad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ mii_serial_read(); // TA_Z
+// if((bit=mii_serial_read()) !=0 ) // TA_0
+// {
+// return(0);
+// }
+ value=0;
+ for (i=0;i<16;i++) { // READ DATA
+ bit=mii_serial_read();
+ value += (bit<<(15-i)) ;
+ }
+
+ mii_serial_write(0); // dumy clock
+ mii_serial_write(0); // dumy clock
+//printk("%s: phy_addr=%x reg_addr=%x value=%x \n",__func__,phyad,regad,value);
+ return(value);
+}
+
+/******************************************
+* Write MII register
+* phyad -> physical address
+* regad -> register address
+* value -> value to be write
+***************************************** */
+void mii_write(unsigned char phyad,unsigned char regad,unsigned int value)
+{
+ unsigned int i;
+ char bit;
+
+printk("%s: phy_addr=%x reg_addr=%x value=%x \n",__func__,phyad,regad,value);
+ if (phyad == GPHY_ADDR)
+ {
+ GPIO_MDC_PIN = G_MDC_PIN; /* assigned MDC pin for giga PHY */
+ GPIO_MDIO_PIN = G_MDIO_PIN; /* assigned MDIO pin for giga PHY */
+ }
+ else
+ {
+ GPIO_MDC_PIN = H_MDC_PIN; /* assigned MDC pin for 10/100 PHY */
+ GPIO_MDIO_PIN = H_MDIO_PIN; /* assigned MDIO pin for 10/100 PHY */
+ }
+ GPIO_MDC = (1<<GPIO_MDC_PIN);
+ GPIO_MDIO = (1<<GPIO_MDIO_PIN);
+
+ mii_pre_st(); // PRE+ST
+ mii_serial_write(0); // OP
+ mii_serial_write(1);
+ for (i=0;i<5;i++) { // PHYAD
+ bit= ((phyad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ for (i=0;i<5;i++) { // REGAD
+ bit= ((regad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+ mii_serial_write(1); // TA_1
+ mii_serial_write(0); // TA_0
+
+ for (i=0;i<16;i++) { // OUT DATA
+ bit= ((value>>(15-i)) & 0x01) ? 1 : 0 ;
+ mii_serial_write(bit);
+ }
+ mii_serial_write(0); // dumy clock
+ mii_serial_write(0); // dumy clock
+}
+
+
+
+
+
+
+
+
+
+/* NOTES
+ * The instruction set of the 93C66/56/46/26/06 chips are as follows:
+ *
+ * Start OP *
+ * Function Bit Code Address** Data Description
+ * -------------------------------------------------------------------
+ * READ 1 10 A7 - A0 Reads data stored in memory,
+ * starting at specified address
+ * EWEN 1 00 11XXXXXX Write enable must precede
+ * all programming modes
+ * ERASE 1 11 A7 - A0 Erase register A7A6A5A4A3A2A1A0
+ * WRITE 1 01 A7 - A0 D15 - D0 Writes register
+ * ERAL 1 00 10XXXXXX Erase all registers
+ * WRAL 1 00 01XXXXXX D15 - D0 Writes to all registers
+ * EWDS 1 00 00XXXXXX Disables all programming
+ * instructions
+ * *Note: A value of X for address is a don't care condition.
+ * **Note: There are 8 address bits for the 93C56/66 chips unlike
+ * the 93C46/26/06 chips which have 6 address bits.
+ *
+ * The 93Cx6 has a four wire interface: clock, chip select, data in, and
+ * data out.While the ADM6996 uning three interface: clock, chip select,and data line.
+ * The input and output are the same pin. ADM6996 can only recognize the write cmd.
+ * In order to perform above functions, you need
+ * 1. to enable the chip select .
+ * 2. send one clock of dummy clock
+ * 3. send start bit and opcode
+ * 4. send 8 bits address and 16 bits data
+ * 5. to disable the chip select.
+ * Jason Lee 2003/07/30
+ */
+
+/***************************************/
+/* define GPIO module base address */
+/***************************************/
+#define GPIO_EECS 0x00400000 /* EECS: GPIO[22] */
+//#define GPIO_MOSI 0x20000000 /* EEDO: GPIO[29] send to 6996*/
+#define GPIO_MISO 0x40000000 /* EEDI: GPIO[30] receive from 6996*/
+#define GPIO_EECK 0x80000000 /* EECK: GPIO[31] */
+
+#define ADM_EECS 0x01
+#define ADM_EECK 0x02
+#define ADM_EDIO 0x04
+/*************************************************************
+* SPI protocol for ADM6996 control
+**************************************************************/
+#define SPI_OP_LEN 0x03 // the length of start bit and opcode
+#define SPI_OPWRITE 0X05 // write
+#define SPI_OPREAD 0X06 // read
+#define SPI_OPERASE 0X07 // erase
+#define SPI_OPWTEN 0X04 // write enable
+#define SPI_OPWTDIS 0X04 // write disable
+#define SPI_OPERSALL 0X04 // erase all
+#define SPI_OPWTALL 0X04 // write all
+
+#define SPI_ADD_LEN 8 // bits of Address
+#define SPI_DAT_LEN 16 // bits of Data
+#define ADM6996_PORT_NO 6 // the port number of ADM6996
+#define ADM6999_PORT_NO 9 // the port number of ADM6999
+#ifdef CONFIG_ADM_6996
+ #define ADM699X_PORT_NO ADM6996_PORT_NO
+#endif
+#ifdef CONFIG_ADM_6999
+ #define ADM699X_PORT_NO ADM6999_PORT_NO
+#endif
+#define LPC_GPIO_SET 3
+#define LPC_BASE_ADDR IO_ADDRESS(IT8712_IO_BASE)
+
+extern int it8712_exist;
+
+#define inb_gpio(x) inb(LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+#define outb_gpio(x, y) outb(y, LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+
+/****************************************/
+/* Function Declare */
+/****************************************/
+/*
+void SPI_write(unsigned char addr,unsigned int value);
+unsigned int SPI_read(unsigned char table,unsigned char addr);
+void SPI_write_bit(char bit_EEDO);
+unsigned int SPI_read_bit(void);
+void SPI_default(void);
+void SPI_reset(unsigned char rstype,unsigned char port_cnt);
+void SPI_pre_st(void);
+void SPI_CS_enable(unsigned char enable);
+void SPI_Set_VLAN(unsigned char LAN,unsigned int port_mask);
+void SPI_Set_tag(unsigned int port,unsigned tag);
+void SPI_Set_PVID(unsigned int PVID,unsigned int port_mask);
+void SPI_mac_lock(unsigned int port, unsigned char lock);
+void SPI_get_port_state(unsigned int port);
+void SPI_port_enable(unsigned int port,unsigned char enable);
+
+void SPI_get_status(unsigned int port);
+*/
+
+struct PORT_CONFIG
+{
+ unsigned char auto_negotiation; // 0:Disable 1:Enable
+ unsigned char speed; // 0:10M 1:100M
+ unsigned char duplex; // 0:Half 1:Full duplex
+ unsigned char Tag; // 0:Untag 1:Tag
+ unsigned char port_disable; // 0:port enable 1:disable
+ unsigned char pvid; // port VLAN ID 0001
+ unsigned char mdix; // Crossover judgement. 0:Disable 1:Enable
+ unsigned char mac_lock; // MAC address Lock 0:Disable 1:Enable
+};
+
+struct PORT_STATUS
+{
+ unsigned char link; // 0:not link 1:link established
+ unsigned char speed; // 0:10M 1:100M
+ unsigned char duplex; // 0:Half 1:Full duplex
+ unsigned char flow_ctl; // 0:flow control disable 1:enable
+ unsigned char mac_lock; // MAC address Lock 0:Disable 1:Enable
+ unsigned char port_disable; // 0:port enable 1:disable
+
+ // Serial Management
+ unsigned long rx_pac_count; //receive packet count
+ unsigned long rx_pac_byte; //receive packet byte count
+ unsigned long tx_pac_count; //transmit packet count
+ unsigned long tx_pac_byte; //transmit packet byte count
+ unsigned long collision_count; //error count
+ unsigned long error_count ;
+
+ unsigned long rx_pac_count_overflow; //overflow flag
+ unsigned long rx_pac_byte_overflow;
+ unsigned long tx_pac_count_overflow;
+ unsigned long tx_pac_byte_overflow;
+ unsigned long collision_count_overflow;
+ unsigned long error_count_overflow;
+};
+
+struct PORT_CONFIG port_config[ADM699X_PORT_NO]; // 0~3:LAN , 4:WAN , 5:MII
+static struct PORT_STATUS port_state[ADM699X_PORT_NO];
+
+/******************************************
+* SPI_write
+* addr -> Write Address
+* value -> value to be write
+***************************************** */
+void SPI_write(unsigned char addr,unsigned int value)
+{
+ int i;
+ char bit;
+#ifdef CONFIG_IT8712_GPIO
+ char status;
+#else
+ int ad1;
+#endif
+
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+ SPI_CS_enable(1);
+
+ SPI_write_bit(0); //dummy clock
+
+ //send write command (0x05)
+ for(i=SPI_OP_LEN-1;i>=0;i--)
+ {
+ bit = (SPI_OPWRITE>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+ // send 8 bits address (MSB first, LSB last)
+ for(i=SPI_ADD_LEN-1;i>=0;i--)
+ {
+ bit = (addr>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+ // send 16 bits data (MSB first, LSB last)
+ for(i=SPI_DAT_LEN-1;i>=0;i--)
+ {
+ bit = (value>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+
+ SPI_CS_enable(0); // CS low
+
+ for(i=0;i<0xFFF;i++) ;
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+}
+
+
+/************************************
+* SPI_write_bit
+* bit_EEDO -> 1 or 0 to be written
+************************************/
+void SPI_write_bit(char bit_EEDO)
+{
+#ifdef CONFIG_IT8712_GPIO
+ unsigned char iomode,status;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode |= (ADM_EECK|ADM_EDIO|ADM_EECS) ; // Set EECK,EDIO,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ if(bit_EEDO)
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= ADM_EDIO ; //EDIO high
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+ else
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+
+ status |= ADM_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(ADM_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+#else
+ unsigned int addr;
+ unsigned int value;
+
+ addr = (GPIO_BASE_ADDR + GPIO_PIN_DIR);
+ value = readl(addr) |GPIO_EECK |GPIO_MISO ; /* set EECK/MISO Pin to output */
+ writel(value,addr);
+ if(bit_EEDO)
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MISO,addr); /* set MISO to 1 */
+ writel(GPIO_EECK,addr); /* set EECK to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); /* set EECK to 0 */
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,addr); /* set MISO to 0 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); /* set EECK to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); /* set EECK to 0 */
+ }
+
+ return ;
+#endif
+}
+
+/**********************************************************************
+* read a bit from ADM6996 register
+***********************************************************************/
+unsigned int SPI_read_bit(void) // read data from
+{
+#ifdef CONFIG_IT8712_GPIO
+ unsigned char iomode,status;
+ unsigned int value ;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode &= ~(ADM_EDIO) ; // Set EDIO input
+ iomode |= (ADM_EECS|ADM_EECK) ; // Set EECK,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= ADM_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(ADM_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+ value = inb_gpio( LPC_GPIO_SET);
+
+ value = value>>2 ;
+ value &= 0x01;
+
+ return value ;
+#else
+ unsigned int addr;
+ unsigned int value;
+
+ addr = (GPIO_BASE_ADDR + GPIO_PIN_DIR);
+ value = readl(addr) & (~GPIO_MISO); // set EECK to output and MISO to input
+ writel(value,addr);
+
+ addr =(GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); // set EECK to 1
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); // set EECK to 0
+
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_IN);
+ value = readl(addr) ;
+ value = value >> 30;
+ return value ;
+#endif
+}
+
+/******************************************
+* SPI_default
+* EEPROM content default value
+*******************************************/
+void SPI_default(void)
+{
+ int i;
+#ifdef CONFIG_ADM_6999
+ SPI_write(0x11,0xFF30);
+ for(i=1;i<8;i++)
+ SPI_write(i,0x840F);
+
+ SPI_write(0x08,0x880F); //port 8 Untag, PVID=2
+ SPI_write(0x09,0x881D); //port 9 Tag, PVID=2 ,10M
+ SPI_write(0x14,0x017F); //Group 0~6,8 as VLAN 1
+ SPI_write(0x15,0x0180); //Group 7,8 as VLAN 2
+#endif
+
+#ifdef CONFIG_ADM_6996
+ SPI_write(0x11,0xFF30);
+ SPI_write(0x01,0x840F); //port 0~3 Untag ,PVID=1 ,100M ,duplex
+ SPI_write(0x03,0x840F);
+ SPI_write(0x05,0x840F);
+ SPI_write(0x07,0x840F);
+ SPI_write(0x08,0x880F); //port 4 Untag, PVID=2
+ SPI_write(0x09,0x881D); //port 5 Tag, PVID=2 ,10M
+ SPI_write(0x14,0x0155); //Group 0~3,5 as VLAN 1
+ SPI_write(0x15,0x0180); //Group 4,5 as VLAN 2
+
+#endif
+
+ for(i=0x16;i<=0x22;i++)
+ SPI_write((unsigned char)i,0x0000); // clean VLAN<41>@map 3~15
+
+ for (i=0;i<NUM_VLAN_IF;i++) // Set VLAN ID map 1,2
+ SPI_Set_PVID( VLAN_conf[i].vid, VLAN_conf[i].portmap);
+
+ for(i=0;i<ADM699X_PORT_NO;i++) // reset count
+ SPI_reset(0,i);
+}
+
+/*************************************************
+* SPI_reset
+* rstype -> reset type
+* 0:reset all count for 'port_cnt' port
+* 1:reset specified count 'port_cnt'
+* port_cnt -> port number or counter index
+***************************************************/
+void SPI_reset(unsigned char rstype,unsigned char port_cnt)
+{
+
+ int i;
+#ifdef CONFIG_IT8712_GPIO
+ char status;
+#else
+ int ad1;
+#endif
+ char bit;
+
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+
+ SPI_CS_enable(0); // CS low
+
+ SPI_pre_st(); // PRE+ST
+ SPI_write_bit(0); // OP
+ SPI_write_bit(1);
+
+ SPI_write_bit(1); // Table select, must be 1 -> reset Counter
+
+ SPI_write_bit(0); // Device Address
+ SPI_write_bit(0);
+
+ rstype &= 0x01;
+ SPI_write_bit(rstype); // Reset type 0:clear dedicate port's all counters 1:clear dedicate counter
+
+ for (i=5;i>=0;i--) // port or cnt index
+ {
+ bit = port_cnt >> i ;
+ bit &= 0x01 ;
+ SPI_write_bit(bit);
+ }
+
+ SPI_write_bit(0); // dumy clock
+ SPI_write_bit(0); // dumy clock
+
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+}
+
+/*****************************************************
+* SPI_pre_st
+* preambler: 32 bits '1' start bit: '01'
+*****************************************************/
+void SPI_pre_st(void)
+{
+ int i;
+
+ for(i=0;i<32;i++) // PREAMBLE
+ SPI_write_bit(1);
+ SPI_write_bit(0); // ST
+ SPI_write_bit(1);
+}
+
+
+/***********************************************************
+* SPI_CS_enable
+* before access ,you have to enable Chip Select. (pull high)
+* When fisish, you should pull low !!
+*************************************************************/
+void SPI_CS_enable(unsigned char enable)
+{
+#ifdef CONFIG_IT8712_GPIO
+
+ unsigned char iomode,status;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode |= (ADM_EECK|ADM_EDIO|ADM_EECS) ; // Set EECK,EDIO,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+
+ status = inb_gpio( LPC_GPIO_SET);
+ if(enable)
+ status |= ADM_EECS ; //EECS high
+ else
+ status &= ~(ADM_EECS) ; //EECS low
+
+ outb_gpio(LPC_GPIO_SET, status);
+
+
+ status |= ADM_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(ADM_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+#else
+ unsigned int addr,value;
+
+ addr = (GPIO_BASE_ADDR + GPIO_PIN_DIR);
+ value = readl(addr) |GPIO_EECS |GPIO_EECK; /* set EECS/EECK Pin to output */
+ writel(value,addr);
+
+ if(enable)
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_EECS,addr); /* set EECS to 1 */
+
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_EECS,addr); /* set EECS to 0 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); /* set EECK to 1 */ // at least one clock after CS low
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); /* set EECK to 0 */
+ }
+#endif
+}
+
+/*********************************************************
+* SPI_Set_VLAN: group ports as VLAN
+* LAN -> VLAN number : 0~16
+* port_mask -> ports which would group as LAN
+* ex. 0x03 = 0000 0011
+* port 0 and port 1
+*********************************************************/
+void SPI_Set_VLAN(unsigned char LAN,unsigned int port_mask)
+{
+ unsigned int i,value=0;
+ unsigned reg_add = 0x13 + LAN ;
+
+ for(i=0;i<ADM6996_PORT_NO;i++)
+ { if(port_mask&0x01)
+ {
+ switch(i)
+ {
+ case 0: value|=0x0001; break; //port0:bit[0]
+ case 1: value|=0x0004; break; //port1:bit[2]
+ case 2: value|=0x0010; break; //port2:bit[4]
+ case 3: value|=0x0040; break; //port3:bit[6]
+ case 4: value|=0x0080; break; //port4:bit[7]
+ case 5: value|=0x0100; break; //port5:bit[8]
+ }
+ }
+ port_mask >>= 1;
+ }
+
+ SPI_write(reg_add,value);
+}
+
+
+/*******************************************
+* SPI_Set_tag
+* port -> port number to set tag or untag
+* tag -> 0/set untag, 1/set tag
+* In general, tag is for MII port. LAN and
+* WAN port is configed as untag!!
+********************************************/
+void SPI_Set_tag(unsigned int port,unsigned tag)
+{
+ unsigned int regadd,value;
+
+ // mapping port's register !! (0,1,2,3,4,5) ==> (1,3,5,7,8,9)
+ if(port<=3)
+ regadd=2*port+1;
+ else if(port==4) regadd = 8 ;
+ else regadd = 9 ;
+
+
+ value = SPI_read(0,regadd); //read original setting
+
+ if(tag)
+ value |= 0x0010 ; // set tag
+ else
+ value &= 0xFFEF ; // set untag
+
+ SPI_write(regadd,value); // write back!!
+}
+
+/************************************************
+* SPI_Set_PVID
+* PVID -> PVID number :
+* port_mask -> ports which would group as LAN
+* ex. 0x0F = 0000 1111 ==> port 0~3
+************************************************/
+void SPI_Set_PVID(unsigned int PVID,unsigned int port_mask)
+{
+ unsigned int i,value=0;
+
+ PVID &= 0x000F ;
+
+ for(i=0;i<ADM699X_PORT_NO;i++)
+ { if(port_mask&0x01)
+ {
+#ifdef CONFIG_ADM_6996
+ switch(i)
+ {
+ case 0:
+ value = SPI_read(0,0x01); // read original value
+ value &= 0xC3FF ; //set PVIC column as 0 first
+ value |= PVID << 10 ; //Set PVID column as PVID
+ SPI_write(0x01,value); //write back
+ break;
+ case 1:
+ value = SPI_read(0,0x03);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write(0x03,value);
+ break;
+ case 2:
+ value = SPI_read(0,0x05);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write(0x05,value);
+ break;
+ case 3:
+ value = SPI_read(0,0x07);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write(0x07,value);
+ break;
+ case 4:
+ value = SPI_read(0,0x08);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write(0x08,value);
+ break;
+ case 5:
+ value = SPI_read(0,0x09);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write(0x09,value);
+ break;
+ }
+#endif
+#ifdef CONFIG_ADM_6999
+ value = SPI_read(0,(unsigned char)i+1);
+ value &= 0xC3FF ;
+ value |= PVID << 10 ;
+ SPI_write((unsigned char)i+1,value);
+#endif
+ }
+ port_mask >>= 1;
+ }
+}
+
+
+/************************************************
+* SPI_get_PVID
+* port -> which ports to VID
+************************************************/
+unsigned int SPI_Get_PVID(unsigned int port)
+{
+ unsigned int value=0;
+
+ if (port>=ADM6996_PORT_NO)
+ return 0;
+
+ switch(port)
+ {
+ case 0:
+ value = SPI_read(0,0x01); // read original value
+ value &= 0x3C00 ; // get VID
+ value = value >> 10 ; // Shift
+ break;
+ case 1:
+ value = SPI_read(0,0x03);
+ value &= 0x3C00 ;
+ value = value >> 10 ;
+ break;
+ case 2:
+ value = SPI_read(0,0x05);
+ value &= 0x3C00 ;
+ value = value >> 10 ;
+ break;
+ case 3:
+ value = SPI_read(0,0x07);
+ value &= 0x3C00 ;
+ value = value >> 10 ;
+ break;
+ case 4:
+ value = SPI_read(0,0x08);
+ value &= 0x3C00 ;
+ value = value >> 10 ;
+ break;
+ case 5:
+ value = SPI_read(0,0x09);
+ value &= 0x3C00 ;
+ value = value >> 10 ;
+ break;
+ }
+ return value ;
+}
+
+
+/**********************************************
+* SPI_mac_clone
+* port -> the port which will lock or unlock
+* lock -> 0/the port will be unlock
+* 1/the port will be locked
+**********************************************/
+void SPI_mac_lock(unsigned int port, unsigned char lock)
+{
+ unsigned int i,value=0;
+
+ value = SPI_read(0,0x12); // read original
+
+ for(i=0;i<ADM6996_PORT_NO;i++)
+ { if(lock) // lock port
+ {
+ switch(port)
+ {
+ case 0: value|=0x0001; break; //port0:bit[0]
+ case 1: value|=0x0004; break; //port1:bit[2]
+ case 2: value|=0x0010; break; //port2:bit[4]
+ case 3: value|=0x0040; break; //port3:bit[6]
+ case 4: value|=0x0080; break; //port4:bit[7]
+ case 5: value|=0x0100; break; //port5:bit[8]
+ }
+ }
+ else
+ {
+ switch(i) // unlock port
+ {
+ case 0: value&=0xFFFE; break;
+ case 1: value&=0xFFFB; break;
+ case 2: value&=0xFFEF; break;
+ case 3: value&=0xFFBF; break;
+ case 4: value&=0xFF7F; break;
+ case 5: value&=0xFEFF; break;
+ }
+ }
+ }
+
+ SPI_write(0x12,value);
+}
+
+
+/***************************************************
+* SPI_learn_pause
+* pause = 01-80-c2-00-00-01
+* DA=distination address
+* forward -> 0: if DA == pause then drop and stop mac learning
+* 1: if DA == pause ,then forward it
+***************************************************/
+void SPI_pause_cmd_forward(unsigned char forward)
+{
+ unsigned int value=0;
+
+ value = SPI_read(0,0x2C); // read original setting
+ if(forward)
+ value |= 0x2000; // set bit[13] '1'
+ else
+ value &= 0xDFFF; // set bit[13] '0'
+
+ SPI_write(0x2C,value);
+
+}
+
+
+/************************************************
+* SPI_read
+* table -> which table to be read: 1/count 0/EEPROM
+* addr -> Address to be read
+* return : Value of the register
+*************************************************/
+unsigned int SPI_read(unsigned char table,unsigned char addr)
+{
+ int i ;
+ unsigned int value=0;
+ unsigned int bit;
+#ifdef CONFIG_IT8712_GPIO
+ unsigned char status;
+#else
+ unsigned int ad1;
+#endif
+
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+
+ SPI_CS_enable(0);
+
+ SPI_pre_st(); // PRE+ST
+ SPI_write_bit(1); // OPCODE '10' for read
+ SPI_write_bit(0);
+
+ (table==1) ? SPI_write_bit(1) : SPI_write_bit(0) ; // table select
+
+ SPI_write_bit(0); // Device Address
+ SPI_write_bit(0);
+
+
+ // send 7 bits address to be read
+ for (i=6;i>=0;i--) {
+ bit= ((addr>>i) & 0x01) ? 1 :0 ;
+ SPI_write_bit(bit);
+ }
+
+
+ // turn around
+ SPI_read_bit(); // TA_Z
+
+ value=0;
+ for (i=31;i>=0;i--) { // READ DATA
+ bit=SPI_read_bit();
+ value |= bit << i ;
+ }
+
+ SPI_read_bit(); // dumy clock
+ SPI_read_bit(); // dumy clock
+
+ if(!table) // EEPROM, only fetch 16 bits data
+ {
+ if(addr&0x01) // odd number content (register,register-1)
+ value >>= 16 ; // so we remove the rear 16bits
+ else // even number content (register+1,register),
+ value &= 0x0000FFFF ; // so we keep the rear 16 bits
+ }
+
+
+ SPI_CS_enable(0);
+
+#ifdef CONFIG_IT8712_GPIO
+ status = inb_gpio(LPC_GPIO_SET);
+ status &= ~(ADM_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+#else
+ ad1 = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MISO,ad1); /* set MISO to 0 */
+#endif
+
+ return(value);
+
+}
+
+
+
+/**************************************************
+* SPI_port_en
+* port -> Number of port to config
+* enable -> 1/ enable this port
+* 0/ disable this port
+**************************************************/
+void SPI_port_enable(unsigned int port,unsigned char enable)
+{
+ unsigned int reg_val ;
+ unsigned char reg_add ;
+
+ if(port<=3)
+ reg_add=2*port+1;
+ else if(port==4) reg_add = 8 ;
+ else reg_add = 9 ;
+
+ reg_val = SPI_read(0,reg_add);
+ if(enable)
+ {
+ reg_val &= 0xFFDF ;
+ SPI_write(reg_add,reg_val);
+ }
+ else
+ {
+ reg_val |= 0x0020 ;
+ SPI_write(reg_add,reg_val);
+ }
+}
+
+/********************************************************
+* get port status
+* port -> specify the port number to get configuration
+*********************************************************/
+void SPI_get_status(unsigned int port)
+{
+/* unsigned int reg_val,add_offset[6];
+ struct PORT_STATUS *status;
+ status = &port_state[port];
+
+ if(port>(ADM6996_PORT_NO-1))
+ return ;
+
+ // Link estabilish , speed, deplex, flow control ?
+ if(port < 5 )
+ {
+ reg_val = SPI_read(1, 1) ;
+ if(port < 4)
+ reg_val >>= port*8 ;
+ else
+ reg_val >>=28 ;
+ status->link = reg_val & 0x00000001 ;
+ status->speed = reg_val & 0x00000002 ;
+ status->duplex = reg_val & 0x00000004 ;
+ status->flow_ctl = reg_val & 0x00000008 ;
+ }
+ else if(port ==5 )
+ {
+ reg_val = SPI_read(1, 2) ;
+ status->link = reg_val & 0x00000001 ;
+ status->speed = reg_val & 0x00000002 ;
+ status->duplex = reg_val & 0x00000008 ;
+ status->flow_ctl = reg_val & 0x00000010 ;
+ }
+
+ // Mac Lock ?
+ reg_val = SPI_read(0,0x12);
+ switch(port)
+ {
+ case 0: status->mac_lock = reg_val & 0x00000001;
+ case 1: status->mac_lock = reg_val & 0x00000004;
+ case 2: status->mac_lock = reg_val & 0x00000010;
+ case 3: status->mac_lock = reg_val & 0x00000040;
+ case 4: status->mac_lock = reg_val & 0x00000080;
+ case 5: status->mac_lock = reg_val & 0x00000100;
+ }
+
+ // port enable ?
+ add_offset[0] = 0x01 ; add_offset[1] = 0x03 ;
+ add_offset[2] = 0x05 ; add_offset[3] = 0x07 ;
+ add_offset[4] = 0x08 ; add_offset[5] = 0x09 ;
+ reg_val = SPI_read(0,add_offset[port]);
+ status->port_disable = reg_val & 0x0020;
+
+
+ // Packet Count ...
+ add_offset[0] = 0x04 ; add_offset[1] = 0x06 ;
+ add_offset[2] = 0x08 ; add_offset[3] = 0x0a ;
+ add_offset[4] = 0x0b ; add_offset[5] = 0x0c ;
+
+ reg_val = SPI_read(1,add_offset[port]);
+ status->rx_pac_count = reg_val ;
+ reg_val = SPI_read(1,add_offset[port]+9);
+ status->rx_pac_byte = reg_val ;
+ reg_val = SPI_read(1,add_offset[port]+18);
+ status->tx_pac_count = reg_val ;
+ reg_val = SPI_read(1,add_offset[port]+27);
+ status->tx_pac_byte = reg_val ;
+ reg_val = SPI_read(1,add_offset[port]+36);
+ status->collision_count = reg_val ;
+ reg_val = SPI_read(1,add_offset[port]+45);
+ status->error_count = reg_val ;
+ reg_val = SPI_read(1, 0x3A);
+ switch(port)
+ {
+ case 0: status->rx_pac_count_overflow = reg_val & 0x00000001;
+ status->rx_pac_byte_overflow = reg_val & 0x00000200 ;
+ case 1: status->rx_pac_count_overflow = reg_val & 0x00000004;
+ status->rx_pac_byte_overflow = reg_val & 0x00000800 ;
+ case 2: status->rx_pac_count_overflow = reg_val & 0x00000010;
+ status->rx_pac_byte_overflow = reg_val & 0x00002000 ;
+ case 3: status->rx_pac_count_overflow = reg_val & 0x00000040;;
+ status->rx_pac_byte_overflow = reg_val & 0x00008000 ;
+ case 4: status->rx_pac_count_overflow = reg_val & 0x00000080;
+ status->rx_pac_byte_overflow = reg_val & 0x00010000 ;
+ case 5: status->rx_pac_count_overflow = reg_val & 0x00000100;
+ status->rx_pac_byte_overflow = reg_val & 0x00020000 ;
+ }
+
+ reg_val = SPI_read(1, 0x3B);
+ switch(port)
+ {
+ case 0: status->tx_pac_count_overflow = reg_val & 0x00000001;
+ status->tx_pac_byte_overflow = reg_val & 0x00000200 ;
+ case 1: status->tx_pac_count_overflow = reg_val & 0x00000004;
+ status->tx_pac_byte_overflow = reg_val & 0x00000800 ;
+ case 2: status->tx_pac_count_overflow = reg_val & 0x00000010;
+ status->tx_pac_byte_overflow = reg_val & 0x00002000 ;
+ case 3: status->tx_pac_count_overflow = reg_val & 0x00000040;;
+ status->tx_pac_byte_overflow = reg_val & 0x00008000 ;
+ case 4: status->tx_pac_count_overflow = reg_val & 0x00000080;
+ status->tx_pac_byte_overflow = reg_val & 0x00010000 ;
+ case 5: status->tx_pac_count_overflow = reg_val & 0x00000100;
+ status->tx_pac_byte_overflow = reg_val & 0x00020000 ;
+ }
+*/
+
+ unsigned int reg_val;
+ struct PORT_STATUS *status;
+ status = &port_state[port];
+
+ if(port>=ADM6999_PORT_NO)
+ return ;
+
+ // Link estabilish , speed, deplex, flow control ?
+ if(port < ADM6999_PORT_NO-1 )
+ {
+ reg_val = SPI_read(1, 0x01) ;
+ reg_val = reg_val >> port*4 ;
+ status->link = reg_val & 0x00000001 ;
+ status->speed = reg_val & 0x00000002 ;
+ status->duplex = reg_val & 0x00000004 ;
+ status->flow_ctl = reg_val & 0x00000008 ;
+ }
+ else if(port == (ADM6999_PORT_NO-1) )
+ {
+ reg_val = SPI_read(1, 0x02) ;
+ status->link = reg_val & 0x00000001 ;
+ status->speed = reg_val & 0x00000002 ;
+ status->duplex = reg_val & 0x00000008 ;
+ status->flow_ctl = reg_val & 0x00000010 ;
+ }
+
+ // Mac Lock ?
+ reg_val = SPI_read(0,0x12);
+ reg_val = reg_val >> port ;
+ reg_val = reg_val & 0x01 ;
+ status->mac_lock = reg_val ? 0x01:0x00 ;
+
+ // port enable ?
+ reg_val = SPI_read(0,(unsigned char)port+1);
+ status->port_disable = reg_val & 0x0020;
+
+ // Packet Count ...
+ reg_val = SPI_read(1,(unsigned char)port+0x04);
+ status->rx_pac_count = reg_val ;
+ reg_val = SPI_read(1,(unsigned char)port+0x0D);
+ status->rx_pac_byte = reg_val ;
+ reg_val = SPI_read(1,(unsigned char)port+0x16);
+ status->tx_pac_count = reg_val ;
+ reg_val = SPI_read(1,(unsigned char)port+0x1F);
+ status->tx_pac_byte = reg_val ;
+ reg_val = SPI_read(1,(unsigned char)port+0x28);
+ status->collision_count = reg_val ;
+ reg_val = SPI_read(1,(unsigned char)port+0x31);
+ status->error_count = reg_val ;
+ reg_val = SPI_read(1, 0x3A);
+ reg_val = reg_val >> port ;
+ status->rx_pac_count_overflow = reg_val & 0x00000001;
+ reg_val = reg_val >> 0x09 ;
+ status->rx_pac_byte_overflow = reg_val & 0x00000001 ;
+
+ reg_val = SPI_read(1, 0x3B);
+ reg_val = reg_val >> port ;
+ status->tx_pac_count_overflow = reg_val & 0x00000001;
+ reg_val = reg_val >> 0x09 ;
+ status->tx_pac_byte_overflow = reg_val & 0x00000001 ;
+
+ reg_val = SPI_read(1, 0x3C);
+ reg_val = reg_val >> port ;
+ status->collision_count_overflow = reg_val & 0x00000001;
+ reg_val = reg_val >> 0x09 ;
+ status->error_count_overflow = reg_val & 0x00000001 ;
+
+}
+
+unsigned int SPI_get_identifier(void)
+{
+ unsigned int flag=0;
+
+#ifdef CONFIG_IT8712_GPIO
+
+ if (!it8712_exist) {
+ return -ENODEV;
+ }
+ printk("it8712_gpio init\n");
+
+ /* initialize registers */
+ // switch all multi-function pins to GPIO
+ LPCSetConfig(LDN_GPIO, 0x28, 0xff);
+
+ // set simple I/O base address
+ LPCSetConfig(LDN_GPIO, 0x62, IT8712_GPIO_BASE >> 8);
+ LPCSetConfig(LDN_GPIO, 0x63, (unsigned char) IT8712_GPIO_BASE >> 8);
+
+ // select GPIO to simple I/O
+ LPCSetConfig(LDN_GPIO, 0xc3, 0xff);
+
+ // enable internal pull-up
+ LPCSetConfig(LDN_GPIO, 0xbb, 0xff);
+
+#endif
+
+ flag = SPI_read(1,0x00);
+ printk("Get ADM identifier %6x\n",flag);
+ if ((flag & 0xFFFF0) == 0x21120) {
+ printk("ADM699X Found\n");
+ return 1;
+ }
+ else {
+ printk("ADM699X not Found\n");
+ return 0;
+ }
+}
+
--- /dev/null
+++ b/drivers/net/sl351x_crc16.c
@@ -0,0 +1,93 @@
+/****************************************************************************
+* Name : sl351x_crc16.c
+* Description :
+* Implement CRC16
+* refer to RFC1662
+* History
+*
+* Date Writer Description
+* ----------- ----------- -------------------------------------------------
+* 09/14/2005 Gary Chen Create
+*
+****************************************************************************/
+
+#define INITFCS16 0xffff /* Initial FCS value */
+#define GOODFCS16 0xf0b8 /* Good final FCS value */
+#define SWAP_WORD(x) (unsigned short)((((unsigned short)x & 0x00FF) << 8) | \
+ (((unsigned short)x & 0xFF00) >> 8))
+
+/*----------------------------------------------------------------------
+* x**0 + x**5 + x**12 + x**16
+*----------------------------------------------------------------------*/
+static const unsigned short crc16_tbl[256] = {
+ 0x0000, 0x1189, 0x2312, 0x329b, 0x4624, 0x57ad, 0x6536, 0x74bf,
+ 0x8c48, 0x9dc1, 0xaf5a, 0xbed3, 0xca6c, 0xdbe5, 0xe97e, 0xf8f7,
+ 0x1081, 0x0108, 0x3393, 0x221a, 0x56a5, 0x472c, 0x75b7, 0x643e,
+ 0x9cc9, 0x8d40, 0xbfdb, 0xae52, 0xdaed, 0xcb64, 0xf9ff, 0xe876,
+ 0x2102, 0x308b, 0x0210, 0x1399, 0x6726, 0x76af, 0x4434, 0x55bd,
+ 0xad4a, 0xbcc3, 0x8e58, 0x9fd1, 0xeb6e, 0xfae7, 0xc87c, 0xd9f5,
+ 0x3183, 0x200a, 0x1291, 0x0318, 0x77a7, 0x662e, 0x54b5, 0x453c,
+ 0xbdcb, 0xac42, 0x9ed9, 0x8f50, 0xfbef, 0xea66, 0xd8fd, 0xc974,
+ 0x4204, 0x538d, 0x6116, 0x709f, 0x0420, 0x15a9, 0x2732, 0x36bb,
+ 0xce4c, 0xdfc5, 0xed5e, 0xfcd7, 0x8868, 0x99e1, 0xab7a, 0xbaf3,
+ 0x5285, 0x430c, 0x7197, 0x601e, 0x14a1, 0x0528, 0x37b3, 0x263a,
+ 0xdecd, 0xcf44, 0xfddf, 0xec56, 0x98e9, 0x8960, 0xbbfb, 0xaa72,
+ 0x6306, 0x728f, 0x4014, 0x519d, 0x2522, 0x34ab, 0x0630, 0x17b9,
+ 0xef4e, 0xfec7, 0xcc5c, 0xddd5, 0xa96a, 0xb8e3, 0x8a78, 0x9bf1,
+ 0x7387, 0x620e, 0x5095, 0x411c, 0x35a3, 0x242a, 0x16b1, 0x0738,
+ 0xffcf, 0xee46, 0xdcdd, 0xcd54, 0xb9eb, 0xa862, 0x9af9, 0x8b70,
+ 0x8408, 0x9581, 0xa71a, 0xb693, 0xc22c, 0xd3a5, 0xe13e, 0xf0b7,
+ 0x0840, 0x19c9, 0x2b52, 0x3adb, 0x4e64, 0x5fed, 0x6d76, 0x7cff,
+ 0x9489, 0x8500, 0xb79b, 0xa612, 0xd2ad, 0xc324, 0xf1bf, 0xe036,
+ 0x18c1, 0x0948, 0x3bd3, 0x2a5a, 0x5ee5, 0x4f6c, 0x7df7, 0x6c7e,
+ 0xa50a, 0xb483, 0x8618, 0x9791, 0xe32e, 0xf2a7, 0xc03c, 0xd1b5,
+ 0x2942, 0x38cb, 0x0a50, 0x1bd9, 0x6f66, 0x7eef, 0x4c74, 0x5dfd,
+ 0xb58b, 0xa402, 0x9699, 0x8710, 0xf3af, 0xe226, 0xd0bd, 0xc134,
+ 0x39c3, 0x284a, 0x1ad1, 0x0b58, 0x7fe7, 0x6e6e, 0x5cf5, 0x4d7c,
+ 0xc60c, 0xd785, 0xe51e, 0xf497, 0x8028, 0x91a1, 0xa33a, 0xb2b3,
+ 0x4a44, 0x5bcd, 0x6956, 0x78df, 0x0c60, 0x1de9, 0x2f72, 0x3efb,
+ 0xd68d, 0xc704, 0xf59f, 0xe416, 0x90a9, 0x8120, 0xb3bb, 0xa232,
+ 0x5ac5, 0x4b4c, 0x79d7, 0x685e, 0x1ce1, 0x0d68, 0x3ff3, 0x2e7a,
+ 0xe70e, 0xf687, 0xc41c, 0xd595, 0xa12a, 0xb0a3, 0x8238, 0x93b1,
+ 0x6b46, 0x7acf, 0x4854, 0x59dd, 0x2d62, 0x3ceb, 0x0e70, 0x1ff9,
+ 0xf78f, 0xe606, 0xd49d, 0xc514, 0xb1ab, 0xa022, 0x92b9, 0x8330,
+ 0x7bc7, 0x6a4e, 0x58d5, 0x495c, 0x3de3, 0x2c6a, 0x1ef1, 0x0f78
+};
+
+/*----------------------------------------------------------------------
+* hash_crc16
+*----------------------------------------------------------------------*/
+unsigned short hash_crc16(unsigned short crc, unsigned char *datap, unsigned long len)
+{
+ while (len--)
+ {
+ crc = (crc >> 8) ^ crc16_tbl[(crc ^ (*datap++)) & 0xff];
+ }
+
+ return (crc);
+
+}
+
+/*----------------------------------------------------------------------
+* hash_check_crc16
+*----------------------------------------------------------------------*/
+unsigned long hash_check_crc16(unsigned char *datap, unsigned long len)
+{
+ unsigned short crc;
+
+ crc = hash_crc16(INITFCS16, datap, len );
+ return (crc == GOODFCS16) ? 0 : 1;
+}
+
+/*----------------------------------------------------------------------
+* hash_gen_crc16
+*----------------------------------------------------------------------*/
+unsigned short hash_gen_crc16(unsigned char *datap, unsigned long len)
+{
+ unsigned short crc;
+
+ crc = hash_crc16(INITFCS16, datap, len);
+ crc ^= 0xffff;
+
+ return(SWAP_WORD(crc));
+}
--- /dev/null
+++ b/drivers/net/sl351x_gmac.c
@@ -0,0 +1,5622 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl351x_gmac.c
+* Description :
+* Ethernet device driver for Storlink SL351x FPGA
+*
+* History
+*
+* Date Writer Description
+* ----------- ----------- -------------------------------------------------
+* 08/22/2005 Gary Chen Create and implement
+* 27/10/2005 CH Hsu Porting to Linux
+*
+****************************************************************************/
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/slab.h>
+#include <linux/mm.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch/irqs.h>
+#include <asm/arch/it8712.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/skbuff.h>
+#include <linux/in.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#include <linux/udp.h>
+
+#include <linux/mtd/kvctl.h>
+
+#define MIDWAY
+#define SL_LEPUS
+#define VITESSE_G5SWITCH 1
+
+#ifndef CONFIG_SL351x_RXTOE
+//#define CONFIG_SL351x_RXTOE 1
+#endif
+#undef CONFIG_SL351x_RXTOE
+
+#include <asm/arch/sl2312.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <asm/arch/sl351x_hash_cfg.h>
+#include <asm/arch/sl351x_nat_cfg.h>
+
+#ifdef CONFIG_SL351x_SYSCTL
+#include <linux/sysctl_storlink.h>
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+#include <asm/arch/sl351x_toe.h>
+#include <net/tcp.h>
+#include <linux/tcp.h>
+#include <linux/ip.h>
+#endif
+
+// #define SL351x_TEST_WORKAROUND
+#ifdef CONFIG_SL351x_NAT
+#define CONFIG_SL_NAPI 1
+#endif
+#define GMAX_TX_INTR_DISABLED 1
+#define DO_HW_CHKSUM 1
+#define ENABLE_TSO 1
+#define GMAC_USE_TXQ0 1
+// #define NAT_WORKAROUND_BY_RESET_GMAC 1
+// #define HW_RXBUF_BY_KMALLOC 1
+//#define _DUMP_TX_TCP_CONTENT 1
+#define br_if_ioctl 1
+#define GMAC_LEN_1_2_ISSUE 1
+
+#define GMAC_EXISTED_FLAG 0x5566abcd
+#define CONFIG_MAC_NUM GMAC_NUM
+#define GMAC0_BASE TOE_GMAC0_BASE
+#define GMAC1_BASE TOE_GMAC1_BASE
+#define PAUSE_SET_HW_FREEQ (TOE_HW_FREEQ_DESC_NUM / 2)
+#define PAUSE_REL_HW_FREEQ ((TOE_HW_FREEQ_DESC_NUM / 2) + 10)
+#define DEFAULT_RXQ_MAX_CNT 256
+#ifdef L2_jumbo_frame
+#define TCPHDRLEN(tcp_hdr) ((ntohs(*((__u16 *)tcp_hdr + 6)) >> 12) & 0x000F)
+#endif
+
+/* define chip information */
+#define DRV_NAME "SL351x"
+#define DRV_VERSION "0.1.4"
+#define SL351x_DRIVER_NAME DRV_NAME " Giga Ethernet driver " DRV_VERSION
+
+#define toe_gmac_enable_interrupt(irq) enable_irq(irq)
+#define toe_gmac_disable_interrupt(irq) disable_irq(irq)
+
+#ifdef SL351x_GMAC_WORKAROUND
+#define GMAC_SHORT_FRAME_THRESHOLD 10
+static struct timer_list gmac_workround_timer_obj;
+void sl351x_poll_gmac_hanged_status(u32 data);
+#ifdef CONFIG_SL351x_NAT
+//#define IxscriptMate_1518 1
+ void sl351x_nat_workaround_init(void);
+ #ifndef NAT_WORKAROUND_BY_RESET_GMAC
+ static void sl351x_nat_workaround_handler(void);
+ #endif
+#endif
+#endif
+
+#ifdef GMAC_LEN_1_2_ISSUE
+ #define _DEBUG_PREFETCH_NUM 256
+static int _debug_prefetch_cnt;
+static char _debug_prefetch_buf[_DEBUG_PREFETCH_NUM][4] __attribute__((aligned(4)));
+#endif
+/*************************************************************
+ * Global Variable
+ *************************************************************/
+static int gmac_initialized = 0;
+TOE_INFO_T toe_private_data;
+//static int do_again = 0;
+spinlock_t gmac_fq_lock;
+unsigned int FLAG_SWITCH;
+
+static unsigned int next_tick = 3 * HZ;
+static unsigned char eth_mac[CONFIG_MAC_NUM][6]= {{0x00,0x11,0x11,0x87,0x87,0x87}, {0x00,0x22,0x22,0xab,0xab,0xab}};
+
+#undef CONFIG_SL351x_RXTOE
+extern NAT_CFG_T nat_cfg;
+
+/************************************************/
+/* function declare */
+/************************************************/
+static int gmac_set_mac_address(struct net_device *dev, void *addr);
+static unsigned int gmac_get_phy_vendor(int phy_addr);
+static void gmac_set_phy_status(struct net_device *dev);
+void gmac_get_phy_status(struct net_device *dev);
+static int gmac_netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
+static void gmac_tx_timeout(struct net_device *dev);
+static int gmac_phy_thread (void *data);
+struct net_device_stats * gmac_get_stats(struct net_device *dev);
+static int gmac_start_xmit(struct sk_buff *skb, struct net_device *dev);
+static void gmac_set_rx_mode(struct net_device *dev);
+static irqreturn_t toe_gmac_interrupt (int irq, void *dev_instance);
+static void toe_gmac_handle_default_rxq(struct net_device *dev, GMAC_INFO_T *tp);
+unsigned int mii_read(unsigned char phyad,unsigned char regad);
+void mii_write(unsigned char phyad,unsigned char regad,unsigned int value);
+void mac_init_drv(void);
+
+static void toe_init_free_queue(void);
+static void toe_init_swtx_queue(void);
+static void toe_init_default_queue(void);
+#ifdef CONFIG_SL351x_RXTOE
+static void toe_init_interrupt_queue(void);
+#endif
+static void toe_init_interrupt_config(void);
+static void toe_gmac_sw_reset(void);
+static int toe_gmac_init_chip(struct net_device *dev);
+static void toe_gmac_enable_tx_rx(struct net_device* dev);
+static void toe_gmac_disable_tx_rx(struct net_device *dev);
+static void toe_gmac_hw_start(struct net_device *dev);
+static void toe_gmac_hw_stop(struct net_device *dev);
+static int toe_gmac_clear_counter(struct net_device *dev);
+static void toe_init_gmac(struct net_device *dev);
+static void toe_gmac_tx_complete(GMAC_INFO_T *tp, unsigned int tx_qid, struct net_device *dev, int interrupt);
+#ifdef CONFIG_SL_NAPI
+static int gmac_rx_poll(struct net_device *dev, int *budget);
+// static void toe_gmac_disable_rx(struct net_device *dev);
+// static void toe_gmac_enable_rx(struct net_device *dev);
+#endif
+
+u32 mac_read_dma_reg(int mac, unsigned int offset);
+void mac_write_dma_reg(int mac, unsigned int offset, u32 data);
+void mac_stop_txdma(struct net_device *dev);
+void mac_get_sw_tx_weight(struct net_device *dev, char *weight);
+void mac_set_sw_tx_weight(struct net_device *dev, char *weight);
+void mac_get_hw_tx_weight(struct net_device *dev, char *weight);
+void mac_set_hw_tx_weight(struct net_device *dev, char *weight);
+static inline void toe_gmac_fill_free_q(void);
+
+#ifdef VITESSE_G5SWITCH
+extern int Get_Set_port_status(void);
+extern int SPI_default(void);
+extern unsigned int SPI_get_identifier(void);
+void gmac_get_switch_status(struct net_device *dev);
+unsigned int Giga_switch=0;
+unsigned int switch_port_no=0;
+unsigned int ever_dwon=0;
+#endif
+
+/************************************************/
+/* GMAC function declare */
+/************************************************/
+static int gmac_open (struct net_device *dev);
+static int gmac_close (struct net_device *dev);
+static void gmac_cleanup_module(void);
+static void gmac_get_mac_address(void);
+
+#ifdef CONFIG_SL351x_NAT
+static void toe_init_hwtx_queue(void);
+extern void sl351x_nat_init(void);
+extern void sl351x_nat_input(struct sk_buff *skb, int port, void *l3off, void *l4off);
+extern int sl351x_nat_output(struct sk_buff *skb, int port);
+extern int sl351x_nat_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+extern void set_toeq_hdr(struct toe_conn* connection, TOE_INFO_T* toe, struct net_device *dev);
+extern void sl351x_toe_init(void);
+extern void toe_gmac_handle_toeq(struct net_device *dev, GMAC_INFO_T* tp, __u32 status);
+extern struct toe_conn* init_toeq(int ipver, void* iph, struct tcphdr* tcp_hdr, TOE_INFO_T* toe, unsigned char* l2hdr);
+#endif
+
+int mac_set_rule_reg(int mac, int rule, int enabled, u32 reg0, u32 reg1, u32 reg2);
+void mac_set_rule_enable_bit(int mac, int rule, int data);
+int mac_set_rule_action(int mac, int rule, int data);
+int mac_get_MRxCRx(int mac, int rule, int ctrlreg);
+void mac_set_MRxCRx(int mac, int rule, int ctrlreg, u32 data);
+
+/*----------------------------------------------------------------------
+* Ethernet Driver init
+*----------------------------------------------------------------------*/
+
+static int __init gmac_init_module(void)
+{
+ GMAC_INFO_T *tp;
+ struct net_device *dev;
+ int i,j;
+ unsigned int chip_id;
+// unsigned int chip_version;
+
+#ifdef CONFIG_SL3516_ASIC
+{
+ unsigned int val;
+ /* set GMAC global register */
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x10);
+ val = val | 0x005f0000;
+ writel(val,GMAC_GLOBAL_BASE_ADDR+0x10);
+// writel(0xb737b737,GMAC_GLOBAL_BASE_ADDR+0x1c); //For Socket Board
+ writel(0x77777777,GMAC_GLOBAL_BASE_ADDR+0x20);
+// writel(0xa737b747,GMAC_GLOBAL_BASE_ADDR+0x1c);//For Mounting Board
+
+ //debug_Aaron
+ //writel(0xa7f0a7f0,GMAC_GLOBAL_BASE_ADDR+0x1c);//For Mounting Board
+ writel(0xa7f0b7f0,GMAC_GLOBAL_BASE_ADDR+0x1c);//For Mounting Board
+
+ writel(0x77777777,GMAC_GLOBAL_BASE_ADDR+0x24);
+ writel(0x09200030,GMAC_GLOBAL_BASE_ADDR+0x2C);
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x04);
+ if((val&(1<<20))==0){ // GMAC1 enable
+ val = readl(GMAC_GLOBAL_BASE_ADDR+0x30);
+ val = (val & 0xe7ffffff) | 0x08000000;
+ writel(val,GMAC_GLOBAL_BASE_ADDR+0x30);
+ }
+}
+#endif
+
+#ifdef VITESSE_G5SWITCH
+ Giga_switch = SPI_get_identifier();
+ if(Giga_switch)
+ switch_port_no = SPI_default();
+#endif
+
+ chip_id = readl(GMAC_GLOBAL_BASE_ADDR+0x0);
+ if (chip_id == 0x3512C1)
+ {
+ writel(0x5787a5f0,GMAC_GLOBAL_BASE_ADDR+0x1c);//For 3512 Switch Board
+ writel(0x55557777,GMAC_GLOBAL_BASE_ADDR+0x20);//For 3512 Switch Board
+ }
+//#endif
+
+ mac_init_drv();
+
+ printk (KERN_INFO SL351x_DRIVER_NAME " built at %s %s\n", __DATE__, __TIME__);
+
+// init_waitqueue_entry(&wait, current);
+
+ // printk("GMAC Init......\n");
+
+ i = 0;
+ for(j = 0; i<CONFIG_MAC_NUM; j++)
+ {
+ i=j;
+ if(Giga_switch){ // if gswitch present, swap eth0/1
+ if(j==0)
+ i=1;
+ else if(j==1)
+ i=0;
+ }
+
+ tp = (GMAC_INFO_T *)&toe_private_data.gmac[i];
+ tp->dev = NULL;
+ if (tp->existed != GMAC_EXISTED_FLAG) continue;
+
+ dev = alloc_etherdev(0);
+ if (dev == NULL)
+ {
+ printk (KERN_ERR "Can't allocate ethernet device #%d .\n",i);
+ return -ENOMEM;
+ }
+
+ dev->priv=tp;
+ tp->dev = dev;
+
+ SET_MODULE_OWNER(dev);
+
+ // spin_lock_init(&tp->lock);
+ spin_lock_init(&gmac_fq_lock);
+ dev->base_addr = tp->base_addr;
+ dev->irq = tp->irq;
+ dev->open = gmac_open;
+ dev->stop = gmac_close;
+ dev->hard_start_xmit = gmac_start_xmit;
+ dev->get_stats = gmac_get_stats;
+ dev->set_multicast_list = gmac_set_rx_mode;
+ dev->set_mac_address = gmac_set_mac_address;
+ dev->do_ioctl = gmac_netdev_ioctl;
+ dev->tx_timeout = gmac_tx_timeout;
+ dev->watchdog_timeo = GMAC_DEV_TX_TIMEOUT;
+#ifdef L2_jumbo_frame
+ dev->mtu = 2018; //2002 ,2018
+#endif
+ if (tp->port_id == 0)
+ dev->tx_queue_len = TOE_GMAC0_SWTXQ_DESC_NUM;
+ else
+ dev->tx_queue_len = TOE_GMAC1_SWTXQ_DESC_NUM;
+
+#ifdef DO_HW_CHKSUM
+ dev->features |= NETIF_F_SG|NETIF_F_HW_CSUM;
+#ifdef ENABLE_TSO
+ dev->features |= NETIF_F_TSO;
+#endif
+#endif
+#ifdef CONFIG_SL_NAPI
+ dev->poll = gmac_rx_poll;
+ dev->weight = 64;
+#endif
+
+ if (register_netdev(dev))
+ {
+ gmac_cleanup_module();
+ return(-1);
+ }
+ }
+
+
+// FLAG_SWITCH = 0 ;
+// FLAG_SWITCH = SPI_get_identifier();
+// if(FLAG_SWITCH)
+// {
+// printk("Configure ADM699X...\n");
+// SPI_default(); //Add by jason for ADM699X configuration
+// }
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* gmac_cleanup_module
+*----------------------------------------------------------------------*/
+
+static void gmac_cleanup_module(void)
+{
+ int i;
+
+#ifdef SL351x_GMAC_WORKAROUND
+ del_timer(&gmac_workround_timer_obj);
+#endif
+
+ for (i=0;i<CONFIG_MAC_NUM;i++)
+ {
+ if (toe_private_data.gmac[i].dev)
+ {
+ unregister_netdev(toe_private_data.gmac[i].dev);
+ toe_private_data.gmac[i].dev = NULL;
+ }
+ }
+ return ;
+}
+
+module_init(gmac_init_module);
+module_exit(gmac_cleanup_module);
+
+
+/*----------------------------------------------------------------------
+* gmac_read_reg
+*----------------------------------------------------------------------*/
+static inline unsigned int gmac_read_reg(unsigned int base, unsigned int offset)
+//static unsigned int gmac_read_reg(unsigned int base, unsigned int offset)
+{
+ volatile unsigned int reg_val;
+
+ reg_val = readl(base + offset);
+ return (reg_val);
+}
+
+/*----------------------------------------------------------------------
+* gmac_write_reg
+*----------------------------------------------------------------------*/
+static inline void gmac_write_reg(unsigned int base, unsigned int offset,unsigned int data,unsigned int bit_mask)
+//static void gmac_write_reg(unsigned int base, unsigned int offset,unsigned int data,unsigned int bit_mask)
+{
+ volatile unsigned int reg_val;
+ unsigned int *addr;
+
+ reg_val = ( gmac_read_reg(base, offset) & (~bit_mask) ) | (data & bit_mask);
+ addr = (unsigned int *)(base + offset);
+ writel(reg_val,addr);
+ return;
+}
+
+/*----------------------------------------------------------------------
+* mac_init_drv
+*----------------------------------------------------------------------*/
+void mac_init_drv(void)
+{
+ TOE_INFO_T *toe;
+ int i;
+ QUEUE_THRESHOLD_T threshold;
+ u32 *destp;
+ unsigned int chip_id,chip_version;
+
+ chip_id = readl(GMAC_GLOBAL_BASE_ADDR+0x0);
+ chip_version = chip_id & 0x1 ;
+
+ if (!gmac_initialized)
+ {
+ gmac_initialized = 1;
+
+ // clear non TOE Queue Header Area
+ destp = (u32 *)TOE_NONTOE_QUE_HDR_BASE;
+ for (; destp < (u32 *)NONTOE_Q_HDR_AREA_END; destp++)
+ *destp = 0x00;
+
+ // clear TOE Queue Header Area
+ destp = (u32 *)TOE_TOE_QUE_HDR_BASE;
+ for (; destp < (u32 *)TOE_Q_HDR_AREA_END; destp++)
+ *destp = 0x00;
+
+ // init private data
+ toe = (TOE_INFO_T *)&toe_private_data;
+ memset((void *)toe, 0, sizeof(TOE_INFO_T));
+ toe->gmac[0].base_addr = GMAC0_BASE;
+ toe->gmac[1].base_addr = GMAC1_BASE;
+ toe->gmac[0].dma_base_addr = TOE_GMAC0_DMA_BASE;
+ toe->gmac[1].dma_base_addr = TOE_GMAC1_DMA_BASE;
+ toe->gmac[0].auto_nego_cfg = 1;
+ toe->gmac[1].auto_nego_cfg = 1;
+#ifdef CONFIG_SL3516_ASIC
+ toe->gmac[0].speed_cfg = GMAC_SPEED_1000;
+ toe->gmac[1].speed_cfg = GMAC_SPEED_1000;
+#else
+ toe->gmac[0].speed_cfg = GMAC_SPEED_100;
+ toe->gmac[1].speed_cfg = GMAC_SPEED_100;
+#endif
+ toe->gmac[0].full_duplex_cfg = 1;
+ toe->gmac[1].full_duplex_cfg = 1;
+#ifdef CONFIG_SL3516_ASIC
+ toe->gmac[0].phy_mode = GMAC_PHY_RGMII_1000;
+ toe->gmac[1].phy_mode = GMAC_PHY_RGMII_1000;
+#else
+ toe->gmac[0].phy_mode = GMAC_PHY_RGMII_100;
+ toe->gmac[1].phy_mode = GMAC_PHY_RGMII_100;
+#endif
+ toe->gmac[0].port_id = GMAC_PORT0;
+ toe->gmac[1].port_id = GMAC_PORT1;
+ toe->gmac[0].phy_addr = 0x1;
+ toe->gmac[1].phy_addr = 2;
+// toe->gmac[0].irq = SL2312_INTERRUPT_GMAC0;
+ toe->gmac[0].irq =1;
+// toe->gmac[1].irq = SL2312_INTERRUPT_GMAC1;
+ toe->gmac[1].irq =2;
+ toe->gmac[0].mac_addr1 = &eth_mac[0][0];
+ toe->gmac[1].mac_addr1 = &eth_mac[1][0];
+
+ for (i=0; i<CONFIG_MAC_NUM; i++)
+ {
+ unsigned int data, phy_vendor;
+ gmac_write_reg(toe->gmac[i].base_addr, GMAC_STA_ADD2, 0x55aa55aa, 0xffffffff);
+ data = gmac_read_reg(toe->gmac[i].base_addr, GMAC_STA_ADD2);
+ if (data == 0x55aa55aa)
+ {
+#ifdef VITESSE_G5SWITCH
+ if(Giga_switch && (i==1)){
+ toe->gmac[i].existed = GMAC_EXISTED_FLAG;
+ break;
+ }
+#endif
+ phy_vendor = gmac_get_phy_vendor(toe->gmac[i].phy_addr);
+ if (phy_vendor != 0 && phy_vendor != 0xffffffff)
+ toe->gmac[i].existed = GMAC_EXISTED_FLAG;
+ }
+ }
+
+ // Write GLOBAL_QUEUE_THRESHOLD_REG
+ threshold.bits32 = 0;
+ threshold.bits.swfq_empty = (TOE_SW_FREEQ_DESC_NUM > 256) ? 255 :
+ TOE_SW_FREEQ_DESC_NUM/2;
+ threshold.bits.hwfq_empty = (TOE_HW_FREEQ_DESC_NUM > 256) ? 256/4 :
+ TOE_HW_FREEQ_DESC_NUM/4;
+ threshold.bits.toe_class = (TOE_TOE_DESC_NUM > 256) ? 256/4 :
+ TOE_TOE_DESC_NUM/4;
+ threshold.bits.intrq = (TOE_INTR_DESC_NUM > 256) ? 256/4 :
+ TOE_INTR_DESC_NUM/4;
+ writel(threshold.bits32, TOE_GLOBAL_BASE + GLOBAL_QUEUE_THRESHOLD_REG);
+
+ FLAG_SWITCH = 0;
+ toe_gmac_sw_reset();
+ toe_init_free_queue();
+ toe_init_swtx_queue();
+#ifdef CONFIG_SL351x_NAT
+ toe_init_hwtx_queue();
+#endif
+ toe_init_default_queue();
+#ifdef CONFIG_SL351x_RXTOE
+ toe_init_interrupt_queue();
+#endif
+ toe_init_interrupt_config();
+
+#if defined(CONFIG_SL351x_NAT) || defined(CONFIG_SL351x_RXTOE)
+ sl351x_hash_init();
+#else
+ {
+ volatile u32 *dp1, *dp2, dword;
+
+ dp1 = (volatile u32 *) TOE_V_BIT_BASE;
+ dp2 = (volatile u32 *) TOE_A_BIT_BASE;
+
+ for (i=0; i<HASH_TOTAL_ENTRIES/32; i++)
+ {
+ *dp1++ = 0;
+ dword = *dp2++; // read-clear
+ }
+ }
+#endif
+ }
+
+#ifdef SL351x_GMAC_WORKAROUND
+#ifdef CONFIG_SL351x_NAT
+ sl351x_nat_workaround_init();
+#endif
+ init_timer(&gmac_workround_timer_obj);
+ if (chip_version == 1)
+ {
+ gmac_workround_timer_obj.expires = jiffies * 50;
+ }
+ else
+ {
+ gmac_workround_timer_obj.expires = jiffies + 2;
+ }
+ gmac_workround_timer_obj.data = (unsigned long)&gmac_workround_timer_obj;
+ gmac_workround_timer_obj.function = (void *)&sl351x_poll_gmac_hanged_status;
+ add_timer(&gmac_workround_timer_obj);
+#endif
+}
+
+/*----------------------------------------------------------------------
+* toe_init_free_queue
+* (1) Initialize the Free Queue Descriptor Base Address & size
+* Register: TOE_GLOBAL_BASE + 0x0004
+* (2) Initialize DMA Read/Write pointer for
+* SW Free Queue and HW Free Queue
+* (3) Initialize DMA Descriptors for
+* SW Free Queue and HW Free Queue,
+*----------------------------------------------------------------------*/
+static void toe_init_free_queue(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ DMA_RWPTR_T rwptr_reg;
+// unsigned int rwptr_addr;
+ unsigned int desc_buf;
+ GMAC_RXDESC_T *sw_desc_ptr;
+ struct sk_buff *skb;
+#ifdef CONFIG_SL351x_NAT
+ GMAC_RXDESC_T *desc_ptr;
+ unsigned int buf_ptr;
+#endif
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_SW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T)),
+ (dma_addr_t *)&toe->sw_freeq_desc_base_dma) ;
+ sw_desc_ptr = (GMAC_RXDESC_T *)desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return;
+ }
+ memset((void *)desc_buf, 0, TOE_SW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T));
+
+ // DMA Queue Base & Size
+ writel((toe->sw_freeq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_SW_FREEQ_DESC_POWER,
+ TOE_GLOBAL_BASE + GLOBAL_SW_FREEQ_BASE_SIZE_REG);
+
+ // init descriptor base
+ toe->swfq_desc_base = desc_buf;
+
+ // SW Free Queue Read/Write Pointer
+ rwptr_reg.bits.wptr = TOE_SW_FREEQ_DESC_NUM - 1;
+ rwptr_reg.bits.rptr = 0;
+ toe->fq_rx_rwptr.bits32 = rwptr_reg.bits32;
+ writel(rwptr_reg.bits32, TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+
+ // SW Free Queue Descriptors
+ for (i=0; i<TOE_SW_FREEQ_DESC_NUM; i++)
+ {
+ sw_desc_ptr->word0.bits.buffer_size = SW_RX_BUF_SIZE;
+ sw_desc_ptr->word1.bits.sw_id = i; // used to locate skb
+ if ( (skb = dev_alloc_skb(SW_RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__); while(1);
+ }
+ REG32(skb->data) = (unsigned int)skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES);
+ // toe->rx_skb[i] = skb;
+ sw_desc_ptr->word2.buf_adr = (unsigned int)__pa(skb->data);
+// consistent_sync((unsigned int)desc_ptr, sizeof(GMAC_RXDESC_T), PCI_DMA_TODEVICE);
+ sw_desc_ptr++;
+ }
+
+#ifdef CONFIG_SL351x_NAT
+ if (sizeof(skb->cb) < 64)
+ {
+ printk("==> %s:: sk structure is incorrect -->Change to cb[64] !\n",__func__); while(1);
+ }
+ // init hardware free queues
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_HW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T)),
+ (dma_addr_t *)&toe->hw_freeq_desc_base_dma) ;
+ desc_ptr = (GMAC_RXDESC_T *)desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return;
+ }
+ memset((void *)desc_buf, 0, TOE_HW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T));
+
+ // DMA Queue Base & Size
+ writel((toe->hw_freeq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_HW_FREEQ_DESC_POWER,
+ TOE_GLOBAL_BASE + GLOBAL_HW_FREEQ_BASE_SIZE_REG);
+
+ // init descriptor base
+ toe->hwfq_desc_base = desc_buf;
+
+ // HW Free Queue Read/Write Pointer
+ rwptr_reg.bits.wptr = TOE_HW_FREEQ_DESC_NUM - 1;
+ rwptr_reg.bits.rptr = 0;
+ writel(rwptr_reg.bits32, TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+#ifndef HW_RXBUF_BY_KMALLOC
+ buf_ptr = (unsigned int)DMA_MALLOC(TOE_HW_FREEQ_DESC_NUM * HW_RX_BUF_SIZE,
+ (dma_addr_t *)&toe->hwfq_buf_base_dma);
+#else
+ buf_ptr = (unsigned int)kmalloc(TOE_HW_FREEQ_DESC_NUM * HW_RX_BUF_SIZE, GFP_KERNEL);
+ toe->hwfq_buf_base_dma = __pa(buf_ptr);
+#endif
+ if (!buf_ptr)
+ {
+ printk("===> %s::Failed to allocate HW TxQ Buffers!\n",__func__);
+ while(1); // could not be happened, if happened, adjust the buffer descriptor number
+ return;
+ }
+
+ toe->hwfq_buf_base = buf_ptr;
+ toe->hwfq_buf_end_dma = toe->hwfq_buf_base_dma + (TOE_HW_FREEQ_DESC_NUM * HW_RX_BUF_SIZE);
+ buf_ptr = (unsigned int)toe->hwfq_buf_base_dma;
+ for (i=0; i<TOE_HW_FREEQ_DESC_NUM; i++)
+ {
+ desc_ptr->word0.bits.buffer_size = HW_RX_BUF_SIZE;
+ desc_ptr->word1.bits.sw_id = i;
+ desc_ptr->word2.buf_adr = (unsigned int)buf_ptr;
+// consistent_sync((unsigned int)desc_ptr, sizeof(GMAC_RXDESC_T), PCI_DMA_TODEVICE);
+ // consistent_sync((unsigned int)buf_ptr, HW_RX_BUF_SIZE, PCI_DMA_TODEVICE);
+ desc_ptr++;
+ buf_ptr += HW_RX_BUF_SIZE;
+ }
+#else
+ // DMA Queue Base & Size
+ writel((0) | TOE_SW_FREEQ_DESC_POWER,
+ TOE_GLOBAL_BASE + GLOBAL_HW_FREEQ_BASE_SIZE_REG);
+ rwptr_reg.bits.wptr = TOE_HW_FREEQ_DESC_NUM - 1;
+ rwptr_reg.bits.rptr = 0;
+ writel(rwptr_reg.bits32, TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+
+#endif
+}
+/*----------------------------------------------------------------------
+* toe_init_swtx_queue
+* (2) Initialize the GMAC 0/1 SW TXQ Queue Descriptor Base Address & sizeup
+* GMAC_SW_TX_QUEUE_BASE_REG(0x0050)
+* (2) Initialize DMA Read/Write pointer for
+* GMAC 0/1 SW TX Q0-5
+*----------------------------------------------------------------------*/
+static void toe_init_swtx_queue(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ DMA_RWPTR_T rwptr_reg;
+ unsigned int rwptr_addr;
+ unsigned int desc_buf;
+
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ // GMAC-0, SW-TXQ
+ // The GMAC-0 and GMAC-0 maybe have different descriptor number
+ // so, not use for instruction
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_GMAC0_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T)),
+ (dma_addr_t *)&toe->gmac[0].swtxq_desc_base_dma) ;
+ toe->gmac[0].swtxq_desc_base = desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_buf, 0, TOE_GMAC0_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T));
+ writel((toe->gmac[0].swtxq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_GMAC0_SWTXQ_DESC_POWER,
+ TOE_GMAC0_DMA_BASE+ GMAC_SW_TX_QUEUE_BASE_REG);
+
+ // GMAC0 SW TX Q0-Q5
+ rwptr_reg.bits.wptr = 0;
+ rwptr_reg.bits.rptr = 0;
+ rwptr_addr = TOE_GMAC0_DMA_BASE + GMAC_SW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_SW_TXQ_NUM; i++)
+ {
+ toe->gmac[0].swtxq[i].rwptr_reg = rwptr_addr;
+ toe->gmac[0].swtxq[i].desc_base = desc_buf;
+ toe->gmac[0].swtxq[i].total_desc_num = TOE_GMAC0_SWTXQ_DESC_NUM;
+ desc_buf += TOE_GMAC0_SWTXQ_DESC_NUM * sizeof(GMAC_TXDESC_T);
+ writel(rwptr_reg.bits32, rwptr_addr);
+ rwptr_addr+=4;
+ }
+
+ // GMAC-1, SW-TXQ
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_GMAC1_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T)),
+ (dma_addr_t *)&toe->gmac[1].swtxq_desc_base_dma) ;
+ toe->gmac[1].swtxq_desc_base = desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_buf, 0, TOE_GMAC1_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T));
+ writel((toe->gmac[1].swtxq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_GMAC1_SWTXQ_DESC_POWER,
+ TOE_GMAC1_DMA_BASE+ GMAC_SW_TX_QUEUE_BASE_REG);
+
+
+ // GMAC1 SW TX Q0-Q5
+ rwptr_reg.bits.wptr = 0;
+ rwptr_reg.bits.rptr = 0;
+ rwptr_addr = TOE_GMAC1_DMA_BASE + GMAC_SW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_SW_TXQ_NUM; i++)
+ {
+ toe->gmac[1].swtxq[i].rwptr_reg = rwptr_addr;
+ toe->gmac[1].swtxq[i].desc_base = desc_buf;
+ toe->gmac[1].swtxq[i].total_desc_num = TOE_GMAC1_SWTXQ_DESC_NUM;
+ desc_buf += TOE_GMAC1_SWTXQ_DESC_NUM * sizeof(GMAC_TXDESC_T);
+ writel(rwptr_reg.bits32, rwptr_addr);
+ rwptr_addr+=4;
+ }
+}
+
+/*----------------------------------------------------------------------
+* toe_init_hwtx_queue
+* (2) Initialize the GMAC 0/1 HW TXQ Queue Descriptor Base Address & size
+* GMAC_HW_TX_QUEUE_BASE_REG(0x0054)
+* (2) Initialize DMA Read/Write pointer for
+* GMAC 0/1 HW TX Q0-5
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static void toe_init_hwtx_queue(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ DMA_RWPTR_T rwptr_reg;
+ unsigned int rwptr_addr;
+ unsigned int desc_buf;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ // GMAC-0, HW-TXQ
+ // The GMAC-0 and GMAC-0 maybe have different descriptor number
+ // so, not use for instruction
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_GMAC0_HWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T)),
+ (dma_addr_t *)&toe->gmac[0].hwtxq_desc_base_dma) ;
+ toe->gmac[0].hwtxq_desc_base = desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_buf, 0, TOE_GMAC0_HWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T));
+ writel((toe->gmac[0].hwtxq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_GMAC0_HWTXQ_DESC_POWER,
+ TOE_GMAC0_DMA_BASE+ GMAC_HW_TX_QUEUE_BASE_REG);
+
+ // GMAC0 HW TX Q0-Q5
+ rwptr_reg.bits.wptr = 0;
+ rwptr_reg.bits.rptr = 0;
+ rwptr_addr = TOE_GMAC0_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ toe->gmac[0].hwtxq[i].desc_base = desc_buf;
+ desc_buf += TOE_GMAC0_HWTXQ_DESC_NUM * sizeof(GMAC_TXDESC_T);
+ writel(rwptr_reg.bits32, rwptr_addr);
+ rwptr_addr+=4;
+ }
+
+ // GMAC-1, HW-TXQ
+ desc_buf = (unsigned int)DMA_MALLOC((TOE_GMAC1_HWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T)),
+ (dma_addr_t *)&toe->gmac[1].hwtxq_desc_base_dma) ;
+ toe->gmac[1].hwtxq_desc_base = desc_buf;
+ if (!desc_buf)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_buf, 0, TOE_GMAC1_HWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T));
+ writel((toe->gmac[1].hwtxq_desc_base_dma & DMA_Q_BASE_MASK) | TOE_GMAC1_HWTXQ_DESC_POWER,
+ TOE_GMAC1_DMA_BASE+ GMAC_HW_TX_QUEUE_BASE_REG);
+
+ // GMAC1 HW TX Q0-Q5
+ rwptr_reg.bits.wptr = 0;
+ rwptr_reg.bits.rptr = 0;
+ rwptr_addr = TOE_GMAC1_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ toe->gmac[1].hwtxq[i].desc_base = desc_buf;
+ desc_buf += TOE_GMAC1_HWTXQ_DESC_NUM * sizeof(GMAC_TXDESC_T);
+ writel(rwptr_reg.bits32, rwptr_addr);
+ rwptr_addr+=4;
+ }
+}
+#endif
+
+/*----------------------------------------------------------------------
+* toe_init_default_queue
+* (1) Initialize the default 0/1 Queue Header
+* Register: TOE_DEFAULT_Q0_HDR_BASE (0x60002000)
+* TOE_DEFAULT_Q1_HDR_BASE (0x60002008)
+* (2) Initialize Descriptors of Default Queue 0/1
+*----------------------------------------------------------------------*/
+static void toe_init_default_queue(void)
+{
+ TOE_INFO_T *toe;
+ volatile NONTOE_QHDR_T *qhdr;
+ GMAC_RXDESC_T *desc_ptr;
+ DMA_SKB_SIZE_T skb_size;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ desc_ptr = (GMAC_RXDESC_T *)DMA_MALLOC((TOE_DEFAULT_Q0_DESC_NUM * sizeof(GMAC_RXDESC_T)),
+ (dma_addr_t *)&toe->gmac[0].default_desc_base_dma);
+ if (!desc_ptr)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_ptr, 0, TOE_DEFAULT_Q0_DESC_NUM * sizeof(GMAC_RXDESC_T));
+ toe->gmac[0].default_desc_base = (unsigned int)desc_ptr;
+ toe->gmac[0].default_desc_num = TOE_DEFAULT_Q0_DESC_NUM;
+ qhdr = (volatile NONTOE_QHDR_T *)TOE_DEFAULT_Q0_HDR_BASE;
+ qhdr->word0.base_size = ((unsigned int)toe->gmac[0].default_desc_base_dma & NONTOE_QHDR0_BASE_MASK) | TOE_DEFAULT_Q0_DESC_POWER;
+ qhdr->word1.bits32 = 0;
+ toe->gmac[0].rx_rwptr.bits32 = 0;
+ toe->gmac[0].default_qhdr = (NONTOE_QHDR_T *)qhdr;
+ desc_ptr = (GMAC_RXDESC_T *)DMA_MALLOC((TOE_DEFAULT_Q1_DESC_NUM * sizeof(GMAC_RXDESC_T)),
+ (dma_addr_t *)&toe->gmac[1].default_desc_base_dma);
+ if (!desc_ptr)
+ {
+ printk("%s::DMA_MALLOC fail !\n",__func__);
+ return ;
+ }
+ memset((void *)desc_ptr, 0, TOE_DEFAULT_Q1_DESC_NUM * sizeof(GMAC_RXDESC_T));
+ toe->gmac[1].default_desc_base = (unsigned int)desc_ptr;
+ toe->gmac[1].default_desc_num = TOE_DEFAULT_Q1_DESC_NUM;
+ qhdr = (volatile NONTOE_QHDR_T *)TOE_DEFAULT_Q1_HDR_BASE;
+ qhdr->word0.base_size = ((unsigned int)toe->gmac[1].default_desc_base_dma & NONTOE_QHDR0_BASE_MASK) | TOE_DEFAULT_Q1_DESC_POWER;
+ qhdr->word1.bits32 = 0;
+ toe->gmac[1].rx_rwptr.bits32 = 0;
+ toe->gmac[1].default_qhdr = (NONTOE_QHDR_T *)qhdr;
+
+ skb_size.bits.hw_skb_size = HW_RX_BUF_SIZE;
+ skb_size.bits.sw_skb_size = SW_RX_BUF_SIZE;
+ writel(skb_size.bits32, TOE_GLOBAL_BASE + GLOBAL_DMA_SKB_SIZE_REG);
+}
+
+/*----------------------------------------------------------------------
+* toe_init_interrupt_queue
+* (1) Initialize the Interrupt Queue Header
+* Register: TOE_INTR_Q_HDR_BASE (0x60002080)
+* (2) Initialize Descriptors of Interrupt Queues
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static void toe_init_interrupt_queue(void)
+{
+ TOE_INFO_T *toe;
+ volatile NONTOE_QHDR_T *qhdr;
+ INTR_QHDR_T *desc_ptr;
+ // unsigned int desc_buf_addr;
+ int i;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ desc_ptr = (INTR_QHDR_T *)DMA_MALLOC((TOE_INTR_QUEUE_NUM * TOE_INTR_DESC_NUM * sizeof(INTR_QHDR_T)),
+ (dma_addr_t *)&toe->intr_desc_base_dma);
+ if (!desc_ptr)
+ {
+ printk("%s::DMA_MALLOC interrupt queue fail !\n",__func__);
+ return ;
+ }
+ /*
+ desc_buf_addr = (unsigned int)DMA_MALLOC((TOE_INTR_DESC_NUM * sizeof(TOE_QHDR_T)),
+ (dma_addr_t *)&toe->intr_buf_base_dma);
+ if (!desc_buf_addr)
+ {
+ printk("%s::DMA_MALLOC interrupt desc fail !\n",__func__);
+ return ;
+ }*/
+ printk("#### %s::Intr Q desc %x\n", __func__, (u32)desc_ptr);
+
+ memset((void *)desc_ptr, 0, TOE_INTR_QUEUE_NUM * TOE_INTR_DESC_NUM * sizeof(INTR_QHDR_T));
+// memset((void *)desc_buf_addr, 0, TOE_INTR_DESC_NUM * sizeof(TOE_QHDR_T));
+ toe->intr_desc_base = (unsigned int)desc_ptr;
+ toe->intr_desc_num = TOE_INTR_DESC_NUM;
+
+ qhdr = (volatile NONTOE_QHDR_T *)TOE_INTR_Q_HDR_BASE;
+// intrq = (INTRQ_INFO_T*) &toe->intrq[0];
+ for (i=0; i<TOE_INTR_QUEUE_NUM; i++, qhdr++)
+ {
+ qhdr->word0.base_size = ((unsigned int)toe->intr_desc_base_dma & NONTOE_QHDR0_BASE_MASK) | TOE_INTR_DESC_POWER;
+ qhdr->word1.bits32 = 0;
+ desc_ptr += TOE_INTR_DESC_NUM;
+ }
+}
+
+#endif
+
+/*----------------------------------------------------------------------
+* toe_init_interrupt_config
+* Interrupt Select Registers are used to map interrupt to int0 or int1
+* Int0 and int1 are wired to CPU 0/1 GMAC 0/1
+* Interrupt Device Inteface data are used to pass device info to
+* upper device deiver or store status/statistics
+* ISR handler
+* (1) If status bit ON but masked, the prinf error message (bug issue)
+* (2) If select bits are for me, handle it, else skip to let
+* the other ISR handles it.
+* Notes:
+* GMACx init routine (for eCOS) or open routine (for Linux)
+* enable the interrupt bits only which are selected for him.
+*
+* Default Setting:
+* GMAC0 intr bits ------> int0 ----> eth0
+* GMAC1 intr bits ------> int1 ----> eth1
+* TOE intr -------------> int0 ----> eth0
+* Classification Intr --> int0 ----> eth0
+* Default Q0 -----------> int0 ----> eth0
+* Default Q1 -----------> int1 ----> eth1
+*----------------------------------------------------------------------*/
+static void toe_init_interrupt_config(void)
+{
+ // clear all status bits
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_0_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_1_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_2_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_3_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+
+ // Init select registers
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_0_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_1_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_2_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_3_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_4_REG);
+
+ // disable all interrupt
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_0_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_2_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_3_REG);
+ writel(0, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_4_REG);
+}
+
+/*----------------------------------------------------------------------
+* toe_init_gmac
+*----------------------------------------------------------------------*/
+static void toe_init_gmac(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ TOE_INFO_T *toe;
+ u32 data;
+
+ if (!gmac_initialized)
+ return ;
+
+ if (!tp->existed)
+ return;
+
+ tp->dev = dev;
+ tp->flow_control_enable = 1;
+ tp->pre_phy_status = LINK_DOWN;
+ tp->full_duplex_status = tp->full_duplex_cfg;
+ tp->speed_status = tp->speed_status;
+
+#if 0
+ /* get mac address from FLASH */
+ gmac_get_mac_address();
+#endif
+
+ /* set PHY register to start autonegition process */
+ gmac_set_phy_status(dev);
+
+ /* GMAC initialization */
+ if ( toe_gmac_init_chip(dev) )
+ {
+ printk ("GMAC %d init fail\n", tp->port_id);
+ }
+
+ /* clear statistic counter */
+ toe_gmac_clear_counter(dev);
+
+ memset((void *)&tp->ifStatics, 0, sizeof(struct net_device_stats));
+
+ /* -----------------------------------------------------------
+ Enable GMAC interrupt & disable loopback
+ Notes:
+ GMACx init routine (for eCOS) or open routine (for Linux)
+ enable the interrupt bits only which are selected for him.
+ --------------------------------------------------------------*/
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ // Enable Interrupt Bits
+ if (tp->port_id == 0)
+ {
+ tp->intr0_selected = GMAC0_TXDERR_INT_BIT | GMAC0_TXPERR_INT_BIT |
+ GMAC0_RXDERR_INT_BIT | GMAC0_RXPERR_INT_BIT |
+ GMAC0_SWTQ05_FIN_INT_BIT | GMAC0_SWTQ05_EOF_INT_BIT |
+ GMAC0_SWTQ04_FIN_INT_BIT | GMAC0_SWTQ04_EOF_INT_BIT |
+ GMAC0_SWTQ03_FIN_INT_BIT | GMAC0_SWTQ03_EOF_INT_BIT |
+ GMAC0_SWTQ02_FIN_INT_BIT | GMAC0_SWTQ02_EOF_INT_BIT |
+ GMAC0_SWTQ01_FIN_INT_BIT | GMAC0_SWTQ01_EOF_INT_BIT |
+ GMAC0_SWTQ00_FIN_INT_BIT | GMAC0_SWTQ00_EOF_INT_BIT;
+
+#ifdef GMAX_TX_INTR_DISABLED
+ tp->intr0_enabled = 0;
+#else
+ tp->intr0_enabled = GMAC0_SWTQ00_FIN_INT_BIT | GMAC0_SWTQ00_EOF_INT_BIT;
+#endif
+
+ tp->intr1_selected = TOE_IQ_ALL_BITS | TOE_CLASS_RX_INT_BITS |
+ GMAC0_HWTQ03_EOF_INT_BIT | GMAC0_HWTQ02_EOF_INT_BIT |
+ GMAC0_HWTQ01_EOF_INT_BIT | GMAC0_HWTQ00_EOF_INT_BIT |
+ DEFAULT_Q0_INT_BIT;
+ tp->intr1_enabled = DEFAULT_Q0_INT_BIT | TOE_IQ_ALL_BITS;
+ tp->intr2_selected = 0xffffffff; // TOE Queue 32-63 FUUL Intr
+ tp->intr2_enabled = 0xffffffff;
+ tp->intr3_selected = 0xffffffff; // TOE Queue 0-31 FUUL Intr
+ tp->intr3_enabled = 0xffffffff;
+ tp->intr4_selected = GMAC0_INT_BITS | CLASS_RX_FULL_INT_BITS |
+ HWFQ_EMPTY_INT_BIT | SWFQ_EMPTY_INT_BIT;
+ tp->intr4_enabled = GMAC0_INT_BITS | SWFQ_EMPTY_INT_BIT;
+
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_0_REG) & ~tp->intr0_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_0_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_1_REG) & ~tp->intr1_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_1_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_2_REG) & ~tp->intr2_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_2_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_3_REG) & ~tp->intr3_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_3_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_4_REG) & ~tp->intr4_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_4_REG);
+ }
+ else
+ {
+ tp->intr0_selected = GMAC1_TXDERR_INT_BIT | GMAC1_TXPERR_INT_BIT |
+ GMAC1_RXDERR_INT_BIT | GMAC1_RXPERR_INT_BIT |
+ GMAC1_SWTQ15_FIN_INT_BIT | GMAC1_SWTQ15_EOF_INT_BIT |
+ GMAC1_SWTQ14_FIN_INT_BIT | GMAC1_SWTQ14_EOF_INT_BIT |
+ GMAC1_SWTQ13_FIN_INT_BIT | GMAC1_SWTQ13_EOF_INT_BIT |
+ GMAC1_SWTQ12_FIN_INT_BIT | GMAC1_SWTQ12_EOF_INT_BIT |
+ GMAC1_SWTQ11_FIN_INT_BIT | GMAC1_SWTQ11_EOF_INT_BIT |
+ GMAC1_SWTQ10_FIN_INT_BIT | GMAC1_SWTQ10_EOF_INT_BIT;
+#ifdef GMAX_TX_INTR_DISABLED
+ tp->intr0_enabled = 0;
+#else
+ tp->intr0_enabled = GMAC1_SWTQ10_FIN_INT_BIT | GMAC1_SWTQ10_EOF_INT_BIT;
+#endif
+
+ tp->intr1_selected = DEFAULT_Q1_INT_BIT;
+ tp->intr1_enabled = DEFAULT_Q1_INT_BIT | TOE_IQ_ALL_BITS;
+ tp->intr2_selected = 0; // TOE Queue 32-63 FUUL Intr
+ tp->intr2_enabled = 0;
+ tp->intr3_selected = 0; // TOE Queue 0-31 FUUL Intr
+ tp->intr3_enabled = 0;
+ tp->intr4_selected = GMAC1_INT_BITS;
+ tp->intr4_enabled = GMAC1_INT_BITS;
+
+ if (toe->gmac[0].existed != GMAC_EXISTED_FLAG)
+ {
+ tp->intr1_selected |= TOE_IQ_ALL_BITS | TOE_CLASS_RX_INT_BITS |
+ GMAC0_HWTQ03_EOF_INT_BIT | GMAC0_HWTQ02_EOF_INT_BIT |
+ GMAC0_HWTQ01_EOF_INT_BIT | GMAC0_HWTQ00_EOF_INT_BIT;
+ tp->intr1_enabled |= TOE_IQ_ALL_BITS;
+ tp->intr2_selected |= 0xffffffff; // TOE Queue 32-63 FUUL Intr
+ tp->intr2_enabled |= 0xffffffff;
+ tp->intr3_selected |= 0xffffffff; // TOE Queue 0-31 FUUL Intr
+ tp->intr3_enabled |= 0xffffffff;
+ tp->intr4_selected |= CLASS_RX_FULL_INT_BITS |
+ HWFQ_EMPTY_INT_BIT | SWFQ_EMPTY_INT_BIT;
+ tp->intr4_enabled |= SWFQ_EMPTY_INT_BIT;
+ }
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_0_REG) | tp->intr0_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_0_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_1_REG) | tp->intr1_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_1_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_2_REG) | tp->intr2_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_2_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_3_REG) | tp->intr3_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_3_REG);
+ data = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_4_REG) | tp->intr4_selected;
+ writel(data, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_SELECT_4_REG);
+ }
+
+ // enable only selected bits
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_0_REG,
+ tp->intr0_enabled, tp->intr0_selected);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_1_REG,
+ tp->intr1_enabled, tp->intr1_selected);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_2_REG,
+ tp->intr2_enabled, tp->intr2_selected);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_3_REG,
+ tp->intr3_enabled, tp->intr3_selected);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_4_REG,
+ tp->intr4_enabled, tp->intr4_selected);
+
+ /* start DMA process */
+ toe_gmac_hw_start(dev);
+
+ /* enable tx/rx register */
+ toe_gmac_enable_tx_rx(dev);
+
+// toe_gmac_enable_interrupt(tp->irq);
+
+ return ;
+}
+
+
+/*----------------------------------------------------------------------
+* toe_gmac_sw_reset
+*----------------------------------------------------------------------*/
+static void toe_gmac_sw_reset(void)
+{
+ unsigned int reg_val;
+ reg_val = readl(GMAC_GLOBAL_BASE_ADDR+GLOBAL_RESET_REG) | 0x00000060; /* GMAC0 S/W reset */
+ writel(reg_val,GMAC_GLOBAL_BASE_ADDR+GLOBAL_RESET_REG);
+ udelay(100);
+ return;
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_init_chip
+*----------------------------------------------------------------------*/
+static int toe_gmac_init_chip(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG2_T config2_val;
+ GMAC_CONFIG0_T config0,config0_mask;
+ GMAC_CONFIG1_T config1;
+ #ifdef CONFIG_SL351x_NAT
+ GMAC_CONFIG3_T config3_val;
+ #endif
+ GMAC_TX_WCR0_T hw_weigh;
+ GMAC_TX_WCR1_T sw_weigh;
+// GMAC_HASH_ENABLE_REG0_T hash_ctrl;
+//
+#if 0 /* mac address will be set in late_initcall */
+ struct sockaddr sock;
+ // GMAC_AHB_WEIGHT_T ahb_weight, ahb_weight_mask;
+
+
+ /* set station MAC address1 and address2 */
+ memcpy(&sock.sa_data[0],&eth_mac[tp->port_id][0],6);
+ gmac_set_mac_address(dev,(void *)&sock);
+#endif
+
+ /* set RX_FLTR register to receive all multicast packet */
+ gmac_write_reg(tp->base_addr, GMAC_RX_FLTR, 0x00000007,0x0000001f);
+ // gmac_write_reg(tp->base_addr, GMAC_RX_FLTR, 0x00000007,0x0000001f);
+ //gmac_write_reg(tp->base_addr, GMAC_RX_FLTR,0x00000007,0x0000001f);
+
+ /* set per packet buffer size */
+ // config1.bits32 = 0x002004; //next version
+ /* set flow control threshold */
+ config1.bits32 = 0;
+ config1.bits.set_threshold = 32 / 2;
+ config1.bits.rel_threshold = 32 / 4 * 3;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG1, config1.bits32, 0xffffffff);
+
+ /* set flow control threshold */
+ config2_val.bits32 = 0;
+ config2_val.bits.set_threshold = TOE_SW_FREEQ_DESC_NUM/2;
+ config2_val.bits.rel_threshold = TOE_SW_FREEQ_DESC_NUM*3/4;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG2, config2_val.bits32,0xffffffff);
+
+ #ifdef CONFIG_SL351x_NAT
+ /* set HW free queue flow control threshold */
+ config3_val.bits32 = 0;
+ config3_val.bits.set_threshold = PAUSE_SET_HW_FREEQ;
+ config3_val.bits.rel_threshold = PAUSE_REL_HW_FREEQ;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG3, config3_val.bits32,0xffffffff);
+ #endif
+ /* set_mcast_filter mask*/
+ // gmac_write_reg(tp->base_addr,GMAC_MCAST_FIL0,0x0,0xffffffff);
+ // gmac_write_reg(tp->base_addr,GMAC_MCAST_FIL1,0x0,0xffffffff);
+
+ /* disable TX/RX and disable internal loop back */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+
+ //debug_Aaron
+#ifdef L2_jumbo_frame
+ config0.bits.max_len = 5;
+#else
+ config0.bits.max_len = 2;
+#endif
+
+ if (tp->flow_control_enable==1)
+ {
+ config0.bits.tx_fc_en = 1; /* enable tx flow control */
+ config0.bits.rx_fc_en = 1; /* enable rx flow control */
+ printk("Enable MAC Flow Control...\n");
+ }
+ else
+ {
+ config0.bits.tx_fc_en = 0; /* disable tx flow control */
+ config0.bits.rx_fc_en = 0; /* disable rx flow control */
+ printk("Disable MAC Flow Control...\n");
+ }
+ config0.bits.dis_rx = 1; /* disable rx */
+ config0.bits.dis_tx = 1; /* disable tx */
+ config0.bits.loop_back = 0; /* enable/disable GMAC loopback */
+ config0.bits.rx_err_detect = 1;
+ config0.bits.rgmii_en = 0;
+ config0.bits.rgmm_edge = 1;
+ config0.bits.rxc_inv = 0;
+ config0.bits.ipv4_rx_chksum = 1; /* enable H/W to check ip checksum */
+ config0.bits.ipv6_rx_chksum = 1; /* enable H/W to check ip checksum */
+ config0.bits.port0_chk_hwq = 1; // GaryChen 3/24/2006 2:26PM
+ config0.bits.port1_chk_hwq = 1; // GaryChen 3/24/2006 2:26PM
+ config0.bits.port0_chk_toeq = 1;
+ config0.bits.port1_chk_toeq = 1;
+ config0.bits.port0_chk_classq = 1;
+ config0.bits.port1_chk_classq = 1;
+
+ config0_mask.bits.max_len = 7;
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ config0_mask.bits.loop_back = 1;
+ config0_mask.bits.rgmii_en = 1;
+ config0_mask.bits.rgmm_edge = 1;
+ config0_mask.bits.rxc_inv = 1;
+ config0_mask.bits.ipv4_rx_chksum = 1;
+ config0_mask.bits.ipv6_rx_chksum = 1;
+ config0_mask.bits.port0_chk_hwq = 1;
+ config0_mask.bits.port1_chk_hwq = 1;
+ config0_mask.bits.port0_chk_toeq = 1;
+ config0_mask.bits.port1_chk_toeq = 1;
+ config0_mask.bits.port0_chk_classq = 1;
+ config0_mask.bits.port1_chk_classq = 1;
+ config0_mask.bits.rx_err_detect = 1;
+
+ #if 0
+ config0.bits.dis_rx = 1; /* disable rx */
+ config0.bits.dis_tx = 1; /* disable tx */
+ config0.bits.loop_back = 0; /* enable/disable GMAC loopback */
+ config0.bits.txc_inv = 0;
+ config0.bits.rgmii_en = 0;
+ config0.bits.rgmm_edge = 1;
+ config0.bits.rxc_inv = 1;
+ config0.bits.ipv4_tss_rx_en = 1; /* enable H/W to check ip checksum */
+ config0.bits.ipv6_tss_rx_en = 1; /* enable H/W to check ip checksum */
+
+ config0_mask.bits.max_len = 3;
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ config0_mask.bits.loop_back = 1;
+ config0_mask.bits.rgmii_en = 1;
+ config0_mask.bits.rgmm_edge = 1;
+ config0_mask.bits.txc_inv = 1;
+ config0_mask.bits.rxc_inv = 1;
+ config0_mask.bits.ipv4_tss_rx_en = 1;
+ config0_mask.bits.ipv6_tss_rx_en = 1;
+ #endif
+
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0, config0.bits32,config0_mask.bits32);
+
+ #if 1
+ hw_weigh.bits32 = 0;
+ hw_weigh.bits.hw_tq3 = 1;
+ hw_weigh.bits.hw_tq2 = 1;
+ hw_weigh.bits.hw_tq1 = 1;
+ hw_weigh.bits.hw_tq0 = 1;
+ gmac_write_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_0_REG, hw_weigh.bits32, 0xffffffff);
+
+ sw_weigh.bits32 = 0;
+ sw_weigh.bits.sw_tq5 = 1;
+ sw_weigh.bits.sw_tq4 = 1;
+ sw_weigh.bits.sw_tq3 = 1;
+ sw_weigh.bits.sw_tq2 = 1;
+ sw_weigh.bits.sw_tq1 = 1;
+ sw_weigh.bits.sw_tq0 = 1;
+ gmac_write_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_1_REG, sw_weigh.bits32, 0xffffffff);
+ #endif
+
+ #if 0
+ ahb_weight.bits32 = 0;
+ ahb_weight_mask.bits32 = 0;
+ ahb_weight.bits.rx_weight = 1;
+ ahb_weight.bits.tx_weight = 1;
+ ahb_weight.bits.hash_weight = 1;
+ ahb_weight.bits.pre_req = 0x1f;
+ ahb_weight.bits.tqDV_threshold = 0;
+ ahb_weight_mask.bits.rx_weight = 0x1f;
+ ahb_weight_mask.bits.tx_weight = 0x1f;
+ ahb_weight_mask.bits.hash_weight = 0x1f;
+ ahb_weight_mask.bits.pre_req = 0x1f;
+ ahb_weight_mask.bits.tqDV_threshold = 0x1f;
+ gmac_write_reg(tp->dma_base_addr, GMAC_AHB_WEIGHT_REG, ahb_weight.bits32, ahb_weight_mask.bits32);
+ #endif
+
+ #if defined(CONFIG_SL351x_NAT) || defined(CONFIG_SL351x_RXTOE)
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR0, IPPROTO_TCP, 0xffffffff);
+ #endif
+ #ifdef CONFIG_SL351x_NAT
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR1, IPPROTO_UDP, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR2, IPPROTO_GRE, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR3, 0xff, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR4, 0xff, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR5, 0xff, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR6, 0xff, 0xffffffff);
+ gmac_write_reg(tp->dma_base_addr, GMAC_SPR7, 0xff, 0xffffffff);
+
+ sl351x_nat_init();
+ #endif
+
+ #ifdef CONFIG_SL351x_RXTOE
+ /* setup matching rule to TOE */
+ sl351x_toe_init();
+ #endif
+
+ // for A1 ASIC version
+// hash_ctrl.bits32 = 0;
+// hash_ctrl.bits.timing = 6;
+// gmac_write_reg(tp->dma_base_addr, GMAC_HASH_ENGINE_REG0, hash_ctrl.bits32, 0xffffffff);
+
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_enable_tx_rx
+*----------------------------------------------------------------------*/
+static void toe_gmac_enable_tx_rx(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 0; /* enable rx */
+ config0.bits.dis_tx = 0; /* enable tx */
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0, config0.bits32,config0_mask.bits32);
+}
+/*----------------------------------------------------------------------
+* toe_gmac_disable_rx
+*----------------------------------------------------------------------*/
+#if 0
+static void toe_gmac_disable_rx(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 1; /* disable rx */
+// config0.bits.dis_tx = 1; /* disable tx */
+ config0_mask.bits.dis_rx = 1;
+// config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0, config0.bits32,config0_mask.bits32);
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_gmac_enable_rx
+*----------------------------------------------------------------------*/
+#if 0
+static void toe_gmac_enable_rx(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 0; /* enable rx */
+// config0.bits.dis_tx = 0; /* enable tx */
+ config0_mask.bits.dis_rx = 1;
+// config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0, config0.bits32,config0_mask.bits32);
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_gmac_disable_tx_rx
+*----------------------------------------------------------------------*/
+static void toe_gmac_disable_tx_rx(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+
+ /* enable TX/RX */
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.dis_rx = 1; /* disable rx */
+ config0.bits.dis_tx = 1; /* disable tx */
+ config0_mask.bits.dis_rx = 1;
+ config0_mask.bits.dis_tx = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0, config0.bits32,config0_mask.bits32);
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_hw_start
+*----------------------------------------------------------------------*/
+static void toe_gmac_hw_start(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+
+
+ /* program dma control register */
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.rd_enable = 1;
+ dma_ctrl.bits.td_enable = 1;
+ dma_ctrl.bits.loopback = 0;
+ dma_ctrl.bits.drop_small_ack = 0;
+ dma_ctrl.bits.rd_prot = 0;
+ dma_ctrl.bits.rd_burst_size = 3;
+ dma_ctrl.bits.rd_insert_bytes = RX_INSERT_BYTES;
+ dma_ctrl.bits.rd_bus = 3;
+ dma_ctrl.bits.td_prot = 0;
+ dma_ctrl.bits.td_burst_size = 3;
+ dma_ctrl.bits.td_bus = 3;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.rd_enable = 1;
+ dma_ctrl_mask.bits.td_enable = 1;
+ dma_ctrl_mask.bits.loopback = 1;
+ dma_ctrl_mask.bits.drop_small_ack = 1;
+ dma_ctrl_mask.bits.rd_prot = 3;
+ dma_ctrl_mask.bits.rd_burst_size = 3;
+ dma_ctrl_mask.bits.rd_insert_bytes = 3;
+ dma_ctrl_mask.bits.rd_bus = 3;
+ dma_ctrl_mask.bits.td_prot = 0x0f;
+ dma_ctrl_mask.bits.td_burst_size = 3;
+ dma_ctrl_mask.bits.td_bus = 3;
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+
+ return;
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_hw_stop
+*----------------------------------------------------------------------*/
+static void toe_gmac_hw_stop(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+
+ /* program dma control register */
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.rd_enable = 0;
+ dma_ctrl.bits.td_enable = 0;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.rd_enable = 1;
+ dma_ctrl_mask.bits.td_enable = 1;
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_clear_counter
+*----------------------------------------------------------------------*/
+static int toe_gmac_clear_counter (struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+ /* clear counter */
+ gmac_read_reg(tp->base_addr, GMAC_IN_DISCARDS);
+ gmac_read_reg(tp->base_addr, GMAC_IN_ERRORS);
+ gmac_read_reg(tp->base_addr, GMAC_IN_MCAST);
+ gmac_read_reg(tp->base_addr, GMAC_IN_BCAST);
+ gmac_read_reg(tp->base_addr, GMAC_IN_MAC1);
+ gmac_read_reg(tp->base_addr, GMAC_IN_MAC2);
+ tp->ifStatics.tx_bytes = 0;
+ tp->ifStatics.tx_packets = 0;
+ tp->ifStatics.tx_errors = 0;
+ tp->ifStatics.rx_bytes = 0;
+ tp->ifStatics.rx_packets = 0;
+ tp->ifStatics.rx_errors = 0;
+ tp->ifStatics.rx_dropped = 0;
+ return (0);
+}
+
+
+/*----------------------------------------------------------------------
+* toe_gmac_tx_complete
+*----------------------------------------------------------------------*/
+static void toe_gmac_tx_complete(GMAC_INFO_T *tp, unsigned int tx_qid,
+ struct net_device *dev, int interrupt)
+{
+ volatile GMAC_TXDESC_T *curr_desc;
+ GMAC_TXDESC_0_T word0;
+ GMAC_TXDESC_1_T word1;
+ unsigned int desc_count;
+// struct net_device_stats *isPtr = (struct net_device_stats *)&tp->ifStatics;
+ GMAC_SWTXQ_T *swtxq;
+ DMA_RWPTR_T rwptr;
+
+ /* get tx H/W completed descriptor virtual address */
+ /* check tx status and accumulate tx statistics */
+ swtxq = &tp->swtxq[tx_qid];
+ swtxq->intr_cnt++;
+ for (;;)
+ {
+ rwptr.bits32 = readl(swtxq->rwptr_reg);
+ if (rwptr.bits.rptr == swtxq->finished_idx)
+ break;
+ curr_desc = (volatile GMAC_TXDESC_T *)swtxq->desc_base + swtxq->finished_idx;
+// consistent_sync((void *)curr_desc, sizeof(GMAC_TXDESC_T), PCI_DMA_FROMDEVICE);
+ word0.bits32 = curr_desc->word0.bits32;
+ word1.bits32 = curr_desc->word1.bits32;
+
+ if (word0.bits.status_tx_ok)
+ {
+ tp->ifStatics.tx_bytes += word1.bits.byte_count;
+ desc_count = word0.bits.desc_count;
+ if (desc_count==0)
+ {
+ printk("%s::Desc 0x%x = 0x%x, desc_count=%d\n",__func__, (u32)curr_desc, word0.bits32, desc_count);
+ while(1);
+ }
+ while (--desc_count)
+ {
+ word0.bits.status_tx_ok = 0;
+ curr_desc->word0.bits32 = word0.bits32;
+ swtxq->finished_idx = RWPTR_ADVANCE_ONE(swtxq->finished_idx, swtxq->total_desc_num);
+ curr_desc = (GMAC_TXDESC_T *)swtxq->desc_base + swtxq->finished_idx;
+ word0.bits32 = curr_desc->word0.bits32;
+#ifdef _DUMP_TX_TCP_CONTENT
+ if (curr_desc->word0.bits.buffer_size < 16)
+ {
+ int a;
+ char *datap;
+ printk("\t Tx Finished Desc 0x%x Len %d Addr 0x%08x: ", (u32)curr_desc, curr_desc->word0.bits.buffer_size, curr_desc->word2.buf_adr);
+ datap = (char *)__va(curr_desc->word2.buf_adr);
+ for (a=0; a<8 && a<curr_desc->word0.bits.buffer_size; a++, datap++)
+ {
+ printk("0x%02x ", *datap);
+ }
+ printk("\n");
+ }
+#endif
+ }
+
+ word0.bits.status_tx_ok = 0;
+ if (swtxq->tx_skb[swtxq->finished_idx])
+ {
+ if (interrupt)
+ dev_kfree_skb_irq(swtxq->tx_skb[swtxq->finished_idx]);
+ else
+ dev_kfree_skb(swtxq->tx_skb[swtxq->finished_idx]);
+ swtxq->tx_skb[swtxq->finished_idx] = NULL;
+ }
+ curr_desc->word0.bits32 = word0.bits32;
+ swtxq->curr_finished_desc = (GMAC_TXDESC_T *)curr_desc;
+ swtxq->total_finished++;
+ tp->ifStatics.tx_packets++;
+ swtxq->finished_idx = RWPTR_ADVANCE_ONE(swtxq->finished_idx, swtxq->total_desc_num);
+ }
+ else
+ {
+ // tp->ifStatics.tx_errors++;
+ // printk("%s::Tx Descriptor is !!!\n",__func__);
+ // wait ready by breaking
+ break;
+ }
+ }
+
+ if (netif_queue_stopped(dev))
+ {
+ netif_wake_queue(dev);
+ }
+}
+
+/*----------------------------------------------------------------------
+* gmac_start_xmit
+*----------------------------------------------------------------------*/
+static int gmac_start_xmit(struct sk_buff *skb, struct net_device *dev)
+{
+ GMAC_INFO_T *tp= dev->priv;
+// static unsigned int pcount = 0;
+// unsigned int tx_qid;
+ DMA_RWPTR_T rwptr;
+ volatile GMAC_TXDESC_T *curr_desc;
+ int snd_pages = skb_shinfo(skb)->nr_frags + 1; /* get number of descriptor */
+ int frag_id = 0;
+ int len, total_len = skb->len;
+ struct net_device_stats *isPtr;
+ unsigned int free_desc;
+ GMAC_SWTXQ_T *swtxq;
+ register unsigned long word0, word1, word2, word3;
+ unsigned short wptr, rptr;
+#ifdef L2_jumbo_frame
+ int header_len = skb->len;
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ int tcp_hdr_len;
+ unsigned char *ptr;
+ int data_len,a;
+ unsigned int val;
+#endif
+
+#ifdef GMAC_LEN_1_2_ISSUE
+ int total_pages;
+ total_pages = snd_pages;
+#endif
+
+ isPtr = (struct net_device_stats *)&tp->ifStatics;
+#if 1
+ if (skb->len >= 0x10000)
+ {
+// spin_unlock(&tp->tx_mutex);
+ isPtr->tx_dropped++;
+ printk("%s::[GMAC %d] skb->len %d >= 64K\n", __func__, tp->port_id, skb->len);
+ netif_stop_queue(dev);
+ return 1;
+ }
+#endif
+
+#if 0
+ if (storlink_ctl.recvfile==2)
+ {
+ printk("snd_pages=%d skb->len=%d\n",snd_pages,skb->len);
+ }
+#endif
+
+#ifdef GMAC_USE_TXQ0
+ #define tx_qid 0
+#endif
+
+ swtxq = &tp->swtxq[tx_qid];
+
+// spin_lock(&tp->tx_mutex);
+ rwptr.bits32 = readl(swtxq->rwptr_reg);
+ wptr = rwptr.bits.wptr;
+ rptr = rwptr.bits.rptr;
+
+ // check finished desc or empty BD
+ // cannot check by read ptr of RW PTR register,
+ // because the HW complete to send but the SW may NOT handle it
+#ifndef GMAX_TX_INTR_DISABLED
+ if (wptr >= swtxq->finished_idx)
+ free_desc = swtxq->total_desc_num - wptr - 1 + swtxq->finished_idx;
+ else
+ free_desc = swtxq->finished_idx - wptr - 1;
+
+ if (free_desc < snd_pages)
+ {
+// spin_unlock(&tp->tx_mutex);
+ isPtr->tx_dropped++;
+// printk("GMAC %d No available descriptor!\n", tp->port_id);
+ netif_stop_queue(dev);
+ return 1;
+ }
+#else
+ toe_gmac_tx_complete(tp, tx_qid, dev, 0);
+
+ if (wptr >= swtxq->finished_idx)
+ free_desc = swtxq->total_desc_num - wptr - 1 + swtxq->finished_idx;
+ else
+ free_desc = swtxq->finished_idx - wptr - 1;
+ if (free_desc < snd_pages)
+ {
+// spin_unlock(&tp->tx_mutex);
+ isPtr->tx_dropped++;
+// printk("GMAC %d No available descriptor!\n", tp->port_id);
+ netif_stop_queue(dev);
+ return 1;
+ }
+
+#if 0
+ printk("1: free_desc=%d, wptr=%d, finished_idx=%d\n", free_desc, wptr, swtxq->finished_idx);
+ if ((free_desc < (snd_pages << 2)) ||
+ (free_desc < (swtxq->total_desc_num >> 2)))
+ {
+ printk("2: free_desc = %d\n", free_desc);
+ toe_gmac_tx_complete(tp, tx_qid, dev, 0);
+ rwptr.bits32 = readl(swtxq->rwptr_reg);
+ wptr = rwptr.bits.wptr;
+ if (wptr>= swtxq->finished_idx)
+ free_desc = swtxq->total_desc_num - wptr -1 + swtxq->finished_idx;
+ else
+ free_desc = swtxq->finished_idx - wptr - 1;
+ }
+#endif
+#endif
+
+#ifdef L2_jumbo_frame
+// data_len = skb->len - 14 - ip_hdr->ihl *4 - tcp_hdr_len;
+// if ((skb->nh.iph->protocol == __constant_htons(ETH_P_IP)) && ((skb->nh.iph->protocol & 0x00ff) == IPPROTO_TCP))
+// if (skb->nh.iph->protocol == 0x006 && (skb->nh.iph->protocol == __constant_htons(ETH_P_IP)))
+ if (((skb->nh.iph->protocol & 0x00ff) == IPPROTO_TCP))
+ {
+ ip_hdr = (struct iphdr*)(skb->nh.iph);
+ tcp_hdr = (struct tcphdr*)(skb->h.th);
+ tcp_hdr_len = TCPHDRLEN(tcp_hdr) * 4;
+ tcp_hdr_len = TCPHDRLEN(tcp_hdr) * 4;
+
+ if ((skb->h.th->syn) && (tcp_hdr_len > 20))
+ {
+ ptr = (unsigned char *)(tcp_hdr+1);
+ if ((ptr[0] == 0x02) && (ptr[1] == 0x04) && (ptr[2] == 0x07) && (ptr[3] == 0xba)) // 0x07 aa=2016-54=1962 ,0x07ba=2032-54=1978
+ {
+ ptr[2]=0x20; //23
+ ptr[3]=0x00; //00
+ printk("-----> Change MSS to 8K \n" );
+ }
+ }
+ }
+// if ((ip_hdr->protocol & 0x00ff) != IPPROTO_TCP)
+// if ((tcp_hdr_len > 20) && (skb->h.th->syn))
+#endif
+
+
+#if 0
+ if (snd_pages > 1)
+ printk("-----> snd_pages=%d\n", snd_pages);
+ if (total_len > 1514)
+ {
+ printk("-----> total_len=%d\n", total_len);
+ }
+#endif
+
+ while (snd_pages)
+ {
+ char *pkt_datap;
+
+ curr_desc = (GMAC_TXDESC_T *)swtxq->desc_base + wptr;
+// consistent_sync((void *)curr_desc, sizeof(GMAC_TXDESC_T), PCI_DMA_FROMDEVICE);
+#if 0
+//#if (GMAC_DEBUG==1)
+ // if curr_desc->word2.buf_adr !=0 means that the ISR does NOT handle it
+ // if (curr_desc->word2.buf_adr)
+ if (swtxq->tx_skb[wptr])
+ {
+ printk("Error! Stop due to TX descriptor's buffer is not freed!\n");
+ while(1);
+ dev_kfree_skb(swtxq->tx_skb[wptr]);
+ swtxq->tx_skb[wptr] = NULL;
+ }
+#endif
+
+ if (frag_id == 0)
+ {
+#if 0
+ int i;
+ pkt_datap = skb->data;
+ len = total_len;
+ for (i=0; i<skb_shinfo(skb)->nr_frags; i++)
+ {
+ skb_frag_t* frag = &skb_shinfo(skb)->frags[i];
+ len -= frag->size;
+ }
+#else
+ pkt_datap = skb->data;
+ len = total_len - skb->data_len;
+#endif
+ }
+ else
+ {
+ skb_frag_t* frag = &skb_shinfo(skb)->frags[frag_id-1];
+ pkt_datap = page_address(frag->page) + frag->page_offset;
+ len = frag->size;
+ if (len > total_len)
+ {
+ printk("===> Fatal Error! Send Frag size %d > Total Size %d!!!!!\n",
+ len, total_len);
+ }
+ }
+
+ /* set TX descriptor */
+ /* copy packet to descriptor buffer address */
+ // curr_desc->word0.bits32 = len; /* total frame byte count */
+ word0 = len;
+#ifdef L2_jumbo_frame
+ word3 = (dev->mtu+14) | EOFIE_BIT; //2016 ,2032
+#else
+ word3 = 1514 | EOFIE_BIT;
+#endif
+
+#ifdef DO_HW_CHKSUM
+#ifdef L2_jumbo_frame
+ if (total_len >= (dev->mtu+14) && (skb->nh.iph->protocol == 0x011) && skb->nh.iph && (skb->nh.iph->frag_off & __constant_htons(0x3fff)))
+#else
+ if (total_len <= 1514 && ip_hdr(skb) && (ip_hdr(skb)->frag_off & __constant_htons(0x3fff)))
+#endif
+ word1 = total_len |
+ TSS_IP_CHKSUM_BIT |
+ TSS_IPV6_ENABLE_BIT |
+ TSS_MTU_ENABLE_BIT;
+ else
+ word1 = total_len |
+ TSS_UDP_CHKSUM_BIT |
+ TSS_TCP_CHKSUM_BIT |
+ TSS_IP_CHKSUM_BIT |
+ TSS_IPV6_ENABLE_BIT |
+ TSS_MTU_ENABLE_BIT;
+#else
+ word1 = total_len | TSS_MTU_ENABLE_BIT;
+#endif
+ word2 = (unsigned long)__pa(pkt_datap);
+
+ if (frag_id == 0)
+ {
+ word3 |= SOF_BIT; // SOF
+ }
+
+ if (snd_pages == 1)
+ {
+ word3 |= EOF_BIT; // EOF
+ swtxq->tx_skb[wptr] = skb;
+#ifdef CONFIG_SL351x_NAT
+ if (nat_cfg.enabled && sl351x_nat_output(skb, tp->port_id))
+ word1 |= TSS_IP_FIXED_LEN_BIT;
+#endif
+ }
+ else
+ swtxq->tx_skb[wptr] = NULL;
+ // word1 |= TSS_IP_FIXED_LEN_BIT;
+#if 1
+#ifdef CONFIG_SL351x_RXTOE
+ // check if this frame has the mission to enable toe hash entry..
+ // if rx_max_pktsize ==0, do not enable RXTOE
+ if (TCP_SKB_CB(skb)->connection && storlink_ctl.rx_max_pktsize) {
+ set_toeq_hdr(TCP_SKB_CB(skb)->connection, &toe_private_data, dev);
+ }
+#endif
+#endif
+#ifdef _DUMP_TX_TCP_CONTENT
+ if (len < 16 && frag_id && skb->h.th && (skb->h.th->source == __constant_htons(445) || skb->h.th->source == __constant_htons(139)))
+ {
+ int a;
+ char *datap;
+ printk("Tx Desc 0x%x Frag %d Len %d [IP-ID 0x%x] 0x%08x: ", (u32)curr_desc, frag_id, len, htons(skb->nh.iph->id), (u32)pkt_datap);
+ datap = (char *)pkt_datap;
+ for (a=0; a<8 && a<len; a++, datap++)
+ {
+ printk("0x%02x ", *datap);
+ }
+ printk("\n");
+ }
+#endif
+
+#ifdef GMAC_LEN_1_2_ISSUE
+ if ((total_pages!=snd_pages) && (len == 1 || len == 2 ) && ((u32)pkt_datap & 0x03))
+ {
+ memcpy((void *)&_debug_prefetch_buf[_debug_prefetch_cnt][0], pkt_datap, len);
+ pkt_datap = (char *)&_debug_prefetch_buf[_debug_prefetch_cnt][0];
+ word2 = (unsigned long)__pa(pkt_datap);
+ _debug_prefetch_cnt++;
+ if (_debug_prefetch_cnt >= _DEBUG_PREFETCH_NUM)
+ _debug_prefetch_cnt = 0;
+ }
+#endif
+
+ consistent_sync((void *)pkt_datap, len, PCI_DMA_TODEVICE);
+ wmb();
+ curr_desc->word0.bits32 = word0;
+ curr_desc->word1.bits32 = word1;
+ curr_desc->word2.bits32 = word2;
+ curr_desc->word3.bits32 = word3;
+ swtxq->curr_tx_desc = (GMAC_TXDESC_T *)curr_desc;
+// consistent_sync((void *)curr_desc, sizeof(GMAC_TXDESC_T), PCI_DMA_TODEVICE);
+#ifdef _DUMP_TX_TCP_CONTENT
+ if (len < 16 && frag_id && skb->h.th && (skb->h.th->source == __constant_htons(445) || skb->h.th->source == __constant_htons(139)))
+ {
+ int a;
+ char *datap;
+ printk("\t 0x%08x: ", (u32)pkt_datap);
+ datap = (char *)pkt_datap;
+ for (a=0; a<8 && a<len; a++, datap++)
+ {
+ printk("0x%02x ", *datap);
+ }
+ printk("\n");
+ }
+#endif
+ free_desc--;
+ wmb();
+ wptr = RWPTR_ADVANCE_ONE(wptr, swtxq->total_desc_num);
+ frag_id++;
+ snd_pages--;
+ }
+
+ swtxq->total_sent++;
+ SET_WPTR(swtxq->rwptr_reg, wptr);
+ dev->trans_start = jiffies;
+
+
+ // printk("MAC %d Qid %d rwptr = 0x%x, curr_desc=0x%x\n", skb->tx_port_id, tx_qid, rwptr.bits32, curr_desc);
+//#ifdef GMAX_TX_INTR_DISABLED
+// toe_gmac_tx_complete(tp, tx_qid, dev, 0);
+//#endif
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* gmac_set_mac_address
+*----------------------------------------------------------------------*/
+
+static int gmac_set_mac_address(struct net_device *dev, void *addr)
+{
+ GMAC_INFO_T *tp= dev->priv;
+ struct sockaddr *sock;
+ unsigned int reg_val;
+ unsigned int i;
+
+ sock = (struct sockaddr *) addr;
+ for (i = 0; i < 6; i++)
+ {
+ dev->dev_addr[i] = sock->sa_data[i];
+ }
+
+ reg_val = dev->dev_addr[0] + (dev->dev_addr[1]<<8) + (dev->dev_addr[2]<<16) + (dev->dev_addr[3]<<24);
+ gmac_write_reg(tp->base_addr,GMAC_STA_ADD0,reg_val,0xffffffff);
+ reg_val = dev->dev_addr[4] + (dev->dev_addr[5]<<8);
+ gmac_write_reg(tp->base_addr,GMAC_STA_ADD1,reg_val,0x0000ffff);
+ memcpy(&eth_mac[tp->port_id][0],&dev->dev_addr[0],6);
+
+ printk("Storlink %s address = ",dev->name);
+ printk("%02x",dev->dev_addr[0]);
+ printk("%02x",dev->dev_addr[1]);
+ printk("%02x",dev->dev_addr[2]);
+ printk("%02x",dev->dev_addr[3]);
+ printk("%02x",dev->dev_addr[4]);
+ printk("%02x\n",dev->dev_addr[5]);
+
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* gmac_get_mac_address
+* get mac address from FLASH
+*----------------------------------------------------------------------*/
+static void gmac_get_mac_address(void)
+{
+#ifdef CONFIG_MTD
+ extern int get_vlaninfo(vlaninfo* vlan);
+ static vlaninfo vlan[2];
+
+ if (get_vlaninfo(&vlan[0]))
+ {
+ memcpy((void *)&eth_mac[0][0],vlan[0].mac,6);
+ // VLAN_conf[0].vid = vlan[0].vlanid;
+ // VLAN_conf[0].portmap = vlan[0].vlanmap;
+ memcpy((void *)&eth_mac[1][0],vlan[1].mac,6);
+ // VLAN_conf[1].vid = vlan[1].vlanid;
+ // VLAN_conf[1].portmap = vlan[1].vlanmap;
+ }
+#else
+ unsigned int reg_val;
+
+ reg_val = readl(IO_ADDRESS(TOE_GMAC0_BASE)+0xac);
+ eth_mac[0][4] = (reg_val & 0xff00) >> 8;
+ eth_mac[0][5] = reg_val & 0x00ff;
+ reg_val = readl(IO_ADDRESS(SL2312_SECURITY_BASE)+0xac);
+ eth_mac[1][4] = (reg_val & 0xff00) >> 8;
+ eth_mac[1][5] = reg_val & 0x00ff;
+#endif
+ return;
+}
+
+
+/*----------------------------------------------------------------------
+* mac_stop_txdma
+*----------------------------------------------------------------------*/
+void mac_stop_txdma(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+ GMAC_TXDMA_FIRST_DESC_T txdma_busy;
+
+ // wait idle
+ do
+ {
+ txdma_busy.bits32 = gmac_read_reg(tp->dma_base_addr, GMAC_DMA_TX_FIRST_DESC_REG);
+ } while (txdma_busy.bits.td_busy);
+
+ /* program dma control register */
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.rd_enable = 0;
+ dma_ctrl.bits.td_enable = 0;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.rd_enable = 1;
+ dma_ctrl_mask.bits.td_enable = 1;
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+}
+
+/*----------------------------------------------------------------------
+* mac_start_txdma
+*----------------------------------------------------------------------*/
+void mac_start_txdma(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+
+ /* program dma control register */
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.rd_enable = 1;
+ dma_ctrl.bits.td_enable = 1;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.rd_enable = 1;
+ dma_ctrl_mask.bits.td_enable = 1;
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+}
+
+
+/*----------------------------------------------------------------------
+* gmac_get_stats
+*----------------------------------------------------------------------*/
+
+struct net_device_stats * gmac_get_stats(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ // unsigned int flags;
+ unsigned int pkt_drop;
+ unsigned int pkt_error;
+
+ if (netif_running(dev))
+ {
+ /* read H/W counter */
+ // spin_lock_irqsave(&tp->lock,flags);
+ pkt_drop = gmac_read_reg(tp->base_addr,GMAC_IN_DISCARDS);
+ pkt_error = gmac_read_reg(tp->base_addr,GMAC_IN_ERRORS);
+ tp->ifStatics.rx_dropped = tp->ifStatics.rx_dropped + pkt_drop;
+ tp->ifStatics.rx_errors = tp->ifStatics.rx_errors + pkt_error;
+ // spin_unlock_irqrestore(&tp->lock,flags);
+ }
+ return &tp->ifStatics;
+}
+
+
+
+/*----------------------------------------------------------------------
+* mac_get_sw_tx_weight
+*----------------------------------------------------------------------*/
+void mac_get_sw_tx_weight(struct net_device *dev, char *weight)
+{
+ GMAC_TX_WCR1_T sw_weigh;
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+ sw_weigh.bits32 = gmac_read_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_1_REG);
+
+ weight[0] = sw_weigh.bits.sw_tq0;
+ weight[1] = sw_weigh.bits.sw_tq1;
+ weight[2] = sw_weigh.bits.sw_tq2;
+ weight[3] = sw_weigh.bits.sw_tq3;
+ weight[4] = sw_weigh.bits.sw_tq4;
+ weight[5] = sw_weigh.bits.sw_tq5;
+}
+
+/*----------------------------------------------------------------------
+* mac_set_sw_tx_weight
+*----------------------------------------------------------------------*/
+void mac_set_sw_tx_weight(struct net_device *dev, char *weight)
+{
+ GMAC_TX_WCR1_T sw_weigh;
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+ sw_weigh.bits32 = 0;
+ sw_weigh.bits.sw_tq0 = weight[0];
+ sw_weigh.bits.sw_tq1 = weight[1];
+ sw_weigh.bits.sw_tq2 = weight[2];
+ sw_weigh.bits.sw_tq3 = weight[3];
+ sw_weigh.bits.sw_tq4 = weight[4];
+ sw_weigh.bits.sw_tq5 = weight[5];
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_1_REG, sw_weigh.bits32, 0xffffffff);
+}
+
+/*----------------------------------------------------------------------
+* mac_get_hw_tx_weight
+*----------------------------------------------------------------------*/
+void mac_get_hw_tx_weight(struct net_device *dev, char *weight)
+{
+ GMAC_TX_WCR0_T hw_weigh;
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+ hw_weigh.bits32 = gmac_read_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_0_REG);
+
+ weight[0] = hw_weigh.bits.hw_tq0;
+ weight[1] = hw_weigh.bits.hw_tq1;
+ weight[2] = hw_weigh.bits.hw_tq2;
+ weight[3] = hw_weigh.bits.hw_tq3;
+}
+
+/*----------------------------------------------------------------------
+* mac_set_hw_tx_weight
+*----------------------------------------------------------------------*/
+void mac_set_hw_tx_weight(struct net_device *dev, char *weight)
+{
+ GMAC_TX_WCR0_T hw_weigh;
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+ hw_weigh.bits32 = 0;
+ hw_weigh.bits.hw_tq0 = weight[0];
+ hw_weigh.bits.hw_tq1 = weight[1];
+ hw_weigh.bits.hw_tq2 = weight[2];
+ hw_weigh.bits.hw_tq3 = weight[3];
+
+ gmac_write_reg(tp->dma_base_addr, GMAC_TX_WEIGHTING_CTRL_0_REG, hw_weigh.bits32, 0xffffffff);
+}
+
+/*----------------------------------------------------------------------
+* mac_start_tx_dma
+*----------------------------------------------------------------------*/
+int mac_start_tx_dma(int mac)
+{
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.td_enable = 1;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.td_enable = 1;
+
+ if (mac == 0)
+ gmac_write_reg(TOE_GMAC0_DMA_BASE, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+ else
+ gmac_write_reg(TOE_GMAC1_DMA_BASE, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+ return 1;
+}
+
+/*----------------------------------------------------------------------
+* mac_stop_tx_dma
+*----------------------------------------------------------------------*/
+int mac_stop_tx_dma(int mac)
+{
+ GMAC_DMA_CTRL_T dma_ctrl, dma_ctrl_mask;
+
+ dma_ctrl.bits32 = 0;
+ dma_ctrl.bits.td_enable = 0;
+
+ dma_ctrl_mask.bits32 = 0;
+ dma_ctrl_mask.bits.td_enable = 1;
+
+ if (mac == 0)
+ gmac_write_reg(TOE_GMAC0_DMA_BASE, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+ else
+ gmac_write_reg(TOE_GMAC1_DMA_BASE, GMAC_DMA_CTRL_REG, dma_ctrl.bits32, dma_ctrl_mask.bits32);
+ return 1;
+}
+
+/*----------------------------------------------------------------------
+* mac_read_reg(int mac, unsigned int offset)
+*----------------------------------------------------------------------*/
+unsigned int mac_read_reg(int mac, unsigned int offset)
+{
+ switch (mac)
+ {
+ case 0:
+ return gmac_read_reg(TOE_GMAC0_BASE, offset);
+ case 1:
+ return gmac_read_reg(TOE_GMAC1_BASE, offset);
+ default:
+ return 0;
+ }
+}
+
+/*----------------------------------------------------------------------
+* mac_write_reg
+*----------------------------------------------------------------------*/
+void mac_write_reg(int mac, unsigned int offset, unsigned data)
+{
+ switch (mac)
+ {
+ case 0:
+ gmac_write_reg(GMAC0_BASE, offset, data, 0xffffffff);
+ break;
+ case 1:
+ gmac_write_reg(GMAC1_BASE, offset, data, 0xffffffff);
+ break;
+ }
+}
+
+/*----------------------------------------------------------------------
+* mac_read_dma_reg(int mac, unsigned int offset)
+*----------------------------------------------------------------------*/
+u32 mac_read_dma_reg(int mac, unsigned int offset)
+{
+ switch (mac)
+ {
+ case 0:
+ return gmac_read_reg(TOE_GMAC0_DMA_BASE, offset);
+ case 1:
+ return gmac_read_reg(TOE_GMAC1_DMA_BASE, offset);
+ default:
+ return 0;
+ }
+}
+
+/*----------------------------------------------------------------------
+* mac_write_dma_reg
+*----------------------------------------------------------------------*/
+void mac_write_dma_reg(int mac, unsigned int offset, u32 data)
+{
+ switch (mac)
+ {
+ case 0:
+ gmac_write_reg(TOE_GMAC0_DMA_BASE, offset, data, 0xffffffff);
+ break;
+ case 1:
+ gmac_write_reg(TOE_GMAC1_DMA_BASE, offset, data, 0xffffffff);
+ break;
+ }
+}
+
+/*----------------------------------------------------------------------
+* ether_crc
+*----------------------------------------------------------------------*/
+static unsigned const ethernet_polynomial = 0x04c11db7U;
+static unsigned int ether_crc (int length, unsigned char *data)
+{
+ int crc = -1;
+ unsigned int i;
+ unsigned int crc_val=0;
+
+ while (--length >= 0) {
+ unsigned char current_octet = *data++;
+ int bit;
+ for (bit = 0; bit < 8; bit++, current_octet >>= 1)
+ crc = (crc << 1) ^ ((crc < 0) ^ (current_octet & 1) ?
+ ethernet_polynomial : 0);
+ }
+ crc = ~crc;
+ for (i=0;i<32;i++)
+ {
+ crc_val = crc_val + (((crc << i) & 0x80000000) >> (31-i));
+ }
+ return crc_val;
+}
+
+
+
+/*----------------------------------------------------------------------
+* mac_set_rx_mode
+*----------------------------------------------------------------------*/
+void mac_set_rx_mode(int pid, unsigned int data)
+{
+ unsigned int base;
+
+ base = (pid == 0) ? GMAC0_BASE : GMAC1_BASE;
+
+ gmac_write_reg(base, GMAC_RX_FLTR, data, 0x0000001f);
+ return;
+}
+
+
+/*----------------------------------------------------------------------
+* gmac_open
+*----------------------------------------------------------------------*/
+
+static int gmac_open (struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ int retval;
+ TOE_INFO_T *toe;
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ /* hook ISR */
+ retval = request_irq (dev->irq, toe_gmac_interrupt, IRQF_DISABLED, dev->name, dev);
+ if (retval)
+ return retval;
+
+ toe_init_gmac(dev);
+
+ if(!FLAG_SWITCH)
+ {
+ init_waitqueue_head (&tp->thr_wait);
+ init_completion(&tp->thr_exited);
+
+ tp->time_to_die = 0;
+ tp->thr_pid = kernel_thread (gmac_phy_thread, dev, CLONE_FS | CLONE_FILES);
+ if (tp->thr_pid < 0)
+ {
+ printk (KERN_WARNING "%s: unable to start kernel thread\n",dev->name);
+ }
+ }
+
+ tp->operation = 1;
+
+ netif_start_queue (dev);
+
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* gmac_close
+*----------------------------------------------------------------------*/
+static int gmac_close(struct net_device *dev)
+{
+ TOE_INFO_T *toe;
+// GMAC_RXDESC_T *sw_desc_ptr,*desc_ptr;
+// unsigned int buf_ptr;
+ GMAC_INFO_T *tp = dev->priv;
+ unsigned int ret;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ tp->operation = 0;
+
+ netif_stop_queue(dev);
+ mdelay(20);
+
+ /* stop tx/rx packet */
+ toe_gmac_disable_tx_rx(dev);
+ mdelay(20);
+
+ /* stop the chip's Tx and Rx DMA processes */
+ toe_gmac_hw_stop(dev);
+
+ toe_gmac_disable_interrupt(tp->irq);
+
+ /* disable interrupts by clearing the interrupt mask */
+ synchronize_irq();
+ free_irq(dev->irq,dev);
+
+// DMA_MFREE(sw_desc_ptr, (TOE_SW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T),(dma_addr_t *)&toe->sw_freeq_desc_base_dma);
+// DMA_MFREE(desc_ptr, TOE_HW_FREEQ_DESC_NUM * sizeof(GMAC_RXDESC_T),(dma_addr_t *)&toe->hw_freeq_desc_base_dma);
+// DMA_MFREE(buf_ptr, TOE_HW_FREEQ_DESC_NUM) * HW_RX_BUF_SIZE),(dma_addr_t *)&toe->hwfq_buf_base_dma);
+// DMA_MFREE(toe->gmac[0].swtxq_desc_base , TOE_GMAC0_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[0].swtxq_desc_base_dma);
+// DMA_MFREE(toe->gmac[1].swtxq_desc_base , TOE_GMAC0_SWTXQ_DESC_NUM * TOE_SW_TXQ_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[1].swtxq_desc_base_dma);
+// DMA_MFREE(toe->gmac[0].hwtxq_desc_base_dma , TOE_GMAC0_HWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[0].hwtxq_desc_base_dma);
+// DMA_MFREE(toe->gmac[1].hwtxq_desc_base_dma , TOE_GMAC0_SWTXQ_DESC_NUM * TOE_HW_TXQ_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[1].hwtxq_desc_base_dma);
+// DMA_MFREE(toe->gmac[0].default_desc_base_dma ,TOE_DEFAULT_Q0_DESC_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[0].default_desc_base_dma);
+// DMA_MFREE(toe->gmac[1].default_desc_base_dma , TOE_DEFAULT_Q0_DESC_NUM * sizeof(GMAC_TXDESC_T),(dma_addr_t *)&toe->gmac[1].default_desc_base_dma);
+// DMA_MFREE(toe->intr_desc_base_dma , TOE_INTR_QUEUE_NUM * TOE_INTR_DESC_NUM * sizeof(GMAC_RXDESC_T),(dma_addr_t *)&toe->intr_desc_base_dma);
+// DMA_MFREE(toe->intr_buf_base_dma , TOE_INTR_DESC_NUM * sizeof(TOE_QHDR_T),(dma_addr_t *)&toe->intr_buf_base_dma);
+
+ if(!FLAG_SWITCH)
+ {
+ if (tp->thr_pid >= 0)
+ {
+ tp->time_to_die = 1;
+ wmb();
+ ret = kill_proc (tp->thr_pid, SIGTERM, 1);
+ if (ret)
+ {
+ printk (KERN_ERR "%s: unable to signal thread\n", dev->name);
+ return ret;
+ }
+// wait_for_completion (&tp->thr_exited);
+ }
+ }
+
+ return (0);
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_fill_free_q
+* allocate buffers for free queue.
+*----------------------------------------------------------------------*/
+static inline void toe_gmac_fill_free_q(void)
+{
+ struct sk_buff *skb;
+ volatile DMA_RWPTR_T fq_rwptr;
+ volatile GMAC_RXDESC_T *fq_desc;
+ unsigned long flags;
+ // unsigned short max_cnt=TOE_SW_FREEQ_DESC_NUM>>1;
+
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ spin_lock_irqsave(&gmac_fq_lock, flags);
+ //while ((max_cnt--) && (unsigned short)RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr,
+ // TOE_SW_FREEQ_DESC_NUM) != fq_rwptr.bits.rptr) {
+ while ((unsigned short)RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr,
+ TOE_SW_FREEQ_DESC_NUM) != fq_rwptr.bits.rptr) {
+ if ((skb = dev_alloc_skb(SW_RX_BUF_SIZE)) == NULL) {
+ printk("%s::skb allocation fail!\n", __func__);
+ //while(1);
+ break;
+ }
+ REG32(skb->data) = (unsigned int)skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES);
+ // fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ fq_rwptr.bits.wptr = RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr,
+ TOE_SW_FREEQ_DESC_NUM);
+ fq_desc = (GMAC_RXDESC_T*)toe_private_data.swfq_desc_base+fq_rwptr.bits.wptr;
+ fq_desc->word2.buf_adr = (unsigned int)__pa(skb->data);
+ SET_WPTR(TOE_GLOBAL_BASE+GLOBAL_SWFQ_RWPTR_REG, fq_rwptr.bits.wptr);
+ toe_private_data.fq_rx_rwptr.bits32 = fq_rwptr.bits32;
+ }
+ spin_unlock_irqrestore(&gmac_fq_lock, flags);
+}
+// EXPORT_SYMBOL(toe_gmac_fill_free_q);
+
+/*----------------------------------------------------------------------
+* toe_gmac_interrupt
+*----------------------------------------------------------------------*/
+static irqreturn_t toe_gmac_interrupt (int irq, void *dev_instance)
+{
+ struct net_device *dev = (struct net_device *)dev_instance;
+ TOE_INFO_T *toe;
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ unsigned int status0;
+ unsigned int status1;
+ unsigned int status2;
+ unsigned int status3;
+ unsigned int status4;
+
+// struct net_device_stats *isPtr = (struct net_device_stats *)&tp->ifStatics;
+ toe = (TOE_INFO_T *)&toe_private_data;
+// handle NAPI
+#ifdef CONFIG_SL_NAPI
+if (storlink_ctl.pauseoff == 1)
+{
+/* disable GMAC interrupt */
+ //toe_gmac_disable_interrupt(tp->irq);
+
+// isPtr->interrupts++;
+ /* read Interrupt status */
+ status0 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_0_REG);
+ status1 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_1_REG);
+ status2 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_2_REG);
+ status3 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_3_REG);
+ status4 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+ // prompt warning if status bit ON but not enabled
+#if 0
+ if (status0 & ~tp->intr0_enabled)
+ printk("Intr 0 Status error. status = 0x%X, enable = 0x%X\n",
+ status0, tp->intr0_enabled);
+ if (status1 & ~tp->intr1_enabled)
+ printk("Intr 1 Status error. status = 0x%X, enable = 0x%X\n",
+ status1, tp->intr1_enabled);
+ if (status2 & ~tp->intr2_enabled)
+ printk("Intr 2 Status error. status = 0x%X, enable = 0x%X\n",
+ status2, tp->intr2_enabled);
+ if (status3 & ~tp->intr3_enabled)
+ printk("Intr 3 Status error. status = 0x%X, enable = 0x%X\n",
+ status3, tp->intr3_enabled);
+ if (status4 & ~tp->intr4_enabled)
+ printk("Intr 4 Status error. status = 0x%X, enable = 0x%X\n",
+ status4, tp->intr4_enabled);
+#endif
+
+ if (status0)
+ writel(status0 & tp->intr0_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_0_REG);
+ if (status1)
+ writel(status1 & tp->intr1_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_1_REG);
+ if (status2)
+ writel(status2 & tp->intr2_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_2_REG);
+ if (status3)
+ writel(status3 & tp->intr3_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_3_REG);
+ if (status4)
+ writel(status4 & tp->intr4_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_4_REG);
+#if 0
+ /* handle freeq interrupt first */
+ if (status4 & tp->intr4_enabled) {
+ if ((status4 & SWFQ_EMPTY_INT_BIT) && (tp->intr4_enabled & SWFQ_EMPTY_INT_BIT))
+ {
+ // unsigned long data = REG32(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ //gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_4_REG,
+ // tp->intr4_enabled & ~SWFQ_EMPTY_INT_BIT, SWFQ_EMPTY_INT_BIT);
+
+ if (toe->gmac[0].dev && netif_running(toe->gmac[0].dev))
+ toe_gmac_handle_default_rxq(toe->gmac[0].dev,&toe->gmac[0]);
+ if (toe->gmac[1].dev && netif_running(toe->gmac[1].dev))
+ toe_gmac_handle_default_rxq(toe->gmac[1].dev,&toe->gmac[1]);
+ printk("\nfreeq int\n");
+ toe_gmac_fill_free_q();
+ tp->sw_fq_empty_cnt++;
+
+ }
+ }
+#endif
+ // Interrupt Status 1
+ if (status1 & tp->intr1_enabled)
+ {
+ #define G1_INTR0_BITS (GMAC1_HWTQ13_EOF_INT_BIT | GMAC1_HWTQ12_EOF_INT_BIT | GMAC1_HWTQ11_EOF_INT_BIT | GMAC1_HWTQ10_EOF_INT_BIT)
+ #define G0_INTR0_BITS (GMAC0_HWTQ03_EOF_INT_BIT | GMAC0_HWTQ02_EOF_INT_BIT | GMAC0_HWTQ01_EOF_INT_BIT | GMAC0_HWTQ00_EOF_INT_BIT)
+ // Handle GMAC 0/1 HW Tx queue 0-3 EOF events
+ // Only count
+ // TOE, Classification, and default queues interrupts are handled by ISR
+ // because they should pass packets to upper layer
+ if (tp->port_id == 0)
+ {
+ if (netif_running(dev) && (status1 & G0_INTR0_BITS) && (tp->intr1_enabled & G0_INTR0_BITS))
+ {
+ if (status1 & GMAC0_HWTQ03_EOF_INT_BIT)
+ tp->hwtxq[3].eof_cnt++;
+ if (status1 & GMAC0_HWTQ02_EOF_INT_BIT)
+ tp->hwtxq[2].eof_cnt++;
+ if (status1 & GMAC0_HWTQ01_EOF_INT_BIT)
+ tp->hwtxq[1].eof_cnt++;
+ if (status1 & GMAC0_HWTQ00_EOF_INT_BIT)
+ tp->hwtxq[0].eof_cnt++;
+ }
+ if (netif_running(dev) && (status1 & DEFAULT_Q0_INT_BIT) && (tp->intr1_enabled & DEFAULT_Q0_INT_BIT))
+ {
+ if (likely(netif_rx_schedule_prep(dev)))
+ {
+ unsigned int data32;
+ // disable GMAC-0 rx interrupt
+ // class-Q & TOE-Q are implemented in future
+ //data32 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //data32 &= ~DEFAULT_Q0_INT_BIT;
+ //writel(data32, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //printk("\%s: DEFAULT_Q0_INT_BIT===================>>>>>>>>>>>>\n",__func__);
+ writel(0x0, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //tp->total_q_cnt_napi=0;
+ //rx_time = jiffies;
+ //rx_old_bytes = isPtr->rx_bytes;
+ __netif_rx_schedule(dev);
+ }
+ }
+ }
+ else if (tp->port_id == 1)
+ {
+ if (netif_running(dev) && (status1 & G1_INTR0_BITS) && (tp->intr1_enabled & G1_INTR0_BITS))
+ {
+ if (status1 & GMAC1_HWTQ13_EOF_INT_BIT)
+ tp->hwtxq[3].eof_cnt++;
+ if (status1 & GMAC1_HWTQ12_EOF_INT_BIT)
+ tp->hwtxq[2].eof_cnt++;
+ if (status1 & GMAC1_HWTQ11_EOF_INT_BIT)
+ tp->hwtxq[1].eof_cnt++;
+ if (status1 & GMAC1_HWTQ10_EOF_INT_BIT)
+ tp->hwtxq[0].eof_cnt++;
+ }
+
+ if (netif_running(dev) && (status1 & DEFAULT_Q1_INT_BIT) && (tp->intr1_enabled & DEFAULT_Q1_INT_BIT))
+ {
+ if (likely(netif_rx_schedule_prep(dev)))
+ {
+ unsigned int data32;
+ // disable GMAC-0 rx interrupt
+ // class-Q & TOE-Q are implemented in future
+ //data32 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //data32 &= ~DEFAULT_Q1_INT_BIT;
+ //writel(data32, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //printk("\%s: 1111111111--->DEFAULT_Q1_INT_BIT===================>>>>>>>>>>>>\n",__func__);
+ writel(0x0, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //tp->total_q_cnt_napi=0;
+ //rx_time = jiffies;
+ //rx_old_bytes = isPtr->rx_bytes;
+ __netif_rx_schedule(dev);
+ }
+ }
+ }
+ }
+
+ // Interrupt Status 0
+ if (status0 & tp->intr0_enabled)
+ {
+ #define ERR_INTR_BITS (GMAC0_TXDERR_INT_BIT | GMAC0_TXPERR_INT_BIT | \
+ GMAC1_TXDERR_INT_BIT | GMAC1_TXPERR_INT_BIT | \
+ GMAC0_RXDERR_INT_BIT | GMAC0_RXPERR_INT_BIT | \
+ GMAC1_RXDERR_INT_BIT | GMAC1_RXPERR_INT_BIT)
+
+ if (status0 & ERR_INTR_BITS)
+ {
+ if ((status0 & GMAC0_TXDERR_INT_BIT) && (tp->intr0_enabled & GMAC0_TXDERR_INT_BIT))
+ {
+ tp->txDerr_cnt[0]++;
+ printk("GMAC0 TX AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC0_TXPERR_INT_BIT) && (tp->intr0_enabled & GMAC0_TXPERR_INT_BIT))
+ {
+ tp->txPerr_cnt[0]++;
+ printk("GMAC0 Tx Descriptor Protocol Error!\n");
+ }
+ if ((status0 & GMAC1_TXDERR_INT_BIT) && (tp->intr0_enabled & GMAC1_TXDERR_INT_BIT))
+ {
+ tp->txDerr_cnt[1]++;
+ printk("GMAC1 Tx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC1_TXPERR_INT_BIT) && (tp->intr0_enabled & GMAC1_TXPERR_INT_BIT))
+ {
+ tp->txPerr_cnt[1]++;
+ printk("GMAC1 Tx Descriptor Protocol Error!\n");
+ }
+
+ if ((status0 & GMAC0_RXDERR_INT_BIT) && (tp->intr0_enabled & GMAC0_RXDERR_INT_BIT))
+ {
+ tp->RxDerr_cnt[0]++;
+ printk("GMAC0 Rx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC0_RXPERR_INT_BIT) && (tp->intr0_enabled & GMAC0_RXPERR_INT_BIT))
+ {
+ tp->RxPerr_cnt[0]++;
+ printk("GMAC0 Rx Descriptor Protocol Error!\n");
+ }
+ if ((status0 & GMAC1_RXDERR_INT_BIT) && (tp->intr0_enabled & GMAC1_RXDERR_INT_BIT))
+ {
+ tp->RxDerr_cnt[1]++;
+ printk("GMAC1 Rx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC1_RXPERR_INT_BIT) && (tp->intr0_enabled & GMAC1_RXPERR_INT_BIT))
+ {
+ tp->RxPerr_cnt[1]++;
+ printk("GMAC1 Rx Descriptor Protocol Error!\n");
+ }
+ }
+
+#ifndef GMAX_TX_INTR_DISABLED
+ if (tp->port_id == 1 && netif_running(dev) &&
+ (((status0 & GMAC1_SWTQ10_FIN_INT_BIT) && (tp->intr0_enabled & GMAC1_SWTQ10_FIN_INT_BIT))
+ ||
+ ((status0 & GMAC1_SWTQ10_EOF_INT_BIT) && (tp->intr0_enabled & GMAC1_SWTQ10_EOF_INT_BIT))))
+ {
+ toe_gmac_tx_complete(&toe_private_data.gmac[1], 0, dev, 1);
+ }
+
+ if (tp->port_id == 0 && netif_running(dev) &&
+ (((status0 & GMAC0_SWTQ00_FIN_INT_BIT) && (tp->intr0_enabled & GMAC0_SWTQ00_FIN_INT_BIT))
+ ||
+ ((status0 & GMAC0_SWTQ00_EOF_INT_BIT) && (tp->intr0_enabled & GMAC0_SWTQ00_EOF_INT_BIT))))
+ {
+ toe_gmac_tx_complete(&toe_private_data.gmac[0], 0, dev, 1);
+ }
+#endif
+ }
+ // Interrupt Status 4
+ if (status4 & tp->intr4_enabled)
+ {
+ #define G1_INTR4_BITS (0xff000000)
+ #define G0_INTR4_BITS (0x00ff0000)
+
+ if (tp->port_id == 0)
+ {
+ if ((status4 & G0_INTR4_BITS) && (tp->intr4_enabled & G0_INTR4_BITS))
+ {
+ if (status4 & GMAC0_RESERVED_INT_BIT)
+ printk("GMAC0_RESERVED_INT_BIT is ON\n");
+ if (status4 & GMAC0_MIB_INT_BIT)
+ tp->mib_full_cnt++;
+ if (status4 & GMAC0_RX_PAUSE_ON_INT_BIT)
+ tp->rx_pause_on_cnt++;
+ if (status4 & GMAC0_TX_PAUSE_ON_INT_BIT)
+ tp->tx_pause_on_cnt++;
+ if (status4 & GMAC0_RX_PAUSE_OFF_INT_BIT)
+ tp->rx_pause_off_cnt++;
+ if (status4 & GMAC0_TX_PAUSE_OFF_INT_BIT)
+ tp->rx_pause_off_cnt++;
+ if (status4 & GMAC0_RX_OVERRUN_INT_BIT)
+ tp->rx_overrun_cnt++;
+ if (status4 & GMAC0_STATUS_CHANGE_INT_BIT)
+ tp->status_changed_cnt++;
+ }
+ }
+ else if (tp->port_id == 1)
+ {
+ if ((status4 & G1_INTR4_BITS) && (tp->intr4_enabled & G1_INTR4_BITS))
+ {
+ if (status4 & GMAC1_RESERVED_INT_BIT)
+ printk("GMAC1_RESERVED_INT_BIT is ON\n");
+ if (status4 & GMAC1_MIB_INT_BIT)
+ tp->mib_full_cnt++;
+ if (status4 & GMAC1_RX_PAUSE_ON_INT_BIT)
+ {
+ printk("Gmac pause on\n");
+ tp->rx_pause_on_cnt++;
+ }
+ if (status4 & GMAC1_TX_PAUSE_ON_INT_BIT)
+ {
+ printk("Gmac pause on\n");
+ tp->tx_pause_on_cnt++;
+ }
+ if (status4 & GMAC1_RX_PAUSE_OFF_INT_BIT)
+ {
+ printk("Gmac pause off\n");
+ tp->rx_pause_off_cnt++;
+ }
+ if (status4 & GMAC1_TX_PAUSE_OFF_INT_BIT)
+ {
+ printk("Gmac pause off\n");
+ tp->rx_pause_off_cnt++;
+ }
+ if (status4 & GMAC1_RX_OVERRUN_INT_BIT)
+ {
+ //printk("Gmac Rx Overrun \n");
+ tp->rx_overrun_cnt++;
+ }
+ if (status4 & GMAC1_STATUS_CHANGE_INT_BIT)
+ tp->status_changed_cnt++;
+ }
+ }
+ }
+
+ //toe_gmac_enable_interrupt(tp->irq);
+#ifdef IxscriptMate_1518
+ if (storlink_ctl.pauseoff == 1)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+#endif
+// enable_irq(gmac_irq[dev_index]);
+ //printk("gmac_interrupt complete!\n\n");
+// return IRQ_RETVAL(handled);
+ return IRQ_RETVAL(1);
+}
+else
+{
+#endif //endif NAPI
+
+
+ /* disable GMAC interrupt */
+ toe_gmac_disable_interrupt(tp->irq);
+
+// isPtr->interrupts++;
+ /* read Interrupt status */
+ status0 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_0_REG);
+ status1 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_1_REG);
+ status2 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_2_REG);
+ status3 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_3_REG);
+ status4 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+ // prompt warning if status bit ON but not enabled
+#if 0
+ if (status0 & ~tp->intr0_enabled)
+ printk("Intr 0 Status error. status = 0x%X, enable = 0x%X\n",
+ status0, tp->intr0_enabled);
+ if (status1 & ~tp->intr1_enabled)
+ printk("Intr 1 Status error. status = 0x%X, enable = 0x%X\n",
+ status1, tp->intr1_enabled);
+ if (status2 & ~tp->intr2_enabled)
+ printk("Intr 2 Status error. status = 0x%X, enable = 0x%X\n",
+ status2, tp->intr2_enabled);
+ if (status3 & ~tp->intr3_enabled)
+ printk("Intr 3 Status error. status = 0x%X, enable = 0x%X\n",
+ status3, tp->intr3_enabled);
+ if (status4 & ~tp->intr4_enabled)
+ printk("Intr 4 Status error. status = 0x%X, enable = 0x%X\n",
+ status4, tp->intr4_enabled);
+#endif
+#define INTERRUPT_SELECT 1
+ if (status0)
+ writel(status0 & tp->intr0_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_0_REG);
+ if (status1)
+ writel(status1 & tp->intr1_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_1_REG);
+ if (status2)
+ writel(status2 & tp->intr2_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_2_REG);
+ if (status3)
+ writel(status3 & tp->intr3_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_3_REG);
+ if (status4)
+ writel(status4 & tp->intr4_enabled, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_STATUS_4_REG);
+
+ /* handle freeq interrupt first */
+ if (status4 & tp->intr4_enabled) {
+ if ((status4 & SWFQ_EMPTY_INT_BIT) && (tp->intr4_enabled & SWFQ_EMPTY_INT_BIT))
+ {
+ // unsigned long data = REG32(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ //gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_4_REG,
+ // tp->intr4_enabled & ~SWFQ_EMPTY_INT_BIT, SWFQ_EMPTY_INT_BIT);
+
+ //gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG,
+ // SWFQ_EMPTY_INT_BIT, SWFQ_EMPTY_INT_BIT);
+ if (toe->gmac[0].dev && netif_running(toe->gmac[0].dev))
+ toe_gmac_handle_default_rxq(toe->gmac[0].dev,&toe->gmac[0]);
+ if (toe->gmac[1].dev && netif_running(toe->gmac[1].dev))
+ toe_gmac_handle_default_rxq(toe->gmac[1].dev,&toe->gmac[1]);
+ printk("\nfreeq int\n");
+ toe_gmac_fill_free_q();
+ tp->sw_fq_empty_cnt++;
+
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG, status4,
+ SWFQ_EMPTY_INT_BIT);
+ }
+ }
+
+ // Interrupt Status 1
+ if (status1 & tp->intr1_enabled)
+ {
+ #define G1_INTR0_BITS (GMAC1_HWTQ13_EOF_INT_BIT | GMAC1_HWTQ12_EOF_INT_BIT | GMAC1_HWTQ11_EOF_INT_BIT | GMAC1_HWTQ10_EOF_INT_BIT)
+ #define G0_INTR0_BITS (GMAC0_HWTQ03_EOF_INT_BIT | GMAC0_HWTQ02_EOF_INT_BIT | GMAC0_HWTQ01_EOF_INT_BIT | GMAC0_HWTQ00_EOF_INT_BIT)
+ // Handle GMAC 0/1 HW Tx queue 0-3 EOF events
+ // Only count
+ // TOE, Classification, and default queues interrupts are handled by ISR
+ // because they should pass packets to upper layer
+ if (tp->port_id == 0)
+ {
+#ifndef INTERRUPT_SELECT
+ if (netif_running(dev) && (status1 & G0_INTR0_BITS) && (tp->intr1_enabled & G0_INTR0_BITS))
+ {
+ if (status1 & GMAC0_HWTQ03_EOF_INT_BIT)
+ tp->hwtxq[3].eof_cnt++;
+ if (status1 & GMAC0_HWTQ02_EOF_INT_BIT)
+ tp->hwtxq[2].eof_cnt++;
+ if (status1 & GMAC0_HWTQ01_EOF_INT_BIT)
+ tp->hwtxq[1].eof_cnt++;
+ if (status1 & GMAC0_HWTQ00_EOF_INT_BIT)
+ tp->hwtxq[0].eof_cnt++;
+#endif //INTERRUPT_SELECT
+#ifndef INTERRUPT_SELECT
+ }
+#endif //INTERRUPT_SELECT
+ if (netif_running(dev) && (status1 & DEFAULT_Q0_INT_BIT) && (tp->intr1_enabled & DEFAULT_Q0_INT_BIT))
+ {
+ tp->default_q_intr_cnt++;
+ toe_gmac_handle_default_rxq(dev, tp);
+ }
+#ifdef CONFIG_SL351x_RXTOE
+ if (netif_running(dev) && (status1 & TOE_IQ_ALL_BITS) &&
+ (tp->intr1_enabled & TOE_IQ_ALL_BITS)) {
+ //printk("status %x, bits %x, slct %x\n", status1, TOE_IQ_ALL_BITS, tp->intr1_selected);
+ toe_gmac_handle_toeq(dev, tp, status1);
+ //toe_gmac_handle_toeq(dev, toe, tp, status1);
+ }
+#endif
+ }
+ else if (tp->port_id == 1)
+ {
+#ifndef INTERRUPT_SELECT
+ if (netif_running(dev) && (status1 & G1_INTR0_BITS) && (tp->intr1_enabled & G1_INTR0_BITS))
+ {
+ if (status1 & GMAC1_HWTQ13_EOF_INT_BIT)
+ tp->hwtxq[3].eof_cnt++;
+ if (status1 & GMAC1_HWTQ12_EOF_INT_BIT)
+ tp->hwtxq[2].eof_cnt++;
+ if (status1 & GMAC1_HWTQ11_EOF_INT_BIT)
+ tp->hwtxq[1].eof_cnt++;
+ if (status1 & GMAC1_HWTQ10_EOF_INT_BIT)
+ tp->hwtxq[0].eof_cnt++;
+#endif //INTERRUPT_SELECT
+#ifndef INTERRUPT_SELECT
+ }
+#endif //INTERRUPT_SELECT
+ if (netif_running(dev) && (status1 & DEFAULT_Q1_INT_BIT) && (tp->intr1_enabled & DEFAULT_Q1_INT_BIT))
+ {
+ tp->default_q_intr_cnt++;
+ toe_gmac_handle_default_rxq(dev, tp);
+ }
+#ifdef CONFIG_SL351x_RXTOE
+ if (netif_running(dev) && (status1 & TOE_IQ_ALL_BITS) &&
+ (tp->intr1_enabled & TOE_IQ_ALL_BITS)) {
+ //printk("status %x, bits %x, slct %x\n", status1, TOE_IQ_ALL_BITS, tp->intr1_selected);
+ toe_gmac_handle_toeq(dev, tp, status1);
+ //toe_gmac_handle_toeq(dev, toe, tp, status1);
+ }
+#endif
+ }
+ }
+
+
+ // Interrupt Status 0
+ if (status0 & tp->intr0_enabled)
+ {
+
+ #define ERR_INTR_BITS (GMAC0_TXDERR_INT_BIT | GMAC0_TXPERR_INT_BIT | \
+ GMAC1_TXDERR_INT_BIT | GMAC1_TXPERR_INT_BIT | \
+ GMAC0_RXDERR_INT_BIT | GMAC0_RXPERR_INT_BIT | \
+ GMAC1_RXDERR_INT_BIT | GMAC1_RXPERR_INT_BIT)
+#ifndef INTERRUPT_SELECT
+ if (status0 & ERR_INTR_BITS)
+ {
+ if ((status0 & GMAC0_TXDERR_INT_BIT) && (tp->intr0_enabled & GMAC0_TXDERR_INT_BIT))
+ {
+ tp->txDerr_cnt[0]++;
+ printk("GMAC0 TX AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC0_TXPERR_INT_BIT) && (tp->intr0_enabled & GMAC0_TXPERR_INT_BIT))
+ {
+ tp->txPerr_cnt[0]++;
+ printk("GMAC0 Tx Descriptor Protocol Error!\n");
+ }
+ if ((status0 & GMAC1_TXDERR_INT_BIT) && (tp->intr0_enabled & GMAC1_TXDERR_INT_BIT))
+ {
+ tp->txDerr_cnt[1]++;
+ printk("GMAC1 Tx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC1_TXPERR_INT_BIT) && (tp->intr0_enabled & GMAC1_TXPERR_INT_BIT))
+ {
+ tp->txPerr_cnt[1]++;
+ printk("GMAC1 Tx Descriptor Protocol Error!\n");
+ }
+
+ if ((status0 & GMAC0_RXDERR_INT_BIT) && (tp->intr0_enabled & GMAC0_RXDERR_INT_BIT))
+ {
+ tp->RxDerr_cnt[0]++;
+ printk("GMAC0 Rx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC0_RXPERR_INT_BIT) && (tp->intr0_enabled & GMAC0_RXPERR_INT_BIT))
+ {
+ tp->RxPerr_cnt[0]++;
+ printk("GMAC0 Rx Descriptor Protocol Error!\n");
+ }
+ if ((status0 & GMAC1_RXDERR_INT_BIT) && (tp->intr0_enabled & GMAC1_RXDERR_INT_BIT))
+ {
+ tp->RxDerr_cnt[1]++;
+ printk("GMAC1 Rx AHB Bus Error!\n");
+ }
+ if ((status0 & GMAC1_RXPERR_INT_BIT) && (tp->intr0_enabled & GMAC1_RXPERR_INT_BIT))
+ {
+ tp->RxPerr_cnt[1]++;
+ printk("GMAC1 Rx Descriptor Protocol Error!\n");
+ }
+ }
+#endif //INTERRUPT_SELECT
+#ifndef GMAX_TX_INTR_DISABLED
+ if (tp->port_id == 1 && netif_running(dev) &&
+ (((status0 & GMAC1_SWTQ10_FIN_INT_BIT) && (tp->intr0_enabled & GMAC1_SWTQ10_FIN_INT_BIT))
+ ||
+ ((status0 & GMAC1_SWTQ10_EOF_INT_BIT) && (tp->intr0_enabled & GMAC1_SWTQ10_EOF_INT_BIT))))
+ {
+ toe_gmac_tx_complete(&toe_private_data.gmac[1], 0, dev, 1);
+ }
+
+ if (tp->port_id == 0 && netif_running(dev) &&
+ (((status0 & GMAC0_SWTQ00_FIN_INT_BIT) && (tp->intr0_enabled & GMAC0_SWTQ00_FIN_INT_BIT))
+ ||
+ ((status0 & GMAC0_SWTQ00_EOF_INT_BIT) && (tp->intr0_enabled & GMAC0_SWTQ00_EOF_INT_BIT))))
+ {
+ toe_gmac_tx_complete(&toe_private_data.gmac[0], 0, dev, 1);
+ }
+#endif
+ // clear enabled status bits
+ }
+ // Interrupt Status 4
+#ifndef INTERRUPT_SELECT
+ if (status4 & tp->intr4_enabled)
+ {
+ #define G1_INTR4_BITS (0xff000000)
+ #define G0_INTR4_BITS (0x00ff0000)
+
+ if (tp->port_id == 0)
+ {
+ if ((status4 & G0_INTR4_BITS) && (tp->intr4_enabled & G0_INTR4_BITS))
+ {
+ if (status4 & GMAC0_RESERVED_INT_BIT)
+ printk("GMAC0_RESERVED_INT_BIT is ON\n");
+ if (status4 & GMAC0_MIB_INT_BIT)
+ tp->mib_full_cnt++;
+ if (status4 & GMAC0_RX_PAUSE_ON_INT_BIT)
+ tp->rx_pause_on_cnt++;
+ if (status4 & GMAC0_TX_PAUSE_ON_INT_BIT)
+ tp->tx_pause_on_cnt++;
+ if (status4 & GMAC0_RX_PAUSE_OFF_INT_BIT)
+ tp->rx_pause_off_cnt++;
+ if (status4 & GMAC0_TX_PAUSE_OFF_INT_BIT)
+ tp->rx_pause_off_cnt++;
+ if (status4 & GMAC0_RX_OVERRUN_INT_BIT)
+ tp->rx_overrun_cnt++;
+ if (status4 & GMAC0_STATUS_CHANGE_INT_BIT)
+ tp->status_changed_cnt++;
+ }
+ }
+ else if (tp->port_id == 1)
+ {
+ if ((status4 & G1_INTR4_BITS) && (tp->intr4_enabled & G1_INTR4_BITS))
+ {
+ if (status4 & GMAC1_RESERVED_INT_BIT)
+ printk("GMAC1_RESERVED_INT_BIT is ON\n");
+ if (status4 & GMAC1_MIB_INT_BIT)
+ tp->mib_full_cnt++;
+ if (status4 & GMAC1_RX_PAUSE_ON_INT_BIT)
+ {
+ //printk("Gmac pause on\n");
+ tp->rx_pause_on_cnt++;
+ }
+ if (status4 & GMAC1_TX_PAUSE_ON_INT_BIT)
+ {
+ //printk("Gmac pause on\n");
+ tp->tx_pause_on_cnt++;
+ }
+ if (status4 & GMAC1_RX_PAUSE_OFF_INT_BIT)
+ {
+ //printk("Gmac pause off\n");
+ tp->rx_pause_off_cnt++;
+ }
+ if (status4 & GMAC1_TX_PAUSE_OFF_INT_BIT)
+ {
+ //printk("Gmac pause off\n");
+ tp->rx_pause_off_cnt++;
+ }
+ if (status4 & GMAC1_RX_OVERRUN_INT_BIT)
+ {
+ //printk("Gmac Rx Overrun \n");
+ tp->rx_overrun_cnt++;
+ }
+ if (status4 & GMAC1_STATUS_CHANGE_INT_BIT)
+ tp->status_changed_cnt++;
+ }
+ }
+#if 0
+ if ((status4 & SWFQ_EMPTY_INT_BIT) && (tp->intr4_enabled & SWFQ_EMPTY_INT_BIT))
+ {
+ // unsigned long data = REG32(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+// mac_stop_rxdma(tp->sc);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_ENABLE_4_REG,
+ tp->intr4_enabled & ~SWFQ_EMPTY_INT_BIT, SWFQ_EMPTY_INT_BIT);
+
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG,
+ SWFQ_EMPTY_INT_BIT, SWFQ_EMPTY_INT_BIT);
+ toe_gmac_fill_free_q();
+ tp->sw_fq_empty_cnt++;
+
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG, status4,
+ SWFQ_EMPTY_INT_BIT);
+//#if 0
+/* if (netif_running(dev))
+ toe_gmac_handle_default_rxq(dev, tp);
+ printk("SWFQ_EMPTY_INT_BIT is ON!\n"); // should not be happened */
+//#endif
+ }
+#endif
+ }
+#endif //INTERRUPT_SELECT
+ toe_gmac_enable_interrupt(tp->irq);
+//enable gmac rx function when do RFC 2544
+#ifdef IxscriptMate_1518
+ if (storlink_ctl.pauseoff == 1)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+#endif
+ //printk("gmac_interrupt complete!\n\n");
+// return IRQ_RETVAL(handled);
+ return IRQ_RETVAL(1);
+#ifdef CONFIG_SL_NAPI
+}
+#endif
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_handle_default_rxq
+* (1) Get rx Buffer for default Rx queue
+* (2) notify or call upper-routine to handle it
+* (3) get a new buffer and insert it into SW free queue
+* (4) Note: The SW free queue Read-Write Pointer should be locked when accessing
+*----------------------------------------------------------------------*/
+//static inline void toe_gmac_handle_default_rxq(struct net_device *dev, GMAC_INFO_T *tp)
+static void toe_gmac_handle_default_rxq(struct net_device *dev, GMAC_INFO_T *tp)
+{
+ TOE_INFO_T *toe;
+ GMAC_RXDESC_T *curr_desc;
+ struct sk_buff *skb;
+ DMA_RWPTR_T rwptr;
+ unsigned int pkt_size;
+ int max_cnt;
+ unsigned int desc_count;
+ unsigned int good_frame, chksum_status, rx_status;
+ struct net_device_stats *isPtr = (struct net_device_stats *)&tp->ifStatics;
+
+//when do ixia RFC 2544 test and packet size is select 1518 bytes,disable gmace rx function immediately after one interrupt come in.
+#ifdef IxscriptMate_1518
+ if (storlink_ctl.pauseoff == 1)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+#endif
+ rwptr.bits32 = readl(&tp->default_qhdr->word1);
+#if 0
+ if (rwptr.bits.rptr != tp->rx_rwptr.bits.rptr)
+ {
+ mac_stop_txdma((struct net_device *)tp->dev);
+ printk("Default Queue HW RD ptr (0x%x) != SW RD Ptr (0x%x)\n",
+ rwptr.bits32, tp->rx_rwptr.bits.rptr);
+ while(1);
+ }
+#endif
+ toe = (TOE_INFO_T *)&toe_private_data;
+ max_cnt = DEFAULT_RXQ_MAX_CNT;
+ while ((--max_cnt) && rwptr.bits.rptr != rwptr.bits.wptr)
+// while (rwptr.bits.rptr != rwptr.bits.wptr)
+ {
+//if packet size is not 1518 for RFC 2544,enable gmac rx function.The other packet size have RX workaround.
+#ifdef IxscriptMate_1518
+ if (storlink_ctl.pauseoff == 1)
+ {
+ if (pkt_size != 1514)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+ }
+#endif
+ curr_desc = (GMAC_RXDESC_T *)tp->default_desc_base + rwptr.bits.rptr;
+// consistent_sync(curr_desc, sizeof(GMAC_RXDESC_T), PCI_DMA_FROMDEVICE);
+ tp->default_q_cnt++;
+ tp->rx_curr_desc = (unsigned int)curr_desc;
+ rx_status = curr_desc->word0.bits.status;
+ chksum_status = curr_desc->word0.bits.chksum_status;
+ tp->rx_status_cnt[rx_status]++;
+ tp->rx_chksum_cnt[chksum_status]++;
+ pkt_size = curr_desc->word1.bits.byte_count; /*total byte count in a frame*/
+ desc_count = curr_desc->word0.bits.desc_count; /* get descriptor count per frame */
+ good_frame=1;
+ if ((curr_desc->word0.bits32 & (GMAC_RXDESC_0_T_derr | GMAC_RXDESC_0_T_perr))
+ || (pkt_size < 60)
+ || (chksum_status & 0x4)
+ || rx_status)
+ {
+ good_frame = 0;
+ if (curr_desc->word0.bits32 & GMAC_RXDESC_0_T_derr)
+ printk("%s::derr (GMAC-%d)!!!\n", __func__, tp->port_id);
+ if (curr_desc->word0.bits32 & GMAC_RXDESC_0_T_perr)
+ printk("%s::perr (GMAC-%d)!!!\n", __func__, tp->port_id);
+ if (rx_status)
+ {
+ if (rx_status == 4 || rx_status == 7)
+ isPtr->rx_crc_errors++;
+// printk("%s::Status=%d (GMAC-%d)!!!\n", __func__, rx_status, tp->port_id);
+ }
+#ifdef SL351x_GMAC_WORKAROUND
+ else if (pkt_size < 60)
+ {
+ if (tp->short_frames_cnt < GMAC_SHORT_FRAME_THRESHOLD)
+ tp->short_frames_cnt++;
+ if (tp->short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+ }
+#endif
+// if (chksum_status)
+// printk("%s::Checksum Status=%d (GMAC-%d)!!!\n", __func__, chksum_status, tp->port_id);
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ dev_kfree_skb_irq(skb);
+ }
+ if (good_frame)
+ {
+ if (curr_desc->word0.bits.drop)
+ printk("%s::Drop (GMAC-%d)!!!\n", __func__, tp->port_id);
+// if (chksum_status)
+// printk("%s::Checksum Status=%d (GMAC-%d)!!!\n", __func__, chksum_status, tp->port_id);
+
+ /* get frame information from the first descriptor of the frame */
+#ifdef SL351x_GMAC_WORKAROUND
+ if (tp->short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+ tp->short_frames_cnt = 0;
+#endif
+ isPtr->rx_packets++;
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr - SKB_RESERVE_BYTES)));
+ if (!skb)
+ {
+ printk("Fatal Error!!skb==NULL\n");
+ goto next_rx;
+ }
+ tp->curr_rx_skb = skb;
+ // consistent_sync((void *)__va(curr_desc->word2.buf_adr), pkt_size, PCI_DMA_FROMDEVICE);
+
+ // curr_desc->word2.buf_adr = 0;
+
+ skb_reserve (skb, RX_INSERT_BYTES); /* 16 byte align the IP fields. */
+ skb_put(skb, pkt_size);
+ skb->dev = dev;
+ if (chksum_status == RX_CHKSUM_IP_UDP_TCP_OK)
+ {
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+#ifdef CONFIG_SL351x_NAT
+ if (nat_cfg.enabled && curr_desc->word3.bits.l3_offset && curr_desc->word3.bits.l4_offset)
+ {
+ struct iphdr *ip_hdr;
+ ip_hdr = (struct iphdr *)&(skb->data[curr_desc->word3.bits.l3_offset]);
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+#endif
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+#if 0
+#ifdef CONFIG_SL351x_RXTOE
+ if (storlink_ctl.rx_max_pktsize) {
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ int ip_hdrlen;
+
+ ip_hdr = (struct iphdr*)&(skb->data[0]);
+ if ((skb->protocol == __constant_htons(ETH_P_IP)) &&
+ ((ip_hdr->protocol & 0x00ff) == IPPROTO_TCP)) {
+ ip_hdrlen = ip_hdr->ihl << 2;
+ tcp_hdr = (struct tcphdr*)&(skb->data[ip_hdrlen]);
+ if (tcp_hdr->syn) {
+ struct toe_conn* connection = init_toeq(ip_hdr->version,
+ ip_hdr, tcp_hdr, toe, &(skb->data[0]) - 14);
+ TCP_SKB_CB(skb)->connection = connection;
+ // hash_dump_entry(TCP_SKB_CB(skb)->connection->hash_entry_index);
+ // printk("%s::skb data %x, conn %x, mode %x\n",
+ // __func__, skb->data, connection, connection->mode);
+ }
+ }
+ }
+#endif
+#endif
+ }
+ else if (chksum_status == RX_CHKSUM_IP_OK_ONLY)
+ {
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+#ifdef CONFIG_SL351x_NAT
+ if (nat_cfg.enabled && curr_desc->word3.bits.l3_offset && curr_desc->word3.bits.l4_offset)
+ {
+ struct iphdr *ip_hdr;
+ //struct tcphdr *tcp_hdr;
+ ip_hdr = (struct iphdr *)&(skb->data[curr_desc->word3.bits.l3_offset]);
+ //tcp_hdr = (struct tcphdr *)&(skb->data[curr_desc->word3.bits.l4_offset]);
+ if (ip_hdr->protocol == IPPROTO_UDP)
+ {
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+ else if (ip_hdr->protocol == IPPROTO_GRE)
+ {
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+ }
+#endif
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+ }
+ else
+ {
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+ }
+
+ netif_rx(skb); /* socket rx */
+ dev->last_rx = jiffies;
+
+ isPtr->rx_bytes += pkt_size;
+
+ }
+
+next_rx:
+ // advance one for Rx default Q 0/1
+ rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, tp->default_desc_num);
+ SET_RPTR(&tp->default_qhdr->word1, rwptr.bits.rptr);
+ tp->rx_rwptr.bits32 = rwptr.bits32;
+
+ toe_gmac_fill_free_q();
+ }
+}
+
+/*----------------------------------------------------------------------
+* gmac_get_phy_vendor
+*----------------------------------------------------------------------*/
+static unsigned int gmac_get_phy_vendor(int phy_addr)
+{
+ unsigned int reg_val;
+ reg_val=(mii_read(phy_addr,0x02) << 16) + mii_read(phy_addr,0x03);
+ return reg_val;
+}
+
+/*----------------------------------------------------------------------
+* gmac_set_phy_status
+*----------------------------------------------------------------------*/
+void gmac_set_phy_status(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_STATUS_T status;
+ unsigned int reg_val, ability,wan_port_id;
+ unsigned int i = 0;
+
+#ifdef VITESSE_G5SWITCH
+ if((tp->port_id == GMAC_PORT1)&&(Giga_switch==1)){
+#if 0
+ rcv_mask = SPI_read(2,0,0x10); // Receive mask
+ rcv_mask |= 0x4F;
+ for(i=0;i<4;i++){
+ reg_val = BIT(26)|(i<<21)|(10<<16);
+ SPI_write(3,0,1,reg_val);
+ msleep(10);
+ reg_val = SPI_read(3,0,2);
+ if(reg_val & 0x0c00){
+ printk("Port%d:Giga mode\n",i);
+ SPI_write(1,i,0x00,0x300701B1);
+ SPI_write(1,i,0x00,0x10070181);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_1000;
+ }
+ else{
+ reg_val = BIT(26)|(i<<21)|(5<<16);
+ SPI_write(3,0,1,reg_val);
+ msleep(10);
+ ability = (reg_val = SPI_read(3,0,2)&0x5e0) >>5;
+ if ((ability & 0x0C)) /* 100M full duplex */
+ {
+ SPI_write(1,i,0x00,0x30050472);
+ SPI_write(1,i,0x00,0x10050442);
+ printk("Port%d:100M\n",i);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_100;
+ }
+ else if((ability & 0x03)) /* 10M full duplex */
+ {
+ SPI_write(1,i,0x00,0x30050473);
+ SPI_write(1,i,0x00,0x10050443);
+ printk("Port%d:10M\n",i);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_10;
+ }
+ else{
+ SPI_write(1,i,0x00,BIT(16)); // disable RX
+ SPI_write(5,0,0x0E,BIT(i)); // dicard packet
+ while((SPI_read(5,0,0x0C)&BIT(i))==0) // wait to be empty
+ msleep(1);
+
+ SPI_write(1,i,0x00,0x20000030); // PORT_RST
+ switch_pre_link[i]=LINK_DOWN;
+ switch_pre_speed[i]=GMAC_SPEED_10;
+ rcv_mask &= ~BIT(i);
+ SPI_write(2,0,0x10,rcv_mask); // Disable Receive
+ }
+ }
+ }
+#endif
+ gmac_get_switch_status(dev);
+ gmac_write_reg(tp->base_addr, GMAC_STATUS, 0x7d, 0x0000007f);
+// SPI_write(2,0,0x10,rcv_mask); // Enable Receive
+ return ;
+ }
+#endif
+
+ reg_val = gmac_get_phy_vendor(tp->phy_addr);
+ printk("GMAC-%d Addr %d Vendor ID: 0x%08x\n", tp->port_id, tp->phy_addr, reg_val);
+
+ switch (tp->phy_mode)
+ {
+ case GMAC_PHY_GMII:
+ mii_write(tp->phy_addr,0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+ #ifdef CONFIG_SL3516_ASIC
+ mii_write(tp->phy_addr,0x09,0x0300); /* advertise 1000M full/half duplex */
+ #else
+ mii_write(tp->phy_addr,0x09,0x0000); /* advertise no 1000M full/half duplex */
+ #endif
+ break;
+ case GMAC_PHY_RGMII_100:
+ mii_write(tp->phy_addr,0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+ mii_write(tp->phy_addr,0x09,0x0000); /* advertise no 1000M */
+ break;
+ case GMAC_PHY_RGMII_1000:
+ mii_write(tp->phy_addr,0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+ #ifdef CONFIG_SL3516_ASIC
+ mii_write(tp->phy_addr,0x09,0x0300); /* advertise 1000M full/half duplex */
+ #else
+ mii_write(tp->phy_addr,0x09,0x0000); /* advertise no 1000M full/half duplex */
+ #endif
+ break;
+ case GMAC_PHY_MII:
+ default:
+ mii_write(tp->phy_addr,0x04,0x05e1); /* advertisement 100M full duplex, pause capable on */
+ mii_write(tp->phy_addr,0x09,0x0000); /* advertise no 1000M */
+ break;
+ }
+
+ mii_write(tp->phy_addr,0x18,0x0041); // Phy active led
+ if (tp->auto_nego_cfg)
+ {
+ reg_val = 0x1200 | (1 << 15);
+ mii_write(tp->phy_addr,0x00,reg_val); /* Enable and Restart Auto-Negotiation */
+ mdelay(500);
+ reg_val &= ~(1 << 15);
+ mii_write(tp->phy_addr, 0x00, reg_val);
+ }
+ else
+ {
+ reg_val = 0;
+ reg_val |= (tp->full_duplex_cfg) ? (1 << 8) : 0;
+ reg_val |= (tp->speed_cfg == GMAC_SPEED_1000) ? (1 << 6) : 0;
+ reg_val |= (tp->speed_cfg == GMAC_SPEED_100) ? (1 << 13) : 0;
+ mii_write(tp->phy_addr, 0x00, reg_val);
+ mdelay(100);
+
+ reg_val |= (1 << 15); // Reset PHY;
+ mii_write(tp->phy_addr, 0x00, reg_val);
+ }
+
+ status.bits32 = 0;
+ /* set PHY operation mode */
+ status.bits.mii_rmii = tp->phy_mode;
+ status.bits.reserved = 1;
+ mdelay(100);
+ while (((reg_val=mii_read(tp->phy_addr,0x01)) & 0x00000004)!=0x04)
+ {
+ msleep(100);
+ i++;
+ if (i > 30)
+ break;
+ }
+ if (i>30)
+ {
+ tp->pre_phy_status = LINK_DOWN;
+ status.bits.link = LINK_DOWN;
+ // clear_bit(__LINK_STATE_START, &dev->state);
+ printk("Link Down (0x%04x) ", reg_val);
+ if(Giga_switch == 1)
+ {
+ wan_port_id = 1;
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ wan_port_id] = 0;
+#endif
+ }
+ else
+ {
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ tp->port_id] = 0;
+#endif
+ }
+ }
+ else
+ {
+ tp->pre_phy_status = LINK_UP;
+ status.bits.link = LINK_UP;
+ // set_bit(__LINK_STATE_START, &dev->state);
+ printk("Link Up (0x%04x) ",reg_val);
+ if(Giga_switch == 1)
+ {
+ wan_port_id = 1;
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ wan_port_id] = 1;
+#endif
+ }
+ else
+ {
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ tp->port_id] = 1;
+#endif
+ }
+ }
+ // value = mii_read(PHY_ADDR,0x05);
+
+ ability = (mii_read(tp->phy_addr,0x05) & 0x05E0) >> 5;
+
+ //#ifdef CONFIG_SL3516_ASIC
+ reg_val = mii_read(tp->phy_addr,10);
+ printk("MII REG 10 = 0x%x\n",reg_val);
+
+ if ((reg_val & 0x0800) == 0x0800)
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 2;
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_100)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_1000;
+
+ printk(" 1000M/Full \n");
+ }
+ else if ((reg_val & 0x0400) == 0x0400)
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 2;
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_100)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_1000;
+
+ printk(" 1000M/Half \n");
+ }
+ //#endif
+ else
+ {
+ #ifdef CONFIG_SL3516_ASIC
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_1000)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_100;
+ #endif
+ printk("MII REG 5 (bit 5:15) = 0x%x\n", ability);
+ if ((ability & 0x08)==0x08) /* 100M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 1;
+ printk(" 100M/Full\n");
+
+ }
+ else if ((ability & 0x04)==0x04) /* 100M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 1;
+ printk(" 100M/Half\n");
+
+ }
+ else if ((ability & 0x02)==0x02) /* 10M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 0;
+ printk(" 10M/Full\n");
+
+ }
+ else if ((ability & 0x01)==0x01) /* 10M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 0;
+ printk(" 10M/Half\n");
+
+ }
+ }
+ if ((ability & 0x20)==0x20)
+ {
+ tp->flow_control_enable = 1;
+ printk("Flow Control Enable.\n");
+ }
+ else
+ {
+ tp->flow_control_enable = 0;
+ printk("Flow Control Disable.\n");
+ }
+ tp->full_duplex_status = status.bits.duplex;
+ tp->speed_status = status.bits.speed;
+ if (!tp->auto_nego_cfg)
+ {
+ status.bits.duplex = tp->full_duplex_cfg;
+ status.bits.speed = tp->speed_cfg;
+ }
+ toe_gmac_disable_tx_rx(dev);
+ mdelay(10);
+ gmac_write_reg(tp->base_addr, GMAC_STATUS, status.bits32, 0x0000007f);
+ toe_gmac_enable_tx_rx(dev);
+}
+
+/*----------------------------------------------------------------------
+* gmac_phy_thread
+*----------------------------------------------------------------------*/
+static int gmac_phy_thread (void *data)
+{
+ struct net_device *dev = data;
+ GMAC_INFO_T *tp = dev->priv;
+ unsigned long timeout;
+
+ daemonize("%s", dev->name);
+ allow_signal(SIGTERM);
+// reparent_to_init();
+// spin_lock_irq(&current->sigmask_lock);
+// sigemptyset(&current->blocked);
+// recalc_sigpending(current);
+// spin_unlock_irq(&current->sigmask_lock);
+// strncpy (current->comm, dev->name, sizeof(current->comm) - 1);
+// current->comm[sizeof(current->comm) - 1] = '\0';
+
+ while (1)
+ {
+ timeout = next_tick;
+ do
+ {
+ timeout = interruptible_sleep_on_timeout (&tp->thr_wait, timeout);
+ } while (!signal_pending (current) && (timeout > 0));
+
+ if (signal_pending (current))
+ {
+// spin_lock_irq(&current->sigmask_lock);
+ flush_signals(current);
+// spin_unlock_irq(&current->sigmask_lock);
+ }
+
+ if (tp->time_to_die)
+ break;
+
+ // printk("%s : Polling MAC %d PHY Status...\n",__func__, tp->port_id);
+ rtnl_lock ();
+ if (tp->auto_nego_cfg){
+#ifdef VITESSE_G5SWITCH
+ if((tp->port_id == GMAC_PORT1)&&(Giga_switch==1))
+ gmac_get_switch_status(dev);
+ else
+#endif
+ gmac_get_phy_status(dev); //temp remove
+ }
+ rtnl_unlock ();
+ }
+ complete_and_exit (&tp->thr_exited, 0);
+}
+
+/*----------------------------------------------------------------------
+* gmac_get_switch_status
+*----------------------------------------------------------------------*/
+#ifdef VITESSE_G5SWITCH
+void gmac_get_switch_status(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+ unsigned int switch_port_id;
+ int get_link=0;
+
+ get_link = Get_Set_port_status();
+ if(get_link){ // link
+ if(ever_dwon){
+ ever_dwon = 0;
+ toe_gmac_enable_tx_rx(dev);
+ netif_wake_queue(dev);
+ set_bit(__LINK_STATE_START, &dev->state);
+ }
+ }
+ else{ // all down
+ //printk("All link down\n");
+ ever_dwon=1;
+ netif_stop_queue(dev);
+ toe_gmac_disable_tx_rx(dev);
+ clear_bit(__LINK_STATE_START, &dev->state);
+ }
+
+ if ( tp->port_id == 1 )
+ switch_port_id = 0;
+#ifdef CONFIG_SL351x_SYSCTL
+ if (get_link)
+ {
+ storlink_ctl.link[switch_port_id] = 1;
+ }
+ else
+ {
+ storlink_ctl.link[switch_port_id] = 0;
+ }
+ if (storlink_ctl.pauseoff == 1)
+ {
+ if (tp->flow_control_enable == 1)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 0; /* disable tx flow control */
+ config0.bits.rx_fc_en = 0; /* disable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+ printk("Disable SWITCH Flow Control...\n");
+ }
+ tp->flow_control_enable = 0;
+ }
+ else
+#endif
+ {
+ if (tp->flow_control_enable == 0)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 1; /* enable tx flow control */
+ config0.bits.rx_fc_en = 1; /* enable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+ printk("Enable SWITCH Flow Control...\n");
+ }
+ tp->flow_control_enable = 1;
+ }
+ return ;
+
+}
+#endif
+
+/*----------------------------------------------------------------------
+* gmac_get_phy_status
+*----------------------------------------------------------------------*/
+void gmac_get_phy_status(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = dev->priv;
+ GMAC_CONFIG0_T config0,config0_mask;
+ GMAC_STATUS_T status, old_status;
+ unsigned int reg_val,ability,wan_port_id;
+
+ old_status.bits32 = status.bits32 = gmac_read_reg(tp->base_addr, GMAC_STATUS);
+
+
+ /* read PHY status register */
+ reg_val = mii_read(tp->phy_addr,0x01);
+ if ((reg_val & 0x0024) == 0x0024) /* link is established and auto_negotiate process completed */
+ {
+ ability = (mii_read(tp->phy_addr,0x05) & 0x05E0) >> 5;
+ /* read PHY Auto-Negotiation Link Partner Ability Register */
+ #ifdef CONFIG_SL3516_ASIC
+ reg_val = mii_read(tp->phy_addr,10);
+ if ((reg_val & 0x0800) == 0x0800)
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 2;
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_100)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_1000;
+ }
+ else if ((reg_val & 0x0400) == 0x0400)
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 2;
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_100)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_1000;
+ }
+ else
+ #endif
+ {
+ #ifdef CONFIG_SL3516_ASIC
+ if (status.bits.mii_rmii == GMAC_PHY_RGMII_1000)
+ status.bits.mii_rmii = GMAC_PHY_RGMII_100;
+ #endif
+ if ((ability & 0x08)==0x08) /* 100M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 1;
+ }
+ else if ((ability & 0x04)==0x04) /* 100M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 1;
+ }
+ else if ((ability & 0x02)==0x02) /* 10M full duplex */
+ {
+ status.bits.duplex = 1;
+ status.bits.speed = 0;
+ }
+ else if ((ability & 0x01)==0x01) /* 10M half duplex */
+ {
+ status.bits.duplex = 0;
+ status.bits.speed = 0;
+ }
+ }
+ status.bits.link = LINK_UP; /* link up */
+ if(Giga_switch==1)
+ {
+ wan_port_id = 1;
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ wan_port_id] = 1;
+ }
+ else
+ {
+ storlink_ctl.link[ tp->port_id] = 1;
+#endif
+ }
+ if ((ability & 0x20)==0x20)
+ {
+ if (tp->flow_control_enable == 0)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 1; /* enable tx flow control */
+ config0.bits.rx_fc_en = 1; /* enable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+ printk("GMAC-%d Flow Control Enable.\n", tp->port_id);
+ }
+ tp->flow_control_enable = 1;
+ }
+ else
+ {
+ if (tp->flow_control_enable == 1)
+ {
+ config0.bits32 = 0;
+ config0_mask.bits32 = 0;
+ config0.bits.tx_fc_en = 0; /* disable tx flow control */
+ config0.bits.rx_fc_en = 0; /* disable rx flow control */
+ config0_mask.bits.tx_fc_en = 1;
+ config0_mask.bits.rx_fc_en = 1;
+ gmac_write_reg(tp->base_addr, GMAC_CONFIG0,config0.bits32,config0_mask.bits32);
+ printk("GMAC-%d Flow Control Disable.\n", tp->port_id);
+ }
+ tp->flow_control_enable = 0;
+ }
+
+ if (tp->pre_phy_status == LINK_DOWN)
+ {
+ printk("GMAC-%d LINK_UP......\n",tp->port_id);
+ tp->pre_phy_status = LINK_UP;
+ }
+ }
+ else
+ {
+ status.bits.link = LINK_DOWN; /* link down */
+ if(Giga_switch == 1)
+ {
+ wan_port_id = 1;
+#ifdef CONFIG_SL351x_SYSCTL
+ storlink_ctl.link[ wan_port_id] = 0;
+ }
+ else
+ {
+ storlink_ctl.link[ tp->port_id] = 0;
+#endif
+ }
+ if (tp->pre_phy_status == LINK_UP)
+ {
+ printk("GMAC-%d LINK_Down......\n",tp->port_id);
+ tp->pre_phy_status = LINK_DOWN;
+ }
+ }
+
+ tp->full_duplex_status = status.bits.duplex;
+ tp->speed_status = status.bits.speed;
+ if (!tp->auto_nego_cfg)
+ {
+ status.bits.duplex = tp->full_duplex_cfg;
+ status.bits.speed = tp->speed_cfg;
+ }
+
+ if (old_status.bits32 != status.bits32)
+ {
+ netif_stop_queue(dev);
+ toe_gmac_disable_tx_rx(dev);
+ clear_bit(__LINK_STATE_START, &dev->state);
+ printk("GMAC-%d Change Status Bits 0x%x-->0x%x\n",tp->port_id, old_status.bits32, status.bits32);
+ mdelay(10); // let GMAC consume packet
+ gmac_write_reg(tp->base_addr, GMAC_STATUS, status.bits32, 0x0000007f);
+ if (status.bits.link == LINK_UP)
+ {
+ toe_gmac_enable_tx_rx(dev);
+ netif_wake_queue(dev);
+ set_bit(__LINK_STATE_START, &dev->state);
+ }
+ }
+}
+
+/***************************************/
+/* define GPIO module base address */
+/***************************************/
+#define GPIO_BASE_ADDR (IO_ADDRESS(SL2312_GPIO_BASE))
+#define GPIO_BASE_ADDR1 (IO_ADDRESS(SL2312_GPIO_BASE1))
+
+/* define GPIO pin for MDC/MDIO */
+#ifdef CONFIG_SL3516_ASIC
+#define H_MDC_PIN 22
+#define H_MDIO_PIN 21
+#define G_MDC_PIN 22
+#define G_MDIO_PIN 21
+#else
+#define H_MDC_PIN 3
+#define H_MDIO_PIN 2
+#define G_MDC_PIN 0
+#define G_MDIO_PIN 1
+#endif
+
+//#define GPIO_MDC 0x80000000
+//#define GPIO_MDIO 0x00400000
+
+static unsigned int GPIO_MDC = 0;
+static unsigned int GPIO_MDIO = 0;
+static unsigned int GPIO_MDC_PIN = 0;
+static unsigned int GPIO_MDIO_PIN = 0;
+
+// For PHY test definition!!
+#define LPC_EECK 0x02
+#define LPC_EDIO 0x04
+#define LPC_GPIO_SET 3
+#define LPC_BASE_ADDR IO_ADDRESS(IT8712_IO_BASE)
+#define inb_gpio(x) inb(LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+#define outb_gpio(x, y) outb(y, LPC_BASE_ADDR + IT8712_GPIO_BASE + x)
+
+enum GPIO_REG
+{
+ GPIO_DATA_OUT = 0x00,
+ GPIO_DATA_IN = 0x04,
+ GPIO_PIN_DIR = 0x08,
+ GPIO_BY_PASS = 0x0c,
+ GPIO_DATA_SET = 0x10,
+ GPIO_DATA_CLEAR = 0x14,
+};
+/***********************/
+/* MDC : GPIO[31] */
+/* MDIO: GPIO[22] */
+/***********************/
+
+/***************************************************
+* All the commands should have the frame structure:
+*<PRE><ST><OP><PHYAD><REGAD><TA><DATA><IDLE>
+****************************************************/
+
+/*****************************************************************
+* Inject a bit to NWay register through CSR9_MDC,MDIO
+*******************************************************************/
+void mii_serial_write(char bit_MDO) // write data into mii PHY
+{
+#ifdef CONFIG_SL2312_LPC_IT8712
+ unsigned char iomode,status;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode |= (LPC_EECK|LPC_EDIO) ; // Set EECK,EDIO,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ if(bit_MDO)
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= LPC_EDIO ; //EDIO high
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+ else
+ {
+ status = inb_gpio( LPC_GPIO_SET);
+ status &= ~(LPC_EDIO) ; //EDIO low
+ outb_gpio(LPC_GPIO_SET, status);
+ }
+
+ status |= LPC_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(LPC_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+#else
+ unsigned int addr;
+ unsigned int value;
+
+ addr = GPIO_BASE_ADDR + GPIO_PIN_DIR;
+ value = readl(addr) | GPIO_MDC | GPIO_MDIO; /* set MDC/MDIO Pin to output */
+ writel(value,addr);
+ if(bit_MDO)
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDIO,addr); /* set MDIO to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDIO,addr); /* set MDIO to 0 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+ }
+
+#endif
+}
+
+/**********************************************************************
+* read a bit from NWay register through CSR9_MDC,MDIO
+***********************************************************************/
+unsigned int mii_serial_read(void) // read data from mii PHY
+{
+#ifdef CONFIG_SL2312_LPC_IT8712
+ unsigned char iomode,status;
+ unsigned int value ;
+
+ iomode = LPCGetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET);
+ iomode &= ~(LPC_EDIO) ; // Set EDIO input
+ iomode |= (LPC_EECK) ; // Set EECK,EECS output
+ LPCSetConfig(LDN_GPIO, 0xc8 + LPC_GPIO_SET, iomode);
+
+ status = inb_gpio( LPC_GPIO_SET);
+ status |= LPC_EECK ; //EECK high
+ outb_gpio(LPC_GPIO_SET, status);
+
+ status &= ~(LPC_EECK) ; //EECK low
+ outb_gpio(LPC_GPIO_SET, status);
+
+ value = inb_gpio( LPC_GPIO_SET);
+
+ value = value>>2 ;
+ value &= 0x01;
+
+ return value ;
+
+#else
+ unsigned int *addr;
+ unsigned int value;
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_PIN_DIR);
+ value = readl(addr) & ~GPIO_MDIO; //0xffbfffff; /* set MDC to output and MDIO to input */
+ writel(value,addr);
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_SET);
+ writel(GPIO_MDC,addr); /* set MDC to 1 */
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_CLEAR);
+ writel(GPIO_MDC,addr); /* set MDC to 0 */
+
+ addr = (unsigned int *)(GPIO_BASE_ADDR + GPIO_DATA_IN);
+ value = readl(addr);
+ value = (value & (1<<GPIO_MDIO_PIN)) >> GPIO_MDIO_PIN;
+ return(value);
+
+#endif
+}
+
+/***************************************
+* preamble + ST
+***************************************/
+void mii_pre_st(void)
+{
+ unsigned char i;
+
+ for(i=0;i<32;i++) // PREAMBLE
+ mii_serial_write(1);
+ mii_serial_write(0); // ST
+ mii_serial_write(1);
+}
+
+
+/******************************************
+* Read MII register
+* phyad -> physical address
+* regad -> register address
+***************************************** */
+unsigned int mii_read(unsigned char phyad,unsigned char regad)
+{
+ unsigned int i,value;
+ unsigned int bit;
+
+ if (phyad == GPHY_ADDR)
+ {
+ GPIO_MDC_PIN = G_MDC_PIN; /* assigned MDC pin for giga PHY */
+ GPIO_MDIO_PIN = G_MDIO_PIN; /* assigned MDIO pin for giga PHY */
+ }
+ else
+ {
+ GPIO_MDC_PIN = H_MDC_PIN; /* assigned MDC pin for 10/100 PHY */
+ GPIO_MDIO_PIN = H_MDIO_PIN; /* assigned MDIO pin for 10/100 PHY */
+ }
+ GPIO_MDC = (1<<GPIO_MDC_PIN);
+ GPIO_MDIO = (1<<GPIO_MDIO_PIN);
+
+ mii_pre_st(); // PRE+ST
+ mii_serial_write(1); // OP
+ mii_serial_write(0);
+
+ for (i=0;i<5;i++) { // PHYAD
+ bit= ((phyad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ for (i=0;i<5;i++) { // REGAD
+ bit= ((regad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ mii_serial_read(); // TA_Z
+// if((bit=mii_serial_read()) !=0 ) // TA_0
+// {
+// return(0);
+// }
+ value=0;
+ for (i=0;i<16;i++) { // READ DATA
+ bit=mii_serial_read();
+ value += (bit<<(15-i)) ;
+ }
+
+ mii_serial_write(0); // dumy clock
+ mii_serial_write(0); // dumy clock
+
+ //printk("%s: phy_addr=0x%x reg_addr=0x%x value=0x%x \n",__func__,phyad,regad,value);
+ return(value);
+}
+
+/******************************************
+* Write MII register
+* phyad -> physical address
+* regad -> register address
+* value -> value to be write
+***************************************** */
+void mii_write(unsigned char phyad,unsigned char regad,unsigned int value)
+{
+ unsigned int i;
+ char bit;
+
+ printk("%s: phy_addr=0x%x reg_addr=0x%x value=0x%x \n",__func__,phyad,regad,value);
+ if (phyad == GPHY_ADDR)
+ {
+ GPIO_MDC_PIN = G_MDC_PIN; /* assigned MDC pin for giga PHY */
+ GPIO_MDIO_PIN = G_MDIO_PIN; /* assigned MDIO pin for giga PHY */
+ }
+ else
+ {
+ GPIO_MDC_PIN = H_MDC_PIN; /* assigned MDC pin for 10/100 PHY */
+ GPIO_MDIO_PIN = H_MDIO_PIN; /* assigned MDIO pin for 10/100 PHY */
+ }
+ GPIO_MDC = (1<<GPIO_MDC_PIN);
+ GPIO_MDIO = (1<<GPIO_MDIO_PIN);
+
+ mii_pre_st(); // PRE+ST
+ mii_serial_write(0); // OP
+ mii_serial_write(1);
+ for (i=0;i<5;i++) { // PHYAD
+ bit= ((phyad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+
+ for (i=0;i<5;i++) { // REGAD
+ bit= ((regad>>(4-i)) & 0x01) ? 1 :0 ;
+ mii_serial_write(bit);
+ }
+ mii_serial_write(1); // TA_1
+ mii_serial_write(0); // TA_0
+
+ for (i=0;i<16;i++) { // OUT DATA
+ bit= ((value>>(15-i)) & 0x01) ? 1 : 0 ;
+ mii_serial_write(bit);
+ }
+ mii_serial_write(0); // dumy clock
+ mii_serial_write(0); // dumy clock
+}
+
+/*----------------------------------------------------------------------
+* gmac_set_rx_mode
+*----------------------------------------------------------------------*/
+static void gmac_set_rx_mode(struct net_device *dev)
+{
+ GMAC_RX_FLTR_T filter;
+ unsigned int mc_filter[2]; /* Multicast hash filter */
+ int bit_nr;
+ unsigned int i;
+ GMAC_INFO_T *tp = dev->priv;
+
+// printk("%s : dev->flags = %x \n",__func__,dev->flags);
+// dev->flags |= IFF_ALLMULTI; /* temp */
+ filter.bits32 = 0;
+ filter.bits.error = 0;
+ if (dev->flags & IFF_PROMISC)
+ {
+ filter.bits.error = 1;
+ filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0xffffffff;
+ }
+ else if (dev->flags & IFF_ALLMULTI)
+ {
+// filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0xffffffff;
+ }
+ else
+ {
+ struct dev_mc_list *mclist;
+
+// filter.bits.promiscuous = 1;
+ filter.bits.broadcast = 1;
+ filter.bits.multicast = 1;
+ filter.bits.unicast = 1;
+ mc_filter[1] = mc_filter[0] = 0;
+ for (i = 0, mclist = dev->mc_list; mclist && i < dev->mc_count;i++, mclist = mclist->next)
+ {
+ bit_nr = ether_crc(ETH_ALEN,mclist->dmi_addr) & 0x0000003f;
+ if (bit_nr < 32)
+ {
+ mc_filter[0] = mc_filter[0] | (1<<bit_nr);
+ }
+ else
+ {
+ mc_filter[1] = mc_filter[1] | (1<<(bit_nr-32));
+ }
+ }
+ }
+ gmac_write_reg(tp->base_addr,GMAC_RX_FLTR,filter.bits32,0xffffffff); //chech base address!!!
+ gmac_write_reg(tp->base_addr,GMAC_MCAST_FIL0,mc_filter[0],0xffffffff);
+ gmac_write_reg(tp->base_addr,GMAC_MCAST_FIL1,mc_filter[1],0xffffffff);
+ return;
+}
+
+#ifdef CONFIG_SL_NAPI
+/*----------------------------------------------------------------------
+* gmac_rx_poll
+*----------------------------------------------------------------------*/
+static int gmac_rx_poll(struct net_device *dev, int *budget)
+{
+ TOE_INFO_T *toe;
+ GMAC_RXDESC_T *curr_desc;
+ struct sk_buff *skb;
+ DMA_RWPTR_T rwptr;
+ unsigned int pkt_size;
+ unsigned int desc_count;
+ unsigned int good_frame, chksum_status, rx_status;
+ int rx_pkts_num = 0;
+ int quota = min(dev->quota, *budget);
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ unsigned int status4;
+ volatile DMA_RWPTR_T fq_rwptr;
+ int max_cnt = TOE_SW_FREEQ_DESC_NUM;//TOE_SW_FREEQ_DESC_NUM = 64
+ //unsigned long rx_old_bytes;
+ struct net_device_stats *isPtr = (struct net_device_stats *)&tp->ifStatics;
+ //unsigned long long rx_time;
+
+
+
+#if 1
+ if (do_again)
+ {
+ toe_gmac_fill_free_q();
+ status4 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ //printk("\n%s:: do_again toe_gmac_fill_free_q =======>status4=0x%x =====fq_rwptr =0x%8x======>JKJKJKJKJKJKJKJKJ \n", __func__,status4,fq_rwptr.bits32);
+ if (fq_rwptr.bits.wptr != fq_rwptr.bits.rptr)
+ {
+ //status4 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+ do_again =0;
+ //netif_rx_complete(dev);
+ gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG, status4, 0x1);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ rwptr.bits32 = readl(&tp->default_qhdr->word1);
+ }
+ else
+ return 1;
+ }
+#endif
+ rwptr.bits32 = readl(&tp->default_qhdr->word1);
+#if 0
+ if (rwptr.bits.rptr != tp->rx_rwptr.bits.rptr)
+ {
+ mac_stop_txdma((struct net_device *)tp->dev);
+ printk("Default Queue HW RD ptr (0x%x) != SW RD Ptr (0x%x)\n",
+ rwptr.bits32, tp->rx_rwptr.bits.rptr);
+ while(1);
+ }
+#endif
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ //printk("%s:---Before-------------->Default Queue HW RW ptr (0x%8x), fq_rwptr =0x%8x \n",__func__,rwptr.bits32,fq_rwptr.bits32 );
+ //printk("%s:---Before while rx_pkts_num=%d------rx_finished_idx=0x%x------->Default_Q [rwptr.bits.rptr(SW)=0x%x, rwptr.bits.wptr(HW) = 0x%x ]---->Free_Q(SW_HW) = 0x%8x \n",__func__,rx_pkts_num,rx_finished_idx,rwptr.bits.rptr,rwptr.bits.wptr,fq_rwptr.bits32 );
+// while ((--max_cnt) && (rwptr.bits.rptr != rwptr.bits.wptr) && (rx_pkts_num < quota))
+
+ while ((rwptr.bits.rptr != rwptr.bits.wptr) && (rx_pkts_num < quota))
+ {
+
+ curr_desc = (GMAC_RXDESC_T *)tp->default_desc_base + rwptr.bits.rptr;
+ tp->default_q_cnt++;
+ tp->rx_curr_desc = (unsigned int)curr_desc;
+ rx_status = curr_desc->word0.bits.status;
+ chksum_status = curr_desc->word0.bits.chksum_status;
+ tp->rx_status_cnt[rx_status]++;
+ tp->rx_chksum_cnt[chksum_status]++;
+ pkt_size = curr_desc->word1.bits.byte_count; /*total byte count in a frame*/
+ desc_count = curr_desc->word0.bits.desc_count; /* get descriptor count per frame */
+ good_frame=1;
+ if ((curr_desc->word0.bits32 & (GMAC_RXDESC_0_T_derr | GMAC_RXDESC_0_T_perr))
+ || (pkt_size < 60)
+ || (chksum_status & 0x4)
+ || rx_status )
+// || rx_status || (rwptr.bits.rptr > rwptr.bits.wptr ))
+ {
+ good_frame = 0;
+ if (curr_desc->word0.bits32 & GMAC_RXDESC_0_T_derr)
+ printk("%s::derr (GMAC-%d)!!!\n", __func__, tp->port_id);
+ if (curr_desc->word0.bits32 & GMAC_RXDESC_0_T_perr)
+ printk("%s::perr (GMAC-%d)!!!\n", __func__, tp->port_id);
+ if (rx_status)
+ {
+ if (rx_status == 4 || rx_status == 7)
+ isPtr->rx_crc_errors++;
+// printk("%s::Status=%d (GMAC-%d)!!!\n", __func__, rx_status, tp->port_id);
+ }
+#ifdef SL351x_GMAC_WORKAROUND
+ else if (pkt_size < 60)
+ {
+ if (tp->short_frames_cnt < GMAC_SHORT_FRAME_THRESHOLD)
+ tp->short_frames_cnt++;
+ if (tp->short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+ }
+#endif
+// if (chksum_status)
+// printk("%s::Checksum Status=%d (GMAC-%d)!!!\n", __func__, chksum_status, tp->port_id);
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ dev_kfree_skb_irq(skb);
+ }
+ if (good_frame)
+ {
+ if (curr_desc->word0.bits.drop)
+ printk("%s::Drop (GMAC-%d)!!!\n", __func__, tp->port_id);
+// if (chksum_status)
+// printk("%s::Checksum Status=%d (GMAC-%d)!!!\n", __func__, chksum_status, tp->port_id);
+
+#ifdef SL351x_GMAC_WORKAROUND
+ if (tp->short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ GMAC_CONFIG0_T config0;
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ }
+ tp->short_frames_cnt = 0;
+#endif
+ /* get frame information from the first descriptor of the frame */
+ isPtr->rx_packets++;
+ //consistent_sync((void *)__va(curr_desc->word2.buf_adr), pkt_size, PCI_DMA_FROMDEVICE);
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ tp->curr_rx_skb = skb;
+ // curr_desc->word2.buf_adr = 0;
+
+ //skb_reserve (skb, SKB_RESERVE_BYTES);
+ skb_reserve (skb, RX_INSERT_BYTES); /* 2 byte align the IP fields. */
+ //if ((skb->tail+pkt_size) > skb->end )
+ //printk("%s::------------->Here skb->len=%d,pkt_size= %d,skb->head=0x%x,skb->tail= 0x%x, skb->end= 0x%x\n", __func__, skb->len, pkt_size,skb->head,skb->tail,skb->end);
+ skb_put(skb, pkt_size);
+
+
+ skb->dev = dev;
+ if (chksum_status == RX_CHKSUM_IP_UDP_TCP_OK)
+ {
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+#ifdef CONFIG_SL351x_NAT
+ if (nat_cfg.enabled && curr_desc->word3.bits.l3_offset && curr_desc->word3.bits.l4_offset)
+ {
+ struct iphdr *ip_hdr;
+ ip_hdr = (struct iphdr *)&(skb->data[curr_desc->word3.bits.l3_offset]);
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+#endif
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+#if 0
+#ifdef CONFIG_SL351x_RXTOE
+ if (storlink_ctl.rx_max_pktsize) {
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ int ip_hdrlen;
+
+ ip_hdr = (struct iphdr*)&(skb->data[0]);
+ if ((skb->protocol == __constant_htons(ETH_P_IP)) &&
+ ((ip_hdr->protocol & 0x00ff) == IPPROTO_TCP)) {
+ ip_hdrlen = ip_hdr->ihl << 2;
+ tcp_hdr = (struct tcphdr*)&(skb->data[ip_hdrlen]);
+ if (tcp_hdr->syn) {
+ struct toe_conn* connection = init_toeq(ip_hdr->version,
+ ip_hdr, tcp_hdr, toe, &(skb->data[0]) - 14);
+ TCP_SKB_CB(skb)->connection = connection;
+ // hash_dump_entry(TCP_SKB_CB(skb)->connection->hash_entry_index);
+ // printk("%s::skb data %x, conn %x, mode %x\n",
+ // __func__, skb->data, connection, connection->mode);
+ }
+ }
+ }
+#endif
+#endif
+ }
+ else if (chksum_status == RX_CHKSUM_IP_OK_ONLY)
+ {
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+#ifdef CONFIG_SL351x_NAT
+ if (nat_cfg.enabled && curr_desc->word3.bits.l3_offset && curr_desc->word3.bits.l4_offset)
+ {
+ struct iphdr *ip_hdr;
+ ip_hdr = (struct iphdr *)&(skb->data[curr_desc->word3.bits.l3_offset]);
+ if (ip_hdr->protocol == IPPROTO_UDP)
+ {
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+ else if (ip_hdr->protocol == IPPROTO_GRE)
+ {
+ sl351x_nat_input(skb,
+ tp->port_id,
+ (void *)curr_desc->word3.bits.l3_offset,
+ (void *)curr_desc->word3.bits.l4_offset);
+ }
+ }
+#endif
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+ }
+ else
+ {
+ skb->protocol = eth_type_trans(skb,dev); /* set skb protocol */
+ }
+ //netif_rx(skb); /* socket rx */
+ netif_receive_skb(skb); //For NAPI
+ dev->last_rx = jiffies;
+
+ isPtr->rx_bytes += pkt_size;
+ //printk("------------------->isPtr->rx_bytes = %d\n",isPtr->rx_bytes);
+
+
+ }
+ // advance one for Rx default Q 0/1
+ rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, tp->default_desc_num);
+ SET_RPTR(&tp->default_qhdr->word1, rwptr.bits.rptr);
+ tp->rx_rwptr.bits32 = rwptr.bits32;
+ rx_pkts_num++;
+ //rwptr.bits32 = readl(&tp->default_qhdr->word1);//try read default_qhdr again
+ //fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ //printk("%s:---Loop -------->rx_pkts_num=%d------------>Default Queue HW RW ptr = (0x%8x), fq_rwptr =0x%8x \n",__func__,rx_pkts_num,rwptr.bits32,fq_rwptr.bits32 );
+#if 0
+ if ((status4 & 0x1) == 0)
+ {
+ //if (!((dev->last_rx <= (rx_time + 2)) && (isPtr->rx_bytes > (rx_old_bytes + 1000000 ))))
+ if (tp->total_q_cnt_napi < 1024)
+ {
+ tp->total_q_cnt_napi++;
+ toe_gmac_fill_free_q(); //for iperf test disable
+ }
+ //else
+ //printk("%s:---isPtr->rx_bytes =%u , rx_old_bytes =%u\n",__func__,isPtr->rx_bytes,rx_old_bytes );
+
+ }
+#endif
+ //rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, tp->default_desc_num);
+ //printk("%s:---Loop -------->rx_pkts_num=%d----rwptr.bits.rptr=0x%x-------->Default Queue HW RW ptr = (0x%8x), fq_rwptr =0x%8x \n",__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits32,fq_rwptr.bits32 );
+ //printk("%s:---Loop rx_pkts_num=%d------rwptr.bits.rptr=0x%x------->Default_Q [rwptr.bits.rptr(SW)=0x%x, rwptr.bits.wptr(HW) = 0x%x ]---->Free_Q(SW_HW) = 0x%8x \n",__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.rptr,rwptr.bits.wptr,fq_rwptr.bits32 );
+ }
+ // advance one for Rx default Q 0/1
+
+ //rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, tp->default_desc_num);
+ //SET_RPTR(&tp->default_qhdr->word1, rwptr.bits.rptr);
+ //tp->rx_rwptr.bits32 = rwptr.bits32;
+ //rwptr.bits.rptr = rwptr.bits.rptr;
+
+ dev->quota -= rx_pkts_num;
+ *budget -= rx_pkts_num;
+
+ status4 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);//try read SWFQ empty again
+ //fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ rwptr.bits32 = readl(&tp->default_qhdr->word1); //try read default_qhdr again
+ //printk("%s:---After rx_pkts_num=%d------rwptr.bits.rptr=0x%x------->Default_Q [rwptr.bits.rptr(SW)=0x%x, rwptr.bits.wptr(HW) = 0x%x ]---->Free_Q(SW_HW) = 0x%8x \n",__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.rptr,rwptr.bits.wptr,fq_rwptr.bits32 );
+// if (rwptr.bits.rptr > rwptr.bits.wptr )
+// {
+ //toe_gmac_disable_rx(dev);
+ //wait_event_interruptible_timeout(freeq_wait,
+ //(rx_pkts_num == 100), CMTP_INTEROP_TIMEOUT);
+ //printk("\n%s:: return 22222=======> rx_pkts_num =%d, rwptr.bits.rptr=%d, rwptr.bits.wptr = %d ====---------=======>JKJKJKJKJK\n",
+ //__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.wptr);
+// return 1;
+// }
+
+ if (rwptr.bits.rptr == rwptr.bits.wptr)
+ {
+ unsigned int data32;
+ //printk("%s:---[rwptr.bits.rptr == rwptr.bits.wptr] rx_pkts_num=%d------rwptr.bits.rptr=0x%x------->Default_Q [rwptr.bits.rptr(SW)=0x%x, rwptr.bits.wptr(HW) = 0x%x ]---->Free_Q(SW_HW) = 0x%8x \n",__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.rptr,rwptr.bits.wptr,fq_rwptr.bits32 );
+
+ /* Receive descriptor is empty now */
+#if 1
+ if (status4 & 0x1)
+ {
+ do_again =1;
+ //writel(0x40400000, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_ENABLE_4_REG); //disable SWFQ empty interrupt
+ //toe_gmac_disable_interrupt(tp->irq);
+ tp->sw_fq_empty_cnt++;
+ //toe_gmac_disable_rx(dev);
+ writel(0x07960202, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ writel(0x07960202, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ //printk("\n%s :: freeq int-----tp->sw_fq_empty_cnt =%d---------====================----------------->\n",__func__,tp->sw_fq_empty_cnt);
+ //while ((fq_rwptr.bits.wptr >= (fq_rwptr.bits.rptr+256)) || (fq_rwptr.bits.wptr <= (fq_rwptr.bits.rptr+256)))
+ //{
+ //gmac_write_reg(TOE_GLOBAL_BASE, GLOBAL_INTERRUPT_STATUS_4_REG, status4,
+ //0x1);
+ //printk("\n%s::fq_rwptr.wrptr = %x =======> ===========>here \n", __func__,fq_rwptr.bits32);
+ //if ((status4 & 0x1) == 0)
+ //break;
+ return 1;
+ //}
+
+ }
+#endif
+ //toe_gmac_fill_free_q();
+ netif_rx_complete(dev);
+ // enable GMAC-0 rx interrupt
+ // class-Q & TOE-Q are implemented in future
+ //data32 = readl(TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //if (tp->port_id == 0)
+ //data32 |= DEFAULT_Q0_INT_BIT;
+ //else
+ //data32 |= DEFAULT_Q1_INT_BIT;
+ //writel(data32, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_ENABLE_1_REG);
+ writel(0x3, TOE_GLOBAL_BASE+GLOBAL_INTERRUPT_ENABLE_1_REG);
+ //printk("\n%s::netif_rx_complete--> rx_pkts_num =%d, rwptr.bits.rptr=0x%x, rwptr.bits.wptr = 0x%x ====---------=======>JKJKJKJKJK\n",
+ //__func__,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.wptr);
+ writel(0x07960200, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ writel(0x07960200, TOE_GMAC1_BASE+GMAC_CONFIG0);
+ return 0;
+ }
+ else
+ {
+ //printk("\n%s:: return 1 -->status4= 0x%x,rx_pkts_num =%d, rwptr.bits.rptr=0x%x, rwptr.bits.wptr = 0x%x ======> \n", __func__,status4,rx_pkts_num,rwptr.bits.rptr,rwptr.bits.wptr);
+ return 1;
+ }
+}
+#endif
+
+/*----------------------------------------------------------------------
+* gmac_tx_timeout
+*----------------------------------------------------------------------*/
+void gmac_tx_timeout(struct net_device *dev)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+
+#ifdef CONFIG_SL351x_SYSCTL
+ if (tp->operation && storlink_ctl.link[tp->port_id])
+#else
+ if (tp->operation)
+#endif
+ {
+ netif_wake_queue(dev);
+ }
+}
+
+
+
+/*----------------------------------------------------------------------
+* mac_set_rule_reg
+*----------------------------------------------------------------------*/
+int mac_set_rule_reg(int mac, int rule, int enabled, u32 reg0, u32 reg1, u32 reg2)
+{
+ int total_key_dwords;
+
+ total_key_dwords = 1;
+
+ if (reg0 & MR_L2_BIT)
+ {
+ if (reg0 & MR_DA_BIT) total_key_dwords += 2;
+ if (reg0 & MR_SA_BIT) total_key_dwords += 2;
+ if ((reg0 & MR_DA_BIT) && ( reg0 & MR_SA_BIT)) total_key_dwords--;
+ if (reg0 & (MR_PPPOE_BIT | MR_VLAN_BIT)) total_key_dwords++;
+ }
+ if (reg0 & MR_L3_BIT)
+ {
+ if (reg0 & (MR_IP_HDR_LEN_BIT | MR_TOS_TRAFFIC_BIT | MR_SPR_BITS))
+ total_key_dwords++;
+ if (reg0 & MR_FLOW_LABLE_BIT) total_key_dwords++;
+ if ((reg0 & MR_IP_VER_BIT) == 0) // IPv4
+ {
+ if (reg1 & 0xff000000) total_key_dwords += 1;
+ if (reg1 & 0x00ff0000) total_key_dwords += 1;
+ }
+ else
+ {
+ if (reg1 & 0xff000000) total_key_dwords += 4;
+ if (reg1 & 0x00ff0000) total_key_dwords += 4;
+ }
+ }
+ if (reg0 & MR_L4_BIT)
+ {
+ if (reg1 & 0x0000f000) total_key_dwords += 1;
+ if (reg1 & 0x00000f00) total_key_dwords += 1;
+ if (reg1 & 0x000000f0) total_key_dwords += 1;
+ if (reg1 & 0x0000000f) total_key_dwords += 1;
+ if (reg2 & 0xf0000000) total_key_dwords += 1;
+ if (reg2 & 0x0f000000) total_key_dwords += 1;
+ }
+ if (reg0 & MR_L7_BIT)
+ {
+ if (reg2 & 0x00f00000) total_key_dwords += 1;
+ if (reg2 & 0x000f0000) total_key_dwords += 1;
+ if (reg2 & 0x0000f000) total_key_dwords += 1;
+ if (reg2 & 0x00000f00) total_key_dwords += 1;
+ if (reg2 & 0x000000f0) total_key_dwords += 1;
+ if (reg2 & 0x0000000f) total_key_dwords += 1;
+ }
+
+ if (total_key_dwords > HASH_MAX_KEY_DWORD)
+ return -1;
+
+ if (total_key_dwords == 0 && enabled)
+ return -2;
+
+ mac_set_rule_enable_bit(mac, rule, 0);
+ if (enabled)
+ {
+ mac_set_MRxCRx(mac, rule, 0, reg0);
+ mac_set_MRxCRx(mac, rule, 1, reg1);
+ mac_set_MRxCRx(mac, rule, 2, reg2);
+ mac_set_rule_action(mac, rule, total_key_dwords);
+ mac_set_rule_enable_bit(mac, rule, enabled);
+ }
+ else
+ {
+ mac_set_rule_action(mac, rule, 0);
+ }
+ return total_key_dwords;
+}
+
+/*----------------------------------------------------------------------
+* mac_get_rule_enable_bit
+*----------------------------------------------------------------------*/
+int mac_get_rule_enable_bit(int mac, int rule)
+{
+ switch (rule)
+ {
+ case 0: return ((mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) >> 15) & 1);
+ case 1: return ((mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) >> 31) & 1);
+ case 2: return ((mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) >> 15) & 1);
+ case 3: return ((mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) >> 31) & 1);
+ default: return 0;
+ }
+}
+
+/*----------------------------------------------------------------------
+* mac_set_rule_enable_bit
+*----------------------------------------------------------------------*/
+void mac_set_rule_enable_bit(int mac, int rule, int data)
+{
+ u32 reg;
+
+ if (data & ~1)
+ return;
+
+ switch (rule)
+ {
+ case 0:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) & ~(1<<15)) | (data << 15);
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG0, reg);
+ break;
+ case 1:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) & ~(1<<31)) | (data << 31);
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG0, reg);
+ break;
+ case 2:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) & ~(1<<15)) | (data << 15);
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG1, reg);
+ break;
+ case 3:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) & ~(1<<31)) | (data << 31);
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG1, reg);
+ }
+}
+
+/*----------------------------------------------------------------------
+* mac_set_rule_action
+*----------------------------------------------------------------------*/
+int mac_set_rule_action(int mac, int rule, int data)
+{
+ u32 reg;
+
+ if (data > 32)
+ return -1;
+
+ if (data)
+ data = (data << 6) | (data + HASH_ACTION_DWORDS);
+ switch (rule)
+ {
+ case 0:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) & ~(0x7ff));
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG0, reg | data);
+ break;
+ case 1:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG0) & ~(0x7ff<<16));
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG0, reg | (data << 16));
+ break;
+ case 2:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) & ~(0x7ff));
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG1, reg | data);
+ break;
+ case 3:
+ reg = (mac_read_dma_reg(mac, GMAC_HASH_ENGINE_REG1) & ~(0x7ff<<16));
+ mac_write_dma_reg(mac, GMAC_HASH_ENGINE_REG1, reg | (data << 16));
+ break;
+ default:
+ return -1;
+ }
+
+ return 0;
+}
+/*----------------------------------------------------------------------
+* mac_get_MRxCRx
+*----------------------------------------------------------------------*/
+int mac_get_MRxCRx(int mac, int rule, int ctrlreg)
+{
+ int reg;
+
+ switch (rule)
+ {
+ case 0: reg = GMAC_MR0CR0 + ctrlreg * 4; break;
+ case 1: reg = GMAC_MR1CR0 + ctrlreg * 4; break;
+ case 2: reg = GMAC_MR2CR0 + ctrlreg * 4; break;
+ case 3: reg = GMAC_MR3CR0 + ctrlreg * 4; break;
+ default: return 0;
+ }
+ return mac_read_dma_reg(mac, reg);
+}
+
+/*----------------------------------------------------------------------
+* mac_set_MRxCRx
+*----------------------------------------------------------------------*/
+void mac_set_MRxCRx(int mac, int rule, int ctrlreg, u32 data)
+{
+ int reg;
+
+ switch (rule)
+ {
+ case 0: reg = GMAC_MR0CR0 + ctrlreg * 4; break;
+ case 1: reg = GMAC_MR1CR0 + ctrlreg * 4; break;
+ case 2: reg = GMAC_MR2CR0 + ctrlreg * 4; break;
+ case 3: reg = GMAC_MR3CR0 + ctrlreg * 4; break;
+ default: return;
+ }
+ mac_write_dma_reg(mac, reg, data);
+}
+
+/*----------------------------------------------------------------------
+* mac_set_rule_priority
+*----------------------------------------------------------------------*/
+void mac_set_rule_priority(int mac, int p0, int p1, int p2, int p3)
+{
+ int i;
+ GMAC_MRxCR0_T reg[4];
+
+ for (i=0; i<4; i++)
+ reg[i].bits32 = mac_get_MRxCRx(mac, i, 0);
+
+ reg[0].bits.priority = p0;
+ reg[1].bits.priority = p1;
+ reg[2].bits.priority = p2;
+ reg[3].bits.priority = p3;
+
+ for (i=0; i<4; i++)
+ mac_set_MRxCRx(mac, i, 0, reg[i].bits32);
+}
+
+/*----------------------------------------------------------------------
+* gmac_netdev_ioctl
+*----------------------------------------------------------------------*/
+static int gmac_netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
+{
+ int rc = 0;
+ unsigned char *hwa = rq->ifr_ifru.ifru_hwaddr.sa_data;
+
+#ifdef br_if_ioctl
+ struct ethtool_cmd ecmd; //br_if.c will call this ioctl
+ GMAC_INFO_T *tp = dev->priv;
+#endif
+
+#ifdef CONFIG_SL351x_NAT
+ if (cmd == SIOCDEVPRIVATE)
+ return sl351x_nat_ioctl(dev, rq, cmd);
+#endif
+
+ switch (cmd) {
+ case SIOCETHTOOL:
+#ifdef br_if_ioctl //br_if.c will call this ioctl
+ if (!netif_running(dev))
+ {
+ printk("Before changing the H/W address,please down the device.\n");
+ return -EINVAL;
+ }
+ memset((void *) &ecmd, 0, sizeof (ecmd));
+ ecmd.supported =
+ SUPPORTED_Autoneg | SUPPORTED_TP | SUPPORTED_MII |
+ SUPPORTED_10baseT_Half | SUPPORTED_10baseT_Full |
+ SUPPORTED_100baseT_Half | SUPPORTED_100baseT_Full;
+ ecmd.port = PORT_TP;
+ ecmd.transceiver = XCVR_EXTERNAL;
+ ecmd.phy_address = tp->phy_addr;
+ switch (tp->speed_status)
+ {
+ case GMAC_SPEED_10: ecmd.speed = SPEED_10; break;
+ case GMAC_SPEED_100: ecmd.speed = SPEED_100; break;
+ case GMAC_SPEED_1000: ecmd.speed = SPEED_1000; break;
+ default: ecmd.speed = SPEED_10; break;
+ }
+ ecmd.duplex = tp->full_duplex_status ? DUPLEX_FULL : DUPLEX_HALF;
+ ecmd.advertising = ADVERTISED_TP;
+ ecmd.advertising |= ADVERTISED_Autoneg;
+ ecmd.autoneg = AUTONEG_ENABLE;
+ if (copy_to_user(rq->ifr_data, &ecmd, sizeof (ecmd)))
+ return -EFAULT;
+#endif
+
+ break;
+
+ case SIOCSIFHWADDR:
+ if (!netif_running(dev))
+ {
+ printk("Before changing the H/W address,please down the device.\n");
+ return -EINVAL;
+ }
+ gmac_set_mac_address(dev,hwa);
+ break;
+
+ case SIOCGMIIPHY: /* Get the address of the PHY in use. */
+ break;
+
+ case SIOCGMIIREG: /* Read the specified MII register. */
+ break;
+
+ case SIOCSMIIREG: /* Write the specified MII register */
+ break;
+
+ default:
+ rc = -EOPNOTSUPP;
+ break;
+ }
+
+ return rc;
+}
+
+#ifdef SL351x_GMAC_WORKAROUND
+
+#define GMAC_TX_STATE_OFFSET 0x60
+#define GMAC_RX_STATE_OFFSET 0x64
+#define GMAC_POLL_HANGED_NUM 200
+#define GMAC_RX_HANGED_STATE 0x4b2000
+#define GMAC_RX_HANGED_MASK 0xdff000
+#define GMAC_TX_HANGED_STATE 0x34012
+#define GMAC_TX_HANGED_MASK 0xfffff
+#define TOE_GLOBAL_REG_SIZE (0x78/sizeof(u32))
+#define TOE_DMA_REG_SIZE (0xd0/sizeof(u32))
+#define TOE_GMAC_REG_SIZE (0x30/sizeof(u32))
+#define GMAC0_RX_HANG_BIT (1 << 0)
+#define GMAC0_TX_HANG_BIT (1 << 1)
+#define GMAC1_RX_HANG_BIT (1 << 2)
+#define GMAC1_TX_HANG_BIT (1 << 3)
+
+int gmac_in_do_workaround;
+#if 0
+int debug_cnt, poll_max_cnt;
+#endif
+u32 gmac_workaround_cnt[4];
+u32 toe_global_reg[TOE_GLOBAL_REG_SIZE];
+u32 toe_dma_reg[GMAC_NUM][TOE_DMA_REG_SIZE];
+u32 toe_gmac_reg[GMAC_NUM][TOE_GMAC_REG_SIZE];
+u32 gmac_short_frame_workaround_cnt[2];
+
+static void sl351x_gmac_release_buffers(void);
+static void sl351x_gmac_release_swtx_q(void);
+static void sl351x_gmac_release_rx_q(void);
+#ifdef _TOEQ_CLASSQ_READY_
+static void sl351x_gmac_release_class_q(void);
+static void sl351x_gmac_release_toe_q(void);
+static void sl351x_gmac_release_intr_q(void);
+#endif
+static void sl351x_gmac_release_sw_free_q(void);
+static void sl351x_gmac_release_hw_free_q(void);
+#ifdef CONFIG_SL351x_NAT
+static int get_free_desc_cnt(unsigned long rwptr, int total);
+static void sl351x_gmac_release_hwtx_q(void);
+u32 sl351x_nat_workaround_cnt;
+#endif
+void sl351x_gmac_save_reg(void);
+void sl351x_gmac_restore_reg(void);
+
+
+/*----------------------------------------------------------------------
+* sl351x_poll_gmac_hanged_status
+* - Called by timer routine, period 10ms
+* - If (state != 0 && state == prev state && )
+*----------------------------------------------------------------------*/
+void sl351x_poll_gmac_hanged_status(u32 data)
+{
+ int i;
+ u32 state;
+ TOE_INFO_T *toe;
+ GMAC_INFO_T *tp;
+ u32 hanged_state;
+ // int old_operation[GMAC_NUM];
+#ifdef CONFIG_SL351x_NAT
+ u32 hw_free_cnt;
+#endif
+
+ if (gmac_in_do_workaround)
+ return;
+
+ gmac_in_do_workaround = 1;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ hanged_state = 0;
+
+#ifdef SL351x_TEST_WORKAROUND
+ if (toe->gmac[0].operation || toe->gmac[1].operation)
+ {
+ debug_cnt++;
+ if (debug_cnt == (30 * HZ))
+ {
+ debug_cnt = 0;
+ hanged_state = GMAC0_RX_HANG_BIT;
+ goto do_workaround;
+ }
+ }
+#endif
+ if (toe->gmac[0].operation)
+ hanged_state |= GMAC0_RX_HANG_BIT | GMAC0_TX_HANG_BIT;
+
+#if (GMAC_NUM > 1)
+ if (toe->gmac[1].operation)
+ hanged_state |= GMAC1_RX_HANG_BIT | GMAC1_TX_HANG_BIT;
+#endif
+
+ for (i=0; i<GMAC_POLL_HANGED_NUM; i++)
+ {
+ if (hanged_state & GMAC0_RX_HANG_BIT)
+ {
+ state = readl(TOE_GMAC0_BASE + GMAC_RX_STATE_OFFSET) & GMAC_RX_HANGED_MASK;
+ if (state != GMAC_RX_HANGED_STATE)
+ hanged_state &= ~GMAC0_RX_HANG_BIT;
+ }
+ if (hanged_state & GMAC0_TX_HANG_BIT)
+ {
+ state = readl(TOE_GMAC0_BASE + GMAC_TX_STATE_OFFSET) & GMAC_TX_HANGED_MASK;
+ if (state != GMAC_TX_HANGED_STATE)
+ hanged_state &= ~GMAC0_TX_HANG_BIT;
+ }
+#if (GMAC_NUM > 1)
+ if (hanged_state & GMAC1_RX_HANG_BIT)
+ {
+ state = readl(TOE_GMAC1_BASE + GMAC_RX_STATE_OFFSET) & GMAC_RX_HANGED_MASK;
+ if (state != GMAC_RX_HANGED_STATE)
+ hanged_state &= ~GMAC1_RX_HANG_BIT;
+ }
+ if (hanged_state & GMAC1_TX_HANG_BIT)
+ {
+ state = readl(TOE_GMAC1_BASE + GMAC_TX_STATE_OFFSET) & GMAC_TX_HANGED_MASK;
+ if (state != GMAC_TX_HANGED_STATE)
+ hanged_state &= ~GMAC1_TX_HANG_BIT;
+ }
+#endif
+ if (!hanged_state)
+ {
+#if 0
+ if (i < poll_max_cnt)
+ poll_max_cnt = i;
+#endif
+ if (toe->gmac[0].short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ gmac_short_frame_workaround_cnt[0]++;
+ toe->gmac[0].short_frames_cnt = 0;
+ goto do_workaround;
+ }
+#if (GMAC_NUM > 1)
+ if (toe->gmac[1].short_frames_cnt >= GMAC_SHORT_FRAME_THRESHOLD)
+ {
+ gmac_short_frame_workaround_cnt[1]++;
+ toe->gmac[1].short_frames_cnt = 0;
+ goto do_workaround;
+ }
+#endif
+
+#ifdef CONFIG_SL351x_NAT
+ hw_free_cnt = readl(TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+ hw_free_cnt = get_free_desc_cnt(hw_free_cnt, TOE_HW_FREEQ_DESC_NUM);
+#ifdef NAT_WORKAROUND_BY_RESET_GMAC
+ if (readl(TOE_GLOBAL_BASE + 0x4084) && (hw_free_cnt <= PAUSE_SET_HW_FREEQ))
+ {
+ sl351x_nat_workaround_cnt++;
+ goto do_workaround;
+ }
+#else
+ if (readl(TOE_GLOBAL_BASE + 0x4084) && (hw_free_cnt <= (PAUSE_SET_HW_FREEQ*2)))
+ {
+ sl351x_nat_workaround_cnt++;
+ sl351x_nat_workaround_handler();
+ }
+#endif
+#endif
+ gmac_in_do_workaround = 0;
+ add_timer(&gmac_workround_timer_obj);
+ return;
+ }
+ }
+
+do_workaround:
+
+ gmac_initialized = 0;
+ if (hanged_state)
+ {
+ if (hanged_state & GMAC0_RX_HANG_BIT) gmac_workaround_cnt[0]++;
+ if (hanged_state & GMAC0_TX_HANG_BIT) gmac_workaround_cnt[1]++;
+ if (hanged_state & GMAC1_RX_HANG_BIT) gmac_workaround_cnt[2]++;
+ if (hanged_state & GMAC1_TX_HANG_BIT) gmac_workaround_cnt[3]++;
+ }
+
+ for (i=0; i<GMAC_NUM; i++)
+ {
+ tp=(GMAC_INFO_T *)&toe->gmac[i];
+ // old_operation[i] = tp->operation;
+ if (tp->operation)
+ {
+ netif_stop_queue(tp->dev);
+ clear_bit(__LINK_STATE_START, &tp->dev->state);
+ toe_gmac_disable_interrupt(tp->irq);
+ toe_gmac_disable_tx_rx(tp->dev);
+ toe_gmac_hw_stop(tp->dev);
+ }
+ }
+
+ // clear all status bits
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_0_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_1_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_2_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_3_REG);
+ writel(0xffffffff, TOE_GLOBAL_BASE + GLOBAL_INTERRUPT_STATUS_4_REG);
+
+#if 0
+ if ((hanged_state & GMAC0_RX_HANG_BIT) &&
+ (readl(TOE_GMAC0_DMA_BASE + 0xdc) & 0xf0))
+ {
+ struct sk_buff *skb;
+ unsigned int buf;
+ buf = readl(TOE_GMAC0_DMA_BASE + 0x68) & ~3;
+#ifdef CONFIG_SL351x_NAT
+ if (buf < toe->hwfq_buf_base_dma || buf > toe->hwfq_buf_end_dma)
+#endif
+ {
+ skb = (struct sk_buff *)(REG32(buf - SKB_RESERVE_BYTES));
+ printk("GMAC-0 free a loss SKB 0x%x\n", (u32)skb);
+ dev_kfree_skb(skb);
+ }
+ }
+ if ((hanged_state & GMAC1_RX_HANG_BIT) &&
+ (readl(TOE_GMAC1_DMA_BASE + 0xdc) & 0xf0))
+ {
+ struct sk_buff *skb;
+ unsigned int buf;
+ buf = readl(TOE_GMAC1_DMA_BASE + 0x68) & ~3;
+#ifdef CONFIG_SL351x_NAT
+ if (buf < toe->hwfq_buf_base_dma || buf > toe->hwfq_buf_end_dma)
+#endif
+ {
+ skb = (struct sk_buff *)(REG32(buf - SKB_RESERVE_BYTES));
+ printk("GMAC-1 free a loss SKB 0x%x\n", (u32)skb);
+ dev_kfree_skb(skb);
+ }
+ }
+#endif
+
+ sl351x_gmac_release_buffers();
+ sl351x_gmac_save_reg();
+ toe_gmac_sw_reset();
+ sl351x_gmac_restore_reg();
+
+ if (toe->gmac[0].default_qhdr->word1.bits32)
+ {
+ // printk("===> toe->gmac[0].default_qhdr->word1 = 0x%x\n", toe->gmac[0].default_qhdr->word1);
+ sl351x_gmac_release_rx_q();
+ writel(0, &toe->gmac[0].default_qhdr->word1);
+ }
+ if (toe->gmac[1].default_qhdr->word1.bits32)
+ {
+ // printk("===> toe->gmac[1].default_qhdr->word1 = 0x%x\n", toe->gmac[1].default_qhdr->word1);
+ sl351x_gmac_release_rx_q();
+ writel(0, &toe->gmac[1].default_qhdr->word1);
+ }
+
+ gmac_initialized = 1;
+
+#ifdef CONFIG_SL351x_NAT
+ writel(0, TOE_GLOBAL_BASE + 0x4084);
+#endif
+
+ for (i=0; i<GMAC_NUM; i++)
+ {
+ tp=(GMAC_INFO_T *)&toe->gmac[i];
+ if (tp->operation)
+ {
+ toe_gmac_enable_interrupt(tp->irq);
+ toe_gmac_hw_start(tp->dev);
+ toe_gmac_enable_tx_rx(tp->dev);
+ netif_wake_queue(tp->dev);
+ set_bit(__LINK_STATE_START, &tp->dev->state);
+ }
+ }
+
+ gmac_in_do_workaround = 0;
+ add_timer(&gmac_workround_timer_obj);
+}
+
+/*----------------------------------------------------------------------
+* get_free_desc_cnt
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static int get_free_desc_cnt(unsigned long rwptr, int total)
+{
+ unsigned short wptr = rwptr & 0xffff;
+ unsigned short rptr = rwptr >> 16;
+
+ if (wptr >= rptr)
+ return (total - wptr + rptr);
+ else
+ return (rptr - wptr);
+}
+#endif
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_buffers
+*----------------------------------------------------------------------*/
+static void sl351x_gmac_release_buffers(void)
+{
+ // Free buffers & Descriptors in all SW Tx Queues
+ sl351x_gmac_release_swtx_q();
+
+ // Free buffers in Default Rx Queues
+ sl351x_gmac_release_rx_q();
+
+#ifdef _TOEQ_CLASSQ_READY_
+ // Free buffers in Classification Queues
+ sl351x_gmac_release_class_q();
+
+ // Free buffers in TOE Queues
+ sl351x_gmac_release_toe_q();
+
+ // Free buffers in Interrupt Queues
+ sl351x_gmac_release_intr_q();
+#endif
+
+ // Free buffers & descriptors in SW free queue
+ sl351x_gmac_release_sw_free_q();
+
+ // Free buffers & descriptors in HW free queue
+ sl351x_gmac_release_hw_free_q();
+
+#ifdef CONFIG_SL351x_NAT
+ // Free buffers & descriptors in HW free queue
+ sl351x_gmac_release_hwtx_q();
+#endif
+}
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_swtx_q
+*----------------------------------------------------------------------*/
+static void sl351x_gmac_release_swtx_q(void)
+{
+ int i, j;
+ GMAC_TXDESC_T *curr_desc;
+ unsigned int desc_count;
+ TOE_INFO_T *toe;
+ GMAC_INFO_T *tp;
+ GMAC_SWTXQ_T *swtxq;
+ DMA_RWPTR_T rwptr;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ tp = (GMAC_INFO_T *)&toe->gmac[0];
+ for (i=0; i<GMAC_NUM; i++, tp++)
+ {
+ if (!tp->existed) continue;
+ swtxq = (GMAC_SWTXQ_T *)&tp->swtxq[0];
+ for (j=0; j<TOE_SW_TXQ_NUM; j++, swtxq++)
+ {
+ for (;;)
+ {
+ rwptr.bits32 = readl(swtxq->rwptr_reg);
+ if (rwptr.bits.rptr == swtxq->finished_idx)
+ break;
+ curr_desc = (GMAC_TXDESC_T *)swtxq->desc_base + swtxq->finished_idx;
+ // if (curr_desc->word0.bits.status_tx_ok)
+ {
+ desc_count = curr_desc->word0.bits.desc_count;
+ while (--desc_count)
+ {
+ curr_desc->word0.bits.status_tx_ok = 0;
+ swtxq->finished_idx = RWPTR_ADVANCE_ONE(swtxq->finished_idx, swtxq->total_desc_num);
+ curr_desc = (GMAC_TXDESC_T *)swtxq->desc_base + swtxq->finished_idx;
+ }
+
+ curr_desc->word0.bits.status_tx_ok = 0;
+ if (swtxq->tx_skb[swtxq->finished_idx])
+ {
+ dev_kfree_skb_irq(swtxq->tx_skb[swtxq->finished_idx]);
+ swtxq->tx_skb[swtxq->finished_idx] = NULL;
+ }
+ }
+ swtxq->finished_idx = RWPTR_ADVANCE_ONE(swtxq->finished_idx, swtxq->total_desc_num);
+ }
+ writel(0, swtxq->rwptr_reg);
+ swtxq->finished_idx = 0;
+ }
+ }
+
+}
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_rx_q
+*----------------------------------------------------------------------*/
+static void sl351x_gmac_release_rx_q(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ GMAC_INFO_T *tp;
+ DMA_RWPTR_T rwptr;
+ volatile GMAC_RXDESC_T *curr_desc;
+ struct sk_buff *skb;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ tp = (GMAC_INFO_T *)&toe->gmac[0];
+ for (i=0; i<GMAC_NUM; i++, tp++)
+ {
+ if (!tp->existed) continue;
+ rwptr.bits32 = readl(&tp->default_qhdr->word1);
+ while (rwptr.bits.rptr != rwptr.bits.wptr)
+ {
+ curr_desc = (GMAC_RXDESC_T *)tp->default_desc_base + rwptr.bits.rptr;
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ dev_kfree_skb_irq(skb);
+ rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, tp->default_desc_num);
+ SET_RPTR(&tp->default_qhdr->word1, rwptr.bits.rptr);
+ rwptr.bits32 = readl(&tp->default_qhdr->word1);
+ } // while
+ writel(0, &tp->default_qhdr->word1);
+ tp->rx_rwptr.bits32 = 0;
+ } // for
+
+}
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_class_q
+*----------------------------------------------------------------------*/
+#ifdef _TOEQ_CLASSQ_READY_
+static void sl351x_gmac_release_class_q(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ CLASSQ_INFO_T *classq;
+ DMA_RWPTR_T rwptr;
+ volatile GMAC_RXDESC_T *curr_desc;
+ struct sk_buff *skb;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ classq = (CLASSQ_INFO_T *)&toe->classq[0];
+ for (i=0; i<TOE_CLASS_QUEUE_NUM; i++, classq++)
+ {
+ rwptr.bits32 = readl(&classq->qhdr->word1);
+ while (rwptr.bits.rptr != rwptr.bits.wptr)
+ {
+ curr_desc = (GMAC_RXDESC_T *)classq->desc_base + rwptr.bits.rptr;
+ skb = (struct sk_buff *)(REG32(__va(curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ dev_kfree_skb_irq(skb);
+ rwptr.bits.rptr = RWPTR_ADVANCE_ONE(rwptr.bits.rptr, classq->desc_num);
+ SET_RPTR(&classq->qhdr->word1, rwptr.bits.rptr);
+ rwptr.bits32 = readl(&classq->qhdr->word1);
+ } // while
+ writel(0, &classq->qhdr->word1);
+ classq->rwptr.bits32 = 0;
+ } // for
+
+}
+#endif
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_toe_q
+*----------------------------------------------------------------------*/
+#ifdef _TOEQ_CLASSQ_READY_
+static void sl351x_gmac_release_toe_q(void)
+{
+ int i;
+ TOE_INFO_T *toe;
+ TOEQ_INFO_T *toeq_info;
+ TOE_QHDR_T *toe_qhdr;
+ DMA_RWPTR_T rwptr;
+ volatile GMAC_RXDESC_T *curr_desc;
+ unsigned int rptr, wptr;
+ GMAC_RXDESC_T *toe_curr_desc;
+ struct sk_buff *skb;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ toe_qhdr = (TOE_QHDR_T *)TOE_TOE_QUE_HDR_BASE;
+ for (i=0; i<TOE_TOE_QUEUE_NUM; i++, toe_qhdr++)
+ {
+ toeq_info = (TOEQ_INFO_T *)&toe->toeq[i];
+ wptr = toe_qhdr->word1.bits.wptr;
+ rptr = toe_qhdr->word1.bits.rptr;
+ while (rptr != wptr)
+ {
+ toe_curr_desc = (GMAC_RXDESC_T *)toeq_info->desc_base + rptr;
+ skb = (struct sk_buff *)(REG32(__va(toe_curr_desc->word2.buf_adr) - SKB_RESERVE_BYTES));
+ dev_kfree_skb_irq(skb);
+ rptr = RWPTR_ADVANCE_ONE(rptr, toeq_info->desc_num);
+ SET_RPTR(&toe_qhdr->word1.bits32, rptr);
+ wptr = toe_qhdr->word1.bits.wptr;
+ rptr = toe_qhdr->word1.bits.rptr;
+ }
+ toe_qhdr->word1.bits32 = 0;
+ toeq_info->rwptr.bits32 = 0;
+ }
+}
+#endif
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_intr_q
+*----------------------------------------------------------------------*/
+#ifdef _TOEQ_CLASSQ_READY_
+static void sl351x_gmac_release_intr_q(void)
+{
+}
+#endif
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_sw_free_q
+*----------------------------------------------------------------------*/
+static void sl351x_gmac_release_sw_free_q(void)
+{
+ TOE_INFO_T *toe;
+ volatile DMA_RWPTR_T fq_rwptr;
+ volatile GMAC_RXDESC_T *fq_desc;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+
+ while ((unsigned short)RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr, TOE_SW_FREEQ_DESC_NUM) != fq_rwptr.bits.rptr)
+ {
+ struct sk_buff *skb;
+ if ((skb = dev_alloc_skb(SW_RX_BUF_SIZE))==NULL) /* allocate socket buffer */
+ {
+ printk("%s::skb buffer allocation fail !\n",__func__); while(1);
+ }
+ // *(unsigned int *)(skb->data) = (unsigned int)skb;
+ REG32(skb->data) = (unsigned long)skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES);
+
+ fq_rwptr.bits.wptr = RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr, TOE_SW_FREEQ_DESC_NUM);
+ fq_desc = (volatile GMAC_RXDESC_T *)toe->swfq_desc_base + fq_rwptr.bits.wptr;
+ fq_desc->word2.buf_adr = (unsigned int)__pa(skb->data);
+ SET_WPTR(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG, fq_rwptr.bits.wptr);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ }
+
+ toe->fq_rx_rwptr.bits.wptr = TOE_SW_FREEQ_DESC_NUM - 1;
+ toe->fq_rx_rwptr.bits.rptr = 0;
+ writel(toe->fq_rx_rwptr.bits32, TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+
+}
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_hw_free_q
+*----------------------------------------------------------------------*/
+static void sl351x_gmac_release_hw_free_q(void)
+{
+ DMA_RWPTR_T rwptr_reg;
+
+#ifdef CONFIG_SL351x_NAT
+ int i;
+ TOE_INFO_T *toe;
+ GMAC_RXDESC_T *desc_ptr;
+ unsigned int buf_ptr;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+ desc_ptr = (GMAC_RXDESC_T *)toe->hwfq_desc_base;
+ buf_ptr = (unsigned int)toe->hwfq_buf_base_dma;
+ for (i=0; i<TOE_HW_FREEQ_DESC_NUM; i++)
+ {
+ desc_ptr->word0.bits.buffer_size = HW_RX_BUF_SIZE;
+ desc_ptr->word1.bits.sw_id = i;
+ desc_ptr->word2.buf_adr = (unsigned int)buf_ptr;
+ desc_ptr++;
+ buf_ptr += HW_RX_BUF_SIZE;
+ }
+#endif
+ rwptr_reg.bits.wptr = TOE_HW_FREEQ_DESC_NUM - 1;
+ rwptr_reg.bits.rptr = 0;
+ writel(rwptr_reg.bits32, TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+}
+
+/*----------------------------------------------------------------------
+* sl351x_gmac_release_hw_free_q
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static void sl351x_gmac_release_hwtx_q(void)
+{
+ int i;
+ unsigned int rwptr_addr;
+
+ rwptr_addr = TOE_GMAC0_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ writel(0, rwptr_addr);
+ rwptr_addr+=4;
+ }
+ rwptr_addr = TOE_GMAC1_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ writel(0, rwptr_addr);
+ rwptr_addr+=4;
+ }
+}
+#endif
+
+/*----------------------------------------------------------------------
+* sl351x_gmac_save_reg
+*----------------------------------------------------------------------*/
+void sl351x_gmac_save_reg(void)
+{
+ int i;
+ volatile u32 *destp;
+ unsigned int srce_addr;
+
+ srce_addr = TOE_GLOBAL_BASE;
+ destp = (volatile u32 *)toe_global_reg;
+ for (i=0; i<TOE_GLOBAL_REG_SIZE; i++, destp++, srce_addr+=4)
+ *destp = readl(srce_addr);
+
+ srce_addr = TOE_GMAC0_DMA_BASE;
+ destp = (volatile u32 *)&toe_dma_reg[0][0];
+ for (i=0; i<TOE_DMA_REG_SIZE; i++, destp++, srce_addr+=4)
+ {
+ if (srce_addr == (TOE_GMAC0_DMA_BASE+0x38))
+ srce_addr = (TOE_GMAC0_DMA_BASE+0x50);
+ if (srce_addr == (TOE_GMAC0_DMA_BASE+0x58))
+ srce_addr = (TOE_GMAC0_DMA_BASE+0x70);
+
+ *destp = readl(srce_addr);
+ }
+ srce_addr = TOE_GMAC1_DMA_BASE;
+ destp = (volatile u32 *)&toe_dma_reg[1][0];
+ for (i=0; i<TOE_DMA_REG_SIZE; i++, destp++, srce_addr+=4)
+ {
+ if (srce_addr == (TOE_GMAC0_DMA_BASE+0x38))
+ srce_addr = (TOE_GMAC0_DMA_BASE+0x50);
+ if (srce_addr == (TOE_GMAC0_DMA_BASE+0x58))
+ srce_addr = (TOE_GMAC0_DMA_BASE+0x70);
+
+ *destp = readl(srce_addr);
+ }
+
+ srce_addr = TOE_GMAC0_BASE;
+ destp = (volatile u32 *)&toe_gmac_reg[0][0];
+ for (i=0; i<TOE_GMAC_REG_SIZE; i++, destp++, srce_addr+=4)
+ *destp = readl(srce_addr);
+
+ srce_addr = TOE_GMAC1_BASE;
+ destp = (volatile u32 *)&toe_gmac_reg[1][0];
+ for (i=0; i<TOE_GMAC_REG_SIZE; i++, destp++, srce_addr+=4)
+ *destp = readl(srce_addr);
+}
+
+/*----------------------------------------------------------------------
+* sl351x_gmac_restore_reg
+*----------------------------------------------------------------------*/
+void sl351x_gmac_restore_reg(void)
+{
+ int i;
+ volatile u32 *srcep;
+ unsigned int dest_addr;
+
+ srcep = (volatile u32 *)&toe_dma_reg[0][0];
+ dest_addr = TOE_GMAC0_DMA_BASE;
+ for (i=0; i<TOE_DMA_REG_SIZE; i++, dest_addr+=4, srcep++)
+ {
+ if (dest_addr == (TOE_GMAC0_DMA_BASE+0x38))
+ dest_addr = (TOE_GMAC0_DMA_BASE+0x50);
+ if (dest_addr == (TOE_GMAC0_DMA_BASE+0x58))
+ dest_addr = (TOE_GMAC0_DMA_BASE+0x70);
+
+ writel(*srcep, dest_addr);
+ // gmac_write_reg(dest_addr, 0, *srcep, 0xffffffff);
+ }
+ srcep = (volatile u32 *)&toe_dma_reg[1][0];
+ dest_addr = TOE_GMAC1_DMA_BASE;
+ for (i=0; i<TOE_DMA_REG_SIZE; i++, dest_addr+=4, srcep++)
+ {
+ if (dest_addr == (TOE_GMAC0_DMA_BASE+0x38))
+ dest_addr = (TOE_GMAC0_DMA_BASE+0x50);
+ if (dest_addr == (TOE_GMAC0_DMA_BASE+0x58))
+ dest_addr = (TOE_GMAC0_DMA_BASE+0x70);
+
+ writel(*srcep, dest_addr);
+ // gmac_write_reg(dest_addr, 0, *srcep, 0xffffffff);
+ }
+
+ srcep = (volatile u32 *)&toe_gmac_reg[0][0];
+ dest_addr = TOE_GMAC0_BASE;
+ for (i=0; i<TOE_GMAC_REG_SIZE; i++, dest_addr+=4, srcep++)
+ writel(*srcep, dest_addr);
+
+ srcep = (volatile u32 *)&toe_gmac_reg[1][0];
+ dest_addr = TOE_GMAC1_BASE;
+ for (i=0; i<TOE_GMAC_REG_SIZE; i++, dest_addr+=4, srcep++)
+ writel(*srcep, dest_addr);
+
+ srcep = (volatile u32 *)toe_global_reg;
+ dest_addr = TOE_GLOBAL_BASE;
+ for (i=0; i<TOE_GLOBAL_REG_SIZE; i++, dest_addr+=4, srcep++)
+ writel(*srcep, dest_addr);
+
+}
+
+#ifdef CONFIG_SL351x_NAT
+/*----------------------------------------------------------------------
+* sl351x_nat_workaround_init
+*----------------------------------------------------------------------*/
+#define NAT_WORAROUND_DESC_POWER (6)
+#define NAT_WORAROUND_DESC_NUM (2 << NAT_WORAROUND_DESC_POWER)
+dma_addr_t sl351x_nat_workaround_desc_dma;
+void sl351x_nat_workaround_init(void)
+{
+ unsigned int desc_buf;
+
+ desc_buf = (unsigned int)DMA_MALLOC((NAT_WORAROUND_DESC_NUM * sizeof(GMAC_RXDESC_T)),
+ (dma_addr_t *)&sl351x_nat_workaround_desc_dma) ;
+ memset((void *)desc_buf, 0, NAT_WORAROUND_DESC_NUM * sizeof(GMAC_RXDESC_T));
+
+ // DMA Queue Base & Size
+ writel((sl351x_nat_workaround_desc_dma & DMA_Q_BASE_MASK) | NAT_WORAROUND_DESC_POWER,
+ TOE_GLOBAL_BASE + 0x4080);
+ writel(0, TOE_GLOBAL_BASE + 0x4084);
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_workaround_handler
+*----------------------------------------------------------------------*/
+#ifndef NAT_WORKAROUND_BY_RESET_GMAC
+static void sl351x_nat_workaround_handler(void)
+{
+ int i;
+ DMA_RWPTR_T rwptr;
+ GMAC_RXDESC_T *desc_ptr;
+ unsigned int buf_ptr;
+ TOE_INFO_T *toe;
+ GMAC_CONFIG0_T config0;
+ unsigned int rwptr_addr;
+
+ toe = (TOE_INFO_T *)&toe_private_data;
+
+ // disable Rx of GMAC-0 & 1
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 1;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+
+ // wait GMAC-0 HW Tx finished
+ rwptr_addr = TOE_GMAC0_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ rwptr.bits32 = readl(rwptr_addr);
+ if (rwptr.bits.rptr != rwptr.bits.wptr)
+ return; // wait the HW to send packets and release buffers
+ rwptr_addr+=4;
+ }
+ rwptr_addr = TOE_GMAC1_DMA_BASE + GMAC_HW_TX_QUEUE0_PTR_REG;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ rwptr.bits32 = readl(rwptr_addr);
+ if (rwptr.bits.rptr != rwptr.bits.wptr)
+ return; // wait the HW to send packets and release buffers
+ rwptr_addr+=4;
+ }
+
+ // printk("sl351x_nat_workaround_handler %d\n", sl351x_nat_workaround_cnt);
+ desc_ptr = (GMAC_RXDESC_T *)toe->hwfq_desc_base;
+ buf_ptr = (unsigned int)toe->hwfq_buf_base_dma;
+ for (i=0; i<TOE_HW_FREEQ_DESC_NUM; i++)
+ {
+ desc_ptr->word0.bits.buffer_size = HW_RX_BUF_SIZE;
+ desc_ptr->word1.bits.sw_id = i;
+ desc_ptr->word2.buf_adr = (unsigned int)buf_ptr;
+ desc_ptr++;
+ buf_ptr += HW_RX_BUF_SIZE;
+ }
+ rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+ rwptr.bits.wptr = RWPTR_RECEDE_ONE(rwptr.bits.rptr, TOE_HW_FREEQ_DESC_NUM);
+ writel(rwptr.bits32, TOE_GLOBAL_BASE + GLOBAL_HWFQ_RWPTR_REG);
+ writel(0, TOE_GLOBAL_BASE + 0x4084);
+
+ // Enable Rx of GMAC-0 & 1
+ config0.bits32 = readl(TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC0_BASE+GMAC_CONFIG0);
+ config0.bits32 = readl(TOE_GMAC1_BASE+GMAC_CONFIG0);
+ config0.bits.dis_rx = 0;
+ writel(config0.bits32, TOE_GMAC1_BASE+GMAC_CONFIG0);
+}
+#endif
+#endif // CONFIG_SL351x_NAT
+
+#endif // SL351x_GMAC_WORKAROUND
+
+/* get the mac addresses from flash
+ *can't do this in module_init because mtd driver is initialized after ethernet
+ */
+static __init int sl351x_mac_address_init(void)
+{
+ GMAC_INFO_T *tp;
+ struct sockaddr sock;
+ int i;
+
+ /* get mac address from FLASH */
+ gmac_get_mac_address();
+
+ for (i = 0; i < GMAC_NUM; i++) {
+ tp = (GMAC_INFO_T *)&toe_private_data.gmac[i];
+ memcpy(&sock.sa_data[0],&eth_mac[tp->port_id][0],6);
+ gmac_set_mac_address(tp->dev,(void *)&sock);
+ }
+
+ return 0;
+}
+late_initcall(sl351x_mac_address_init);
+
+
--- /dev/null
+++ b/drivers/net/sl351x_hash.c
@@ -0,0 +1,713 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl351x_hash.c
+* Description :
+* Handle Storlink SL351x Hash Functions
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* 03/13/2006 Gary Chen Create and implement
+*
+****************************************************************************/
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch/irqs.h>
+#include <asm/arch/it8712.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/skbuff.h>
+#include <linux/in.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#include <linux/list.h>
+#define MIDWAY
+#define SL_LEPUS
+
+#include <asm/arch/sl2312.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <asm/arch/sl351x_hash_cfg.h>
+
+#ifndef RXTOE_DEBUG
+#define RXTOE_DEBUG
+#endif
+#undef RXTOE_DEBUG
+
+/*----------------------------------------------------------------------
+* Definition
+*----------------------------------------------------------------------*/
+#define hash_printf printk
+
+#define HASH_TIMER_PERIOD (30) // seconds
+#define HASH_ILLEGAL_INDEX 0xffff
+
+/*----------------------------------------------------------------------
+* Variables
+*----------------------------------------------------------------------*/
+u32 hash_nat_owner_bits[HASH_TOTAL_ENTRIES/32];
+char hash_tables[HASH_TOTAL_ENTRIES][HASH_MAX_BYTES] __attribute__ ((aligned(16)));
+static struct timer_list hash_timer_obj;
+LIST_HEAD(hash_timeout_list);
+
+/*----------------------------------------------------------------------
+* Functions
+*----------------------------------------------------------------------*/
+void dm_long(u32 location, int length);
+static void hash_timer_func(u32 data);
+
+/*----------------------------------------------------------------------
+* hash_init
+*----------------------------------------------------------------------*/
+void sl351x_hash_init(void)
+{
+ int i;
+ volatile u32 *dp1, *dp2, dword;
+
+ dp1 = (volatile u32 *) TOE_V_BIT_BASE;
+ dp2 = (volatile u32 *) TOE_A_BIT_BASE;
+
+ for (i=0; i<HASH_TOTAL_ENTRIES/32; i++)
+ {
+ *dp1++ = 0;
+ dword = *dp2++; // read-clear
+ }
+ memset((void *)&hash_nat_owner_bits, 0, sizeof(hash_nat_owner_bits));
+ memset((void *)&hash_tables, 0, sizeof(hash_tables));
+
+ init_timer(&hash_timer_obj);
+ hash_timer_obj.expires = jiffies + (HASH_TIMER_PERIOD * HZ);
+ hash_timer_obj.data = (unsigned long)&hash_timer_obj;
+ hash_timer_obj.function = (void *)&hash_timer_func;
+ add_timer(&hash_timer_obj);
+
+#if (HASH_MAX_BYTES == 128)
+ writel((unsigned long)__pa(&hash_tables) | 3, // 32 words
+ TOE_GLOBAL_BASE + GLOBAL_HASH_TABLE_BASE_REG);
+#elif (HASH_MAX_BYTES == 64)
+ writel((unsigned long)__pa(&hash_tables) | 2, // 16 words
+ TOE_GLOBAL_BASE + GLOBAL_HASH_TABLE_BASE_REG);
+#else
+ #error Incorrect setting for HASH_MAX_BYTES
+#endif
+
+}
+/*----------------------------------------------------------------------
+* hash_add_entry
+*----------------------------------------------------------------------*/
+int hash_add_entry(HASH_ENTRY_T *entry)
+{
+ int rc;
+ u32 key[HASH_MAX_DWORDS];
+ rc = hash_build_keys((u32 *)&key, entry);
+ if (rc < 0)
+ return -1;
+ hash_write_entry(entry, (unsigned char*) &key[0]);
+// hash_set_valid_flag(entry->index, 1);
+// printk("Dump hash key!\n");
+// dump_hash_key(entry);
+ return entry->index;
+}
+
+/*----------------------------------------------------------------------
+* hash_set_valid_flag
+*----------------------------------------------------------------------*/
+void hash_set_valid_flag(int index, int valid)
+{
+ register u32 reg32;
+
+ reg32 = TOE_V_BIT_BASE + (index/32) * 4;
+
+ if (valid)
+ {
+ writel(readl(reg32) | (1 << (index%32)), reg32);
+ }
+ else
+ {
+ writel(readl(reg32) & ~(1 << (index%32)), reg32);
+ }
+}
+
+/*----------------------------------------------------------------------
+* hash_set_nat_owner_flag
+*----------------------------------------------------------------------*/
+void hash_set_nat_owner_flag(int index, int valid)
+{
+ if (valid)
+ {
+ hash_nat_owner_bits[index/32] |= (1 << (index % 32));
+ }
+ else
+ {
+ hash_nat_owner_bits[index/32] &= ~(1 << (index % 32));
+ }
+}
+
+
+/*----------------------------------------------------------------------
+* hash_build_keys
+*----------------------------------------------------------------------*/
+int hash_build_keys(u32 *destp, HASH_ENTRY_T *entry)
+{
+ u32 data;
+ unsigned char *cp;
+ int i, j;
+ unsigned short index;
+ int total;
+
+ memset((void *)destp, 0, HASH_MAX_BYTES);
+ cp = (unsigned char *)destp;
+
+ if (entry->key_present.port || entry->key_present.Ethertype)
+ {
+ HASH_PUSH_WORD(cp, entry->key.Ethertype); // word 0
+ HASH_PUSH_BYTE(cp, entry->key.port); // Byte 2
+ HASH_PUSH_BYTE(cp, 0); // Byte 3
+ }
+ else
+ {
+ HASH_PUSH_DWORD(cp, 0);
+ }
+
+ if (entry->key_present.da || entry->key_present.sa)
+ {
+ unsigned char mac[4];
+ if (entry->key_present.da)
+ {
+ for (i=0; i<4; i++)
+ HASH_PUSH_BYTE(cp, entry->key.da[i]);
+ }
+ mac[0] = (entry->key_present.da) ? entry->key.da[4] : 0;
+ mac[1] = (entry->key_present.da) ? entry->key.da[5] : 0;
+ mac[2] = (entry->key_present.sa) ? entry->key.sa[0] : 0;
+ mac[3] = (entry->key_present.sa) ? entry->key.sa[1] : 0;
+ data = mac[0] + (mac[1]<<8) + (mac[2]<<16) + (mac[3]<<24);
+ HASH_PUSH_DWORD(cp, data);
+ if (entry->key_present.sa)
+ {
+ for (i=2; i<6; i++)
+ HASH_PUSH_BYTE(cp, entry->key.sa[i]);
+ }
+ }
+
+ if (entry->key_present.pppoe_sid || entry->key_present.vlan_id)
+ {
+ HASH_PUSH_WORD(cp, entry->key.vlan_id); // low word
+ HASH_PUSH_WORD(cp, entry->key.pppoe_sid); // high word
+ }
+ if (entry->key_present.ipv4_hdrlen || entry->key_present.ip_tos || entry->key_present.ip_protocol)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.ip_protocol); // Byte 0
+ HASH_PUSH_BYTE(cp, entry->key.ip_tos); // Byte 1
+ HASH_PUSH_BYTE(cp, entry->key.ipv4_hdrlen); // Byte 2
+ HASH_PUSH_BYTE(cp, 0); // Byte 3
+ }
+
+ if (entry->key_present.ipv6_flow_label)
+ {
+ HASH_PUSH_DWORD(cp, entry->key.ipv6_flow_label); // low word
+ }
+ if (entry->key_present.sip)
+ {
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[i]);
+ if (entry->key.ipv6)
+ {
+ for (i=4; i<16; i+=4)
+ {
+ for (j=i+3; j>=i; j--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[j]);
+ }
+ }
+ }
+ if (entry->key_present.dip)
+ {
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[i]);
+ if (entry->key.ipv6)
+ {
+ for (i=4; i<16; i+=4)
+ {
+ for (j=i+3; j>=i; j--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[j]);
+ }
+ }
+ }
+
+ if (entry->key_present.l4_bytes_0_3)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[3]);
+ }
+ if (entry->key_present.l4_bytes_4_7)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[4]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[5]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[6]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[7]);
+ }
+ if (entry->key_present.l4_bytes_8_11)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[8]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[9]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[10]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[11]);
+ }
+ if (entry->key_present.l4_bytes_12_15)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[12]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[13]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[14]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[15]);
+ }
+ if (entry->key_present.l4_bytes_16_19)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[16]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[17]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[18]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[19]);
+ }
+ if (entry->key_present.l4_bytes_20_23)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[20]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[21]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[22]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[23]);
+ }
+ if (entry->key_present.l7_bytes_0_3)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[3]);
+ }
+ if (entry->key_present.l7_bytes_4_7)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[4]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[5]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[6]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[7]);
+ }
+ if (entry->key_present.l7_bytes_8_11)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[8]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[9]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[10]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[11]);
+ }
+ if (entry->key_present.l7_bytes_12_15)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[12]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[13]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[14]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[15]);
+ }
+ if (entry->key_present.l7_bytes_16_19)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[16]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[17]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[18]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[19]);
+ }
+ if (entry->key_present.l7_bytes_20_23)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[20]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[21]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[22]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[23]);
+ }
+
+ // get hash index
+ total = (u32)((u32)cp - (u32)destp) / (sizeof(u32));
+
+ if (total > HASH_MAX_KEY_DWORD)
+ {
+ //hash_printf("Total key words (%d) is too large (> %d)!\n",
+ // total, HASH_MAX_KEY_DWORD);
+ return -1;
+ }
+
+ if (entry->key_present.port || entry->key_present.Ethertype)
+ index = hash_gen_crc16((unsigned char *)destp, total * 4);
+ else
+ {
+ if (total == 1)
+ {
+ hash_printf("No key is assigned!\n");
+ return -1;
+ }
+
+ index = hash_gen_crc16((unsigned char *)(destp+1), (total-1) * 4);
+ }
+
+ entry->index = index & HASH_BITS_MASK;
+
+ //hash_printf("Total key words = %d, Hash Index= %d\n",
+ // total, entry->index);
+
+ cp = (unsigned char *)destp;
+ cp+=3;
+ HASH_PUSH_BYTE(cp, entry->rule); // rule
+
+ entry->total_dwords = total;
+
+ return total;
+}
+
+/*----------------------------------------------------------------------
+* hash_build_nat_keys
+*----------------------------------------------------------------------*/
+void hash_build_nat_keys(u32 *destp, HASH_ENTRY_T *entry)
+{
+ unsigned char *cp;
+ int i;
+ unsigned short index;
+ int total;
+
+ memset((void *)destp, 0, HASH_MAX_BYTES);
+
+ cp = (unsigned char *)destp + 2;
+ HASH_PUSH_BYTE(cp, entry->key.port);
+ cp++;
+
+ if (entry->key_present.pppoe_sid || entry->key_present.vlan_id)
+ {
+ HASH_PUSH_WORD(cp, entry->key.vlan_id); // low word
+ HASH_PUSH_WORD(cp, entry->key.pppoe_sid); // high word
+ }
+
+ HASH_PUSH_BYTE(cp, entry->key.ip_protocol);
+ cp+=3;
+
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[i]);
+
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[i]);
+
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[3]);
+
+ // get hash index
+ total = (u32)((u32)cp - (u32)destp) / (sizeof(u32));
+
+ index = hash_gen_crc16((unsigned char *)destp, total * 4);
+ entry->index = index & ((1 << HASH_BITS) - 1);
+
+ cp = (unsigned char *)destp;
+ cp+=3;
+ HASH_PUSH_BYTE(cp, entry->rule); // rule
+
+ entry->total_dwords = total;
+}
+
+/*----------------------------------------------------------------------
+* hash_build_toe_keys
+*----------------------------------------------------------------------*/
+int hash_build_toe_keys(u32 *destp, HASH_ENTRY_T *entry)
+{
+ unsigned long data;
+ unsigned char *cp;
+ unsigned short index;
+ int i;
+ int total;
+ //printk("%s\n", __func__);
+ memset((void*)destp, 0, HASH_MAX_BYTES);
+ cp = (unsigned char*)destp;
+
+ if(entry->key_present.port || entry->key_present.Ethertype) {
+ data = (entry->key.port << 16) + entry->key.Ethertype;
+ HASH_PUSH_DWORD(cp, data);
+ } else
+ HASH_PUSH_DWORD(cp, 0);
+
+ if (entry->key_present.da || entry->key_present.sa) {
+ unsigned char mac[4];
+ if (entry->key_present.da) {
+ data = (entry->key.da[0]) + (entry->key.da[1] << 8) +
+ (entry->key.da[2] << 16) + (entry->key.da[3] <<24);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ mac[0] = (entry->key_present.da) ? entry->key.da[4] : 0;
+ mac[1] = (entry->key_present.da) ? entry->key.da[5] : 0;
+ mac[2] = (entry->key_present.sa) ? entry->key.sa[0] : 0;
+ mac[3] = (entry->key_present.sa) ? entry->key.sa[1] : 0;
+ data = mac[0] + (mac[1]<<8) + (mac[2]<<16) + (mac[3]<<24);
+ HASH_PUSH_DWORD(cp, data);
+ if (entry->key_present.sa) {
+ data = (entry->key.sa[2]) + (entry->key.sa[3] << 8) +
+ (entry->key.sa[4] << 16) + (entry->key.sa[5] <<24);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ }
+
+ if (entry->key_present.ip_protocol) {
+ unsigned char ip_protocol;
+ ip_protocol = entry->key.ip_protocol;
+ data = ip_protocol;
+ HASH_PUSH_DWORD(cp, data);
+ }
+
+ if (entry->key_present.ipv6_flow_label) {
+ unsigned long flow_label;
+ flow_label = entry->key.ipv6_flow_label;
+ data = flow_label & 0xfffff;
+ HASH_PUSH_DWORD(cp, data);
+ }
+
+ if (entry->key_present.sip) {
+ {
+ data = IPIV(entry->key.sip[0], entry->key.sip[1],
+ entry->key.sip[2], entry->key.sip[3]);
+ HASH_PUSH_DWORD(cp, data);
+ if (entry->key.ipv6) {
+ for (i=4; i<16; i+=4) {
+ data = IPIV(entry->key.sip[i+0], entry->key.sip[i+1],
+ entry->key.sip[i+2], entry->key.sip[i+3]);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ }
+ }
+ }
+
+ if (entry->key_present.dip) {
+ {
+ data = IPIV(entry->key.dip[0], entry->key.dip[1],
+ entry->key.dip[2], entry->key.dip[3]);
+ HASH_PUSH_DWORD(cp, data);
+ if (entry->key.ipv6) {
+ for (i=4; i<16; i+=4) {
+ data = IPIV(entry->key.dip[i+0], entry->key.dip[i+1],
+ entry->key.dip[i+2], entry->key.dip[i+3]);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ }
+ }
+ }
+ if (entry->key_present.l4_bytes_0_3)
+ {
+ unsigned char *datap;
+ datap = &entry->key.l4_bytes[0];
+ data = datap[0] + (datap[1] << 8) + (datap[2] << 16) + (datap[3] << 24);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ if (entry->key_present.l7_bytes_0_3)
+ {
+ unsigned char *datap;
+ datap = &entry->key.l7_bytes[0];
+ data = datap[0] + (datap[1] << 8) + (datap[2] << 16) + (datap[3] << 24);
+ HASH_PUSH_DWORD(cp, data);
+ }
+ if (entry->key_present.l7_bytes_4_7)
+ {
+ unsigned char *datap;
+ datap = &entry->key.l7_bytes[4];
+ data = datap[0] + (datap[1] << 8) + (datap[2] << 16) + (datap[3] << 24);
+ HASH_PUSH_DWORD(cp, data);
+ }
+
+ total = (unsigned long)((unsigned long)cp - (unsigned long)destp) / (sizeof(u32));
+ if (total > HASH_MAX_KEY_DWORD) {
+ //printf("Total key words (%d) is too large (> %d)!\n",
+ // total, HASH_MAX_KEY_DWORD);
+ return -1;
+ }
+ index = hash_gen_crc16((unsigned char*)(destp + 1), (total-1)*4);
+ entry->index = index & ((1 << HASH_BITS)-1);
+
+ cp = (unsigned char*) destp;
+ cp += 3;
+ HASH_PUSH_BYTE(cp, entry->rule);
+ entry->total_dwords = total;
+ return total;
+}
+
+/*----------------------------------------------------------------------
+* hash_add_toe_entry
+*----------------------------------------------------------------------*/
+int hash_add_toe_entry(HASH_ENTRY_T *entry)
+{
+ int rc;
+ u32 key[HASH_MAX_DWORDS];
+
+ rc = hash_build_toe_keys((u32 *)&key, entry);
+ if (rc < 0)
+ return -1;
+ hash_write_entry(entry, (unsigned char*) &key[0]);
+ //hash_dump_entry(entry->index);
+// hash_set_valid_flag(entry->index, 1);
+// printk("Dump hash key!\n");
+// dump_hash_key(entry);
+ return entry->index;
+}
+
+
+/*----------------------------------------------------------------------
+* hash_write_entry
+*----------------------------------------------------------------------*/
+int hash_write_entry(HASH_ENTRY_T *entry, unsigned char *key)
+{
+ int i;
+ u32 *srcep, *destp, *destp2;
+
+ srcep = (u32 *)key;
+ destp2 = destp = (u32 *)&hash_tables[entry->index][0];
+
+ for (i=0; i<(entry->total_dwords); i++, srcep++, destp++)
+ *destp = *srcep;
+
+ srcep = (u32 *)&entry->action;
+ *destp++ = *srcep;
+
+ srcep = (u32 *)&entry->param;
+ for (i=0; i<(sizeof(ENTRY_PARAM_T)/sizeof(*destp)); i++, srcep++, destp++)
+ *destp = *srcep;
+
+ memset(destp, 0, (HASH_MAX_DWORDS-entry->total_dwords-HASH_ACTION_DWORDS) * sizeof(u32));
+
+ consistent_sync(destp2, (entry->total_dwords+HASH_ACTION_DWORDS) * 4, PCI_DMA_TODEVICE);
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* hash_timer_func
+*----------------------------------------------------------------------*/
+static void hash_timer_func(u32 data)
+{
+ int i, j, idx;
+ volatile u32 *own_p, *valid_p;
+ u32 own_bits, a_bits;
+ int period = HASH_TIMER_PERIOD;
+
+ valid_p = (volatile u32 *)TOE_V_BIT_BASE;
+ own_p = (volatile u32 *)hash_nat_owner_bits;
+ for (i=0, idx=0; i<(HASH_TOTAL_ENTRIES/32); i++, own_p++, valid_p++, idx+=32)
+ {
+ a_bits = readl(TOE_A_BIT_BASE + (i*4));
+ own_bits = *own_p;
+ if (own_bits)
+ {
+ for (j=0; own_bits && j<32; j++)
+ {
+ if (own_bits & 1)
+ {
+ short *counter_p, *interval_p;
+ NAT_HASH_ENTRY_T *nat_entry;
+ GRE_HASH_ENTRY_T *gre_entry;
+ nat_entry = (NAT_HASH_ENTRY_T *)hash_get_entry(idx+j);
+ gre_entry = (GRE_HASH_ENTRY_T *)nat_entry;
+ if (nat_entry->key.ip_protocol == IPPROTO_GRE)
+ {
+ counter_p = (short *)&gre_entry->tmo.counter;
+ interval_p = (short *)&gre_entry->tmo.interval;
+ }
+ else
+ {
+ counter_p = (short *)&nat_entry->tmo.counter;
+ interval_p = (short *)&nat_entry->tmo.interval;
+ }
+ if (a_bits & 1)
+ {
+ *counter_p = *interval_p;
+ }
+ else
+ {
+ *counter_p -= HASH_TIMER_PERIOD;
+ if (*counter_p <= 0)
+ {
+ *valid_p &= ~(1 << j); // invalidate it
+ *own_p &= ~(1 << j); // release ownership for NAT
+ *counter_p = 0;
+ // hash_printf("%lu %s: Clear hash index: %d\n", jiffies/HZ, __func__, i*32+j);
+ }
+ else if (period > *counter_p)
+ {
+ period = *counter_p;
+ }
+ }
+ }
+ a_bits >>= 1;
+ own_bits >>=1;
+ }
+ }
+ }
+
+ hash_timer_obj.expires = jiffies + (period * HZ);
+ add_timer((struct timer_list *)data);
+}
+
+/*----------------------------------------------------------------------
+* dm_long
+*----------------------------------------------------------------------*/
+void dm_long(u32 location, int length)
+{
+ u32 *start_p, *curr_p, *end_p;
+ u32 *datap, data;
+ int i;
+
+ //if (length > 1024)
+ // length = 1024;
+
+ start_p = (u32 *)location;
+ end_p = (u32 *)location + length;
+ curr_p = (u32 *)((u32)location & 0xfffffff0);
+ datap = (u32 *)location;
+ while (curr_p < end_p)
+ {
+ hash_printf("0x%08x: ",(u32)curr_p & 0xfffffff0);
+ for (i=0; i<4; i++)
+ {
+ if (curr_p < start_p || curr_p >= end_p)
+ hash_printf(" ");
+ else
+ {
+ data = *datap;
+ hash_printf("%08X ", data);
+ }
+ if (i==1)
+ hash_printf("- ");
+
+ curr_p++;
+ datap++;
+ }
+ hash_printf("\n");
+ }
+}
+
+/*----------------------------------------------------------------------
+* hash_dump_entry
+*----------------------------------------------------------------------*/
+void hash_dump_entry(int index)
+{
+ hash_printf("Hash Index %d:\n", index);
+ dm_long((u32)&hash_tables[index][0], HASH_MAX_DWORDS);
+}
+
+
--- /dev/null
+++ b/drivers/net/sl351x_nat.c
@@ -0,0 +1,1736 @@
+/****************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*----------------------------------------------------------------------------
+* Name : sl351x_nat.c
+* Description :
+* Handle Storlink SL351x NAT Functions
+*
+*
+* Packet Flow:
+*
+* (xmit)+<--- SW NAT -->+(xmit)
+* | ^^ |
+* | || |
+* | || |
+* Client <---> GMAC-x HW-NAT GMAC-y <---> Server
+*
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* 03/13/2006 Gary Chen Create and implement
+*
+*
+****************************************************************************/
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch/irqs.h>
+#include <asm/arch/it8712.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/skbuff.h>
+#include <linux/if_ether.h>
+#include <linux/if_pppox.h>
+#include <linux/in.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#include <linux/udp.h>
+#include <linux/ppp_defs.h>
+
+#define MIDWAY
+#define SL_LEPUS
+
+#include <asm/arch/sl2312.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <asm/arch/sl351x_hash_cfg.h>
+#include <asm/arch/sl351x_nat_cfg.h>
+#ifdef CONFIG_NETFILTER
+// #include <linux/netfilter/nf_conntrack.h>
+#include <linux/netfilter/nf_conntrack_tcp.h>
+#endif
+
+//#define NAT_DEBUG_MSG 1
+#define _NOT_CHECK_SIP_DIP
+//#define SL351x_NAT_TEST_BY_SMARTBITS 1 // Initialize 32 hash entries and test by SmartBITS
+#define VITESSE_G5SWITCH 1
+
+#ifdef CONFIG_SL351x_NAT
+
+/*----------------------------------------------------------------------
+* Definition
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL3516_ASIC
+#define CONFIG_SL351x_NAT_TCP_UDP
+#define CONFIG_SL351x_NAT_GRE
+#define CONFIG_SL351x_TCP_UDP_RULE_ID 0
+#define CONFIG_SL351x_GRE_RULE_ID 1
+#else
+#define CONFIG_SL351x_NAT_TCP_UDP
+//#define CONFIG_SL351x_NAT_GRE
+#define CONFIG_SL351x_TCP_UDP_RULE_ID 0
+#define CONFIG_SL351x_GRE_RULE_ID 0
+#endif
+
+#define nat_printf printk
+#define NAT_FTP_CTRL_PORT (21) // TCP
+#define NAT_H323_PORT (1720) // TCP
+#define NAT_T120_PORT (1503) // TCP
+#define NAT_PPTP_PORT (1723) // TCP
+#define NAT_TFTP_PORT (69) // UDP
+#define NAT_DNS_PORT (53) // UDP
+#define NAT_NTP_PORT (123) // UDP
+#define NAT_RAS_PORT (1719) // UDP
+#define NAT_BOOTP67_PORT (67) // UDP
+#define NAT_BOOTP68_PORT (68) // UDP
+
+#define NAT_TCP_PORT_MAX 64
+#define NAT_UDP_PORT_MAX 64
+
+#define GRE_PROTOCOL (0x880b)
+#define GRE_PROTOCOL_SWAP __constant_htons(0x880b)
+
+#ifdef VITESSE_G5SWITCH
+extern int Giga_switch;
+#endif
+
+typedef struct
+{
+ u16 flags_ver;
+ u16 protocol;
+ u16 payload_length;
+ u16 call_id;
+ u32 seq;
+ u32 ack;
+} GRE_PKTHDR_T;
+
+/*----------------------------------------------------------------------
+* NAT Configuration
+*
+* Note: Any change for network setting, the NAT configuration should
+* be changed also.
+* cfg->lan_port 0 if GMAC-0, 1: if GMAC-1
+* cfg->wan_port 0 if GMAC-0, 1: if GMAC-1
+* cfg->lan_ipaddr, cfg->lan_gateway, cfg->lan_netmask
+* cfg->wan_ipaddr, cfg->wan_gateway, cfg->wan_netmask
+*
+*----------------------------------------------------------------------*/
+NAT_CFG_T nat_cfg;
+static int nat_initialized;
+u32 nat_collision;
+
+#ifdef CONFIG_SL351x_NAT_TCP_UDP
+static u16 fixed_tcp_port_list[]={NAT_FTP_CTRL_PORT,
+ NAT_H323_PORT,
+ // NAT_T120_PORT,
+ NAT_PPTP_PORT,
+ 0};
+static u16 fixed_udp_port_list[]={NAT_DNS_PORT,
+ NAT_NTP_PORT,
+ NAT_TFTP_PORT,
+ NAT_RAS_PORT,
+ NAT_BOOTP67_PORT,
+ NAT_BOOTP68_PORT,
+ 0};
+#endif
+
+// #define _HAVE_DYNAMIC_PORT_LIST
+#ifdef _HAVE_DYNAMIC_PORT_LIST
+static u16 dynamic_tcp_port_list[NAT_TCP_PORT_MAX+1];
+static u16 dynamic_udp_port_list[NAT_UDP_PORT_MAX+1]};
+#endif
+
+/*----------------------------------------------------------------------
+* Functions
+*----------------------------------------------------------------------*/
+int sl351x_nat_tcp_udp_output(struct sk_buff *skb, int port);
+int sl351x_nat_udp_output(struct sk_buff *skb, int port);
+int sl351x_nat_gre_output(struct sk_buff *skb, int port);
+
+extern int mac_set_rule_reg(int mac, int rule, int enabled, u32 reg0, u32 reg1, u32 reg2);
+extern void hash_dump_entry(int index);
+extern void mac_get_hw_tx_weight(struct net_device *dev, char *weight);
+extern void mac_set_hw_tx_weight(struct net_device *dev, char *weight);
+
+#ifdef SL351x_NAT_TEST_BY_SMARTBITS
+static void nat_init_test_entry(void);
+#endif
+/*----------------------------------------------------------------------
+* sl351x_nat_init
+* initialize a NAT matching rule
+* Called by SL351x Driver
+* key : port, protocol, Sip, Dip, Sport, Dport
+* Action : Srce Q: HW Free Queue,
+* Dest Q: HW TxQ
+* Change DA
+* Change SA
+* Change Sip or Dip
+* Change Sport or Dport
+*----------------------------------------------------------------------*/
+void sl351x_nat_init(void)
+{
+ int rc;
+ GMAC_MRxCR0_T mrxcr0;
+ GMAC_MRxCR1_T mrxcr1;
+ GMAC_MRxCR2_T mrxcr2;
+ NAT_CFG_T *cfg;
+
+ if (nat_initialized)
+ return;
+
+ nat_initialized = 1;
+
+ if ((sizeof(NAT_HASH_ENTRY_T) > HASH_MAX_BYTES) ||
+ (sizeof(GRE_HASH_ENTRY_T) > HASH_MAX_BYTES))
+ {
+ nat_printf("NAT_HASH_ENTRY_T structure Size is too larger!\n");
+ while(1);
+ }
+
+ cfg = (NAT_CFG_T *)&nat_cfg;
+ memset((void *)cfg, 0, sizeof(NAT_CFG_T));
+#ifdef _HAVE_DYNAMIC_PORT_LIST
+ memset((void *)dynamic_tcp_port_list, 0, sizeof(dynamic_tcp_port_list));
+ memset((void *)dynamic_udp_port_list, 0, sizeof(dynamic_udp_port_list));
+#endif
+
+#ifdef VITESSE_G5SWITCH
+ if(Giga_switch)
+ {
+ cfg->enabled = 1;
+ cfg->tcp_udp_rule_id = CONFIG_SL351x_TCP_UDP_RULE_ID;
+ cfg->gre_rule_id = CONFIG_SL351x_GRE_RULE_ID;
+ cfg->lan_port = 1;
+ cfg->wan_port = 0;
+ cfg->default_hw_txq = 3;
+ cfg->tcp_tmo_interval = 60;
+ cfg->udp_tmo_interval = 180;
+ cfg->gre_tmo_interval = 60;
+ }
+ else
+ {
+ cfg->enabled = 1;
+ cfg->tcp_udp_rule_id = CONFIG_SL351x_TCP_UDP_RULE_ID;
+ cfg->gre_rule_id = CONFIG_SL351x_GRE_RULE_ID;
+ cfg->lan_port = 0;
+ cfg->wan_port = 1;
+ cfg->default_hw_txq = 3;
+ cfg->tcp_tmo_interval = 60;
+ cfg->udp_tmo_interval = 180;
+ cfg->gre_tmo_interval = 60;
+
+ }
+#endif
+
+#if 1 // debug purpose
+ cfg->ipcfg[0].total = 1;
+ cfg->ipcfg[0].entry[0].ipaddr = IPIV(192,168,2,92);
+ cfg->ipcfg[0].entry[0].netmask = IPIV(255,255,255,0);
+ cfg->ipcfg[1].total = 1;
+ cfg->ipcfg[1].entry[0].ipaddr = IPIV(192,168,1,200);
+ cfg->ipcfg[1].entry[0].netmask = IPIV(255,255,255,0);
+#endif
+
+#if 1
+ cfg->xport.total = 0;
+#else
+ cfg->xport.total = 4;
+
+ // H.323/H.225 Call setup
+ cfg->xport.entry[0].protocol = IPPROTO_TCP;
+ cfg->xport.entry[0].sport_start = 0;
+ cfg->xport.entry[0].sport_end = 0;
+ cfg->xport.entry[0].dport_start = 1720;
+ cfg->xport.entry[0].dport_end = 1720;
+ cfg->xport.entry[1].protocol = IPPROTO_TCP;
+ cfg->xport.entry[1].sport_start = 1720;
+ cfg->xport.entry[1].sport_end = 1720;
+ cfg->xport.entry[1].dport_start = 0;
+ cfg->xport.entry[1].dport_end = 0;
+
+ // RAS Setup
+ cfg->xport.entry[2].protocol = IPPROTO_UDP;
+ cfg->xport.entry[2].sport_start = 0;
+ cfg->xport.entry[2].sport_end = 0;
+ cfg->xport.entry[2].dport_start = 1719;
+ cfg->xport.entry[2].dport_end = 1719;
+ cfg->xport.entry[3].protocol = IPPROTO_UDP;
+ cfg->xport.entry[3].sport_start = 1719;
+ cfg->xport.entry[3].sport_end = 1719;
+ cfg->xport.entry[3].dport_start = 0;
+ cfg->xport.entry[3].dport_end = 0;
+#endif
+
+#ifdef CONFIG_SL351x_NAT_TCP_UDP
+ mrxcr0.bits32 = 0;
+ mrxcr1.bits32 = 0;
+ mrxcr2.bits32 = 0;
+ mrxcr0.bits.port = 1;
+ mrxcr0.bits.l3 = 1;
+ mrxcr0.bits.l4 = 1;
+ mrxcr1.bits.sip = 1;
+ mrxcr1.bits.dip = 1;
+ mrxcr1.bits.l4_byte0_15 = 0x0f; // Byte 0-3
+ mrxcr0.bits.sprx = 3;
+
+ rc = mac_set_rule_reg(cfg->lan_port, cfg->tcp_udp_rule_id, 1, mrxcr0.bits32, mrxcr1.bits32, mrxcr2.bits32);
+ if (rc < 0)
+ {
+ nat_printf("NAT Failed to set MAC-%d Rule %d!\n", cfg->lan_port, cfg->tcp_udp_rule_id);
+ }
+
+ if (cfg->lan_port != cfg->wan_port)
+ {
+ rc = mac_set_rule_reg(cfg->wan_port, cfg->tcp_udp_rule_id, 1, mrxcr0.bits32, mrxcr1.bits32, mrxcr2.bits32);
+ if (rc < 0)
+ {
+ nat_printf("NAT Failed to set MAC-%d Rule %d!\n", cfg->wan_port, cfg->tcp_udp_rule_id);
+ }
+ }
+#endif
+
+#ifdef CONFIG_SL351x_NAT_GRE
+ mrxcr0.bits32 = 0;
+ mrxcr1.bits32 = 0;
+ mrxcr2.bits32 = 0;
+ mrxcr0.bits.port = 1;
+ mrxcr0.bits.l3 = 1;
+ mrxcr0.bits.l4 = 1;
+ mrxcr1.bits.sip = 1;
+ mrxcr1.bits.dip = 1;
+ mrxcr1.bits.l4_byte0_15 = 0xcc; // Byte 2, 3, 6, 7
+ mrxcr0.bits.sprx = 4; // see GMAC driver about SPR
+
+ rc = mac_set_rule_reg(cfg->lan_port, cfg->gre_rule_id, 1, mrxcr0.bits32, mrxcr1.bits32, mrxcr2.bits32);
+ if (rc < 0)
+ {
+ nat_printf("NAT Failed to set MAC-%d Rule %d!\n", cfg->lan_port, cfg->gre_rule_id);
+ }
+
+ if (cfg->lan_port != cfg->wan_port)
+ {
+ rc = mac_set_rule_reg(cfg->wan_port, cfg->gre_rule_id, 1, mrxcr0.bits32, mrxcr1.bits32, mrxcr2.bits32);
+ if (rc < 0)
+ {
+ nat_printf("NAT Failed to set MAC-%d Rule %d!\n", cfg->wan_port, cfg->gre_rule_id);
+ }
+ }
+#endif
+
+#ifdef SL351x_NAT_TEST_BY_SMARTBITS
+ nat_init_test_entry();
+#endif
+}
+
+/*----------------------------------------------------------------------
+* nat_build_keys
+* Note: To call this routine, the key->rule_id MUST be zero
+*----------------------------------------------------------------------*/
+static inline int nat_build_keys(NAT_KEY_T *key)
+{
+ return hash_gen_crc16((unsigned char *)key, NAT_KEY_SIZE) & HASH_BITS_MASK;
+}
+
+/*----------------------------------------------------------------------
+* gre_build_keys
+* Note: To call this routine, the key->rule_id MUST be zero
+*----------------------------------------------------------------------*/
+static inline int gre_build_keys(GRE_KEY_T *key)
+{
+ return hash_gen_crc16((unsigned char *)key, GRE_KEY_SIZE) & HASH_BITS_MASK;
+}
+
+/*----------------------------------------------------------------------
+* nat_write_hash_entry
+*----------------------------------------------------------------------*/
+static inline int nat_write_hash_entry(int index, void *hash_entry)
+{
+ int i;
+ u32 *srcep, *destp, *destp2;
+
+ srcep = (u32 *)hash_entry;
+ destp = destp2 = (u32 *)&hash_tables[index][0];
+
+ for (i=0; i<(NAT_HASH_ENTRY_SIZE/sizeof(u32)); i++)
+ *destp++ = *srcep++;
+
+ consistent_sync(destp2, NAT_HASH_ENTRY_SIZE, PCI_DMA_TODEVICE);
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* gre_write_hash_entry
+*----------------------------------------------------------------------*/
+static inline int gre_write_hash_entry(int index, void *hash_entry)
+{
+ int i;
+ u32 *srcep, *destp, *destp2;
+
+ srcep = (u32 *)hash_entry;
+ destp = destp2 = (u32 *)&hash_tables[index][0];
+
+ for (i=0; i<(GRE_HASH_ENTRY_SIZE/sizeof(u32)); i++)
+ *destp++ = *srcep++;
+
+ consistent_sync(destp2, GRE_HASH_ENTRY_SIZE, PCI_DMA_TODEVICE);
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_find_ipcfg
+* return NULL if not found
+*----------------------------------------------------------------------*/
+static NAT_IP_ENTRY_T *sl351x_nat_find_ipcfg(u32 ipaddr, int port)
+{
+ int i;
+ NAT_IP_ENTRY_T *ipcfg;
+
+ ipcfg = (NAT_IP_ENTRY_T *)&nat_cfg.ipcfg[port].entry[0];
+ for (i=0; i<nat_cfg.ipcfg[port].total; i++, ipcfg++)
+ {
+ if (ipaddr == ipcfg->ipaddr)
+ {
+ return ipcfg;
+ }
+ }
+ return NULL;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_assign_qid
+*----------------------------------------------------------------------*/
+static int sl351x_nat_assign_qid(u8 proto, u32 sip, u32 dip, u16 sport, u16 dport)
+{
+ int i, total, qid;
+ NAT_WRULE_ENTRY_T *entry;
+
+ for (qid = 0; qid<CONFIG_NAT_TXQ_NUM; qid++)
+ {
+ if (qid == nat_cfg.default_hw_txq)
+ continue;
+
+ entry = (NAT_WRULE_ENTRY_T *)&nat_cfg.wrule[qid].entry[0];
+ total = nat_cfg.wrule[qid].total;
+ for (i=0; i<total; i++, entry++)
+ {
+ if (!entry->protocol || entry->protocol==proto)
+ {
+ //if (!entry->sip_start && !entry->dip_start && !entry->sport_start && !entry->dport_start)
+ // continue; // UI take care
+ if (entry->sip_start && !((sip >= entry->sip_start) &&
+ (sip <= entry->sip_end)))
+ continue;
+ if (entry->dip_start && !((dip >= entry->dip_start) &&
+ (dip <= entry->dip_end)))
+ continue;
+ if (entry->sport_start && !((sport >= entry->sport_start) &&
+ (sport <= entry->sport_end)))
+ continue;
+ if (entry->dport_start && !((dport >= entry->dport_start)
+ && (dport <= entry->dport_end)))
+ continue;
+ return qid;
+ }
+ }
+ }
+ return nat_cfg.default_hw_txq;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_input
+* Handle NAT input frames
+* Called by SL351x Driver - Handle Default Rx Queue
+* Notes: The caller must make sure that the l3off & l4offset should not be zero.
+* SL351x NAT Frames should meet the following conditions:
+* 1. TCP or UDP frame
+* 2. Cannot be special ALGs ports which TCP/UDP data is updated
+* 3. LAN-IN Frames:
+* Source IP is in the LAN subnet and Destination is not in the LAN subnet
+* 4. WAN-IN Frames
+* Destination IP is in the WAN port IP
+*
+* Example Ports
+* 1. TCP/UDP data is updated
+* (a) FTP Control Packet
+* (b) VoIP Packets
+* (c) etc. (add in future)
+* 2. UDP Low packet rate, not worth
+* (b) TFTP Destination Port is 69
+* (b) DNS 53
+* (c) NTP 123
+* (d) etc. (add in future)
+*----------------------------------------------------------------------*/
+void sl351x_nat_input(struct sk_buff *skb, int port, void *l3off, void *l4off)
+{
+ int i, found;
+ u32 sip, dip;
+ u16 sport, dport;
+ struct ethhdr *ether_hdr;
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ struct pppoe_hdr *pppoe_hdr;
+ NAT_CB_T *nat_cb;
+ u8 proto, pppoe_frame=0;
+ NAT_CFG_T *cfg;
+ u16 ppp_proto;
+ NAT_IP_ENTRY_T *ipcfg;
+ NAT_XPORT_ENTRY_T *xentry;
+ GRE_PKTHDR_T *gre_hdr;
+#ifdef CONFIG_SL351x_NAT_TCP_UDP
+ u16 *port_ptr;
+#endif
+
+ cfg = (NAT_CFG_T *)&nat_cfg;
+ if (!cfg->enabled || !cfg->ipcfg[port].total)
+ return;
+
+ ip_hdr = (struct iphdr *)&(skb->data[(u32)l3off]);
+ proto = ip_hdr->protocol;
+
+ tcp_hdr = (struct tcphdr *)&(skb->data[(u32)l4off]);
+ gre_hdr = (GRE_PKTHDR_T *)tcp_hdr;
+ sport = ntohs(tcp_hdr->source);
+ dport = ntohs(tcp_hdr->dest);
+
+ sip = ntohl(ip_hdr->saddr);
+ dip = ntohl(ip_hdr->daddr);
+
+ if (dip == IPIV(255,255,255,255))
+ return;
+
+ if (port == cfg->lan_port)
+ {
+ ipcfg = (NAT_IP_ENTRY_T *)&cfg->ipcfg[port].entry[0];
+ for (i=0, found=0; i<cfg->ipcfg[port].total; i++, ipcfg++)
+ {
+ u32 subnet = ipcfg->ipaddr & ipcfg->netmask;
+ if (((sip & ipcfg->netmask) == subnet) &&
+ ((dip & ipcfg->netmask) != subnet))
+ {
+ found = 1;
+ break;
+ }
+ }
+ if (!found)
+ return;
+ }
+ else
+ {
+#ifndef _NOT_CHECK_SIP_DIP // enable it if know and get the wan ip address
+ if (!sl351x_nat_find_ipcfg(dip, port))
+ {
+ printk("WAN->LAN Incorrect Dip %d.%d.%d.%d\n", HIPQUAD(dip));
+ return;
+ }
+#endif
+ ether_hdr = (struct ethhdr *)skb->data;
+ pppoe_hdr = (struct pppoe_hdr *)(ether_hdr + 1);
+ ppp_proto = *(u16 *)&pppoe_hdr->tag[0];
+ if (ether_hdr->h_proto == __constant_htons(ETH_P_PPP_SES) // 0x8864
+ && ppp_proto == __constant_htons(PPP_IP) ) // 0x21
+ {
+ pppoe_frame = 1;
+ }
+ }
+
+#ifdef CONFIG_SL351x_NAT_TCP_UDP
+ if (proto == IPPROTO_TCP)
+ {
+#ifdef NAT_DEBUG_MSG
+ nat_printf("From GMAC-%d: 0x%-4X TCP %d.%d.%d.%d [%d] --> %d.%d.%d.%d [%d]",
+ port, ntohs(ip_hdr->id),
+ NIPQUAD(ip_hdr->saddr), sport,
+ NIPQUAD(ip_hdr->daddr), dport);
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_SYN) nat_printf(" SYN");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_FIN) nat_printf(" FIN");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_RST) nat_printf(" RST");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_ACK) nat_printf(" ACK");
+ nat_printf("\n");
+#endif
+ // if (tcp_flag_word(tcp_hdr) & (TCP_FLAG_SYN | TCP_FLAG_FIN | TCP_FLAG_RST))
+ if (tcp_flag_word(tcp_hdr) & (TCP_FLAG_SYN))
+ {
+ return;
+ }
+ port_ptr = fixed_tcp_port_list;
+ for (i=0; *port_ptr; i++, port_ptr++)
+ {
+ if (sport == *port_ptr || dport == *port_ptr)
+ return;
+ }
+#ifdef _HAVE_DYNAMIC_PORT_LIST
+ port_ptr = dynamic_tcp_port_list;
+ for (i=0; *port_ptr; i++, port_ptr++)
+ {
+ if (sport == *port_ptr || dport == *port_ptr)
+ return;
+ }
+#endif
+ }
+ else if (proto == IPPROTO_UDP)
+ {
+#ifdef NAT_DEBUG_MSG
+ nat_printf("From GMAC-%d: 0x%-4X UDP %d.%d.%d.%d [%d] --> %d.%d.%d.%d [%d]",
+ port, ntohs(ip_hdr->id),
+ NIPQUAD(ip_hdr->saddr), sport,
+ NIPQUAD(ip_hdr->daddr), dport);
+ nat_printf("\n");
+#endif
+ port_ptr = fixed_udp_port_list;
+ for (i=0; *port_ptr; i++, port_ptr++)
+ {
+ if (sport == *port_ptr || dport == *port_ptr)
+ return;
+ }
+#ifdef _HAVE_DYNAMIC_PORT_LIST
+ port_ptr = dynamic_udp_port_list;
+ for (i=0; *port_ptr; i++, port_ptr++)
+ {
+ if (sport == *port_ptr || dport == *port_ptr)
+ return;
+ }
+#endif
+ }
+ else
+#endif // CONFIG_SL351x_NAT_TCP_UDP
+#ifdef CONFIG_SL351x_NAT_GRE
+ if (proto == IPPROTO_GRE)
+ {
+ if (gre_hdr->protocol != GRE_PROTOCOL_SWAP)
+ return;
+#ifdef NAT_DEBUG_MSG
+ nat_printf("From GMAC-%d: 0x%-4X GRE %d.%d.%d.%d [%d] --> %d.%d.%d.%d",
+ port, ntohs(ip_hdr->id),
+ NIPQUAD(ip_hdr->saddr), ntohs(gre_hdr->call_id),
+ NIPQUAD(ip_hdr->daddr));
+ nat_printf("\n");
+#endif
+ }
+ else
+#endif
+ return;
+
+
+ // check xport list
+ xentry = (NAT_XPORT_ENTRY_T *)&cfg->xport.entry[0];
+ for (i=0; i<cfg->xport.total; i++, xentry++)
+ {
+ if (!xentry->protocol || xentry->protocol == proto)
+ {
+ //if (!xentry->sport_start && !xentry->dport_start) // UI take care
+ // continue;
+ if (xentry->sport_start && !((sport >= xentry->sport_start) &&
+ (sport <= xentry->sport_end)))
+ continue;
+ if (xentry->dport_start && !((dport >= xentry->dport_start)
+ && (dport <= xentry->dport_end)))
+ continue;
+ return;
+ }
+ }
+
+ nat_cb = NAT_SKB_CB(skb);
+ if (((u32)nat_cb & 3))
+ {
+ nat_printf("%s ERROR! nat_cb is not alignment!!!!!!\n", __func__);
+ return;
+ }
+ nat_cb->tag = NAT_CB_TAG;
+ memcpy(nat_cb->sa, skb->data+6, 6);
+ nat_cb->sip = ip_hdr->saddr;
+ nat_cb->dip = ip_hdr->daddr;
+ if (proto == IPPROTO_GRE)
+ {
+ nat_cb->sport = gre_hdr->protocol;
+ nat_cb->dport = gre_hdr->call_id;
+ }
+ else
+ {
+ nat_cb->sport = tcp_hdr->source;
+ nat_cb->dport = tcp_hdr->dest;
+ }
+ nat_cb->pppoe_frame = pppoe_frame;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_output
+* Handle NAT output frames
+* Called by SL351x Driver - Transmit
+*
+* 1. If not SL351x NAT frames, return FALSE
+* 2. LAN-to-WAN frames
+* (1) Sip must be WAN IP
+* 3. If TCP SY/RST/FIN frame, return
+* 4. Build the hash key and get the hash index
+* 5. If V-Bit is ON, return.
+* 6. Write hash entry and validate it
+*
+*----------------------------------------------------------------------*/
+int sl351x_nat_output(struct sk_buff *skb, int port)
+{
+ struct iphdr *ip_hdr;
+ u8 proto;
+ NAT_CB_T *nat_cb;
+
+ nat_cb = NAT_SKB_CB(skb);
+ if (nat_cb->tag != NAT_CB_TAG)
+ return 0;
+
+ if (((u32)nat_cb & 3))
+ {
+ nat_printf("%s ERROR! nat_cb is not alignment!!!!!!\n", __func__);
+ return 0;
+ }
+ ip_hdr = (struct iphdr *)skb->h.ipiph;
+ proto = ip_hdr->protocol;
+
+ switch (proto)
+ {
+ case IPPROTO_TCP:
+ case IPPROTO_UDP:
+ return sl351x_nat_tcp_udp_output(skb, port);
+ case IPPROTO_GRE:
+ return sl351x_nat_gre_output(skb, port);
+ }
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_tcp_udp_output
+* Handle NAT TCP/UDP output frames
+*----------------------------------------------------------------------*/
+int sl351x_nat_tcp_udp_output(struct sk_buff *skb, int port)
+{
+ u32 sip, dip;
+ struct ethhdr *ether_hdr;
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ struct pppoe_hdr *pppoe_hdr;
+ NAT_CB_T *nat_cb;
+ NAT_CFG_T *cfg;
+ u8 proto;
+ u16 sport, dport, ppp_proto;
+ u32 hash_data[HASH_MAX_DWORDS];
+ NAT_HASH_ENTRY_T *hash_entry;
+ int hash_index;
+ struct ip_conntrack *nat_ip_conntrack;
+ enum ip_conntrack_info ctinfo;
+
+ nat_cb = NAT_SKB_CB(skb);
+ cfg = (NAT_CFG_T *)&nat_cfg;
+
+ ether_hdr = (struct ethhdr *)skb->data;
+ ip_hdr = (struct iphdr *)skb->h.ipiph;
+ tcp_hdr = (struct tcphdr *)((u32)ip_hdr + (ip_hdr->ihl<<2));
+ sip = ntohl(ip_hdr->saddr);
+ dip = ntohl(ip_hdr->daddr);
+ proto = ip_hdr->protocol;
+ sport = ntohs(tcp_hdr->source);
+ dport = ntohs(tcp_hdr->dest);
+
+#ifdef NAT_DEBUG_MSG
+ {
+ nat_printf("To GMAC-%d: 0x%-4X [%d] %d.%d.%d.%d [%d] --> %d.%d.%d.%d [%d]",
+ port, ntohs(ip_hdr->id), proto,
+ NIPQUAD(ip_hdr->saddr), sport,
+ NIPQUAD(ip_hdr->daddr), dport);
+ if (proto == IPPROTO_TCP)
+ {
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_SYN) nat_printf(" SYN");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_FIN) nat_printf(" FIN");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_RST) nat_printf(" RST");
+ if (tcp_flag_word(tcp_hdr) & TCP_FLAG_ACK) nat_printf(" ACK");
+ }
+ nat_printf("\n");
+ }
+#endif
+ nat_ip_conntrack = ip_conntrack_get(skb, &ctinfo);
+ if (!nat_ip_conntrack)
+ {
+ nat_printf("IP conntrack info is not found!\n");
+ return 0;
+ }
+ // nat_printf("nat_ip_conntrack = 0x%x, status=0x%lx, ctinfo=%d\n", (u32)nat_ip_conntrack, nat_ip_conntrack->status, ctinfo);
+ // if (nat_ip_conntrack->master || nat_ip_conntrack->helper)
+ if (nat_ip_conntrack->helper)
+ {
+ nat_printf("Sport=%d Dport=%d master=0x%x, helper=0x%x\n", sport, dport, (u32)nat_ip_conntrack->master, (u32)nat_ip_conntrack->helper);
+ return 0;
+ }
+
+ //if (proto == IPPROTO_TCP && !(nat_ip_conntrack->status & IPS_ASSURED))
+ // return 0;
+
+#ifdef NAT_DEBUG_MSG
+ nat_printf("nat_ip_conntrack=0x%x, nat_cb->state=%d\n", (u32)nat_ip_conntrack, nat_cb->state);
+ nat_printf("lan2wan_hash_index=%d, wan2lan_hash_index=%d\n", nat_ip_conntrack->lan2wan_hash_index, nat_ip_conntrack->wan2lan_hash_index);
+ nat_printf("lan2wan_collision=%d, wan2lan_collision=%d\n", nat_ip_conntrack->lan2wan_collision, nat_ip_conntrack->wan2lan_collision);
+#endif
+ if (proto == IPPROTO_TCP)
+ {
+ if (nat_cb->state >= TCP_CONNTRACK_FIN_WAIT && nat_cb->state <= TCP_CONNTRACK_CLOSE)
+ {
+ if (nat_ip_conntrack->lan2wan_hash_index)
+ {
+#ifdef NAT_DEBUG_MSG
+ nat_printf("Invalidate LAN->WAN hash entry %d\n", nat_ip_conntrack->lan2wan_hash_index - 1);
+#endif
+ hash_nat_disable_owner(nat_ip_conntrack->lan2wan_hash_index - 1);
+ hash_invalidate_entry(nat_ip_conntrack->lan2wan_hash_index - 1);
+ nat_ip_conntrack->lan2wan_hash_index = 0;
+ }
+ if (nat_ip_conntrack->wan2lan_hash_index)
+ {
+#ifdef NAT_DEBUG_MSG
+ nat_printf("Invalidate WAN->LAN hash entry %d\n", nat_ip_conntrack->wan2lan_hash_index - 1);
+#endif
+ hash_nat_disable_owner(nat_ip_conntrack->wan2lan_hash_index - 1);
+ hash_invalidate_entry(nat_ip_conntrack->wan2lan_hash_index - 1);
+ nat_ip_conntrack->wan2lan_hash_index = 0;
+ }
+ return 0;
+
+ }
+ else if (nat_cb->state != TCP_CONNTRACK_ESTABLISHED)
+ {
+ return 0;
+ }
+ }
+ if (proto == IPPROTO_TCP && (tcp_flag_word(tcp_hdr) & (TCP_FLAG_SYN | TCP_FLAG_FIN | TCP_FLAG_RST)))
+ // if (proto == IPPROTO_TCP && (tcp_flag_word(tcp_hdr) & (TCP_FLAG_SYN)))
+ return 0;
+
+ hash_entry = (NAT_HASH_ENTRY_T *)&hash_data;
+ if (port == cfg->wan_port) // LAN-to-WAN
+ {
+ if (nat_ip_conntrack->lan2wan_hash_index || nat_ip_conntrack->lan2wan_collision)
+ return 0;
+#ifndef _NOT_CHECK_SIP_DIP // enable it if know and get the wan ip address
+ if (!sl351x_nat_find_ipcfg(sip, port))
+ {
+ printk("LAN->WAN Incorrect Sip %d.%d.%d.%d\n", HIPQUAD(sip));
+ return 0;
+ }
+#endif
+ // Note: unused fields (including rule_id) MUST be zero
+ hash_entry->key.Ethertype = 0;
+ hash_entry->key.port_id = cfg->lan_port;
+ hash_entry->key.rule_id = 0;
+ hash_entry->key.ip_protocol = proto;
+ hash_entry->key.reserved1 = 0;
+ hash_entry->key.reserved2 = 0;
+ hash_entry->key.sip = ntohl(nat_cb->sip);
+ hash_entry->key.dip = ntohl(nat_cb->dip);
+ hash_entry->key.sport = nat_cb->sport;
+ hash_entry->key.dport = nat_cb->dport;
+
+ hash_index = nat_build_keys(&hash_entry->key);
+
+#ifdef NAT_DEBUG_LAN_HASH_TIMEOUT
+ if (hash_get_nat_owner_flag(hash_index))
+ return 0;
+#endif
+ if (hash_get_valid_flag(hash_index))
+ {
+ nat_ip_conntrack->lan2wan_collision = 1;
+ nat_collision++;
+#if 0
+ if (proto == IPPROTO_TCP && (tcp_flag_word(tcp_hdr) & (TCP_FLAG_FIN | TCP_FLAG_RST)))
+ {
+ if (memcmp((void *)&hash_entry->key, hash_get_entry(hash_index), sizeof(NAT_KEY_T)) == 0)
+ {
+ hash_nat_disable_owner(hash_index);
+ hash_invalidate_entry(hash_index); // Must last one, else HW Tx fast SW
+ // nat_printf("Invalidate nat hash entry %d\n", hash_index);
+ }
+ }
+#endif
+ return 0;
+ }
+
+ // write hash entry
+ hash_entry->key.rule_id = cfg->tcp_udp_rule_id;
+ memcpy(hash_entry->param.da, skb->data, 6);
+ memcpy(hash_entry->param.sa, skb->data+6, 6);
+ hash_entry->param.Sip = sip;
+ hash_entry->param.Dip = dip;
+ hash_entry->param.Sport = sport;
+ hash_entry->param.Dport = dport;
+ hash_entry->param.vlan = 0;
+ hash_entry->param.sw_id = 0;
+ hash_entry->param.mtu = 0;
+ // check PPPoE
+ pppoe_hdr = (struct pppoe_hdr *)(ether_hdr + 1);
+ ppp_proto = *(u16 *)&pppoe_hdr->tag[0];
+ if (ether_hdr->h_proto == __constant_htons(ETH_P_PPP_SES) // 0x8864
+ && ppp_proto == __constant_htons(PPP_IP) ) // 0x21
+ {
+ hash_entry->action.dword = NAT_PPPOE_LAN2WAN_ACTIONS;
+ hash_entry->param.pppoe = htons(pppoe_hdr->sid);
+ }
+ else
+ {
+ hash_entry->action.dword = NAT_LAN2WAN_ACTIONS;
+ hash_entry->param.pppoe = 0;
+ }
+ hash_entry->action.bits.dest_qid = sl351x_nat_assign_qid(proto, sip, dip, sport, dport);
+ hash_entry->action.bits.dest_qid += (cfg->wan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;
+ hash_entry->tmo.counter = hash_entry->tmo.interval =
+ (proto == IPPROTO_TCP) ? cfg->tcp_tmo_interval : cfg->udp_tmo_interval;
+ nat_write_hash_entry(hash_index, hash_entry);
+ // nat_printf("%lu Validate a LAN hash entry %d\n", jiffies/HZ, hash_index);
+ // hash_dump_entry(hash_index);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast than SW
+ nat_ip_conntrack->lan2wan_hash_index = hash_index + 1;
+ nat_ip_conntrack->hw_nat |= 1;
+ return 0;
+ }
+ else // WAN-to-LAN
+ {
+ if (nat_ip_conntrack->wan2lan_hash_index || nat_ip_conntrack->wan2lan_collision)
+ return 0;
+
+ // Note: unused fields (including rule_id) MUST be zero
+ hash_entry->key.Ethertype = 0;
+ hash_entry->key.port_id = cfg->wan_port;
+ hash_entry->key.rule_id = 0;
+ hash_entry->key.ip_protocol = proto;
+ hash_entry->key.reserved1 = 0;
+ hash_entry->key.reserved2 = 0;
+ hash_entry->key.sip = ntohl(nat_cb->sip);
+ hash_entry->key.dip = ntohl(nat_cb->dip);
+ hash_entry->key.sport = nat_cb->sport;
+ hash_entry->key.dport = nat_cb->dport;
+
+ hash_index = nat_build_keys(&hash_entry->key);
+
+#ifdef NAT_DEBUG_WAN_HASH_TIMEOUT
+ if (hash_get_nat_owner_flag(hash_index))
+ return 0;
+#endif
+ if (hash_get_valid_flag(hash_index))
+ {
+ nat_ip_conntrack->wan2lan_collision = 1;
+ nat_collision++;
+#if 0
+ if (proto == IPPROTO_TCP && (tcp_flag_word(tcp_hdr) & (TCP_FLAG_FIN | TCP_FLAG_RST)))
+ {
+ if (memcmp((void *)&hash_entry->key, hash_get_entry(hash_index), sizeof(NAT_KEY_T)) == 0)
+ {
+ hash_nat_disable_owner(hash_index);
+ hash_invalidate_entry(hash_index); // Must last one, else HW Tx fast SW
+ // nat_printf("Invalidate nat hash entry %d\n", hash_index);
+ }
+ }
+#endif
+ return 0;
+ }
+
+ // write hash entry
+ hash_entry->key.rule_id = cfg->tcp_udp_rule_id;
+ memcpy(hash_entry->param.da, skb->data, 6);
+ memcpy(hash_entry->param.sa, skb->data+6, 6);
+ hash_entry->param.Sip = sip;
+ hash_entry->param.Dip = dip;
+ hash_entry->param.Sport = sport;
+ hash_entry->param.Dport = dport;
+ hash_entry->param.vlan = 0;
+ hash_entry->param.pppoe = 0;
+ hash_entry->param.sw_id = 0;
+ hash_entry->param.mtu = 0;
+ hash_entry->action.dword = (nat_cb->pppoe_frame) ? NAT_PPPOE_WAN2LAN_ACTIONS : NAT_WAN2LAN_ACTIONS;
+ hash_entry->action.bits.dest_qid = sl351x_nat_assign_qid(proto, sip, dip, sport, dport);
+ hash_entry->action.bits.dest_qid += (cfg->lan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;;
+ hash_entry->tmo.counter = hash_entry->tmo.interval =
+ (proto == IPPROTO_TCP) ? cfg->tcp_tmo_interval : cfg->udp_tmo_interval;
+ nat_write_hash_entry(hash_index, hash_entry);
+
+ // nat_printf("%lu Validate a WAN hash entry %d\n", jiffies/HZ, hash_index);
+ // hash_dump_entry(hash_index);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast SW
+ nat_ip_conntrack->wan2lan_hash_index = hash_index + 1;
+ nat_ip_conntrack->hw_nat |= 2;
+ return 0;
+ }
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* sl351x_nat_gre_output
+* Handle NAT GRE output frames
+*----------------------------------------------------------------------*/
+int sl351x_nat_gre_output(struct sk_buff *skb, int port)
+{
+ u32 sip, dip;
+ struct ethhdr *ether_hdr;
+ struct iphdr *ip_hdr;
+ struct pppoe_hdr *pppoe_hdr;
+ GRE_PKTHDR_T *gre_hdr;
+ NAT_CB_T *nat_cb;
+ NAT_CFG_T *cfg;
+ u16 ppp_proto;
+ u32 hash_data[HASH_MAX_DWORDS];
+ GRE_HASH_ENTRY_T *hash_entry;
+ int hash_index;
+ struct ip_conntrack *nat_ip_conntrack;
+ enum ip_conntrack_info ctinfo;
+
+ nat_cb = NAT_SKB_CB(skb);
+ cfg = (NAT_CFG_T *)&nat_cfg;
+
+ ether_hdr = (struct ethhdr *)skb->data;
+ ip_hdr = (struct iphdr *)skb->h.ipiph;
+ gre_hdr = (GRE_PKTHDR_T *)((u32)ip_hdr + (ip_hdr->ihl<<2));
+ sip = ntohl(ip_hdr->saddr);
+ dip = ntohl(ip_hdr->daddr);
+
+#ifdef NAT_DEBUG_MSG
+ {
+ nat_printf("To GMAC-%d: 0x%-4X GRE %d.%d.%d.%d [%d] --> %d.%d.%d.%d",
+ port, ntohs(ip_hdr->id),
+ NIPQUAD(ip_hdr->saddr), ntohs(gre_hdr->call_id),
+ NIPQUAD(ip_hdr->daddr));
+ nat_printf("\n");
+ }
+#endif
+ nat_ip_conntrack = ip_conntrack_get(skb, &ctinfo);
+ if (nat_ip_conntrack)
+ {
+ // if (nat_ip_conntrack->master || nat_ip_conntrack->helper)
+ if (nat_ip_conntrack->helper)
+ {
+ nat_printf("GRE Call-ID=%d, master=0x%x, helper=0x%x\n", ntohs(gre_hdr->call_id), (u32)nat_ip_conntrack->master, (u32)nat_ip_conntrack->helper);
+ return 0;
+ }
+ if (!(nat_ip_conntrack->status & IPS_ASSURED))
+ return 0;
+ }
+
+ hash_entry = (GRE_HASH_ENTRY_T *)&hash_data;
+ if (port == cfg->wan_port) // LAN-to-WAN
+ {
+#ifdef _NOT_CHECK_SIP_DIP // enable it if know and get the wan ip address
+ if (!sl351x_nat_find_ipcfg(sip, port))
+ {
+ printk("LAN->WAN Incorrect Sip %d.%d.%d.%d\n", HIPQUAD(sip));
+ return 0;
+ }
+#endif
+ // Note: unused fields (including rule_id) MUST be zero
+ hash_entry->key.Ethertype = 0;
+ hash_entry->key.port_id = cfg->lan_port;
+ hash_entry->key.rule_id = 0;
+ hash_entry->key.ip_protocol = IPPROTO_GRE;
+ hash_entry->key.reserved1 = 0;
+ hash_entry->key.reserved2 = 0;
+ hash_entry->key.reserved3 = 0;
+ hash_entry->key.reserved4 = 0;
+ hash_entry->key.sip = ntohl(nat_cb->sip);
+ hash_entry->key.dip = ntohl(nat_cb->dip);
+ hash_entry->key.protocol = nat_cb->sport;
+ hash_entry->key.call_id = nat_cb->dport;
+
+ hash_index = gre_build_keys(&hash_entry->key);
+
+#ifdef NAT_DEBUG_LAN_HASH_TIMEOUT
+ if (hash_get_nat_owner_flag(hash_index))
+ return 0;
+#endif
+ if (hash_get_valid_flag(hash_index))
+ {
+ return 0;
+ }
+
+ // write hash entry
+ hash_entry->key.rule_id = cfg->gre_rule_id;
+ memcpy(hash_entry->param.da, skb->data, 6);
+ memcpy(hash_entry->param.sa, skb->data+6, 6);
+ hash_entry->param.Sip = sip;
+ hash_entry->param.Dip = dip;
+ hash_entry->param.Sport = 0;
+ hash_entry->param.Dport = ntohs(gre_hdr->call_id);
+ hash_entry->param.vlan = 0;
+ hash_entry->param.sw_id = 0;
+ hash_entry->param.mtu = 0;
+ // check PPPoE
+ pppoe_hdr = (struct pppoe_hdr *)(ether_hdr + 1);
+ ppp_proto = *(u16 *)&pppoe_hdr->tag[0];
+ if (ether_hdr->h_proto == __constant_htons(ETH_P_PPP_SES) // 0x8864
+ && ppp_proto == __constant_htons(PPP_IP) ) // 0x21
+ {
+ hash_entry->action.dword = NAT_PPPOE_PPTP_LAN2WAN_ACTIONS;
+ hash_entry->param.pppoe = htons(pppoe_hdr->sid);
+ }
+ else
+ {
+ hash_entry->action.dword = NAT_PPTP_LAN2WAN_ACTIONS;
+ hash_entry->param.pppoe = 0;
+ }
+ hash_entry->action.bits.dest_qid = sl351x_nat_assign_qid(IPPROTO_GRE, sip, dip, 0, ntohs(gre_hdr->call_id));
+ hash_entry->action.bits.dest_qid += (cfg->wan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;
+ hash_entry->tmo.counter = hash_entry->tmo.interval = cfg->gre_tmo_interval;
+ gre_write_hash_entry(hash_index, hash_entry);
+ // nat_printf("%lu Validate a LAN hash entry %d\n", jiffies/HZ, hash_index);
+ // hash_dump_entry(hash_index);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast than SW
+ return 0;
+ }
+ else // WAN-to-LAN
+ {
+ // Note: unused fields (including rule_id) MUST be zero
+ hash_entry->key.Ethertype = 0;
+ hash_entry->key.port_id = cfg->wan_port;
+ hash_entry->key.rule_id = 0;
+ hash_entry->key.ip_protocol = IPPROTO_GRE;
+ hash_entry->key.reserved1 = 0;
+ hash_entry->key.reserved2 = 0;
+ hash_entry->key.reserved3 = 0;
+ hash_entry->key.reserved4 = 0;
+ hash_entry->key.sip = ntohl(nat_cb->sip);
+ hash_entry->key.dip = ntohl(nat_cb->dip);
+ hash_entry->key.protocol = nat_cb->sport;
+ hash_entry->key.call_id = nat_cb->dport;
+
+ hash_index = gre_build_keys(&hash_entry->key);
+
+#ifdef NAT_DEBUG_WAN_HASH_TIMEOUT
+ if (hash_get_nat_owner_flag(hash_index))
+ return 0;
+#endif
+ if (hash_get_valid_flag(hash_index))
+ {
+ return 0;
+ }
+
+ // write hash entry
+ hash_entry->key.rule_id = cfg->gre_rule_id;
+ memcpy(hash_entry->param.da, skb->data, 6);
+ memcpy(hash_entry->param.sa, skb->data+6, 6);
+ hash_entry->param.Sip = sip;
+ hash_entry->param.Dip = dip;
+ hash_entry->param.Sport = 0;
+ hash_entry->param.Dport = ntohs(gre_hdr->call_id);
+ hash_entry->param.vlan = 0;
+ hash_entry->param.pppoe = 0;
+ hash_entry->param.sw_id = 0;
+ hash_entry->param.mtu = 0;
+ hash_entry->action.dword = (nat_cb->pppoe_frame) ? NAT_PPPOE_PPTP_WAN2LAN_ACTIONS : NAT_PPTP_WAN2LAN_ACTIONS;
+ hash_entry->action.bits.dest_qid = sl351x_nat_assign_qid(IPPROTO_GRE, sip, dip, 0, ntohs(gre_hdr->call_id));
+ hash_entry->action.bits.dest_qid += (cfg->lan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;;
+ hash_entry->tmo.counter = hash_entry->tmo.interval = cfg->gre_tmo_interval;
+ gre_write_hash_entry(hash_index, hash_entry);
+
+ // nat_printf("%lu Validate a WAN hash entry %d\n", jiffies/HZ, hash_index);
+ // hash_dump_entry(hash_index);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast SW
+ return 0;
+ }
+ return 0;
+}
+
+
+#ifdef _HAVE_DYNAMIC_PORT_LIST
+/*----------------------------------------------------------------------
+* sl_nat_add_port
+*----------------------------------------------------------------------*/
+void sl_nat_add_port(u8 protocol, u16 port)
+{
+ int i;
+ u16 *port_ptr;
+
+ if (protocol == IPPROTO_TCP)
+ port_ptr = dynamic_tcp_port_list;
+ else if (protocol == IPPROTO_UDP)
+ port_ptr = dynamic_udp_port_list;
+ else
+ return;
+
+ for (i=0; *port_ptr; i++)
+ {
+ if (port == *port_ptr)
+ return;
+ port_ptr++;
+ }
+ port_ptr++;
+ *port_ptr = port;
+}
+
+/*----------------------------------------------------------------------
+* sl_nat_remove_port
+*----------------------------------------------------------------------*/
+void sl_nat_remove_port(u8 protocol, u16 port)
+{
+ int i, j;
+ u16 *port_ptr, *next;
+
+ if (protocol == IPPROTO_TCP)
+ port_ptr = dynamic_tcp_port_list;
+ else if (protocol == IPPROTO_UDP)
+ port_ptr = dynamic_udp_port_list;
+ else
+ return;
+
+ for (i=0; *port_ptr; i++, port_ptr++)
+ {
+ if (port == *port_ptr)
+ {
+ port_next = port_ptr + 1;
+ for (j=i+1; *port_next; i++, j++)
+ *port_ptr++ = *port_next++;
+ *port_ptr = 0;
+ return;
+ }
+ }
+}
+#endif
+
+/*----------------------------------------------------------------------
+* sl351x_nat_ioctl
+*----------------------------------------------------------------------*/
+int sl351x_nat_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
+{
+ GMAC_INFO_T *tp = (GMAC_INFO_T *)dev->priv;
+ int i, j, port_id;
+ NATCMD_HDR_T nat_hdr;
+ NAT_REQ_E ctrl;
+ unsigned char *req_datap;
+ NAT_IP_ENTRY_T *ipcfg;
+ NAT_XPORT_ENTRY_T *xport_entry;
+ NAT_WRULE_ENTRY_T *wrule_entry;
+ unsigned int qid;
+
+ if (copy_from_user((void *)&nat_hdr, rq->ifr_data, sizeof(nat_hdr)))
+ return -EFAULT;
+ req_datap = (unsigned char *)rq->ifr_data + sizeof(nat_hdr);
+ port_id = tp->port_id;
+ switch (nat_hdr.cmd) {
+ case NATSSTATUS:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_STATUS_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.status, req_datap, sizeof(ctrl.status)))
+ return -EFAULT;
+ if (ctrl.status.enable != 0 && ctrl.status.enable != 1)
+ return -EPERM;
+ // sl351x_nat_set_enabled_flag(ctrl.status.enable);
+ if (nat_cfg.enabled && (ctrl.status.enable == 0))
+ {
+ for (i=0; i<HASH_TOTAL_ENTRIES; i++)
+ {
+ if (hash_get_nat_owner_flag(i))
+ {
+ hash_nat_disable_owner(i);
+ hash_invalidate_entry(i);
+ }
+ }
+ }
+ nat_cfg.enabled = ctrl.status.enable;
+ break;
+ case NATGSTATUS:
+ if (nat_hdr.len != sizeof(NAT_STATUS_T))
+ return -EPERM;
+ ctrl.status.enable = nat_cfg.enabled;
+ if (copy_to_user(req_datap, (void *)&ctrl.status, sizeof(ctrl.status)))
+ return -EFAULT;
+ break;
+ case NATSETPORT:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_PORTCFG_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.portcfg, req_datap, sizeof(ctrl.portcfg)))
+ return -EFAULT;
+ if (ctrl.portcfg.portmap == 0)
+ nat_cfg.lan_port = port_id;
+ else if (ctrl.portcfg.portmap == 1)
+ nat_cfg.wan_port = port_id;
+ else
+ return -EPERM;
+ break;
+ case NATGETPORT:
+ if (nat_hdr.len != sizeof(NAT_PORTCFG_T))
+ return -EPERM;
+ if (nat_cfg.lan_port == port_id)
+ ctrl.portcfg.portmap = 0;
+ else if (nat_cfg.wan_port == port_id)
+ ctrl.portcfg.portmap = 1;
+ else
+ return -EPERM;
+ if (copy_to_user(req_datap, (void *)&ctrl.portcfg, sizeof(ctrl.portcfg)))
+ return -EFAULT;
+ break;
+ case NATADDIP:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_IPCFG_T))
+ return -EPERM;
+ i = nat_cfg.ipcfg[port_id].total;
+ if (i >= CONFIG_NAT_MAX_IP_NUM)
+ return -E2BIG;
+ if (copy_from_user((void *)&nat_cfg.ipcfg[port_id].entry[i], req_datap, sizeof(NAT_IPCFG_T)))
+ return -EFAULT;
+ nat_cfg.ipcfg[port_id].total++;
+ break;
+ case NATDELIP:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_IPCFG_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.ipcfg, req_datap, sizeof(ctrl.ipcfg)))
+ return -EFAULT;
+ ipcfg = (NAT_IP_ENTRY_T *)&nat_cfg.ipcfg[port_id].entry[0];
+ for (i=0; i<nat_cfg.ipcfg[port_id].total; i++, ipcfg++)
+ {
+ if (ipcfg->ipaddr == ctrl.ipcfg.entry.ipaddr)
+ {
+ NAT_IP_ENTRY_T *ipcfg_next;
+ ipcfg_next = ipcfg + 1;
+ for (j=i+1; j < nat_cfg.ipcfg[port_id].total; i++, j++)
+ {
+ memcpy((void *)ipcfg, (void *)ipcfg_next, sizeof(NAT_IP_ENTRY_T));
+ ipcfg++;
+ ipcfg_next++;
+ }
+ ipcfg->ipaddr = 0;
+ ipcfg->netmask = 0;
+ nat_cfg.ipcfg[port_id].total--;
+ return 0;
+ }
+ }
+ return -ENOENT;
+ case NATGETIP:
+ if (nat_hdr.len != sizeof(NAT_IPCFG_ALL_T))
+ return -EPERM;
+ if (copy_to_user(req_datap, (void *)&nat_cfg.ipcfg[port_id], sizeof(NAT_IPCFG_ALL_T)))
+ return -EFAULT;
+ break;
+ case NATAXPORT:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_XPORT_T))
+ return -EPERM;
+ i = nat_cfg.xport.total;
+ if (i >= CONFIG_NAT_MAX_XPORT)
+ return -E2BIG;
+ if (copy_from_user((void *)&nat_cfg.xport.entry[i], req_datap, sizeof(NAT_XPORT_T)))
+ return -EFAULT;
+ nat_cfg.xport.total++;
+ break;
+ case NATDXPORT:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_XPORT_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.xport, req_datap, sizeof(NAT_XPORT_T)))
+ return -EFAULT;
+ xport_entry = (NAT_XPORT_ENTRY_T *)&nat_cfg.xport.entry[0];
+ for (i=0; i<nat_cfg.xport.total; i++, xport_entry++)
+ {
+ if (memcmp((void *)xport_entry, (void *)&ctrl.xport, sizeof(NAT_XPORT_ENTRY_T)) == 0)
+ {
+ NAT_XPORT_ENTRY_T *xport_next;
+ xport_next = xport_entry + 1;
+ for (j=i+1; j < nat_cfg.xport.total; i++, j++)
+ {
+ memcpy((void *)xport_entry, (void *)xport_next, sizeof(NAT_XPORT_ENTRY_T));
+ xport_entry++;
+ xport_next++;
+ }
+ memset((void *)xport_entry, 0, sizeof(NAT_XPORT_ENTRY_T));
+ nat_cfg.xport.total--;
+ return 0;
+ }
+ }
+ return -ENOENT;
+ case NATGXPORT:
+ if (nat_hdr.len != sizeof(NAT_XPORT_ALL_T))
+ return -EPERM;
+ if (copy_to_user(req_datap, (void *)&nat_cfg.xport, sizeof(NAT_XPORT_ALL_T)))
+ return -EFAULT;
+ break;
+ case NATSWEIGHT:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_WEIGHT_T))
+ return -EPERM;
+ if (copy_from_user((void *)&nat_cfg.weight, req_datap, sizeof(NAT_WEIGHT_T)))
+ return -EFAULT;
+ mac_set_hw_tx_weight(dev, (char *)&nat_cfg.weight);
+ break;
+ case NATGWEIGHT:
+ if (nat_hdr.len != sizeof(NAT_WEIGHT_T))
+ return -EPERM;
+ mac_get_hw_tx_weight(dev, (char *)&nat_cfg.weight);
+ if (copy_to_user(req_datap, (void *)&nat_cfg.weight, sizeof(NAT_WEIGHT_T)))
+ return -EFAULT;
+ break;
+ case NATAWRULE:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_WRULE_T))
+ return -EPERM;
+ if (copy_from_user((void *)&qid, req_datap, sizeof(qid)))
+ return -EFAULT;
+ if (qid > CONFIG_NAT_TXQ_NUM)
+ return -EPERM;
+ i = nat_cfg.wrule[qid].total;
+ if (i >= CONFIG_NAT_MAX_WRULE)
+ return -E2BIG;
+ if (copy_from_user((void *)&nat_cfg.wrule[qid].entry[i], req_datap+sizeof(qid), sizeof(NAT_WRULE_T)))
+ return -EFAULT;
+ nat_cfg.wrule[qid].total++;
+ break;
+ case NATDWRULE:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_WRULE_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.wrule, req_datap, sizeof(NAT_WRULE_T)))
+ return -EFAULT;
+ qid = ctrl.wrule.qid;
+ if (qid >= CONFIG_NAT_TXQ_NUM)
+ return -EPERM;
+ wrule_entry = (NAT_WRULE_ENTRY_T *)&nat_cfg.wrule[qid].entry[0];
+ for (i=0; i<nat_cfg.wrule[qid].total; i++, wrule_entry++)
+ {
+ if (memcmp((void *)wrule_entry, (void *)&ctrl.wrule.entry, sizeof(NAT_WRULE_ENTRY_T)) == 0)
+ {
+ NAT_WRULE_ENTRY_T *wrule_next;
+ wrule_next = wrule_entry + 1;
+ for (j=i+1; j < nat_cfg.wrule[qid].total; i++, j++)
+ {
+ memcpy((void *)wrule_entry, (void *)wrule_next, sizeof(NAT_WRULE_ENTRY_T));
+ wrule_entry++;
+ wrule_next++;
+ }
+ memset((void *)wrule_entry, 0, sizeof(NAT_WRULE_ENTRY_T));
+ nat_cfg.wrule[qid].total--;
+ return 0;
+ }
+ }
+ return -ENOENT;
+ case NATGWRULE:
+ if (nat_hdr.len != sizeof(NAT_WRULE_ALL_T))
+ return -EPERM;
+ if (copy_from_user((void *)&qid, req_datap, sizeof(qid)))
+ return -EFAULT;
+ if (qid >= CONFIG_NAT_TXQ_NUM)
+ return -EPERM;
+ if (copy_to_user(req_datap, (void *)&nat_cfg.wrule[qid], sizeof(NAT_WRULE_ALL_T)))
+ return -EFAULT;
+ break;
+ case NATSDEFQ:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_QUEUE_T))
+ return -EPERM;
+ if (copy_from_user((void *)&nat_cfg.default_hw_txq, req_datap, sizeof(u32)))
+ return -EFAULT;
+ break;
+ case NATGDEFQ:
+ if (nat_hdr.len != sizeof(NAT_QUEUE_T))
+ return -EPERM;
+ if (copy_to_user(req_datap, (void *)&nat_cfg.default_hw_txq, sizeof(u32)))
+ return -EFAULT;
+ case NATRMIPCFG:
+ nat_cfg.ipcfg[port_id].total = 0;
+ break;
+ case NATTESTENTRY:
+ if (!capable(CAP_NET_ADMIN))
+ return -EPERM;
+ if (nat_hdr.len != sizeof(NAT_TESTENTRY_T))
+ return -EPERM;
+ if (copy_from_user((void *)&ctrl.init_entry, req_datap, sizeof(ctrl.init_entry)))
+ return -EFAULT;
+ if (ctrl.init_entry.init_enable != 0 && ctrl.init_entry.init_enable != 1)
+ return -EPERM;
+ nat_cfg.init_enabled = ctrl.init_entry.init_enable;
+ break;
+
+ default:
+ return -EPERM;
+ }
+
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* nat_init_test_entry
+* Initialize NAT test hash entries
+*
+* SmartBits P1 -----> Lepus GMAC 0 --------------+
+* |
+* |
+* P3 <----- Lepus GMAC 1 -- HW TxQ0 <--+
+* -- HW TxQ1 <--+
+* -- HW TxQ2 <--+
+* -- HW TxQ3 <--+
+*
+* SmartBits P1 <----- Lepus GMAC 0 -- HW TxQ0 <--+
+* -- HW TxQ1 <--+
+* -- HW TxQ2 <--+
+* -- HW TxQ3 <--+
+* |
+* |
+* P3 -----> Lepus GMAC 1 --------------+
+*
+* LAN GMAC0 <--------------------------------------------> GMAC1 WAN
+* 192.168.[x].[y]:50 --> 168.95.[x].[y]:80 ---TXQ[y-1]---> 192.168.2.254:200[y] --> 168.95.[x].[y]:80
+* 192.168.[x].[y]:50 <-- 168.95.[x].[y]:80 <--TXQ[y-1]---- 192.168.2.254:200[y] <-- 168.95.[x].[y]:80
+* where:
+* [x] : Packet Type
+* [y] : Tx Queue, 1 for TxQ0, 2 for TxQ1, 3 for TxQ2, 4 for TxQ3,
+*
+*
+* Packet Type:
+* 1. TCP Frames <---> TCP Frames
+* LAN GMAC0 <--------------------------------> GMAC1 WAN
+* 192.168.1.1:50 --> 168.95.1.1:80 ---TXQ0---> 192.168.2.254:2001 --> 168.95.1.1:80
+* 192.168.1.1:50 <-- 168.95.1.1:80 <--TXQ0---- 192.168.2.254:2001 <-- 168.95.1.1:80
+*
+* 192.168.1.2:50 --> 168.95.1.2:80 ---TXQ1---> 192.168.2.254:2002 --> 168.95.1.2:80
+* 192.168.1.2:50 <-- 168.95.1.2:80 <--TXQ1---- 192.168.2.254:2002 <-- 168.95.1.2:80
+*
+* 192.168.1.3:50 --> 168.95.1.3:80 ---TXQ2---> 192.168.2.254:2003 --> 168.95.1.3:80
+* 192.168.1.3:50 <-- 168.95.1.3:80 <--TXQ2---- 192.168.2.254:2003 <-- 168.95.1.3:80
+*
+* 192.168.1.4:50 --> 168.95.1.4:80 ---TXQ3---> 192.168.2.254:2004 --> 168.95.1.4:80
+* 192.168.1.4:50 <-- 168.95.1.4:80 <--TXQ3---- 192.168.2.254:2004 <-- 168.95.1.4:80
+*
+* 2 TCP Frames <----> PPPoE + TCP Frames
+* LAN GMAC0 <--------------------------------> GMAC1 WAN
+* 192.168.2.1:50 --> 168.95.2.1:80 ---TXQ0---> 192.168.2.254:2001 --> 168.95.2.1:80
+* 192.168.2.1:50 <-- 168.95.2.1:80 <--TXQ0---- 192.168.2.254:2001 <-- 168.95.2.1:80
+*
+* 192.168.2.2:50 --> 168.95.2.2:80 ---TXQ1---> 192.168.2.254:2002 --> 168.95.2.2:80
+* 192.168.2.2:50 <-- 168.95.2.2:80 <--TXQ1---- 192.168.2.254:2002 <-- 168.95.2.2:80
+*
+* 192.168.2.3:50 --> 168.95.2.3:80 ---TXQ2---> 192.168.2.254:2003 --> 168.95.2.3:80
+* 192.168.2.3:50 <-- 168.95.2.3:80 <--TXQ2---- 192.168.2.254:2003 <-- 168.95.2.3:80
+*
+* 192.168.2.4:50 --> 168.95.2.4:80 ---TXQ3---> 192.168.2.254:2004 --> 168.95.2.4:80
+* 192.168.2.4:50 <-- 168.95.2.4:80 <--TXQ3---- 192.168.2.254:2004 <-- 168.95.2.4:80
+*
+* 3 TCP Frames <----> VLAN + PPPoE + TCP Frames
+* LAN GMAC0 <--------------------------------> GMAC1 WAN
+* 192.168.3.1:50 --> 168.95.3.1:80 ---TXQ0---> 192.168.2.254:2001 --> 168.95.3.1:80
+* 192.168.3.1:50 <-- 168.95.3.1:80 <--TXQ0---- 192.168.2.254:2001 <-- 168.95.3.1:80
+*
+* 192.168.3.2:50 --> 168.95.3.2:80 ---TXQ1---> 192.168.2.254:2002 --> 168.95.3.2:80
+* 192.168.3.2:50 <-- 168.95.3.2:80 <--TXQ1---- 192.168.2.254:2002 <-- 168.95.3.2:80
+*
+* 192.168.3.3:50 --> 168.95.3.3:80 ---TXQ2---> 192.168.2.254:2003 --> 168.95.3.3:80
+* 192.168.3.3:50 <-- 168.95.3.3:80 <--TXQ2---- 192.168.2.254:2003 <-- 168.95.3.3:80
+*
+* 192.168.3.4:50 --> 168.95.3.4:80 ---TXQ3---> 192.168.2.254:2004 --> 168.95.3.4:80
+* 192.168.3.4:50 <-- 168.95.3.4:80 <--TXQ3---- 192.168.2.254:2004 <-- 168.95.3.4:80
+*
+* 4 VLAN-A + TCP Frames <----> VLAN-B + PPPoE + TCP Frames
+* LAN GMAC0 <--------------------------------> GMAC1 WAN
+* 192.168.4.1:50 --> 168.95.4.1:80 ---TXQ0---> 192.168.2.254:2001 --> 168.95.4.1:80
+* 192.168.4.1:50 <-- 168.95.4.1:80 <--TXQ0---- 192.168.2.254:2001 <-- 168.95.4.1:80
+*
+* 192.168.4.2:50 --> 168.95.4.2:80 ---TXQ1---> 192.168.2.254:2002 --> 168.95.4.2:80
+* 192.168.4.2:50 <-- 168.95.4.2:80 <--TXQ1---- 192.168.2.254:2002 <-- 168.95.4.2:80
+*
+* 192.168.4.3:50 --> 168.95.4.3:80 ---TXQ2---> 192.168.2.254:2003 --> 168.95.4.3:80
+* 192.168.4.3:50 <-- 168.95.4.3:80 <--TXQ2---- 192.168.2.254:2003 <-- 168.95.4.3:80
+*
+* 192.168.4.4:50 --> 168.95.4.4:80 ---TXQ3---> 192.168.2.254:2004 --> 168.95.4.4:80
+* 192.168.4.4:50 <-- 168.95.4.4:80 <--TXQ3---- 192.168.2.254:2004 <-- 168.95.4.4:80
+*
+*
+*
+*----------------------------------------------------------------------*/
+#ifdef SL351x_NAT_TEST_BY_SMARTBITS
+#define NAT_IPIV(a,b,c,d) ((a<<24)+(b<<16)+(c<<8)+d)
+#define NAT_TEST_CLIENT_IP NAT_IPIV(192,168,1,1)
+#define NAT_TEST_SERVER_IP NAT_IPIV(168,95,1,1)
+#define NAT_TEST_LAN_IP NAT_IPIV(192,168,1,254)
+#define NAT_TEST_WAN_IP NAT_IPIV(192,168,2,254)
+#define NAT_TEST_MAP_PORT_BASE 2001
+#define NAT_TEST_SPORT 50
+#define NAT_TEST_DPORT 80
+#define NAT_TEST_PROTOCOL 6
+u8 nat_test_lan_target_da[6]={0x00,0x11,0x22,0x33,0x44,0x55};
+u8 nat_test_wan_target_da[6]={0x00,0xaa,0xbb,0xcc,0xdd,0xee};
+u8 nat_test_lan_my_da[6]={0x00,0x11,0x11,0x11,0x11,0x11};
+u8 nat_test_wan_my_da[6]={0x00,0x22,0x22,0x22,0x22,0x22};
+static void nat_init_test_entry(void)
+{
+ int i, j ;
+ NAT_HASH_ENTRY_T *hash_entry;
+ u32 sip, dip;
+ u32 hash_data[HASH_MAX_DWORDS];
+ NAT_CFG_T *cfg;
+ int hash_index;
+
+ cfg = (NAT_CFG_T *)&nat_cfg;
+ hash_entry = (NAT_HASH_ENTRY_T *)&hash_data;
+ hash_entry->key.Ethertype = 0;
+ hash_entry->key.rule_id = 0;
+ hash_entry->key.ip_protocol = IPPROTO_TCP;
+ hash_entry->key.reserved1 = 0;
+ hash_entry->key.reserved2 = 0;
+ // hash_entry->key.sip = NAT_TEST_CLIENT_IP;
+ // hash_entry->key.dip = NAT_TEST_SERVER_IP;
+ hash_entry->key.sport = htons(NAT_TEST_SPORT);
+ hash_entry->key.dport = htons(NAT_TEST_DPORT);
+ hash_entry->key.rule_id = cfg->tcp_udp_rule_id;
+ hash_entry->action.dword = NAT_LAN2WAN_ACTIONS;
+
+ sip = NAT_TEST_CLIENT_IP;
+ dip = NAT_TEST_SERVER_IP;
+
+ // Init TCP <------> TCP hash entries
+ // LAN --> WAN
+ // (1) TCP --> TCP
+ // (2) TCP --> PPPoE + TCP
+ // (3) TCP --> VLAN-B + PPPoE + TCP
+ // (4) TCP + VLAN-A --> VLAN-B + PPPoE + TCP
+ memcpy(hash_entry->param.da, nat_test_wan_target_da, 6);
+ memcpy(hash_entry->param.sa, nat_test_wan_my_da, 6);
+ hash_entry->key.port_id = cfg->lan_port;
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ if (i < 2)
+ {
+ hash_entry->action.bits.dest_qid = i+2;
+ }
+ else
+ {
+ hash_entry->action.bits.dest_qid = i;
+ }
+ hash_entry->action.bits.dest_qid += (cfg->wan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;
+ hash_entry->param.Sport = NAT_TEST_MAP_PORT_BASE+i;
+ hash_entry->param.Dport = NAT_TEST_DPORT;
+ for (j=0; j<4; j++)
+ {
+ hash_entry->key.sip = sip + i + j*0x100;
+ hash_entry->key.dip = dip + i + j*0x100;
+ hash_entry->param.Dip = hash_entry->key.dip;
+ hash_entry->param.Sip = NAT_TEST_WAN_IP;
+ switch (j)
+ {
+ case 0:
+ hash_entry->action.bits.pppoe = 0;
+ hash_entry->param.pppoe = 0;
+ hash_entry->action.bits.vlan = 0;
+ hash_entry->param.vlan = 0;
+ break;
+ case 1:
+ hash_entry->action.bits.pppoe = 1;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 0;
+ hash_entry->param.vlan = 0;
+ break;
+ case 2:
+ hash_entry->action.bits.pppoe = 1;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 1;
+ hash_entry->param.vlan = i+10;
+ break;
+ case 3:
+ hash_entry->action.bits.pppoe = 1;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 1;
+ hash_entry->param.vlan = i+10;
+ break;
+ }
+ hash_entry->tmo.counter = hash_entry->tmo.interval = 0x7fff;
+ hash_index = nat_build_keys(&hash_entry->key);
+ nat_write_hash_entry(hash_index, hash_entry);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast than SW
+ }
+ }
+
+
+ // WAN --> LAN
+ hash_entry->key.port_id = cfg->wan_port;
+ hash_entry->key.sport = htons(NAT_TEST_DPORT);
+ hash_entry->key.dport = htons(NAT_TEST_DPORT);
+ hash_entry->key.rule_id = cfg->tcp_udp_rule_id;
+ hash_entry->action.dword = NAT_WAN2LAN_ACTIONS;
+ hash_entry->key.sport = htons(NAT_TEST_DPORT);
+ memcpy(hash_entry->param.da, nat_test_lan_target_da, 6);
+ memcpy(hash_entry->param.sa, nat_test_lan_my_da, 6);
+ for (i=0; i<TOE_HW_TXQ_NUM; i++)
+ {
+ hash_entry->key.dport = htons(NAT_TEST_MAP_PORT_BASE + i);
+ if (i < 2)
+ {
+ hash_entry->action.bits.dest_qid = i+2;
+ }
+ else
+ {
+ hash_entry->action.bits.dest_qid = i;
+ }
+ hash_entry->action.bits.dest_qid += (cfg->lan_port==0) ? TOE_GMAC0_HW_TXQ0_QID : TOE_GMAC1_HW_TXQ0_QID;
+ hash_entry->param.Dport = NAT_TEST_SPORT;
+ hash_entry->param.Sport = NAT_TEST_DPORT;
+ hash_entry->param.da[5] = i;
+ for (j=0; j<4; j++)
+ {
+ hash_entry->key.sip = (dip + i + j*0x100);
+ hash_entry->key.dip = (NAT_TEST_WAN_IP);
+ hash_entry->param.Sip = hash_entry->key.sip;
+ hash_entry->param.Dip = sip + i + j*0x100;
+ switch (j)
+ {
+ case 0:
+ hash_entry->action.bits.pppoe = 0;
+ hash_entry->param.pppoe = 0;
+ hash_entry->action.bits.vlan = 0;
+ hash_entry->param.vlan = 0;
+ break;
+ case 1:
+ hash_entry->action.bits.pppoe = 2;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 0;
+ hash_entry->param.vlan = 0;
+ break;
+ case 2:
+ hash_entry->action.bits.pppoe = 2;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 2;
+ hash_entry->param.vlan = i+5;
+ break;
+ case 3:
+ hash_entry->action.bits.pppoe = 1;
+ hash_entry->param.pppoe = i+1;
+ hash_entry->action.bits.vlan = 1;
+ hash_entry->param.vlan = i+5;
+ break;
+ }
+ hash_entry->tmo.counter = hash_entry->tmo.interval = 0x7fff;
+ hash_index = nat_build_keys(&hash_entry->key);
+ nat_write_hash_entry(hash_index, hash_entry);
+ hash_nat_enable_owner(hash_index);
+ hash_validate_entry(hash_index); // Must last one, else HW Tx fast than SW
+ }
+ }
+}
+#endif // SL351x_NAT_TEST_BY_SMARTBITS
+
+#endif // CONFIG_SL351x_NAT
+
--- /dev/null
+++ b/drivers/net/sl351x_proc.c
@@ -0,0 +1,578 @@
+/****************************************************************************
+* Copyright 2006 Storlink Corp. All rights reserved.
+*----------------------------------------------------------------------------
+* Name : sl351x_proc.c
+* Description :
+* Handle Proc Routines for Storlink SL351x Platform
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* 04/13/2006 Gary Chen Create and implement
+*
+*
+****************************************************************************/
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch/irqs.h>
+#include <asm/arch/it8712.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/skbuff.h>
+#include <linux/if_ether.h>
+#include <linux/if_pppox.h>
+#include <linux/in.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#include <linux/ppp_defs.h>
+#ifdef CONFIG_NETFILTER
+// #include <linux/netfilter_ipv4/ip_conntrack.h>
+#endif
+#include <linux/proc_fs.h>
+#include <linux/seq_file.h>
+#include <linux/percpu.h>
+#ifdef CONFIG_SYSCTL
+#include <linux/sysctl.h>
+#endif
+
+#define MIDWAY
+#define SL_LEPUS
+
+// #define PROC_DEBUG_MSG 1
+
+#include <asm/arch/sl2312.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <asm/arch/sl351x_hash_cfg.h>
+#include <asm/arch/sl351x_nat_cfg.h>
+#include <asm/arch/sl351x_toe.h>
+
+#ifdef CONFIG_PROC_FS
+/*----------------------------------------------------------------------
+* Definition
+*----------------------------------------------------------------------*/
+#define proc_printf printk
+#define SL351x_GMAC_PROC_NAME "sl351x_gmac"
+#define SL351x_NAT_PROC_NAME "sl351x_nat"
+#define SL351x_TOE_PROC_NAME "sl351x_toe"
+
+/*----------------------------------------------------------------------
+* Function Definition
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static int nat_ct_open(struct inode *inode, struct file *file);
+static void *nat_ct_seq_start(struct seq_file *s, loff_t *pos);
+static void nat_ct_seq_stop(struct seq_file *s, void *v);
+static void *nat_ct_seq_next(struct seq_file *s, void *v, loff_t *pos);
+static int nat_ct_seq_show(struct seq_file *s, void *v);
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+static int toe_ct_open(struct inode *inode, struct file *file);
+static void *toe_ct_seq_start(struct seq_file *s, loff_t *pos);
+static void toe_ct_seq_stop(struct seq_file *s, void *v);
+static void *toe_ct_seq_next(struct seq_file *s, void *v, loff_t *pos);
+static int toe_ct_seq_show(struct seq_file *s, void *v);
+extern int sl351x_get_toe_conn_flag(int index);
+extern struct toe_conn * sl351x_get_toe_conn_info(int index);
+#endif
+
+static int gmac_ct_open(struct inode *inode, struct file *file);
+static void *gmac_ct_seq_start(struct seq_file *s, loff_t *pos);
+static void gmac_ct_seq_stop(struct seq_file *s, void *v);
+static void *gmac_ct_seq_next(struct seq_file *s, void *v, loff_t *pos);
+static int gmac_ct_seq_show(struct seq_file *s, void *v);
+
+
+/*----------------------------------------------------------------------
+* Data
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SYSCTL
+// static struct ctl_table_header *nat_ct_sysctl_header;
+#endif
+
+#ifdef CONFIG_SL351x_NAT
+static struct seq_operations nat_ct_seq_ops = {
+ .start = nat_ct_seq_start,
+ .next = nat_ct_seq_next,
+ .stop = nat_ct_seq_stop,
+ .show = nat_ct_seq_show
+};
+
+static struct file_operations nat_file_ops= {
+ .owner = THIS_MODULE,
+ .open = nat_ct_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = seq_release
+};
+#endif // CONFIG_SL351x_NAT
+
+#ifdef CONFIG_SL351x_RXTOE
+static struct seq_operations toe_ct_seq_ops = {
+ .start = toe_ct_seq_start,
+ .next = toe_ct_seq_next,
+ .stop = toe_ct_seq_stop,
+ .show = toe_ct_seq_show
+};
+
+static struct file_operations toe_file_ops= {
+ .owner = THIS_MODULE,
+ .open = toe_ct_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = seq_release
+};
+#endif
+
+static struct seq_operations gmac_ct_seq_ops = {
+ .start = gmac_ct_seq_start,
+ .next = gmac_ct_seq_next,
+ .stop = gmac_ct_seq_stop,
+ .show = gmac_ct_seq_show
+};
+
+static struct file_operations gmac_file_ops= {
+ .owner = THIS_MODULE,
+ .open = gmac_ct_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = seq_release
+};
+
+#ifdef SL351x_GMAC_WORKAROUND
+extern u32 gmac_workaround_cnt[4];
+extern u32 gmac_short_frame_workaround_cnt[2];
+#ifdef CONFIG_SL351x_NAT
+ extern u32 sl351x_nat_workaround_cnt;
+#endif
+#endif
+/*----------------------------------------------------------------------
+* nat_ct_open
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static int nat_ct_open(struct inode *inode, struct file *file)
+{
+ return seq_open(file, &nat_ct_seq_ops);
+}
+#endif // CONFIG_SL351x_NAT
+/*----------------------------------------------------------------------
+* nat_ct_seq_start
+* find the first
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static void *nat_ct_seq_start(struct seq_file *s, loff_t *pos)
+{
+ int i;
+
+ // proc_printf("%s: *pos=%d\n", __func__, (int)*pos);
+ for (i=*pos; i<HASH_TOTAL_ENTRIES; i++)
+ {
+ if (hash_get_nat_owner_flag(i))
+ {
+ *pos = i;
+ return (void *)(i+1);
+ }
+ }
+ return NULL;
+}
+#endif // CONFIG_SL351x_NAT
+/*----------------------------------------------------------------------
+* nat_ct_seq_stop
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static void nat_ct_seq_stop(struct seq_file *s, void *v)
+{
+}
+#endif // CONFIG_SL351x_NAT
+/*----------------------------------------------------------------------
+* nat_ct_seq_next
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static void *nat_ct_seq_next(struct seq_file *s, void *v, loff_t *pos)
+{
+ int i;
+
+ // proc_printf("%s: *pos=%d\n", __func__, (int)*pos);
+ (*pos)++;
+ for (i=*pos; i<HASH_TOTAL_ENTRIES; i++)
+ {
+ if (hash_get_nat_owner_flag(i))
+ {
+ *pos = i;
+ return (void *)(i+1);
+ }
+ }
+ return NULL;
+}
+#endif // CONFIG_SL351x_NAT
+/*----------------------------------------------------------------------
+* nat_ct_seq_show
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_NAT
+static int nat_ct_seq_show(struct seq_file *s, void *v)
+{
+ int idx;
+ NAT_HASH_ENTRY_T *nat_entry;
+ GRE_HASH_ENTRY_T *gre_entry;
+
+ idx = (int)v;
+ if (idx<=0 || idx >HASH_TOTAL_ENTRIES)
+ return -ENOSPC;
+
+ idx--;
+ nat_entry = (NAT_HASH_ENTRY_T *)&hash_tables[idx];
+ gre_entry = (GRE_HASH_ENTRY_T *)nat_entry;
+ if (nat_entry->key.ip_protocol == IPPROTO_GRE)
+ {
+ if (seq_printf(s, "%4d: KEY MAC-%d [%d] %u.%u.%u.%u [%u]-->%u.%u.%u.%u\n",
+ idx, gre_entry->key.port_id, gre_entry->key.ip_protocol,
+ HIPQUAD(gre_entry->key.sip), ntohs(gre_entry->key.call_id),
+ HIPQUAD(gre_entry->key.dip)))
+ return -ENOSPC;
+ if (seq_printf(s, " PARAMETER: %u.%u.%u.%u -->%u.%u.%u.%u [%u] Timeout:%ds\n",
+ HIPQUAD(gre_entry->param.Sip),
+ HIPQUAD(gre_entry->param.Dip), gre_entry->param.Dport,
+ gre_entry->tmo.counter))
+ return -ENOSPC;
+ }
+ else
+ {
+ if (seq_printf(s, "%4d: KEY MAC-%d [%d] %u.%u.%u.%u [%u]-->%u.%u.%u.%u [%u]\n",
+ idx, nat_entry->key.port_id, nat_entry->key.ip_protocol,
+ HIPQUAD(nat_entry->key.sip), ntohs(nat_entry->key.sport),
+ HIPQUAD(nat_entry->key.dip), ntohs(nat_entry->key.dport)))
+ return -ENOSPC;
+ if (seq_printf(s, " PARAMETER: %u.%u.%u.%u [%u]-->%u.%u.%u.%u [%u] Timeout:%ds\n",
+ HIPQUAD(nat_entry->param.Sip), nat_entry->param.Sport,
+ HIPQUAD(nat_entry->param.Dip), nat_entry->param.Dport,
+ nat_entry->tmo.counter))
+ return -ENOSPC;
+ }
+ return 0;
+}
+#endif // CONFIG_SL351x_NAT
+
+/*----------------------------------------------------------------------
+* toe_ct_open
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static int toe_ct_open(struct inode *inode, struct file *file)
+{
+ return seq_open(file, &toe_ct_seq_ops);
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_ct_seq_start
+* find the first
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static void *toe_ct_seq_start(struct seq_file *s, loff_t *pos)
+{
+ int i;
+
+ // proc_printf("%s: *pos=%d\n", __func__, (int)*pos);
+ for (i=*pos; i<TOE_TOE_QUEUE_NUM; i++)
+ {
+ if (sl351x_get_toe_conn_flag(i))
+ {
+ *pos = i;
+ return (void *)(i+1);
+ }
+ }
+ return NULL;
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_ct_seq_stop
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static void toe_ct_seq_stop(struct seq_file *s, void *v)
+{
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_ct_seq_next
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static void *toe_ct_seq_next(struct seq_file *s, void *v, loff_t *pos)
+{
+ int i;
+
+ // proc_printf("%s: *pos=%d\n", __func__, (int)*pos);
+ (*pos)++;
+ for (i=*pos; i<TOE_TOE_QUEUE_NUM; i++)
+ {
+ if (sl351x_get_toe_conn_flag(i))
+ {
+ *pos = i;
+ return (void *)(i+1);
+ }
+ }
+ return NULL;
+}
+#endif
+/*----------------------------------------------------------------------
+* toe_ct_seq_show
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_SL351x_RXTOE
+static int toe_ct_seq_show(struct seq_file *s, void *v)
+{
+ int idx;
+ struct toe_conn *toe_entry;
+
+ idx = (int)v;
+ if (idx<=0 || idx >TOE_TOE_QUEUE_NUM)
+ return -ENOSPC;
+
+ idx--;
+ toe_entry = (struct toe_conn *)sl351x_get_toe_conn_info(idx);
+ if (!toe_entry)
+ return -ENOSPC;
+
+ if (seq_printf(s, "%4d: Qid %d MAC-%d TCP %u.%u.%u.%u [%u]-->%u.%u.%u.%u [%u]\n",
+ idx, toe_entry->qid, toe_entry->gmac->port_id,
+ NIPQUAD(toe_entry->saddr[0]), ntohs(toe_entry->source),
+ NIPQUAD(toe_entry->daddr[0]), ntohs(toe_entry->dest)))
+ return -ENOSPC;
+ return 0;
+}
+#endif
+/*----------------------------------------------------------------------
+* gmac_ct_open
+*----------------------------------------------------------------------*/
+static int gmac_ct_open(struct inode *inode, struct file *file)
+{
+ return seq_open(file, &gmac_ct_seq_ops);
+}
+
+/*----------------------------------------------------------------------
+* gmac_ct_seq_start
+* find the first
+*----------------------------------------------------------------------*/
+static void *gmac_ct_seq_start(struct seq_file *s, loff_t *pos)
+{
+ int i;
+ i = (int)*pos + 1;;
+
+ if (i > 9)
+ return NULL;
+ else
+ return (void *)i;
+}
+
+/*----------------------------------------------------------------------
+* gmac_ct_seq_stop
+*----------------------------------------------------------------------*/
+static void gmac_ct_seq_stop(struct seq_file *s, void *v)
+{
+}
+
+/*----------------------------------------------------------------------
+* gmac_ct_seq_next
+*----------------------------------------------------------------------*/
+static void *gmac_ct_seq_next(struct seq_file *s, void *v, loff_t *pos)
+{
+ int i;
+
+ // proc_printf("%s: *pos=%d\n", __func__, (int)*pos);
+
+ (*pos)++;
+ i = (int)*pos + 1;;
+
+ if (i > 9)
+ return NULL;
+ else
+ return (void *)i;
+}
+
+/*----------------------------------------------------------------------
+* seq_dm_long
+*----------------------------------------------------------------------*/
+static void seq_dm_long(struct seq_file *s, u32 location, int length)
+{
+ u32 *start_p, *curr_p, *end_p;
+ u32 *datap, data;
+ int i;
+
+ //if (length > 1024)
+ // length = 1024;
+
+ start_p = (u32 *)location;
+ end_p = (u32 *)location + length;
+ curr_p = (u32 *)((u32)location & 0xfffffff0);
+ datap = (u32 *)location;
+ while (curr_p < end_p)
+ {
+ cond_resched();
+ seq_printf(s, "0x%08x: ",(u32)curr_p & 0xfffffff0);
+ for (i=0; i<4; i++)
+ {
+ if (curr_p < start_p || curr_p >= end_p)
+ seq_printf(s, " ");
+ else
+ {
+ data = *datap;
+ seq_printf(s, "%08X ", data);
+ }
+ if (i==1)
+ seq_printf(s, "- ");
+
+ curr_p++;
+ datap++;
+ }
+ seq_printf(s, "\n");
+ }
+}
+
+/*----------------------------------------------------------------------
+* gmac_ct_seq_show
+*----------------------------------------------------------------------*/
+static int gmac_ct_seq_show(struct seq_file *s, void *v)
+{
+ switch ((int)v)
+ {
+ case 1:
+ seq_printf(s, "\nGMAC Global Registers\n");
+ seq_dm_long(s, TOE_GLOBAL_BASE, 32);
+ break;
+ case 2:
+ seq_printf(s, "\nGMAC Non-TOE Queue Header\n");
+ seq_dm_long(s, TOE_NONTOE_QUE_HDR_BASE, 12);
+ break;
+ case 3:
+ seq_printf(s, "\nGMAC TOE Queue Header\n");
+ seq_dm_long(s, TOE_TOE_QUE_HDR_BASE, 12);
+ break;
+ case 4:
+ seq_printf(s, "\nGMAC-0 DMA Registers\n");
+ seq_dm_long(s, TOE_GMAC0_DMA_BASE, 52);
+ break;
+ case 5:
+ seq_printf(s, "\nGMAC-0 Registers\n");
+ seq_dm_long(s, TOE_GMAC0_BASE, 32);
+ break;
+ case 6:
+ seq_printf(s, "\nGMAC-1 DMA Registers\n");
+ seq_dm_long(s, TOE_GMAC1_DMA_BASE, 52);
+ break;
+ case 7:
+ seq_printf(s, "\nGMAC-1 Registers\n");
+ seq_dm_long(s, TOE_GMAC1_BASE, 32);
+ break;
+ case 8:
+ seq_printf(s, "\nGLOBAL Registers\n");
+ seq_dm_long(s, GMAC_GLOBAL_BASE_ADDR, 16);
+ break;
+ case 9:
+#ifdef SL351x_GMAC_WORKAROUND
+ seq_printf(s, "\nGMAC-0 Rx/Tx/Short Workaround: %u, %u, %u\n", gmac_workaround_cnt[0], gmac_workaround_cnt[1], gmac_short_frame_workaround_cnt[0]);
+ seq_printf(s, "GMAC-1 Rx/Tx/Short Workaround: %u, %u, %u\n", gmac_workaround_cnt[2], gmac_workaround_cnt[3], gmac_short_frame_workaround_cnt[1]);
+#ifdef CONFIG_SL351x_NAT
+ seq_printf(s, "NAT Workaround: %u\n", sl351x_nat_workaround_cnt);
+#endif
+#endif
+ break;
+ default:
+ return -ENOSPC;
+ }
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* init
+*----------------------------------------------------------------------*/
+static int __init init(void)
+{
+ struct proc_dir_entry *proc_gmac=NULL;
+
+#ifdef CONFIG_SL351x_NAT
+ struct proc_dir_entry *proc_nat=NULL;
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+ struct proc_dir_entry *proc_toe=NULL;
+#endif
+
+#ifdef CONFIG_SYSCTL
+ // nat_ct_sysctl_header = NULL;
+#endif
+ proc_gmac = proc_net_fops_create(SL351x_GMAC_PROC_NAME, 0440, &gmac_file_ops);
+ if (!proc_gmac) goto init_bad;
+
+#ifdef CONFIG_SL351x_NAT
+ proc_nat = proc_net_fops_create(SL351x_NAT_PROC_NAME, 0440, &nat_file_ops);
+ if (!proc_nat) goto init_bad;
+#endif // CONFIG_SL351x_NAT
+
+#ifdef CONFIG_SL351x_RXTOE
+ proc_toe = proc_net_fops_create(SL351x_TOE_PROC_NAME, 0440, &toe_file_ops);
+ if (!proc_toe) goto init_bad;
+#endif
+
+#ifdef CONFIG_SYSCTL
+ // nat_ct_sysctl_header = register_sysctl_table(nat_ct_net_table, 0);
+ // if (!nat_ct_sysctl_header) goto init_bad;
+#endif
+
+ return 0;
+
+init_bad:
+ if (proc_gmac) proc_net_remove(SL351x_GMAC_PROC_NAME);
+
+#ifdef CONFIG_SL351x_NAT
+ if (proc_nat) proc_net_remove(SL351x_NAT_PROC_NAME);
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+ if (proc_toe) proc_net_remove(SL351x_NAT_PROC_NAME);
+#endif
+
+#ifdef CONFIG_SYSCTL
+ // if (nat_ct_sysctl_header) unregister_sysctl_table(nat_ct_sysctl_header);
+#endif
+ proc_printf("SL351x NAT Proc: can't create proc or register sysctl.\n");
+ return -ENOMEM;
+}
+
+/*----------------------------------------------------------------------
+* fini
+*----------------------------------------------------------------------*/
+static void __exit fini(void)
+{
+ proc_net_remove(SL351x_GMAC_PROC_NAME);
+
+#ifdef CONFIG_SL351x_NAT
+ proc_net_remove(SL351x_NAT_PROC_NAME);
+#endif
+
+#ifdef CONFIG_SL351x_RXTOE
+ proc_net_remove(SL351x_TOE_PROC_NAME);
+#endif
+
+#ifdef CONFIG_SYSCTL
+ // unregister_sysctl_table(nat_ct_sysctl_header);
+#endif
+}
+
+/*----------------------------------------------------------------------
+* module
+*----------------------------------------------------------------------*/
+module_init(init);
+module_exit(fini);
+
+#endif // CONFIG_PROC_FS
--- /dev/null
+++ b/drivers/net/sl351x_toe.c
@@ -0,0 +1,1083 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl351x_toe.c
+* Description :
+* Provide TOE routines for SL351x
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* Xiaochong
+*
+****************************************************************************/
+
+#include <linux/pci.h>
+#include <linux/ip.h>
+#include <linux/ipv6.h>
+#include <linux/tcp.h>
+#include <linux/slab.h>
+#include <linux/etherdevice.h>
+#include <asm/io.h>
+#include <linux/sysctl_storlink.h>
+#include <net/tcp.h>
+#include <linux/if_ether.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <asm/arch/sl351x_toe.h>
+#include <asm/arch/sl351x_hash_cfg.h>
+#include <asm/arch/sl351x_nat_cfg.h>
+
+static int in_toe_isr;
+static int toe_initialized=0;
+
+static struct toe_conn toe_connections[TOE_TOE_QUEUE_NUM];
+EXPORT_SYMBOL(toe_connections);
+static __u32 toe_connection_bits[TOE_TOE_QUEUE_NUM/32] __attribute__ ((aligned(16)));
+struct sk_buff* gen_pure_ack(struct toe_conn* connection, TOE_QHDR_T* toe_qhdr, INTR_QHDR_T *intr_curr_desc);
+
+extern struct storlink_sysctl storlink_ctl;
+extern TOE_INFO_T toe_private_data;
+extern spinlock_t gmac_fq_lock;
+extern void mac_write_dma_reg(int mac, unsigned int offset, u32 data);
+extern int mac_set_rule_reg(int mac, int rule, int enabled, u32 reg0, u32 reg1, u32 reg2);
+extern int hash_add_toe_entry(HASH_ENTRY_T *entry);
+extern void toe_gmac_fill_free_q(void);
+
+#define _DEBUG_SKB_ 1
+#ifdef _DEBUG_SKB_
+/*---------------------------------------------------------------------------
+ * _debug_skb
+ *-------------------------------------------------------------------------*/
+static inline void _debug_skb(struct sk_buff *skb, GMAC_RXDESC_T *toe_curr_desc, u32 data)
+{
+ if ((u32)skb < 0x1000)
+ {
+ printk("%s skb=%x\n", __func__, (u32)skb);
+ while(1);
+ }
+ REG32(__va(toe_curr_desc->word2.buf_adr)-SKB_RESERVE_BYTES) = data;
+}
+#else
+#define _debug_skb(x, y, z)
+#endif
+
+/*---------------------------------------------------------------------------
+ * get_connection_seq_num
+ *-------------------------------------------------------------------------*/
+u32 get_connection_seq_num(unsigned short qid)
+{
+ TOE_QHDR_T *toe_qhdr;
+
+ toe_qhdr = (TOE_QHDR_T*)TOE_TOE_QUE_HDR_BASE;
+ toe_qhdr += qid;
+ return (u32)toe_qhdr->word3.seq_num;
+}
+EXPORT_SYMBOL(get_connection_seq_num);
+
+/*---------------------------------------------------------------------------
+ * get_connection_ack_num
+ *-------------------------------------------------------------------------*/
+u32 get_connection_ack_num(unsigned short qid)
+{
+ TOE_QHDR_T *toe_qhdr;
+
+ toe_qhdr = (TOE_QHDR_T*)TOE_TOE_QUE_HDR_BASE;
+ toe_qhdr += qid;
+ return (u32)toe_qhdr->word4.ack_num;
+}
+EXPORT_SYMBOL(get_connection_ack_num);
+
+/*---------------------------------------------------------------------------
+ * dump_toe_qhdr
+ *-------------------------------------------------------------------------*/
+void dump_toe_qhdr(TOE_QHDR_T *toe_qhdr)
+{
+ printk("TOE w1 %x, w2 %x, w3 %x\n", toe_qhdr->word1.bits32,
+ toe_qhdr->word2.bits32, toe_qhdr->word3.bits32);
+ printk("w4 %x, w5 %x, w6 %x\n", toe_qhdr->word4.bits32,
+ toe_qhdr->word5.bits32, toe_qhdr->word6.bits32);
+}
+
+/*---------------------------------------------------------------------------
+ * dump_intrq_desc
+ *-------------------------------------------------------------------------*/
+void dump_intrq_desc(INTR_QHDR_T *intr_curr_desc)
+{
+ printk("INTR w0 %x, w1 %x, seq %x\n", intr_curr_desc->word0.bits32,
+ intr_curr_desc->word1.bits32, intr_curr_desc->word2.bits32);
+ printk("ack %x, w4 %x\n", intr_curr_desc->word3.bits32,
+ intr_curr_desc->word4.bits32);
+}
+
+/*---------------------------------------------------------------------------
+ * This routine will initialize a TOE matching rule
+ * called by SL351x GMAC driver.
+ *-------------------------------------------------------------------------*/
+void sl351x_toe_init(void)
+{
+ GMAC_MRxCR0_T mrxcr0;
+ GMAC_MRxCR1_T mrxcr1;
+ GMAC_MRxCR2_T mrxcr2;
+ int rule, rc;
+
+ if (toe_initialized)
+ return;
+
+ toe_initialized = 1;
+
+#ifndef CONFIG_SL351x_NAT
+ mrxcr0.bits32 = 0;
+ mrxcr1.bits32 = 0;
+ mrxcr2.bits32 = 0;
+ mrxcr0.bits.l3 = 1;
+ mrxcr0.bits.l4 = 1;
+ mrxcr1.bits.sip = 1;
+ mrxcr1.bits.dip = 1;
+ mrxcr1.bits.l4_byte0_15 = 0x0f;
+ mrxcr0.bits.sprx = 1;
+ rule = 0;
+ rc = mac_set_rule_reg(0, rule, 1, mrxcr0.bits32, mrxcr1.bits32,
+ mrxcr2.bits32);
+ if (rc<0) {
+ printk("%s::Set MAC 0 rule fail!\n", __func__);
+ }
+ rc = mac_set_rule_reg(1, rule, 1, mrxcr0.bits32, mrxcr1.bits32,
+ mrxcr2.bits32);
+ if (rc<0) {
+ printk("%s::Set MAC 1 rule fail!\n", __func__);
+ }
+#endif // CONFIG_SL351x_NAT
+}
+
+/*---------------------------------------------------------------------------
+ * dump_intrq_desc
+ * assign an interrupt queue number to a give tcp queue
+ *-------------------------------------------------------------------------*/
+int get_interrupt_queue_id(int tcp_qid)
+{
+ return (int)(tcp_qid & 0x0003);
+}
+
+/*---------------------------------------------------------------------------
+ * reset_connection_index
+ * reset the connection bit by given index
+ *-------------------------------------------------------------------------*/
+void reset_connection_index(__u8 index)
+{
+ __u32 mask = ~(0xffffffff & (1<< (index&0x1f)));
+ toe_connection_bits[index>>5] = toe_connection_bits[index>>5] & mask;
+}
+
+/*---------------------------------------------------------------------------
+ * update_timer
+ *-------------------------------------------------------------------------*/
+void update_timer(struct toe_conn* connection)
+{
+// if (time_before(jiffies, connection->last_rx_jiffies+3))
+// if ((jiffies + 0xffffffff - connection->last_rx_jiffies) & 0x3)
+// if (connection->last_rx_jiffies > jiffies)
+// printk("%s::jif %g, last_rx_jif %g\n", __func__, jiffies, connection->last_rx_jiffies);
+/* if ((long)(jiffies + 2)< 3) { // overflow...
+ printk("%s::jiffies %x\n", __func__, jiffies);
+ } */
+// if ((long)(jiffies - connection->last_rx_jiffies)< 2)
+// return;
+ connection->last_rx_jiffies = jiffies;
+ // gary chen mod_timer(&connection->rx_timer, jiffies+2);
+ connection->rx_timer.expires = jiffies + 2;
+ add_timer(&connection->rx_timer);
+// printk("%s::nt %x, lj %x\n", __func__, (jiffies+2), connection->last_rx_jiffies);
+}
+
+/*---------------------------------------------------------------------------
+ * gen_pure_ack
+ *-------------------------------------------------------------------------*/
+struct sk_buff* gen_pure_ack(struct toe_conn* connection, TOE_QHDR_T* toe_qhdr,
+INTR_QHDR_T *intr_curr_desc)
+{
+ struct sk_buff *skb;
+ struct iphdr *ip_hdr;
+ struct tcphdr *tcp_hdr;
+ struct ethhdr *eth_hdr;
+
+ if ((skb= dev_alloc_skb(RX_BUF_SIZE))==NULL) {
+ printk("%s::alloc pure ack fail!\n", __func__);
+ return NULL;
+ }
+ skb_reserve(skb, RX_INSERT_BYTES);
+ memset(skb->data, 0, 60);
+
+ eth_hdr = (struct ethhdr*)&(skb->data[0]);
+ memcpy(eth_hdr, &connection->l2_hdr, sizeof(struct ethhdr));
+
+ ip_hdr = (struct iphdr*)&(skb->data[14]);
+ ip_hdr->version = connection->ip_ver;
+ ip_hdr->ihl = 20>>2;
+ ip_hdr->tot_len = ntohs(40);
+ ip_hdr->frag_off = htons(IP_DF);
+ ip_hdr->ttl = 128;
+ ip_hdr->protocol = 0x06;
+ ip_hdr->saddr = connection->saddr[0];
+ ip_hdr->daddr = connection->daddr[0];
+// printk("%s ip sa %x, da %x\n",
+// __func__, ntohl(ip_hdr->saddr), ntohl(ip_hdr->daddr));
+
+ tcp_hdr = (struct tcphdr*)&(skb->data[34]);
+ tcp_hdr->source = connection->source;
+ tcp_hdr->dest = connection->dest;
+ if (intr_curr_desc) {
+ tcp_hdr->seq = htonl(intr_curr_desc->word2.seq_num);
+ tcp_hdr->ack_seq = htonl(intr_curr_desc->word3.ack_num);
+ tcp_hdr->window = htons(intr_curr_desc->word0.bits.win_size);
+ } else {
+ tcp_hdr->seq = htonl(toe_qhdr->word3.seq_num);
+ tcp_hdr->ack_seq = htonl(toe_qhdr->word4.ack_num);
+ tcp_hdr->window = htons(toe_qhdr->word6.bits.WinSize);
+ }
+ tcp_hdr->ack = 1;
+ tcp_hdr->doff = 20 >> 2;
+#if 0
+ if (!intr_curr_desc) {
+ unsigned char byte;
+ for (i=0; i<20; i++) {
+ byte = skb->data[34+i];
+ printk("%x ", byte);
+ }
+ printk("\n");
+ }
+#endif
+ TCP_SKB_CB(skb)->connection = connection;
+ return skb;
+}
+
+/*---------------------------------------------------------------------------
+ * connection_rx_timer
+ *-------------------------------------------------------------------------*/
+void connection_rx_timer(unsigned long *data)
+{
+ struct toe_conn *connection = (struct toe_conn*)data;
+ unsigned int tcp_qid, toeq_wptr;
+ unsigned int pkt_size, desc_count;
+ struct sk_buff *skb;
+ GMAC_RXDESC_T *toe_curr_desc;
+ TOE_QHDR_T *toe_qhdr;
+ struct net_device *dev;
+ unsigned long conn_flags;
+ DMA_RWPTR_T toeq_rwptr;
+ unsigned short timeout_descs;
+
+ if (in_toe_isr)
+ printk("%s::in_toe_isr=%d!\n", __func__, in_toe_isr);
+
+ if (connection) {
+ /* should we disable gmac interrupt first? */
+ if (!connection->gmac)
+ printk("%s::conn gmac %x!\n", __func__, (u32)connection->gmac);
+ local_irq_save(conn_flags);
+ if (!spin_trylock(&connection->conn_lock)) {
+ local_irq_restore(conn_flags);
+ // timer should be updated by the toeq isr. So no need to update here.
+ printk("%s::conn_lock is held by ISR!\n", __func__);
+ return;
+ }
+ disable_irq(connection->gmac->irq);
+
+ /* disable hash entry and get toeq desc. */
+ hash_set_valid_flag(connection->hash_entry_index, 0);
+ do{} while(0); /* wait until HW finish */
+
+ dev = connection->dev;
+ if (!dev)
+ printk("%s::conn dev NULL!\n", __func__);
+ tcp_qid = connection->qid;
+ toe_qhdr = (TOE_QHDR_T *)(TOE_TOE_QUE_HDR_BASE +
+ tcp_qid * sizeof(TOE_QHDR_T));
+ toeq_rwptr.bits32 = readl(&toe_qhdr->word1);
+ toeq_wptr = toe_qhdr->word1.bits.wptr;
+ timeout_descs = toeq_wptr - toeq_rwptr.bits.rptr;
+
+ if (toeq_rwptr.bits.rptr == toeq_wptr) {
+ if (toe_qhdr->word5.bits32) {
+ // shall we check toe_qhdr->word2.bits?
+ skb = gen_pure_ack(connection, toe_qhdr, (INTR_QHDR_T *)NULL);
+ skb_put(skb, 54);
+ skb->dev = connection->dev;
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+ skb->protocol = eth_type_trans(skb, connection->dev);
+ netif_rx(skb);
+ connection->dev->last_rx = jiffies;
+ }
+ } else {
+ while (toeq_rwptr.bits.rptr != toeq_rwptr.bits.wptr) {
+ /* we just simply send those packets to tcp? */
+ toe_curr_desc = (GMAC_RXDESC_T*)(toe_private_data.toe_desc_base[tcp_qid]
+ + toeq_rwptr.bits.rptr * sizeof(GMAC_RXDESC_T));
+ connection->curr_desc = toe_curr_desc;
+ if (toe_curr_desc->word3.bits.ctrl_flag) {
+ printk("%s::ctrl flag! %x, conn rptr %d, to %d, jif %x, conn_jif %x\n",
+ __func__, toe_curr_desc->word3.bits32,
+ connection->toeq_rwptr.bits.rptr, timeout_descs,
+ (u32)jiffies, (u32)connection->last_rx_jiffies);
+ }
+ desc_count = toe_curr_desc->word0.bits.desc_count;
+ pkt_size = toe_curr_desc->word1.bits.byte_count;
+ consistent_sync((void*)__va(toe_curr_desc->word2.buf_adr), pkt_size,
+ PCI_DMA_FROMDEVICE);
+ skb = (struct sk_buff*)(REG32(__va(toe_curr_desc->word2.buf_adr)-
+ SKB_RESERVE_BYTES));
+ _debug_skb(skb, (GMAC_RXDESC_T *)toe_curr_desc, 0x02);
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, RX_INSERT_BYTES);
+ skb_put(skb, pkt_size);
+ skb->dev = dev;
+ skb->protocol = eth_type_trans(skb, dev);
+ {
+ struct iphdr* ip_hdr = (struct iphdr*)&(skb->data[0]);
+ if (toe_curr_desc->word3.bits.ctrl_flag)
+ printk("%s::ip id %x\n", __func__, ntohs(ip_hdr->id));
+ }
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+
+ netif_rx(skb);
+ dev->last_rx = jiffies;
+#if 0
+ if ((skb = dev_alloc_skb(RX_BUF_SIZE)) == NULL) {
+ printk("%s::alloc buf fail!\n", __func__);
+ }
+ *(unsigned int*)(skb->data) = (unsigned int)skb;
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES);
+ spin_lock_irqsave(&connection->gmac->rx_mutex, flags);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ if (toe_private_data.fq_rx_rwptr.bits.wptr != fq_rwptr.bits.wptr) {
+ mac_stop_txdma((struct net_device*)connection->dev);
+ spin_unlock_irqrestore(&connection->gmac->rx_mutex, flags);
+ while(1);
+ }
+ fq_desc = (GMAC_RXDESC_T*)toe_private_data.swfq_desc_base + fq_rwptr.bits.wptr;
+ fq_desc->word2.buf_adr = (unsigned int)__pa(skb->data);
+ fq_rwptr.bits.wptr = RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr, TOE_SW_FREEQ_DESC_NUM);
+ SET_WPTR(TOE_GLOBAL_BASE+GLOBAL_SWFQ_RWPTR_REG, fq_rwptr.bits.wptr);
+ toe_private_data.fq_rx_rwptr.bits32 = fq_rwptr.bits32;
+ spin_unlock_irqrestore(&connection->gmac->rx_mutex, flags);
+#endif
+// spin_lock_irqsave(&connection->gmac->rx_mutex, flags);
+ toeq_rwptr.bits.rptr = RWPTR_ADVANCE_ONE(toeq_rwptr.bits.rptr, TOE_TOE_DESC_NUM);
+ SET_RPTR(&toe_qhdr->word1, toeq_rwptr.bits.rptr);
+// spin_unlock_irqrestore(&connection->gmac->rx_mutex, flags);
+ connection->toeq_rwptr.bits32 = toeq_rwptr.bits32;
+ }
+ toeq_rwptr.bits32 = readl(&toe_qhdr->word1);
+// toe_gmac_fill_free_q();
+ }
+ connection->last_rx_jiffies = jiffies;
+ if (connection->status != TCP_CONN_CLOSED)
+ mod_timer(&connection->rx_timer, jiffies+2);
+ if (connection->status != TCP_CONN_ESTABLISHED)
+ printk("%s::conn status %x\n", __func__, connection->status);
+ hash_set_valid_flag(connection->hash_entry_index, 1);
+ enable_irq(connection->gmac->irq);
+ // Gary Chen spin_unlock_irqrestore(&connection->conn_lock, conn_flags);
+ }
+}
+
+/*---------------------------------------------------------------------------
+ * free_toeq_descs
+ *-------------------------------------------------------------------------*/
+void free_toeq_descs(int qid, TOE_INFO_T *toe)
+{
+ void *desc_ptr;
+
+ desc_ptr = (void*)toe->toe_desc_base[qid];
+ pci_free_consistent(NULL, TOE_TOE_DESC_NUM*sizeof(GMAC_RXDESC_T), desc_ptr,
+ (dma_addr_t)toe->toe_desc_base_dma[qid]);
+ toe->toe_desc_base[qid] = 0;
+}
+
+/*---------------------------------------------------------------------------
+ * set_toeq_hdr
+ *-------------------------------------------------------------------------*/
+void set_toeq_hdr(struct toe_conn* connection, TOE_INFO_T* toe, struct net_device *dev)
+{
+ volatile TOE_QHDR_T *toe_qhdr;
+ volatile unsigned int toeq_wptr; // toeq_rptr
+ volatile GMAC_RXDESC_T *toe_curr_desc;
+ struct sk_buff *skb;
+ unsigned int pkt_size;
+ DMA_RWPTR_T toeq_rwptr;
+
+ if (connection->status == TCP_CONN_CLOSING) {
+ connection->status = TCP_CONN_CLOSED;
+ hash_set_valid_flag(connection->hash_entry_index, 0);
+ // remove timer first.
+ // del_timer_sync(&(connection->rx_timer));
+ // check if any queued frames last time.
+ toe_qhdr = (volatile TOE_QHDR_T*)TOE_TOE_QUE_HDR_BASE;
+ toe_qhdr += connection->qid;
+ toeq_rwptr.bits32 = readl(&toe_qhdr->word1);
+
+ //toeq_rptr = toe_qhdr->word1.bits.rptr;
+ toeq_wptr = toe_qhdr->word1.bits.wptr;
+ while (toeq_rwptr.bits.rptr != toeq_wptr) {
+ printk("%s::pending frames in TOE Queue before closing!\n", __func__);
+ toe_curr_desc = (GMAC_RXDESC_T*)(toe->toe_desc_base[connection->qid] +
+ toe_qhdr->word1.bits.rptr*sizeof(GMAC_RXDESC_T));
+ connection->curr_desc = (GMAC_RXDESC_T *)toe_curr_desc;
+ pkt_size = toe_curr_desc->word1.bits.byte_count;
+ consistent_sync((void*)__va(toe_curr_desc->word2.buf_adr), pkt_size,
+ PCI_DMA_FROMDEVICE);
+ skb = (struct sk_buff*)(REG32(__va(toe_curr_desc->word2.buf_adr) -
+ SKB_RESERVE_BYTES));
+ _debug_skb(skb, (GMAC_RXDESC_T *)toe_curr_desc, 0x03);
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, RX_INSERT_BYTES);
+ skb_put(skb, pkt_size);
+ skb->dev = connection->dev;
+ skb->protocol = eth_type_trans(skb, connection->dev);
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+ netif_rx(skb);
+ connection->dev->last_rx = jiffies;
+
+ toeq_rwptr.bits.rptr = RWPTR_ADVANCE_ONE(toeq_rwptr.bits.rptr, TOE_TOE_DESC_NUM);
+ SET_RPTR(&toe_qhdr->word1, toeq_rwptr.bits.rptr);
+ }
+ free_toeq_descs(connection->qid, toe);
+ // shall we re-fill free queue?
+
+ reset_connection_index(connection->qid);
+ //memset(connection, 0, sizeof(struct toe_conn));
+ printk(" del timer and close connection %x, qid %d\n", (u32)connection, connection->qid);
+ return;
+ }
+ /* enable or setup toe queue header */
+ if (connection->status == TCP_CONN_CONNECTING && storlink_ctl.rx_max_pktsize) {
+ volatile TOE_QHDR_T *qhdr;
+ int iq_id;
+ connection->status = TCP_CONN_ESTABLISHED;
+ qhdr = (volatile TOE_QHDR_T*)((unsigned int)TOE_TOE_QUE_HDR_BASE +
+ connection->qid * sizeof(TOE_QHDR_T));
+
+ iq_id = get_interrupt_queue_id(connection->qid);
+ connection->dev = dev;
+ connection->gmac = dev->priv;
+ connection->toeq_rwptr.bits32 = 0;
+
+// qhdr->word6.bits.iq_num = iq_id;
+ qhdr->word6.bits.MaxPktSize = (connection->max_pktsize)>>2; // in word.
+ qhdr->word7.bits.AckThreshold = connection->ack_threshold;
+ qhdr->word7.bits.SeqThreshold = connection->seq_threshold;
+
+ // init timer.
+#if 1
+ init_timer(&connection->rx_timer);
+ connection->rx_timer.expires = jiffies + 5;
+ connection->rx_timer.data = (unsigned long)connection;
+ connection->rx_timer.function = (void *)&connection_rx_timer;
+ add_timer(&connection->rx_timer);
+ connection->last_rx_jiffies = jiffies;
+ printk("init_timer %x\n", (u32)jiffies);
+#endif
+ hash_set_valid_flag(connection->hash_entry_index, 1);
+ return;
+ } else {
+ printk("%s::conn status %x, rx_pktsize %d\n",
+ __func__, connection->status, storlink_ctl.rx_max_pktsize);
+ }
+}
+
+/*---------------------------------------------------------------------------
+ * get_connection_index
+ * get_connection_index will find an available index for the connection,
+ * when allocate a new connection is needed.
+ * we find available Qid from AV bits and write to hash_table, so that when RxTOE
+ * packet is received, sw_id from ToeQ descriptor is also the Qid of conneciton Q.
+ *-------------------------------------------------------------------------*/
+int get_connection_index(void)
+{
+ int i=0, j=0, index=-1;
+ __u32 connection_bits;
+
+ for (i = 0; i< TOE_TOE_QUEUE_NUM/32; i++) {
+ connection_bits = ~(toe_connection_bits[i]);
+ if (connection_bits == 0)
+ // all 32 bits are used.
+ continue;
+
+ for (j=0; j<32; j++) {
+ if (connection_bits & 0x01) {
+ index = i*32 + j;
+ return index;
+ }
+ connection_bits = connection_bits >> 1;
+ }
+ }
+ return index;
+}
+
+/*---------------------------------------------------------------------------
+ * set_toe_connection
+ *-------------------------------------------------------------------------*/
+void set_toe_connection(int index, int val)
+{
+ if (val) {
+ toe_connection_bits[index/32] |= (1<<(index%32));
+ } else {
+ toe_connection_bits[index/32] &= (~(1<<(index%32)));
+ }
+}
+
+/*---------------------------------------------------------------------------
+ * sl351x_get_toe_conn_flag
+ *-------------------------------------------------------------------------*/
+int sl351x_get_toe_conn_flag(int index)
+{
+ if (index < TOE_TOE_QUEUE_NUM)
+ return (toe_connection_bits[index/32] & (1 << (index %32)));
+ else
+ return 0;
+}
+
+/*---------------------------------------------------------------------------
+ * sl351x_get_toe_conn_info
+ *-------------------------------------------------------------------------*/
+struct toe_conn * sl351x_get_toe_conn_info(int index)
+{
+ if (index < TOE_TOE_QUEUE_NUM)
+ return (struct toe_conn *)&toe_connections[index];
+ else
+ return NULL;
+}
+
+/*---------------------------------------------------------------------------
+ * create_sw_toe_connection
+ *-------------------------------------------------------------------------*/
+struct toe_conn* create_sw_toe_connection(int qid, int ip_ver, void* ip_hdr,
+ struct tcphdr* tcp_hdr)
+{
+ struct toe_conn* connection = &(toe_connections[qid]);
+
+ connection->ip_ver = (__u8)ip_ver;
+ connection->qid = (__u8)qid;
+ connection->source = (__u16)tcp_hdr->source;
+ connection->dest = (__u16)tcp_hdr->dest;
+ if (ip_ver == 4) {
+ struct iphdr* iph = (struct iphdr*) ip_hdr;
+ connection->saddr[0] = (__u32)iph->saddr;
+ connection->daddr[0] = (__u32)iph->daddr;
+// printk("%s::saddr %x, daddr %x\n", __func__,
+// ntohl(connection->saddr[0]), ntohl(connection->daddr[0]));
+ } else if (ip_ver == 6) {
+ struct ipv6hdr *iph = (struct ipv6hdr*)ip_hdr;
+ int i=0;
+ for (i=0; i<4; i++) {
+ connection->saddr[i] = (__u32)iph->saddr.in6_u.u6_addr32[i];
+ connection->daddr[i] = (__u32)iph->daddr.in6_u.u6_addr32[i];
+ }
+ }
+ connection->status = TCP_CONN_CREATION;
+ return connection;
+}
+
+/*---------------------------------------------------------------------------
+ * fill_toeq_buf
+ *-------------------------------------------------------------------------*/
+int fill_toeq_buf(int index, TOE_INFO_T* toe)
+{
+ volatile TOE_QHDR_T *qhdr;
+ //struct toe_conn* connection;
+ GMAC_RXDESC_T *desc_ptr;
+
+ if (!toe->toe_desc_base[index]) {
+ // first time. init.
+ desc_ptr = (GMAC_RXDESC_T*)(pci_alloc_consistent(NULL, TOE_TOE_DESC_NUM
+ *sizeof(GMAC_RXDESC_T), (dma_addr_t*)&toe->toe_desc_base_dma[index]));
+
+ toe->toe_desc_num = TOE_TOE_DESC_NUM;
+ toe->toe_desc_base[index] = (unsigned int)desc_ptr;
+ }
+ qhdr = (volatile TOE_QHDR_T*)((unsigned int)TOE_TOE_QUE_HDR_BASE +
+ index*sizeof(TOE_QHDR_T));
+ //connection = (struct toe_conn*)&(toe_connections[index]);
+
+ qhdr->word0.base_size = ((unsigned int)toe->toe_desc_base_dma[index]&TOE_QHDR0_BASE_MASK)
+ | TOE_TOE_DESC_POWER;
+ qhdr->word1.bits32 = 0;
+ qhdr->word2.bits32 = 0;
+ qhdr->word3.bits32 = 0;
+ qhdr->word4.bits32 = 0;
+ qhdr->word5.bits32 = 0;
+ return 1;
+}
+
+/*---------------------------------------------------------------------------
+ * create_toe_hash_entry_smb
+ * add SMB header in hash entry.
+ *-------------------------------------------------------------------------*/
+int create_toe_hash_entry_smb(int ip_ver, void* ip_hdr, struct tcphdr* tcp_hdr,
+ int sw_id)
+{
+ HASH_ENTRY_T hash_entry, *entry;
+ int hash_entry_index;
+ int i;
+
+ entry = (HASH_ENTRY_T*)&hash_entry;
+ memset((void*)entry, 0, sizeof(HASH_ENTRY_T));
+ entry->rule = 0;
+
+ /* enable fields of hash key */
+ entry->key_present.ip_protocol = 1;
+ entry->key_present.sip = 1;
+ entry->key_present.dip = 1;
+ entry->key_present.l4_bytes_0_3 = 1; // src port and dest port
+ entry->key_present.l7_bytes_0_3 = 0; // do we need to enable NETBIOS? how?
+ entry->key_present.l7_bytes_4_7 = 1; // "SMB" header
+
+ /* hash key */
+ entry->key.ip_protocol = IPPROTO_TCP;
+ if (ip_ver == 4) {
+ struct iphdr *iph = (struct iphdr*)ip_hdr;
+ memcpy(entry->key.sip, &iph->saddr, 4);
+ memcpy(entry->key.dip, &iph->daddr, 4);
+ } else if (ip_ver == 6) {
+ struct ipv6hdr *iph = (struct ipv6hdr*)ip_hdr;
+ for (i=0; i<4; i++) {
+ memcpy(&(entry->key.sip[i*4]), &(iph->saddr.in6_u.u6_addr32[i]), 4);
+ memcpy(&(entry->key.dip[i*4]), &(iph->daddr.in6_u.u6_addr32[i]), 4);
+ }
+ }
+ *(__u16*)&entry->key.l4_bytes[0] = tcp_hdr->source;
+ *(__u16*)&entry->key.l4_bytes[2] = tcp_hdr->dest;
+
+ entry->key.l7_bytes[4] = 0xff;
+ entry->key.l7_bytes[5] = 0x53;
+ entry->key.l7_bytes[6] = 0x4d;
+ entry->key.l7_bytes[7] = 0x42;
+
+ /* action of hash entry match */
+ entry->action.sw_id = 1;
+ entry->action.dest_qid = (__u8)TOE_TOE_QID(sw_id);
+ entry->action.srce_qid = 0;
+ hash_entry_index = hash_add_toe_entry(entry);
+
+ return hash_entry_index;
+}
+
+// best performance of tcp streaming.
+/*---------------------------------------------------------------------------
+ * create_toe_hash_entry_smb
+ * add SMB header in hash entry.
+ *-------------------------------------------------------------------------*/
+int create_toe_hash_entry_ftp(int ip_ver, void* ip_hdr, struct tcphdr* tcphdr)
+{
+ return 0;
+}
+
+// is hash entry for nfs needed?
+
+/*
+ * Create a TOE hash entry by given ip addresses and tcp port numbers.
+ * hash entry index will be saved in sw connection.
+ */
+/*---------------------------------------------------------------------------
+ * create_toe_hash_entry
+ *-------------------------------------------------------------------------*/
+int create_toe_hash_entry(int ip_ver, void* ip_hdr, struct tcphdr* tcp_hdr, int sw_id)
+{
+ HASH_ENTRY_T hash_entry, *entry;
+// unsigned long hash_key[HASH_MAX_DWORDS];
+ int hash_entry_index;
+
+ entry = (HASH_ENTRY_T*) &hash_entry;
+ memset((void*)entry, 0, sizeof(HASH_ENTRY_T));
+ entry->rule = 0;
+ /* enable fields of hash key */
+ entry->key_present.ip_protocol = 1;
+ entry->key_present.sip = 1;
+ entry->key_present.dip = 1;
+ entry->key_present.l4_bytes_0_3 = 1; // src port and dest port
+
+ /* hash key */
+ entry->key.ip_protocol = IPPROTO_TCP;
+ if (ip_ver == 4) {
+ // key of ipv4
+ struct iphdr* iph = (struct iphdr*)ip_hdr;
+ memcpy(entry->key.sip, &iph->saddr, 4);
+ memcpy(entry->key.dip, &iph->daddr, 4);
+ } else if (ip_ver == 6) {
+ // key of ipv6
+ int i=0;
+ struct ipv6hdr *iph = (struct ipv6hdr*)ip_hdr;
+ for (i=0; i<4; i++) {
+ memcpy(&(entry->key.sip[i*4]), &(iph->saddr.in6_u.u6_addr32[i]), 4);
+ memcpy(&(entry->key.dip[i*4]), &(iph->daddr.in6_u.u6_addr32[i]), 4);
+ }
+ }
+ *(__u16*)&entry->key.l4_bytes[0] = tcp_hdr->source;
+ *(__u16*)&entry->key.l4_bytes[2] = tcp_hdr->dest;
+ // is it necessary to write ip version to hash key?
+
+ /* action of hash entry match */
+ entry->action.sw_id = 1;
+ entry->action.dest_qid = (__u8)TOE_TOE_QID(sw_id);
+ entry->action.srce_qid = 0; // 0 for SW FreeQ. 1 for HW FreeQ.
+ hash_entry_index = hash_add_toe_entry(entry);
+// printk("\n%s. sw_id %d, hash_entry index %x\n",
+// __func__, TOE_TOE_QID(sw_id), hash_entry_index);
+ return hash_entry_index;
+}
+
+/*---------------------------------------------------------------------------
+ * init_toeq
+ * 1. Reserve a TOE Queue id first, to get the sw toe_connection.
+ * 2. Setup the hash entry with given iphdr and tcphdr, save hash entry index
+ * in sw toe_connection.
+ * 3. Prepare sw toe_connection and allocate buffers.
+ * 4. Validate hash entry.
+ *-------------------------------------------------------------------------*/
+struct toe_conn* init_toeq(int ipver, void* iph, struct tcphdr* tcp_hdr,
+ TOE_INFO_T* toe, unsigned char* l2hdr)
+{
+// printk("\t*** %s, ipver %d\n", __func__, ipver);
+ int qid=-1;
+ struct toe_conn* connection;
+ int hash_entry_index;
+ // int i=0;
+ unsigned short dest_port = ntohs(tcp_hdr->dest);
+
+ if (dest_port == 445) {
+ printk("%s::SMB/CIFS connection\n", __func__);
+ } else if (dest_port == 20) {
+ printk("%s::ftp-data connection\n", __func__);
+ } else if (dest_port == 2049) {
+ printk("%s::nfs daemon connection\n", __func__);
+ }
+ qid = get_connection_index();
+ if (qid<0)
+ return 0; // setup toeq failure
+ set_toe_connection(qid, 1); // reserve this sw toeq.
+
+ //connection = (struct toe_conn*)&(toe_connections[qid]);
+ hash_entry_index = create_toe_hash_entry(ipver, iph, tcp_hdr, qid);
+ if (hash_entry_index <0) {
+ printk("%s::release toe hash entry!\n", __func__);
+ set_toe_connection(qid, 0); // release this sw toeq.
+ return 0;
+ }
+ connection = create_sw_toe_connection(qid, ipver, iph, tcp_hdr);
+ connection->hash_entry_index = (__u16) hash_entry_index;
+
+ fill_toeq_buf(qid, toe);
+ memcpy(&connection->l2_hdr, l2hdr, sizeof(struct ethhdr));
+ spin_lock_init(&connection->conn_lock);
+
+ return connection;
+}
+
+#if 0
+/*----------------------------------------------------------------------
+* toe_init_toe_queue
+* (1) Initialize the TOE Queue Header
+* Register: TOE_TOE_QUE_HDR_BASE (0x60003000)
+* (2) Initialize Descriptors of TOE Queues
+*----------------------------------------------------------------------*/
+void toe_init_toe_queue(TOE_INFO_T* toe)
+{
+}
+EXPORT_SYMBOL(toe_init_toe_queue);
+#endif
+
+/*---------------------------------------------------------------------------
+ * dump_jumbo_skb
+ *-------------------------------------------------------------------------*/
+void dump_jumbo_skb(struct jumbo_frame *jumbo_skb)
+{
+ if (jumbo_skb->skb0) {
+// printk("%s. jumbo skb %x, len %d\n",
+// __func__, jumbo_skb->skb0->data, jumbo_skb->skb0->len);
+ netif_rx(jumbo_skb->skb0);
+ }
+ jumbo_skb->skb0 = 0;
+ jumbo_skb->tail = 0;
+ jumbo_skb->iphdr0 = 0;
+ jumbo_skb->tcphdr0 = 0;
+}
+
+/* ---------------------------------------------------------------------
+ * Append skb to skb0. skb0 is the jumbo frame that will be passed to
+ * kernel tcp.
+ * --------------------------------------------------------------------*/
+void rx_append_skb(struct jumbo_frame *jumbo_skb, struct sk_buff* skb, int payload_len)
+{
+ struct iphdr* iphdr0 = (struct iphdr*)&(skb->data[0]);
+ int ip_hdrlen = iphdr0->ihl << 2;
+ struct tcphdr* tcphdr0 = (struct tcphdr*)&(skb->data[ip_hdrlen]);
+
+ if (!jumbo_skb->skb0) {
+ // head of the jumbo frame.
+ jumbo_skb->skb0 = skb;
+ jumbo_skb->tail = 0;
+ jumbo_skb->iphdr0 = iphdr0;
+ jumbo_skb->tcphdr0 = tcphdr0;
+ } else {
+ if (!jumbo_skb->tail)
+ skb_shinfo(jumbo_skb->skb0)->frag_list = skb;
+ else
+ (jumbo_skb->tail)->next = skb;
+ jumbo_skb->tail = skb;
+
+ // do we need to change truesize as well?
+ jumbo_skb->skb0->len += payload_len;
+ jumbo_skb->skb0->data_len += payload_len;
+
+ jumbo_skb->iphdr0->tot_len = htons(ntohs(jumbo_skb->iphdr0->tot_len)+payload_len);
+ jumbo_skb->tcphdr0->ack_seq = tcphdr0->ack_seq;
+ jumbo_skb->tcphdr0->window = tcphdr0->window;
+
+ skb->len += payload_len;
+ skb->data_len = 0;
+ skb->data += ntohs(iphdr0->tot_len) - payload_len;
+ }
+}
+
+/*----------------------------------------------------------------------
+* toe_gmac_handle_toeq
+* (1) read interrupt Queue to get TOE Q.
+* (2) get packet fro TOE Q and send to upper layer handler.
+* (3) allocate new buffers and put to TOE Q. Intr Q buffer is recycled.
+*----------------------------------------------------------------------*/
+void toe_gmac_handle_toeq(struct net_device *dev, GMAC_INFO_T* tp, __u32 status)
+{
+ //volatile INTRQ_INFO_T *intrq_info;
+ //TOEQ_INFO_T *toeq_info;
+ volatile NONTOE_QHDR_T *intr_qhdr;
+ volatile TOE_QHDR_T *toe_qhdr;
+ volatile INTR_QHDR_T *intr_curr_desc;
+ TOE_INFO_T *toe = &toe_private_data;
+
+ volatile GMAC_RXDESC_T *toe_curr_desc; // , *fq_desc;// *tmp_desc;
+ volatile DMA_RWPTR_T intr_rwptr, toeq_rwptr; // fq_rwptr;
+
+ unsigned int pkt_size, desc_count, tcp_qid;
+ volatile unsigned int toeq_wptr;
+ struct toe_conn* connection;
+ int i, frag_id = 0;
+ // unsigned long toeq_flags;
+ struct jumbo_frame jumbo_skb;
+ struct sk_buff *skb;
+ __u32 interrupt_status;
+
+ in_toe_isr++;
+
+ interrupt_status = status >> 24;
+ // get interrupt queue header
+ intr_qhdr = (volatile NONTOE_QHDR_T*)TOE_INTR_Q_HDR_BASE;
+ memset(&jumbo_skb, 0, sizeof(struct jumbo_frame));
+
+ for (i=0; i<TOE_INTR_QUEUE_NUM; i++, intr_qhdr++) {
+ if (!(interrupt_status & 0x0001)) {
+ // no interrupt of this IntQ
+ interrupt_status = interrupt_status >> 1;
+ continue;
+ }
+ interrupt_status = interrupt_status >> 1;
+ intr_rwptr.bits32 = readl(&intr_qhdr->word1);
+
+ while ( intr_rwptr.bits.rptr != intr_rwptr.bits.wptr) {
+ int max_pktsize = 1;
+ // get interrupt queue descriptor.
+ intr_curr_desc = (INTR_QHDR_T*)toe->intr_desc_base +
+ i* TOE_INTR_DESC_NUM + intr_rwptr.bits.rptr;
+// printk("%s::int %x\n", __func__, intr_curr_desc->word1.bits32);
+ // get toeq id
+ tcp_qid = (u8)intr_curr_desc->word1.bits.tcp_qid - (u8)TOE_TOE_QID(0);
+ // get toeq queue header
+ toe_qhdr = (volatile TOE_QHDR_T*) TOE_TOE_QUE_HDR_BASE;
+ toe_qhdr += tcp_qid;
+ connection = &toe_connections[tcp_qid];
+ del_timer(&connection->rx_timer);
+ // Gary Chen spin_lock_irqsave(&connection->conn_lock, toeq_flags);
+ // handling interrupts of this TOE Q.
+ if (intr_curr_desc->word1.bits.ctl || intr_curr_desc->word1.bits.osq ||
+ intr_curr_desc->word1.bits.abn)
+ max_pktsize = 0;
+ if (!max_pktsize || intr_curr_desc->word1.bits.TotalPktSize) {
+ desc_count=0;
+ // wptr in intl queue is where this TOE interrupt should stop.
+ toeq_rwptr.bits32 = readl(&toe_qhdr->word1);
+ toeq_wptr = intr_curr_desc->word0.bits.wptr;
+ if (connection->toeq_rwptr.bits.rptr != toeq_rwptr.bits.rptr)
+ printk("conn rptr %d, hw rptr %d\n",
+ connection->toeq_rwptr.bits.rptr, toeq_rwptr.bits.rptr);
+
+ if (intr_curr_desc->word1.bits.ctl &&
+ (toeq_rwptr.bits.rptr == toeq_wptr)) {
+ printk("\nctrl frame, but not in TOE queue! conn rptr %d, hw wptr %d\n",
+ connection->toeq_rwptr.bits.rptr, toeq_wptr);
+// dump_toe_qhdr(toe_qhdr);
+// dump_intrq_desc(intr_curr_desc);
+ }
+ // while (toeq_rwptr.bits.rptr != intr_curr_desc->word0.bits.wptr) {
+ while (toe_qhdr->word1.bits.rptr != intr_curr_desc->word0.bits.wptr) {
+ frag_id++;
+ toe_curr_desc = (volatile GMAC_RXDESC_T *)(toe->toe_desc_base[tcp_qid]
+ + toe_qhdr->word1.bits.rptr *sizeof(GMAC_RXDESC_T));
+ connection->curr_desc = (GMAC_RXDESC_T *)toe_curr_desc;
+ desc_count = toe_curr_desc->word0.bits.desc_count;
+ pkt_size = toe_curr_desc->word1.bits.byte_count;
+ consistent_sync((void*)__va(toe_curr_desc->word2.buf_adr), pkt_size,
+ PCI_DMA_FROMDEVICE);
+ skb = (struct sk_buff*)(REG32(__va(toe_curr_desc->word2.buf_adr)-
+ SKB_RESERVE_BYTES));
+ _debug_skb(skb, (GMAC_RXDESC_T *)toe_curr_desc, 0x01);
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, RX_INSERT_BYTES);
+ if ((skb->len + pkt_size) > (1514+16))
+ {
+ printk("skb->len=%d, pkt_size=%d\n",skb->len, pkt_size);
+ while(1);
+ }
+
+ skb_put(skb, pkt_size);
+ skb->dev = dev;
+ skb->protocol = eth_type_trans(skb, dev);
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+
+ if (toe_curr_desc->word3.bits32 & 0x1b000000)
+ dump_jumbo_skb(&jumbo_skb);
+
+ rx_append_skb(&jumbo_skb, skb, pkt_size-toe_curr_desc->word3.bits.l7_offset);
+// spin_lock_irqsave(&gmac_fq_lock, flags);
+ toeq_rwptr.bits.rptr = RWPTR_ADVANCE_ONE(toeq_rwptr.bits.rptr, TOE_TOE_DESC_NUM);
+ SET_RPTR(&toe_qhdr->word1, toeq_rwptr.bits.rptr);
+// spin_unlock_irqrestore(&gmac_fq_lock, flags);
+ if (storlink_ctl.fqint_threshold)
+ continue;
+#if 0
+//#if (HANDLE_FREEQ_METHOD == HANDLE_FREEQ_INDIVIDUAL)
+ if ((skb = dev_alloc_skb(RX_BUF_SIZE)) == NULL) {
+ printk("%s::toe queue alloc buffer ", __func__);
+ }
+ *(unsigned int*)(skb->data) = (unsigned int)skb;
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES);
+
+ spin_lock_irqsave(&gmac_fq_lock, flags);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+ if (toe->fq_rx_rwptr.bits.wptr != fq_rwptr.bits.wptr) {
+ printk("%s::fq_rx_rwptr %x\n", __func__, toe->fq_rx_rwptr.bits32);
+ mac_stop_txdma((struct net_device*) tp->dev);
+ spin_unlock_irqrestore(&gmac_fq_lock, flags);
+ while(1);
+ }
+ fq_desc = (GMAC_RXDESC_T*)toe->swfq_desc_base + fq_rwptr.bits.wptr;
+ fq_desc->word2.buf_adr = (unsigned int)__pa(skb->data);
+
+ fq_rwptr.bits.wptr = RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr, TOE_SW_FREEQ_DESC_NUM);
+ SET_WPTR(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG, fq_rwptr.bits.wptr);
+ toe->fq_rx_rwptr.bits32 = fq_rwptr.bits32;
+ spin_unlock_irqrestore(&gmac_fq_lock, flags);
+#endif
+ } // end of this multi-desc.
+ dump_jumbo_skb(&jumbo_skb);
+ dev->last_rx = jiffies;
+ connection->toeq_rwptr.bits32 = toeq_rwptr.bits32;
+ } else if (intr_curr_desc->word1.bits.sat) {
+ toeq_rwptr.bits32 = readl(&toe_qhdr->word1);
+ toeq_wptr = intr_curr_desc->word0.bits.wptr;
+ if (connection->toeq_rwptr.bits.rptr != toeq_rwptr.bits.rptr)
+ printk("SAT. conn rptr %d, hw rptr %d\n",
+ connection->toeq_rwptr.bits.rptr, toeq_rwptr.bits.rptr);
+/*
+ printk("%s::SAT int!, ackcnt %x, seqcnt %x, rptr %d, wptr %d, ack %x, qhack %x\n",
+ __func__, intr_curr_desc->word4.bits.AckCnt, intr_curr_desc->word4.bits.SeqCnt,
+ toeq_rptr, toeq_wptr, intr_curr_desc->word3.ack_num, toe_qhdr->word4.ack_num);*/
+ /* pure ack */
+ if (toeq_rwptr.bits.rptr == toeq_wptr) {
+ if (intr_curr_desc->word4.bits32) {
+ skb = gen_pure_ack(connection, (TOE_QHDR_T *)toe_qhdr, (INTR_QHDR_T *)intr_curr_desc);
+ skb_put(skb, 60);
+ skb->dev = connection->dev;
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+ skb->protocol = eth_type_trans(skb, connection->dev);
+ netif_rx(skb);
+ } else
+ printk("%s::SAT Interrupt!. But cnt is 0!\n", __func__);
+ } else {
+ // while (toeq_rwptr.bits.rptr != toeq_wptr) {
+ while (toe_qhdr->word1.bits.rptr != intr_curr_desc->word0.bits.wptr) {
+ toe_curr_desc = (volatile GMAC_RXDESC_T*)(toe->toe_desc_base[tcp_qid]
+ + toe_qhdr->word1.bits.rptr * sizeof(GMAC_RXDESC_T));
+ connection->curr_desc = (GMAC_RXDESC_T *)toe_curr_desc;
+ desc_count = toe_curr_desc->word0.bits.desc_count;
+ pkt_size = toe_curr_desc->word1.bits.byte_count;
+ consistent_sync((void*)__va(toe_curr_desc->word2.buf_adr), pkt_size,
+ PCI_DMA_FROMDEVICE);
+ // if ( ((toeq_rwptr.bits.rptr +1)&(TOE_TOE_DESC_NUM-1)) == toeq_wptr) {
+ if ( RWPTR_ADVANCE_ONE(toe_qhdr->word1.bits.rptr, TOE_TOE_DESC_NUM) == toeq_wptr) {
+ skb = (struct sk_buff*)(REG32(__va(toe_curr_desc->word2.buf_adr) -
+ SKB_RESERVE_BYTES));
+ _debug_skb(skb, (GMAC_RXDESC_T *)toe_curr_desc, 0x04);
+ connection->curr_rx_skb = skb;
+ skb_reserve(skb, RX_INSERT_BYTES);
+ skb_put(skb, pkt_size);
+ skb->dev = dev;
+ skb->protocol = eth_type_trans(skb, dev);
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+ // printk("toeq_rptr %d, wptr %d\n", toeq_rptr, toeq_wptr);
+ netif_rx(skb);
+ dev->last_rx = jiffies;
+/*
+ if ((skb = dev_alloc_skb(RX_BUF_SIZE)) == NULL) {
+
+ }
+ *(unsigned int*)(skb->data) = (unsigned int) skb;
+ skb_reserve(skb, SKB_RESERVE_BYTES); */
+ } else {
+ // reuse this skb, append to free queue..
+ skb = (struct sk_buff*)(REG32(__va(toe_curr_desc->word2.buf_adr)-
+ SKB_RESERVE_BYTES));
+ _debug_skb(skb, (GMAC_RXDESC_T *)toe_curr_desc, 0x05);
+ connection->curr_rx_skb = skb;
+ dev_kfree_skb_irq(skb);
+ }
+#if 0
+ spin_lock_irqsave(&gmac_fq_lock, flags);
+ fq_rwptr.bits32 = readl(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG);
+/* if (toe->fq_rx_rwptr.bits.wptr != fq_rwptr.bits.wptr) {
+ printk("%s::fq_rx_rwptr %x\n", __func__, toe->fq_rx_rwptr.bits32);
+ mac_stop_txdma((struct net_device*) tp->dev);
+ spin_unlock_irqrestore(&gmac_fq_lock, flags);
+ while(1);
+ } */
+ fq_desc = (GMAC_RXDESC_T*)toe->swfq_desc_base + fq_rwptr.bits.wptr;
+ fq_desc->word2.buf_adr = (unsigned int)__pa(skb->data);
+
+ fq_rwptr.bits.wptr = RWPTR_ADVANCE_ONE(fq_rwptr.bits.wptr, TOE_SW_FREEQ_DESC_NUM);
+ SET_WPTR(TOE_GLOBAL_BASE + GLOBAL_SWFQ_RWPTR_REG, fq_rwptr.bits.wptr);
+ toe->fq_rx_rwptr.bits32 = fq_rwptr.bits32;
+ // spin_unlock_irqrestore(&gmac_fq_lock, flags);
+#endif
+// spin_lock_irqsave(&gmac_fq_lock, flags);
+ toeq_rwptr.bits.rptr = RWPTR_ADVANCE_ONE(toeq_rwptr.bits.rptr, TOE_TOE_DESC_NUM);
+ SET_RPTR(&toe_qhdr->word1, toeq_rwptr.bits.rptr);
+// spin_unlock_irqrestore(&gmac_fq_lock, flags);
+ }
+ } // end of ACK with options.
+ connection->toeq_rwptr.bits32 = toeq_rwptr.bits32;
+ // Gary Chen spin_unlock_irqrestore(&connection->conn_lock, toeq_flags);
+// }
+ };
+ update_timer(connection);
+ // any protection against interrupt queue header?
+ intr_rwptr.bits.rptr = RWPTR_ADVANCE_ONE(intr_rwptr.bits.rptr, TOE_INTR_DESC_NUM);
+ SET_RPTR(&intr_qhdr->word1, intr_rwptr.bits.rptr);
+ intr_rwptr.bits32 = readl(&intr_qhdr->word1);
+ toe_gmac_fill_free_q();
+ } // end of this interrupt Queue processing.
+ } // end of all interrupt Queues.
+
+ in_toe_isr = 0;
+}
+
+
--- /dev/null
+++ b/drivers/net/sl_lepus_hash.c
@@ -0,0 +1,553 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl_lepus_hash.c
+* Description :
+* Handle Storlink Lepus Hash Functions
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* 03/13/2006 Gary Chen Create and implement
+*
+****************************************************************************/
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/compiler.h>
+#include <linux/pci.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/rtnetlink.h>
+#include <linux/delay.h>
+#include <linux/ethtool.h>
+#include <linux/mii.h>
+#include <linux/completion.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+#include <asm/irq.h>
+#include <asm/semaphore.h>
+#include <asm/arch/irqs.h>
+#include <asm/arch/it8712.h>
+#include <linux/mtd/kvctl.h>
+#include <linux/skbuff.h>
+#include <linux/ip.h>
+#include <linux/tcp.h>
+#include <linux/list.h>
+#define MIDWAY
+#define SL_LEPUS
+
+#include <asm/arch/sl2312.h>
+#include <asm/arch/sl_lepus_gmac.h>
+#include <asm/arch/sl_hash_cfg.h>
+
+#ifndef RXTOE_DEBUG
+#define RXTOE_DEBUG
+#endif
+#undef RXTOE_DEBUG
+
+/*----------------------------------------------------------------------
+* Definition
+*----------------------------------------------------------------------*/
+#define hash_printf printk
+
+#define HASH_TIMER_PERIOD (60*HZ) // seconds
+#define HASH_ILLEGAL_INDEX 0xffff
+
+/*----------------------------------------------------------------------
+* Variables
+*----------------------------------------------------------------------*/
+u32 hash_activate_bits[HASH_TOTAL_ENTRIES/32];
+u32 hash_nat_owner_bits[HASH_TOTAL_ENTRIES/32];
+char hash_tables[HASH_TOTAL_ENTRIES][HASH_MAX_BYTES] __attribute__ ((aligned(16)));
+static struct timer_list hash_timer_obj;
+LIST_HEAD(hash_timeout_list);
+
+/*----------------------------------------------------------------------
+* Functions
+*----------------------------------------------------------------------*/
+void dm_long(u32 location, int length);
+static void hash_timer_func(u32 data);
+
+/*----------------------------------------------------------------------
+* hash_init
+*----------------------------------------------------------------------*/
+void hash_init(void)
+{
+ int i;
+ volatile u32 *dp1, *dp2, dword;
+
+ dp1 = (volatile u32 *) TOE_V_BIT_BASE;
+ dp2 = (volatile u32 *) TOE_A_BIT_BASE;
+
+ for (i=0; i<HASH_TOTAL_ENTRIES/32; i++)
+ {
+ *dp1++ = 0;
+ dword = *dp2++; // read-clear
+ }
+ memset((void *)&hash_nat_owner_bits, 0, sizeof(hash_nat_owner_bits));
+ memset((void *)&hash_tables, 0, sizeof(hash_tables));
+
+ init_timer(&hash_timer_obj);
+ hash_timer_obj.expires = jiffies + HASH_TIMER_PERIOD;
+ hash_timer_obj.data = (unsigned long)&hash_timer_obj;
+ hash_timer_obj.function = (void *)&hash_timer_func;
+ add_timer(&hash_timer_obj);
+
+#if (HASH_MAX_BYTES == 128)
+ writel((unsigned long)__pa(&hash_tables) | 3, // 32 words
+ TOE_GLOBAL_BASE + GLOBAL_HASH_TABLE_BASE_REG);
+#elif (HASH_MAX_BYTES == 64)
+ writel((unsigned long)__pa(&hash_tables) | 2, // 16 words
+ TOE_GLOBAL_BASE + GLOBAL_HASH_TABLE_BASE_REG);
+#else
+ #error Incorrect setting for HASH_MAX_BYTES
+#endif
+
+}
+/*----------------------------------------------------------------------
+* hash_add_entry
+*----------------------------------------------------------------------*/
+int hash_add_entry(HASH_ENTRY_T *entry)
+{
+ int rc;
+ u32 key[HASH_MAX_DWORDS];
+ rc = hash_build_keys((u32 *)&key, entry);
+ if (rc < 0)
+ return -1;
+ hash_write_entry(entry, (unsigned char*) &key[0]);
+// hash_set_valid_flag(entry->index, 1);
+// printk("Dump hash key!\n");
+// dump_hash_key(entry);
+ return entry->index;
+}
+
+/*----------------------------------------------------------------------
+* hash_set_valid_flag
+*----------------------------------------------------------------------*/
+void hash_set_valid_flag(int index, int valid)
+{
+ register u32 reg32;
+
+ reg32 = TOE_V_BIT_BASE + (index/32) * 4;
+
+ if (valid)
+ {
+ writel(readl(reg32) | (1 << (index%32)), reg32);
+ }
+ else
+ {
+ writel(readl(reg32) & ~(1 << (index%32)), reg32);
+ }
+}
+
+/*----------------------------------------------------------------------
+* hash_set_nat_owner_flag
+*----------------------------------------------------------------------*/
+void hash_set_nat_owner_flag(int index, int valid)
+{
+ if (valid)
+ {
+ hash_nat_owner_bits[index/32] |= (1 << (index % 32));
+ }
+ else
+ {
+ hash_nat_owner_bits[index/32] &= ~(1 << (index % 32));
+ }
+}
+
+
+/*----------------------------------------------------------------------
+* hash_build_keys
+*----------------------------------------------------------------------*/
+int hash_build_keys(u32 *destp, HASH_ENTRY_T *entry)
+{
+ u32 data;
+ unsigned char *cp;
+ int i, j;
+ unsigned short index;
+ int total;
+
+ memset((void *)destp, 0, HASH_MAX_BYTES);
+ cp = (unsigned char *)destp;
+
+ if (entry->key_present.port || entry->key_present.Ethertype)
+ {
+ HASH_PUSH_WORD(cp, entry->key.Ethertype); // word 0
+ HASH_PUSH_BYTE(cp, entry->key.port); // Byte 2
+ HASH_PUSH_BYTE(cp, 0); // Byte 3
+ }
+ else
+ {
+ HASH_PUSH_DWORD(cp, 0);
+ }
+
+ if (entry->key_present.da || entry->key_present.sa)
+ {
+ unsigned char mac[4];
+ if (entry->key_present.da)
+ {
+ for (i=0; i<4; i++)
+ HASH_PUSH_BYTE(cp, entry->key.da[i]);
+ }
+ mac[0] = (entry->key_present.da) ? entry->key.da[4] : 0;
+ mac[1] = (entry->key_present.da) ? entry->key.da[5] : 0;
+ mac[2] = (entry->key_present.sa) ? entry->key.sa[0] : 0;
+ mac[3] = (entry->key_present.sa) ? entry->key.sa[1] : 0;
+ data = mac[0] + (mac[1]<<8) + (mac[2]<<16) + (mac[3]<<24);
+ HASH_PUSH_DWORD(cp, data);
+ if (entry->key_present.sa)
+ {
+ for (i=2; i<6; i++)
+ HASH_PUSH_BYTE(cp, entry->key.sa[i]);
+ }
+ }
+
+ if (entry->key_present.pppoe_sid || entry->key_present.vlan_id)
+ {
+ HASH_PUSH_WORD(cp, entry->key.vlan_id); // low word
+ HASH_PUSH_WORD(cp, entry->key.pppoe_sid); // high word
+ }
+ if (entry->key_present.ipv4_hdrlen || entry->key_present.ip_tos || entry->key_present.ip_protocol)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.ip_protocol); // Byte 0
+ HASH_PUSH_BYTE(cp, entry->key.ip_tos); // Byte 1
+ HASH_PUSH_BYTE(cp, entry->key.ipv4_hdrlen); // Byte 2
+ HASH_PUSH_BYTE(cp, 0); // Byte 3
+ }
+
+ if (entry->key_present.ipv6_flow_label)
+ {
+ HASH_PUSH_DWORD(cp, entry->key.ipv6_flow_label); // low word
+ }
+ if (entry->key_present.sip)
+ {
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[i]);
+ if (entry->key.ipv6)
+ {
+ for (i=4; i<16; i+=4)
+ {
+ for (j=i+3; j>=i; j--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[j]);
+ }
+ }
+ }
+ if (entry->key_present.dip)
+ {
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[i]);
+ if (entry->key.ipv6)
+ {
+ for (i=4; i<16; i+=4)
+ {
+ for (j=i+3; j>=i; j--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[j]);
+ }
+ }
+ }
+
+ if (entry->key_present.l4_bytes_0_3)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[3]);
+ }
+ if (entry->key_present.l4_bytes_4_7)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[4]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[5]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[6]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[7]);
+ }
+ if (entry->key_present.l4_bytes_8_11)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[8]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[9]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[10]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[11]);
+ }
+ if (entry->key_present.l4_bytes_12_15)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[12]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[13]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[14]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[15]);
+ }
+ if (entry->key_present.l4_bytes_16_19)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[16]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[17]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[18]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[19]);
+ }
+ if (entry->key_present.l4_bytes_20_23)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[20]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[21]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[22]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[23]);
+ }
+ if (entry->key_present.l7_bytes_0_3)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[3]);
+ }
+ if (entry->key_present.l7_bytes_4_7)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[4]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[5]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[6]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[7]);
+ }
+ if (entry->key_present.l7_bytes_8_11)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[8]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[9]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[10]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[11]);
+ }
+ if (entry->key_present.l7_bytes_12_15)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[12]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[13]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[14]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[15]);
+ }
+ if (entry->key_present.l7_bytes_16_19)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[16]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[17]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[18]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[19]);
+ }
+ if (entry->key_present.l7_bytes_20_23)
+ {
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[20]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[21]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[22]);
+ HASH_PUSH_BYTE(cp, entry->key.l7_bytes[23]);
+ }
+
+ // get hash index
+ total = (u32)((u32)cp - (u32)destp) / (sizeof(u32));
+
+ if (total > HASH_MAX_KEY_DWORD)
+ {
+ //hash_printf("Total key words (%d) is too large (> %d)!\n",
+ // total, HASH_MAX_KEY_DWORD);
+ return -1;
+ }
+
+ if (entry->key_present.port || entry->key_present.Ethertype)
+ index = hash_gen_crc16((unsigned char *)destp, total * 4);
+ else
+ {
+ if (total == 1)
+ {
+ hash_printf("No key is assigned!\n");
+ return -1;
+ }
+
+ index = hash_gen_crc16((unsigned char *)(destp+1), (total-1) * 4);
+ }
+
+ entry->index = index & HASH_BITS_MASK;
+
+ //hash_printf("Total key words = %d, Hash Index= %d\n",
+ // total, entry->index);
+
+ cp = (unsigned char *)destp;
+ cp+=3;
+ HASH_PUSH_BYTE(cp, entry->rule); // rule
+
+ entry->total_dwords = total;
+
+ return total;
+}
+
+/*----------------------------------------------------------------------
+* hash_build_nat_keys
+*----------------------------------------------------------------------*/
+void hash_build_nat_keys(u32 *destp, HASH_ENTRY_T *entry)
+{
+ unsigned char *cp;
+ int i;
+ unsigned short index;
+ int total;
+
+ memset((void *)destp, 0, HASH_MAX_BYTES);
+
+ cp = (unsigned char *)destp + 2;
+ HASH_PUSH_BYTE(cp, entry->key.port);
+ cp++;
+
+ if (entry->key_present.pppoe_sid || entry->key_present.vlan_id)
+ {
+ HASH_PUSH_WORD(cp, entry->key.vlan_id); // low word
+ HASH_PUSH_WORD(cp, entry->key.pppoe_sid); // high word
+ }
+
+ HASH_PUSH_BYTE(cp, entry->key.ip_protocol);
+ cp+=3;
+
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.sip[i]);
+
+ // input (entry->key.sip[i]) is network-oriented
+ // output (hash key) is host-oriented
+ for (i=3; i>=0; i--)
+ HASH_PUSH_BYTE(cp, entry->key.dip[i]);
+
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[0]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[1]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[2]);
+ HASH_PUSH_BYTE(cp, entry->key.l4_bytes[3]);
+
+ // get hash index
+ total = (u32)((u32)cp - (u32)destp) / (sizeof(u32));
+
+ index = hash_gen_crc16((unsigned char *)destp, total * 4);
+ entry->index = index & ((1 << HASH_BITS) - 1);
+
+ cp = (unsigned char *)destp;
+ cp+=3;
+ HASH_PUSH_BYTE(cp, entry->rule); // rule
+
+ entry->total_dwords = total;
+}
+
+
+/*----------------------------------------------------------------------
+* hash_write_entry
+*----------------------------------------------------------------------*/
+int hash_write_entry(HASH_ENTRY_T *entry, unsigned char *key)
+{
+ int i;
+ u32 *srcep, *destp, *destp2;
+
+ srcep = (u32 *)key;
+ destp2 = destp = (u32 *)&hash_tables[entry->index][0];
+
+ for (i=0; i<(entry->total_dwords); i++, srcep++, destp++)
+ *destp = *srcep;
+
+ srcep = (u32 *)&entry->action;
+ *destp++ = *srcep;
+
+ srcep = (u32 *)&entry->param;
+ for (i=0; i<(sizeof(ENTRY_PARAM_T)/sizeof(*destp)); i++, srcep++, destp++)
+ *destp = *srcep;
+
+ memset(destp, 0, (HASH_MAX_DWORDS-entry->total_dwords-HASH_ACTION_DWORDS) * sizeof(u32));
+
+ consistent_sync(destp2, (entry->total_dwords+HASH_ACTION_DWORDS) * 4, PCI_DMA_TODEVICE);
+ return 0;
+}
+
+/*----------------------------------------------------------------------
+* hash_timer_func
+*----------------------------------------------------------------------*/
+static void hash_timer_func(u32 data)
+{
+ int i, j;
+ volatile u32 *active_p, *own_p, *valid_p;
+ u32 a_bits, own_bits;
+
+ valid_p = (volatile u32 *)TOE_V_BIT_BASE;
+ active_p = (volatile u32 *)hash_activate_bits;
+ own_p = (volatile u32 *)hash_nat_owner_bits;
+ for (i=0; i<(HASH_TOTAL_ENTRIES/32); i++, own_p++, active_p++, valid_p++)
+ {
+ *active_p |= readl(TOE_A_BIT_BASE + (i*4));
+ a_bits = *active_p;
+ own_bits = *own_p;
+ if (own_bits)
+ {
+#ifndef DEBUG_NAT_MIXED_HW_SW_TX
+ a_bits = own_bits & ~a_bits;
+#else
+ a_bits = own_bits & a_bits;
+#endif
+ for (j=0; a_bits && j<32; j++)
+ {
+ if (a_bits & 1)
+ {
+ *valid_p &= ~(1 << j); // invalidate it
+#if !(defined(NAT_DEBUG_LAN_HASH_TIMEOUT) || defined(NAT_DEBUG_WAN_HASH_TIMEOUT))
+ *own_p &= ~(1 << j); // release ownership for NAT
+#endif
+// #ifdef DEBUG_NAT_MIXED_HW_SW_TX
+#if 0
+ hash_printf("%lu %s: Clear hash index: %d\n", jiffies/HZ, __func__, i*32+j);
+#endif
+ }
+ a_bits >>= 1;
+ }
+ *active_p &= ~own_bits; // deactivate it for next polling
+ }
+ }
+
+ hash_timer_obj.expires = jiffies + HASH_TIMER_PERIOD;
+ add_timer((struct timer_list *)data);
+}
+
+/*----------------------------------------------------------------------
+* dm_long
+*----------------------------------------------------------------------*/
+void dm_long(u32 location, int length)
+{
+ u32 *start_p, *curr_p, *end_p;
+ u32 *datap, data;
+ int i;
+
+ //if (length > 1024)
+ // length = 1024;
+
+ start_p = (u32 *)location;
+ end_p = (u32 *)location + length;
+ curr_p = (u32 *)((u32)location & 0xfffffff0);
+ datap = (u32 *)location;
+ while (curr_p < end_p)
+ {
+ hash_printf("0x%08x: ",(u32)curr_p & 0xfffffff0);
+ for (i=0; i<4; i++)
+ {
+ if (curr_p < start_p || curr_p >= end_p)
+ hash_printf(" ");
+ else
+ {
+ data = *datap;
+ hash_printf("%08X ", data);
+ }
+ if (i==1)
+ hash_printf("- ");
+
+ curr_p++;
+ datap++;
+ }
+ hash_printf("\n");
+ }
+}
+
+/*----------------------------------------------------------------------
+* hash_dump_entry
+*----------------------------------------------------------------------*/
+void hash_dump_entry(int index)
+{
+ hash_printf("Hash Index %d:\n", index);
+ dm_long((u32)&hash_tables[index][0], HASH_MAX_DWORDS);
+}
+
+
--- /dev/null
+++ b/drivers/net/sl_switch.c
@@ -0,0 +1,650 @@
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/init.h>
+#include <linux/ioport.h>
+#include <linux/delay.h>
+#include <asm/hardware.h>
+#include <asm/io.h>
+
+#define GMAC_GLOBAL_BASE_ADDR (IO_ADDRESS(SL2312_GLOBAL_BASE))
+#define GPIO_BASE_ADDR1 (IO_ADDRESS(SL2312_GPIO_BASE1))
+enum GPIO_REG
+{
+ GPIO_DATA_OUT = 0x00,
+ GPIO_DATA_IN = 0x04,
+ GPIO_PIN_DIR = 0x08,
+ GPIO_BY_PASS = 0x0c,
+ GPIO_DATA_SET = 0x10,
+ GPIO_DATA_CLEAR = 0x14,
+};
+
+#define GMAC_SPEED_10 0
+#define GMAC_SPEED_100 1
+#define GMAC_SPEED_1000 2
+
+enum phy_state
+{
+ LINK_DOWN = 0,
+ LINK_UP = 1
+};
+
+#ifndef BIT
+#define BIT(x) (1 << (x))
+#endif
+
+//int Get_Set_port_status();
+unsigned int SPI_read_bit(void);
+void SPI_write_bit(char bit_EEDO);
+void SPI_write(unsigned char block,unsigned char subblock,unsigned char addr,unsigned int value);
+unsigned int SPI_read(unsigned char block,unsigned char subblock,unsigned char addr);
+int SPI_default(void);
+void SPI_CS_enable(unsigned char enable);
+unsigned int SPI_get_identifier(void);
+void phy_write(unsigned char port_no,unsigned char reg,unsigned int val);
+unsigned int phy_read(unsigned char port_no,unsigned char reg);
+void phy_write_masked(unsigned char port_no,unsigned char reg,unsigned int val,unsigned int mask);
+void init_seq_7385(unsigned char port_no) ;
+void phy_receiver_init (unsigned char port_no);
+
+#define PORT_NO 4
+int switch_pre_speed[PORT_NO]={0,0,0,0};
+int switch_pre_link[PORT_NO]={0,0,0,0};
+
+
+
+
+
+/* NOTES
+ * The Protocol of the SPI are as follows:
+ *
+ * Bit7 Bit6 Bit5 Bit4 Bit3 Bit2 Bit1 Bit0
+ * byte0 | Block id | r/w | sub-block |
+ * byte1 | Address |
+ * byte2 | Data |
+ * byte3 | Data |
+ * byte4 | Data |
+ * byte5 | Data |
+ */
+
+
+
+
+/***************************************/
+/* define GPIO module base address */
+/***************************************/
+#define GPIO_EECS 0x80000000 /* EECS: GPIO[22] */
+#define GPIO_MOSI 0x20000000 /* EEDO: GPIO[29] send to 6996*/
+#define GPIO_MISO 0x40000000 /* EEDI: GPIO[30] receive from 6996*/
+#define GPIO_EECK 0x10000000 /* EECK: GPIO[31] */
+
+/*************************************************************
+* SPI protocol for ADM6996 control
+**************************************************************/
+#define SPI_OP_LEN 0x08 // the length of start bit and opcode
+#define SPI_OPWRITE 0X05 // write
+#define SPI_OPREAD 0X06 // read
+#define SPI_OPERASE 0X07 // erase
+#define SPI_OPWTEN 0X04 // write enable
+#define SPI_OPWTDIS 0X04 // write disable
+#define SPI_OPERSALL 0X04 // erase all
+#define SPI_OPWTALL 0X04 // write all
+
+#define SPI_ADD_LEN 8 // bits of Address
+#define SPI_DAT_LEN 32 // bits of Data
+
+
+/****************************************/
+/* Function Declare */
+/****************************************/
+
+//unsigned int SPI_read_bit(void);
+//void SPI_write_bit(char bit_EEDO);
+//unsigned int SPI_read_bit(void);
+/******************************************
+* SPI_write
+* addr -> Write Address
+* value -> value to be write
+***************************************** */
+void phy_receiver_init (unsigned char port_no)
+{
+ phy_write(port_no,31,0x2a30);
+ phy_write_masked(port_no, 12, 0x0200, 0x0300);
+ phy_write(port_no,31,0);
+}
+
+void phy_write(unsigned char port_no,unsigned char reg,unsigned int val)
+{
+ unsigned int cmd;
+
+ cmd = (port_no<<21)|(reg<<16)|val;
+ SPI_write(3,0,1,cmd);
+}
+
+unsigned int phy_read(unsigned char port_no,unsigned char reg)
+{
+ unsigned int cmd,reg_val;
+
+ cmd = BIT(26)|(port_no<<21)|(reg<<16);
+ SPI_write(3,0,1,cmd);
+ msleep(2);
+ reg_val = SPI_read(3,0,2);
+ return reg_val;
+}
+
+void phy_write_masked(unsigned char port_no,unsigned char reg,unsigned int val,unsigned int mask)
+{
+ unsigned int cmd,reg_val;
+
+ cmd = BIT(26)|(port_no<<21)|(reg<<16); // Read reg_val
+ SPI_write(3,0,1,cmd);
+ mdelay(2);
+ reg_val = SPI_read(3,0,2);
+ reg_val &= ~mask; // Clear masked bit
+ reg_val |= (val&mask) ; // set masked bit ,if true
+ cmd = (port_no<<21)|(reg<<16)|reg_val;
+ SPI_write(3,0,1,cmd);
+}
+
+void init_seq_7385(unsigned char port_no)
+{
+ unsigned char rev;
+
+ phy_write(port_no, 31, 0x2a30);
+ phy_write_masked(port_no, 8, 0x0200, 0x0200);
+ phy_write(port_no, 31, 0x52b5);
+ phy_write(port_no, 16, 0xb68a);
+ phy_write_masked(port_no, 18, 0x0003, 0xff07);
+ phy_write_masked(port_no, 17, 0x00a2, 0x00ff);
+ phy_write(port_no, 16, 0x968a);
+ phy_write(port_no, 31, 0x2a30);
+ phy_write_masked(port_no, 8, 0x0000, 0x0200);
+ phy_write(port_no, 31, 0x0000); /* Read revision */
+ rev = phy_read(port_no, 3) & 0x000f;
+ if (rev == 0)
+ {
+ phy_write(port_no, 31, 0x2a30);
+ phy_write_masked(port_no, 8, 0x0200, 0x0200);
+ phy_write(port_no, 31, 0x52b5);
+ phy_write(port_no, 18, 0x0000);
+ phy_write(port_no, 17, 0x0689);
+ phy_write(port_no, 16, 0x8f92);
+ phy_write(port_no, 31, 0x52B5);
+ phy_write(port_no, 18, 0x0000);
+ phy_write(port_no, 17, 0x0E35);
+ phy_write(port_no, 16, 0x9786);
+ phy_write(port_no, 31, 0x2a30);
+ phy_write_masked(port_no, 8, 0x0000, 0x0200);
+ phy_write(port_no, 23, 0xFF80);
+ phy_write(port_no, 23, 0x0000);
+ }
+ phy_write(port_no, 31, 0x0000);
+ phy_write(port_no, 18, 0x0048);
+ if (rev == 0)
+ {
+ phy_write(port_no, 31, 0x2a30);
+ phy_write(port_no, 20, 0x6600);
+ phy_write(port_no, 31, 0x0000);
+ phy_write(port_no, 24, 0xa24e);
+ }
+ else
+ {
+ phy_write(port_no, 31, 0x2a30);
+ phy_write_masked(port_no, 22, 0x0240, 0x0fc0);
+ phy_write_masked(port_no, 20, 0x4000, 0x6000);
+ phy_write(port_no, 31, 1);
+ phy_write_masked(port_no, 20, 0x6000, 0xe000);
+ phy_write(port_no, 31, 0x0000);
+ }
+}
+
+int Get_Set_port_status()
+{
+ unsigned int reg_val,ability,rcv_mask,mac_config;
+ int is_link=0;
+ int i;
+
+ rcv_mask = SPI_read(2,0,0x10); // Receive mask
+
+ for(i=0;i<4;i++){
+ reg_val = phy_read(i,1);
+ if ((reg_val & 0x0024) == 0x0024) /* link is established and auto_negotiate process completed */
+ {
+ is_link=1;
+ if(switch_pre_link[i]==LINK_DOWN){ // Link Down ==> Link up
+
+ rcv_mask |= BIT(i); // Enable receive
+
+ reg_val = phy_read(i,10);
+ if(reg_val & 0x0c00){
+ printk("Port%d:Giga mode\n",i);
+// SPI_write(1,i,0x00,0x300701B1);
+ mac_config = 0x00060004|(6<<6);
+
+ SPI_write(1,i,0x00,((mac_config & 0xfffffff8) | 1) | 0x20000030); // reset port
+ mac_config |= (( BIT(i) << 19) | 0x08000000);
+ SPI_write(1,i,0x00,mac_config);
+ SPI_write(1,i,0x04,0x000300ff); // flow control
+
+ reg_val = SPI_read(5,0,0x12);
+ reg_val &= ~BIT(i);
+ SPI_write(5,0,0x12,reg_val);
+
+ reg_val = SPI_read(1,i,0x00);
+ reg_val |= 0x10010000;
+ SPI_write(1,i,0x00,reg_val);
+// SPI_write(1,i,0x00,0x10070181);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_1000;
+ }
+ else{
+ reg_val = phy_read(i,5);
+ ability = (reg_val&0x5e0) >>5;
+ if ((ability & 0x0C)) /* 100M */
+ {
+// SPI_write(1,i,0x00,0x30050472);
+ if((ability&0x08)==0) // Half
+ mac_config = 0x00040004 |(17<<6);
+ else // Full
+ mac_config = 0x00040004 |(17<<6);
+
+ SPI_write(1,i,0x00,((mac_config & 0xfffffff8) | 1) | 0x20000030); // reset port
+ mac_config |= (( BIT(i) << 19) | 0x08000000);
+ SPI_write(1,i,0x00,mac_config);
+ SPI_write(1,i,0x04,0x000300ff); // flow control
+
+ reg_val = SPI_read(5,0,0x12);
+ reg_val &= ~BIT(i);
+ SPI_write(5,0,0x12,reg_val);
+
+ reg_val = SPI_read(1,i,0x00);
+ reg_val &= ~0x08000000;
+ reg_val |= 0x10010000;
+ SPI_write(1,i,0x00,reg_val);
+// SPI_write(1,i,0x00,0x10050442);
+ printk("Port%d:100M\n",i);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_100;
+ }
+ else if((ability & 0x03)) /* 10M */
+ {
+// SPI_write(1,i,0x00,0x30050473);
+ if((ability&0x2)==0) // Half
+ mac_config = 0x00040004 |(17<<6);
+ else // Full
+ mac_config = 0x00040004 |(17<<6);
+
+ SPI_write(1,i,0x00,((mac_config & 0xfffffff8) | 1) | 0x20000030); // reset port
+ mac_config |= (( BIT(i) << 19) | 0x08000000);
+ SPI_write(1,i,0x00,mac_config);
+ SPI_write(1,i,0x04,0x000300ff); // flow control
+
+ reg_val = SPI_read(5,0,0x12);
+ reg_val &= ~BIT(i);
+ SPI_write(5,0,0x12,reg_val);
+
+ reg_val = SPI_read(1,i,0x00);
+ reg_val &= ~0x08000000;
+ reg_val |= 0x10010000;
+ SPI_write(1,i,0x00,reg_val);
+// SPI_write(1,i,0x00,0x10050443);
+ printk("Port%d:10M\n",i);
+ switch_pre_link[i]=LINK_UP;
+ switch_pre_speed[i]=GMAC_SPEED_10;
+ }
+ else{
+ SPI_write(1,i,0x00,0x20000030);
+ printk("Port%d:Unknown mode\n",i);
+ switch_pre_link[i]=LINK_DOWN;
+ switch_pre_speed[i]=GMAC_SPEED_10;
+ }
+ }
+ }
+ else{ // Link up ==> Link UP
+
+ }
+ }
+ else{ // Link Down
+ if(switch_pre_link[i]==LINK_UP){
+ printk("Port%d:Link Down\n",i);
+ //phy_receiver_init(i);
+ reg_val = SPI_read(1,i,0);
+ reg_val &= ~BIT(16);
+ SPI_write(1,i,0x00,reg_val); // disable RX
+ SPI_write(5,0,0x0E,BIT(i)); // dicard packet
+ while((SPI_read(5,0,0x0C)&BIT(i))==0) // wait to be empty
+ msleep(1);
+ SPI_write(1,i,0x00,0x20000030); // PORT_RST
+ SPI_write(5,0,0x0E,SPI_read(5,0,0x0E) & ~BIT(i));// accept packet
+
+ reg_val = SPI_read(5,0,0x12);
+ reg_val |= BIT(i);
+ SPI_write(5,0,0x12,reg_val);
+ }
+ switch_pre_link[i]=LINK_DOWN;
+ rcv_mask &= ~BIT(i); // disable receive
+ }
+ }
+
+ SPI_write(2,0,0x10,rcv_mask); // Receive mask
+ return is_link;
+
+}
+EXPORT_SYMBOL(Get_Set_port_status);
+
+void SPI_write(unsigned char block,unsigned char subblock,unsigned char addr,unsigned int value)
+{
+ int i;
+ char bit;
+ unsigned int data;
+
+ SPI_CS_enable(1);
+
+ data = (block<<5) | 0x10 | subblock;
+
+ //send write command
+ for(i=SPI_OP_LEN-1;i>=0;i--)
+ {
+ bit = (data>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+
+ // send 8 bits address (MSB first, LSB last)
+ for(i=SPI_ADD_LEN-1;i>=0;i--)
+ {
+ bit = (addr>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+ // send 32 bits data (MSB first, LSB last)
+ for(i=SPI_DAT_LEN-1;i>=0;i--)
+ {
+ bit = (value>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+
+ SPI_CS_enable(0); // CS low
+
+}
+
+
+/************************************
+* SPI_write_bit
+* bit_EEDO -> 1 or 0 to be written
+************************************/
+void SPI_write_bit(char bit_EEDO)
+{
+ unsigned int addr;
+ unsigned int value;
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_PIN_DIR);
+ value = readl(addr) |GPIO_EECK |GPIO_MOSI ; /* set EECK/MISO Pin to output */
+ writel(value,addr);
+ if(bit_EEDO)
+ {
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_SET);
+ writel(GPIO_MOSI,addr); /* set MISO to 1 */
+
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(GPIO_MOSI,addr); /* set MISO to 0 */
+ }
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); /* set EECK to 1 */
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); /* set EECK to 0 */
+
+ //return ;
+}
+
+/**********************************************************************
+* read a bit from ADM6996 register
+***********************************************************************/
+unsigned int SPI_read_bit(void) // read data from
+{
+ unsigned int addr;
+ unsigned int value;
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_PIN_DIR);
+ value = readl(addr) & (~GPIO_MISO); // set EECK to output and MISO to input
+ writel(value,addr);
+
+ addr =(GPIO_BASE_ADDR1 + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); // set EECK to 1
+
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_IN);
+ value = readl(addr) ;
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); // set EECK to 0
+
+
+ value = value >> 30;
+ return value ;
+}
+
+/******************************************
+* SPI_default
+* EEPROM content default value
+*******************************************/
+int SPI_default(void)
+{
+ int i;
+ unsigned reg_val,cmd;
+
+#if 0
+ SPI_write(7,0,0x1C,0x01); // map code space to 0
+
+ reg_val = SPI_read(7,0,0x10);
+ reg_val |= 0x0146;
+ reg_val &= ~0x0001;
+ SPI_write(7,0,0x10,reg_val); // reset iCPU and enable ext_access
+ SPI_write(7,0,0x11,0x0000); // start address
+ for(i=0;i<sizeof(vts_img);i++){
+ SPI_write(7,0,0x12,vts_img[i]); // fill in ROM data
+ }
+ reg_val |= BIT(0)|BIT(3);
+ SPI_write(7,0,0x10,reg_val); // release iCPU
+ SPI_write(7,0,0x10,SPI_read(7,0,0x10)&~BIT(7)); // release iCPU
+ return ;
+#endif
+
+
+ for(i=0;i<15;i++){
+ if(i!=6 && i!=7)
+ SPI_write(3,2,0,0x1010400+i); // Initial memory
+ mdelay(1);
+ }
+
+ mdelay(30);
+
+ SPI_write(2,0,0xB0,0x05); // Clear MAC table
+ SPI_write(2,0,0xD0,0x03); // Clear VLAN
+
+ //for(i=0;i<5;i++)
+ SPI_write(1,6,0x19,0x2C); // Double Data rate
+
+ for(i=0;i<4;i++){
+ SPI_write(1,i,0x00,0x30050472); // MAC configure
+ SPI_write(1,i,0x00,0x10050442); // MAC configure
+ SPI_write(1,i,0x10,0x5F4); // Max length
+ SPI_write(1,i,0x04,0x00030000); // Flow control
+ SPI_write(1,i,0xDF,0x00000001); // Flow control
+ SPI_write(1,i,0x08,0x000050c2); // Flow control mac high
+ SPI_write(1,i,0x0C,0x002b00f1); // Flow control mac low
+ SPI_write(1,i,0x6E,BIT(3)); // forward pause frame
+ }
+ SPI_write(1,i,0x00,0x20000030); // set port 4 as reset
+
+ SPI_write(1,6,0x00,0x300701B1); // MAC configure
+ SPI_write(1,6,0x00,0x10070181); // MAC configure
+ SPI_write(1,6,0x10,0x5F4); // Max length
+ SPI_write(1,6,0x04,0x00030000); // Flow control
+ SPI_write(1,6,0xDF,0x00000002); // Flow control
+ SPI_write(1,6,0x08,0x000050c2); // Flow control mac high
+ SPI_write(1,6,0x0C,0x002b00f1); // Flow control mac low
+ SPI_write(1,6,0x6E,BIT(3)); // forward pause frame
+
+
+ //SPI_write(7,0,0x05,0x31); // MII delay for loader
+ //SPI_write(7,0,0x05,0x01); // MII delay for kernel
+ SPI_write(7,0,0x05,0x33);
+
+ SPI_write(2,0,0x10,0x4F); // Receive mask
+
+ mdelay(50);
+
+ SPI_write(7,0,0x14,0x02); // Release Reset
+
+ mdelay(3);
+
+ for(i=0;i<4;i++){
+ init_seq_7385(i);
+ phy_receiver_init(i);
+ cmd = BIT(26)|(i<<21)|(0x1B<<16); // Config LED
+ SPI_write(3,0,1,cmd);
+ mdelay(10);
+ reg_val = SPI_read(3,0,2);
+ reg_val &= 0xFF00;
+ reg_val |= 0x61;
+ cmd = (i<<21)|(0x1B<<16)|reg_val;
+ SPI_write(3,0,1,cmd);
+
+ cmd = BIT(26)|(i<<21)|(0x04<<16); // Pause enable
+ SPI_write(3,0,1,cmd);
+ mdelay(10);
+ reg_val = SPI_read(3,0,2);
+ reg_val |= BIT(10)|BIT(11);
+ cmd = (i<<21)|(0x04<<16)|reg_val;
+ SPI_write(3,0,1,cmd);
+
+ cmd = BIT(26)|(i<<21)|(0x0<<16); // collision test and re-negotiation
+ SPI_write(3,0,1,cmd);
+ mdelay(10);
+ reg_val = SPI_read(3,0,2);
+ reg_val |= BIT(7)|BIT(8)|BIT(9);
+ cmd = (i<<21)|(0x0<<16)|reg_val;
+ SPI_write(3,0,1,cmd);
+ }
+ init_seq_7385(i);
+ writel(0x5787a7f0,GMAC_GLOBAL_BASE_ADDR+0x1c);//For switch timing
+ return 4; // return port_no
+}
+EXPORT_SYMBOL(SPI_default);
+
+/***********************************************************
+* SPI_CS_enable
+* before access ,you have to enable Chip Select. (pull high)
+* When fisish, you should pull low !!
+*************************************************************/
+void SPI_CS_enable(unsigned char enable)
+{
+
+ unsigned int addr,value;
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_PIN_DIR);
+ value = readl(addr) |GPIO_EECS |GPIO_EECK; /* set EECS/EECK Pin to output */
+ writel(value,addr);
+
+ if(enable)
+ {
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(GPIO_EECK,addr); /* set EECK to 0 */ // pull low clk first
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(GPIO_EECS,addr); /* set EECS to 0 */
+
+ }
+ else
+ {
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_SET);
+ writel(GPIO_EECK,addr); /* set EECK to 1 */ // pull high clk before disable
+ writel(GPIO_EECS,addr); /* set EECS to 1 */
+ }
+}
+
+
+/************************************************
+* SPI_read
+* table -> which table to be read: 1/count 0/EEPROM
+* addr -> Address to be read
+* return : Value of the register
+*************************************************/
+unsigned int SPI_read(unsigned char block,unsigned char subblock,unsigned char addr)
+{
+ int i;
+ char bit;
+ unsigned int data,value=0;
+
+ SPI_CS_enable(1);
+
+ data = (block<<5) | subblock;
+
+ //send write command
+ for(i=SPI_OP_LEN-1;i>=0;i--)
+ {
+ bit = (data>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+
+ // send 8 bits address (MSB first, LSB last)
+ for(i=SPI_ADD_LEN-1;i>=0;i--)
+ {
+ bit = (addr>>i)& 0x01;
+ SPI_write_bit(bit);
+ }
+
+ // dummy read for chip ready
+ for(i=0;i<8;i++)
+ SPI_read_bit();
+
+
+ // read 32 bits data (MSB first, LSB last)
+ for(i=SPI_DAT_LEN-1;i>=0;i--)
+ {
+ bit = SPI_read_bit();
+ value |= bit<<i;
+ }
+
+ SPI_CS_enable(0); // CS low
+ return(value);
+
+}
+
+void pull_low_gpio(unsigned int val)
+{
+
+ unsigned int addr,value;
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_DATA_CLEAR);
+ writel(val,addr); /* set pin low to save power*/
+
+ addr = (GPIO_BASE_ADDR1 + GPIO_PIN_DIR);
+ value = readl(addr) & ~ val; /* set Pin to input */
+ writel(value,addr);
+
+// value = readl(GMAC_GLOBAL_BASE_ADDR+0x0C); // reset GPIO1 module(self clear)
+// value |= BIT(21);
+// writel(value,GMAC_GLOBAL_BASE_ADDR+0x0C);
+}
+
+unsigned int SPI_get_identifier(void)
+{
+ unsigned int flag=0;
+
+ SPI_write(7,0,0x01,0x01);
+ flag = SPI_read(7,0,0x18); // chip id
+ if((flag & 0x0ffff000)==0x07385000){
+ printk("Get VSC-switch ID 0x%08x\n",flag);
+ //Giga_switch = 1;;
+ return 1;
+ }
+ else{
+ printk("VSC-switch not found\n");
+ //Giga_switch = 0;
+ pull_low_gpio(GPIO_EECK|GPIO_MOSI|GPIO_MISO|GPIO_EECS); // reduce power consume
+ return 0;
+ }
+}
+EXPORT_SYMBOL(SPI_get_identifier);
+
--- /dev/null
+++ b/include/asm-arm/arch-sl2312/sl351x_gmac.h
@@ -0,0 +1,2223 @@
+/****************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl351x_gmac.h
+* Description :
+* Define for device driver of Storlink SL351x network Engine
+*
+* Historych
+*
+* Date Writer Description
+* ----------- ----------- -------------------------------------------------
+* 08/22/2005 Gary Chen Create and implement
+*
+****************************************************************************/
+#ifndef _GMAC_SL351x_H
+#define _GMAC_SL351x_H
+#include <linux/skbuff.h>
+
+#define SL351x_GMAC_WORKAROUND 1
+
+#undef BIG_ENDIAN
+#define BIG_ENDIAN 0
+#define GMAC_DEBUG 1
+#define GMAC_NUM 2
+//#define L2_jumbo_frame 1
+
+#define _PACKED_ __attribute__ ((aligned(1), packed))
+
+#ifndef BIT
+#define BIT(x) (1 << (x))
+#endif
+
+#define REG32(addr) (*(volatile unsigned long * const)(addr))
+
+#define DMA_MALLOC(size,handle) pci_alloc_consistent(NULL,size,handle)
+#define DMA_MFREE(mem,size,handle) pci_free_consistent(NULL,size,mem,handle)
+
+// Define frame size
+#define ETHER_ADDR_LEN 6
+#define GMAC_MAX_ETH_FRAME_SIZE 1514
+#define GMAC_TX_BUF_SIZE ((GMAC_MAX_ETH_FRAME_SIZE + 31) & (~31))
+#define MAX_ISR_WORK 20
+
+#ifdef L2_jumbo_frame
+#define SW_RX_BUF_SIZE 9234 // 2048 ,9234
+#else
+#define SW_RX_BUF_SIZE 1536 // 2048
+#endif
+
+#define HW_RX_BUF_SIZE 1536 // 2048
+
+#define GMAC_DEV_TX_TIMEOUT (10*HZ) //add by CH
+#define SKB_RESERVE_BYTES 16
+
+/**********************************************************************
+ * Base Register
+ **********************************************************************/
+#define TOE_BASE (IO_ADDRESS(SL2312_TOE_BASE))
+#define GMAC_GLOBAL_BASE_ADDR (IO_ADDRESS(SL2312_GLOBAL_BASE))
+
+#define TOE_GLOBAL_BASE (TOE_BASE + 0x0000)
+#define TOE_NONTOE_QUE_HDR_BASE (TOE_BASE + 0x2000)
+#define TOE_TOE_QUE_HDR_BASE (TOE_BASE + 0x3000)
+#define TOE_V_BIT_BASE (TOE_BASE + 0x4000)
+#define TOE_A_BIT_BASE (TOE_BASE + 0x6000)
+#define TOE_GMAC0_DMA_BASE (TOE_BASE + 0x8000)
+#define TOE_GMAC0_BASE (TOE_BASE + 0xA000)
+#define TOE_GMAC1_DMA_BASE (TOE_BASE + 0xC000)
+#define TOE_GMAC1_BASE (TOE_BASE + 0xE000)
+
+/**********************************************************************
+ * Queue ID
+ **********************************************************************/
+#define TOE_SW_FREE_QID 0x00
+#define TOE_HW_FREE_QID 0x01
+#define TOE_GMAC0_SW_TXQ0_QID 0x02
+#define TOE_GMAC0_SW_TXQ1_QID 0x03
+#define TOE_GMAC0_SW_TXQ2_QID 0x04
+#define TOE_GMAC0_SW_TXQ3_QID 0x05
+#define TOE_GMAC0_SW_TXQ4_QID 0x06
+#define TOE_GMAC0_SW_TXQ5_QID 0x07
+#define TOE_GMAC0_HW_TXQ0_QID 0x08
+#define TOE_GMAC0_HW_TXQ1_QID 0x09
+#define TOE_GMAC0_HW_TXQ2_QID 0x0A
+#define TOE_GMAC0_HW_TXQ3_QID 0x0B
+#define TOE_GMAC1_SW_TXQ0_QID 0x12
+#define TOE_GMAC1_SW_TXQ1_QID 0x13
+#define TOE_GMAC1_SW_TXQ2_QID 0x14
+#define TOE_GMAC1_SW_TXQ3_QID 0x15
+#define TOE_GMAC1_SW_TXQ4_QID 0x16
+#define TOE_GMAC1_SW_TXQ5_QID 0x17
+#define TOE_GMAC1_HW_TXQ0_QID 0x18
+#define TOE_GMAC1_HW_TXQ1_QID 0x19
+#define TOE_GMAC1_HW_TXQ2_QID 0x1A
+#define TOE_GMAC1_HW_TXQ3_QID 0x1B
+#define TOE_GMAC0_DEFAULT_QID 0x20
+#define TOE_GMAC1_DEFAULT_QID 0x21
+#define TOE_CLASSIFICATION_QID(x) (0x22 + x) // 0x22 ~ 0x2F
+#define TOE_TOE_QID(x) (0x40 + x) // 0x40 ~ 0x7F
+
+/**********************************************************************
+ * TOE DMA Queue Number should be 2^n, n = 6...12
+ * TOE DMA Queues are the following queue types:
+ * SW Free Queue, HW Free Queue,
+ * GMAC 0/1 SW TX Q0-5, and GMAC 0/1 HW TX Q0-5
+ * They have same descriptor numbers.
+ * The base address and descriptor number are configured at
+ * DMA Queues Descriptor Ring Base Address/Size Register (offset 0x0004)
+ **********************************************************************/
+#define TOE_SW_FREEQ_DESC_POWER 10
+#define TOE_SW_FREEQ_DESC_NUM (1<<TOE_SW_FREEQ_DESC_POWER)
+#define TOE_HW_FREEQ_DESC_POWER 8
+#define TOE_HW_FREEQ_DESC_NUM (1<<TOE_HW_FREEQ_DESC_POWER)
+#define TOE_GMAC0_SWTXQ_DESC_POWER 8
+#define TOE_GMAC0_SWTXQ_DESC_NUM (1<<TOE_GMAC0_SWTXQ_DESC_POWER)
+#define TOE_GMAC0_HWTXQ_DESC_POWER 8
+#define TOE_GMAC0_HWTXQ_DESC_NUM (1<<TOE_GMAC0_HWTXQ_DESC_POWER)
+#define TOE_GMAC1_SWTXQ_DESC_POWER 8
+#define TOE_GMAC1_SWTXQ_DESC_NUM (1<<TOE_GMAC1_SWTXQ_DESC_POWER)
+#define TOE_GMAC1_HWTXQ_DESC_POWER 8
+#define TOE_GMAC1_HWTXQ_DESC_NUM (1<<TOE_GMAC1_HWTXQ_DESC_POWER)
+#define TOE_DEFAULT_Q0_DESC_POWER 8
+#define TOE_DEFAULT_Q0_DESC_NUM (1<<TOE_DEFAULT_Q0_DESC_POWER)
+#define TOE_DEFAULT_Q1_DESC_POWER 8
+#define TOE_DEFAULT_Q1_DESC_NUM (1<<TOE_DEFAULT_Q1_DESC_POWER)
+#define TOE_TOE_DESC_POWER 8
+#define TOE_TOE_DESC_NUM (1<<TOE_TOE_DESC_POWER)
+#define TOE_CLASS_DESC_POWER 8
+#define TOE_CLASS_DESC_NUM (1<<TOE_CLASS_DESC_POWER)
+#define TOE_INTR_DESC_POWER 8
+#define TOE_INTR_DESC_NUM (1<<TOE_INTR_DESC_POWER)
+
+#define TOE_TOE_QUEUE_MAX 64
+#define TOE_TOE_QUEUE_NUM 64
+#define TOE_CLASS_QUEUE_MAX 14
+#define TOE_CLASS_QUEUE_NUM 14
+#define TOE_INTR_QUEUE_MAX 4
+#define TOE_INTR_QUEUE_NUM 4
+#define TOE_SW_TXQ_MAX 6
+#define TOE_SW_TXQ_NUM 1
+#define TOE_HW_TXQ_MAX 4
+#define TOE_HW_TXQ_NUM 4
+#define _max(x,y) ((x>y) ? x :y)
+#define TX_DESC_NUM _max(TOE_GMAC0_SWTXQ_DESC_NUM, TOE_GMAC1_SWTXQ_DESC_NUM)
+
+#define RWPTR_ADVANCE_ONE(x, max) ((x == (max -1)) ? 0 : x+1)
+#define RWPTR_RECEDE_ONE(x, max) ((x == 0) ? (max -1) : x-1)
+#define SET_WPTR(addr, data) (*(volatile u16 * const)((u32)(addr)+2) = (u16)data)
+#define SET_RPTR(addr, data) (*(volatile u16 * const)((u32)(addr)) = (u16)data)
+
+/**********************************************************************
+ * Global registers
+ * #define TOE_GLOBAL_BASE (TOE_BASE + 0x0000)
+ * Base 0x60000000
+ **********************************************************************/
+#define GLOBAL_TOE_VERSION_REG 0x0000
+#define GLOBAL_SW_FREEQ_BASE_SIZE_REG 0x0004
+#define GLOBAL_HW_FREEQ_BASE_SIZE_REG 0x0008
+#define GLOBAL_DMA_SKB_SIZE_REG 0x0010
+#define GLOBAL_SWFQ_RWPTR_REG 0x0014
+#define GLOBAL_HWFQ_RWPTR_REG 0x0018
+#define GLOBAL_INTERRUPT_STATUS_0_REG 0x0020
+#define GLOBAL_INTERRUPT_ENABLE_0_REG 0x0024
+#define GLOBAL_INTERRUPT_SELECT_0_REG 0x0028
+#define GLOBAL_INTERRUPT_STATUS_1_REG 0x0030
+#define GLOBAL_INTERRUPT_ENABLE_1_REG 0x0034
+#define GLOBAL_INTERRUPT_SELECT_1_REG 0x0038
+#define GLOBAL_INTERRUPT_STATUS_2_REG 0x0040
+#define GLOBAL_INTERRUPT_ENABLE_2_REG 0x0044
+#define GLOBAL_INTERRUPT_SELECT_2_REG 0x0048
+#define GLOBAL_INTERRUPT_STATUS_3_REG 0x0050
+#define GLOBAL_INTERRUPT_ENABLE_3_REG 0x0054
+#define GLOBAL_INTERRUPT_SELECT_3_REG 0x0058
+#define GLOBAL_INTERRUPT_STATUS_4_REG 0x0060
+#define GLOBAL_INTERRUPT_ENABLE_4_REG 0x0064
+#define GLOBAL_INTERRUPT_SELECT_4_REG 0x0068
+#define GLOBAL_HASH_TABLE_BASE_REG 0x006C
+#define GLOBAL_QUEUE_THRESHOLD_REG 0x0070
+
+/**********************************************************************
+ * GMAC 0/1 DMA/TOE register
+ * #define TOE_GMAC0_DMA_BASE (TOE_BASE + 0x8000)
+ * #define TOE_GMAC1_DMA_BASE (TOE_BASE + 0xC000)
+ * Base 0x60008000 or 0x6000C000
+ **********************************************************************/
+#define GMAC_DMA_CTRL_REG 0x0000
+#define GMAC_TX_WEIGHTING_CTRL_0_REG 0x0004
+#define GMAC_TX_WEIGHTING_CTRL_1_REG 0x0008
+#define GMAC_SW_TX_QUEUE0_PTR_REG 0x000C
+#define GMAC_SW_TX_QUEUE1_PTR_REG 0x0010
+#define GMAC_SW_TX_QUEUE2_PTR_REG 0x0014
+#define GMAC_SW_TX_QUEUE3_PTR_REG 0x0018
+#define GMAC_SW_TX_QUEUE4_PTR_REG 0x001C
+#define GMAC_SW_TX_QUEUE5_PTR_REG 0x0020
+#define GMAC_HW_TX_QUEUE0_PTR_REG 0x0024
+#define GMAC_HW_TX_QUEUE1_PTR_REG 0x0028
+#define GMAC_HW_TX_QUEUE2_PTR_REG 0x002C
+#define GMAC_HW_TX_QUEUE3_PTR_REG 0x0030
+#define GMAC_DMA_TX_FIRST_DESC_REG 0x0038
+#define GMAC_DMA_TX_CURR_DESC_REG 0x003C
+#define GMAC_DMA_TX_DESC_WORD0_REG 0x0040
+#define GMAC_DMA_TX_DESC_WORD1_REG 0x0044
+#define GMAC_DMA_TX_DESC_WORD2_REG 0x0048
+#define GMAC_DMA_TX_DESC_WORD3_REG 0x004C
+#define GMAC_SW_TX_QUEUE_BASE_REG 0x0050
+#define GMAC_HW_TX_QUEUE_BASE_REG 0x0054
+#define GMAC_DMA_RX_FIRST_DESC_REG 0x0058
+#define GMAC_DMA_RX_CURR_DESC_REG 0x005C
+#define GMAC_DMA_RX_DESC_WORD0_REG 0x0060
+#define GMAC_DMA_RX_DESC_WORD1_REG 0x0064
+#define GMAC_DMA_RX_DESC_WORD2_REG 0x0068
+#define GMAC_DMA_RX_DESC_WORD3_REG 0x006C
+#define GMAC_HASH_ENGINE_REG0 0x0070
+#define GMAC_HASH_ENGINE_REG1 0x0074
+#define GMAC_MR0CR0 0x0078 // matching rule 0 Control register 0
+#define GMAC_MR0CR1 0x007C // matching rule 0 Control register 1
+#define GMAC_MR0CR2 0x0080 // matching rule 0 Control register 2
+#define GMAC_MR1CR0 0x0084 // matching rule 1 Control register 0
+#define GMAC_MR1CR1 0x0088 // matching rule 1 Control register 1
+#define GMAC_MR1CR2 0x008C // matching rule 1 Control register 2
+#define GMAC_MR2CR0 0x0090 // matching rule 2 Control register 0
+#define GMAC_MR2CR1 0x0094 // matching rule 2 Control register 1
+#define GMAC_MR2CR2 0x0098 // matching rule 2 Control register 2
+#define GMAC_MR3CR0 0x009C // matching rule 3 Control register 0
+#define GMAC_MR3CR1 0x00A0 // matching rule 3 Control register 1
+#define GMAC_MR3CR2 0x00A4 // matching rule 3 Control register 2
+#define GMAC_SPR0 0x00A8 // Support Protocol Regsister 0
+#define GMAC_SPR1 0x00AC // Support Protocol Regsister 1
+#define GMAC_SPR2 0x00B0 // Support Protocol Regsister 2
+#define GMAC_SPR3 0x00B4 // Support Protocol Regsister 3
+#define GMAC_SPR4 0x00B8 // Support Protocol Regsister 4
+#define GMAC_SPR5 0x00BC // Support Protocol Regsister 5
+#define GMAC_SPR6 0x00C0 // Support Protocol Regsister 6
+#define GMAC_SPR7 0x00C4 // Support Protocol Regsister 7
+#define GMAC_AHB_WEIGHT_REG 0x00C8 // GMAC Hash/Rx/Tx AHB Weighting register
+
+/**********************************************************************
+ * TOE GMAC 0/1 register
+ * #define TOE_GMAC0_BASE (TOE_BASE + 0xA000)
+ * #define TOE_GMAC1_BASE (TOE_BASE + 0xE000)
+ * Base 0x6000A000 or 0x6000E000
+ **********************************************************************/
+enum GMAC_REGISTER {
+ GMAC_STA_ADD0 = 0x0000,
+ GMAC_STA_ADD1 = 0x0004,
+ GMAC_STA_ADD2 = 0x0008,
+ GMAC_RX_FLTR = 0x000c,
+ GMAC_MCAST_FIL0 = 0x0010,
+ GMAC_MCAST_FIL1 = 0x0014,
+ GMAC_CONFIG0 = 0x0018,
+ GMAC_CONFIG1 = 0x001c,
+ GMAC_CONFIG2 = 0x0020,
+ GMAC_CONFIG3 = 0x0024,
+ GMAC_RESERVED = 0x0028,
+ GMAC_STATUS = 0x002c,
+ GMAC_IN_DISCARDS= 0x0030,
+ GMAC_IN_ERRORS = 0x0034,
+ GMAC_IN_MCAST = 0x0038,
+ GMAC_IN_BCAST = 0x003c,
+ GMAC_IN_MAC1 = 0x0040, // for STA 1 MAC Address
+ GMAC_IN_MAC2 = 0x0044 // for STA 2 MAC Address
+};
+/**********************************************************************
+ * TOE version Register (offset 0x0000)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int reserved : 15; // bit 31:17
+ unsigned int v_bit_mode : 1; // bit 16 1: 128-entry
+ unsigned int device_id : 12; // bit 15:4 Device ID
+ unsigned int revision_id : 4; // bit 3:0 Revision ID
+#else
+ unsigned int revision_id : 4; // bit 3:0 Revision ID
+ unsigned int device_id : 12; // bit 15:4 Device ID
+ unsigned int v_bit_mode : 1; // bit 16 1: 128-entry
+ unsigned int reserved : 15; // bit 31:17
+#endif
+ } bits;
+} TOE_VERSION_T;
+
+
+/**********************************************************************
+ * DMA Queues description Ring Base Address/Size Register (offset 0x0004)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int base_size;
+} DMA_Q_BASE_SIZE_T;
+#define DMA_Q_BASE_MASK (~0x0f)
+
+/**********************************************************************
+ * DMA SKB Buffer register (offset 0x0008)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0008
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int hw_skb_size : 16; // bit 31:16 HW Free poll SKB Size
+ unsigned int sw_skb_size : 16; // bit 15:0 SW Free poll SKB Size
+#else
+ unsigned int sw_skb_size : 16; // bit 15:0 SW Free poll SKB Size
+ unsigned int hw_skb_size : 16; // bit 31:16 HW Free poll SKB Size
+#endif
+ } bits;
+} DMA_SKB_SIZE_T;
+
+/**********************************************************************
+ * DMA SW Free Queue Read/Write Pointer Register (offset 0x000C)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_000c
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int wptr : 16; // bit 31:16 Write Ptr, RW
+ unsigned int rptr : 16; // bit 15:0 Read Ptr, RO
+#else
+ unsigned int rptr : 16; // bit 15:0 Read Ptr, RO
+ unsigned int wptr : 16; // bit 31:16 Write Ptr, RW
+#endif
+ } bits;
+} DMA_RWPTR_T;
+
+/**********************************************************************
+ * DMA HW Free Queue Read/Write Pointer Register (offset 0x0010)
+ **********************************************************************/
+// see DMA_RWPTR_T structure
+
+/**********************************************************************
+ * Interrupt Status Register 0 (offset 0x0020)
+ * Interrupt Mask Register 0 (offset 0x0024)
+ * Interrupt Select Register 0 (offset 0x0028)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0020
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int txDerr1 : 1; // bit 31 GMAC1 AHB Bus Error while Tx
+ unsigned int txPerr1 : 1; // bit 30 GMAC1 Tx Descriptor Protocol Error
+ unsigned int txDerr0 : 1; // bit 29 GMAC0 AHB Bus Error while Tx
+ unsigned int txPerr0 : 1; // bit 28 GMAC0 Tx Descriptor Protocol Error
+ unsigned int rxDerr1 : 1; // bit 27 GMAC1 AHB Bus Error while Rx
+ unsigned int rxPerr1 : 1; // bit 26 GMAC1 Rx Descriptor Protocol Error
+ unsigned int rxDerr0 : 1; // bit 25 GMAC0 AHB Bus Error while Rx
+ unsigned int rxPerr0 : 1; // bit 24 GMAC0 Rx Descriptor Protocol Error
+ unsigned int swtq15_fin : 1; // bit 23 GMAC1 SW Tx Queue 5 Finish Interrupt
+ unsigned int swtq14_fin : 1; // bit 22 GMAC1 SW Tx Queue 4 Finish Interrupt
+ unsigned int swtq13_fin : 1; // bit 21 GMAC1 SW Tx Queue 3 Finish Interrupt
+ unsigned int swtq12_fin : 1; // bit 20 GMAC1 SW Tx Queue 2 Finish Interrupt
+ unsigned int swtq11_fin : 1; // bit 19 GMAC1 SW Tx Queue 1 Finish Interrupt
+ unsigned int swtq10_fin : 1; // bit 18 GMAC1 SW Tx Queue 0 Finish Interrupt
+ unsigned int swtq05_fin : 1; // bit 17 GMAC0 SW Tx Queue 5 Finish Interrupt
+ unsigned int swtq04_fin : 1; // bit 16 GMAC0 SW Tx Queue 4 Finish Interrupt
+ unsigned int swtq03_fin : 1; // bit 15 GMAC0 SW Tx Queue 3 Finish Interrupt
+ unsigned int swtq02_fin : 1; // bit 14 GMAC0 SW Tx Queue 2 Finish Interrupt
+ unsigned int swtq01_fin : 1; // bit 13 GMAC0 SW Tx Queue 1 Finish Interrupt
+ unsigned int swtq00_fin : 1; // bit 12 GMAC0 SW Tx Queue 0 Finish Interrupt
+ unsigned int swtq15_eof : 1; // bit 11 GMAC1 SW Tx Queue 5 EOF Interrupt
+ unsigned int swtq14_eof : 1; // bit 10 GMAC1 SW Tx Queue 4 EOF Interrupt
+ unsigned int swtq13_eof : 1; // bit 9 GMAC1 SW Tx Queue 3 EOF Interrupt
+ unsigned int swtq12_eof : 1; // bit 8 GMAC1 SW Tx Queue 2 EOF Interrupt
+ unsigned int swtq11_eof : 1; // bit 7 GMAC1 SW Tx Queue 1 EOF Interrupt
+ unsigned int swtq10_eof : 1; // bit 6 GMAC1 SW Tx Queue 0 EOF Interrupt
+ unsigned int swtq05_eof : 1; // bit 5 GMAC0 SW Tx Queue 5 EOF Interrupt
+ unsigned int swtq04_eof : 1; // bit 4 GMAC0 SW Tx Queue 4 EOF Interrupt
+ unsigned int swtq03_eof : 1; // bit 3 GMAC0 SW Tx Queue 3 EOF Interrupt
+ unsigned int swtq02_eof : 1; // bit 2 GMAC0 SW Tx Queue 2 EOF Interrupt
+ unsigned int swtq01_eof : 1; // bit 1 GMAC0 SW Tx Queue 1 EOF Interrupt
+ unsigned int swtq00_eof : 1; // bit 0 GMAC0 SW Tx Queue 0 EOF Interrupt
+#else
+ unsigned int swtq00_eof : 1; // bit 0 GMAC0 SW Tx Queue 0 EOF Interrupt
+ unsigned int swtq01_eof : 1; // bit 1 GMAC0 SW Tx Queue 1 EOF Interrupt
+ unsigned int swtq02_eof : 1; // bit 2 GMAC0 SW Tx Queue 2 EOF Interrupt
+ unsigned int swtq03_eof : 1; // bit 3 GMAC0 SW Tx Queue 3 EOF Interrupt
+ unsigned int swtq04_eof : 1; // bit 4 GMAC0 SW Tx Queue 4 EOF Interrupt
+ unsigned int swtq05_eof : 1; // bit 5 GMAC0 SW Tx Queue 5 EOF Interrupt
+ unsigned int swtq10_eof : 1; // bit 6 GMAC1 SW Tx Queue 0 EOF Interrupt
+ unsigned int swtq11_eof : 1; // bit 7 GMAC1 SW Tx Queue 1 EOF Interrupt
+ unsigned int swtq12_eof : 1; // bit 8 GMAC1 SW Tx Queue 2 EOF Interrupt
+ unsigned int swtq13_eof : 1; // bit 9 GMAC1 SW Tx Queue 3 EOF Interrupt
+ unsigned int swtq14_eof : 1; // bit 10 GMAC1 SW Tx Queue 4 EOF Interrupt
+ unsigned int swtq15_eof : 1; // bit 11 GMAC1 SW Tx Queue 5 EOF Interrupt
+ unsigned int swtq00_fin : 1; // bit 12 GMAC0 SW Tx Queue 0 Finish Interrupt
+ unsigned int swtq01_fin : 1; // bit 13 GMAC0 SW Tx Queue 1 Finish Interrupt
+ unsigned int swtq02_fin : 1; // bit 14 GMAC0 SW Tx Queue 2 Finish Interrupt
+ unsigned int swtq03_fin : 1; // bit 15 GMAC0 SW Tx Queue 3 Finish Interrupt
+ unsigned int swtq04_fin : 1; // bit 16 GMAC0 SW Tx Queue 4 Finish Interrupt
+ unsigned int swtq05_fin : 1; // bit 17 GMAC0 SW Tx Queue 5 Finish Interrupt
+ unsigned int swtq10_fin : 1; // bit 18 GMAC1 SW Tx Queue 0 Finish Interrupt
+ unsigned int swtq11_fin : 1; // bit 19 GMAC1 SW Tx Queue 1 Finish Interrupt
+ unsigned int swtq12_fin : 1; // bit 20 GMAC1 SW Tx Queue 2 Finish Interrupt
+ unsigned int swtq13_fin : 1; // bit 21 GMAC1 SW Tx Queue 3 Finish Interrupt
+ unsigned int swtq14_fin : 1; // bit 22 GMAC1 SW Tx Queue 4 Finish Interrupt
+ unsigned int swtq15_fin : 1; // bit 23 GMAC1 SW Tx Queue 5 Finish Interrupt
+ unsigned int rxPerr0 : 1; // bit 24 GMAC0 Rx Descriptor Protocol Error
+ unsigned int rxDerr0 : 1; // bit 25 GMAC0 AHB Bus Error while Rx
+ unsigned int rxPerr1 : 1; // bit 26 GMAC1 Rx Descriptor Protocol Error
+ unsigned int rxDerr1 : 1; // bit 27 GMAC1 AHB Bus Error while Rx
+ unsigned int txPerr0 : 1; // bit 28 GMAC0 Tx Descriptor Protocol Error
+ unsigned int txDerr0 : 1; // bit 29 GMAC0 AHB Bus Error while Tx
+ unsigned int txPerr1 : 1; // bit 30 GMAC1 Tx Descriptor Protocol Error
+ unsigned int txDerr1 : 1; // bit 31 GMAC1 AHB Bus Error while Tx
+#endif
+ } bits;
+} INTR_REG0_T;
+
+#define GMAC1_TXDERR_INT_BIT BIT(31)
+#define GMAC1_TXPERR_INT_BIT BIT(30)
+#define GMAC0_TXDERR_INT_BIT BIT(29)
+#define GMAC0_TXPERR_INT_BIT BIT(28)
+#define GMAC1_RXDERR_INT_BIT BIT(27)
+#define GMAC1_RXPERR_INT_BIT BIT(26)
+#define GMAC0_RXDERR_INT_BIT BIT(25)
+#define GMAC0_RXPERR_INT_BIT BIT(24)
+#define GMAC1_SWTQ15_FIN_INT_BIT BIT(23)
+#define GMAC1_SWTQ14_FIN_INT_BIT BIT(22)
+#define GMAC1_SWTQ13_FIN_INT_BIT BIT(21)
+#define GMAC1_SWTQ12_FIN_INT_BIT BIT(20)
+#define GMAC1_SWTQ11_FIN_INT_BIT BIT(19)
+#define GMAC1_SWTQ10_FIN_INT_BIT BIT(18)
+#define GMAC0_SWTQ05_FIN_INT_BIT BIT(17)
+#define GMAC0_SWTQ04_FIN_INT_BIT BIT(16)
+#define GMAC0_SWTQ03_FIN_INT_BIT BIT(15)
+#define GMAC0_SWTQ02_FIN_INT_BIT BIT(14)
+#define GMAC0_SWTQ01_FIN_INT_BIT BIT(13)
+#define GMAC0_SWTQ00_FIN_INT_BIT BIT(12)
+#define GMAC1_SWTQ15_EOF_INT_BIT BIT(11)
+#define GMAC1_SWTQ14_EOF_INT_BIT BIT(10)
+#define GMAC1_SWTQ13_EOF_INT_BIT BIT(9)
+#define GMAC1_SWTQ12_EOF_INT_BIT BIT(8)
+#define GMAC1_SWTQ11_EOF_INT_BIT BIT(7)
+#define GMAC1_SWTQ10_EOF_INT_BIT BIT(6)
+#define GMAC0_SWTQ05_EOF_INT_BIT BIT(5)
+#define GMAC0_SWTQ04_EOF_INT_BIT BIT(4)
+#define GMAC0_SWTQ03_EOF_INT_BIT BIT(3)
+#define GMAC0_SWTQ02_EOF_INT_BIT BIT(2)
+#define GMAC0_SWTQ01_EOF_INT_BIT BIT(1)
+#define GMAC0_SWTQ00_EOF_INT_BIT BIT(0)
+
+
+/**********************************************************************
+ * Interrupt Status Register 1 (offset 0x0030)
+ * Interrupt Mask Register 1 (offset 0x0034)
+ * Interrupt Select Register 1 (offset 0x0038)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0030
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int toe_iq3_full : 1; // bit 31 TOE Interrupt Queue 3 Full Interrupt
+ unsigned int toe_iq2_full : 1; // bit 30 TOE Interrupt Queue 2 Full Interrupt
+ unsigned int toe_iq1_full : 1; // bit 29 TOE Interrupt Queue 1 Full Interrupt
+ unsigned int toe_iq0_full : 1; // bit 28 TOE Interrupt Queue 0 Full Interrupt
+ unsigned int toe_iq3_intr : 1; // bit 27 TOE Interrupt Queue 3 with Interrupts
+ unsigned int toe_iq2_intr : 1; // bit 26 TOE Interrupt Queue 2 with Interrupts
+ unsigned int toe_iq1_intr : 1; // bit 25 TOE Interrupt Queue 1 with Interrupts
+ unsigned int toe_iq0_intr : 1; // bit 24 TOE Interrupt Queue 0 with Interrupts
+ unsigned int hwtq13_eof : 1; // bit 23 GMAC1 HW Tx Queue3 EOF Interrupt
+ unsigned int hwtq12_eof : 1; // bit 22 GMAC1 HW Tx Queue2 EOF Interrupt
+ unsigned int hwtq11_eof : 1; // bit 21 GMAC1 HW Tx Queue1 EOF Interrupt
+ unsigned int hwtq10_eof : 1; // bit 20 GMAC1 HW Tx Queue0 EOF Interrupt
+ unsigned int hwtq03_eof : 1; // bit 19 GMAC0 HW Tx Queue3 EOF Interrupt
+ unsigned int hwtq02_eof : 1; // bit 18 GMAC0 HW Tx Queue2 EOF Interrupt
+ unsigned int hwtq01_eof : 1; // bit 17 GMAC0 HW Tx Queue1 EOF Interrupt
+ unsigned int hwtq00_eof : 1; // bit 16 GMAC0 HW Tx Queue0 EOF Interrupt
+ unsigned int class_rx : 14; // bit 15:2 Classification Queue Rx Interrupt
+ unsigned int default_q1_eof : 1; // bit 1 Default Queue 1 EOF Interrupt
+ unsigned int default_q0_eof : 1; // bit 0 Default Queue 0 EOF Interrupt
+#else
+ unsigned int default_q0_eof : 1; // bit 0 Default Queue 0 EOF Interrupt
+ unsigned int default_q1_eof : 1; // bit 1 Default Queue 1 EOF Interrupt
+ unsigned int class_rx : 14; // bit 15:2 Classification Queue Rx Interrupt
+ unsigned int hwtq00_eof : 1; // bit 16 GMAC0 HW Tx Queue0 EOF Interrupt
+ unsigned int hwtq01_eof : 1; // bit 17 GMAC0 HW Tx Queue1 EOF Interrupt
+ unsigned int hwtq02_eof : 1; // bit 18 GMAC0 HW Tx Queue2 EOF Interrupt
+ unsigned int hwtq03_eof : 1; // bit 19 GMAC0 HW Tx Queue3 EOF Interrupt
+ unsigned int hwtq10_eof : 1; // bit 20 GMAC1 HW Tx Queue0 EOF Interrupt
+ unsigned int hwtq11_eof : 1; // bit 21 GMAC1 HW Tx Queue1 EOF Interrupt
+ unsigned int hwtq12_eof : 1; // bit 22 GMAC1 HW Tx Queue2 EOF Interrupt
+ unsigned int hwtq13_eof : 1; // bit 23 GMAC1 HW Tx Queue3 EOF Interrupt
+ unsigned int toe_iq0_intr : 1; // bit 24 TOE Interrupt Queue 0 with Interrupts
+ unsigned int toe_iq1_intr : 1; // bit 25 TOE Interrupt Queue 1 with Interrupts
+ unsigned int toe_iq2_intr : 1; // bit 26 TOE Interrupt Queue 2 with Interrupts
+ unsigned int toe_iq3_intr : 1; // bit 27 TOE Interrupt Queue 3 with Interrupts
+ unsigned int toe_iq0_full : 1; // bit 28 TOE Interrupt Queue 0 Full Interrupt
+ unsigned int toe_iq1_full : 1; // bit 29 TOE Interrupt Queue 1 Full Interrupt
+ unsigned int toe_iq2_full : 1; // bit 30 TOE Interrupt Queue 2 Full Interrupt
+ unsigned int toe_iq3_full : 1; // bit 31 TOE Interrupt Queue 3 Full Interrupt
+#endif
+ } bits;
+} INTR_REG1_T;
+
+#define TOE_IQ3_FULL_INT_BIT BIT(31)
+#define TOE_IQ2_FULL_INT_BIT BIT(30)
+#define TOE_IQ1_FULL_INT_BIT BIT(29)
+#define TOE_IQ0_FULL_INT_BIT BIT(28)
+#define TOE_IQ3_INT_BIT BIT(27)
+#define TOE_IQ2_INT_BIT BIT(26)
+#define TOE_IQ1_INT_BIT BIT(25)
+#define TOE_IQ0_INT_BIT BIT(24)
+#define GMAC1_HWTQ13_EOF_INT_BIT BIT(23)
+#define GMAC1_HWTQ12_EOF_INT_BIT BIT(22)
+#define GMAC1_HWTQ11_EOF_INT_BIT BIT(21)
+#define GMAC1_HWTQ10_EOF_INT_BIT BIT(20)
+#define GMAC0_HWTQ03_EOF_INT_BIT BIT(19)
+#define GMAC0_HWTQ02_EOF_INT_BIT BIT(18)
+#define GMAC0_HWTQ01_EOF_INT_BIT BIT(17)
+#define GMAC0_HWTQ00_EOF_INT_BIT BIT(16)
+#define CLASS_RX_INT_BIT(x) BIT((x+2))
+#define DEFAULT_Q1_INT_BIT BIT(1)
+#define DEFAULT_Q0_INT_BIT BIT(0)
+
+#define TOE_IQ_INT_BITS (TOE_IQ0_INT_BIT | TOE_IQ1_INT_BIT | \
+ TOE_IQ2_INT_BIT | TOE_IQ3_INT_BIT)
+#define TOE_IQ_FULL_BITS (TOE_IQ0_FULL_INT_BIT | TOE_IQ1_FULL_INT_BIT | \
+ TOE_IQ2_FULL_INT_BIT | TOE_IQ3_FULL_INT_BIT)
+#define TOE_IQ_ALL_BITS (TOE_IQ_INT_BITS | TOE_IQ_FULL_BITS)
+#define TOE_CLASS_RX_INT_BITS 0xfffc
+
+/**********************************************************************
+ * Interrupt Status Register 2 (offset 0x0040)
+ * Interrupt Mask Register 2 (offset 0x0044)
+ * Interrupt Select Register 2 (offset 0x0048)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0040
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int toe_q31_full : 1; // bit 31 TOE Queue 31 Full Interrupt
+ unsigned int toe_q30_full : 1; // bit 30 TOE Queue 30 Full Interrupt
+ unsigned int toe_q29_full : 1; // bit 29 TOE Queue 29 Full Interrupt
+ unsigned int toe_q28_full : 1; // bit 28 TOE Queue 28 Full Interrupt
+ unsigned int toe_q27_full : 1; // bit 27 TOE Queue 27 Full Interrupt
+ unsigned int toe_q26_full : 1; // bit 26 TOE Queue 26 Full Interrupt
+ unsigned int toe_q25_full : 1; // bit 25 TOE Queue 25 Full Interrupt
+ unsigned int toe_q24_full : 1; // bit 24 TOE Queue 24 Full Interrupt
+ unsigned int toe_q23_full : 1; // bit 23 TOE Queue 23 Full Interrupt
+ unsigned int toe_q22_full : 1; // bit 22 TOE Queue 22 Full Interrupt
+ unsigned int toe_q21_full : 1; // bit 21 TOE Queue 21 Full Interrupt
+ unsigned int toe_q20_full : 1; // bit 20 TOE Queue 20 Full Interrupt
+ unsigned int toe_q19_full : 1; // bit 19 TOE Queue 19 Full Interrupt
+ unsigned int toe_q18_full : 1; // bit 18 TOE Queue 18 Full Interrupt
+ unsigned int toe_q17_full : 1; // bit 17 TOE Queue 17 Full Interrupt
+ unsigned int toe_q16_full : 1; // bit 16 TOE Queue 16 Full Interrupt
+ unsigned int toe_q15_full : 1; // bit 15 TOE Queue 15 Full Interrupt
+ unsigned int toe_q14_full : 1; // bit 14 TOE Queue 14 Full Interrupt
+ unsigned int toe_q13_full : 1; // bit 13 TOE Queue 13 Full Interrupt
+ unsigned int toe_q12_full : 1; // bit 12 TOE Queue 12 Full Interrupt
+ unsigned int toe_q11_full : 1; // bit 11 TOE Queue 11 Full Interrupt
+ unsigned int toe_q10_full : 1; // bit 10 TOE Queue 10 Full Interrupt
+ unsigned int toe_q9_full : 1; // bit 9 TOE Queue 9 Full Interrupt
+ unsigned int toe_q8_full : 1; // bit 8 TOE Queue 8 Full Interrupt
+ unsigned int toe_q7_full : 1; // bit 7 TOE Queue 7 Full Interrupt
+ unsigned int toe_q6_full : 1; // bit 6 TOE Queue 6 Full Interrupt
+ unsigned int toe_q5_full : 1; // bit 5 TOE Queue 5 Full Interrupt
+ unsigned int toe_q4_full : 1; // bit 4 TOE Queue 4 Full Interrupt
+ unsigned int toe_q3_full : 1; // bit 3 TOE Queue 3 Full Interrupt
+ unsigned int toe_q2_full : 1; // bit 2 TOE Queue 2 Full Interrupt
+ unsigned int toe_q1_full : 1; // bit 1 TOE Queue 1 Full Interrupt
+ unsigned int toe_q0_full : 1; // bit 0 TOE Queue 0 Full Interrupt
+#else
+ unsigned int toe_q0_full : 1; // bit 0 TOE Queue 0 Full Interrupt
+ unsigned int toe_q1_full : 1; // bit 1 TOE Queue 1 Full Interrupt
+ unsigned int toe_q2_full : 1; // bit 2 TOE Queue 2 Full Interrupt
+ unsigned int toe_q3_full : 1; // bit 3 TOE Queue 3 Full Interrupt
+ unsigned int toe_q4_full : 1; // bit 4 TOE Queue 4 Full Interrupt
+ unsigned int toe_q5_full : 1; // bit 5 TOE Queue 5 Full Interrupt
+ unsigned int toe_q6_full : 1; // bit 6 TOE Queue 6 Full Interrupt
+ unsigned int toe_q7_full : 1; // bit 7 TOE Queue 7 Full Interrupt
+ unsigned int toe_q8_full : 1; // bit 8 TOE Queue 8 Full Interrupt
+ unsigned int toe_q9_full : 1; // bit 9 TOE Queue 9 Full Interrupt
+ unsigned int toe_q10_full : 1; // bit 10 TOE Queue 10 Full Interrupt
+ unsigned int toe_q11_full : 1; // bit 11 TOE Queue 11 Full Interrupt
+ unsigned int toe_q12_full : 1; // bit 12 TOE Queue 12 Full Interrupt
+ unsigned int toe_q13_full : 1; // bit 13 TOE Queue 13 Full Interrupt
+ unsigned int toe_q14_full : 1; // bit 14 TOE Queue 14 Full Interrupt
+ unsigned int toe_q15_full : 1; // bit 15 TOE Queue 15 Full Interrupt
+ unsigned int toe_q16_full : 1; // bit 16 TOE Queue 16 Full Interrupt
+ unsigned int toe_q17_full : 1; // bit 17 TOE Queue 17 Full Interrupt
+ unsigned int toe_q18_full : 1; // bit 18 TOE Queue 18 Full Interrupt
+ unsigned int toe_q19_full : 1; // bit 19 TOE Queue 19 Full Interrupt
+ unsigned int toe_q20_full : 1; // bit 20 TOE Queue 20 Full Interrupt
+ unsigned int toe_q21_full : 1; // bit 21 TOE Queue 21 Full Interrupt
+ unsigned int toe_q22_full : 1; // bit 22 TOE Queue 22 Full Interrupt
+ unsigned int toe_q23_full : 1; // bit 23 TOE Queue 23 Full Interrupt
+ unsigned int toe_q24_full : 1; // bit 24 TOE Queue 24 Full Interrupt
+ unsigned int toe_q25_full : 1; // bit 25 TOE Queue 25 Full Interrupt
+ unsigned int toe_q26_full : 1; // bit 26 TOE Queue 26 Full Interrupt
+ unsigned int toe_q27_full : 1; // bit 27 TOE Queue 27 Full Interrupt
+ unsigned int toe_q28_full : 1; // bit 28 TOE Queue 28 Full Interrupt
+ unsigned int toe_q29_full : 1; // bit 29 TOE Queue 29 Full Interrupt
+ unsigned int toe_q30_full : 1; // bit 30 TOE Queue 30 Full Interrupt
+ unsigned int toe_q31_full : 1; // bit 31 TOE Queue 31 Full Interrupt
+#endif
+ } bits;
+} INTR_REG2_T;
+
+#define TOE_QL_FULL_INT_BIT(x) BIT(x)
+
+/**********************************************************************
+ * Interrupt Status Register 3 (offset 0x0050)
+ * Interrupt Mask Register 3 (offset 0x0054)
+ * Interrupt Select Register 3 (offset 0x0058)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0050
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int toe_q63_full : 1; // bit 63 TOE Queue 63 Full Interrupt
+ unsigned int toe_q62_full : 1; // bit 62 TOE Queue 62 Full Interrupt
+ unsigned int toe_q61_full : 1; // bit 61 TOE Queue 61 Full Interrupt
+ unsigned int toe_q60_full : 1; // bit 60 TOE Queue 60 Full Interrupt
+ unsigned int toe_q59_full : 1; // bit 59 TOE Queue 59 Full Interrupt
+ unsigned int toe_q58_full : 1; // bit 58 TOE Queue 58 Full Interrupt
+ unsigned int toe_q57_full : 1; // bit 57 TOE Queue 57 Full Interrupt
+ unsigned int toe_q56_full : 1; // bit 56 TOE Queue 56 Full Interrupt
+ unsigned int toe_q55_full : 1; // bit 55 TOE Queue 55 Full Interrupt
+ unsigned int toe_q54_full : 1; // bit 54 TOE Queue 54 Full Interrupt
+ unsigned int toe_q53_full : 1; // bit 53 TOE Queue 53 Full Interrupt
+ unsigned int toe_q52_full : 1; // bit 52 TOE Queue 52 Full Interrupt
+ unsigned int toe_q51_full : 1; // bit 51 TOE Queue 51 Full Interrupt
+ unsigned int toe_q50_full : 1; // bit 50 TOE Queue 50 Full Interrupt
+ unsigned int toe_q49_full : 1; // bit 49 TOE Queue 49 Full Interrupt
+ unsigned int toe_q48_full : 1; // bit 48 TOE Queue 48 Full Interrupt
+ unsigned int toe_q47_full : 1; // bit 47 TOE Queue 47 Full Interrupt
+ unsigned int toe_q46_full : 1; // bit 46 TOE Queue 46 Full Interrupt
+ unsigned int toe_q45_full : 1; // bit 45 TOE Queue 45 Full Interrupt
+ unsigned int toe_q44_full : 1; // bit 44 TOE Queue 44 Full Interrupt
+ unsigned int toe_q43_full : 1; // bit 43 TOE Queue 43 Full Interrupt
+ unsigned int toe_q42_full : 1; // bit 42 TOE Queue 42 Full Interrupt
+ unsigned int toe_q41_full : 1; // bit 41 TOE Queue 41 Full Interrupt
+ unsigned int toe_q40_full : 1; // bit 40 TOE Queue 40 Full Interrupt
+ unsigned int toe_q39_full : 1; // bit 39 TOE Queue 39 Full Interrupt
+ unsigned int toe_q38_full : 1; // bit 38 TOE Queue 38 Full Interrupt
+ unsigned int toe_q37_full : 1; // bit 37 TOE Queue 37 Full Interrupt
+ unsigned int toe_q36_full : 1; // bit 36 TOE Queue 36 Full Interrupt
+ unsigned int toe_q35_full : 1; // bit 35 TOE Queue 35 Full Interrupt
+ unsigned int toe_q34_full : 1; // bit 34 TOE Queue 34 Full Interrupt
+ unsigned int toe_q33_full : 1; // bit 33 TOE Queue 33 Full Interrupt
+ unsigned int toe_q32_full : 1; // bit 32 TOE Queue 32 Full Interrupt
+#else
+ unsigned int toe_q32_full : 1; // bit 32 TOE Queue 32 Full Interrupt
+ unsigned int toe_q33_full : 1; // bit 33 TOE Queue 33 Full Interrupt
+ unsigned int toe_q34_full : 1; // bit 34 TOE Queue 34 Full Interrupt
+ unsigned int toe_q35_full : 1; // bit 35 TOE Queue 35 Full Interrupt
+ unsigned int toe_q36_full : 1; // bit 36 TOE Queue 36 Full Interrupt
+ unsigned int toe_q37_full : 1; // bit 37 TOE Queue 37 Full Interrupt
+ unsigned int toe_q38_full : 1; // bit 38 TOE Queue 38 Full Interrupt
+ unsigned int toe_q39_full : 1; // bit 39 TOE Queue 39 Full Interrupt
+ unsigned int toe_q40_full : 1; // bit 40 TOE Queue 40 Full Interrupt
+ unsigned int toe_q41_full : 1; // bit 41 TOE Queue 41 Full Interrupt
+ unsigned int toe_q42_full : 1; // bit 42 TOE Queue 42 Full Interrupt
+ unsigned int toe_q43_full : 1; // bit 43 TOE Queue 43 Full Interrupt
+ unsigned int toe_q44_full : 1; // bit 44 TOE Queue 44 Full Interrupt
+ unsigned int toe_q45_full : 1; // bit 45 TOE Queue 45 Full Interrupt
+ unsigned int toe_q46_full : 1; // bit 46 TOE Queue 46 Full Interrupt
+ unsigned int toe_q47_full : 1; // bit 47 TOE Queue 47 Full Interrupt
+ unsigned int toe_q48_full : 1; // bit 48 TOE Queue 48 Full Interrupt
+ unsigned int toe_q49_full : 1; // bit 49 TOE Queue 49 Full Interrupt
+ unsigned int toe_q50_full : 1; // bit 50 TOE Queue 50 Full Interrupt
+ unsigned int toe_q51_full : 1; // bit 51 TOE Queue 51 Full Interrupt
+ unsigned int toe_q52_full : 1; // bit 52 TOE Queue 52 Full Interrupt
+ unsigned int toe_q53_full : 1; // bit 53 TOE Queue 53 Full Interrupt
+ unsigned int toe_q54_full : 1; // bit 54 TOE Queue 54 Full Interrupt
+ unsigned int toe_q55_full : 1; // bit 55 TOE Queue 55 Full Interrupt
+ unsigned int toe_q56_full : 1; // bit 56 TOE Queue 56 Full Interrupt
+ unsigned int toe_q57_full : 1; // bit 57 TOE Queue 57 Full Interrupt
+ unsigned int toe_q58_full : 1; // bit 58 TOE Queue 58 Full Interrupt
+ unsigned int toe_q59_full : 1; // bit 59 TOE Queue 59 Full Interrupt
+ unsigned int toe_q60_full : 1; // bit 60 TOE Queue 60 Full Interrupt
+ unsigned int toe_q61_full : 1; // bit 61 TOE Queue 61 Full Interrupt
+ unsigned int toe_q62_full : 1; // bit 62 TOE Queue 62 Full Interrupt
+ unsigned int toe_q63_full : 1; // bit 63 TOE Queue 63 Full Interrupt
+#endif
+ } bits;
+} INTR_REG3_T;
+
+#define TOE_QH_FULL_INT_BIT(x) BIT(x-32)
+
+/**********************************************************************
+ * Interrupt Status Register 4 (offset 0x0060)
+ * Interrupt Mask Register 4 (offset 0x0064)
+ * Interrupt Select Register 4 (offset 0x0068)
+ **********************************************************************/
+typedef union
+{
+ unsigned char byte;
+ struct bit_0060
+ {
+#if (BIG_ENDIAN==1)
+ unsigned char reserved : 1; //
+ unsigned char cnt_full : 1; // MIB counters half full interrupt
+ unsigned char rx_pause_on : 1; // received pause on frame interrupt
+ unsigned char tx_pause_on : 1; // transmit pause on frame interrupt
+ unsigned char rx_pause_off : 1; // received pause off frame interrupt
+ unsigned char tx_pause_off : 1; // received pause off frame interrupt
+ unsigned char rx_overrun : 1; // GMAC Rx FIFO overrun interrupt
+ unsigned char status_changed: 1; // Status Changed Intr for RGMII Mode
+#else
+ unsigned char status_changed: 1; // Status Changed Intr for RGMII Mode
+ unsigned char rx_overrun : 1; // GMAC Rx FIFO overrun interrupt
+ unsigned char tx_pause_off : 1; // received pause off frame interrupt
+ unsigned char rx_pause_off : 1; // received pause off frame interrupt
+ unsigned char tx_pause_on : 1; // transmit pause on frame interrupt
+ unsigned char rx_pause_on : 1; // received pause on frame interrupt
+ unsigned char cnt_full : 1; // MIB counters half full interrupt
+ unsigned char reserved : 1; //
+#endif
+ } _PACKED_ bits;
+} _PACKED_ GMAC_INTR_T;
+
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0060_2
+ {
+#if (BIG_ENDIAN==1)
+ GMAC_INTR_T gmac1;
+ GMAC_INTR_T gmac0;
+ unsigned int class_qf_int: 14; // bit 15:2 Classification Rx Queue13-0 Full Intr.
+ unsigned int hwfq_empty : 1; // bit 1 Hardware Free Queue Empty Intr.
+ unsigned int swfq_empty : 1; // bit 0 Software Free Queue Empty Intr.
+#else
+#endif
+ unsigned int swfq_empty : 1; // bit 0 Software Free Queue Empty Intr.
+ unsigned int hwfq_empty : 1; // bit 1 Hardware Free Queue Empty Intr.
+ unsigned int class_qf_int: 14; // bit 15:2 Classification Rx Queue13-0 Full Intr.
+ GMAC_INTR_T gmac0;
+ GMAC_INTR_T gmac1;
+ } bits;
+} INTR_REG4_T;
+
+#define GMAC1_RESERVED_INT_BIT BIT(31)
+#define GMAC1_MIB_INT_BIT BIT(30)
+#define GMAC1_RX_PAUSE_ON_INT_BIT BIT(29)
+#define GMAC1_TX_PAUSE_ON_INT_BIT BIT(28)
+#define GMAC1_RX_PAUSE_OFF_INT_BIT BIT(27)
+#define GMAC1_TX_PAUSE_OFF_INT_BIT BIT(26)
+#define GMAC1_RX_OVERRUN_INT_BIT BIT(25)
+#define GMAC1_STATUS_CHANGE_INT_BIT BIT(24)
+#define GMAC0_RESERVED_INT_BIT BIT(23)
+#define GMAC0_MIB_INT_BIT BIT(22)
+#define GMAC0_RX_PAUSE_ON_INT_BIT BIT(21)
+#define GMAC0_TX_PAUSE_ON_INT_BIT BIT(20)
+#define GMAC0_RX_PAUSE_OFF_INT_BIT BIT(19)
+#define GMAC0_TX_PAUSE_OFF_INT_BIT BIT(18)
+#define GMAC0_RX_OVERRUN_INT_BIT BIT(17)
+#define GMAC0_STATUS_CHANGE_INT_BIT BIT(16)
+#define CLASS_RX_FULL_INT_BIT(x) BIT((x+2))
+#define HWFQ_EMPTY_INT_BIT BIT(1)
+#define SWFQ_EMPTY_INT_BIT BIT(0)
+
+#if 1
+#define GMAC0_INT_BITS (GMAC0_MIB_INT_BIT)
+#define GMAC1_INT_BITS (GMAC1_MIB_INT_BIT)
+#else
+#define GMAC0_INT_BITS (GMAC0_RESERVED_INT_BIT | GMAC0_MIB_INT_BIT | \
+ GMAC0_RX_PAUSE_ON_INT_BIT | GMAC0_TX_PAUSE_ON_INT_BIT | \
+ GMAC0_RX_PAUSE_OFF_INT_BIT | GMAC0_TX_PAUSE_OFF_INT_BIT | \
+ GMAC0_RX_OVERRUN_INT_BIT | GMAC0_STATUS_CHANGE_INT_BIT)
+#define GMAC1_INT_BITS (GMAC1_RESERVED_INT_BIT | GMAC1_MIB_INT_BIT | \
+ GMAC1_RX_PAUSE_ON_INT_BIT | GMAC1_TX_PAUSE_ON_INT_BIT | \
+ GMAC1_RX_PAUSE_OFF_INT_BIT | GMAC1_TX_PAUSE_OFF_INT_BIT | \
+ GMAC1_RX_OVERRUN_INT_BIT | GMAC1_STATUS_CHANGE_INT_BIT)
+#endif
+
+#define CLASS_RX_FULL_INT_BITS 0xfffc
+
+/**********************************************************************
+ * GLOBAL_QUEUE_THRESHOLD_REG (offset 0x0070)
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_0070_2
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int toe_class : 8; // 31:24
+ unsigned int intrq : 8; // 23:16
+ unsigned int hwfq_empty : 8; // 15:8 Hardware Free Queue Empty Threshold
+ unsigned int swfq_empty : 8; // 7:0 Software Free Queue Empty Threshold
+#else
+#endif
+ unsigned int swfq_empty : 8; // 7:0 Software Free Queue Empty Threshold
+ unsigned int hwfq_empty : 8; // 15:8 Hardware Free Queue Empty Threshold
+ unsigned int intrq : 8; // 23:16
+ unsigned int toe_class : 8; // 31:24
+ } bits;
+} QUEUE_THRESHOLD_T;
+
+
+/**********************************************************************
+ * GMAC DMA Control Register
+ * GMAC0 offset 0x8000
+ * GMAC1 offset 0xC000
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8000
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rd_enable : 1; // bit 31 Rx DMA Enable
+ unsigned int td_enable : 1; // bit 30 Tx DMA Enable
+ unsigned int loopback : 1; // bit 29 Loopback TxDMA to RxDMA
+ unsigned int drop_small_ack : 1; // bit 28 1: Drop, 0: Accept
+ unsigned int reserved : 10; // bit 27:18
+ unsigned int rd_insert_bytes : 2; // bit 17:16
+ unsigned int rd_prot : 4; // bit 15:12 DMA Protection Control
+ unsigned int rd_burst_size : 2; // bit 11:10 DMA max burst size for every AHB request
+ unsigned int rd_bus : 2; // bit 9:8 Peripheral Bus Width
+ unsigned int td_prot : 4; // bit 7:4 TxDMA protection control
+ unsigned int td_burst_size : 2; // bit 3:2 TxDMA max burst size for every AHB request
+ unsigned int td_bus : 2; // bit 1:0 Peripheral Bus Width
+#else
+ unsigned int td_bus : 2; // bit 1:0 Peripheral Bus Width
+ unsigned int td_burst_size : 2; // bit 3:2 TxDMA max burst size for every AHB request
+ unsigned int td_prot : 4; // bit 7:4 TxDMA protection control
+ unsigned int rd_bus : 2; // bit 9:8 Peripheral Bus Width
+ unsigned int rd_burst_size : 2; // bit 11:10 DMA max burst size for every AHB request
+ unsigned int rd_prot : 4; // bit 15:12 DMA Protection Control
+ unsigned int rd_insert_bytes : 2; // bit 17:16
+ unsigned int reserved : 10; // bit 27:18
+ unsigned int drop_small_ack : 1; // bit 28 1: Drop, 0: Accept
+ unsigned int loopback : 1; // bit 29 Loopback TxDMA to RxDMA
+ unsigned int td_enable : 1; // bit 30 Tx DMA Enable
+ unsigned int rd_enable : 1; // bit 31 Rx DMA Enable
+#endif
+ } bits;
+} GMAC_DMA_CTRL_T;
+
+/**********************************************************************
+ * GMAC Tx Weighting Control Register 0
+ * GMAC0 offset 0x8004
+ * GMAC1 offset 0xC004
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8004
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 8; // bit 31:24
+ unsigned int hw_tq3 : 6; // bit 23:18 HW TX Queue 0
+ unsigned int hw_tq2 : 6; // bit 17:12 HW TX Queue 1
+ unsigned int hw_tq1 : 6; // bit 11:6 HW TX Queue 2
+ unsigned int hw_tq0 : 6; // bit 5:0 HW TX Queue 3
+#else
+ unsigned int hw_tq0 : 6; // bit 5:0 HW TX Queue 3
+ unsigned int hw_tq1 : 6; // bit 11:6 HW TX Queue 2
+ unsigned int hw_tq2 : 6; // bit 17:12 HW TX Queue 1
+ unsigned int hw_tq3 : 6; // bit 23:18 HW TX Queue 0
+ unsigned int reserved : 8; // bit 31:24
+#endif
+ } bits;
+} GMAC_TX_WCR0_T; // Weighting Control Register 0
+
+/**********************************************************************
+ * GMAC Tx Weighting Control Register 1
+ * GMAC0 offset 0x8008
+ * GMAC1 offset 0xC008
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8008
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 2; // bit 31:30
+ unsigned int sw_tq5 : 5; // bit 29:25 SW TX Queue 5
+ unsigned int sw_tq4 : 5; // bit 24:20 SW TX Queue 4
+ unsigned int sw_tq3 : 5; // bit 19:15 SW TX Queue 3
+ unsigned int sw_tq2 : 5; // bit 14:10 SW TX Queue 2
+ unsigned int sw_tq1 : 5; // bit 9:5 SW TX Queue 1
+ unsigned int sw_tq0 : 5; // bit 4:0 SW TX Queue 0
+#else
+ unsigned int sw_tq0 : 5; // bit 4:0 SW TX Queue 0
+ unsigned int sw_tq1 : 5; // bit 9:5 SW TX Queue 1
+ unsigned int sw_tq2 : 5; // bit 14:10 SW TX Queue 2
+ unsigned int sw_tq3 : 5; // bit 19:15 SW TX Queue 3
+ unsigned int sw_tq4 : 5; // bit 24:20 SW TX Queue 4
+ unsigned int sw_tq5 : 5; // bit 29:25 SW TX Queue 5
+ unsigned int reserved : 2; // bit 31:30
+#endif
+ } bits;
+} GMAC_TX_WCR1_T; // Weighting Control Register 1
+
+/**********************************************************************
+ * Queue Read/Write Pointer
+ * GMAC SW TX Queue 0~5 Read/Write Pointer register
+ * GMAC0 offset 0x800C ~ 0x8020
+ * GMAC1 offset 0xC00C ~ 0xC020
+ * GMAC HW TX Queue 0~3 Read/Write Pointer register
+ * GMAC0 offset 0x8024 ~ 0x8030
+ * GMAC1 offset 0xC024 ~ 0xC030
+ **********************************************************************/
+// see DMA_RWPTR_T structure
+
+/**********************************************************************
+ * GMAC DMA Tx First Description Address Register
+ * GMAC0 offset 0x8038
+ * GMAC1 offset 0xC038
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8038
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int td_first_des_ptr : 28; // bit 31:4 first descriptor address
+ unsigned int td_busy : 1; // bit 3 1: TxDMA busy; 0: TxDMA idle
+ unsigned int reserved : 3;
+#else
+ unsigned int reserved : 3;
+ unsigned int td_busy : 1; // bit 3 1: TxDMA busy; 0: TxDMA idle
+ unsigned int td_first_des_ptr : 28; // bit 31:4 first descriptor address
+#endif
+ } bits;
+} GMAC_TXDMA_FIRST_DESC_T;
+
+/**********************************************************************
+ * GMAC DMA Tx Current Description Address Register
+ * GMAC0 offset 0x803C
+ * GMAC1 offset 0xC03C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_803C
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int td_curr_desc_ptr : 28; // bit 31:4 current descriptor address
+ unsigned int reserved : 4;
+#else
+ unsigned int reserved : 4;
+ unsigned int td_curr_desc_ptr : 28; // bit 31:4 current descriptor address
+#endif
+ } bits;
+} GMAC_TXDMA_CURR_DESC_T;
+
+/**********************************************************************
+ * GMAC DMA Tx Description Word 0 Register
+ * GMAC0 offset 0x8040
+ * GMAC1 offset 0xC040
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8040
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 1; // bit 31
+ unsigned int derr : 1; // bit 30 data error during processing this descriptor
+ unsigned int perr : 1; // bit 29 protocol error during processing this descriptor
+ unsigned int status_rvd : 6; // bit 28:23 Tx Status, Reserved bits
+ unsigned int status_tx_ok : 1; // bit 22 Tx Status, 1: Successful 0: Failed
+ unsigned int desc_count : 6; // bit 21:16 number of descriptors used for the current frame
+ unsigned int buffer_size : 16; // bit 15:0 Transfer size
+#else
+ unsigned int buffer_size : 16; // bit 15:0 Transfer size
+ unsigned int desc_count : 6; // bit 21:16 number of descriptors used for the current frame
+ unsigned int status_tx_ok : 1; // bit 22 Tx Status, 1: Successful 0: Failed
+ unsigned int status_rvd : 6; // bit 28:23 Tx Status, Reserved bits
+ unsigned int perr : 1; // bit 29 protocol error during processing this descriptor
+ unsigned int derr : 1; // bit 30 data error during processing this descriptor
+ unsigned int reserved : 1; // bit 31
+#endif
+ } bits;
+} GMAC_TXDESC_0_T;
+
+/**********************************************************************
+ * GMAC DMA Tx Description Word 1 Register
+ * GMAC0 offset 0x8044
+ * GMAC1 offset 0xC044
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct txdesc_word1
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 9; // bit 31:23 Tx Flag, Reserved
+ unsigned int ip_fixed_len: 1; // bit 22
+ unsigned int bypass_tss : 1; // bit 21
+ unsigned int udp_chksum : 1; // bit 20 UDP Checksum Enable
+ unsigned int tcp_chksum : 1; // bit 19 TCP Checksum Enable
+ unsigned int ipv6_enable : 1; // bit 18 IPV6 Tx Enable
+ unsigned int ip_chksum : 1; // bit 17 IPV4 Header Checksum Enable
+ unsigned int mtu_enable : 1; // bit 16 TSS segmentation use MTU setting
+ unsigned int byte_count : 16; // bit 15: 0 Tx Frame Byte Count
+#else
+ unsigned int byte_count : 16; // bit 15: 0 Tx Frame Byte Count
+ unsigned int mtu_enable : 1; // bit 16 TSS segmentation use MTU setting
+ unsigned int ip_chksum : 1; // bit 17 IPV4 Header Checksum Enable
+ unsigned int ipv6_enable : 1; // bit 18 IPV6 Tx Enable
+ unsigned int tcp_chksum : 1; // bit 19 TCP Checksum Enable
+ unsigned int udp_chksum : 1; // bit 20 UDP Checksum Enable
+ unsigned int bypass_tss : 1; // bit 21
+ unsigned int ip_fixed_len: 1; // bit 22
+ unsigned int reserved : 9; // bit 31:23 Tx Flag, Reserved
+#endif
+ } bits;
+} GMAC_TXDESC_1_T;
+
+#define TSS_IP_FIXED_LEN_BIT BIT(22)
+#define TSS_UDP_CHKSUM_BIT BIT(20)
+#define TSS_TCP_CHKSUM_BIT BIT(19)
+#define TSS_IPV6_ENABLE_BIT BIT(18)
+#define TSS_IP_CHKSUM_BIT BIT(17)
+#define TSS_MTU_ENABLE_BIT BIT(16)
+
+/**********************************************************************
+ * GMAC DMA Tx Description Word 2 Register
+ * GMAC0 offset 0x8048
+ * GMAC1 offset 0xC048
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int buf_adr;
+} GMAC_TXDESC_2_T;
+
+/**********************************************************************
+ * GMAC DMA Tx Description Word 3 Register
+ * GMAC0 offset 0x804C
+ * GMAC1 offset 0xC04C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct txdesc_word3
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int sof_eof : 2; // bit 31:30 11: only one, 10: first, 01: last, 00: linking
+ unsigned int eofie : 1; // bit 29 End of frame interrupt enable
+ unsigned int reserved : 18; // bit 28:11
+ unsigned int mtu_size : 11; // bit 10: 0 Tx Frame Byte Count
+#else
+ unsigned int mtu_size : 11; // bit 10: 0 Tx Frame Byte Count
+ unsigned int reserved : 18; // bit 28:11
+ unsigned int eofie : 1; // bit 29 End of frame interrupt enable
+ unsigned int sof_eof : 2; // bit 31:30 11: only one, 10: first, 01: last, 00: linking
+#endif
+ } bits;
+} GMAC_TXDESC_3_T;
+#define SOF_EOF_BIT_MASK 0x3fffffff
+#define SOF_BIT 0x80000000
+#define EOF_BIT 0x40000000
+#define EOFIE_BIT BIT(29)
+#define MTU_SIZE_BIT_MASK 0x7ff
+
+/**********************************************************************
+ * GMAC Tx Descriptor
+ **********************************************************************/
+typedef struct
+{
+ GMAC_TXDESC_0_T word0;
+ GMAC_TXDESC_1_T word1;
+ GMAC_TXDESC_2_T word2;
+ GMAC_TXDESC_3_T word3;
+} GMAC_TXDESC_T;
+
+
+/**********************************************************************
+ * GMAC DMA Rx First Description Address Register
+ * GMAC0 offset 0x8058
+ * GMAC1 offset 0xC058
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8058
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rd_first_des_ptr : 28; // bit 31:4 first descriptor address
+ unsigned int rd_busy : 1; // bit 3 1-RxDMA busy; 0-RxDMA idle
+ unsigned int reserved : 3; // bit 2:0
+#else
+ unsigned int reserved : 3; // bit 2:0
+ unsigned int rd_busy : 1; // bit 3 1-RxDMA busy; 0-RxDMA idle
+ unsigned int rd_first_des_ptr : 28; // bit 31:4 first descriptor address
+#endif
+ } bits;
+} GMAC_RXDMA_FIRST_DESC_T;
+
+/**********************************************************************
+ * GMAC DMA Rx Current Description Address Register
+ * GMAC0 offset 0x805C
+ * GMAC1 offset 0xC05C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_805C
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rd_curr_des_ptr : 28; // bit 31:4 current descriptor address
+ unsigned int reserved : 4; // bit 3:0
+#else
+ unsigned int reserved : 4; // bit 3:0
+ unsigned int rd_curr_des_ptr : 28; // bit 31:4 current descriptor address
+#endif
+ } bits;
+} GMAC_RXDMA_CURR_DESC_T;
+
+/**********************************************************************
+ * GMAC DMA Rx Description Word 0 Register
+ * GMAC0 offset 0x8060
+ * GMAC1 offset 0xC060
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8060
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int drop : 1; // bit 31 TOE/CIS Queue Full dropped packet to default queue
+ unsigned int derr : 1; // bit 30 data error during processing this descriptor
+ unsigned int perr : 1; // bit 29 protocol error during processing this descriptor
+ unsigned int chksum_status : 3; // bit 28:26 Check Sum Status
+ unsigned int status : 4; // bit 24:22 Status of rx frame
+ unsigned int desc_count : 6; // bit 21:16 number of descriptors used for the current frame
+ unsigned int buffer_size : 16; // bit 15:0 number of descriptors used for the current frame
+#else
+ unsigned int buffer_size : 16; // bit 15:0 number of descriptors used for the current frame
+ unsigned int desc_count : 6; // bit 21:16 number of descriptors used for the current frame
+ unsigned int status : 4; // bit 24:22 Status of rx frame
+ unsigned int chksum_status : 3; // bit 28:26 Check Sum Status
+ unsigned int perr : 1; // bit 29 protocol error during processing this descriptor
+ unsigned int derr : 1; // bit 30 data error during processing this descriptor
+ unsigned int drop : 1; // bit 31 TOE/CIS Queue Full dropped packet to default queue
+#endif
+ } bits;
+} GMAC_RXDESC_0_T;
+
+#define GMAC_RXDESC_0_T_derr BIT(30)
+#define GMAC_RXDESC_0_T_perr BIT(29)
+#define GMAC_RXDESC_0_T_chksum_status(x) BIT((x+26))
+#define GMAC_RXDESC_0_T_status(x) BIT((x+22))
+#define GMAC_RXDESC_0_T_desc_count(x) BIT((x+16))
+
+#define RX_CHKSUM_IP_UDP_TCP_OK 0
+#define RX_CHKSUM_IP_OK_ONLY 1
+#define RX_CHKSUM_NONE 2
+#define RX_CHKSUM_IP_ERR_UNKNOWN 4
+#define RX_CHKSUM_IP_ERR 5
+#define RX_CHKSUM_TCP_UDP_ERR 6
+#define RX_CHKSUM_NUM 8
+
+#define RX_STATUS_GOOD_FRAME 0
+#define RX_STATUS_TOO_LONG_GOOD_CRC 1
+#define RX_STATUS_RUNT_FRAME 2
+#define RX_STATUS_SFD_NOT_FOUND 3
+#define RX_STATUS_CRC_ERROR 4
+#define RX_STATUS_TOO_LONG_BAD_CRC 5
+#define RX_STATUS_ALIGNMENT_ERROR 6
+#define RX_STATUS_TOO_LONG_BAD_ALIGN 7
+#define RX_STATUS_RX_ERR 8
+#define RX_STATUS_DA_FILTERED 9
+#define RX_STATUS_BUFFER_FULL 10
+#define RX_STATUS_NUM 16
+
+
+/**********************************************************************
+ * GMAC DMA Rx Description Word 1 Register
+ * GMAC0 offset 0x8064
+ * GMAC1 offset 0xC064
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct rxdesc_word1
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int sw_id : 16; // bit 31:16 Software ID
+ unsigned int byte_count : 16; // bit 15: 0 Rx Frame Byte Count
+#else
+ unsigned int byte_count : 16; // bit 15: 0 Rx Frame Byte Count
+ unsigned int sw_id : 16; // bit 31:16 Software ID
+#endif
+ } bits;
+} GMAC_RXDESC_1_T;
+
+/**********************************************************************
+ * GMAC DMA Rx Description Word 2 Register
+ * GMAC0 offset 0x8068
+ * GMAC1 offset 0xC068
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int buf_adr;
+} GMAC_RXDESC_2_T;
+
+#define RX_INSERT_NONE 0
+#define RX_INSERT_1_BYTE 1
+#define RX_INSERT_2_BYTE 2
+#define RX_INSERT_3_BYTE 3
+
+#define RX_INSERT_BYTES RX_INSERT_2_BYTE
+/**********************************************************************
+ * GMAC DMA Rx Description Word 3 Register
+ * GMAC0 offset 0x806C
+ * GMAC1 offset 0xC06C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct rxdesc_word3
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int sof_eof : 2; // bit 31:30 11: only one, 10: first, 01: last, 00: linking
+ unsigned int eofie : 1; // bit 29 End of frame interrupt enable
+ unsigned int ctrl_flag : 1; // bit 28 Control Flag is present
+ unsigned int out_of_seq : 1; // bit 27 Out of Sequence packet
+ unsigned int option : 1; // bit 26 IPV4 option or IPV6 extension header
+ unsigned int abnormal : 1; // bit 25 abnormal case found
+ unsigned int dup_ack : 1; // bit 24 Duplicated ACK detected
+ unsigned int l7_offset : 8; // bit 23: 16 L7 data offset
+ unsigned int l4_offset : 8; // bit 15: 8 L4 data offset
+ unsigned int l3_offset : 8; // bit 7: 0 L3 data offset
+#else
+ unsigned int l3_offset : 8; // bit 7: 0 L3 data offset
+ unsigned int l4_offset : 8; // bit 15: 8 L4 data offset
+ unsigned int l7_offset : 8; // bit 23: 16 L7 data offset
+ unsigned int dup_ack : 1; // bit 24 Duplicated ACK detected
+ unsigned int abnormal : 1; // bit 25 abnormal case found
+ unsigned int option : 1; // bit 26 IPV4 option or IPV6 extension header
+ unsigned int out_of_seq : 1; // bit 27 Out of Sequence packet
+ unsigned int ctrl_flag : 1; // bit 28 Control Flag is present
+ unsigned int eofie : 1; // bit 29 End of frame interrupt enable
+ unsigned int sof_eof : 2; // bit 31:30 11: only one, 10: first, 01: last, 00: linking
+#endif
+ } bits;
+} GMAC_RXDESC_3_T;
+
+/**********************************************************************
+ * GMAC Rx Descriptor
+ **********************************************************************/
+typedef struct
+{
+ GMAC_RXDESC_0_T word0;
+ GMAC_RXDESC_1_T word1;
+ GMAC_RXDESC_2_T word2;
+ GMAC_RXDESC_3_T word3;
+} GMAC_RXDESC_T;
+
+/**********************************************************************
+ * GMAC Hash Engine Enable/Action Register 0 Offset Register
+ * GMAC0 offset 0x8070
+ * GMAC1 offset 0xC070
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8070
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int mr1en : 1; // bit 31 Enable Matching Rule 1
+ unsigned int reserved1 : 1; // bit 30
+ unsigned int timing : 3; // bit 29:27
+ unsigned int mr1_action : 5; // bit 26:22 Matching Rule 1 action offset
+ unsigned int mr1hel : 6; // bit 21:16 match rule 1 hash entry size
+ unsigned int mr0en : 1; // bit 15 Enable Matching Rule 0
+ unsigned int reserved0 : 4; // bit 14:11
+ unsigned int mr0_action : 5; // bit 10:6 Matching Rule 0 action offset
+ unsigned int mr0hel : 6; // bit 5:0 match rule 0 hash entry size
+#else
+ unsigned int mr0hel : 6; // bit 5:0 match rule 0 hash entry size
+ unsigned int mr0_action : 5; // bit 10:6 Matching Rule 0 action offset
+ unsigned int reserved0 : 4; // bit 14:11
+ unsigned int mr0en : 1; // bit 15 Enable Matching Rule 0
+ unsigned int mr1hel : 6; // bit 21:16 match rule 1 hash entry size
+ unsigned int mr1_action : 5; // bit 26:22 Matching Rule 1 action offset
+ unsigned int timing : 3; // bit 29:27
+ unsigned int reserved1 : 1; // bit 30
+ unsigned int mr1en : 1; // bit 31 Enable Matching Rule 1
+#endif
+ } bits;
+} GMAC_HASH_ENABLE_REG0_T;
+
+/**********************************************************************
+ * GMAC Hash Engine Enable/Action Register 1 Offset Register
+ * GMAC0 offset 0x8074
+ * GMAC1 offset 0xC074
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8074
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int mr3en : 1; // bit 31 Enable Matching Rule 3
+ unsigned int reserved3 : 4; // bit 30:27
+ unsigned int mr3_action : 5; // bit 26:22 Matching Rule 3 action offset
+ unsigned int mr3hel : 6; // bit 21:16 match rule 3 hash entry size
+ unsigned int mr2en : 1; // bit 15 Enable Matching Rule 2
+ unsigned int reserved2 : 4; // bit 14:11
+ unsigned int mr2_action : 5; // bit 10:6 Matching Rule 2 action offset
+ unsigned int mr2hel : 6; // bit 5:0 match rule 2 hash entry size
+#else
+ unsigned int mr2hel : 6; // bit 5:0 match rule 2 hash entry size
+ unsigned int mr2_action : 5; // bit 10:6 Matching Rule 2 action offset
+ unsigned int reserved2 : 4; // bit 14:11
+ unsigned int mr2en : 1; // bit 15 Enable Matching Rule 2
+ unsigned int mr3hel : 6; // bit 21:16 match rule 3 hash entry size
+ unsigned int mr3_action : 5; // bit 26:22 Matching Rule 3 action offset
+ unsigned int reserved1 : 4; // bit 30:27
+ unsigned int mr3en : 1; // bit 31 Enable Matching Rule 3
+#endif
+ } bits;
+} GMAC_HASH_ENABLE_REG1_T;
+
+
+/**********************************************************************
+ * GMAC Matching Rule Control Register 0
+ * GMAC0 offset 0x8078
+ * GMAC1 offset 0xC078
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_8078
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int l2 : 1; // bit 31 L2 matching enable
+ unsigned int l3 : 1; // bit 30 L3 matching enable
+ unsigned int l4 : 1; // bit 29 L4 matching enable
+ unsigned int l7 : 1; // bit 28 L7 matching enable
+ unsigned int port : 1; // bit 27 PORT ID matching enable
+ unsigned int priority : 3; // bit 26:24 priority if multi-rules matched
+ unsigned int da : 1; // bit 23 MAC DA enable
+ unsigned int sa : 1; // bit 22 MAC SA enable
+ unsigned int ether_type : 1; // bit 21 Ethernet type enable
+ unsigned int vlan : 1; // bit 20 VLAN ID enable
+ unsigned int pppoe : 1; // bit 19 PPPoE Session ID enable
+ unsigned int reserved1 : 3; // bit 18:16
+ unsigned int ip_version : 1; // bit 15 0: IPV4, 1: IPV6
+ unsigned int ip_hdr_len : 1; // bit 14 IPV4 Header length
+ unsigned int flow_lable : 1; // bit 13 IPV6 Flow label
+ unsigned int tos_traffic : 1; // bit 12 IPV4 TOS or IPV6 Traffice Class
+ unsigned int reserved2 : 4; // bit 11:8
+ unsigned int sprx : 8; // bit 7:0 Support Protocol Register 7:0
+#else
+ unsigned int sprx : 8; // bit 7:0 Support Protocol Register 7:0
+ unsigned int reserved2 : 4; // bit 11:8
+ unsigned int tos_traffic : 1; // bit 12 IPV4 TOS or IPV6 Traffice Class
+ unsigned int flow_lable : 1; // bit 13 IPV6 Flow label
+ unsigned int ip_hdr_len : 1; // bit 14 IPV4 Header length
+ unsigned int ip_version : 1; // bit 15 0: IPV4, 1: IPV6
+ unsigned int reserved1 : 3; // bit 18:16
+ unsigned int pppoe : 1; // bit 19 PPPoE Session ID enable
+ unsigned int vlan : 1; // bit 20 VLAN ID enable
+ unsigned int ether_type : 1; // bit 21 Ethernet type enable
+ unsigned int sa : 1; // bit 22 MAC SA enable
+ unsigned int da : 1; // bit 23 MAC DA enable
+ unsigned int priority : 3; // bit 26:24 priority if multi-rules matched
+ unsigned int port : 1; // bit 27 PORT ID matching enable
+ unsigned int l7 : 1; // bit 28 L7 matching enable
+ unsigned int l4 : 1; // bit 29 L4 matching enable
+ unsigned int l3 : 1; // bit 30 L3 matching enable
+ unsigned int l2 : 1; // bit 31 L2 matching enable
+#endif
+ } bits;
+} GMAC_MRxCR0_T;
+
+#define MR_L2_BIT BIT(31)
+#define MR_L3_BIT BIT(30)
+#define MR_L4_BIT BIT(29)
+#define MR_L7_BIT BIT(28)
+#define MR_PORT_BIT BIT(27)
+#define MR_PRIORITY_BIT BIT(26)
+#define MR_DA_BIT BIT(23)
+#define MR_SA_BIT BIT(22)
+#define MR_ETHER_TYPE_BIT BIT(21)
+#define MR_VLAN_BIT BIT(20)
+#define MR_PPPOE_BIT BIT(19)
+#define MR_IP_VER_BIT BIT(15)
+#define MR_IP_HDR_LEN_BIT BIT(14)
+#define MR_FLOW_LABLE_BIT BIT(13)
+#define MR_TOS_TRAFFIC_BIT BIT(12)
+#define MR_SPR_BIT(x) BIT(x)
+#define MR_SPR_BITS 0xff
+
+/**********************************************************************
+ * GMAC Matching Rule Control Register 1
+ * GMAC0 offset 0x807C
+ * GMAC1 offset 0xC07C
+ **********************************************************************/
+ typedef union
+{
+ unsigned int bits32;
+ struct bit_807C
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int sip : 1; // bit 31 Srce IP
+ unsigned int sip_netmask : 7; // bit 30:24 Srce IP net mask, number of mask bits
+ unsigned int dip : 1; // bit 23 Dest IP
+ unsigned int dip_netmask : 7; // bit 22:16 Dest IP net mask, number of mask bits
+ unsigned int l4_byte0_15 : 16; // bit 15: 0
+#else
+ unsigned int l4_byte0_15 : 16; // bit 15: 0
+ unsigned int dip_netmask : 7; // bit 22:16 Dest IP net mask, number of mask bits
+ unsigned int dip : 1; // bit 23 Dest IP
+ unsigned int sip_netmask : 7; // bit 30:24 Srce IP net mask, number of mask bits
+ unsigned int sip : 1; // bit 31 Srce IP
+#endif
+ } bits;
+} GMAC_MRxCR1_T;
+
+/**********************************************************************
+ * GMAC Matching Rule Control Register 2
+ * GMAC0 offset 0x8080
+ * GMAC1 offset 0xC080
+ **********************************************************************/
+ typedef union
+{
+ unsigned int bits32;
+ struct bit_8080
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int l4_byte16_24: 8; // bit 31: 24
+ unsigned int l7_byte0_23 : 24; // bit 23:0
+#else
+ unsigned int l7_byte0_23 : 24; // bit 23:0
+ unsigned int l4_byte16_24: 8; // bit 31: 24
+#endif
+ } bits;
+} GMAC_MRxCR2_T;
+
+
+/**********************************************************************
+ * GMAC Support registers
+ * GMAC0 offset 0x80A8
+ * GMAC1 offset 0xC0A8
+ **********************************************************************/
+ typedef union
+{
+ unsigned int bits32;
+ struct bit_80A8
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved: 21; // bit 31:11
+ unsigned int swap : 3; // bit 10:8 Swap
+ unsigned int protocol: 8; // bit 7:0 Supported protocol
+#else
+ unsigned int protocol: 8; // bit 7:0 Supported protocol
+ unsigned int swap : 3; // bit 10:8 Swap
+ unsigned int reserved: 21; // bit 31:11
+#endif
+ } bits;
+} GMAC_SPR_T;
+
+/**********************************************************************
+ * GMAC_AHB_WEIGHT registers
+ * GMAC0 offset 0x80C8
+ * GMAC1 offset 0xC0C8
+ **********************************************************************/
+ typedef union
+{
+ unsigned int bits32;
+ struct bit_80C8
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 7; // 31:25
+ unsigned int tqDV_threshold : 5; // 24:20 DMA TqCtrl to Start tqDV FIFO Threshold
+ unsigned int pre_req : 5; // 19:15 Rx Data Pre Request FIFO Threshold
+ unsigned int tx_weight : 5; // 14:10
+ unsigned int rx_weight : 5; // 9:5
+ unsigned int hash_weight : 5; // 4:0
+#else
+ unsigned int hash_weight : 5; // 4:0
+ unsigned int rx_weight : 5; // 9:5
+ unsigned int tx_weight : 5; // 14:10
+ unsigned int pre_req : 5; // 19:15 Rx Data Pre Request FIFO Threshold
+ unsigned int tqDV_threshold : 5; // 24:20 DMA TqCtrl to Start tqDV FIFO Threshold
+ unsigned int reserved : 7; // 31:25
+#endif
+ } bits;
+} GMAC_AHB_WEIGHT_T;
+/**********************************************************************
+ * the register structure of GMAC
+ **********************************************************************/
+
+/**********************************************************************
+ * GMAC RX FLTR
+ * GMAC0 Offset 0xA00C
+ * GMAC1 Offset 0xE00C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_000c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 27;
+ unsigned int error : 1; /* enable receive of all error frames */
+ unsigned int promiscuous : 1; /* enable receive of all frames */
+ unsigned int broadcast : 1; /* enable receive of broadcast frames */
+ unsigned int multicast : 1; /* enable receive of multicast frames that pass multicast filter */
+ unsigned int unicast : 1; /* enable receive of unicast frames that are sent to STA address */
+#else
+ unsigned int unicast : 1; /* enable receive of unicast frames that are sent to STA address */
+ unsigned int multicast : 1; /* enable receive of multicast frames that pass multicast filter */
+ unsigned int broadcast : 1; /* enable receive of broadcast frames */
+ unsigned int promiscuous : 1; /* enable receive of all frames */
+ unsigned int error : 1; /* enable receive of all error frames */
+ unsigned int : 27;
+#endif
+ } bits;
+} GMAC_RX_FLTR_T;
+
+/**********************************************************************
+ * GMAC Configuration 0
+ * GMAC0 Offset 0xA018
+ * GMAC1 Offset 0xE018
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0018
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 2; // 31
+ unsigned int port1_chk_classq : 1; // 29
+ unsigned int port0_chk_classq : 1; // 28
+ unsigned int port1_chk_toeq : 1; // 27
+ unsigned int port0_chk_toeq : 1; // 26
+ unsigned int port1_chk_hwq : 1; // 25
+ unsigned int port0_chk_hwq : 1; // 24
+ unsigned int rx_err_detect : 1; // 23
+ unsigned int ipv6_exthdr_order: 1; // 22
+ unsigned int rxc_inv : 1; // 21
+ unsigned int rgmm_edge : 1; // 20
+ unsigned int rx_tag_remove : 1; /* 19: Remove Rx VLAN tag */
+ unsigned int ipv6_rx_chksum : 1; /* 18: IPv6 RX Checksum enable */
+ unsigned int ipv4_rx_chksum : 1; /* 17: IPv4 RX Checksum enable */
+ unsigned int rgmii_en : 1; /* 16: RGMII in-band status enable */
+ unsigned int tx_fc_en : 1; /* 15: TX flow control enable */
+ unsigned int rx_fc_en : 1; /* 14: RX flow control enable */
+ unsigned int sim_test : 1; /* 13: speed up timers in simulation */
+ unsigned int dis_col : 1; /* 12: disable 16 collisions abort function */
+ unsigned int dis_bkoff : 1; /* 11: disable back-off function */
+ unsigned int max_len : 3; /* 8-10 maximum receive frame length allowed */
+ unsigned int adj_ifg : 4; /* 4-7: adjust IFG from 96+/-56 */
+ unsigned int flow_ctrl : 1; /* 3: flow control also trigged by Rx queues */
+ unsigned int loop_back : 1; /* 2: transmit data loopback enable */
+ unsigned int dis_rx : 1; /* 1: disable receive */
+ unsigned int dis_tx : 1; /* 0: disable transmit */
+#else
+ unsigned int dis_tx : 1; /* 0: disable transmit */
+ unsigned int dis_rx : 1; /* 1: disable receive */
+ unsigned int loop_back : 1; /* 2: transmit data loopback enable */
+ unsigned int flow_ctrl : 1; /* 3: flow control also trigged by Rx queues */
+ unsigned int adj_ifg : 4; /* 4-7: adjust IFG from 96+/-56 */
+ unsigned int max_len : 3; /* 8-10 maximum receive frame length allowed */
+ unsigned int dis_bkoff : 1; /* 11: disable back-off function */
+ unsigned int dis_col : 1; /* 12: disable 16 collisions abort function */
+ unsigned int sim_test : 1; /* 13: speed up timers in simulation */
+ unsigned int rx_fc_en : 1; /* 14: RX flow control enable */
+ unsigned int tx_fc_en : 1; /* 15: TX flow control enable */
+ unsigned int rgmii_en : 1; /* 16: RGMII in-band status enable */
+ unsigned int ipv4_rx_chksum : 1; /* 17: IPv4 RX Checksum enable */
+ unsigned int ipv6_rx_chksum : 1; /* 18: IPv6 RX Checksum enable */
+ unsigned int rx_tag_remove : 1; /* 19: Remove Rx VLAN tag */
+ unsigned int rgmm_edge : 1; // 20
+ unsigned int rxc_inv : 1; // 21
+ unsigned int ipv6_exthdr_order: 1; // 22
+ unsigned int rx_err_detect : 1; // 23
+ unsigned int port0_chk_hwq : 1; // 24
+ unsigned int port1_chk_hwq : 1; // 25
+ unsigned int port0_chk_toeq : 1; // 26
+ unsigned int port1_chk_toeq : 1; // 27
+ unsigned int port0_chk_classq : 1; // 28
+ unsigned int port1_chk_classq : 1; // 29
+ unsigned int reserved : 2; // 31
+#endif
+ } bits;
+} GMAC_CONFIG0_T;
+
+/**********************************************************************
+ * GMAC Configuration 1
+ * GMAC0 Offset 0xA01C
+ * GMAC1 Offset 0xE01C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_001c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int reserved : 16;
+ unsigned int rel_threshold : 8; /* flow control release threshold */
+ unsigned int set_threshold : 8; /* flow control set threshold */
+#else
+ unsigned int set_threshold : 8; /* flow control set threshold */
+ unsigned int rel_threshold : 8; /* flow control release threshold */
+ unsigned int reserved : 16;
+#endif
+ } bits;
+} GMAC_CONFIG1_T;
+
+#define GMAC_FLOWCTRL_SET_MAX 32
+#define GMAC_FLOWCTRL_SET_MIN 0
+#define GMAC_FLOWCTRL_RELEASE_MAX 32
+#define GMAC_FLOWCTRL_RELEASE_MIN 0
+
+/**********************************************************************
+ * GMAC Configuration 2
+ * GMAC0 Offset 0xA020
+ * GMAC1 Offset 0xE020
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0020
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+ unsigned int set_threshold : 16; /* flow control set threshold */
+#else
+ unsigned int set_threshold : 16; /* flow control set threshold */
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+#endif
+ } bits;
+} GMAC_CONFIG2_T;
+
+/**********************************************************************
+ * GMAC Configuration 3
+ * GMAC0 Offset 0xA024
+ * GMAC1 Offset 0xE024
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_0024
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+ unsigned int set_threshold : 16; /* flow control set threshold */
+#else
+ unsigned int set_threshold : 16; /* flow control set threshold */
+ unsigned int rel_threshold : 16; /* flow control release threshold */
+#endif
+ } bits;
+} GMAC_CONFIG3_T;
+
+
+/**********************************************************************
+ * GMAC STATUS
+ * GMAC0 Offset 0xA02C
+ * GMAC1 Offset 0xE02C
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit1_002c
+ {
+#if (BIG_ENDIAN==1)
+ unsigned int : 25;
+ unsigned int mii_rmii : 2; /* PHY interface type */
+ unsigned int reserved : 1;
+ unsigned int duplex : 1; /* duplex mode */
+ unsigned int speed : 2; /* link speed(00->2.5M 01->25M 10->125M) */
+ unsigned int link : 1; /* link status */
+#else
+ unsigned int link : 1; /* link status */
+ unsigned int speed : 2; /* link speed(00->2.5M 01->25M 10->125M) */
+ unsigned int duplex : 1; /* duplex mode */
+ unsigned int reserved : 1;
+ unsigned int mii_rmii : 2; /* PHY interface type */
+ unsigned int : 25;
+#endif
+ } bits;
+} GMAC_STATUS_T;
+
+#define GMAC_SPEED_10 0
+#define GMAC_SPEED_100 1
+#define GMAC_SPEED_1000 2
+
+#define GMAC_PHY_MII 0
+#define GMAC_PHY_GMII 1
+#define GMAC_PHY_RGMII_100 2
+#define GMAC_PHY_RGMII_1000 3
+
+/**********************************************************************
+ * Queue Header
+ * (1) TOE Queue Header
+ * (2) Non-TOE Queue Header
+ * (3) Interrupt Queue Header
+ *
+ * memory Layout
+ * TOE Queue Header
+ * 0x60003000 +---------------------------+ 0x0000
+ * | TOE Queue 0 Header |
+ * | 8 * 4 Bytes |
+ * +---------------------------+ 0x0020
+ * | TOE Queue 1 Header |
+ * | 8 * 4 Bytes |
+ * +---------------------------+ 0x0040
+ * | ...... |
+ * | |
+ * +---------------------------+
+ *
+ * Non TOE Queue Header
+ * 0x60002000 +---------------------------+ 0x0000
+ * | Default Queue 0 Header |
+ * | 2 * 4 Bytes |
+ * +---------------------------+ 0x0008
+ * | Default Queue 1 Header |
+ * | 2 * 4 Bytes |
+ * +---------------------------+ 0x0010
+ * | Classification Queue 0 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+
+ * | Classification Queue 1 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+ (n * 8 + 0x10)
+ * | ... |
+ * | 2 * 4 Bytes |
+ * +---------------------------+ (13 * 8 + 0x10)
+ * | Classification Queue 13 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+ 0x80
+ * | Interrupt Queue 0 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+
+ * | Interrupt Queue 1 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+
+ * | Interrupt Queue 2 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+
+ * | Interrupt Queue 3 |
+ * | 2 * 4 Bytes |
+ * +---------------------------+
+ *
+ **********************************************************************/
+#define TOE_QUEUE_HDR_ADDR(n) (TOE_TOE_QUE_HDR_BASE + n * 32)
+#define TOE_Q_HDR_AREA_END (TOE_QUEUE_HDR_ADDR(TOE_TOE_QUEUE_MAX+1))
+#define TOE_DEFAULT_Q0_HDR_BASE (TOE_NONTOE_QUE_HDR_BASE + 0x00)
+#define TOE_DEFAULT_Q1_HDR_BASE (TOE_NONTOE_QUE_HDR_BASE + 0x08)
+#define TOE_CLASS_Q_HDR_BASE (TOE_NONTOE_QUE_HDR_BASE + 0x10)
+#define TOE_INTR_Q_HDR_BASE (TOE_NONTOE_QUE_HDR_BASE + 0x80)
+#define INTERRUPT_QUEUE_HDR_ADDR(n) (TOE_INTR_Q_HDR_BASE + n * 8)
+#define NONTOE_Q_HDR_AREA_END (INTERRUPT_QUEUE_HDR_ADDR(TOE_INTR_QUEUE_MAX+1))
+/**********************************************************************
+ * TOE Queue Header Word 0
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int base_size;
+} TOE_QHDR0_T;
+
+#define TOE_QHDR0_BASE_MASK (~0x0f)
+
+/**********************************************************************
+ * TOE Queue Header Word 1
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_qhdr1
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int wptr : 16; // bit 31:16
+ unsigned int rptr : 16; // bit 15:0
+#else
+ unsigned int rptr : 16; // bit 15:0
+ unsigned int wptr : 16; // bit 31:16
+#endif
+ } bits;
+} TOE_QHDR1_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 2
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_qhdr2
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int usd : 1; // bit 31 0: if no data assembled yet
+ unsigned int ctl : 1; // bit 30 1: have control flag bits (except ack)
+ unsigned int osq : 1; // bit 29 1: out of sequence
+ unsigned int sat : 1; // bit 28 1: SeqCnt > SeqThreshold, or AckCnt > AckThreshold
+ unsigned int ip_opt : 1; // bit 27 1: have IPV4 option or IPV6 Extension header
+ unsigned int tcp_opt : 1; // bit 26 1: Have TCP option
+ unsigned int abn : 1; // bit 25 1: Abnormal case Found
+ unsigned int dack : 1; // bit 24 1: Duplicated ACK
+ unsigned int reserved : 7; // bit 23:17
+ unsigned int TotalPktSize : 17; // bit 16: 0 Total packet size
+#else
+ unsigned int TotalPktSize : 17; // bit 16: 0 Total packet size
+ unsigned int reserved : 7; // bit 23:17
+ unsigned int dack : 1; // bit 24 1: Duplicated ACK
+ unsigned int abn : 1; // bit 25 1: Abnormal case Found
+ unsigned int tcp_opt : 1; // bit 26 1: Have TCP option
+ unsigned int ip_opt : 1; // bit 27 1: have IPV4 option or IPV6 Extension header
+ unsigned int sat : 1; // bit 28 1: SeqCnt > SeqThreshold, or AckCnt > AckThreshold
+ unsigned int osq : 1; // bit 29 1: out of sequence
+ unsigned int ctl : 1; // bit 30 1: have control flag bits (except ack)
+ unsigned int usd : 1; // bit 31 0: if no data assembled yet
+#endif
+ } bits;
+} TOE_QHDR2_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 3
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int seq_num;
+} TOE_QHDR3_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 4
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int ack_num;
+} TOE_QHDR4_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 5
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_qhdr5
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int SeqCnt : 16; // bit 31:16
+ unsigned int AckCnt : 16; // bit 15:0
+#else
+ unsigned int AckCnt : 16; // bit 15:0
+ unsigned int SeqCnt : 16; // bit 31:16
+#endif
+ } bits;
+} TOE_QHDR5_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 6
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_qhdr6
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int MaxPktSize : 14; // bit 31:18
+ unsigned int iq_num : 2; // bit 17:16
+ unsigned int WinSize : 16; // bit 15:0
+#else
+ unsigned int WinSize : 16; // bit 15:0
+ unsigned int iq_num : 2; // bit 17:16
+ unsigned int MaxPktSize : 14; // bit 31:18
+#endif
+ } bits;
+} TOE_QHDR6_T;
+
+/**********************************************************************
+ * TOE Queue Header Word 7
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_qhdr7
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int SeqThreshold : 16; // bit 31:16
+ unsigned int AckThreshold : 16; // bit 15:0
+#else
+ unsigned int AckThreshold : 16; // bit 15:0
+ unsigned int SeqThreshold : 16; // bit 31:16
+#endif
+ } bits;
+} TOE_QHDR7_T;
+
+/**********************************************************************
+ * TOE Queue Header
+ **********************************************************************/
+typedef struct
+{
+ TOE_QHDR0_T word0;
+ TOE_QHDR1_T word1;
+ TOE_QHDR2_T word2;
+ TOE_QHDR3_T word3;
+ TOE_QHDR4_T word4;
+ TOE_QHDR5_T word5;
+ TOE_QHDR6_T word6;
+ TOE_QHDR7_T word7;
+} TOE_QHDR_T;
+
+/**********************************************************************
+ * NONTOE Queue Header Word 0
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int base_size;
+} NONTOE_QHDR0_T;
+
+#define NONTOE_QHDR0_BASE_MASK (~0x0f)
+
+/**********************************************************************
+ * NONTOE Queue Header Word 1
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_nonqhdr1
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int wptr : 16; // bit 31:16
+ unsigned int rptr : 16; // bit 15:0
+#else
+ unsigned int rptr : 16; // bit 15:0
+ unsigned int wptr : 16; // bit 31:16
+#endif
+ } bits;
+} NONTOE_QHDR1_T;
+
+/**********************************************************************
+ * Non-TOE Queue Header
+ **********************************************************************/
+typedef struct
+{
+ NONTOE_QHDR0_T word0;
+ NONTOE_QHDR1_T word1;
+} NONTOE_QHDR_T;
+
+/**********************************************************************
+ * Interrupt Queue Header Word 0
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_intrqhdr0
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int wptr : 16; // bit 31:16 Write Pointer where hw stopped
+ unsigned int win_size : 16; // bit 15:0 Descriptor Ring Size
+#else
+ unsigned int win_size : 16; // bit 15:0 Descriptor Ring Size
+ unsigned int wptr : 16; // bit 31:16 Write Pointer where hw stopped
+#endif
+ } bits;
+} INTR_QHDR0_T;
+
+/**********************************************************************
+ * Interrupt Queue Header Word 1
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_intrqhdr1
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int ctl : 1; // bit 31 1: have control flag bits (except ack)
+ unsigned int osq : 1; // bit 30 1: out of sequence
+ unsigned int sat : 1; // bit 29 1: SeqCnt > SeqThreshold, or AckCnt > AckThreshold
+ unsigned int ip_opt : 1; // bit 28 1: have IPV4 option or IPV6 Extension header
+ unsigned int tcp_opt : 1; // bit 27 1: Have TCP option
+ unsigned int abn : 1; // bit 26 1: Abnormal case Found
+ unsigned int dack : 1; // bit 25 1: Duplicated ACK
+ unsigned int tcp_qid : 8; // bit 24:17 TCP Queue ID
+ unsigned int TotalPktSize : 17; // bit 16: 0 Total packet size
+#else
+ unsigned int TotalPktSize : 17; // bit 16: 0 Total packet size
+ unsigned int tcp_qid : 8; // bit 24:17 TCP Queue ID
+ unsigned int dack : 1; // bit 25 1: Duplicated ACK
+ unsigned int abn : 1; // bit 26 1: Abnormal case Found
+ unsigned int tcp_opt : 1; // bit 27 1: Have TCP option
+ unsigned int ip_opt : 1; // bit 28 1: have IPV4 option or IPV6 Extension header
+ unsigned int sat : 1; // bit 29 1: SeqCnt > SeqThreshold, or AckCnt > AckThreshold
+ unsigned int osq : 1; // bit 30 1: out of sequence
+ unsigned int ctl : 1; // bit 31 1: have control flag bits (except ack)
+#endif
+ } bits;
+} INTR_QHDR1_T;
+
+/**********************************************************************
+ * Interrupt Queue Header Word 2
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int seq_num;
+} INTR_QHDR2_T;
+
+/**********************************************************************
+ * Interrupt Queue Header Word 3
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ unsigned int ack_num;
+} INTR_QHDR3_T;
+
+/**********************************************************************
+ * Interrupt Queue Header Word 4
+ **********************************************************************/
+typedef union
+{
+ unsigned int bits32;
+ struct bit_intrqhdr4
+ {
+#if (BIG_ENDIAN==1)
+
+ unsigned int SeqCnt : 16; // bit 31:16 Seq# change since last seq# intr.
+ unsigned int AckCnt : 16; // bit 15:0 Ack# change since last ack# intr.
+#else
+ unsigned int AckCnt : 16; // bit 15:0 Ack# change since last ack# intr.
+ unsigned int SeqCnt : 16; // bit 31:16 Seq# change since last seq# intr.
+#endif
+ } bits;
+} INTR_QHDR4_T;
+
+/**********************************************************************
+ * Interrupt Queue Header
+ **********************************************************************/
+typedef struct
+{
+ INTR_QHDR0_T word0;
+ INTR_QHDR1_T word1;
+ INTR_QHDR2_T word2;
+ INTR_QHDR3_T word3;
+ INTR_QHDR4_T word4;
+ unsigned int word5;
+ unsigned int word6;
+ unsigned int word7;
+} INTR_QHDR_T;
+
+/**********************************************************************
+ * GMAC Conf
+ **********************************************************************/
+typedef struct gmac_conf {
+ struct net_device *dev;
+ int portmap;
+ int vid;
+ int flag; /* 1: active 0: non-active */
+} sys_gmac_conf;
+
+/**********************************************************************
+ * GMAC private data
+ **********************************************************************/
+typedef struct {
+ unsigned int rwptr_reg;
+ unsigned int desc_base;
+ unsigned int total_desc_num;
+ unsigned short finished_idx;
+ GMAC_TXDESC_T *curr_tx_desc;
+ GMAC_TXDESC_T *curr_finished_desc;
+ struct sk_buff *tx_skb[TX_DESC_NUM];
+ unsigned long total_sent;
+ unsigned long total_finished;
+ unsigned long intr_cnt;
+} GMAC_SWTXQ_T;
+
+typedef struct {
+ unsigned int desc_base;
+ unsigned long eof_cnt;
+} GMAC_HWTXQ_T;
+
+typedef struct gmac_private{
+ struct net_device *dev;
+ unsigned int existed;
+ unsigned int port_id; // 0 or 1
+ unsigned int base_addr;
+ unsigned int dma_base_addr;
+ unsigned char *mac_addr1;
+ unsigned char *mac_addr2;
+ unsigned int swtxq_desc_base;
+ unsigned int hwtxq_desc_base;
+ GMAC_SWTXQ_T swtxq[TOE_SW_TXQ_NUM];
+ GMAC_HWTXQ_T hwtxq[TOE_HW_TXQ_NUM];
+ NONTOE_QHDR_T *default_qhdr;
+ unsigned int default_desc_base;
+ unsigned int default_desc_num;
+ unsigned int rx_curr_desc;
+ DMA_RWPTR_T rx_rwptr;
+ struct sk_buff *curr_rx_skb;
+ dma_addr_t default_desc_base_dma;
+ dma_addr_t swtxq_desc_base_dma;
+ dma_addr_t hwtxq_desc_base_dma;
+ unsigned int irq;
+ unsigned int flow_control_enable ;
+ unsigned int pre_phy_status;
+ unsigned int full_duplex_cfg;
+ unsigned int speed_cfg;
+ unsigned int auto_nego_cfg;
+ unsigned int full_duplex_status;
+ unsigned int speed_status;
+ unsigned int phy_mode; /* 0->MII 1->GMII 2->RGMII(10/100) 3->RGMII(1000) */
+ unsigned int phy_addr;
+ unsigned int intr0_enabled; // 1: enabled
+ unsigned int intr1_enabled; // 1: enabled
+ unsigned int intr2_enabled; // 1: enabled
+ unsigned int intr3_enabled; // 1: enabled
+ unsigned int intr4_enabled; // 1: enabled
+// unsigned int intr4_enabled_1; // 1: enabled
+ unsigned int intr0_selected; // 1: selected
+ unsigned int intr1_selected; // 1: selected
+ unsigned int intr2_selected; // 1: selected
+ unsigned int intr3_selected; // 1: selected
+ unsigned int intr4_selected; // 1: selected
+ // void (*gmac_rcv_handler)(struct sk_buff *, int);
+ struct net_device_stats ifStatics;
+ unsigned long txDerr_cnt[GMAC_NUM];
+ unsigned long txPerr_cnt[GMAC_NUM];
+ unsigned long RxDerr_cnt[GMAC_NUM];
+ unsigned long RxPerr_cnt[GMAC_NUM];
+ unsigned int isr_rx_cnt;
+ unsigned int isr_tx_cnt;
+ unsigned long rx_discard;
+ unsigned long rx_error;
+ unsigned long rx_mcast;
+ unsigned long rx_bcast;
+ unsigned long rx_status_cnt[8];
+ unsigned long rx_chksum_cnt[8];
+ unsigned long rx_sta1_ucast; // for STA 1 MAC Address
+ unsigned long rx_sta2_ucast; // for STA 2 MAC Address
+ unsigned long mib_full_cnt;
+ unsigned long rx_pause_on_cnt;
+ unsigned long tx_pause_on_cnt;
+ unsigned long rx_pause_off_cnt;
+ unsigned long tx_pause_off_cnt;
+ unsigned long rx_overrun_cnt;
+ unsigned long status_changed_cnt;
+ unsigned long default_q_cnt;
+ unsigned long hw_fq_empty_cnt;
+ unsigned long sw_fq_empty_cnt;
+ unsigned long default_q_intr_cnt;
+ pid_t thr_pid;
+ wait_queue_head_t thr_wait;
+ struct completion thr_exited;
+ spinlock_t lock;
+ int time_to_die;
+ int operation;
+#ifdef SL351x_GMAC_WORKAROUND
+ unsigned long short_frames_cnt;
+#endif
+}GMAC_INFO_T ;
+
+typedef struct toe_private {
+ unsigned int swfq_desc_base;
+ unsigned int hwfq_desc_base;
+ unsigned int hwfq_buf_base;
+// unsigned int toe_desc_base[TOE_TOE_QUEUE_NUM];
+// unsigned int toe_desc_num;
+// unsigned int class_desc_base;
+// unsigned int class_desc_num;
+// unsigned int intr_desc_base;
+// unsigned int intr_desc_num;
+// unsigned int intr_buf_base;
+ DMA_RWPTR_T fq_rx_rwptr;
+ GMAC_INFO_T gmac[GMAC_NUM];
+ dma_addr_t sw_freeq_desc_base_dma;
+ dma_addr_t hw_freeq_desc_base_dma;
+ dma_addr_t hwfq_buf_base_dma;
+ dma_addr_t hwfq_buf_end_dma;
+// dma_addr_t toe_desc_base_dma[TOE_TOE_QUEUE_NUM];
+// dma_addr_t class_desc_base_dma;
+// dma_addr_t intr_desc_base_dma;
+// dma_addr_t intr_buf_base_dma;
+// unsigned long toe_iq_intr_full_cnt[TOE_INTR_QUEUE_NUM];
+// unsigned long toe_iq_intr_cnt[TOE_INTR_QUEUE_NUM];
+// unsigned long toe_q_intr_full_cnt[TOE_TOE_QUEUE_NUM];
+// unsigned long class_q_intr_full_cnt[TOE_CLASS_QUEUE_NUM];
+// unsigned long class_q_intr_cnt[TOE_CLASS_QUEUE_NUM];
+} TOE_INFO_T;
+
+extern TOE_INFO_T toe_private_data;
+
+#define GMAC_PORT0 0
+#define GMAC_PORT1 1
+/**********************************************************************
+ * PHY Definition
+ **********************************************************************/
+#define HPHY_ADDR 0x01
+#define GPHY_ADDR 0x02
+
+enum phy_state
+{
+ LINK_DOWN = 0,
+ LINK_UP = 1
+};
+
+/* transmit timeout value */
+
+#endif //_GMAC_SL351x_H
--- /dev/null
+++ b/include/asm-arm/arch-sl2312/sl351x_hash_cfg.h
@@ -0,0 +1,365 @@
+/*-----------------------------------------------------------------------------------
+* sl351x_hash_cfg.h
+*
+* Description:
+*
+* History:
+*
+* 9/14/2005 Gary Chen Create
+*
+*-------------------------------------------------------------------------------------*/
+#ifndef _SL351x_HASH_CFG_H_
+#define _SL351x_HASH_CFG_H_ 1
+
+// #define NAT_DEBUG_MSG 1
+// #define DEBUG_NAT_MIXED_HW_SW_TX 1
+#ifdef DEBUG_NAT_MIXED_HW_SW_TX
+ // #define NAT_DEBUG_LAN_HASH_TIMEOUT 1
+ // #define NAT_DEBUG_WAN_HASH_TIMEOUT 1
+#endif
+
+#define IPIV(a,b,c,d) ((a<<24)+(b<<16)+(c<<8)+d)
+#define IPIV1(a) ((a>>24)&0xff)
+#define IPIV2(a) ((a>>16)&0xff)
+#define IPIV3(a) ((a>>8)&0xff)
+#define IPIV4(a) ((a)&0xff)
+
+#define HASH_MAX_BYTES 64 // 128
+#define HASH_ACTION_DWORDS 9
+#define HASH_MAX_DWORDS (HASH_MAX_BYTES / sizeof(u32))
+#define HASH_MAX_KEY_DWORD (HASH_MAX_DWORDS - HASH_ACTION_DWORDS)
+#define HASH_INIT_KEY 0x534C4F52
+#define HASH_BITS 12 // 12 : Normal, 7: Simulation
+#define HASH_TOTAL_ENTRIES (1 << HASH_BITS)
+#define HASH_MAX_ENTRIES (1 << 12)
+#define HASH_TOE_ENTRIES (HASH_TOTAL_ENTRIES >> 5)
+#define HASH_BITS_MASK ((1 << HASH_BITS) - 1)
+
+#define hash_lock(lock) // spin_lock_bh(lock)
+#define hash_unlock(lock) // spin_unlock_bh(lock)
+
+/*----------------------------------------------------------------------
+ * special macro
+ ----------------------------------------------------------------------*/
+#define HASH_PUSH_WORD(cp, data) {*cp++ = (((u16)(data)) ) & 0xff; \
+ *cp++ = (((u16)(data)) >> 8) & 0xff;}
+#define HASH_PUSH_DWORD(cp, data) {*cp++ = (u8)(((u32)(data)) ) & 0xff; \
+ *cp++ = (u8)(((u32)(data)) >> 8) & 0xff; \
+ *cp++ = (u8)(((u32)(data)) >> 16) & 0xff; \
+ *cp++ = (u8)(((u32)(data)) >> 24) & 0xff;}
+#define HASH_PUSH_BYTE(cp, data) {*cp++ = ((u8)(data)) & 0xff;}
+
+/*----------------------------------------------------------------------
+ * key
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u8 port;
+ u16 Ethertype;
+ u8 da[6];
+ u8 sa[6];
+ u16 pppoe_sid;
+ u16 vlan_id;
+ u8 ipv4_hdrlen;
+ u8 ip_tos;
+ u8 ip_protocol;
+ u32 ipv6_flow_label;
+ u8 sip[16];
+ u8 dip[16];
+ //__u32 sip[4];
+ //__u32 dip[4];
+ u8 l4_bytes[24];
+ u8 l7_bytes[24];
+ u8 ipv6; // 1: IPv6, 0: IPV4
+} ENTRY_KEY_T;
+
+/*----------------------------------------------------------------------
+ * key for NAT
+ * Note: packed
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u16 Ethertype; // not used
+ u8 port_id;
+ u8 rule_id;
+ u8 ip_protocol;
+ u8 reserved1; // ip_tos, not used
+ u16 reserved2; // not used
+ u32 sip;
+ u32 dip;
+ u16 sport;
+ u16 dport;
+} NAT_KEY_T;
+
+#define NAT_KEY_DWORD_SIZE (sizeof(NAT_KEY_T)/sizeof(u32))
+#define NAT_KEY_SIZE (sizeof(NAT_KEY_T))
+
+/*----------------------------------------------------------------------
+ * key for NAT
+ * Note: packed
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u16 Ethertype; // not used
+ u8 port_id;
+ u8 rule_id;
+ u8 ip_protocol;
+ u8 reserved1; // ip_tos, not used
+ u16 reserved2; // not used
+ u32 sip;
+ u32 dip;
+ u16 reserved3;
+ u16 protocol;
+ u16 reserved4;
+ u16 call_id;
+} GRE_KEY_T;
+
+#define GRE_KEY_DWORD_SIZE (sizeof(GRE_KEY_T)/sizeof(u32))
+#define GRE_KEY_SIZE (sizeof(GRE_KEY_T))
+/*----------------------------------------------------------------------
+ * key present or not
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u32 port : 1;
+ u32 Ethertype : 1;
+ u32 da : 1;
+ u32 sa : 1;
+ u32 pppoe_sid : 1;
+ u32 vlan_id : 1;
+ u32 ipv4_hdrlen : 1;
+ u32 ip_tos : 1;
+ u32 ip_protocol : 1;
+ u32 ipv6_flow_label : 1;
+ u32 sip : 1;
+ u32 dip : 1;
+ u32 l4_bytes_0_3 : 1;
+ u32 l4_bytes_4_7 : 1;
+ u32 l4_bytes_8_11 : 1;
+ u32 l4_bytes_12_15 : 1;
+ u32 l4_bytes_16_19 : 1;
+ u32 l4_bytes_20_23 : 1;
+ u32 l7_bytes_0_3 : 1;
+ u32 l7_bytes_4_7 : 1;
+ u32 l7_bytes_8_11 : 1;
+ u32 l7_bytes_12_15 : 1;
+ u32 l7_bytes_16_19 : 1;
+ u32 l7_bytes_20_23 : 1;
+ u32 reserved : 8;
+} KEY_FIELD_T;
+
+/*----------------------------------------------------------------------
+ * action
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u32 reserved0 : 5; // bit 0:4
+ u32 pppoe : 2; // bit 5:6
+ u32 vlan : 2; // bit 7:8
+ u32 sa : 1; // bit 9
+ u32 da : 1; // bit 10
+ u32 Dport : 1; // bit 11
+ u32 Sport : 1; // bit 12
+ u32 Dip : 1; // bit 13
+ u32 Sip : 1; // bit 14
+ u32 sw_id : 1; // bit 15
+ u32 frag : 1; // bit 16
+ u32 option : 1; // bit 17
+ u32 ttl_0 : 1; // bit 18
+ u32 ttl_1 : 1; // bit 19
+ u32 mtu : 1; // bit 20
+ u32 exception : 1; // bit 21
+ u32 srce_qid : 1; // bit 22
+ u32 discard : 1; // bit 23
+ u32 dest_qid : 8; // bit 24:31
+} ENTRY_ACTION_T;
+
+#define ACTION_DISCARD_BIT BIT(23)
+#define ACTION_SRCE_QID_BIT BIT(22)
+#define ACTION_EXCEPTION_BIT BIT(21)
+#define ACTION_MTU_BIT BIT(20)
+#define ACTION_TTL_1_BIT BIT(19)
+#define ACTION_TTL_0_BIT BIT(18)
+#define ACTION_IP_OPTION BIT(17)
+#define ACTION_FRAG_BIT BIT(16)
+#define ACTION_SWID_BIT BIT(15)
+#define ACTION_SIP_BIT BIT(14)
+#define ACTION_DIP_BIT BIT(13)
+#define ACTION_SPORT_BIT BIT(12)
+#define ACTION_DPORT_BIT BIT(11)
+#define ACTION_DA_BIT BIT(10)
+#define ACTION_SA_BIT BIT(9)
+#define ACTION_VLAN_DEL_BIT BIT(8)
+#define ACTION_VLAN_INS_BIT BIT(7)
+#define ACTION_PPPOE_DEL_BIT BIT(6)
+#define ACTION_PPPOE_INS_BIT BIT(5)
+#define ACTION_L4_THIRD_BIT BIT(4)
+#define ACTION_L4_FOURTH_BIT BIT(3)
+
+#define NAT_ACTION_BITS (ACTION_SRCE_QID_BIT | ACTION_EXCEPTION_BIT | \
+ ACTION_TTL_1_BIT | ACTION_TTL_0_BIT | \
+ ACTION_IP_OPTION | ACTION_FRAG_BIT | \
+ ACTION_DA_BIT | ACTION_SA_BIT)
+#define NAT_LAN2WAN_ACTIONS (NAT_ACTION_BITS | ACTION_SIP_BIT | ACTION_SPORT_BIT)
+#define NAT_WAN2LAN_ACTIONS (NAT_ACTION_BITS | ACTION_DIP_BIT | ACTION_DPORT_BIT)
+#define NAT_PPPOE_LAN2WAN_ACTIONS (NAT_LAN2WAN_ACTIONS | ACTION_PPPOE_INS_BIT)
+#define NAT_PPPOE_WAN2LAN_ACTIONS (NAT_WAN2LAN_ACTIONS | ACTION_PPPOE_DEL_BIT)
+#define NAT_PPTP_LAN2WAN_ACTIONS (NAT_ACTION_BITS | ACTION_SIP_BIT | ACTION_L4_FOURTH_BIT)
+#define NAT_PPTP_WAN2LAN_ACTIONS (NAT_ACTION_BITS | ACTION_DIP_BIT | ACTION_L4_FOURTH_BIT)
+#define NAT_PPPOE_PPTP_LAN2WAN_ACTIONS (NAT_PPTP_LAN2WAN_ACTIONS | ACTION_PPPOE_INS_BIT)
+#define NAT_PPPOE_PPTP_WAN2LAN_ACTIONS (NAT_PPTP_WAN2LAN_ACTIONS | ACTION_PPPOE_DEL_BIT)
+
+/*----------------------------------------------------------------------
+ * parameter
+ ----------------------------------------------------------------------*/
+typedef struct {
+ u8 da[6];
+ u8 sa[6];
+ u16 vlan;
+ u16 pppoe;
+ u32 Sip;
+ u32 Dip;
+ u16 Sport;
+ u16 Dport;
+ u16 sw_id;
+ u16 mtu;
+} ENTRY_PARAM_T;
+
+/*----------------------------------------------------------------------
+ * Hash Entry
+ ----------------------------------------------------------------------*/
+typedef struct {
+ char rule;
+ ENTRY_KEY_T key;
+ KEY_FIELD_T key_present;
+ ENTRY_ACTION_T action;
+ ENTRY_PARAM_T param;
+ int index;
+ int total_dwords;
+} HASH_ENTRY_T;
+
+/*----------------------------------------------------------------------
+ * NAT Hash Entry
+ ----------------------------------------------------------------------*/
+typedef struct {
+ short counter;
+ short interval;
+} HASH_TIMEOUT_T;
+
+/*----------------------------------------------------------------------
+ * NAT Hash Entry for TCP/UDP protocol
+ ----------------------------------------------------------------------*/
+typedef struct {
+ NAT_KEY_T key;
+ union {
+ u32 dword;
+ ENTRY_ACTION_T bits;
+ } action;
+ ENTRY_PARAM_T param;
+ HASH_TIMEOUT_T tmo; // used by software only, to use memory space efficiently
+} NAT_HASH_ENTRY_T;
+
+#define NAT_HASH_ENTRY_SIZE (sizeof(NAT_HASH_ENTRY_T))
+
+/*----------------------------------------------------------------------
+ * GRE Hash Entry for PPTP/GRE protocol
+ ----------------------------------------------------------------------*/
+typedef struct {
+ GRE_KEY_T key;
+ union {
+ u32 dword;
+ ENTRY_ACTION_T bits;
+ } action;
+ ENTRY_PARAM_T param;
+ HASH_TIMEOUT_T tmo; // used by software only, to use memory space efficiently
+} GRE_HASH_ENTRY_T;
+
+#define GRE_HASH_ENTRY_SIZE (sizeof(GRE_HASH_ENTRY_T))
+
+/*----------------------------------------------------------------------
+ * External Variables
+ ----------------------------------------------------------------------*/
+extern char hash_tables[HASH_TOTAL_ENTRIES][HASH_MAX_BYTES] __attribute__ ((aligned(16)));
+extern u32 hash_nat_owner_bits[HASH_TOTAL_ENTRIES/32];
+/*----------------------------------------------------------------------
+* hash_get_valid_flag
+*----------------------------------------------------------------------*/
+static inline int hash_get_valid_flag(int index)
+{
+ volatile u32 *hash_valid_bits_ptr = (volatile u32 *)TOE_V_BIT_BASE;
+
+#ifdef SL351x_GMAC_WORKAROUND
+ if (index >= (0x80 * 8) && index < (0x8c * 8))
+ return 1;
+#endif
+ return (hash_valid_bits_ptr[index/32] & (1 << (index %32)));
+}
+
+/*----------------------------------------------------------------------
+* hash_get_nat_owner_flag
+*----------------------------------------------------------------------*/
+static inline int hash_get_nat_owner_flag(int index)
+{
+ return (hash_nat_owner_bits[index/32] & (1 << (index %32)));
+}
+
+/*----------------------------------------------------------------------
+* hash_validate_entry
+*----------------------------------------------------------------------*/
+static inline void hash_validate_entry(int index)
+{
+ volatile u32 *hash_valid_bits_ptr = (volatile u32 *)TOE_V_BIT_BASE;
+ register int ptr = index/32, bits = 1 << (index %32);
+
+ hash_valid_bits_ptr[ptr] |= bits;
+}
+
+/*----------------------------------------------------------------------
+* hash_invalidate_entry
+*----------------------------------------------------------------------*/
+static inline void hash_invalidate_entry(int index)
+{
+ volatile u32 *hash_valid_bits_ptr = (volatile u32 *)TOE_V_BIT_BASE;
+ register int ptr = index/32, bits = 1 << (index %32);
+
+ hash_valid_bits_ptr[ptr] &= ~(bits);
+}
+
+/*----------------------------------------------------------------------
+* hash_nat_enable_owner
+*----------------------------------------------------------------------*/
+static inline void hash_nat_enable_owner(int index)
+{
+ hash_nat_owner_bits[index/32] |= (1 << (index % 32));
+}
+
+/*----------------------------------------------------------------------
+* hash_nat_disable_owner
+*----------------------------------------------------------------------*/
+static inline void hash_nat_disable_owner(int index)
+{
+ hash_nat_owner_bits[index/32] &= ~(1 << (index % 32));
+}
+
+/*----------------------------------------------------------------------
+* hash_get_entry
+*----------------------------------------------------------------------*/
+static inline void *hash_get_entry(int index)
+{
+ return (void*) &hash_tables[index][0];
+}
+
+/*----------------------------------------------------------------------
+* Functions
+*----------------------------------------------------------------------*/
+extern int hash_add_entry(HASH_ENTRY_T *entry);
+extern void sl351x_hash_init(void);
+extern void hash_set_valid_flag(int index, int valid);
+extern void hash_set_nat_owner_flag(int index, int valid);
+extern void *hash_get_entry(int index);
+extern int hash_build_keys(u32 *destp, HASH_ENTRY_T *entry);
+extern void hash_build_nat_keys(u32 *destp, HASH_ENTRY_T *entry);
+extern int hash_write_entry(HASH_ENTRY_T *entry, u8 *key);
+extern int hash_add_entry(HASH_ENTRY_T *entry);
+extern u16 hash_crc16(u16 crc, u8 *datap, u32 len);
+extern u16 hash_gen_crc16(u8 *datap, u32 len);
+
+#endif // _SL351x_HASH_CFG_H_
+
+
+
--- /dev/null
+++ b/include/asm-arm/arch-sl2312/sl351x_nat_cfg.h
@@ -0,0 +1,211 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* sl_nat_cfg.h
+*
+* Description:
+* - Define the Device Control Commands for NAT Configuration
+*
+* History:
+*
+* 4/28/2006 Gary Chen Create
+*
+*-----------------------------------------------------------------------------*/
+#ifndef _SL351x_NAT_CFG_H_
+#define _SL351x_NAT_CFG_H_ 1
+
+/*----------------------------------------------------------------------
+* Confiuration
+*----------------------------------------------------------------------*/
+#ifdef CONFIG_NETFILTER
+#define CONFIG_SL351x_NAT 1
+#undef CONFIG_SL351x_NAT
+#undef CONFIG_SL351x_SYSCTL
+#endif
+#define CONFIG_NAT_MAX_IP_NUM 4 // per device (eth0 or eth1)
+#define CONFIG_NAT_MAX_XPORT 64
+#define CONFIG_NAT_MAX_WRULE 16 // per Queue
+#define CONFIG_NAT_TXQ_NUM 4
+/*----------------------------------------------------------------------
+* Command set
+*----------------------------------------------------------------------*/
+#define SIOCDEVSL351x SIOCDEVPRIVATE // 0x89F0
+#define NATSSTATUS 0
+#define NATGSTATUS 1
+#define NATSETPORT 2
+#define NATGETPORT 3
+#define NATADDIP 4
+#define NATDELIP 5
+#define NATGETIP 6
+#define NATAXPORT 7
+#define NATDXPORT 8
+#define NATGXPORT 9
+#define NATSWEIGHT 10
+#define NATGWEIGHT 11
+#define NATAWRULE 12
+#define NATDWRULE 13
+#define NATGWRULE 14
+#define NATSDEFQ 15
+#define NATGDEFQ 16
+#define NATRMIPCFG 17 // remove IP config
+#define NATTESTENTRY 18
+#define NATSETMEM 19
+#define NATSHOWMEM 20
+/*----------------------------------------------------------------------
+* Command Structure
+*----------------------------------------------------------------------*/
+// Common Header
+typedef struct {
+ unsigned short cmd; // command ID
+ unsigned short len; // data length, excluding this header
+} NATCMD_HDR_T;
+
+// NATSSTATUS & NATGSTATUS commands
+typedef struct {
+ unsigned char enable;
+} NAT_STATUS_T;
+
+// NATSETPORT & NATGETPORT commands
+typedef struct {
+ unsigned char portmap;
+} NAT_PORTCFG_T;
+
+typedef struct {
+ unsigned int ipaddr;
+ unsigned int netmask;
+} NAT_IP_ENTRY_T;
+
+// NATADDIP & NATDELIP commands
+typedef struct {
+ NAT_IP_ENTRY_T entry;
+} NAT_IPCFG_T;
+
+// NATGETIP command
+typedef struct {
+ unsigned int total;
+ NAT_IP_ENTRY_T entry[CONFIG_NAT_MAX_IP_NUM];
+} NAT_IPCFG_ALL_T;
+
+typedef struct {
+ unsigned int protocol;
+ unsigned short sport_start;
+ unsigned short sport_end;
+ unsigned short dport_start;
+ unsigned short dport_end;
+} NAT_XPORT_ENTRY_T;
+
+// NATAXPORT & NATDXPORT Commands
+typedef struct {
+ NAT_XPORT_ENTRY_T entry;
+} NAT_XPORT_T;
+
+// NATGXPORT Command
+typedef struct {
+ unsigned int total;
+ NAT_XPORT_ENTRY_T entry[CONFIG_NAT_MAX_XPORT];
+} NAT_XPORT_ALL_T;
+
+// NATSWEIGHT & NATGWEIGHT Commands
+typedef struct {
+ unsigned char weight[CONFIG_NAT_TXQ_NUM];
+} NAT_WEIGHT_T;
+
+typedef struct {
+ unsigned int protocol;
+ unsigned int sip_start;
+ unsigned int sip_end;
+ unsigned int dip_start;
+ unsigned int dip_end;
+ unsigned short sport_start;
+ unsigned short sport_end;
+ unsigned short dport_start;
+ unsigned short dport_end;
+} NAT_WRULE_ENTRY_T;
+
+// NATAWRULE & NATDWRULE Commands
+typedef struct {
+ unsigned int qid;
+ NAT_WRULE_ENTRY_T entry;
+} NAT_WRULE_T;
+
+// NATGWRULE Command
+typedef struct {
+ unsigned int total;
+ NAT_WRULE_ENTRY_T entry[CONFIG_NAT_MAX_WRULE];
+} NAT_WRULE_ALL_T;
+
+// NATSDEFQ & NATGDEFQ commands
+typedef struct {
+ unsigned int qid;
+} NAT_QUEUE_T;
+
+// NATTESTENTRY
+typedef struct {
+ u_int16_t cmd; // command ID
+ u_int16_t len; // data length, excluding this header
+ u_int8_t init_enable;
+} NAT_TESTENTRY_T;
+
+typedef union
+{
+ NAT_STATUS_T status;
+ NAT_PORTCFG_T portcfg;
+ NAT_IPCFG_T ipcfg;
+ NAT_XPORT_T xport;
+ NAT_WEIGHT_T weight;
+ NAT_WRULE_T wrule;
+ NAT_QUEUE_T queue;
+ NAT_TESTENTRY_T init_entry;
+} NAT_REQ_E;
+
+/*----------------------------------------------------------------------
+* NAT Configuration
+* - Used by driver only
+*----------------------------------------------------------------------*/
+typedef struct {
+ unsigned int enabled;
+ unsigned int init_enabled;
+ unsigned int tcp_udp_rule_id;
+ unsigned int gre_rule_id;
+ unsigned int lan_port;
+ unsigned int wan_port;
+ unsigned int default_hw_txq;
+ short tcp_tmo_interval;
+ short udp_tmo_interval;
+ short gre_tmo_interval;
+ NAT_IPCFG_ALL_T ipcfg[2]; // LAN/WAN port
+ NAT_XPORT_ALL_T xport;
+ NAT_WEIGHT_T weight;
+ NAT_WRULE_ALL_T wrule[CONFIG_NAT_TXQ_NUM];
+} NAT_CFG_T;
+
+/*----------------------------------------------------------------------
+* NAT Control Block
+* - Used by driver only
+* - Stores LAN-IN or WAN-IN information
+* - WAN-OUT and LAN-OUT driver use them to build up a hash entry
+* - NOTES: To update this data structure, MUST take care of alignment issue
+* - MUST make sure that the size of skbuff structure must
+* be larger than (40 + sizof(NAT_CB_T))
+*----------------------------------------------------------------------*/
+typedef struct {
+ unsigned short tag;
+ unsigned char sa[6];
+ unsigned int sip;
+ unsigned int dip;
+ unsigned short sport;
+ unsigned short dport;
+ unsigned char pppoe_frame;
+ unsigned char state; // same to enum tcp_conntrack
+ unsigned char reserved[2];
+} NAT_CB_T;
+
+#define NAT_CB_TAG 0x4C53 // "SL"
+#define NAT_CB_SIZE sizeof(NAT_CB_T)
+// #define NAT_SKB_CB(skb) (NAT_CB_T *)(((unsigned int)&((skb)->cb[40]) + 3) & ~3) // for align 4
+#define NAT_SKB_CB(skb) (NAT_CB_T *)&((skb)->cb[40]) // for align 4
+
+#endif // _SL351x_NAT_CFG_H_
+
+
+
--- /dev/null
+++ b/include/asm-arm/arch-sl2312/sl351x_toe.h
@@ -0,0 +1,88 @@
+/**************************************************************************
+* Copyright 2006 StorLink Semiconductors, Inc. All rights reserved.
+*--------------------------------------------------------------------------
+* Name : sl351x_toe.h
+* Description :
+* Define for TOE driver of Storlink SL351x
+*
+* History
+*
+* Date Writer Description
+*----------------------------------------------------------------------------
+* Xiaochong Create
+*
+****************************************************************************/
+#ifndef __SL351x_TOE_H
+#define __SL351x_TOE_H 1
+#include <net/sock.h>
+#include <asm/arch/sl351x_gmac.h>
+#include <linux/timer.h>
+#include <linux/netdevice.h>
+#include <linux/ip.h>
+#include <linux/if_ether.h>
+/*
+ * TOE_CONN_T is data structure of tcp connection info, used at both
+ * device layer and kernel tcp layer
+ * skb is the jumbo frame
+ */
+
+struct toe_conn{
+ __u8 qid; // connection qid 0~63.
+ __u8 ip_ver; // 0: not used; 4: ipv4; 6: ipv6.
+ /* hash key of the connection */
+ __u16 source;
+ __u16 dest;
+ __u32 saddr[4];
+ __u32 daddr[4];
+
+ __u32 seq;
+ __u32 ack_seq;
+
+ /* these fields are used to set TOE QHDR */
+ __u32 ack_threshold;
+ __u32 seq_threshold;
+ __u16 max_pktsize;
+
+ /* used by sw toe, accumulated ack_seq of ack frames */
+ __u16 ack_cnt;
+ /* used by sw toe, accumulated data frames held at driver */
+ __u16 cur_pktsize;
+
+ __u8 status;
+#define TCP_CONN_UNDEFINE 0X00
+#define TCP_CONN_CREATION 0X01
+#define TCP_CONN_CONNECTING 0X02
+#define TCP_CONN_ESTABLISHED 0X04
+#define TCP_CONN_RESET 0X08 // this is used for out-of-order
+ // or congestion window is small
+#define TCP_CONN_CLOSING 0X10
+#define TCP_CONN_CLOSED 0x11
+
+ __u16 hash_entry_index; /* associated hash entry */
+
+ // one timer per connection. Otherwise all connections should be scanned
+ // in a timeout interrupt, and timeout interrupt is triggered no matter
+ // a connection is actually timeout or not.
+ struct timer_list rx_timer;
+ unsigned long last_rx_jiffies;
+ GMAC_INFO_T *gmac;
+ struct net_device *dev;
+
+ // for generating pure ack frame.
+ struct ethhdr l2_hdr;
+ struct iphdr l3_hdr;
+
+ spinlock_t conn_lock;
+ DMA_RWPTR_T toeq_rwptr;
+ GMAC_RXDESC_T *curr_desc;
+ struct sk_buff *curr_rx_skb;
+};
+
+struct jumbo_frame {
+ struct sk_buff *skb0; // the head of jumbo frame
+ struct sk_buff *tail; // the tail of jumbo frame
+ struct iphdr *iphdr0; // the ip hdr of skb0.
+ struct tcphdr *tcphdr0; // the tcp hdr of skb0.
+};
+
+#endif // __SL351x_TOE_H
--- a/drivers/net/Kconfig
+++ b/drivers/net/Kconfig
@@ -2131,6 +2131,42 @@
The safe and default value for this is N.
+config NET_GMAC
+ tristate "Storlink Gigabit Ethernet support"
+ depends on ARCH_SL2312
+ help
+ This driver supports Storlink dual Gigabit Ethernet.
+
+config NET_SL2312
+ tristate "Storlink Gigabit Ethernet support"
+ depends on NET_GMAC
+ help
+ This driver supports Storlink dual Gigabit Ethernet.
+
+config NET_SL351X
+ tristate "Storlink Lepus Gigabit Ethernet support"
+ depends on NET_GMAC
+ help
+ This driver supports Storlink TOE and NAT dual Gigabit Ethernet.
+
+config SL2312_TSO
+ bool "Tx Segmentation Enable"
+ depends on NET_GMAC
+ help
+ TBD
+
+config SL2312_MPAGE
+ bool "Tx Multipage Enable"
+ depends on NET_GMAC
+ help
+ TBD
+
+config SL2312_RECVFILE
+ bool "Rx Multipage Enable"
+ depends on NET_GMAC
+ help
+ TBD
+
config DL2K
tristate "D-Link DL2000-based Gigabit Ethernet support"
depends on PCI
--- a/drivers/net/Makefile
+++ b/drivers/net/Makefile
@@ -236,4 +236,8 @@
obj-$(CONFIG_FS_ENET) += fs_enet/
-obj-$(CONFIG_NETXEN_NIC) += netxen/
+
+obj-$(CONFIG_NET_SL351X)+= sl351x_gmac.o sl351x_nat.o sl351x_hash.o sl351x_crc16.o sl351x_proc.o sl_switch.o
+obj-$(CONFIG_NET_SL2312)+= sl2312_emac.o
+
+