diff options
Diffstat (limited to 'target/linux/ixp4xx/patches-2.6.24/033-velocity_be.patch')
-rw-r--r-- | target/linux/ixp4xx/patches-2.6.24/033-velocity_be.patch | 897 |
1 files changed, 897 insertions, 0 deletions
diff --git a/target/linux/ixp4xx/patches-2.6.24/033-velocity_be.patch b/target/linux/ixp4xx/patches-2.6.24/033-velocity_be.patch new file mode 100644 index 0000000000..e1d212d9c0 --- /dev/null +++ b/target/linux/ixp4xx/patches-2.6.24/033-velocity_be.patch @@ -0,0 +1,897 @@ +diff -Naur linux-2.6.24.orig/drivers/net/via-velocity.c linux-2.6.24/drivers/net/via-velocity.c +--- linux-2.6.24.orig/drivers/net/via-velocity.c 2008-01-31 23:11:26.000000000 -0600 ++++ linux-2.6.24/drivers/net/via-velocity.c 2008-02-01 01:12:15.000000000 -0600 +@@ -254,11 +254,31 @@ + MODULE_LICENSE("GPL"); + MODULE_DESCRIPTION("VIA Networking Velocity Family Gigabit Ethernet Adapter Driver"); + ++/* Valid values for vdebug (additive, this is a bitmask): ++ * 0x00 => off ++ * 0x01 => always on ++ * 0x02 => additional detail on tx (rx, too, if anyone implements same) ++ * 0x04 => detail the initialization process ++ * 0x08 => spot debug detail; to be used as developers see fit ++ */ ++static int vdebug = 0; ++ ++/* HAIL - these macros are for the normal 0x01-type tracing... */ ++#define HAIL(S) \ ++ if (vdebug&1) printk(KERN_NOTICE "%s\n", (S)); ++#define HAILS(S,T) \ ++ if (vdebug&1) printk(KERN_NOTICE "%s -> status=0x%x\n", (S), (T)); ++ + #define VELOCITY_PARAM(N,D) \ + static int N[MAX_UNITS]=OPTION_DEFAULT;\ + module_param_array(N, int, NULL, 0); \ + MODULE_PARM_DESC(N, D); + ++#define VELO_DEBUG_MIN 0 ++#define VELO_DEBUG_MAX 255 ++#define VELO_DEBUG_DEF 0 ++VELOCITY_PARAM(velo_debug, "Debug level"); ++ + #define RX_DESC_MIN 64 + #define RX_DESC_MAX 255 + #define RX_DESC_DEF 64 +@@ -557,12 +577,12 @@ + if (val == -1) + *opt |= (def ? flag : 0); + else if (val < 0 || val > 1) { +- printk(KERN_NOTICE "%s: the value of parameter %s is invalid, the valid range is (0-1)\n", +- devname, name); ++ printk(KERN_NOTICE "via-velocity: the value of parameter %s is invalid, the valid range is (0-1)\n", ++ name); + *opt |= (def ? flag : 0); + } else { +- printk(KERN_INFO "%s: set parameter %s to %s\n", +- devname, name, val ? "TRUE" : "FALSE"); ++ printk(KERN_INFO "via-velocity: set parameter %s to %s\n", ++ name, val ? "TRUE" : "FALSE"); + *opt |= (val ? flag : 0); + } + } +@@ -580,6 +600,7 @@ + static void __devinit velocity_get_options(struct velocity_opt *opts, int index, char *devname) + { + ++ velocity_set_int_opt(&opts->velo_debug, velo_debug[index], VELO_DEBUG_MIN, VELO_DEBUG_MAX, VELO_DEBUG_DEF, "velo_debug", devname); + velocity_set_int_opt(&opts->rx_thresh, rx_thresh[index], RX_THRESH_MIN, RX_THRESH_MAX, RX_THRESH_DEF, "rx_thresh", devname); + velocity_set_int_opt(&opts->DMA_length, DMA_length[index], DMA_LENGTH_MIN, DMA_LENGTH_MAX, DMA_LENGTH_DEF, "DMA_length", devname); + velocity_set_int_opt(&opts->numrx, RxDescriptors[index], RX_DESC_MIN, RX_DESC_MAX, RX_DESC_DEF, "RxDescriptors", devname); +@@ -593,6 +614,7 @@ + velocity_set_int_opt((int *) &opts->wol_opts, wol_opts[index], WOL_OPT_MIN, WOL_OPT_MAX, WOL_OPT_DEF, "Wake On Lan options", devname); + velocity_set_int_opt((int *) &opts->int_works, int_works[index], INT_WORKS_MIN, INT_WORKS_MAX, INT_WORKS_DEF, "Interrupt service works", devname); + opts->numrx = (opts->numrx & ~3); ++ vdebug = opts->velo_debug; + } + + /** +@@ -608,6 +630,8 @@ + struct mac_regs __iomem * regs = vptr->mac_regs; + unsigned short vid; + ++ HAIL("velocity_init_cam_filter"); ++ + /* Turn on MCFG_PQEN, turn off MCFG_RTGOPT */ + WORD_REG_BITS_SET(MCFG_PQEN, MCFG_RTGOPT, ®s->MCFG); + WORD_REG_BITS_ON(MCFG_VIDFR, ®s->MCFG); +@@ -636,8 +660,10 @@ + } else { + u16 temp = 0; + mac_set_vlan_cam(regs, 0, (u8 *) &temp); +- temp = 1; +- mac_set_vlan_cam_mask(regs, (u8 *) &temp); ++ /* temp = 1; BE */ ++ /* mac_set_vlan_cam_mask(regs, (u8 *) &temp); BE */ ++ vptr->vCAMmask[0] |= 1; /* BE */ ++ mac_set_vlan_cam_mask(regs, vptr->vCAMmask); /* BE */ + } + } + +@@ -675,13 +701,15 @@ + struct mac_regs __iomem * regs = vptr->mac_regs; + int i; + ++ HAIL("velocity_rx_reset"); + vptr->rd_dirty = vptr->rd_filled = vptr->rd_curr = 0; + + /* + * Init state, all RD entries belong to the NIC + */ + for (i = 0; i < vptr->options.numrx; ++i) +- vptr->rd_ring[i].rdesc0.owner = OWNED_BY_NIC; ++ /* vptr->rd_ring[i].rdesc0.owner = OWNED_BY_NIC; BE */ ++ vptr->rd_ring[i].rdesc0 |= cpu_to_le32(BE_OWNED_BY_NIC); /* BE */ + + writew(vptr->options.numrx, ®s->RBRDU); + writel(vptr->rd_pool_dma, ®s->RDBaseLo); +@@ -704,12 +732,15 @@ + struct mac_regs __iomem * regs = vptr->mac_regs; + int i, mii_status; + ++ if (vdebug&5) printk(KERN_NOTICE "velocity_init_registers: entering\n"); ++ + mac_wol_reset(regs); + + switch (type) { + case VELOCITY_INIT_RESET: + case VELOCITY_INIT_WOL: + ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: RESET or WOL\n"); + netif_stop_queue(vptr->dev); + + /* +@@ -737,12 +768,13 @@ + + case VELOCITY_INIT_COLD: + default: ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: COLD or default\n"); + /* + * Do reset + */ + velocity_soft_reset(vptr); ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: soft reset complete.\n"); + mdelay(5); +- + mac_eeprom_reload(regs); + for (i = 0; i < 6; i++) { + writeb(vptr->dev->dev_addr[i], &(regs->PAR[i])); +@@ -760,11 +792,16 @@ + */ + BYTE_REG_BITS_SET(CFGB_OFSET, (CFGB_CRANDOM | CFGB_CAP | CFGB_MBA | CFGB_BAKOPT), ®s->CFGB); + ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: Initializing CAM filter\n"); + /* + * Init CAM filter + */ ++ if (vdebug&8) printk(KERN_NOTICE "velocity: spot debug: about to init CAM filters\n"); ++ mdelay(5); /* MJW - ARM processors, kernel 2.6.19 - this fixes oopses and hangs */ + velocity_init_cam_filter(vptr); ++ if (vdebug&8) printk(KERN_NOTICE "velocity: spot debug: init CAM filters complete\n"); + ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: Setting packet filter\n"); + /* + * Set packet filter: Receive directed and broadcast address + */ +@@ -774,10 +811,12 @@ + * Enable MII auto-polling + */ + enable_mii_autopoll(regs); ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: enable_mii_autopoll complete.\n"); + + vptr->int_mask = INT_MASK_DEF; + +- writel(cpu_to_le32(vptr->rd_pool_dma), ®s->RDBaseLo); ++ /* writel(cpu_to_le32(vptr->rd_pool_dma), ®s->RDBaseLo); BE */ ++ writel((vptr->rd_pool_dma), ®s->RDBaseLo); /* BE */ + writew(vptr->options.numrx - 1, ®s->RDCSize); + mac_rx_queue_run(regs); + mac_rx_queue_wake(regs); +@@ -785,10 +824,13 @@ + writew(vptr->options.numtx - 1, ®s->TDCSize); + + for (i = 0; i < vptr->num_txq; i++) { +- writel(cpu_to_le32(vptr->td_pool_dma[i]), &(regs->TDBaseLo[i])); ++ /* writel(cpu_to_le32(vptr->td_pool_dma[i]), &(regs->TDBaseLo[i])); BE */ ++ writel((vptr->td_pool_dma[i]), &(regs->TDBaseLo[i])); /* BE */ + mac_tx_queue_run(regs, i); + } + ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: DMA settings complete.\n"); ++ + init_flow_control_register(vptr); + + writel(CR0_STOP, ®s->CR0Clr); +@@ -807,8 +849,10 @@ + + enable_flow_control_ability(vptr); + mac_hw_mibs_init(regs); ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: Set interrupt mask\n"); + mac_write_int_mask(vptr->int_mask, regs); + mac_clear_isr(regs); ++ if (vdebug&4) printk(KERN_NOTICE "velocity_init_registers: complete.\n"); + + } + } +@@ -826,6 +870,7 @@ + struct mac_regs __iomem * regs = vptr->mac_regs; + int i = 0; + ++ HAIL("velocity_soft_reset"); + writel(CR0_SFRST, ®s->CR0Set); + + for (i = 0; i < W_MAX_TIMEOUT; i++) { +@@ -888,6 +933,7 @@ + VELOCITY_FULL_DRV_NAM, VELOCITY_VERSION); + printk(KERN_INFO "Copyright (c) 2002, 2003 VIA Networking Technologies, Inc.\n"); + printk(KERN_INFO "Copyright (c) 2004 Red Hat Inc.\n"); ++ printk(KERN_INFO "BE support, misc. fixes MJW 01Jan2007 - may be unstable\n"); + first = 0; + } + +@@ -1104,6 +1150,7 @@ + dma_addr_t pool_dma; + u8 *pool; + ++ HAIL("velocity_init_rings"); + /* + * Allocate all RD/TD rings a single pool + */ +@@ -1166,6 +1213,7 @@ + static void velocity_free_rings(struct velocity_info *vptr) + { + int size; ++ HAIL("velocity_free_rings"); + + size = vptr->options.numrx * sizeof(struct rx_desc) + + vptr->options.numtx * sizeof(struct tx_desc) * vptr->num_txq; +@@ -1182,6 +1230,7 @@ + struct mac_regs __iomem *regs = vptr->mac_regs; + int avail, dirty, unusable; + ++ HAIL("velocity_give_many_rx_descs"); + /* + * RD number must be equal to 4X per hardware spec + * (programming guide rev 1.20, p.13) +@@ -1195,7 +1244,8 @@ + dirty = vptr->rd_dirty - unusable; + for (avail = vptr->rd_filled & 0xfffc; avail; avail--) { + dirty = (dirty > 0) ? dirty - 1 : vptr->options.numrx - 1; +- vptr->rd_ring[dirty].rdesc0.owner = OWNED_BY_NIC; ++ /* vptr->rd_ring[dirty].rdesc0.owner = OWNED_BY_NIC; BE */ ++ vptr->rd_ring[dirty].rdesc0 |= cpu_to_le32(BE_OWNED_BY_NIC); /* BE */ + } + + writew(vptr->rd_filled & 0xfffc, ®s->RBRDU); +@@ -1205,12 +1255,14 @@ + static int velocity_rx_refill(struct velocity_info *vptr) + { + int dirty = vptr->rd_dirty, done = 0, ret = 0; ++ HAIL("velocity_rx_refill"); + + do { + struct rx_desc *rd = vptr->rd_ring + dirty; + + /* Fine for an all zero Rx desc at init time as well */ +- if (rd->rdesc0.owner == OWNED_BY_NIC) ++ /* if (rd->rdesc0.owner == OWNED_BY_NIC) BE */ ++ if (rd->rdesc0 & cpu_to_le32(BE_OWNED_BY_NIC)) /* BE */ + break; + + if (!vptr->rd_info[dirty].skb) { +@@ -1244,6 +1296,7 @@ + int ret; + int mtu = vptr->dev->mtu; + ++ HAIL("velocity_init_rd_ring"); + vptr->rx_buf_sz = (mtu <= ETH_DATA_LEN) ? PKT_BUF_SZ : mtu + 32; + + vptr->rd_info = kcalloc(vptr->options.numrx, +@@ -1275,6 +1328,7 @@ + { + int i; + ++ HAIL("velocity_free_rd_ring"); + if (vptr->rd_info == NULL) + return; + +@@ -1314,6 +1368,7 @@ + struct tx_desc *td; + struct velocity_td_info *td_info; + ++ HAIL("velocity_init_td_ring"); + /* Init the TD ring entries */ + for (j = 0; j < vptr->num_txq; j++) { + curr = vptr->td_pool_dma[j]; +@@ -1350,6 +1405,7 @@ + struct velocity_td_info * td_info = &(vptr->td_infos[q][n]); + int i; + ++ HAIL("velocity_free_td_ring_entry"); + if (td_info == NULL) + return; + +@@ -1379,6 +1435,7 @@ + { + int i, j; + ++ HAIL("velocity_free_td_ring"); + for (j = 0; j < vptr->num_txq; j++) { + if (vptr->td_infos[j] == NULL) + continue; +@@ -1406,34 +1463,42 @@ + struct net_device_stats *stats = &vptr->stats; + int rd_curr = vptr->rd_curr; + int works = 0; ++ u16 wRSR; /* BE */ + ++ HAILS("velocity_rx_srv", status); + do { + struct rx_desc *rd = vptr->rd_ring + rd_curr; + + if (!vptr->rd_info[rd_curr].skb) + break; + +- if (rd->rdesc0.owner == OWNED_BY_NIC) ++ /* if (rd->rdesc0.owner == OWNED_BY_NIC) BE */ ++ if (rd->rdesc0 & cpu_to_le32(BE_OWNED_BY_NIC)) /* BE */ + break; + + rmb(); + ++ wRSR = (u16)(cpu_to_le32(rd->rdesc0)); /* BE */ + /* + * Don't drop CE or RL error frame although RXOK is off + */ +- if ((rd->rdesc0.RSR & RSR_RXOK) || (!(rd->rdesc0.RSR & RSR_RXOK) && (rd->rdesc0.RSR & (RSR_CE | RSR_RL)))) { ++ /* if ((rd->rdesc0.RSR & RSR_RXOK) || (!(rd->rdesc0.RSR & RSR_RXOK) && (rd->rdesc0.RSR & (RSR_CE | RSR_RL)))) { BE */ ++ if ((wRSR & RSR_RXOK) || (!(wRSR & RSR_RXOK) && (wRSR & (RSR_CE | RSR_RL)))) { /* BE */ + if (velocity_receive_frame(vptr, rd_curr) < 0) + stats->rx_dropped++; + } else { +- if (rd->rdesc0.RSR & RSR_CRC) ++ /* if (rd->rdesc0.RSR & RSR_CRC) BE */ ++ if (wRSR & RSR_CRC) /* BE */ + stats->rx_crc_errors++; +- if (rd->rdesc0.RSR & RSR_FAE) ++ /* if (rd->rdesc0.RSR & RSR_FAE) BE */ ++ if (wRSR & RSR_FAE) /* BE */ + stats->rx_frame_errors++; + + stats->rx_dropped++; + } + +- rd->inten = 1; ++ /* rd->inten = 1; BE */ ++ rd->ltwo |= cpu_to_le32(BE_INT_ENABLE); /* BE */ + + vptr->dev->last_rx = jiffies; + +@@ -1464,13 +1529,21 @@ + + static inline void velocity_rx_csum(struct rx_desc *rd, struct sk_buff *skb) + { ++ u8 bCSM; ++ HAIL("velocity_rx_csum"); + skb->ip_summed = CHECKSUM_NONE; + +- if (rd->rdesc1.CSM & CSM_IPKT) { +- if (rd->rdesc1.CSM & CSM_IPOK) { +- if ((rd->rdesc1.CSM & CSM_TCPKT) || +- (rd->rdesc1.CSM & CSM_UDPKT)) { +- if (!(rd->rdesc1.CSM & CSM_TUPOK)) { ++// if (rd->rdesc1.CSM & CSM_IPKT) { ++// if (rd->rdesc1.CSM & CSM_IPOK) { ++// if ((rd->rdesc1.CSM & CSM_TCPKT) || ++// (rd->rdesc1.CSM & CSM_UDPKT)) { ++// if (!(rd->rdesc1.CSM & CSM_TUPOK)) { ++ bCSM = (u8)(cpu_to_le32(rd->rdesc1) >> 16); /* BE */ ++ if (bCSM & CSM_IPKT) { ++ if (bCSM & CSM_IPOK) { ++ if ((bCSM & CSM_TCPKT) || ++ (bCSM & CSM_UDPKT)) { ++ if (!(bCSM & CSM_TUPOK)) { /* BE */ + return; + } + } +@@ -1496,9 +1569,11 @@ + { + int ret = -1; + ++ HAIL("velocity_rx_copy"); + if (pkt_size < rx_copybreak) { + struct sk_buff *new_skb; + ++ HAIL("velocity_rx_copy (working...)"); + new_skb = dev_alloc_skb(pkt_size + 2); + if (new_skb) { + new_skb->dev = vptr->dev; +@@ -1529,10 +1604,12 @@ + static inline void velocity_iph_realign(struct velocity_info *vptr, + struct sk_buff *skb, int pkt_size) + { ++ HAIL("velocity_iph_realign"); + /* FIXME - memmove ? */ + if (vptr->flags & VELOCITY_FLAGS_IP_ALIGN) { + int i; + ++ HAIL("velocity_iph_realign (working...)"); + for (i = pkt_size; i >= 0; i--) + *(skb->data + i + 2) = *(skb->data + i); + skb_reserve(skb, 2); +@@ -1551,19 +1628,27 @@ + static int velocity_receive_frame(struct velocity_info *vptr, int idx) + { + void (*pci_action)(struct pci_dev *, dma_addr_t, size_t, int); ++ u16 pkt_len; /* BE */ ++ u16 wRSR; /* BE */ ++ struct sk_buff *skb; + struct net_device_stats *stats = &vptr->stats; + struct velocity_rd_info *rd_info = &(vptr->rd_info[idx]); + struct rx_desc *rd = &(vptr->rd_ring[idx]); +- int pkt_len = rd->rdesc0.len; +- struct sk_buff *skb; ++ /* int pkt_len = rd->rdesc0.len BE */; ++ ++ pkt_len = ((cpu_to_le32(rd->rdesc0) >> 16) & 0x00003FFFUL); /* BE */ ++ wRSR = (u16)(cpu_to_le32(rd->rdesc0)); /* BE */ + +- if (rd->rdesc0.RSR & (RSR_STP | RSR_EDP)) { ++ HAIL("velocity_receive_frame"); ++ /* if (rd->rdesc0.RSR & (RSR_STP | RSR_EDP)) { BE */ ++ if (wRSR & (RSR_STP | RSR_EDP)) { /* BE */ + VELOCITY_PRT(MSG_LEVEL_VERBOSE, KERN_ERR " %s : the received frame span multple RDs.\n", vptr->dev->name); + stats->rx_length_errors++; + return -EINVAL; + } + +- if (rd->rdesc0.RSR & RSR_MAR) ++ /* if (rd->rdesc0.RSR & RSR_MAR) BE */ ++ if (wRSR & RSR_MAR) /* BE */ + vptr->stats.multicast++; + + skb = rd_info->skb; +@@ -1576,7 +1661,8 @@ + */ + + if (vptr->flags & VELOCITY_FLAGS_VAL_PKT_LEN) { +- if (rd->rdesc0.RSR & RSR_RL) { ++ /* if (rd->rdesc0.RSR & RSR_RL) { BE */ ++ if (wRSR & RSR_RL) { /* BE */ + stats->rx_length_errors++; + return -EINVAL; + } +@@ -1620,6 +1706,7 @@ + struct rx_desc *rd = &(vptr->rd_ring[idx]); + struct velocity_rd_info *rd_info = &(vptr->rd_info[idx]); + ++ HAIL("velocity_alloc_rx_buf"); + rd_info->skb = dev_alloc_skb(vptr->rx_buf_sz + 64); + if (rd_info->skb == NULL) + return -ENOMEM; +@@ -1637,10 +1724,14 @@ + */ + + *((u32 *) & (rd->rdesc0)) = 0; +- rd->len = cpu_to_le32(vptr->rx_buf_sz); +- rd->inten = 1; ++ /* rd->len = cpu_to_le32(vptr->rx_buf_sz); BE */ ++ /* rd->inten = 1; BE */ + rd->pa_low = cpu_to_le32(rd_info->skb_dma); +- rd->pa_high = 0; ++ /* rd->pa_high = 0; BE */ ++ rd->ltwo &= cpu_to_le32(0xC000FFFFUL); /* BE */ ++ rd->ltwo |= cpu_to_le32((vptr->rx_buf_sz << 16)); /* BE */ ++ rd->ltwo |= cpu_to_le32(BE_INT_ENABLE); /* BE */ ++ rd->ltwo &= cpu_to_le32(0xFFFF0000UL); /* BE */ + return 0; + } + +@@ -1661,9 +1752,11 @@ + int full = 0; + int idx; + int works = 0; ++ u16 wTSR; /* BE */ + struct velocity_td_info *tdinfo; + struct net_device_stats *stats = &vptr->stats; + ++ HAILS("velocity_tx_srv", status); + for (qnum = 0; qnum < vptr->num_txq; qnum++) { + for (idx = vptr->td_tail[qnum]; vptr->td_used[qnum] > 0; + idx = (idx + 1) % vptr->options.numtx) { +@@ -1674,22 +1767,29 @@ + td = &(vptr->td_rings[qnum][idx]); + tdinfo = &(vptr->td_infos[qnum][idx]); + +- if (td->tdesc0.owner == OWNED_BY_NIC) ++ /* if (td->tdesc0.owner == OWNED_BY_NIC) BE */ ++ if (td->tdesc0 & cpu_to_le32(BE_OWNED_BY_NIC)) /* BE */ + break; + + if ((works++ > 15)) + break; + +- if (td->tdesc0.TSR & TSR0_TERR) { ++ wTSR = (u16)cpu_to_le32(td->tdesc0); ++ /* if (td->tdesc0.TSR & TSR0_TERR) { BE */ ++ if (wTSR & TSR0_TERR) { /* BE */ + stats->tx_errors++; + stats->tx_dropped++; +- if (td->tdesc0.TSR & TSR0_CDH) ++ /* if (td->tdesc0.TSR & TSR0_CDH) BE */ ++ if (wTSR & TSR0_CDH) /* BE */ + stats->tx_heartbeat_errors++; +- if (td->tdesc0.TSR & TSR0_CRS) ++ /* if (td->tdesc0.TSR & TSR0_CRS) BE */ ++ if (wTSR & TSR0_CRS) /* BE */ + stats->tx_carrier_errors++; +- if (td->tdesc0.TSR & TSR0_ABT) ++ /* if (td->tdesc0.TSR & TSR0_ABT) BE */ ++ if (wTSR & TSR0_ABT) /* BE */ + stats->tx_aborted_errors++; +- if (td->tdesc0.TSR & TSR0_OWC) ++ /* if (td->tdesc0.TSR & TSR0_OWC) BE */ ++ if (wTSR & TSR0_OWC) /* BE */ + stats->tx_window_errors++; + } else { + stats->tx_packets++; +@@ -1778,6 +1878,7 @@ + + static void velocity_error(struct velocity_info *vptr, int status) + { ++ HAILS("velocity_error", status); + + if (status & ISR_TXSTLI) { + struct mac_regs __iomem * regs = vptr->mac_regs; +@@ -1867,6 +1968,7 @@ + struct sk_buff *skb = tdinfo->skb; + int i; + ++ HAIL("velocity_free_tx_buf"); + /* + * Don't unmap the pre-allocated tx_bufs + */ +@@ -2067,6 +2169,7 @@ + struct velocity_td_info *tdinfo; + unsigned long flags; + int index; ++ u32 lbufsz; /* BE */ + + int pktlen = skb->len; + +@@ -2083,9 +2186,18 @@ + td_ptr = &(vptr->td_rings[qnum][index]); + tdinfo = &(vptr->td_infos[qnum][index]); + +- td_ptr->tdesc1.TCPLS = TCPLS_NORMAL; +- td_ptr->tdesc1.TCR = TCR0_TIC; +- td_ptr->td_buf[0].queue = 0; ++ td_ptr->tdesc0 = 0x00000000UL; /* BE */ ++ td_ptr->tdesc1 = 0x00000000UL; /* BE */ ++ ++ /* td_ptr->tdesc1.TCPLS = TCPLS_NORMAL; BE */ ++ td_ptr->tdesc1 &= cpu_to_le32(0xfcffffffUL); /* BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(((u32)TCPLS_NORMAL) << 24); /* BE */ ++ ++ /* td_ptr->tdesc1.TCR = TCR0_TIC; BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(BE_TCR_TIC); /* BE */ ++ ++ /* td_ptr->td_buf[0].queue = 0; BE */ ++ td_ptr->td_buf[0].ltwo &= cpu_to_le32(~BE_QUEUE_ENABLE); /* BE */ + + /* + * Pad short frames. +@@ -2097,20 +2209,36 @@ + memset(tdinfo->buf + skb->len, 0, ETH_ZLEN - skb->len); + tdinfo->skb = skb; + tdinfo->skb_dma[0] = tdinfo->buf_dma; +- td_ptr->tdesc0.pktsize = pktlen; ++ /* td_ptr->tdesc0.pktsize = pktlen; */ ++ td_ptr->tdesc0 &= cpu_to_le32(0xc000ffffUL); /* BE */ ++ lbufsz = pktlen; /* Assign, and make sure it's unsigned 32 bits - BE */ ++ lbufsz = lbufsz << 16; /* BE - shift over */ ++ td_ptr->tdesc0 |= cpu_to_le32(lbufsz); /* BE */ ++ + td_ptr->td_buf[0].pa_low = cpu_to_le32(tdinfo->skb_dma[0]); +- td_ptr->td_buf[0].pa_high = 0; +- td_ptr->td_buf[0].bufsize = td_ptr->tdesc0.pktsize; ++ /* td_ptr->td_buf[0].pa_high = 0; */ ++ /* td_ptr->td_buf[0].bufsize = td_ptr->tdesc0.pktsize; */ ++ td_ptr->td_buf[0].ltwo = cpu_to_le32(lbufsz); /* BE */ + tdinfo->nskb_dma = 1; +- td_ptr->tdesc1.CMDZ = 2; ++ /* td_ptr->tdesc1.CMDZ = 2; */ ++ td_ptr->tdesc1 &= cpu_to_le32(0x0fffffffUL); /* BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(((u32)0x2) << 28); /* BE */ + } else + #ifdef VELOCITY_ZERO_COPY_SUPPORT ++ /* ++ * BE - NOTE on the VELOCITY_ZERO_COPY_SUPPORT: ++ * This block of code has NOT been patched up for BE support, as ++ * it is certainly broken -- if it compiles at all. Since the BE ++ * fixes depend on the broken code, attempts to convert to BE support ++ * would almost certainly confuse more than help. ++ */ + if (skb_shinfo(skb)->nr_frags > 0) { + int nfrags = skb_shinfo(skb)->nr_frags; + tdinfo->skb = skb; + if (nfrags > 6) { + skb_copy_from_linear_data(skb, tdinfo->buf, skb->len); + tdinfo->skb_dma[0] = tdinfo->buf_dma; ++ /* BE: Er, exactly what value are we assigning in this next line? */ + td_ptr->tdesc0.pktsize = + td_ptr->td_buf[0].pa_low = cpu_to_le32(tdinfo->skb_dma[0]); + td_ptr->td_buf[0].pa_high = 0; +@@ -2127,6 +2255,7 @@ + /* FIXME: support 48bit DMA later */ + td_ptr->td_buf[i].pa_low = cpu_to_le32(tdinfo->skb_dma); + td_ptr->td_buf[i].pa_high = 0; ++ /* BE: This next line can't be right: */ + td_ptr->td_buf[i].bufsize = skb->len->skb->data_len; + + for (i = 0; i < nfrags; i++) { +@@ -2144,7 +2273,7 @@ + } + + } else +-#endif ++#endif /* (broken) VELOCITY_ZERO_COPY_SUPPORT */ + { + /* + * Map the linear network buffer into PCI space and +@@ -2152,19 +2281,29 @@ + */ + tdinfo->skb = skb; + tdinfo->skb_dma[0] = pci_map_single(vptr->pdev, skb->data, pktlen, PCI_DMA_TODEVICE); +- td_ptr->tdesc0.pktsize = pktlen; ++ /* td_ptr->tdesc0.pktsize = pktlen; BE */ ++ td_ptr->tdesc0 &= cpu_to_le32(0xc000ffffUL); /* BE */ ++ lbufsz = pktlen; /* Assign, and make sure it's unsigned 32 bits - BE */ ++ lbufsz = lbufsz << 16; /* BE */ ++ td_ptr->tdesc0 |= cpu_to_le32(lbufsz); /* BE */ + td_ptr->td_buf[0].pa_low = cpu_to_le32(tdinfo->skb_dma[0]); +- td_ptr->td_buf[0].pa_high = 0; +- td_ptr->td_buf[0].bufsize = td_ptr->tdesc0.pktsize; ++ /* td_ptr->td_buf[0].pa_high = 0; BE */ ++ /* td_ptr->td_buf[0].bufsize = td_ptr->tdesc0.pktsize; BE */ ++ td_ptr->td_buf[0].ltwo = cpu_to_le32(lbufsz); /* BE */ + tdinfo->nskb_dma = 1; +- td_ptr->tdesc1.CMDZ = 2; ++ /* td_ptr->tdesc1.CMDZ = 2; BE */ ++ td_ptr->tdesc1 &= cpu_to_le32(0x0fffffffUL); /* BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(((u32)0x2) << 28); /* BE */ + } + + if (vptr->vlgrp && vlan_tx_tag_present(skb)) { +- td_ptr->tdesc1.pqinf.VID = vlan_tx_tag_get(skb); +- td_ptr->tdesc1.pqinf.priority = 0; +- td_ptr->tdesc1.pqinf.CFI = 0; +- td_ptr->tdesc1.TCR |= TCR0_VETAG; ++ /* td_ptr->tdesc1.pqinf.VID = vlan_tx_tag_get(skb); BE */ ++ /* td_ptr->tdesc1.pqinf.priority = 0; BE */ ++ /* td_ptr->tdesc1.pqinf.CFI = 0; BE */ ++ /* td_ptr->tdesc1.TCR |= TCR0_VETAG; BE */ ++ td_ptr->tdesc1 &= cpu_to_le32(0xFFFF0000UL); /* BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(vlan_tx_tag_get(skb)); /* BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(BE_TCR_VETAG); /* BE */ + } + + /* +@@ -2174,26 +2313,34 @@ + && (skb->ip_summed == CHECKSUM_PARTIAL)) { + const struct iphdr *ip = ip_hdr(skb); + if (ip->protocol == IPPROTO_TCP) +- td_ptr->tdesc1.TCR |= TCR0_TCPCK; ++ /* td_ptr->tdesc1.TCR |= TCR0_TCPCK; BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(BE_TCR_TCPCK); /* BE */ + else if (ip->protocol == IPPROTO_UDP) +- td_ptr->tdesc1.TCR |= (TCR0_UDPCK); +- td_ptr->tdesc1.TCR |= TCR0_IPCK; +- } ++ /* td_ptr->tdesc1.TCR |= (TCR0_UDPCK); BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(BE_TCR_UDPCK); /* BE */ ++ /* td_ptr->tdesc1.TCR |= TCR0_IPCK; BE */ ++ td_ptr->tdesc1 |= cpu_to_le32(BE_TCR_IPCK); /* BE */ ++ } + { + + int prev = index - 1; + + if (prev < 0) + prev = vptr->options.numtx - 1; +- td_ptr->tdesc0.owner = OWNED_BY_NIC; ++ /* td_ptr->tdesc0.owner = OWNED_BY_NIC; BE */ ++ td_ptr->tdesc0 |= cpu_to_le32(BE_OWNED_BY_NIC); /* BE */ + vptr->td_used[qnum]++; + vptr->td_curr[qnum] = (index + 1) % vptr->options.numtx; + + if (AVAIL_TD(vptr, qnum) < 1) + netif_stop_queue(dev); + +- td_ptr = &(vptr->td_rings[qnum][prev]); +- td_ptr->td_buf[0].queue = 1; ++ td_ptr = &(vptr->td_rings[qnum][prev]); ++ /* td_ptr->td_buf[0].queue = 1; BE */ ++ td_ptr->td_buf[0].ltwo |= cpu_to_le32(BE_QUEUE_ENABLE); /* BE */ ++ if (vdebug&2) printk(KERN_NOTICE "velocity_xmit: (%s) len=%d idx=%d tdesc0=0x%x tdesc1=0x%x ltwo=0x%x\n", ++ (pktlen<ETH_ZLEN) ? "short" : "normal", pktlen, index, ++ td_ptr->tdesc0, td_ptr->tdesc1, td_ptr->td_buf[0].ltwo); + mac_tx_queue_wake(vptr->mac_regs, qnum); + } + dev->trans_start = jiffies; +@@ -2219,7 +2366,7 @@ + u32 isr_status; + int max_count = 0; + +- ++ HAIL("velocity_intr"); + spin_lock(&vptr->lock); + isr_status = mac_read_isr(vptr->mac_regs); + +@@ -2238,7 +2385,10 @@ + + while (isr_status != 0) { + mac_write_isr(vptr->mac_regs, isr_status); +- if (isr_status & (~(ISR_PRXI | ISR_PPRXI | ISR_PTXI | ISR_PPTXI))) ++ HAILS("velocity_intr",isr_status); ++ /* MJW - velocity_error is ALWAYS called; need to mask off some other flags */ ++ /* if (isr_status & (~(ISR_PRXI | ISR_PPRXI | ISR_PTXI | ISR_PPTXI))) */ ++ if (isr_status & (~(ISR_PRXI | ISR_PPRXI | ISR_PTXI | ISR_PPTXI | ISR_PTX0I | ISR_ISR0))) + velocity_error(vptr, isr_status); + if (isr_status & (ISR_PRXI | ISR_PPRXI)) + max_count += velocity_rx_srv(vptr, isr_status); +@@ -2276,6 +2426,7 @@ + int i; + struct dev_mc_list *mclist; + ++ HAIL("velocity_set_multi"); + if (dev->flags & IFF_PROMISC) { /* Set promiscuous. */ + writel(0xffffffff, ®s->MARCAM[0]); + writel(0xffffffff, ®s->MARCAM[4]); +@@ -2319,6 +2470,7 @@ + { + struct velocity_info *vptr = netdev_priv(dev); + ++ HAIL("net_device_stats"); + /* If the hardware is down, don't touch MII */ + if(!netif_running(dev)) + return &vptr->stats; +@@ -2363,6 +2515,7 @@ + struct velocity_info *vptr = netdev_priv(dev); + int ret; + ++ HAIL("velocity_ioctl"); + /* If we are asked for information and the device is power + saving then we need to bring the device back up to talk to it */ + +@@ -2581,6 +2734,8 @@ + { + u16 ww; + ++ HAIL("velocity_mii_read"); ++ HAIL("velocity_mii_write"); + /* + * Disable MIICR_MAUTO, so that mii addr can be set normally + */ +diff -Naur linux-2.6.24.orig/drivers/net/via-velocity.h linux-2.6.24/drivers/net/via-velocity.h +--- linux-2.6.24.orig/drivers/net/via-velocity.h 2008-01-31 23:11:26.000000000 -0600 ++++ linux-2.6.24/drivers/net/via-velocity.h 2008-02-01 01:12:15.000000000 -0600 +@@ -196,64 +196,70 @@ + * Receive descriptor + */ + +-struct rdesc0 { +- u16 RSR; /* Receive status */ +- u16 len:14; /* Received packet length */ +- u16 reserved:1; +- u16 owner:1; /* Who owns this buffer ? */ +-}; +- +-struct rdesc1 { +- u16 PQTAG; +- u8 CSM; +- u8 IPKT; +-}; ++//struct rdesc0 { ++// u16 RSR; /* Receive status */ ++// u16 len:14; /* Received packet length */ ++// u16 reserved:1; ++// u16 owner:1; /* Who owns this buffer ? */ ++//}; ++ ++//struct rdesc1 { ++// u16 PQTAG; ++// u8 CSM; ++// u8 IPKT; ++//}; + + struct rx_desc { +- struct rdesc0 rdesc0; +- struct rdesc1 rdesc1; ++// struct rdesc0 rdesc0; ++// struct rdesc1 rdesc1; ++ u32 rdesc0; ++ u32 rdesc1; + u32 pa_low; /* Low 32 bit PCI address */ +- u16 pa_high; /* Next 16 bit PCI address (48 total) */ +- u16 len:15; /* Frame size */ +- u16 inten:1; /* Enable interrupt */ ++// u16 pa_high; /* Next 16 bit PCI address (48 total) */ ++// u16 len:15; /* Frame size */ ++// u16 inten:1; /* Enable interrupt */ ++ u32 ltwo; + } __attribute__ ((__packed__)); + + /* + * Transmit descriptor + */ + +-struct tdesc0 { +- u16 TSR; /* Transmit status register */ +- u16 pktsize:14; /* Size of frame */ +- u16 reserved:1; +- u16 owner:1; /* Who owns the buffer */ +-}; +- +-struct pqinf { /* Priority queue info */ +- u16 VID:12; +- u16 CFI:1; +- u16 priority:3; +-} __attribute__ ((__packed__)); +- +-struct tdesc1 { +- struct pqinf pqinf; +- u8 TCR; +- u8 TCPLS:2; +- u8 reserved:2; +- u8 CMDZ:4; +-} __attribute__ ((__packed__)); ++//struct tdesc0 { ++// u16 TSR; /* Transmit status register */ ++// u16 pktsize:14; /* Size of frame */ ++// u16 reserved:1; ++// u16 owner:1; /* Who owns the buffer */ ++//}; ++ ++//struct pqinf { /* Priority queue info */ ++// u16 VID:12; ++// u16 CFI:1; ++// u16 priority:3; ++//} __attribute__ ((__packed__)); ++ ++//struct tdesc1 { ++// struct pqinf pqinf; ++// u8 TCR; ++// u8 TCPLS:2; ++// u8 reserved:2; ++// u8 CMDZ:4; ++//} __attribute__ ((__packed__)); + + struct td_buf { + u32 pa_low; +- u16 pa_high; +- u16 bufsize:14; +- u16 reserved:1; +- u16 queue:1; ++// u16 pa_high; ++// u16 bufsize:14; ++// u16 reserved:1; ++// u16 queue:1; ++ u32 ltwo; + } __attribute__ ((__packed__)); + + struct tx_desc { +- struct tdesc0 tdesc0; +- struct tdesc1 tdesc1; ++// struct tdesc0 tdesc0; ++// struct tdesc1 tdesc1; ++ u32 tdesc0; ++ u32 tdesc1; + struct td_buf td_buf[7]; + }; + +@@ -279,6 +285,16 @@ + OWNED_BY_NIC = 1 + }; + ++/* Constants added for the BE fixes */ ++#define BE_OWNED_BY_NIC 0x80000000UL ++#define BE_INT_ENABLE 0x80000000UL ++#define BE_QUEUE_ENABLE 0x80000000UL ++#define BE_TCR_TIC 0x00800000UL ++#define BE_TCR_VETAG 0x00200000UL ++#define BE_TCR_TCPCK 0x00040000UL ++#define BE_TCR_UDPCK 0x00080000UL ++#define BE_TCR_IPCK 0x00100000UL ++ + + /* + * MAC registers and macros. +@@ -1512,6 +1528,7 @@ + }; + + struct velocity_opt { ++ int velo_debug; /* debug flag */ + int numrx; /* Number of RX descriptors */ + int numtx; /* Number of TX descriptors */ + enum speed_opt spd_dpx; /* Media link mode */ |