Linux-2.6.12-rc2
Initial git repository build. I'm not bothering with the full history, even though we have it. We can create a separate "historical" git archive of that later if we want to, and in the meantime it's about 3.2GB when imported into git - space that would just make the early git days unnecessarily complicated, when we don't have a lot of good infrastructure for it. Let it rip!
This commit is contained in:
786
drivers/net/tulip/interrupt.c
Normal file
786
drivers/net/tulip/interrupt.c
Normal file
@@ -0,0 +1,786 @@
|
||||
/*
|
||||
drivers/net/tulip/interrupt.c
|
||||
|
||||
Maintained by Jeff Garzik <jgarzik@pobox.com>
|
||||
Copyright 2000,2001 The Linux Kernel Team
|
||||
Written/copyright 1994-2001 by Donald Becker.
|
||||
|
||||
This software may be used and distributed according to the terms
|
||||
of the GNU General Public License, incorporated herein by reference.
|
||||
|
||||
Please refer to Documentation/DocBook/tulip-user.{pdf,ps,html}
|
||||
for more information on this driver, or visit the project
|
||||
Web page at http://sourceforge.net/projects/tulip/
|
||||
|
||||
*/
|
||||
|
||||
#include <linux/pci.h>
|
||||
#include "tulip.h"
|
||||
#include <linux/config.h>
|
||||
#include <linux/etherdevice.h>
|
||||
|
||||
int tulip_rx_copybreak;
|
||||
unsigned int tulip_max_interrupt_work;
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
|
||||
#define MIT_SIZE 15
|
||||
#define MIT_TABLE 15 /* We use 0 or max */
|
||||
|
||||
static unsigned int mit_table[MIT_SIZE+1] =
|
||||
{
|
||||
/* CRS11 21143 hardware Mitigation Control Interrupt
|
||||
We use only RX mitigation we other techniques for
|
||||
TX intr. mitigation.
|
||||
|
||||
31 Cycle Size (timer control)
|
||||
30:27 TX timer in 16 * Cycle size
|
||||
26:24 TX No pkts before Int.
|
||||
23:20 RX timer in Cycle size
|
||||
19:17 RX No pkts before Int.
|
||||
16 Continues Mode (CM)
|
||||
*/
|
||||
|
||||
0x0, /* IM disabled */
|
||||
0x80150000, /* RX time = 1, RX pkts = 2, CM = 1 */
|
||||
0x80150000,
|
||||
0x80270000,
|
||||
0x80370000,
|
||||
0x80490000,
|
||||
0x80590000,
|
||||
0x80690000,
|
||||
0x807B0000,
|
||||
0x808B0000,
|
||||
0x809D0000,
|
||||
0x80AD0000,
|
||||
0x80BD0000,
|
||||
0x80CF0000,
|
||||
0x80DF0000,
|
||||
// 0x80FF0000 /* RX time = 16, RX pkts = 7, CM = 1 */
|
||||
0x80F10000 /* RX time = 16, RX pkts = 0, CM = 1 */
|
||||
};
|
||||
#endif
|
||||
|
||||
|
||||
int tulip_refill_rx(struct net_device *dev)
|
||||
{
|
||||
struct tulip_private *tp = netdev_priv(dev);
|
||||
int entry;
|
||||
int refilled = 0;
|
||||
|
||||
/* Refill the Rx ring buffers. */
|
||||
for (; tp->cur_rx - tp->dirty_rx > 0; tp->dirty_rx++) {
|
||||
entry = tp->dirty_rx % RX_RING_SIZE;
|
||||
if (tp->rx_buffers[entry].skb == NULL) {
|
||||
struct sk_buff *skb;
|
||||
dma_addr_t mapping;
|
||||
|
||||
skb = tp->rx_buffers[entry].skb = dev_alloc_skb(PKT_BUF_SZ);
|
||||
if (skb == NULL)
|
||||
break;
|
||||
|
||||
mapping = pci_map_single(tp->pdev, skb->tail, PKT_BUF_SZ,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
tp->rx_buffers[entry].mapping = mapping;
|
||||
|
||||
skb->dev = dev; /* Mark as being used by this device. */
|
||||
tp->rx_ring[entry].buffer1 = cpu_to_le32(mapping);
|
||||
refilled++;
|
||||
}
|
||||
tp->rx_ring[entry].status = cpu_to_le32(DescOwned);
|
||||
}
|
||||
if(tp->chip_id == LC82C168) {
|
||||
if(((ioread32(tp->base_addr + CSR5)>>17)&0x07) == 4) {
|
||||
/* Rx stopped due to out of buffers,
|
||||
* restart it
|
||||
*/
|
||||
iowrite32(0x01, tp->base_addr + CSR2);
|
||||
}
|
||||
}
|
||||
return refilled;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI
|
||||
|
||||
void oom_timer(unsigned long data)
|
||||
{
|
||||
struct net_device *dev = (struct net_device *)data;
|
||||
netif_rx_schedule(dev);
|
||||
}
|
||||
|
||||
int tulip_poll(struct net_device *dev, int *budget)
|
||||
{
|
||||
struct tulip_private *tp = netdev_priv(dev);
|
||||
int entry = tp->cur_rx % RX_RING_SIZE;
|
||||
int rx_work_limit = *budget;
|
||||
int received = 0;
|
||||
|
||||
if (!netif_running(dev))
|
||||
goto done;
|
||||
|
||||
if (rx_work_limit > dev->quota)
|
||||
rx_work_limit = dev->quota;
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
|
||||
|
||||
/* that one buffer is needed for mit activation; or might be a
|
||||
bug in the ring buffer code; check later -- JHS*/
|
||||
|
||||
if (rx_work_limit >=RX_RING_SIZE) rx_work_limit--;
|
||||
#endif
|
||||
|
||||
if (tulip_debug > 4)
|
||||
printk(KERN_DEBUG " In tulip_rx(), entry %d %8.8x.\n", entry,
|
||||
tp->rx_ring[entry].status);
|
||||
|
||||
do {
|
||||
if (ioread32(tp->base_addr + CSR5) == 0xffffffff) {
|
||||
printk(KERN_DEBUG " In tulip_poll(), hardware disappeared.\n");
|
||||
break;
|
||||
}
|
||||
/* Acknowledge current RX interrupt sources. */
|
||||
iowrite32((RxIntr | RxNoBuf), tp->base_addr + CSR5);
|
||||
|
||||
|
||||
/* If we own the next entry, it is a new packet. Send it up. */
|
||||
while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
|
||||
s32 status = le32_to_cpu(tp->rx_ring[entry].status);
|
||||
|
||||
|
||||
if (tp->dirty_rx + RX_RING_SIZE == tp->cur_rx)
|
||||
break;
|
||||
|
||||
if (tulip_debug > 5)
|
||||
printk(KERN_DEBUG "%s: In tulip_rx(), entry %d %8.8x.\n",
|
||||
dev->name, entry, status);
|
||||
if (--rx_work_limit < 0)
|
||||
goto not_done;
|
||||
|
||||
if ((status & 0x38008300) != 0x0300) {
|
||||
if ((status & 0x38000300) != 0x0300) {
|
||||
/* Ingore earlier buffers. */
|
||||
if ((status & 0xffff) != 0x7fff) {
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_WARNING "%s: Oversized Ethernet frame "
|
||||
"spanned multiple buffers, status %8.8x!\n",
|
||||
dev->name, status);
|
||||
tp->stats.rx_length_errors++;
|
||||
}
|
||||
} else if (status & RxDescFatalErr) {
|
||||
/* There was a fatal error. */
|
||||
if (tulip_debug > 2)
|
||||
printk(KERN_DEBUG "%s: Receive error, Rx status %8.8x.\n",
|
||||
dev->name, status);
|
||||
tp->stats.rx_errors++; /* end of a packet.*/
|
||||
if (status & 0x0890) tp->stats.rx_length_errors++;
|
||||
if (status & 0x0004) tp->stats.rx_frame_errors++;
|
||||
if (status & 0x0002) tp->stats.rx_crc_errors++;
|
||||
if (status & 0x0001) tp->stats.rx_fifo_errors++;
|
||||
}
|
||||
} else {
|
||||
/* Omit the four octet CRC from the length. */
|
||||
short pkt_len = ((status >> 16) & 0x7ff) - 4;
|
||||
struct sk_buff *skb;
|
||||
|
||||
#ifndef final_version
|
||||
if (pkt_len > 1518) {
|
||||
printk(KERN_WARNING "%s: Bogus packet size of %d (%#x).\n",
|
||||
dev->name, pkt_len, pkt_len);
|
||||
pkt_len = 1518;
|
||||
tp->stats.rx_length_errors++;
|
||||
}
|
||||
#endif
|
||||
/* Check if the packet is long enough to accept without copying
|
||||
to a minimally-sized skbuff. */
|
||||
if (pkt_len < tulip_rx_copybreak
|
||||
&& (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
|
||||
skb->dev = dev;
|
||||
skb_reserve(skb, 2); /* 16 byte align the IP header */
|
||||
pci_dma_sync_single_for_cpu(tp->pdev,
|
||||
tp->rx_buffers[entry].mapping,
|
||||
pkt_len, PCI_DMA_FROMDEVICE);
|
||||
#if ! defined(__alpha__)
|
||||
eth_copy_and_sum(skb, tp->rx_buffers[entry].skb->tail,
|
||||
pkt_len, 0);
|
||||
skb_put(skb, pkt_len);
|
||||
#else
|
||||
memcpy(skb_put(skb, pkt_len),
|
||||
tp->rx_buffers[entry].skb->tail,
|
||||
pkt_len);
|
||||
#endif
|
||||
pci_dma_sync_single_for_device(tp->pdev,
|
||||
tp->rx_buffers[entry].mapping,
|
||||
pkt_len, PCI_DMA_FROMDEVICE);
|
||||
} else { /* Pass up the skb already on the Rx ring. */
|
||||
char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
|
||||
pkt_len);
|
||||
|
||||
#ifndef final_version
|
||||
if (tp->rx_buffers[entry].mapping !=
|
||||
le32_to_cpu(tp->rx_ring[entry].buffer1)) {
|
||||
printk(KERN_ERR "%s: Internal fault: The skbuff addresses "
|
||||
"do not match in tulip_rx: %08x vs. %08llx %p / %p.\n",
|
||||
dev->name,
|
||||
le32_to_cpu(tp->rx_ring[entry].buffer1),
|
||||
(unsigned long long)tp->rx_buffers[entry].mapping,
|
||||
skb->head, temp);
|
||||
}
|
||||
#endif
|
||||
|
||||
pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
|
||||
PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
|
||||
|
||||
tp->rx_buffers[entry].skb = NULL;
|
||||
tp->rx_buffers[entry].mapping = 0;
|
||||
}
|
||||
skb->protocol = eth_type_trans(skb, dev);
|
||||
|
||||
netif_receive_skb(skb);
|
||||
|
||||
dev->last_rx = jiffies;
|
||||
tp->stats.rx_packets++;
|
||||
tp->stats.rx_bytes += pkt_len;
|
||||
}
|
||||
received++;
|
||||
|
||||
entry = (++tp->cur_rx) % RX_RING_SIZE;
|
||||
if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/4)
|
||||
tulip_refill_rx(dev);
|
||||
|
||||
}
|
||||
|
||||
/* New ack strategy... irq does not ack Rx any longer
|
||||
hopefully this helps */
|
||||
|
||||
/* Really bad things can happen here... If new packet arrives
|
||||
* and an irq arrives (tx or just due to occasionally unset
|
||||
* mask), it will be acked by irq handler, but new thread
|
||||
* is not scheduled. It is major hole in design.
|
||||
* No idea how to fix this if "playing with fire" will fail
|
||||
* tomorrow (night 011029). If it will not fail, we won
|
||||
* finally: amount of IO did not increase at all. */
|
||||
} while ((ioread32(tp->base_addr + CSR5) & RxIntr));
|
||||
|
||||
done:
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI_HW_MITIGATION
|
||||
|
||||
/* We use this simplistic scheme for IM. It's proven by
|
||||
real life installations. We can have IM enabled
|
||||
continuesly but this would cause unnecessary latency.
|
||||
Unfortunely we can't use all the NET_RX_* feedback here.
|
||||
This would turn on IM for devices that is not contributing
|
||||
to backlog congestion with unnecessary latency.
|
||||
|
||||
We monitor the the device RX-ring and have:
|
||||
|
||||
HW Interrupt Mitigation either ON or OFF.
|
||||
|
||||
ON: More then 1 pkt received (per intr.) OR we are dropping
|
||||
OFF: Only 1 pkt received
|
||||
|
||||
Note. We only use min and max (0, 15) settings from mit_table */
|
||||
|
||||
|
||||
if( tp->flags & HAS_INTR_MITIGATION) {
|
||||
if( received > 1 ) {
|
||||
if( ! tp->mit_on ) {
|
||||
tp->mit_on = 1;
|
||||
iowrite32(mit_table[MIT_TABLE], tp->base_addr + CSR11);
|
||||
}
|
||||
}
|
||||
else {
|
||||
if( tp->mit_on ) {
|
||||
tp->mit_on = 0;
|
||||
iowrite32(0, tp->base_addr + CSR11);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#endif /* CONFIG_TULIP_NAPI_HW_MITIGATION */
|
||||
|
||||
dev->quota -= received;
|
||||
*budget -= received;
|
||||
|
||||
tulip_refill_rx(dev);
|
||||
|
||||
/* If RX ring is not full we are out of memory. */
|
||||
if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) goto oom;
|
||||
|
||||
/* Remove us from polling list and enable RX intr. */
|
||||
|
||||
netif_rx_complete(dev);
|
||||
iowrite32(tulip_tbl[tp->chip_id].valid_intrs, tp->base_addr+CSR7);
|
||||
|
||||
/* The last op happens after poll completion. Which means the following:
|
||||
* 1. it can race with disabling irqs in irq handler
|
||||
* 2. it can race with dise/enabling irqs in other poll threads
|
||||
* 3. if an irq raised after beginning loop, it will be immediately
|
||||
* triggered here.
|
||||
*
|
||||
* Summarizing: the logic results in some redundant irqs both
|
||||
* due to races in masking and due to too late acking of already
|
||||
* processed irqs. But it must not result in losing events.
|
||||
*/
|
||||
|
||||
return 0;
|
||||
|
||||
not_done:
|
||||
if (!received) {
|
||||
|
||||
received = dev->quota; /* Not to happen */
|
||||
}
|
||||
dev->quota -= received;
|
||||
*budget -= received;
|
||||
|
||||
if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/2 ||
|
||||
tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
|
||||
tulip_refill_rx(dev);
|
||||
|
||||
if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) goto oom;
|
||||
|
||||
return 1;
|
||||
|
||||
|
||||
oom: /* Executed with RX ints disabled */
|
||||
|
||||
|
||||
/* Start timer, stop polling, but do not enable rx interrupts. */
|
||||
mod_timer(&tp->oom_timer, jiffies+1);
|
||||
|
||||
/* Think: timer_pending() was an explicit signature of bug.
|
||||
* Timer can be pending now but fired and completed
|
||||
* before we did netif_rx_complete(). See? We would lose it. */
|
||||
|
||||
/* remove ourselves from the polling list */
|
||||
netif_rx_complete(dev);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
#else /* CONFIG_TULIP_NAPI */
|
||||
|
||||
static int tulip_rx(struct net_device *dev)
|
||||
{
|
||||
struct tulip_private *tp = netdev_priv(dev);
|
||||
int entry = tp->cur_rx % RX_RING_SIZE;
|
||||
int rx_work_limit = tp->dirty_rx + RX_RING_SIZE - tp->cur_rx;
|
||||
int received = 0;
|
||||
|
||||
if (tulip_debug > 4)
|
||||
printk(KERN_DEBUG " In tulip_rx(), entry %d %8.8x.\n", entry,
|
||||
tp->rx_ring[entry].status);
|
||||
/* If we own the next entry, it is a new packet. Send it up. */
|
||||
while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) {
|
||||
s32 status = le32_to_cpu(tp->rx_ring[entry].status);
|
||||
|
||||
if (tulip_debug > 5)
|
||||
printk(KERN_DEBUG "%s: In tulip_rx(), entry %d %8.8x.\n",
|
||||
dev->name, entry, status);
|
||||
if (--rx_work_limit < 0)
|
||||
break;
|
||||
if ((status & 0x38008300) != 0x0300) {
|
||||
if ((status & 0x38000300) != 0x0300) {
|
||||
/* Ingore earlier buffers. */
|
||||
if ((status & 0xffff) != 0x7fff) {
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_WARNING "%s: Oversized Ethernet frame "
|
||||
"spanned multiple buffers, status %8.8x!\n",
|
||||
dev->name, status);
|
||||
tp->stats.rx_length_errors++;
|
||||
}
|
||||
} else if (status & RxDescFatalErr) {
|
||||
/* There was a fatal error. */
|
||||
if (tulip_debug > 2)
|
||||
printk(KERN_DEBUG "%s: Receive error, Rx status %8.8x.\n",
|
||||
dev->name, status);
|
||||
tp->stats.rx_errors++; /* end of a packet.*/
|
||||
if (status & 0x0890) tp->stats.rx_length_errors++;
|
||||
if (status & 0x0004) tp->stats.rx_frame_errors++;
|
||||
if (status & 0x0002) tp->stats.rx_crc_errors++;
|
||||
if (status & 0x0001) tp->stats.rx_fifo_errors++;
|
||||
}
|
||||
} else {
|
||||
/* Omit the four octet CRC from the length. */
|
||||
short pkt_len = ((status >> 16) & 0x7ff) - 4;
|
||||
struct sk_buff *skb;
|
||||
|
||||
#ifndef final_version
|
||||
if (pkt_len > 1518) {
|
||||
printk(KERN_WARNING "%s: Bogus packet size of %d (%#x).\n",
|
||||
dev->name, pkt_len, pkt_len);
|
||||
pkt_len = 1518;
|
||||
tp->stats.rx_length_errors++;
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Check if the packet is long enough to accept without copying
|
||||
to a minimally-sized skbuff. */
|
||||
if (pkt_len < tulip_rx_copybreak
|
||||
&& (skb = dev_alloc_skb(pkt_len + 2)) != NULL) {
|
||||
skb->dev = dev;
|
||||
skb_reserve(skb, 2); /* 16 byte align the IP header */
|
||||
pci_dma_sync_single_for_cpu(tp->pdev,
|
||||
tp->rx_buffers[entry].mapping,
|
||||
pkt_len, PCI_DMA_FROMDEVICE);
|
||||
#if ! defined(__alpha__)
|
||||
eth_copy_and_sum(skb, tp->rx_buffers[entry].skb->tail,
|
||||
pkt_len, 0);
|
||||
skb_put(skb, pkt_len);
|
||||
#else
|
||||
memcpy(skb_put(skb, pkt_len),
|
||||
tp->rx_buffers[entry].skb->tail,
|
||||
pkt_len);
|
||||
#endif
|
||||
pci_dma_sync_single_for_device(tp->pdev,
|
||||
tp->rx_buffers[entry].mapping,
|
||||
pkt_len, PCI_DMA_FROMDEVICE);
|
||||
} else { /* Pass up the skb already on the Rx ring. */
|
||||
char *temp = skb_put(skb = tp->rx_buffers[entry].skb,
|
||||
pkt_len);
|
||||
|
||||
#ifndef final_version
|
||||
if (tp->rx_buffers[entry].mapping !=
|
||||
le32_to_cpu(tp->rx_ring[entry].buffer1)) {
|
||||
printk(KERN_ERR "%s: Internal fault: The skbuff addresses "
|
||||
"do not match in tulip_rx: %08x vs. %Lx %p / %p.\n",
|
||||
dev->name,
|
||||
le32_to_cpu(tp->rx_ring[entry].buffer1),
|
||||
(long long)tp->rx_buffers[entry].mapping,
|
||||
skb->head, temp);
|
||||
}
|
||||
#endif
|
||||
|
||||
pci_unmap_single(tp->pdev, tp->rx_buffers[entry].mapping,
|
||||
PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
|
||||
|
||||
tp->rx_buffers[entry].skb = NULL;
|
||||
tp->rx_buffers[entry].mapping = 0;
|
||||
}
|
||||
skb->protocol = eth_type_trans(skb, dev);
|
||||
|
||||
netif_rx(skb);
|
||||
|
||||
dev->last_rx = jiffies;
|
||||
tp->stats.rx_packets++;
|
||||
tp->stats.rx_bytes += pkt_len;
|
||||
}
|
||||
received++;
|
||||
entry = (++tp->cur_rx) % RX_RING_SIZE;
|
||||
}
|
||||
return received;
|
||||
}
|
||||
#endif /* CONFIG_TULIP_NAPI */
|
||||
|
||||
static inline unsigned int phy_interrupt (struct net_device *dev)
|
||||
{
|
||||
#ifdef __hppa__
|
||||
struct tulip_private *tp = netdev_priv(dev);
|
||||
int csr12 = ioread32(tp->base_addr + CSR12) & 0xff;
|
||||
|
||||
if (csr12 != tp->csr12_shadow) {
|
||||
/* ack interrupt */
|
||||
iowrite32(csr12 | 0x02, tp->base_addr + CSR12);
|
||||
tp->csr12_shadow = csr12;
|
||||
/* do link change stuff */
|
||||
spin_lock(&tp->lock);
|
||||
tulip_check_duplex(dev);
|
||||
spin_unlock(&tp->lock);
|
||||
/* clear irq ack bit */
|
||||
iowrite32(csr12 & ~0x02, tp->base_addr + CSR12);
|
||||
|
||||
return 1;
|
||||
}
|
||||
#endif
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* The interrupt handler does all of the Rx thread work and cleans up
|
||||
after the Tx thread. */
|
||||
irqreturn_t tulip_interrupt(int irq, void *dev_instance, struct pt_regs *regs)
|
||||
{
|
||||
struct net_device *dev = (struct net_device *)dev_instance;
|
||||
struct tulip_private *tp = netdev_priv(dev);
|
||||
void __iomem *ioaddr = tp->base_addr;
|
||||
int csr5;
|
||||
int missed;
|
||||
int rx = 0;
|
||||
int tx = 0;
|
||||
int oi = 0;
|
||||
int maxrx = RX_RING_SIZE;
|
||||
int maxtx = TX_RING_SIZE;
|
||||
int maxoi = TX_RING_SIZE;
|
||||
#ifdef CONFIG_TULIP_NAPI
|
||||
int rxd = 0;
|
||||
#else
|
||||
int entry;
|
||||
#endif
|
||||
unsigned int work_count = tulip_max_interrupt_work;
|
||||
unsigned int handled = 0;
|
||||
|
||||
/* Let's see whether the interrupt really is for us */
|
||||
csr5 = ioread32(ioaddr + CSR5);
|
||||
|
||||
if (tp->flags & HAS_PHY_IRQ)
|
||||
handled = phy_interrupt (dev);
|
||||
|
||||
if ((csr5 & (NormalIntr|AbnormalIntr)) == 0)
|
||||
return IRQ_RETVAL(handled);
|
||||
|
||||
tp->nir++;
|
||||
|
||||
do {
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI
|
||||
|
||||
if (!rxd && (csr5 & (RxIntr | RxNoBuf))) {
|
||||
rxd++;
|
||||
/* Mask RX intrs and add the device to poll list. */
|
||||
iowrite32(tulip_tbl[tp->chip_id].valid_intrs&~RxPollInt, ioaddr + CSR7);
|
||||
netif_rx_schedule(dev);
|
||||
|
||||
if (!(csr5&~(AbnormalIntr|NormalIntr|RxPollInt|TPLnkPass)))
|
||||
break;
|
||||
}
|
||||
|
||||
/* Acknowledge the interrupt sources we handle here ASAP
|
||||
the poll function does Rx and RxNoBuf acking */
|
||||
|
||||
iowrite32(csr5 & 0x0001ff3f, ioaddr + CSR5);
|
||||
|
||||
#else
|
||||
/* Acknowledge all of the current interrupt sources ASAP. */
|
||||
iowrite32(csr5 & 0x0001ffff, ioaddr + CSR5);
|
||||
|
||||
|
||||
if (csr5 & (RxIntr | RxNoBuf)) {
|
||||
rx += tulip_rx(dev);
|
||||
tulip_refill_rx(dev);
|
||||
}
|
||||
|
||||
#endif /* CONFIG_TULIP_NAPI */
|
||||
|
||||
if (tulip_debug > 4)
|
||||
printk(KERN_DEBUG "%s: interrupt csr5=%#8.8x new csr5=%#8.8x.\n",
|
||||
dev->name, csr5, ioread32(ioaddr + CSR5));
|
||||
|
||||
|
||||
if (csr5 & (TxNoBuf | TxDied | TxIntr | TimerInt)) {
|
||||
unsigned int dirty_tx;
|
||||
|
||||
spin_lock(&tp->lock);
|
||||
|
||||
for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0;
|
||||
dirty_tx++) {
|
||||
int entry = dirty_tx % TX_RING_SIZE;
|
||||
int status = le32_to_cpu(tp->tx_ring[entry].status);
|
||||
|
||||
if (status < 0)
|
||||
break; /* It still has not been Txed */
|
||||
|
||||
/* Check for Rx filter setup frames. */
|
||||
if (tp->tx_buffers[entry].skb == NULL) {
|
||||
/* test because dummy frames not mapped */
|
||||
if (tp->tx_buffers[entry].mapping)
|
||||
pci_unmap_single(tp->pdev,
|
||||
tp->tx_buffers[entry].mapping,
|
||||
sizeof(tp->setup_frame),
|
||||
PCI_DMA_TODEVICE);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (status & 0x8000) {
|
||||
/* There was an major error, log it. */
|
||||
#ifndef final_version
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_DEBUG "%s: Transmit error, Tx status %8.8x.\n",
|
||||
dev->name, status);
|
||||
#endif
|
||||
tp->stats.tx_errors++;
|
||||
if (status & 0x4104) tp->stats.tx_aborted_errors++;
|
||||
if (status & 0x0C00) tp->stats.tx_carrier_errors++;
|
||||
if (status & 0x0200) tp->stats.tx_window_errors++;
|
||||
if (status & 0x0002) tp->stats.tx_fifo_errors++;
|
||||
if ((status & 0x0080) && tp->full_duplex == 0)
|
||||
tp->stats.tx_heartbeat_errors++;
|
||||
} else {
|
||||
tp->stats.tx_bytes +=
|
||||
tp->tx_buffers[entry].skb->len;
|
||||
tp->stats.collisions += (status >> 3) & 15;
|
||||
tp->stats.tx_packets++;
|
||||
}
|
||||
|
||||
pci_unmap_single(tp->pdev, tp->tx_buffers[entry].mapping,
|
||||
tp->tx_buffers[entry].skb->len,
|
||||
PCI_DMA_TODEVICE);
|
||||
|
||||
/* Free the original skb. */
|
||||
dev_kfree_skb_irq(tp->tx_buffers[entry].skb);
|
||||
tp->tx_buffers[entry].skb = NULL;
|
||||
tp->tx_buffers[entry].mapping = 0;
|
||||
tx++;
|
||||
}
|
||||
|
||||
#ifndef final_version
|
||||
if (tp->cur_tx - dirty_tx > TX_RING_SIZE) {
|
||||
printk(KERN_ERR "%s: Out-of-sync dirty pointer, %d vs. %d.\n",
|
||||
dev->name, dirty_tx, tp->cur_tx);
|
||||
dirty_tx += TX_RING_SIZE;
|
||||
}
|
||||
#endif
|
||||
|
||||
if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2)
|
||||
netif_wake_queue(dev);
|
||||
|
||||
tp->dirty_tx = dirty_tx;
|
||||
if (csr5 & TxDied) {
|
||||
if (tulip_debug > 2)
|
||||
printk(KERN_WARNING "%s: The transmitter stopped."
|
||||
" CSR5 is %x, CSR6 %x, new CSR6 %x.\n",
|
||||
dev->name, csr5, ioread32(ioaddr + CSR6), tp->csr6);
|
||||
tulip_restart_rxtx(tp);
|
||||
}
|
||||
spin_unlock(&tp->lock);
|
||||
}
|
||||
|
||||
/* Log errors. */
|
||||
if (csr5 & AbnormalIntr) { /* Abnormal error summary bit. */
|
||||
if (csr5 == 0xffffffff)
|
||||
break;
|
||||
if (csr5 & TxJabber) tp->stats.tx_errors++;
|
||||
if (csr5 & TxFIFOUnderflow) {
|
||||
if ((tp->csr6 & 0xC000) != 0xC000)
|
||||
tp->csr6 += 0x4000; /* Bump up the Tx threshold */
|
||||
else
|
||||
tp->csr6 |= 0x00200000; /* Store-n-forward. */
|
||||
/* Restart the transmit process. */
|
||||
tulip_restart_rxtx(tp);
|
||||
iowrite32(0, ioaddr + CSR1);
|
||||
}
|
||||
if (csr5 & (RxDied | RxNoBuf)) {
|
||||
if (tp->flags & COMET_MAC_ADDR) {
|
||||
iowrite32(tp->mc_filter[0], ioaddr + 0xAC);
|
||||
iowrite32(tp->mc_filter[1], ioaddr + 0xB0);
|
||||
}
|
||||
}
|
||||
if (csr5 & RxDied) { /* Missed a Rx frame. */
|
||||
tp->stats.rx_missed_errors += ioread32(ioaddr + CSR8) & 0xffff;
|
||||
tp->stats.rx_errors++;
|
||||
tulip_start_rxtx(tp);
|
||||
}
|
||||
/*
|
||||
* NB: t21142_lnk_change() does a del_timer_sync(), so be careful if this
|
||||
* call is ever done under the spinlock
|
||||
*/
|
||||
if (csr5 & (TPLnkPass | TPLnkFail | 0x08000000)) {
|
||||
if (tp->link_change)
|
||||
(tp->link_change)(dev, csr5);
|
||||
}
|
||||
if (csr5 & SytemError) {
|
||||
int error = (csr5 >> 23) & 7;
|
||||
/* oops, we hit a PCI error. The code produced corresponds
|
||||
* to the reason:
|
||||
* 0 - parity error
|
||||
* 1 - master abort
|
||||
* 2 - target abort
|
||||
* Note that on parity error, we should do a software reset
|
||||
* of the chip to get it back into a sane state (according
|
||||
* to the 21142/3 docs that is).
|
||||
* -- rmk
|
||||
*/
|
||||
printk(KERN_ERR "%s: (%lu) System Error occurred (%d)\n",
|
||||
dev->name, tp->nir, error);
|
||||
}
|
||||
/* Clear all error sources, included undocumented ones! */
|
||||
iowrite32(0x0800f7ba, ioaddr + CSR5);
|
||||
oi++;
|
||||
}
|
||||
if (csr5 & TimerInt) {
|
||||
|
||||
if (tulip_debug > 2)
|
||||
printk(KERN_ERR "%s: Re-enabling interrupts, %8.8x.\n",
|
||||
dev->name, csr5);
|
||||
iowrite32(tulip_tbl[tp->chip_id].valid_intrs, ioaddr + CSR7);
|
||||
tp->ttimer = 0;
|
||||
oi++;
|
||||
}
|
||||
if (tx > maxtx || rx > maxrx || oi > maxoi) {
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_WARNING "%s: Too much work during an interrupt, "
|
||||
"csr5=0x%8.8x. (%lu) (%d,%d,%d)\n", dev->name, csr5, tp->nir, tx, rx, oi);
|
||||
|
||||
/* Acknowledge all interrupt sources. */
|
||||
iowrite32(0x8001ffff, ioaddr + CSR5);
|
||||
if (tp->flags & HAS_INTR_MITIGATION) {
|
||||
/* Josip Loncaric at ICASE did extensive experimentation
|
||||
to develop a good interrupt mitigation setting.*/
|
||||
iowrite32(0x8b240000, ioaddr + CSR11);
|
||||
} else if (tp->chip_id == LC82C168) {
|
||||
/* the LC82C168 doesn't have a hw timer.*/
|
||||
iowrite32(0x00, ioaddr + CSR7);
|
||||
mod_timer(&tp->timer, RUN_AT(HZ/50));
|
||||
} else {
|
||||
/* Mask all interrupting sources, set timer to
|
||||
re-enable. */
|
||||
iowrite32(((~csr5) & 0x0001ebef) | AbnormalIntr | TimerInt, ioaddr + CSR7);
|
||||
iowrite32(0x0012, ioaddr + CSR11);
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
||||
work_count--;
|
||||
if (work_count == 0)
|
||||
break;
|
||||
|
||||
csr5 = ioread32(ioaddr + CSR5);
|
||||
|
||||
#ifdef CONFIG_TULIP_NAPI
|
||||
if (rxd)
|
||||
csr5 &= ~RxPollInt;
|
||||
} while ((csr5 & (TxNoBuf |
|
||||
TxDied |
|
||||
TxIntr |
|
||||
TimerInt |
|
||||
/* Abnormal intr. */
|
||||
RxDied |
|
||||
TxFIFOUnderflow |
|
||||
TxJabber |
|
||||
TPLnkFail |
|
||||
SytemError )) != 0);
|
||||
#else
|
||||
} while ((csr5 & (NormalIntr|AbnormalIntr)) != 0);
|
||||
|
||||
tulip_refill_rx(dev);
|
||||
|
||||
/* check if the card is in suspend mode */
|
||||
entry = tp->dirty_rx % RX_RING_SIZE;
|
||||
if (tp->rx_buffers[entry].skb == NULL) {
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_WARNING "%s: in rx suspend mode: (%lu) (tp->cur_rx = %u, ttimer = %d, rx = %d) go/stay in suspend mode\n", dev->name, tp->nir, tp->cur_rx, tp->ttimer, rx);
|
||||
if (tp->chip_id == LC82C168) {
|
||||
iowrite32(0x00, ioaddr + CSR7);
|
||||
mod_timer(&tp->timer, RUN_AT(HZ/50));
|
||||
} else {
|
||||
if (tp->ttimer == 0 || (ioread32(ioaddr + CSR11) & 0xffff) == 0) {
|
||||
if (tulip_debug > 1)
|
||||
printk(KERN_WARNING "%s: in rx suspend mode: (%lu) set timer\n", dev->name, tp->nir);
|
||||
iowrite32(tulip_tbl[tp->chip_id].valid_intrs | TimerInt,
|
||||
ioaddr + CSR7);
|
||||
iowrite32(TimerInt, ioaddr + CSR5);
|
||||
iowrite32(12, ioaddr + CSR11);
|
||||
tp->ttimer = 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
#endif /* CONFIG_TULIP_NAPI */
|
||||
|
||||
if ((missed = ioread32(ioaddr + CSR8) & 0x1ffff)) {
|
||||
tp->stats.rx_dropped += missed & 0x10000 ? 0x10000 : missed;
|
||||
}
|
||||
|
||||
if (tulip_debug > 4)
|
||||
printk(KERN_DEBUG "%s: exiting interrupt, csr5=%#4.4x.\n",
|
||||
dev->name, ioread32(ioaddr + CSR5));
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
Reference in New Issue
Block a user