Commit e958d4a0 authored by Felipe Damasio's avatar Felipe Damasio Committed by Jeff Garzik

[netdrvr 3c527] remove cli/sti


    Richard Procter and I worked to remove cli/sti to add proper SMP support (I did the original stuff and Richard did the actual current code :)).

    Besides that, Richard did a great jog improving the perfomance of the driver quite a bit:

    - Improve mc32_command by 770% (438% non-inlined) over the semaphore version (at a cost of 1 sem + 2 completions per driver).

    - Removed mutex covering mc32_send_packet (hard_start_xmit). This lets the interrupt handler operate concurrently and removes unnecessary locking. It makes the code only slightly more brittle

    Original post:

http://marc.theaimsgroup.com/?l=linux-netdev&m=106438449315202&w=2

    Since it didn't apply cleanly against 2.6.0-test6, I forward ported it. Patch attached.

    Jeff, please consider applying,

    Thanks.
parent 4576b4fa
/* 3c527.c: 3Com Etherlink/MC32 driver for Linux 2.4 /* 3c527.c: 3Com Etherlink/MC32 driver for Linux 2.4 and 2.6.
* *
* (c) Copyright 1998 Red Hat Software Inc * (c) Copyright 1998 Red Hat Software Inc
* Written by Alan Cox. * Written by Alan Cox.
* Further debugging by Carl Drougge. * Further debugging by Carl Drougge.
* Modified by Richard Procter (rnp@netlink.co.nz) * Initial SMP support by Felipe W Damasio <felipewd@terra.com.br>
* Heavily modified by Richard Procter <rnp@paradise.net.nz>
* *
* Based on skeleton.c written 1993-94 by Donald Becker and ne2.c * Based on skeleton.c written 1993-94 by Donald Becker and ne2.c
* (for the MCA stuff) written by Wim Dumon. * (for the MCA stuff) written by Wim Dumon.
...@@ -17,11 +18,11 @@ ...@@ -17,11 +18,11 @@
*/ */
#define DRV_NAME "3c527" #define DRV_NAME "3c527"
#define DRV_VERSION "0.6a" #define DRV_VERSION "0.7-SMP"
#define DRV_RELDATE "2001/11/17" #define DRV_RELDATE "2003/10/06"
static const char *version = static const char *version =
DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " Richard Proctor (rnp@netlink.co.nz)\n"; DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " Richard Procter <rnp@paradise.net.nz>\n";
/** /**
* DOC: Traps for the unwary * DOC: Traps for the unwary
...@@ -100,7 +101,9 @@ DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " Richard Proctor (rnp@netlink.co.nz ...@@ -100,7 +101,9 @@ DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " Richard Proctor (rnp@netlink.co.nz
#include <linux/string.h> #include <linux/string.h>
#include <linux/wait.h> #include <linux/wait.h>
#include <linux/ethtool.h> #include <linux/ethtool.h>
#include <linux/completion.h>
#include <asm/semaphore.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/bitops.h> #include <asm/bitops.h>
...@@ -154,8 +157,10 @@ struct mc32_ring_desc ...@@ -154,8 +157,10 @@ struct mc32_ring_desc
/* Information that needs to be kept for each board. */ /* Information that needs to be kept for each board. */
struct mc32_local struct mc32_local
{ {
struct net_device_stats net_stats;
int slot; int slot;
u32 base;
struct net_device_stats net_stats;
volatile struct mc32_mailbox *rx_box; volatile struct mc32_mailbox *rx_box;
volatile struct mc32_mailbox *tx_box; volatile struct mc32_mailbox *tx_box;
volatile struct mc32_mailbox *exec_box; volatile struct mc32_mailbox *exec_box;
...@@ -165,22 +170,22 @@ struct mc32_local ...@@ -165,22 +170,22 @@ struct mc32_local
u16 tx_len; /* Transmit list count */ u16 tx_len; /* Transmit list count */
u16 rx_len; /* Receive list count */ u16 rx_len; /* Receive list count */
u32 base; u16 xceiver_desired_state; /* HALTED or RUNNING */
u16 exec_pending; u16 cmd_nonblocking; /* Thread is uninterested in command result */
u16 mc_reload_wait; /* a multicast load request is pending */ u16 mc_reload_wait; /* A multicast load request is pending */
u32 mc_list_valid; /* True when the mclist is set */ u32 mc_list_valid; /* True when the mclist is set */
u16 xceiver_state; /* Current transceiver state. bitmapped */
u16 desired_state; /* The state we want the transceiver to be in */
atomic_t tx_count; /* buffers left */
wait_queue_head_t event;
struct mc32_ring_desc tx_ring[TX_RING_LEN]; /* Host Transmit ring */ struct mc32_ring_desc tx_ring[TX_RING_LEN]; /* Host Transmit ring */
struct mc32_ring_desc rx_ring[RX_RING_LEN]; /* Host Receive ring */ struct mc32_ring_desc rx_ring[RX_RING_LEN]; /* Host Receive ring */
u16 tx_ring_tail; /* index to tx de-queue end */ atomic_t tx_count; /* buffers left */
u16 tx_ring_head; /* index to tx en-queue end */ volatile u16 tx_ring_head; /* index to tx en-queue end */
u16 tx_ring_tail; /* index to tx de-queue end */
u16 rx_ring_tail; /* index to rx de-queue end */ u16 rx_ring_tail; /* index to rx de-queue end */
struct semaphore cmd_mutex; /* Serialises issuing of execute commands */
struct completion execution_cmd; /* Card has completed an execute command */
struct completion xceiver_cmd; /* Card has completed a tx or rx command */
}; };
/* The station (ethernet) address prefix, used for a sanity check. */ /* The station (ethernet) address prefix, used for a sanity check. */
...@@ -236,7 +241,6 @@ int __init mc32_probe(struct net_device *dev) ...@@ -236,7 +241,6 @@ int __init mc32_probe(struct net_device *dev)
{ {
static int current_mca_slot = -1; static int current_mca_slot = -1;
int i; int i;
int adapter_found = 0;
SET_MODULE_OWNER(dev); SET_MODULE_OWNER(dev);
...@@ -247,11 +251,11 @@ int __init mc32_probe(struct net_device *dev) ...@@ -247,11 +251,11 @@ int __init mc32_probe(struct net_device *dev)
Autodetecting MCA cards is extremely simple. Autodetecting MCA cards is extremely simple.
Just search for the card. */ Just search for the card. */
for(i = 0; (mc32_adapters[i].name != NULL) && !adapter_found; i++) { for(i = 0; (mc32_adapters[i].name != NULL); i++) {
current_mca_slot = current_mca_slot =
mca_find_unused_adapter(mc32_adapters[i].id, 0); mca_find_unused_adapter(mc32_adapters[i].id, 0);
if((current_mca_slot != MCA_NOTFOUND) && !adapter_found) { if(current_mca_slot != MCA_NOTFOUND) {
if(!mc32_probe1(dev, current_mca_slot)) if(!mc32_probe1(dev, current_mca_slot))
{ {
mca_set_adapter_name(current_mca_slot, mca_set_adapter_name(current_mca_slot,
...@@ -409,7 +413,7 @@ static int __init mc32_probe1(struct net_device *dev, int slot) ...@@ -409,7 +413,7 @@ static int __init mc32_probe1(struct net_device *dev, int slot)
* Grab the IRQ * Grab the IRQ
*/ */
i = request_irq(dev->irq, &mc32_interrupt, SA_SHIRQ, dev->name, dev); i = request_irq(dev->irq, &mc32_interrupt, SA_SHIRQ | SA_SAMPLE_RANDOM, dev->name, dev);
if (i) { if (i) {
release_region(dev->base_addr, MC32_IO_EXTENT); release_region(dev->base_addr, MC32_IO_EXTENT);
printk(KERN_ERR "%s: unable to get IRQ %d.\n", dev->name, dev->irq); printk(KERN_ERR "%s: unable to get IRQ %d.\n", dev->name, dev->irq);
...@@ -498,7 +502,9 @@ static int __init mc32_probe1(struct net_device *dev, int slot) ...@@ -498,7 +502,9 @@ static int __init mc32_probe1(struct net_device *dev, int slot)
lp->tx_len = lp->exec_box->data[9]; /* Transmit list count */ lp->tx_len = lp->exec_box->data[9]; /* Transmit list count */
lp->rx_len = lp->exec_box->data[11]; /* Receive list count */ lp->rx_len = lp->exec_box->data[11]; /* Receive list count */
init_waitqueue_head(&lp->event); init_MUTEX_LOCKED(&lp->cmd_mutex);
init_completion(&lp->execution_cmd);
init_completion(&lp->xceiver_cmd);
printk("%s: Firmware Rev %d. %d RX buffers, %d TX buffers. Base of 0x%08X.\n", printk("%s: Firmware Rev %d. %d RX buffers, %d TX buffers. Base of 0x%08X.\n",
dev->name, lp->exec_box->data[12], lp->rx_len, lp->tx_len, lp->base); dev->name, lp->exec_box->data[12], lp->rx_len, lp->tx_len, lp->base);
...@@ -512,10 +518,6 @@ static int __init mc32_probe1(struct net_device *dev, int slot) ...@@ -512,10 +518,6 @@ static int __init mc32_probe1(struct net_device *dev, int slot)
dev->watchdog_timeo = HZ*5; /* Board does all the work */ dev->watchdog_timeo = HZ*5; /* Board does all the work */
dev->ethtool_ops = &netdev_ethtool_ops; dev->ethtool_ops = &netdev_ethtool_ops;
lp->xceiver_state = HALTED;
lp->tx_ring_tail=lp->tx_ring_head=0;
/* Fill in the fields of the device structure with ethernet values. */ /* Fill in the fields of the device structure with ethernet values. */
ether_setup(dev); ether_setup(dev);
...@@ -539,7 +541,7 @@ static int __init mc32_probe1(struct net_device *dev, int slot) ...@@ -539,7 +541,7 @@ static int __init mc32_probe1(struct net_device *dev, int slot)
* status of any pending commands and takes very little time at all. * status of any pending commands and takes very little time at all.
*/ */
static void mc32_ready_poll(struct net_device *dev) static inline void mc32_ready_poll(struct net_device *dev)
{ {
int ioaddr = dev->base_addr; int ioaddr = dev->base_addr;
while(!(inb(ioaddr+HOST_STATUS)&HOST_STATUS_CRR)); while(!(inb(ioaddr+HOST_STATUS)&HOST_STATUS_CRR));
...@@ -554,31 +556,36 @@ static void mc32_ready_poll(struct net_device *dev) ...@@ -554,31 +556,36 @@ static void mc32_ready_poll(struct net_device *dev)
* @len: Length of the data block * @len: Length of the data block
* *
* Send a command from interrupt state. If there is a command * Send a command from interrupt state. If there is a command
* currently being executed then we return an error of -1. It simply * currently being executed then we return an error of -1. It
* isn't viable to wait around as commands may be slow. Providing we * simply isn't viable to wait around as commands may be
* get in, we busy wait for the board to become ready to accept the * slow. This can theoretically be starved on SMP, but it's hard
* command and issue it. We do not wait for the command to complete * to see a realistic situation. We do not wait for the command
* --- the card will interrupt us when it's done. * to complete --- we rely on the interrupt handler to tidy up
* after us.
*/ */
static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int len) static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int len)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
int ioaddr = dev->base_addr; int ioaddr = dev->base_addr;
int ret = -1;
if(lp->exec_pending)
return -1;
lp->exec_pending=3; if (down_trylock(&lp->cmd_mutex) == 0)
lp->exec_box->mbox=0; {
lp->exec_box->mbox=cmd; lp->cmd_nonblocking=1;
memcpy((void *)lp->exec_box->data, data, len); lp->exec_box->mbox=0;
barrier(); /* the memcpy forgot the volatile so be sure */ lp->exec_box->mbox=cmd;
memcpy((void *)lp->exec_box->data, data, len);
barrier(); /* the memcpy forgot the volatile so be sure */
/* Send the command */
mc32_ready_poll(dev);
outb(1<<6, ioaddr+HOST_CMD);
/* Send the command */ ret = 0;
while(!(inb(ioaddr+HOST_STATUS)&HOST_STATUS_CRR)); /* Interrupt handler will signal mutex on completion */
outb(1<<6, ioaddr+HOST_CMD); }
return 0; return ret;
} }
...@@ -592,7 +599,7 @@ static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int ...@@ -592,7 +599,7 @@ static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int
* Sends exec commands in a user context. This permits us to wait around * Sends exec commands in a user context. This permits us to wait around
* for the replies and also to wait for the command buffer to complete * for the replies and also to wait for the command buffer to complete
* from a previous command before we execute our command. After our * from a previous command before we execute our command. After our
* command completes we will complete any pending multicast reload * command completes we will attempt any pending multicast reload
* we blocked off by hogging the exec buffer. * we blocked off by hogging the exec buffer.
* *
* You feed the card a command, you wait, it interrupts you get a * You feed the card a command, you wait, it interrupts you get a
...@@ -600,66 +607,37 @@ static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int ...@@ -600,66 +607,37 @@ static int mc32_command_nowait(struct net_device *dev, u16 cmd, void *data, int
* commands for filter list changes which come in at bh level from things * commands for filter list changes which come in at bh level from things
* like IPV6 group stuff. * like IPV6 group stuff.
* *
* We have a simple state machine
*
* 0 - nothing issued
*
* 1 - command issued, wait reply
*
* 2 - reply waiting - reader then goes to state 0
*
* 3 - command issued, trash reply. In which case the irq
* takes it back to state 0
*
*/ */
static int mc32_command(struct net_device *dev, u16 cmd, void *data, int len) static int mc32_command(struct net_device *dev, u16 cmd, void *data, int len)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
int ioaddr = dev->base_addr; int ioaddr = dev->base_addr;
unsigned long flags;
int ret = 0; int ret = 0;
down(&lp->cmd_mutex);
/* /*
* Wait for a command * My turn
*/
save_flags(flags);
cli();
while(lp->exec_pending)
sleep_on(&lp->event);
/*
* Issue mine
*/ */
lp->exec_pending=1; lp->cmd_nonblocking=0;
restore_flags(flags);
lp->exec_box->mbox=0; lp->exec_box->mbox=0;
lp->exec_box->mbox=cmd; lp->exec_box->mbox=cmd;
memcpy((void *)lp->exec_box->data, data, len); memcpy((void *)lp->exec_box->data, data, len);
barrier(); /* the memcpy forgot the volatile so be sure */ barrier(); /* the memcpy forgot the volatile so be sure */
/* Send the command */ mc32_ready_poll(dev);
while(!(inb(ioaddr+HOST_STATUS)&HOST_STATUS_CRR)); outb(1<<6, ioaddr+HOST_CMD);
outb(1<<6, ioaddr+HOST_CMD);
save_flags(flags);
cli();
while(lp->exec_pending!=2) wait_for_completion(&lp->execution_cmd);
sleep_on(&lp->event);
lp->exec_pending=0;
restore_flags(flags);
if(lp->exec_box->mbox&(1<<13)) if(lp->exec_box->mbox&(1<<13))
ret = -1; ret = -1;
up(&lp->cmd_mutex);
/* /*
* A multicast set got blocked - do it now * A multicast set got blocked - try it now
*/ */
if(lp->mc_reload_wait) if(lp->mc_reload_wait)
...@@ -678,10 +656,8 @@ static int mc32_command(struct net_device *dev, u16 cmd, void *data, int len) ...@@ -678,10 +656,8 @@ static int mc32_command(struct net_device *dev, u16 cmd, void *data, int len)
* This may be called from the interrupt state, where it is used * This may be called from the interrupt state, where it is used
* to restart the rx ring if the card runs out of rx buffers. * to restart the rx ring if the card runs out of rx buffers.
* *
* First, we check if it's ok to start the transceiver. We then show * We must first check if it's ok to (re)start the transceiver. See
* the card where to start in the rx ring and issue the * mc32_close for details.
* commands to start reception and transmission. We don't wait
* around for these to complete.
*/ */
static void mc32_start_transceiver(struct net_device *dev) { static void mc32_start_transceiver(struct net_device *dev) {
...@@ -690,24 +666,20 @@ static void mc32_start_transceiver(struct net_device *dev) { ...@@ -690,24 +666,20 @@ static void mc32_start_transceiver(struct net_device *dev) {
int ioaddr = dev->base_addr; int ioaddr = dev->base_addr;
/* Ignore RX overflow on device closure */ /* Ignore RX overflow on device closure */
if (lp->desired_state==HALTED) if (lp->xceiver_desired_state==HALTED)
return; return;
/* Give the card the offset to the post-EOL-bit RX descriptor */
mc32_ready_poll(dev); mc32_ready_poll(dev);
lp->tx_box->mbox=0;
lp->rx_box->mbox=0; lp->rx_box->mbox=0;
/* Give the card the offset to the post-EOL-bit RX descriptor */
lp->rx_box->data[0]=lp->rx_ring[prev_rx(lp->rx_ring_tail)].p->next; lp->rx_box->data[0]=lp->rx_ring[prev_rx(lp->rx_ring_tail)].p->next;
outb(HOST_CMD_START_RX, ioaddr+HOST_CMD); outb(HOST_CMD_START_RX, ioaddr+HOST_CMD);
mc32_ready_poll(dev); mc32_ready_poll(dev);
lp->tx_box->mbox=0;
outb(HOST_CMD_RESTRT_TX, ioaddr+HOST_CMD); /* card ignores this on RX restart */ outb(HOST_CMD_RESTRT_TX, ioaddr+HOST_CMD); /* card ignores this on RX restart */
/* We are not interrupted on start completion */ /* We are not interrupted on start completion */
lp->xceiver_state=RUNNING;
} }
...@@ -727,24 +699,18 @@ static void mc32_halt_transceiver(struct net_device *dev) ...@@ -727,24 +699,18 @@ static void mc32_halt_transceiver(struct net_device *dev)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
int ioaddr = dev->base_addr; int ioaddr = dev->base_addr;
unsigned long flags;
mc32_ready_poll(dev); mc32_ready_poll(dev);
lp->tx_box->mbox=0;
lp->rx_box->mbox=0; lp->rx_box->mbox=0;
outb(HOST_CMD_SUSPND_RX, ioaddr+HOST_CMD); outb(HOST_CMD_SUSPND_RX, ioaddr+HOST_CMD);
wait_for_completion(&lp->xceiver_cmd);
mc32_ready_poll(dev); mc32_ready_poll(dev);
lp->tx_box->mbox=0;
outb(HOST_CMD_SUSPND_TX, ioaddr+HOST_CMD); outb(HOST_CMD_SUSPND_TX, ioaddr+HOST_CMD);
wait_for_completion(&lp->xceiver_cmd);
save_flags(flags);
cli();
while(lp->xceiver_state!=HALTED)
sleep_on(&lp->event);
restore_flags(flags);
} }
...@@ -756,7 +722,7 @@ static void mc32_halt_transceiver(struct net_device *dev) ...@@ -756,7 +722,7 @@ static void mc32_halt_transceiver(struct net_device *dev)
* the point where mc32_start_transceiver() can be called. * the point where mc32_start_transceiver() can be called.
* *
* The card sets up the receive ring for us. We are required to use the * The card sets up the receive ring for us. We are required to use the
* ring it provides although we can change the size of the ring. * ring it provides, although the size of the ring is configurable.
* *
* We allocate an sk_buff for each ring entry in turn and * We allocate an sk_buff for each ring entry in turn and
* initalise its house-keeping info. At the same time, we read * initalise its house-keeping info. At the same time, we read
...@@ -777,7 +743,7 @@ static int mc32_load_rx_ring(struct net_device *dev) ...@@ -777,7 +743,7 @@ static int mc32_load_rx_ring(struct net_device *dev)
rx_base=lp->rx_chain; rx_base=lp->rx_chain;
for(i=0;i<RX_RING_LEN;i++) for(i=0; i<RX_RING_LEN; i++)
{ {
lp->rx_ring[i].skb=alloc_skb(1532, GFP_KERNEL); lp->rx_ring[i].skb=alloc_skb(1532, GFP_KERNEL);
skb_reserve(lp->rx_ring[i].skb, 18); skb_reserve(lp->rx_ring[i].skb, 18);
...@@ -814,21 +780,20 @@ static int mc32_load_rx_ring(struct net_device *dev) ...@@ -814,21 +780,20 @@ static int mc32_load_rx_ring(struct net_device *dev)
* *
* Free the buffer for each ring slot. This may be called * Free the buffer for each ring slot. This may be called
* before mc32_load_rx_ring(), eg. on error in mc32_open(). * before mc32_load_rx_ring(), eg. on error in mc32_open().
* Requires rx skb pointers to point to a valid skb, or NULL.
*/ */
static void mc32_flush_rx_ring(struct net_device *dev) static void mc32_flush_rx_ring(struct net_device *dev)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
struct sk_buff *skb;
int i; int i;
for(i=0; i < RX_RING_LEN; i++) for(i=0; i < RX_RING_LEN; i++)
{ {
skb = lp->rx_ring[i].skb; if (lp->rx_ring[i].skb) {
if (skb!=NULL) { dev_kfree_skb(lp->rx_ring[i].skb);
kfree_skb(skb); lp->rx_ring[i].skb = NULL;
skb=NULL;
} }
lp->rx_ring[i].p=NULL; lp->rx_ring[i].p=NULL;
} }
...@@ -860,7 +825,7 @@ static void mc32_load_tx_ring(struct net_device *dev) ...@@ -860,7 +825,7 @@ static void mc32_load_tx_ring(struct net_device *dev)
tx_base=lp->tx_box->data[0]; tx_base=lp->tx_box->data[0];
for(i=0;i<lp->tx_len;i++) for(i=0 ; i<TX_RING_LEN ; i++)
{ {
p=isa_bus_to_virt(lp->base+tx_base); p=isa_bus_to_virt(lp->base+tx_base);
lp->tx_ring[i].p=p; lp->tx_ring[i].p=p;
...@@ -870,8 +835,8 @@ static void mc32_load_tx_ring(struct net_device *dev) ...@@ -870,8 +835,8 @@ static void mc32_load_tx_ring(struct net_device *dev)
} }
/* -1 so that tx_ring_head cannot "lap" tx_ring_tail, */ /* -1 so that tx_ring_head cannot "lap" tx_ring_tail, */
/* which would be bad news for mc32_tx_ring as cur. implemented */ /* see mc32_tx_ring */
atomic_set(&lp->tx_count, TX_RING_LEN-1); atomic_set(&lp->tx_count, TX_RING_LEN-1);
lp->tx_ring_head=lp->tx_ring_tail=0; lp->tx_ring_head=lp->tx_ring_tail=0;
} }
...@@ -880,45 +845,27 @@ static void mc32_load_tx_ring(struct net_device *dev) ...@@ -880,45 +845,27 @@ static void mc32_load_tx_ring(struct net_device *dev)
/** /**
* mc32_flush_tx_ring - free transmit ring * mc32_flush_tx_ring - free transmit ring
* @lp: Local data of 3c527 to flush the tx ring of * @lp: Local data of 3c527 to flush the tx ring of
* *
* We have to consider two cases here. We want to free the pending * If the ring is non-empty, zip over the it, freeing any
* buffers only. If the ring buffer head is past the start then the * allocated skb_buffs. The tx ring house-keeping variables are
* ring segment we wish to free wraps through zero. The tx ring * then reset. Requires rx skb pointers to point to a valid skb,
* house-keeping variables are then reset. * or NULL.
*/ */
static void mc32_flush_tx_ring(struct net_device *dev) static void mc32_flush_tx_ring(struct net_device *dev)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
int i;
if(lp->tx_ring_tail!=lp->tx_ring_head) for (i=0; i < TX_RING_LEN; i++)
{ {
int i; if (lp->tx_ring[i].skb)
if(lp->tx_ring_tail < lp->tx_ring_head)
{
for(i=lp->tx_ring_tail;i<lp->tx_ring_head;i++)
{
dev_kfree_skb(lp->tx_ring[i].skb);
lp->tx_ring[i].skb=NULL;
lp->tx_ring[i].p=NULL;
}
}
else
{ {
for(i=lp->tx_ring_tail; i<TX_RING_LEN; i++) dev_kfree_skb(lp->tx_ring[i].skb);
{ lp->tx_ring[i].skb = NULL;
dev_kfree_skb(lp->tx_ring[i].skb); }
lp->tx_ring[i].skb=NULL; }
lp->tx_ring[i].p=NULL;
}
for(i=0; i<lp->tx_ring_head; i++)
{
dev_kfree_skb(lp->tx_ring[i].skb);
lp->tx_ring[i].skb=NULL;
lp->tx_ring[i].p=NULL;
}
}
}
atomic_set(&lp->tx_count, 0); atomic_set(&lp->tx_count, 0);
lp->tx_ring_tail=lp->tx_ring_head=0; lp->tx_ring_tail=lp->tx_ring_head=0;
...@@ -957,8 +904,13 @@ static int mc32_open(struct net_device *dev) ...@@ -957,8 +904,13 @@ static int mc32_open(struct net_device *dev)
regs=inb(ioaddr+HOST_CTRL); regs=inb(ioaddr+HOST_CTRL);
regs|=HOST_CTRL_INTE; regs|=HOST_CTRL_INTE;
outb(regs, ioaddr+HOST_CTRL); outb(regs, ioaddr+HOST_CTRL);
/*
* Allow ourselves to issue commands
*/
up(&lp->cmd_mutex);
/* /*
* Send the indications on command * Send the indications on command
*/ */
...@@ -1010,7 +962,7 @@ static int mc32_open(struct net_device *dev) ...@@ -1010,7 +962,7 @@ static int mc32_open(struct net_device *dev)
return -ENOBUFS; return -ENOBUFS;
} }
lp->desired_state = RUNNING; lp->xceiver_desired_state = RUNNING;
/* And finally, set the ball rolling... */ /* And finally, set the ball rolling... */
mc32_start_transceiver(dev); mc32_start_transceiver(dev);
...@@ -1047,61 +999,64 @@ static void mc32_timeout(struct net_device *dev) ...@@ -1047,61 +999,64 @@ static void mc32_timeout(struct net_device *dev)
* Transmit a buffer. This normally means throwing the buffer onto * Transmit a buffer. This normally means throwing the buffer onto
* the transmit queue as the queue is quite large. If the queue is * the transmit queue as the queue is quite large. If the queue is
* full then we set tx_busy and return. Once the interrupt handler * full then we set tx_busy and return. Once the interrupt handler
* gets messages telling it to reclaim transmit queue entries we will * gets messages telling it to reclaim transmit queue entries, we will
* clear tx_busy and the kernel will start calling this again. * clear tx_busy and the kernel will start calling this again.
* *
* We use cli rather than spinlocks. Since I have no access to an SMP * We do not disable interrupts or acquire any locks; this can
* MCA machine I don't plan to change it. It is probably the top * run concurrently with mc32_tx_ring(), and the function itself
* performance hit for this driver on SMP however. * is serialised at a higher layer. However, this makes it
* crucial that we update lp->tx_ring_head only after we've
* established a valid packet in the tx ring (and is why we mark
* tx_ring_head volatile).
*/ */
static int mc32_send_packet(struct sk_buff *skb, struct net_device *dev) static int mc32_send_packet(struct sk_buff *skb, struct net_device *dev)
{ {
struct mc32_local *lp = (struct mc32_local *)dev->priv; struct mc32_local *lp = (struct mc32_local *)dev->priv;
unsigned long flags;
u16 head = lp->tx_ring_head;
volatile struct skb_header *p, *np; volatile struct skb_header *p, *np;
netif_stop_queue(dev); netif_stop_queue(dev);
save_flags(flags); if(atomic_read(&lp->tx_count)==0) {
cli();
if(atomic_read(&lp->tx_count)==0)
{
restore_flags(flags);
return 1; return 1;
} }
skb = skb_padto(skb, ETH_ZLEN);
if (skb == NULL) {
netif_wake_queue(dev);
return 0;
}
atomic_dec(&lp->tx_count); atomic_dec(&lp->tx_count);
/* P is the last sending/sent buffer as a pointer */ /* P is the last sending/sent buffer as a pointer */
p=lp->tx_ring[lp->tx_ring_head].p; p=lp->tx_ring[head].p;
lp->tx_ring_head=next_tx(lp->tx_ring_head); head = next_tx(head);
/* NP is the buffer we will be loading */ /* NP is the buffer we will be loading */
np=lp->tx_ring[lp->tx_ring_head].p; np=lp->tx_ring[head].p;
/* We will need this to flush the buffer out */
lp->tx_ring[head].skb=skb;
if (skb->len < ETH_ZLEN) { np->length = unlikely(skb->len < ETH_ZLEN) ? ETH_ZLEN : skb->len;
skb = skb_padto(skb, ETH_ZLEN);
if (skb == NULL)
goto out;
}
/* We will need this to flush the buffer out */
lp->tx_ring[lp->tx_ring_head].skb = skb;
np->length = (skb->len < ETH_ZLEN) ? ETH_ZLEN : skb->len;
np->data = isa_virt_to_bus(skb->data); np->data = isa_virt_to_bus(skb->data);
np->status = 0; np->status = 0;
np->control = CONTROL_EOP | CONTROL_EOL; np->control = CONTROL_EOP | CONTROL_EOL;
wmb(); wmb();
p->control &= ~CONTROL_EOL; /* Clear EOL on p */ /*
out: * The new frame has been setup; we can now
restore_flags(flags); * let the card and interrupt handler "see" it
*/
p->control &= ~CONTROL_EOL;
lp->tx_ring_head= head;
netif_wake_queue(dev); netif_wake_queue(dev);
return 0; return 0;
...@@ -1182,11 +1137,11 @@ static void mc32_rx_ring(struct net_device *dev) ...@@ -1182,11 +1137,11 @@ static void mc32_rx_ring(struct net_device *dev)
{ {
struct mc32_local *lp=dev->priv; struct mc32_local *lp=dev->priv;
volatile struct skb_header *p; volatile struct skb_header *p;
u16 rx_ring_tail = lp->rx_ring_tail; u16 rx_ring_tail;
u16 rx_old_tail = rx_ring_tail; u16 rx_old_tail;
int x=0; int x=0;
rx_old_tail = rx_ring_tail = lp->rx_ring_tail;
do do
{ {
p=lp->rx_ring[rx_ring_tail].p; p=lp->rx_ring[rx_ring_tail].p;
...@@ -1275,8 +1230,13 @@ static void mc32_tx_ring(struct net_device *dev) ...@@ -1275,8 +1230,13 @@ static void mc32_tx_ring(struct net_device *dev)
struct mc32_local *lp=(struct mc32_local *)dev->priv; struct mc32_local *lp=(struct mc32_local *)dev->priv;
volatile struct skb_header *np; volatile struct skb_header *np;
/* NB: lp->tx_count=TX_RING_LEN-1 so that tx_ring_head cannot "lap" tail here */ /*
* We rely on head==tail to mean 'queue empty'.
* This is why lp->tx_count=TX_RING_LEN-1: in order to prevent
* tx_ring_head wrapping to tail and confusing a 'queue empty'
* condition with 'queue full'
*/
while (lp->tx_ring_tail != lp->tx_ring_head) while (lp->tx_ring_tail != lp->tx_ring_head)
{ {
u16 t; u16 t;
...@@ -1388,8 +1348,7 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs) ...@@ -1388,8 +1348,7 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs)
break; break;
case 3: /* Halt */ case 3: /* Halt */
case 4: /* Abort */ case 4: /* Abort */
lp->xceiver_state |= TX_HALTED; complete(&lp->xceiver_cmd);
wake_up(&lp->event);
break; break;
default: default:
printk("%s: strange tx ack %d\n", dev->name, status&7); printk("%s: strange tx ack %d\n", dev->name, status&7);
...@@ -1404,8 +1363,7 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs) ...@@ -1404,8 +1363,7 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs)
break; break;
case 3: /* Halt */ case 3: /* Halt */
case 4: /* Abort */ case 4: /* Abort */
lp->xceiver_state |= RX_HALTED; complete(&lp->xceiver_cmd);
wake_up(&lp->event);
break; break;
case 6: case 6:
/* Out of RX buffers stat */ /* Out of RX buffers stat */
...@@ -1421,26 +1379,17 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs) ...@@ -1421,26 +1379,17 @@ static irqreturn_t mc32_interrupt(int irq, void *dev_id, struct pt_regs * regs)
status>>=3; status>>=3;
if(status&1) if(status&1)
{ {
/*
* No thread is waiting: we need to tidy
* up ourself.
*/
/* 0=no 1=yes 2=replied, get cmd, 3 = wait reply & dump it */ if (lp->cmd_nonblocking) {
up(&lp->cmd_mutex);
if(lp->exec_pending!=3) {
lp->exec_pending=2;
wake_up(&lp->event);
}
else
{
lp->exec_pending=0;
/* A new multicast set may have been
blocked while the old one was
running. If so, do it now. */
if (lp->mc_reload_wait) if (lp->mc_reload_wait)
mc32_reset_multicast_list(dev); mc32_reset_multicast_list(dev);
else
wake_up(&lp->event);
} }
else complete(&lp->execution_cmd);
} }
if(status&2) if(status&2)
{ {
...@@ -1498,7 +1447,7 @@ static int mc32_close(struct net_device *dev) ...@@ -1498,7 +1447,7 @@ static int mc32_close(struct net_device *dev)
u8 regs; u8 regs;
u16 one=1; u16 one=1;
lp->desired_state = HALTED; lp->xceiver_desired_state = HALTED;
netif_stop_queue(dev); netif_stop_queue(dev);
/* /*
...@@ -1510,12 +1459,10 @@ static int mc32_close(struct net_device *dev) ...@@ -1510,12 +1459,10 @@ static int mc32_close(struct net_device *dev)
/* Shut down the transceiver */ /* Shut down the transceiver */
mc32_halt_transceiver(dev); mc32_halt_transceiver(dev);
/* Ensure we issue no more commands beyond this point */
down(&lp->cmd_mutex);
/* Catch any waiting commands */
while(lp->exec_pending==1)
sleep_on(&lp->event);
/* Ok the card is now stopping */ /* Ok the card is now stopping */
regs=inb(ioaddr+HOST_CTRL); regs=inb(ioaddr+HOST_CTRL);
...@@ -1542,12 +1489,10 @@ static int mc32_close(struct net_device *dev) ...@@ -1542,12 +1489,10 @@ static int mc32_close(struct net_device *dev)
static struct net_device_stats *mc32_get_stats(struct net_device *dev) static struct net_device_stats *mc32_get_stats(struct net_device *dev)
{ {
struct mc32_local *lp; struct mc32_local *lp = (struct mc32_local *)dev->priv;;
mc32_update_stats(dev); mc32_update_stats(dev);
lp = (struct mc32_local *)dev->priv;
return &lp->net_stats; return &lp->net_stats;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment