Commit b5b45b3c authored by Arik Nemtsov's avatar Arik Nemtsov Committed by Luciano Coelho

wlcore: refactor threaded IRQ routine

Separate the threaded IRQ handling routine into two functions.
The outer function takes the mutex and calls recovery on errors. It also
performs a Tx-path optimization to avoid redundant works.

The inner function is simplified - all calls to recovery are removed and
it assumes the lock is taken. The locked variant will be reused elsewhere.
Signed-off-by: default avatarArik Nemtsov <arik@wizery.com>
Signed-off-by: default avatarLuciano Coelho <coelho@ti.com>
parent c439a1ca
...@@ -494,20 +494,15 @@ static void wl1271_netstack_work(struct work_struct *work) ...@@ -494,20 +494,15 @@ static void wl1271_netstack_work(struct work_struct *work)
#define WL1271_IRQ_MAX_LOOPS 256 #define WL1271_IRQ_MAX_LOOPS 256
static irqreturn_t wl1271_irq(int irq, void *cookie) static int wlcore_irq_locked(struct wl1271 *wl)
{ {
int ret; int ret = 0;
u32 intr; u32 intr;
int loopcount = WL1271_IRQ_MAX_LOOPS; int loopcount = WL1271_IRQ_MAX_LOOPS;
struct wl1271 *wl = (struct wl1271 *)cookie;
bool done = false; bool done = false;
unsigned int defer_count; unsigned int defer_count;
unsigned long flags; unsigned long flags;
/* TX might be handled here, avoid redundant work */
set_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
cancel_work_sync(&wl->tx_work);
/* /*
* In case edge triggered interrupt must be used, we cannot iterate * In case edge triggered interrupt must be used, we cannot iterate
* more than once without introducing race conditions with the hardirq. * more than once without introducing race conditions with the hardirq.
...@@ -515,8 +510,6 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -515,8 +510,6 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ) if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ)
loopcount = 1; loopcount = 1;
mutex_lock(&wl->mutex);
wl1271_debug(DEBUG_IRQ, "IRQ work"); wl1271_debug(DEBUG_IRQ, "IRQ work");
if (unlikely(wl->state == WL1271_STATE_OFF)) if (unlikely(wl->state == WL1271_STATE_OFF))
...@@ -536,10 +529,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -536,10 +529,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
smp_mb__after_clear_bit(); smp_mb__after_clear_bit();
ret = wlcore_fw_status(wl, wl->fw_status_1, wl->fw_status_2); ret = wlcore_fw_status(wl, wl->fw_status_1, wl->fw_status_2);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
wlcore_hw_tx_immediate_compl(wl); wlcore_hw_tx_immediate_compl(wl);
...@@ -553,7 +544,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -553,7 +544,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
if (unlikely(intr & WL1271_ACX_INTR_WATCHDOG)) { if (unlikely(intr & WL1271_ACX_INTR_WATCHDOG)) {
wl1271_error("HW watchdog interrupt received! starting recovery."); wl1271_error("HW watchdog interrupt received! starting recovery.");
wl->watchdog_recovery = true; wl->watchdog_recovery = true;
wl12xx_queue_recovery_work(wl); ret = -EIO;
/* restarting the chip. ignore any other interrupt. */ /* restarting the chip. ignore any other interrupt. */
goto out; goto out;
...@@ -563,7 +554,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -563,7 +554,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
wl1271_error("SW watchdog interrupt received! " wl1271_error("SW watchdog interrupt received! "
"starting recovery."); "starting recovery.");
wl->watchdog_recovery = true; wl->watchdog_recovery = true;
wl12xx_queue_recovery_work(wl); ret = -EIO;
/* restarting the chip. ignore any other interrupt. */ /* restarting the chip. ignore any other interrupt. */
goto out; goto out;
...@@ -573,10 +564,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -573,10 +564,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_DATA"); wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_DATA");
ret = wlcore_rx(wl, wl->fw_status_1); ret = wlcore_rx(wl, wl->fw_status_1);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
/* Check if any tx blocks were freed */ /* Check if any tx blocks were freed */
spin_lock_irqsave(&wl->wl_lock, flags); spin_lock_irqsave(&wl->wl_lock, flags);
...@@ -588,20 +577,16 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -588,20 +577,16 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
* call the work function directly. * call the work function directly.
*/ */
ret = wlcore_tx_work_locked(wl); ret = wlcore_tx_work_locked(wl);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
} else { } else {
spin_unlock_irqrestore(&wl->wl_lock, flags); spin_unlock_irqrestore(&wl->wl_lock, flags);
} }
/* check for tx results */ /* check for tx results */
ret = wlcore_hw_tx_delayed_compl(wl); ret = wlcore_hw_tx_delayed_compl(wl);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
/* Make sure the deferred queues don't get too long */ /* Make sure the deferred queues don't get too long */
defer_count = skb_queue_len(&wl->deferred_tx_queue) + defer_count = skb_queue_len(&wl->deferred_tx_queue) +
...@@ -613,19 +598,15 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -613,19 +598,15 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
if (intr & WL1271_ACX_INTR_EVENT_A) { if (intr & WL1271_ACX_INTR_EVENT_A) {
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_A"); wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_A");
ret = wl1271_event_handle(wl, 0); ret = wl1271_event_handle(wl, 0);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
} }
if (intr & WL1271_ACX_INTR_EVENT_B) { if (intr & WL1271_ACX_INTR_EVENT_B) {
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_B"); wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_B");
ret = wl1271_event_handle(wl, 1); ret = wl1271_event_handle(wl, 1);
if (ret < 0) { if (ret < 0)
wl12xx_queue_recovery_work(wl);
goto out; goto out;
}
} }
if (intr & WL1271_ACX_INTR_INIT_COMPLETE) if (intr & WL1271_ACX_INTR_INIT_COMPLETE)
...@@ -639,6 +620,25 @@ static irqreturn_t wl1271_irq(int irq, void *cookie) ...@@ -639,6 +620,25 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
wl1271_ps_elp_sleep(wl); wl1271_ps_elp_sleep(wl);
out: out:
return ret;
}
static irqreturn_t wlcore_irq(int irq, void *cookie)
{
int ret;
unsigned long flags;
struct wl1271 *wl = cookie;
/* TX might be handled here, avoid redundant work */
set_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
cancel_work_sync(&wl->tx_work);
mutex_lock(&wl->mutex);
ret = wlcore_irq_locked(wl);
if (ret)
wl12xx_queue_recovery_work(wl);
spin_lock_irqsave(&wl->wl_lock, flags); spin_lock_irqsave(&wl->wl_lock, flags);
/* In case TX was not handled here, queue TX work */ /* In case TX was not handled here, queue TX work */
clear_bit(WL1271_FLAG_TX_PENDING, &wl->flags); clear_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
...@@ -1748,7 +1748,7 @@ static int wl1271_op_resume(struct ieee80211_hw *hw) ...@@ -1748,7 +1748,7 @@ static int wl1271_op_resume(struct ieee80211_hw *hw)
/* don't talk to the HW if recovery is pending */ /* don't talk to the HW if recovery is pending */
if (!pending_recovery) if (!pending_recovery)
wl1271_irq(0, wl); wlcore_irq(0, wl);
wlcore_enable_interrupts(wl); wlcore_enable_interrupts(wl);
} }
...@@ -5489,7 +5489,7 @@ int __devinit wlcore_probe(struct wl1271 *wl, struct platform_device *pdev) ...@@ -5489,7 +5489,7 @@ int __devinit wlcore_probe(struct wl1271 *wl, struct platform_device *pdev)
else else
irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT; irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT;
ret = request_threaded_irq(wl->irq, wl12xx_hardirq, wl1271_irq, ret = request_threaded_irq(wl->irq, wl12xx_hardirq, wlcore_irq,
irqflags, irqflags,
pdev->name, wl); pdev->name, wl);
if (ret < 0) { if (ret < 0) {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment