wlcore: refactor threaded IRQ routine
Separate the threaded IRQ handling routine into two functions. The outer function takes the mutex and calls recovery on errors. It also performs a Tx-path optimization to avoid redundant works. The inner function is simplified - all calls to recovery are removed and it assumes the lock is taken. The locked variant will be reused elsewhere. Signed-off-by: Arik Nemtsov <arik@wizery.com> Signed-off-by: Luciano Coelho <coelho@ti.com>
This commit is contained in:
committed by
Luciano Coelho
parent
c439a1ca3b
commit
b5b45b3cbd
@@ -494,20 +494,15 @@ static void wl1271_netstack_work(struct work_struct *work)
|
|||||||
|
|
||||||
#define WL1271_IRQ_MAX_LOOPS 256
|
#define WL1271_IRQ_MAX_LOOPS 256
|
||||||
|
|
||||||
static irqreturn_t wl1271_irq(int irq, void *cookie)
|
static int wlcore_irq_locked(struct wl1271 *wl)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret = 0;
|
||||||
u32 intr;
|
u32 intr;
|
||||||
int loopcount = WL1271_IRQ_MAX_LOOPS;
|
int loopcount = WL1271_IRQ_MAX_LOOPS;
|
||||||
struct wl1271 *wl = (struct wl1271 *)cookie;
|
|
||||||
bool done = false;
|
bool done = false;
|
||||||
unsigned int defer_count;
|
unsigned int defer_count;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
/* TX might be handled here, avoid redundant work */
|
|
||||||
set_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
|
|
||||||
cancel_work_sync(&wl->tx_work);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* In case edge triggered interrupt must be used, we cannot iterate
|
* In case edge triggered interrupt must be used, we cannot iterate
|
||||||
* more than once without introducing race conditions with the hardirq.
|
* more than once without introducing race conditions with the hardirq.
|
||||||
@@ -515,8 +510,6 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ)
|
if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ)
|
||||||
loopcount = 1;
|
loopcount = 1;
|
||||||
|
|
||||||
mutex_lock(&wl->mutex);
|
|
||||||
|
|
||||||
wl1271_debug(DEBUG_IRQ, "IRQ work");
|
wl1271_debug(DEBUG_IRQ, "IRQ work");
|
||||||
|
|
||||||
if (unlikely(wl->state == WL1271_STATE_OFF))
|
if (unlikely(wl->state == WL1271_STATE_OFF))
|
||||||
@@ -536,10 +529,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
smp_mb__after_clear_bit();
|
smp_mb__after_clear_bit();
|
||||||
|
|
||||||
ret = wlcore_fw_status(wl, wl->fw_status_1, wl->fw_status_2);
|
ret = wlcore_fw_status(wl, wl->fw_status_1, wl->fw_status_2);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
|
|
||||||
wlcore_hw_tx_immediate_compl(wl);
|
wlcore_hw_tx_immediate_compl(wl);
|
||||||
|
|
||||||
@@ -553,7 +544,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
if (unlikely(intr & WL1271_ACX_INTR_WATCHDOG)) {
|
if (unlikely(intr & WL1271_ACX_INTR_WATCHDOG)) {
|
||||||
wl1271_error("HW watchdog interrupt received! starting recovery.");
|
wl1271_error("HW watchdog interrupt received! starting recovery.");
|
||||||
wl->watchdog_recovery = true;
|
wl->watchdog_recovery = true;
|
||||||
wl12xx_queue_recovery_work(wl);
|
ret = -EIO;
|
||||||
|
|
||||||
/* restarting the chip. ignore any other interrupt. */
|
/* restarting the chip. ignore any other interrupt. */
|
||||||
goto out;
|
goto out;
|
||||||
@@ -563,7 +554,7 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
wl1271_error("SW watchdog interrupt received! "
|
wl1271_error("SW watchdog interrupt received! "
|
||||||
"starting recovery.");
|
"starting recovery.");
|
||||||
wl->watchdog_recovery = true;
|
wl->watchdog_recovery = true;
|
||||||
wl12xx_queue_recovery_work(wl);
|
ret = -EIO;
|
||||||
|
|
||||||
/* restarting the chip. ignore any other interrupt. */
|
/* restarting the chip. ignore any other interrupt. */
|
||||||
goto out;
|
goto out;
|
||||||
@@ -573,10 +564,8 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_DATA");
|
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_DATA");
|
||||||
|
|
||||||
ret = wlcore_rx(wl, wl->fw_status_1);
|
ret = wlcore_rx(wl, wl->fw_status_1);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
|
|
||||||
/* Check if any tx blocks were freed */
|
/* Check if any tx blocks were freed */
|
||||||
spin_lock_irqsave(&wl->wl_lock, flags);
|
spin_lock_irqsave(&wl->wl_lock, flags);
|
||||||
@@ -588,20 +577,16 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
* call the work function directly.
|
* call the work function directly.
|
||||||
*/
|
*/
|
||||||
ret = wlcore_tx_work_locked(wl);
|
ret = wlcore_tx_work_locked(wl);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
spin_unlock_irqrestore(&wl->wl_lock, flags);
|
spin_unlock_irqrestore(&wl->wl_lock, flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* check for tx results */
|
/* check for tx results */
|
||||||
ret = wlcore_hw_tx_delayed_compl(wl);
|
ret = wlcore_hw_tx_delayed_compl(wl);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
|
|
||||||
/* Make sure the deferred queues don't get too long */
|
/* Make sure the deferred queues don't get too long */
|
||||||
defer_count = skb_queue_len(&wl->deferred_tx_queue) +
|
defer_count = skb_queue_len(&wl->deferred_tx_queue) +
|
||||||
@@ -613,19 +598,15 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
if (intr & WL1271_ACX_INTR_EVENT_A) {
|
if (intr & WL1271_ACX_INTR_EVENT_A) {
|
||||||
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_A");
|
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_A");
|
||||||
ret = wl1271_event_handle(wl, 0);
|
ret = wl1271_event_handle(wl, 0);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (intr & WL1271_ACX_INTR_EVENT_B) {
|
if (intr & WL1271_ACX_INTR_EVENT_B) {
|
||||||
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_B");
|
wl1271_debug(DEBUG_IRQ, "WL1271_ACX_INTR_EVENT_B");
|
||||||
ret = wl1271_event_handle(wl, 1);
|
ret = wl1271_event_handle(wl, 1);
|
||||||
if (ret < 0) {
|
if (ret < 0)
|
||||||
wl12xx_queue_recovery_work(wl);
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (intr & WL1271_ACX_INTR_INIT_COMPLETE)
|
if (intr & WL1271_ACX_INTR_INIT_COMPLETE)
|
||||||
@@ -639,6 +620,25 @@ static irqreturn_t wl1271_irq(int irq, void *cookie)
|
|||||||
wl1271_ps_elp_sleep(wl);
|
wl1271_ps_elp_sleep(wl);
|
||||||
|
|
||||||
out:
|
out:
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
static irqreturn_t wlcore_irq(int irq, void *cookie)
|
||||||
|
{
|
||||||
|
int ret;
|
||||||
|
unsigned long flags;
|
||||||
|
struct wl1271 *wl = cookie;
|
||||||
|
|
||||||
|
/* TX might be handled here, avoid redundant work */
|
||||||
|
set_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
|
||||||
|
cancel_work_sync(&wl->tx_work);
|
||||||
|
|
||||||
|
mutex_lock(&wl->mutex);
|
||||||
|
|
||||||
|
ret = wlcore_irq_locked(wl);
|
||||||
|
if (ret)
|
||||||
|
wl12xx_queue_recovery_work(wl);
|
||||||
|
|
||||||
spin_lock_irqsave(&wl->wl_lock, flags);
|
spin_lock_irqsave(&wl->wl_lock, flags);
|
||||||
/* In case TX was not handled here, queue TX work */
|
/* In case TX was not handled here, queue TX work */
|
||||||
clear_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
|
clear_bit(WL1271_FLAG_TX_PENDING, &wl->flags);
|
||||||
@@ -1748,7 +1748,7 @@ static int wl1271_op_resume(struct ieee80211_hw *hw)
|
|||||||
|
|
||||||
/* don't talk to the HW if recovery is pending */
|
/* don't talk to the HW if recovery is pending */
|
||||||
if (!pending_recovery)
|
if (!pending_recovery)
|
||||||
wl1271_irq(0, wl);
|
wlcore_irq(0, wl);
|
||||||
|
|
||||||
wlcore_enable_interrupts(wl);
|
wlcore_enable_interrupts(wl);
|
||||||
}
|
}
|
||||||
@@ -5489,7 +5489,7 @@ int __devinit wlcore_probe(struct wl1271 *wl, struct platform_device *pdev)
|
|||||||
else
|
else
|
||||||
irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT;
|
irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT;
|
||||||
|
|
||||||
ret = request_threaded_irq(wl->irq, wl12xx_hardirq, wl1271_irq,
|
ret = request_threaded_irq(wl->irq, wl12xx_hardirq, wlcore_irq,
|
||||||
irqflags,
|
irqflags,
|
||||||
pdev->name, wl);
|
pdev->name, wl);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
|
Reference in New Issue
Block a user