Subject: [RFC] ath9k: Implement op_flush()

When op_flush() is called with no drop (drop=false), the driver
tries to tx as many frames as possible in 100ms on every hw queue.
During this time period frames from sw queue are also scheduled on
to respective hw queue.

Signed-off-by: Vasanthakumar Thiagarajan <[email protected]>
---
drivers/net/wireless/ath/ath9k/ath9k.h | 1 +
drivers/net/wireless/ath/ath9k/main.c | 72 ++++++++++++++++++++++++++++++++
drivers/net/wireless/ath/ath9k/xmit.c | 27 +++++++-----
3 files changed, 89 insertions(+), 11 deletions(-)

diff --git a/drivers/net/wireless/ath/ath9k/ath9k.h b/drivers/net/wireless/ath/ath9k/ath9k.h
index 9272278..704521f 100644
--- a/drivers/net/wireless/ath/ath9k/ath9k.h
+++ b/drivers/net/wireless/ath/ath9k/ath9k.h
@@ -192,6 +192,7 @@ struct ath_txq {
u32 axq_ampdu_depth;
bool stopped;
bool axq_tx_inprogress;
+ bool txq_flush_inprogress;
struct list_head axq_acq;
struct list_head txq_fifo[ATH_TXFIFO_DEPTH];
struct list_head txq_fifo_pending;
diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
index 4ed43b2..cd07779 100644
--- a/drivers/net/wireless/ath/ath9k/main.c
+++ b/drivers/net/wireless/ath/ath9k/main.c
@@ -53,6 +53,21 @@ static u8 parse_mpdudensity(u8 mpdudensity)
}
}

+static bool ath9k_is_pending_frames(struct ath_softc *sc, struct ath_txq *txq)
+{
+ bool pending = false;
+
+ spin_lock_bh(&txq->axq_lock);
+
+ if (txq->axq_depth || !list_empty(&txq->axq_acq))
+ pending = true;
+ else if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
+ pending = !list_empty(&txq->txq_fifo_pending);
+
+ spin_unlock_bh(&txq->axq_lock);
+ return pending;
+}
+
bool ath9k_setpower(struct ath_softc *sc, enum ath9k_power_mode mode)
{
unsigned long flags;
@@ -2122,6 +2137,62 @@ static void ath9k_set_coverage_class(struct ieee80211_hw *hw, u8 coverage_class)
mutex_unlock(&sc->mutex);
}

+static void ath9k_flush(struct ieee80211_hw *hw, bool drop)
+{
+#define ATH_FLUSH_TIMEOUT 100 /* ms */
+ struct ath_softc *sc = hw->priv;
+ struct ath_txq *txq;
+ struct ath_hw *ah = sc->sc_ah;
+ struct ath_common *common = ath9k_hw_common(ah);
+ int i, j, npend = 0;
+
+ mutex_lock(&sc->mutex);
+
+ cancel_delayed_work_sync(&sc->tx_complete_work);
+
+ for (i = 0; i < ATH9K_NUM_TX_QUEUES; i++) {
+ if (!ATH_TXQ_SETUP(sc, i))
+ continue;
+ txq = &sc->tx.txq[i];
+
+ if (!drop) {
+ for (j = 0; j < ATH_FLUSH_TIMEOUT; i++) {
+ if (!ath9k_is_pending_frames(sc, txq))
+ break;
+ mdelay(1);
+ }
+ }
+
+ if (drop || ath9k_is_pending_frames(sc, txq)) {
+ ath_dbg(common, ATH_DBG_QUEUE, "Drop frames from hw queue:%d\n",
+ txq->axq_qnum);
+ spin_lock_bh(&txq->axq_lock);
+ txq->txq_flush_inprogress = true;
+ spin_unlock_bh(&txq->axq_lock);
+
+ ath9k_ps_wakeup(sc);
+ ath9k_hw_stoptxdma(ah, txq->axq_qnum);
+ npend = ath9k_hw_numtxpending(ah, txq->axq_qnum);
+ ath9k_ps_restore(sc);
+ if (npend) {
+ ath_reset(sc, false);
+ spin_lock_bh(&txq->axq_lock);
+ txq->txq_flush_inprogress = false;
+ spin_unlock_bh(&txq->axq_lock);
+ break;
+ }
+
+ ath_draintxq(sc, txq, false);
+ spin_lock_bh(&txq->axq_lock);
+ txq->txq_flush_inprogress = false;
+ spin_unlock_bh(&txq->axq_lock);
+ }
+ }
+
+ ieee80211_queue_delayed_work(hw, &sc->tx_complete_work, 0);
+ mutex_unlock(&sc->mutex);
+}
+
struct ieee80211_ops ath9k_ops = {
.tx = ath9k_tx,
.start = ath9k_start,
@@ -2143,4 +2214,5 @@ struct ieee80211_ops ath9k_ops = {
.get_survey = ath9k_get_survey,
.rfkill_poll = ath9k_rfkill_poll_state,
.set_coverage_class = ath9k_set_coverage_class,
+ .flush = ath9k_flush,
};
diff --git a/drivers/net/wireless/ath/ath9k/xmit.c b/drivers/net/wireless/ath/ath9k/xmit.c
index 9f4e755..c6680ed 100644
--- a/drivers/net/wireless/ath/ath9k/xmit.c
+++ b/drivers/net/wireless/ath/ath9k/xmit.c
@@ -2013,7 +2013,8 @@ static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq)
spin_lock_bh(&txq->axq_lock);
if (list_empty(&txq->axq_q)) {
txq->axq_link = NULL;
- if (sc->sc_flags & SC_OP_TXAGGR)
+ if (sc->sc_flags & SC_OP_TXAGGR &&
+ !txq->txq_flush_inprogress)
ath_txq_schedule(sc, txq);
spin_unlock_bh(&txq->axq_lock);
break;
@@ -2070,6 +2071,7 @@ static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq)

if (bf_is_ampdu_not_probing(bf))
txq->axq_ampdu_depth--;
+
spin_unlock_bh(&txq->axq_lock);

if (bf_held)
@@ -2093,7 +2095,7 @@ static void ath_tx_processq(struct ath_softc *sc, struct ath_txq *txq)

spin_lock_bh(&txq->axq_lock);

- if (sc->sc_flags & SC_OP_TXAGGR)
+ if (sc->sc_flags & SC_OP_TXAGGR && !txq->txq_flush_inprogress)
ath_txq_schedule(sc, txq);
spin_unlock_bh(&txq->axq_lock);
}
@@ -2264,15 +2266,18 @@ void ath_tx_edma_tasklet(struct ath_softc *sc)

spin_lock_bh(&txq->axq_lock);

- if (!list_empty(&txq->txq_fifo_pending)) {
- INIT_LIST_HEAD(&bf_head);
- bf = list_first_entry(&txq->txq_fifo_pending,
- struct ath_buf, list);
- list_cut_position(&bf_head, &txq->txq_fifo_pending,
- &bf->bf_lastbf->list);
- ath_tx_txqaddbuf(sc, txq, &bf_head);
- } else if (sc->sc_flags & SC_OP_TXAGGR)
- ath_txq_schedule(sc, txq);
+ if (!txq->txq_flush_inprogress) {
+ if (!list_empty(&txq->txq_fifo_pending)) {
+ INIT_LIST_HEAD(&bf_head);
+ bf = list_first_entry(&txq->txq_fifo_pending,
+ struct ath_buf, list);
+ list_cut_position(&bf_head,
+ &txq->txq_fifo_pending,
+ &bf->bf_lastbf->list);
+ ath_tx_txqaddbuf(sc, txq, &bf_head);
+ } else if (sc->sc_flags & SC_OP_TXAGGR)
+ ath_txq_schedule(sc, txq);
+ }
spin_unlock_bh(&txq->axq_lock);
}
}
--
1.7.0.4



2011-02-09 14:56:24

by Vasanth Thiagarajan

[permalink] [raw]
Subject: RE: [RFC] ath9k: Implement op_flush()


________________________________________
From: [email protected] [[email protected]] On Behalf Of Vasanthakumar Thiagarajan [[email protected]]
Sent: Wednesday, February 09, 2011 5:49 AM
To: [email protected]
Cc: [email protected]
Subject: [RFC] ath9k: Implement op_flush()

When op_flush() is called with no drop (drop=false), the driver
tries to tx as many frames as possible in 100ms on every hw queue.
During this time period frames from sw queue are also scheduled on
to respective hw queue.

Signed-off-by: Vasanthakumar Thiagarajan <[email protected]>
---
drivers/net/wireless/ath/ath9k/ath9k.h | 1 +
drivers/net/wireless/ath/ath9k/main.c | 72 ++++++++++++++++++++++++++++++++
drivers/net/wireless/ath/ath9k/xmit.c | 27 +++++++-----
3 files changed, 89 insertions(+), 11 deletions(-)

diff --git a/drivers/net/wireless/ath/ath9k/ath9k.h b/drivers/net/wireless/ath/ath9k/ath9k.h
index 9272278..704521f 100644
--- a/drivers/net/wireless/ath/ath9k/ath9k.h
+++ b/drivers/net/wireless/ath/ath9k/ath9k.h
@@ -192,6 +192,7 @@ struct ath_txq {
u32 axq_ampdu_depth;
bool stopped;
bool axq_tx_inprogress;
+ bool txq_flush_inprogress;
struct list_head axq_acq;
struct list_head txq_fifo[ATH_TXFIFO_DEPTH];
struct list_head txq_fifo_pending;
diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
index 4ed43b2..cd07779 100644
--- a/drivers/net/wireless/ath/ath9k/main.c
+++ b/drivers/net/wireless/ath/ath9k/main.c
@@ -53,6 +53,21 @@ static u8 parse_mpdudensity(u8 mpdudensity)
}
}

+static bool ath9k_is_pending_frames(struct ath_softc *sc, struct ath_txq *txq)
+{
+ bool pending = false;
+
+ spin_lock_bh(&txq->axq_lock);
+
+ if (txq->axq_depth || !list_empty(&txq->axq_acq))
+ pending = true;
+ else if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
+ pending = !list_empty(&txq->txq_fifo_pending);
+
+ spin_unlock_bh(&txq->axq_lock);
+ return pending;
+}
+
bool ath9k_setpower(struct ath_softc *sc, enum ath9k_power_mode mode)
{
unsigned long flags;
@@ -2122,6 +2137,62 @@ static void ath9k_set_coverage_class(struct ieee80211_hw *hw, u8 coverage_class)
mutex_unlock(&sc->mutex);
}

+static void ath9k_flush(struct ieee80211_hw *hw, bool drop)
+{
+#define ATH_FLUSH_TIMEOUT 100 /* ms */
+ struct ath_softc *sc = hw->priv;
+ struct ath_txq *txq;
+ struct ath_hw *ah = sc->sc_ah;
+ struct ath_common *common = ath9k_hw_common(ah);
+ int i, j, npend = 0;
+
+ mutex_lock(&sc->mutex);
+
+ cancel_delayed_work_sync(&sc->tx_complete_work);
+
+ for (i = 0; i < ATH9K_NUM_TX_QUEUES; i++) {
+ if (!ATH_TXQ_SETUP(sc, i))
+ continue;
+ txq = &sc->tx.txq[i];
+
+ if (!drop) {
+ for (j = 0; j < ATH_FLUSH_TIMEOUT; i++) {

of course it is j++, fixed it locally but forgot to integrate.






2011-02-09 15:04:49

by Vasanth Thiagarajan

[permalink] [raw]
Subject: RE: [RFC] ath9k: Implement op_flush()


________________________________________
From: Felix Fietkau [[email protected]]
Sent: Wednesday, February 09, 2011 6:15 AM
To: Vasanth Thiagarajan
Cc: [email protected]; [email protected]
Subject: Re: [RFC] ath9k: Implement op_flush()

On 2011-02-09 2:49 PM, Vasanthakumar Thiagarajan wrote:
> When op_flush() is called with no drop (drop=false), the driver
> tries to tx as many frames as possible in 100ms on every hw queue.
> During this time period frames from sw queue are also scheduled on
> to respective hw queue.
>
> Signed-off-by: Vasanthakumar Thiagarajan <[email protected]>
> ---
> drivers/net/wireless/ath/ath9k/ath9k.h | 1 +
> drivers/net/wireless/ath/ath9k/main.c | 72 ++++++++++++++++++++++++++++++++
> drivers/net/wireless/ath/ath9k/xmit.c | 27 +++++++-----
> 3 files changed, 89 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/net/wireless/ath/ath9k/ath9k.h b/drivers/net/wireless/ath/ath9k/ath9k.h
> index 9272278..704521f 100644
> --- a/drivers/net/wireless/ath/ath9k/ath9k.h
> +++ b/drivers/net/wireless/ath/ath9k/ath9k.h
> @@ -192,6 +192,7 @@ struct ath_txq {
> u32 axq_ampdu_depth;
> bool stopped;
> bool axq_tx_inprogress;
> + bool txq_flush_inprogress;
> struct list_head axq_acq;
> struct list_head txq_fifo[ATH_TXFIFO_DEPTH];
> struct list_head txq_fifo_pending;
> diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
> index 4ed43b2..cd07779 100644
> --- a/drivers/net/wireless/ath/ath9k/main.c
> +++ b/drivers/net/wireless/ath/ath9k/main.c
> @@ -53,6 +53,21 @@ static u8 parse_mpdudensity(u8 mpdudensity)
> }
> }
>
ath9k_has_pending_frames would be a better name for this

yeah, thanks.

> +static bool ath9k_is_pending_frames(struct ath_softc *sc, struct ath_txq *txq)
> +{
> + bool pending = false;
> +
> + spin_lock_bh(&txq->axq_lock);
> +
> + if (txq->axq_depth || !list_empty(&txq->axq_acq))
> + pending = true;
> + else if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
> + pending = !list_empty(&txq->txq_fifo_pending);
> +
> + spin_unlock_bh(&txq->axq_lock);
> + return pending;
> +}
> +
> bool ath9k_setpower(struct ath_softc *sc, enum ath9k_power_mode mode)
> {
> unsigned long flags;
> @@ -2122,6 +2137,62 @@ static void ath9k_set_coverage_class(struct ieee80211_hw *hw, u8 coverage_class)
> mutex_unlock(&sc->mutex);
> }
>
> +static void ath9k_flush(struct ieee80211_hw *hw, bool drop)
> +{
> +#define ATH_FLUSH_TIMEOUT 100 /* ms */
> + struct ath_softc *sc = hw->priv;
> + struct ath_txq *txq;
> + struct ath_hw *ah = sc->sc_ah;
> + struct ath_common *common = ath9k_hw_common(ah);
> + int i, j, npend = 0;
> +
> + mutex_lock(&sc->mutex);
> +
> + cancel_delayed_work_sync(&sc->tx_complete_work);
> +
> + for (i = 0; i < ATH9K_NUM_TX_QUEUES; i++) {
> + if (!ATH_TXQ_SETUP(sc, i))
> + continue;
> + txq = &sc->tx.txq[i];
> +
> + if (!drop) {
> + for (j = 0; j < ATH_FLUSH_TIMEOUT; i++) {
> + if (!ath9k_is_pending_frames(sc, txq))
> + break;

Do not use mdelay, it blocks the CPU. msleep is better.

sure.
> + mdelay(1);
> + }
> + }
> +
> + if (drop || ath9k_is_pending_frames(sc, txq)) {
> + ath_dbg(common, ATH_DBG_QUEUE, "Drop frames from hw queue:%d\n",
> + txq->axq_qnum);
> + spin_lock_bh(&txq->axq_lock);
> + txq->txq_flush_inprogress = true;
> + spin_unlock_bh(&txq->axq_lock);
> +
> + ath9k_ps_wakeup(sc);
> + ath9k_hw_stoptxdma(ah, txq->axq_qnum);
> + npend = ath9k_hw_numtxpending(ah, txq->axq_qnum);
> + ath9k_ps_restore(sc);

Please move the reset outside of the for loop. Also, I think you can
leave out the spinlocks for the txq_flush_inprogress assignment below.

You mean the following code?, if we decided to do a reset, then we need not proceed
with the next queue. yeah, it is unnecessary to lock to access txq->txq_flush_inprogres
after a reset.
> + if (npend) {
> + ath_reset(sc, false);
> + spin_lock_bh(&txq->axq_lock);
> + txq->txq_flush_inprogress = false;
> + spin_unlock_bh(&txq->axq_lock);
> + break;
> + }

2011-02-09 14:15:27

by Felix Fietkau

[permalink] [raw]
Subject: Re: [RFC] ath9k: Implement op_flush()

On 2011-02-09 2:49 PM, Vasanthakumar Thiagarajan wrote:
> When op_flush() is called with no drop (drop=false), the driver
> tries to tx as many frames as possible in 100ms on every hw queue.
> During this time period frames from sw queue are also scheduled on
> to respective hw queue.
>
> Signed-off-by: Vasanthakumar Thiagarajan <[email protected]>
> ---
> drivers/net/wireless/ath/ath9k/ath9k.h | 1 +
> drivers/net/wireless/ath/ath9k/main.c | 72 ++++++++++++++++++++++++++++++++
> drivers/net/wireless/ath/ath9k/xmit.c | 27 +++++++-----
> 3 files changed, 89 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/net/wireless/ath/ath9k/ath9k.h b/drivers/net/wireless/ath/ath9k/ath9k.h
> index 9272278..704521f 100644
> --- a/drivers/net/wireless/ath/ath9k/ath9k.h
> +++ b/drivers/net/wireless/ath/ath9k/ath9k.h
> @@ -192,6 +192,7 @@ struct ath_txq {
> u32 axq_ampdu_depth;
> bool stopped;
> bool axq_tx_inprogress;
> + bool txq_flush_inprogress;
> struct list_head axq_acq;
> struct list_head txq_fifo[ATH_TXFIFO_DEPTH];
> struct list_head txq_fifo_pending;
> diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
> index 4ed43b2..cd07779 100644
> --- a/drivers/net/wireless/ath/ath9k/main.c
> +++ b/drivers/net/wireless/ath/ath9k/main.c
> @@ -53,6 +53,21 @@ static u8 parse_mpdudensity(u8 mpdudensity)
> }
> }
>
ath9k_has_pending_frames would be a better name for this
> +static bool ath9k_is_pending_frames(struct ath_softc *sc, struct ath_txq *txq)
> +{
> + bool pending = false;
> +
> + spin_lock_bh(&txq->axq_lock);
> +
> + if (txq->axq_depth || !list_empty(&txq->axq_acq))
> + pending = true;
> + else if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
> + pending = !list_empty(&txq->txq_fifo_pending);
> +
> + spin_unlock_bh(&txq->axq_lock);
> + return pending;
> +}
> +
> bool ath9k_setpower(struct ath_softc *sc, enum ath9k_power_mode mode)
> {
> unsigned long flags;
> @@ -2122,6 +2137,62 @@ static void ath9k_set_coverage_class(struct ieee80211_hw *hw, u8 coverage_class)
> mutex_unlock(&sc->mutex);
> }
>
> +static void ath9k_flush(struct ieee80211_hw *hw, bool drop)
> +{
> +#define ATH_FLUSH_TIMEOUT 100 /* ms */
> + struct ath_softc *sc = hw->priv;
> + struct ath_txq *txq;
> + struct ath_hw *ah = sc->sc_ah;
> + struct ath_common *common = ath9k_hw_common(ah);
> + int i, j, npend = 0;
> +
> + mutex_lock(&sc->mutex);
> +
> + cancel_delayed_work_sync(&sc->tx_complete_work);
> +
> + for (i = 0; i < ATH9K_NUM_TX_QUEUES; i++) {
> + if (!ATH_TXQ_SETUP(sc, i))
> + continue;
> + txq = &sc->tx.txq[i];
> +
> + if (!drop) {
> + for (j = 0; j < ATH_FLUSH_TIMEOUT; i++) {
> + if (!ath9k_is_pending_frames(sc, txq))
> + break;

Do not use mdelay, it blocks the CPU. msleep is better.
> + mdelay(1);
> + }
> + }
> +
> + if (drop || ath9k_is_pending_frames(sc, txq)) {
> + ath_dbg(common, ATH_DBG_QUEUE, "Drop frames from hw queue:%d\n",
> + txq->axq_qnum);
> + spin_lock_bh(&txq->axq_lock);
> + txq->txq_flush_inprogress = true;
> + spin_unlock_bh(&txq->axq_lock);
> +
> + ath9k_ps_wakeup(sc);
> + ath9k_hw_stoptxdma(ah, txq->axq_qnum);
> + npend = ath9k_hw_numtxpending(ah, txq->axq_qnum);
> + ath9k_ps_restore(sc);

Please move the reset outside of the for loop. Also, I think you can
leave out the spinlocks for the txq_flush_inprogress assignment below.
> + if (npend) {
> + ath_reset(sc, false);
> + spin_lock_bh(&txq->axq_lock);
> + txq->txq_flush_inprogress = false;
> + spin_unlock_bh(&txq->axq_lock);
> + break;
> + }
> +
> + ath_draintxq(sc, txq, false);
> + spin_lock_bh(&txq->axq_lock);
> + txq->txq_flush_inprogress = false;
> + spin_unlock_bh(&txq->axq_lock);
> + }
> + }
> +
> + ieee80211_queue_delayed_work(hw, &sc->tx_complete_work, 0);
> + mutex_unlock(&sc->mutex);
> +}
> +
> struct ieee80211_ops ath9k_ops = {
> .tx = ath9k_tx,
> .start = ath9k_start,
> @@ -2143,4 +2214,5 @@ struct ieee80211_ops ath9k_ops = {
> .get_survey = ath9k_get_survey,
> .rfkill_poll = ath9k_rfkill_poll_state,
> .set_coverage_class = ath9k_set_coverage_class,
> + .flush = ath9k_flush,
> };