From 36e6d4ca53495970956c697c8549d16ceb998242 Mon Sep 17 00:00:00 2001 From: Felix Fietkau Date: Mon, 29 Jul 2013 19:44:05 +0000 Subject: ath9k: fix several issues in the tx queueing rework Signed-off-by: Felix Fietkau SVN-Revision: 37616 --- .../patches/570-ath9k_tid_queue_non_aggr.patch | 416 +++++++++++++++++++++ 1 file changed, 416 insertions(+) create mode 100644 package/kernel/mac80211/patches/570-ath9k_tid_queue_non_aggr.patch (limited to 'package/kernel/mac80211/patches/570-ath9k_tid_queue_non_aggr.patch') diff --git a/package/kernel/mac80211/patches/570-ath9k_tid_queue_non_aggr.patch b/package/kernel/mac80211/patches/570-ath9k_tid_queue_non_aggr.patch new file mode 100644 index 0000000000..690c90972c --- /dev/null +++ b/package/kernel/mac80211/patches/570-ath9k_tid_queue_non_aggr.patch @@ -0,0 +1,416 @@ +--- a/drivers/net/wireless/ath/ath9k/xmit.c ++++ b/drivers/net/wireless/ath/ath9k/xmit.c +@@ -198,6 +198,41 @@ static struct sk_buff *ath_tid_dequeue(s + return skb; + } + ++/* ++ * ath_tx_tid_change_state: ++ * - clears a-mpdu flag of previous session ++ * - force sequence number allocation to fix next BlockAck Window ++ */ ++static void ++ath_tx_tid_change_state(struct ath_softc *sc, struct ath_atx_tid *tid) ++{ ++ struct ath_txq *txq = tid->ac->txq; ++ struct ieee80211_tx_info *tx_info; ++ struct sk_buff *skb, *tskb; ++ struct ath_buf *bf; ++ struct ath_frame_info *fi; ++ ++ skb_queue_walk_safe(&tid->buf_q, skb, tskb) { ++ fi = get_frame_info(skb); ++ bf = fi->bf; ++ ++ tx_info = IEEE80211_SKB_CB(skb); ++ tx_info->flags &= ~IEEE80211_TX_CTL_AMPDU; ++ ++ if (bf) ++ continue; ++ ++ bf = ath_tx_setup_buffer(sc, txq, tid, skb); ++ if (!bf) { ++ __skb_unlink(skb, &tid->buf_q); ++ ath_txq_skb_done(sc, txq, skb); ++ ieee80211_free_txskb(sc->hw, skb); ++ continue; ++ } ++ } ++ ++} ++ + static void ath_tx_flush_tid(struct ath_softc *sc, struct ath_atx_tid *tid) + { + struct ath_txq *txq = tid->ac->txq; +@@ -212,28 +247,22 @@ static void ath_tx_flush_tid(struct ath_ + + memset(&ts, 0, sizeof(ts)); + +- while ((skb = ath_tid_dequeue(tid))) { ++ while ((skb = __skb_dequeue(&tid->retry_q))) { + fi = get_frame_info(skb); + bf = fi->bf; +- + if (!bf) { +- bf = ath_tx_setup_buffer(sc, txq, tid, skb); +- if (!bf) { +- ath_txq_skb_done(sc, txq, skb); +- ieee80211_free_txskb(sc->hw, skb); +- continue; +- } ++ ath_txq_skb_done(sc, txq, skb); ++ ieee80211_free_txskb(sc->hw, skb); ++ continue; + } + + if (fi->baw_tracked) { +- list_add_tail(&bf->list, &bf_head); + ath_tx_update_baw(sc, tid, bf->bf_state.seqno); +- ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0); + sendbar = true; +- } else { +- ath_set_rates(tid->an->vif, tid->an->sta, bf); +- ath_tx_send_normal(sc, txq, NULL, skb); + } ++ ++ list_add_tail(&bf->list, &bf_head); ++ ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0); + } + + if (sendbar) { +@@ -911,50 +940,39 @@ ath_tx_get_tid_subframe(struct ath_softc + return NULL; + } + +-static enum ATH_AGGR_STATUS ath_tx_form_aggr(struct ath_softc *sc, +- struct ath_txq *txq, +- struct ath_atx_tid *tid, +- struct list_head *bf_q, +- int *aggr_len) ++static bool ++ath_tx_form_aggr(struct ath_softc *sc, struct ath_txq *txq, ++ struct ath_atx_tid *tid, struct list_head *bf_q, ++ struct ath_buf *bf_first, struct sk_buff_head *tid_q, ++ int *aggr_len) + { + #define PADBYTES(_len) ((4 - ((_len) % 4)) % 4) +- struct ath_buf *bf, *bf_first = NULL, *bf_prev = NULL; ++ struct ath_buf *bf = bf_first, *bf_prev = NULL; + int nframes = 0, ndelim; + u16 aggr_limit = 0, al = 0, bpad = 0, + al_delta, h_baw = tid->baw_size / 2; +- enum ATH_AGGR_STATUS status = ATH_AGGR_DONE; + struct ieee80211_tx_info *tx_info; + struct ath_frame_info *fi; + struct sk_buff *skb; +- struct sk_buff_head *tid_q; ++ bool closed = false; + +- do { +- bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q); +- if (!bf) { +- status = ATH_AGGR_BAW_CLOSED; +- break; +- } ++ bf = bf_first; ++ aggr_limit = ath_lookup_rate(sc, bf, tid); + ++ do { + skb = bf->bf_mpdu; + fi = get_frame_info(skb); + +- if (!bf_first) { +- bf_first = bf; +- ath_set_rates(tid->an->vif, tid->an->sta, bf); +- aggr_limit = ath_lookup_rate(sc, bf, tid); +- } +- + /* do not exceed aggregation limit */ + al_delta = ATH_AGGR_DELIM_SZ + fi->framelen; + if (nframes) { + if (aggr_limit < al + bpad + al_delta || +- ath_lookup_legacy(bf) || nframes >= h_baw) { +- status = ATH_AGGR_LIMITED; ++ ath_lookup_legacy(bf) || nframes >= h_baw) + break; +- } + + tx_info = IEEE80211_SKB_CB(bf->bf_mpdu); +- if (tx_info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE) ++ if ((tx_info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE) || ++ !(tx_info->flags & IEEE80211_TX_CTL_AMPDU)) + break; + } + +@@ -984,11 +1002,26 @@ static enum ATH_AGGR_STATUS ath_tx_form_ + + bf_prev = bf; + ++ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q); ++ if (!bf) { ++ closed = true; ++ break; ++ } + } while (ath_tid_has_buffered(tid)); + ++ bf = bf_first; ++ bf->bf_lastbf = bf_prev; ++ ++ if (bf == bf_prev) { ++ al = get_frame_info(bf->bf_mpdu)->framelen; ++ bf->bf_state.bf_type = BUF_AMPDU; ++ } else { ++ TX_STAT_INC(txq->axq_qnum, a_aggr); ++ } ++ + *aggr_len = al; + +- return status; ++ return closed; + #undef PADBYTES + } + +@@ -1249,14 +1282,50 @@ static void ath_tx_fill_desc(struct ath_ + } + } + ++static void ++ath_tx_form_burst(struct ath_softc *sc, struct ath_txq *txq, ++ struct ath_atx_tid *tid, struct list_head *bf_q, ++ struct ath_buf *bf_first, struct sk_buff_head *tid_q) ++{ ++ struct ath_buf *bf = bf_first, *bf_prev = NULL; ++ struct sk_buff *skb; ++ int nframes = 0; ++ ++ do { ++ struct ieee80211_tx_info *tx_info; ++ skb = bf->bf_mpdu; ++ ++ nframes++; ++ __skb_unlink(skb, tid_q); ++ list_add_tail(&bf->list, bf_q); ++ if (bf_prev) ++ bf_prev->bf_next = bf; ++ bf_prev = bf; ++ ++ if (nframes >= 2) ++ break; ++ ++ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q); ++ if (!bf) ++ break; ++ ++ tx_info = IEEE80211_SKB_CB(bf->bf_mpdu); ++ if (tx_info->flags & IEEE80211_TX_CTL_AMPDU) ++ break; ++ ++ ath_set_rates(tid->an->vif, tid->an->sta, bf); ++ } while (1); ++} ++ + static void ath_tx_sched_aggr(struct ath_softc *sc, struct ath_txq *txq, + struct ath_atx_tid *tid) + { + struct ath_buf *bf; +- enum ATH_AGGR_STATUS status; + struct ieee80211_tx_info *tx_info; ++ struct sk_buff_head *tid_q; + struct list_head bf_q; +- int aggr_len; ++ int aggr_len = 0; ++ bool aggr, last = true; + + do { + if (!ath_tid_has_buffered(tid)) +@@ -1264,38 +1333,34 @@ static void ath_tx_sched_aggr(struct ath + + INIT_LIST_HEAD(&bf_q); + +- status = ath_tx_form_aggr(sc, txq, tid, &bf_q, &aggr_len); +- +- /* +- * no frames picked up to be aggregated; +- * block-ack window is not open. +- */ +- if (list_empty(&bf_q)) ++ bf = ath_tx_get_tid_subframe(sc, txq, tid, &tid_q); ++ if (!bf) + break; + +- bf = list_first_entry(&bf_q, struct ath_buf, list); +- bf->bf_lastbf = list_entry(bf_q.prev, struct ath_buf, list); + tx_info = IEEE80211_SKB_CB(bf->bf_mpdu); ++ aggr = !!(tx_info->flags & IEEE80211_TX_CTL_AMPDU); ++ if ((aggr && txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) || ++ (!aggr && txq->axq_depth >= ATH_NON_AGGR_MIN_QDEPTH)) ++ break; ++ ++ ath_set_rates(tid->an->vif, tid->an->sta, bf); ++ if (aggr) ++ last = ath_tx_form_aggr(sc, txq, tid, &bf_q, bf, ++ tid_q, &aggr_len); ++ else ++ ath_tx_form_burst(sc, txq, tid, &bf_q, bf, tid_q); ++ ++ if (list_empty(&bf_q)) ++ return; + + if (tid->ac->clear_ps_filter) { + tid->ac->clear_ps_filter = false; + tx_info->flags |= IEEE80211_TX_CTL_CLEAR_PS_FILT; +- } else { +- tx_info->flags &= ~IEEE80211_TX_CTL_CLEAR_PS_FILT; +- } +- +- /* if only one frame, send as non-aggregate */ +- if (bf == bf->bf_lastbf) { +- aggr_len = get_frame_info(bf->bf_mpdu)->framelen; +- bf->bf_state.bf_type = BUF_AMPDU; +- } else { +- TX_STAT_INC(txq->axq_qnum, a_aggr); + } + + ath_tx_fill_desc(sc, bf, txq, aggr_len); + ath_tx_txqaddbuf(sc, txq, &bf_q, false); +- } while (txq->axq_ampdu_depth < ATH_AGGR_MIN_QDEPTH && +- status != ATH_AGGR_BAW_CLOSED); ++ } while (!last); + } + + int ath_tx_aggr_start(struct ath_softc *sc, struct ieee80211_sta *sta, +@@ -1319,6 +1384,9 @@ int ath_tx_aggr_start(struct ath_softc * + an->mpdudensity = density; + } + ++ /* force sequence number allocation for pending frames */ ++ ath_tx_tid_change_state(sc, txtid); ++ + txtid->active = true; + txtid->paused = true; + *ssn = txtid->seq_start = txtid->seq_next; +@@ -1340,6 +1408,7 @@ void ath_tx_aggr_stop(struct ath_softc * + txtid->active = false; + txtid->paused = false; + ath_tx_flush_tid(sc, txtid); ++ ath_tx_tid_change_state(sc, txtid); + ath_txq_unlock_complete(sc, txq); + } + +@@ -1854,58 +1923,6 @@ static void ath_tx_txqaddbuf(struct ath_ + } + } + +-static void ath_tx_send_ampdu(struct ath_softc *sc, struct ath_txq *txq, +- struct ath_atx_tid *tid, struct sk_buff *skb, +- struct ath_tx_control *txctl) +-{ +- struct ath_frame_info *fi = get_frame_info(skb); +- struct list_head bf_head; +- struct ath_buf *bf; +- +- /* +- * Do not queue to h/w when any of the following conditions is true: +- * - there are pending frames in software queue +- * - the TID is currently paused for ADDBA/BAR request +- * - seqno is not within block-ack window +- * - h/w queue depth exceeds low water mark +- */ +- if ((ath_tid_has_buffered(tid) || tid->paused || +- !BAW_WITHIN(tid->seq_start, tid->baw_size, tid->seq_next) || +- txq->axq_ampdu_depth >= ATH_AGGR_MIN_QDEPTH) && +- txq != sc->tx.uapsdq) { +- /* +- * Add this frame to software queue for scheduling later +- * for aggregation. +- */ +- TX_STAT_INC(txq->axq_qnum, a_queued_sw); +- __skb_queue_tail(&tid->buf_q, skb); +- if (!txctl->an || !txctl->an->sleeping) +- ath_tx_queue_tid(txq, tid); +- return; +- } +- +- bf = ath_tx_setup_buffer(sc, txq, tid, skb); +- if (!bf) { +- ath_txq_skb_done(sc, txq, skb); +- ieee80211_free_txskb(sc->hw, skb); +- return; +- } +- +- ath_set_rates(tid->an->vif, tid->an->sta, bf); +- bf->bf_state.bf_type = BUF_AMPDU; +- INIT_LIST_HEAD(&bf_head); +- list_add(&bf->list, &bf_head); +- +- /* Add sub-frame to BAW */ +- ath_tx_addto_baw(sc, tid, bf); +- +- /* Queue to h/w without aggregation */ +- TX_STAT_INC(txq->axq_qnum, a_queued_hw); +- bf->bf_lastbf = bf; +- ath_tx_fill_desc(sc, bf, txq, fi->framelen); +- ath_tx_txqaddbuf(sc, txq, &bf_head, false); +-} +- + static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq, + struct ath_atx_tid *tid, struct sk_buff *skb) + { +@@ -2131,20 +2148,25 @@ int ath_tx_start(struct ieee80211_hw *hw + ath_txq_unlock(sc, txq); + txq = sc->tx.uapsdq; + ath_txq_lock(sc, txq); +- } +- +- if (txctl->an && ieee80211_is_data_qos(hdr->frame_control)) { ++ } else if (txctl->an && ++ ieee80211_is_data_present(hdr->frame_control)) { + tid = ath_get_skb_tid(sc, txctl->an, skb); + + WARN_ON(tid->ac->txq != txctl->txq); +- } + +- if ((info->flags & IEEE80211_TX_CTL_AMPDU) && tid) { ++ if (info->flags & IEEE80211_TX_CTL_CLEAR_PS_FILT) ++ tid->ac->clear_ps_filter = true; ++ + /* +- * Try aggregation if it's a unicast data frame +- * and the destination is HT capable. ++ * Add this frame to software queue for scheduling later ++ * for aggregation. + */ +- ath_tx_send_ampdu(sc, txq, tid, skb, txctl); ++ TX_STAT_INC(txq->axq_qnum, a_queued_sw); ++ __skb_queue_tail(&tid->buf_q, skb); ++ if (!txctl->an->sleeping) ++ ath_tx_queue_tid(txq, tid); ++ ++ ath_txq_schedule(sc, txq); + goto out; + } + +--- a/drivers/net/wireless/ath/ath9k/ath9k.h ++++ b/drivers/net/wireless/ath/ath9k/ath9k.h +@@ -137,6 +137,8 @@ int ath_descdma_setup(struct ath_softc * + #define ATH_AGGR_ENCRYPTDELIM 10 + /* minimum h/w qdepth to be sustained to maximize aggregation */ + #define ATH_AGGR_MIN_QDEPTH 2 ++/* minimum h/w qdepth for non-aggregated traffic */ ++#define ATH_NON_AGGR_MIN_QDEPTH 8 + + #define IEEE80211_SEQ_SEQ_SHIFT 4 + #define IEEE80211_SEQ_MAX 4096 +@@ -173,12 +175,6 @@ int ath_descdma_setup(struct ath_softc * + + #define ATH_TX_COMPLETE_POLL_INT 1000 + +-enum ATH_AGGR_STATUS { +- ATH_AGGR_DONE, +- ATH_AGGR_BAW_CLOSED, +- ATH_AGGR_LIMITED, +-}; +- + #define ATH_TXFIFO_DEPTH 8 + struct ath_txq { + int mac80211_qnum; /* mac80211 queue number, -1 means not mac80211 Q */ -- cgit v1.2.3