When station connected to AP, and run TX traffic such as TCP/UDP, and system enter suspend state, then mac80211 call ath10k_flush with set drop flag, recently it only send wmi peer flush to firmware and firmware will flush all pending TX packets, for PCIe, firmware will indicate the TX packets status to ath10k, and then ath10k indicate to mac80211 TX complete with the status, then all the packets has been flushed at this moment. For SDIO chip, it is different, its TX complete indication is disabled by default, and it has a tx queue in ath10k, and its tx credit control is enabled, total tx credit is 96, when its credit is not sufficient, then the packets will buffered in the tx queue of ath10k, max packets is TARGET_TLV_NUM_MSDU_DESC_HL which is 1024, for SDIO, when mac80211 call ath10k_flush with set drop flag, maybe it have pending packets in tx queue of ath10k, and if it does not have sufficient tx credit, the packets will stay in queue untill tx credit report from firmware, if it is a noisy environment, tx speed is low and the tx credit report from firmware will delay more time, then the num_pending_tx will remain > 0 untill all packets send to firmware. After the 1st ath10k_flush, mac80211 will call the 2nd ath10k_flush without set drop flag immediately, then it will call to ath10k_mac_wait_tx_complete, and it wait untill num_pending_tx become to 0, in noisy environment, it is esay to wait about near 5 seconds, then it cause the suspend take long time. 1st and 2nd callstack of ath10k_flush [ 303.740427] ath10k_sdio mmc1:0001:1: ath10k_flush drop:1, pending:0-0 [ 303.740495] ------------[ cut here ]------------ [ 303.740739] WARNING: CPU: 1 PID: 3921 at /mnt/host/source/src/third_party/kernel/v4.19/drivers/net/wireless/ath/ath10k/mac.c:7025 ath10k_flush+0x54/0x104 [ath10k_core] [ 303.740757] Modules linked in: bridge stp llc ath10k_sdio ath10k_core rfcomm uinput cros_ec_rpmsg mtk_seninf mtk_cam_isp mtk_vcodec_enc mtk_fd mtk_vcodec_dec mtk_vcodec_common mtk_dip mtk_mdp3 videobuf2_dma_contig videobuf2_memops v4l2_mem2mem videobuf2_v4l2 videobuf2_common hid_google_hammer hci_uart btqca bluetooth dw9768 ov8856 ecdh_generic ov02a10 v4l2_fwnode mtk_scp mtk_rpmsg rpmsg_core mtk_scp_ipi ipt_MASQUERADE fuse iio_trig_sysfs cros_ec_sensors_ring cros_ec_sensors_sync cros_ec_light_prox cros_ec_sensors industrialio_triggered_buffer [ 303.740914] kfifo_buf cros_ec_activity cros_ec_sensors_core lzo_rle lzo_compress ath mac80211 zram cfg80211 joydev [last unloaded: ath10k_core] [ 303.741009] CPU: 1 PID: 3921 Comm: kworker/u16:10 Tainted: G W 4.19.95 #2 [ 303.741027] Hardware name: MediaTek krane sku176 board (DT) [ 303.741061] Workqueue: events_unbound async_run_entry_fn [ 303.741086] pstate: 60000005 (nZCv daif -PAN -UAO) [ 303.741166] pc : ath10k_flush+0x54/0x104 [ath10k_core] [ 303.741244] lr : ath10k_flush+0x54/0x104 [ath10k_core] [ 303.741260] sp : ffffffdf080e77a0 [ 303.741276] x29: ffffffdf080e77a0 x28: ffffffdef3730040 [ 303.741300] x27: ffffff907c2240a0 x26: ffffffde6ff39afc [ 303.741321] x25: ffffffdef3730040 x24: ffffff907bf61018 [ 303.741343] x23: ffffff907c2240a0 x22: ffffffde6ff39a50 [ 303.741364] x21: 0000000000000001 x20: ffffffde6ff39a50 [ 303.741385] x19: ffffffde6bac2420 x18: 0000000000017200 [ 303.741407] x17: ffffff907c24a000 x16: 0000000000000037 [ 303.741428] x15: ffffff907b49a568 x14: ffffff907cf332c1 [ 303.741476] x13: 00000000000922e4 x12: 0000000000000000 [ 303.741497] x11: 0000000000000001 x10: 0000000000000007 [ 303.741518] x9 : f2256b8c1de4bc00 x8 : f2256b8c1de4bc00 [ 303.741539] x7 : ffffff907ab5e764 x6 : 0000000000000000 [ 303.741560] x5 : 0000000000000080 x4 : 0000000000000001 [ 303.741582] x3 : ffffffdf080e74a8 x2 : ffffff907aa91244 [ 303.741603] x1 : ffffffdf080e74a8 x0 : 0000000000000024 [ 303.741624] Call trace: [ 303.741701] ath10k_flush+0x54/0x104 [ath10k_core] [ 303.741941] __ieee80211_flush_queues+0x1dc/0x358 [mac80211] [ 303.742098] ieee80211_flush_queues+0x34/0x44 [mac80211] [ 303.742253] ieee80211_set_disassoc+0xc0/0x5ec [mac80211] [ 303.742399] ieee80211_mgd_deauth+0x720/0x7d4 [mac80211] [ 303.742535] ieee80211_deauth+0x24/0x30 [mac80211] [ 303.742720] cfg80211_mlme_deauth+0x250/0x3bc [cfg80211] [ 303.742849] cfg80211_mlme_down+0x90/0xd0 [cfg80211] [ 303.742971] cfg80211_disconnect+0x340/0x3a0 [cfg80211] [ 303.743087] __cfg80211_leave+0xe4/0x17c [cfg80211] [ 303.743203] cfg80211_leave+0x38/0x50 [cfg80211] [ 303.743319] wiphy_suspend+0x84/0x5bc [cfg80211] [ 303.743335] dpm_run_callback+0x170/0x304 [ 303.743346] __device_suspend+0x2dc/0x3e8 [ 303.743356] async_suspend+0x2c/0xb0 [ 303.743370] async_run_entry_fn+0x48/0xf8 [ 303.743383] process_one_work+0x304/0x604 [ 303.743394] worker_thread+0x248/0x3f4 [ 303.743403] kthread+0x120/0x130 [ 303.743416] ret_from_fork+0x10/0x18 [ 303.743812] ath10k_sdio mmc1:0001:1: ath10k_flush drop:0, pending:0-0 [ 303.743858] ------------[ cut here ]------------ [ 303.744057] WARNING: CPU: 1 PID: 3921 at /mnt/host/source/src/third_party/kernel/v4.19/drivers/net/wireless/ath/ath10k/mac.c:7025 ath10k_flush+0x54/0x104 [ath10k_core] [ 303.744075] Modules linked in: bridge stp llc ath10k_sdio ath10k_core rfcomm uinput cros_ec_rpmsg mtk_seninf mtk_cam_isp mtk_vcodec_enc mtk_fd mtk_vcodec_dec mtk_vcodec_common mtk_dip mtk_mdp3 videobuf2_dma_contig videobuf2_memops v4l2_mem2mem videobuf2_v4l2 videobuf2_common hid_google_hammer hci_uart btqca bluetooth dw9768 ov8856 ecdh_generic ov02a10 v4l2_fwnode mtk_scp mtk_rpmsg rpmsg_core mtk_scp_ipi ipt_MASQUERADE fuse iio_trig_sysfs cros_ec_sensors_ring cros_ec_sensors_sync cros_ec_light_prox cros_ec_sensors industrialio_triggered_buffer kfifo_buf cros_ec_activity cros_ec_sensors_core lzo_rle lzo_compress ath mac80211 zram cfg80211 joydev [last unloaded: ath10k_core] [ 303.744256] CPU: 1 PID: 3921 Comm: kworker/u16:10 Tainted: G W 4.19.95 #2 [ 303.744273] Hardware name: MediaTek krane sku176 board (DT) [ 303.744301] Workqueue: events_unbound async_run_entry_fn [ 303.744325] pstate: 60000005 (nZCv daif -PAN -UAO) [ 303.744403] pc : ath10k_flush+0x54/0x104 [ath10k_core] [ 303.744480] lr : ath10k_flush+0x54/0x104 [ath10k_core] [ 303.744496] sp : ffffffdf080e77a0 [ 303.744512] x29: ffffffdf080e77a0 x28: ffffffdef3730040 [ 303.744534] x27: ffffff907c2240a0 x26: ffffffde6ff39afc [ 303.744556] x25: ffffffdef3730040 x24: ffffff907bf61018 [ 303.744577] x23: ffffff907c2240a0 x22: ffffffde6ff39a50 [ 303.744598] x21: 0000000000000000 x20: ffffffde6ff39a50 [ 303.744620] x19: ffffffde6bac2420 x18: 000000000001831c [ 303.744641] x17: ffffff907c24a000 x16: 0000000000000037 [ 303.744662] x15: ffffff907b49a568 x14: ffffff907cf332c1 [ 303.744683] x13: 00000000000922ea x12: 0000000000000000 [ 303.744704] x11: 0000000000000001 x10: 0000000000000007 [ 303.744747] x9 : f2256b8c1de4bc00 x8 : f2256b8c1de4bc00 [ 303.744768] x7 : ffffff907ab5e764 x6 : 0000000000000000 [ 303.744789] x5 : 0000000000000080 x4 : 0000000000000001 [ 303.744810] x3 : ffffffdf080e74a8 x2 : ffffff907aa91244 [ 303.744831] x1 : ffffffdf080e74a8 x0 : 0000000000000024 [ 303.744853] Call trace: [ 303.744929] ath10k_flush+0x54/0x104 [ath10k_core] [ 303.745098] __ieee80211_flush_queues+0x1dc/0x358 [mac80211] [ 303.745277] ieee80211_flush_queues+0x34/0x44 [mac80211] [ 303.745424] ieee80211_set_disassoc+0x108/0x5ec [mac80211] [ 303.745569] ieee80211_mgd_deauth+0x720/0x7d4 [mac80211] [ 303.745706] ieee80211_deauth+0x24/0x30 [mac80211] [ 303.745853] cfg80211_mlme_deauth+0x250/0x3bc [cfg80211] [ 303.745979] cfg80211_mlme_down+0x90/0xd0 [cfg80211] [ 303.746103] cfg80211_disconnect+0x340/0x3a0 [cfg80211] [ 303.746219] __cfg80211_leave+0xe4/0x17c [cfg80211] [ 303.746335] cfg80211_leave+0x38/0x50 [cfg80211] [ 303.746452] wiphy_suspend+0x84/0x5bc [cfg80211] [ 303.746467] dpm_run_callback+0x170/0x304 [ 303.746477] __device_suspend+0x2dc/0x3e8 [ 303.746487] async_suspend+0x2c/0xb0 [ 303.746498] async_run_entry_fn+0x48/0xf8 [ 303.746510] process_one_work+0x304/0x604 [ 303.746521] worker_thread+0x248/0x3f4 [ 303.746530] kthread+0x120/0x130 [ 303.746542] ret_from_fork+0x10/0x18 one sample's debugging log: it wait 3190 ms(5000 - 1810). 1st ath10k_flush, it has 120 packets in tx queue of ath10k: <...>-1513 [000] .... 25374.786005: ath10k_log_err: ath10k_sdio mmc1:0001:1 ath10k_flush drop:1, pending:120-0 <...>-1513 [000] ...1 25374.788375: ath10k_log_warn: ath10k_sdio mmc1:0001:1 ath10k_htt_tx_mgmt_inc_pending htt->num_pending_mgmt_tx:0 <...>-1500 [001] .... 25374.790143: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:121 2st ath10k_flush, it has 121 packets in tx queue of ath10k: <...>-1513 [000] .... 25374.790571: ath10k_log_err: ath10k_sdio mmc1:0001:1 ath10k_flush drop:0, pending:121-0 <...>-1513 [000] .... 25374.791990: ath10k_log_err: ath10k_sdio mmc1:0001:1 ath10k_mac_wait_tx_complete state:1 pending:121-0 <...>-1508 [001] .... 25374.792696: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:46 <...>-1508 [001] .... 25374.792700: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:46 <...>-1508 [001] .... 25374.792729: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:121 <...>-1508 [001] .... 25374.792937: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:88, count:32, len:49792 <...>-1508 [001] .... 25374.793031: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:75, count:14, len:21784 kworker/u16:0-25773 [003] .... 25374.793701: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx complete, eid:1, pending complete count:46 <...>-1881 [000] .... 25375.073178: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:24 <...>-1881 [000] .... 25375.073182: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:24 <...>-1881 [000] .... 25375.073429: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:75 <...>-1879 [001] .... 25375.074090: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx complete, eid:1, pending complete count:24 <...>-1881 [000] .... 25375.074123: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:51, count:24, len:37344 <...>-1879 [001] .... 25375.270126: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:26 <...>-1879 [001] .... 25375.270130: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:26 <...>-1488 [000] .... 25375.270174: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:51 <...>-1488 [000] .... 25375.270529: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:25, count:26, len:40456 <...>-1879 [001] .... 25375.270693: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx complete, eid:1, pending complete count:26 <...>-1488 [001] .... 25377.775885: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:12 <...>-1488 [001] .... 25377.775890: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:12 <...>-1488 [001] .... 25377.775933: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:25 <...>-1488 [001] .... 25377.776059: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:13, count:12, len:18672 <...>-1879 [001] .... 25377.776100: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx complete, eid:1, pending complete count:12 <...>-1488 [001] .... 25377.878079: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:15 <...>-1488 [001] .... 25377.878087: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:15 <...>-1879 [000] .... 25377.878323: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:13 <...>-1879 [000] .... 25377.878487: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx status:0, eid:1, req count:0, count:13, len:20228 <...>-1879 [000] .... 25377.878497: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx complete, eid:1, pending complete count:13 <...>-1488 [001] .... 25377.919927: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit update: delta:11 <...>-1488 [001] .... 25377.919932: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 credit total:13 <...>-1488 [001] .... 25377.919976: ath10k_log_dbg: ath10k_sdio mmc1:0001:1 bundle tx work, eid:1, count:0 <...>-1881 [000] .... 25377.982645: ath10k_log_warn: ath10k_sdio mmc1:0001:1 HTT_T2H_MSG_TYPE_MGMT_TX_COMPLETION status:0 <...>-1513 [001] .... 25377.982973: ath10k_log_err: ath10k_sdio mmc1:0001:1 ath10k_mac_wait_tx_complete time_left:1810, pending:0-0 Flush all pending TX packets for the 1st ath10k_flush reduced the wait time of the 2nd ath10k_flush and then suspend take short time. This Patch only effect SDIO chips. Tested with QCA6174 SDIO with firmware WLAN.RMH.4.4.1-00042. Signed-off-by: Wen Gong <wgong@codeaurora.org> Signed-off-by: Kalle Valo <kvalo@codeaurora.org> Link: https://lore.kernel.org/r/20200415233730.10581-1-wgong@codeaurora.org
1847 lines
50 KiB
C
1847 lines
50 KiB
C
// SPDX-License-Identifier: ISC
|
|
/*
|
|
* Copyright (c) 2005-2011 Atheros Communications Inc.
|
|
* Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
|
|
*/
|
|
|
|
#include <linux/etherdevice.h>
|
|
#include "htt.h"
|
|
#include "mac.h"
|
|
#include "hif.h"
|
|
#include "txrx.h"
|
|
#include "debug.h"
|
|
|
|
static u8 ath10k_htt_tx_txq_calc_size(size_t count)
|
|
{
|
|
int exp;
|
|
int factor;
|
|
|
|
exp = 0;
|
|
factor = count >> 7;
|
|
|
|
while (factor >= 64 && exp < 4) {
|
|
factor >>= 3;
|
|
exp++;
|
|
}
|
|
|
|
if (exp == 4)
|
|
return 0xff;
|
|
|
|
if (count > 0)
|
|
factor = max(1, factor);
|
|
|
|
return SM(exp, HTT_TX_Q_STATE_ENTRY_EXP) |
|
|
SM(factor, HTT_TX_Q_STATE_ENTRY_FACTOR);
|
|
}
|
|
|
|
static void __ath10k_htt_tx_txq_recalc(struct ieee80211_hw *hw,
|
|
struct ieee80211_txq *txq)
|
|
{
|
|
struct ath10k *ar = hw->priv;
|
|
struct ath10k_sta *arsta;
|
|
struct ath10k_vif *arvif = (void *)txq->vif->drv_priv;
|
|
unsigned long frame_cnt;
|
|
unsigned long byte_cnt;
|
|
int idx;
|
|
u32 bit;
|
|
u16 peer_id;
|
|
u8 tid;
|
|
u8 count;
|
|
|
|
lockdep_assert_held(&ar->htt.tx_lock);
|
|
|
|
if (!ar->htt.tx_q_state.enabled)
|
|
return;
|
|
|
|
if (ar->htt.tx_q_state.mode != HTT_TX_MODE_SWITCH_PUSH_PULL)
|
|
return;
|
|
|
|
if (txq->sta) {
|
|
arsta = (void *)txq->sta->drv_priv;
|
|
peer_id = arsta->peer_id;
|
|
} else {
|
|
peer_id = arvif->peer_id;
|
|
}
|
|
|
|
tid = txq->tid;
|
|
bit = BIT(peer_id % 32);
|
|
idx = peer_id / 32;
|
|
|
|
ieee80211_txq_get_depth(txq, &frame_cnt, &byte_cnt);
|
|
count = ath10k_htt_tx_txq_calc_size(byte_cnt);
|
|
|
|
if (unlikely(peer_id >= ar->htt.tx_q_state.num_peers) ||
|
|
unlikely(tid >= ar->htt.tx_q_state.num_tids)) {
|
|
ath10k_warn(ar, "refusing to update txq for peer_id %hu tid %hhu due to out of bounds\n",
|
|
peer_id, tid);
|
|
return;
|
|
}
|
|
|
|
ar->htt.tx_q_state.vaddr->count[tid][peer_id] = count;
|
|
ar->htt.tx_q_state.vaddr->map[tid][idx] &= ~bit;
|
|
ar->htt.tx_q_state.vaddr->map[tid][idx] |= count ? bit : 0;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt tx txq state update peer_id %hu tid %hhu count %hhu\n",
|
|
peer_id, tid, count);
|
|
}
|
|
|
|
static void __ath10k_htt_tx_txq_sync(struct ath10k *ar)
|
|
{
|
|
u32 seq;
|
|
size_t size;
|
|
|
|
lockdep_assert_held(&ar->htt.tx_lock);
|
|
|
|
if (!ar->htt.tx_q_state.enabled)
|
|
return;
|
|
|
|
if (ar->htt.tx_q_state.mode != HTT_TX_MODE_SWITCH_PUSH_PULL)
|
|
return;
|
|
|
|
seq = le32_to_cpu(ar->htt.tx_q_state.vaddr->seq);
|
|
seq++;
|
|
ar->htt.tx_q_state.vaddr->seq = cpu_to_le32(seq);
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt tx txq state update commit seq %u\n",
|
|
seq);
|
|
|
|
size = sizeof(*ar->htt.tx_q_state.vaddr);
|
|
dma_sync_single_for_device(ar->dev,
|
|
ar->htt.tx_q_state.paddr,
|
|
size,
|
|
DMA_TO_DEVICE);
|
|
}
|
|
|
|
void ath10k_htt_tx_txq_recalc(struct ieee80211_hw *hw,
|
|
struct ieee80211_txq *txq)
|
|
{
|
|
struct ath10k *ar = hw->priv;
|
|
|
|
spin_lock_bh(&ar->htt.tx_lock);
|
|
__ath10k_htt_tx_txq_recalc(hw, txq);
|
|
spin_unlock_bh(&ar->htt.tx_lock);
|
|
}
|
|
|
|
void ath10k_htt_tx_txq_sync(struct ath10k *ar)
|
|
{
|
|
spin_lock_bh(&ar->htt.tx_lock);
|
|
__ath10k_htt_tx_txq_sync(ar);
|
|
spin_unlock_bh(&ar->htt.tx_lock);
|
|
}
|
|
|
|
void ath10k_htt_tx_txq_update(struct ieee80211_hw *hw,
|
|
struct ieee80211_txq *txq)
|
|
{
|
|
struct ath10k *ar = hw->priv;
|
|
|
|
spin_lock_bh(&ar->htt.tx_lock);
|
|
__ath10k_htt_tx_txq_recalc(hw, txq);
|
|
__ath10k_htt_tx_txq_sync(ar);
|
|
spin_unlock_bh(&ar->htt.tx_lock);
|
|
}
|
|
|
|
void ath10k_htt_tx_dec_pending(struct ath10k_htt *htt)
|
|
{
|
|
lockdep_assert_held(&htt->tx_lock);
|
|
|
|
htt->num_pending_tx--;
|
|
if (htt->num_pending_tx == htt->max_num_pending_tx - 1)
|
|
ath10k_mac_tx_unlock(htt->ar, ATH10K_TX_PAUSE_Q_FULL);
|
|
}
|
|
|
|
int ath10k_htt_tx_inc_pending(struct ath10k_htt *htt)
|
|
{
|
|
lockdep_assert_held(&htt->tx_lock);
|
|
|
|
if (htt->num_pending_tx >= htt->max_num_pending_tx)
|
|
return -EBUSY;
|
|
|
|
htt->num_pending_tx++;
|
|
if (htt->num_pending_tx == htt->max_num_pending_tx)
|
|
ath10k_mac_tx_lock(htt->ar, ATH10K_TX_PAUSE_Q_FULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ath10k_htt_tx_mgmt_inc_pending(struct ath10k_htt *htt, bool is_mgmt,
|
|
bool is_presp)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
|
|
lockdep_assert_held(&htt->tx_lock);
|
|
|
|
if (!is_mgmt || !ar->hw_params.max_probe_resp_desc_thres)
|
|
return 0;
|
|
|
|
if (is_presp &&
|
|
ar->hw_params.max_probe_resp_desc_thres < htt->num_pending_mgmt_tx)
|
|
return -EBUSY;
|
|
|
|
htt->num_pending_mgmt_tx++;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ath10k_htt_tx_mgmt_dec_pending(struct ath10k_htt *htt)
|
|
{
|
|
lockdep_assert_held(&htt->tx_lock);
|
|
|
|
if (!htt->ar->hw_params.max_probe_resp_desc_thres)
|
|
return;
|
|
|
|
htt->num_pending_mgmt_tx--;
|
|
}
|
|
|
|
int ath10k_htt_tx_alloc_msdu_id(struct ath10k_htt *htt, struct sk_buff *skb)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
int ret;
|
|
|
|
spin_lock_bh(&htt->tx_lock);
|
|
ret = idr_alloc(&htt->pending_tx, skb, 0,
|
|
htt->max_num_pending_tx, GFP_ATOMIC);
|
|
spin_unlock_bh(&htt->tx_lock);
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt tx alloc msdu_id %d\n", ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void ath10k_htt_tx_free_msdu_id(struct ath10k_htt *htt, u16 msdu_id)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
|
|
lockdep_assert_held(&htt->tx_lock);
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt tx free msdu_id %hu\n", msdu_id);
|
|
|
|
idr_remove(&htt->pending_tx, msdu_id);
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_cont_txbuf_32(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
if (!htt->txbuf.vaddr_txbuff_32)
|
|
return;
|
|
|
|
size = htt->txbuf.size;
|
|
dma_free_coherent(ar->dev, size, htt->txbuf.vaddr_txbuff_32,
|
|
htt->txbuf.paddr);
|
|
htt->txbuf.vaddr_txbuff_32 = NULL;
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_cont_txbuf_32(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct ath10k_htt_txbuf_32);
|
|
|
|
htt->txbuf.vaddr_txbuff_32 = dma_alloc_coherent(ar->dev, size,
|
|
&htt->txbuf.paddr,
|
|
GFP_KERNEL);
|
|
if (!htt->txbuf.vaddr_txbuff_32)
|
|
return -ENOMEM;
|
|
|
|
htt->txbuf.size = size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_cont_txbuf_64(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
if (!htt->txbuf.vaddr_txbuff_64)
|
|
return;
|
|
|
|
size = htt->txbuf.size;
|
|
dma_free_coherent(ar->dev, size, htt->txbuf.vaddr_txbuff_64,
|
|
htt->txbuf.paddr);
|
|
htt->txbuf.vaddr_txbuff_64 = NULL;
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_cont_txbuf_64(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct ath10k_htt_txbuf_64);
|
|
|
|
htt->txbuf.vaddr_txbuff_64 = dma_alloc_coherent(ar->dev, size,
|
|
&htt->txbuf.paddr,
|
|
GFP_KERNEL);
|
|
if (!htt->txbuf.vaddr_txbuff_64)
|
|
return -ENOMEM;
|
|
|
|
htt->txbuf.size = size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_cont_frag_desc_32(struct ath10k_htt *htt)
|
|
{
|
|
size_t size;
|
|
|
|
if (!htt->frag_desc.vaddr_desc_32)
|
|
return;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct htt_msdu_ext_desc);
|
|
|
|
dma_free_coherent(htt->ar->dev,
|
|
size,
|
|
htt->frag_desc.vaddr_desc_32,
|
|
htt->frag_desc.paddr);
|
|
|
|
htt->frag_desc.vaddr_desc_32 = NULL;
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_cont_frag_desc_32(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
if (!ar->hw_params.continuous_frag_desc)
|
|
return 0;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct htt_msdu_ext_desc);
|
|
htt->frag_desc.vaddr_desc_32 = dma_alloc_coherent(ar->dev, size,
|
|
&htt->frag_desc.paddr,
|
|
GFP_KERNEL);
|
|
if (!htt->frag_desc.vaddr_desc_32) {
|
|
ath10k_err(ar, "failed to alloc fragment desc memory\n");
|
|
return -ENOMEM;
|
|
}
|
|
htt->frag_desc.size = size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_cont_frag_desc_64(struct ath10k_htt *htt)
|
|
{
|
|
size_t size;
|
|
|
|
if (!htt->frag_desc.vaddr_desc_64)
|
|
return;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct htt_msdu_ext_desc_64);
|
|
|
|
dma_free_coherent(htt->ar->dev,
|
|
size,
|
|
htt->frag_desc.vaddr_desc_64,
|
|
htt->frag_desc.paddr);
|
|
|
|
htt->frag_desc.vaddr_desc_64 = NULL;
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_cont_frag_desc_64(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
if (!ar->hw_params.continuous_frag_desc)
|
|
return 0;
|
|
|
|
size = htt->max_num_pending_tx *
|
|
sizeof(struct htt_msdu_ext_desc_64);
|
|
|
|
htt->frag_desc.vaddr_desc_64 = dma_alloc_coherent(ar->dev, size,
|
|
&htt->frag_desc.paddr,
|
|
GFP_KERNEL);
|
|
if (!htt->frag_desc.vaddr_desc_64) {
|
|
ath10k_err(ar, "failed to alloc fragment desc memory\n");
|
|
return -ENOMEM;
|
|
}
|
|
htt->frag_desc.size = size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_txq(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
|
|
if (!test_bit(ATH10K_FW_FEATURE_PEER_FLOW_CONTROL,
|
|
ar->running_fw->fw_file.fw_features))
|
|
return;
|
|
|
|
size = sizeof(*htt->tx_q_state.vaddr);
|
|
|
|
dma_unmap_single(ar->dev, htt->tx_q_state.paddr, size, DMA_TO_DEVICE);
|
|
kfree(htt->tx_q_state.vaddr);
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_txq(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
size_t size;
|
|
int ret;
|
|
|
|
if (!test_bit(ATH10K_FW_FEATURE_PEER_FLOW_CONTROL,
|
|
ar->running_fw->fw_file.fw_features))
|
|
return 0;
|
|
|
|
htt->tx_q_state.num_peers = HTT_TX_Q_STATE_NUM_PEERS;
|
|
htt->tx_q_state.num_tids = HTT_TX_Q_STATE_NUM_TIDS;
|
|
htt->tx_q_state.type = HTT_Q_DEPTH_TYPE_BYTES;
|
|
|
|
size = sizeof(*htt->tx_q_state.vaddr);
|
|
htt->tx_q_state.vaddr = kzalloc(size, GFP_KERNEL);
|
|
if (!htt->tx_q_state.vaddr)
|
|
return -ENOMEM;
|
|
|
|
htt->tx_q_state.paddr = dma_map_single(ar->dev, htt->tx_q_state.vaddr,
|
|
size, DMA_TO_DEVICE);
|
|
ret = dma_mapping_error(ar->dev, htt->tx_q_state.paddr);
|
|
if (ret) {
|
|
ath10k_warn(ar, "failed to dma map tx_q_state: %d\n", ret);
|
|
kfree(htt->tx_q_state.vaddr);
|
|
return -EIO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_tx_free_txdone_fifo(struct ath10k_htt *htt)
|
|
{
|
|
WARN_ON(!kfifo_is_empty(&htt->txdone_fifo));
|
|
kfifo_free(&htt->txdone_fifo);
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_txdone_fifo(struct ath10k_htt *htt)
|
|
{
|
|
int ret;
|
|
size_t size;
|
|
|
|
size = roundup_pow_of_two(htt->max_num_pending_tx);
|
|
ret = kfifo_alloc(&htt->txdone_fifo, size, GFP_KERNEL);
|
|
return ret;
|
|
}
|
|
|
|
static int ath10k_htt_tx_alloc_buf(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
int ret;
|
|
|
|
ret = ath10k_htt_alloc_txbuff(htt);
|
|
if (ret) {
|
|
ath10k_err(ar, "failed to alloc cont tx buffer: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
ret = ath10k_htt_alloc_frag_desc(htt);
|
|
if (ret) {
|
|
ath10k_err(ar, "failed to alloc cont frag desc: %d\n", ret);
|
|
goto free_txbuf;
|
|
}
|
|
|
|
ret = ath10k_htt_tx_alloc_txq(htt);
|
|
if (ret) {
|
|
ath10k_err(ar, "failed to alloc txq: %d\n", ret);
|
|
goto free_frag_desc;
|
|
}
|
|
|
|
ret = ath10k_htt_tx_alloc_txdone_fifo(htt);
|
|
if (ret) {
|
|
ath10k_err(ar, "failed to alloc txdone fifo: %d\n", ret);
|
|
goto free_txq;
|
|
}
|
|
|
|
return 0;
|
|
|
|
free_txq:
|
|
ath10k_htt_tx_free_txq(htt);
|
|
|
|
free_frag_desc:
|
|
ath10k_htt_free_frag_desc(htt);
|
|
|
|
free_txbuf:
|
|
ath10k_htt_free_txbuff(htt);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int ath10k_htt_tx_start(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
int ret;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_BOOT, "htt tx max num pending tx %d\n",
|
|
htt->max_num_pending_tx);
|
|
|
|
spin_lock_init(&htt->tx_lock);
|
|
idr_init(&htt->pending_tx);
|
|
|
|
if (htt->tx_mem_allocated)
|
|
return 0;
|
|
|
|
if (ar->bus_param.dev_type == ATH10K_DEV_TYPE_HL)
|
|
return 0;
|
|
|
|
ret = ath10k_htt_tx_alloc_buf(htt);
|
|
if (ret)
|
|
goto free_idr_pending_tx;
|
|
|
|
htt->tx_mem_allocated = true;
|
|
|
|
return 0;
|
|
|
|
free_idr_pending_tx:
|
|
idr_destroy(&htt->pending_tx);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ath10k_htt_tx_clean_up_pending(int msdu_id, void *skb, void *ctx)
|
|
{
|
|
struct ath10k *ar = ctx;
|
|
struct ath10k_htt *htt = &ar->htt;
|
|
struct htt_tx_done tx_done = {0};
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "force cleanup msdu_id %hu\n", msdu_id);
|
|
|
|
tx_done.msdu_id = msdu_id;
|
|
tx_done.status = HTT_TX_COMPL_STATE_DISCARD;
|
|
|
|
ath10k_txrx_tx_unref(htt, &tx_done);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ath10k_htt_tx_destroy(struct ath10k_htt *htt)
|
|
{
|
|
if (!htt->tx_mem_allocated)
|
|
return;
|
|
|
|
ath10k_htt_free_txbuff(htt);
|
|
ath10k_htt_tx_free_txq(htt);
|
|
ath10k_htt_free_frag_desc(htt);
|
|
ath10k_htt_tx_free_txdone_fifo(htt);
|
|
htt->tx_mem_allocated = false;
|
|
}
|
|
|
|
static void ath10k_htt_flush_tx_queue(struct ath10k_htt *htt)
|
|
{
|
|
ath10k_htc_stop_hl(htt->ar);
|
|
idr_for_each(&htt->pending_tx, ath10k_htt_tx_clean_up_pending, htt->ar);
|
|
}
|
|
|
|
void ath10k_htt_tx_stop(struct ath10k_htt *htt)
|
|
{
|
|
ath10k_htt_flush_tx_queue(htt);
|
|
idr_destroy(&htt->pending_tx);
|
|
}
|
|
|
|
void ath10k_htt_tx_free(struct ath10k_htt *htt)
|
|
{
|
|
ath10k_htt_tx_stop(htt);
|
|
ath10k_htt_tx_destroy(htt);
|
|
}
|
|
|
|
void ath10k_htt_op_ep_tx_credits(struct ath10k *ar)
|
|
{
|
|
queue_work(ar->workqueue, &ar->bundle_tx_work);
|
|
}
|
|
|
|
void ath10k_htt_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb)
|
|
{
|
|
struct ath10k_htt *htt = &ar->htt;
|
|
struct htt_tx_done tx_done = {0};
|
|
struct htt_cmd_hdr *htt_hdr;
|
|
struct htt_data_tx_desc *desc_hdr = NULL;
|
|
u16 flags1 = 0;
|
|
u8 msg_type = 0;
|
|
|
|
if (htt->disable_tx_comp) {
|
|
htt_hdr = (struct htt_cmd_hdr *)skb->data;
|
|
msg_type = htt_hdr->msg_type;
|
|
|
|
if (msg_type == HTT_H2T_MSG_TYPE_TX_FRM) {
|
|
desc_hdr = (struct htt_data_tx_desc *)
|
|
(skb->data + sizeof(*htt_hdr));
|
|
flags1 = __le16_to_cpu(desc_hdr->flags1);
|
|
}
|
|
}
|
|
|
|
dev_kfree_skb_any(skb);
|
|
|
|
if ((!htt->disable_tx_comp) || (msg_type != HTT_H2T_MSG_TYPE_TX_FRM))
|
|
return;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT,
|
|
"htt tx complete msdu id:%u ,flags1:%x\n",
|
|
__le16_to_cpu(desc_hdr->id), flags1);
|
|
|
|
if (flags1 & HTT_DATA_TX_DESC_FLAGS1_TX_COMPLETE)
|
|
return;
|
|
|
|
tx_done.status = HTT_TX_COMPL_STATE_ACK;
|
|
tx_done.msdu_id = __le16_to_cpu(desc_hdr->id);
|
|
ath10k_txrx_tx_unref(&ar->htt, &tx_done);
|
|
}
|
|
|
|
void ath10k_htt_hif_tx_complete(struct ath10k *ar, struct sk_buff *skb)
|
|
{
|
|
dev_kfree_skb_any(skb);
|
|
}
|
|
EXPORT_SYMBOL(ath10k_htt_hif_tx_complete);
|
|
|
|
int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
int len = 0;
|
|
int ret;
|
|
|
|
len += sizeof(cmd->hdr);
|
|
len += sizeof(cmd->ver_req);
|
|
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_VERSION_REQ;
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ath10k_htt_h2t_stats_req(struct ath10k_htt *htt, u32 mask, u32 reset_mask,
|
|
u64 cookie)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct htt_stats_req *req;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
int len = 0, ret;
|
|
|
|
len += sizeof(cmd->hdr);
|
|
len += sizeof(cmd->stats_req);
|
|
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_STATS_REQ;
|
|
|
|
req = &cmd->stats_req;
|
|
|
|
memset(req, 0, sizeof(*req));
|
|
|
|
/* currently we support only max 24 bit masks so no need to worry
|
|
* about endian support
|
|
*/
|
|
memcpy(req->upload_types, &mask, 3);
|
|
memcpy(req->reset_types, &reset_mask, 3);
|
|
req->stat_type = HTT_STATS_REQ_CFG_STAT_TYPE_INVALID;
|
|
req->cookie_lsb = cpu_to_le32(cookie & 0xffffffff);
|
|
req->cookie_msb = cpu_to_le32((cookie & 0xffffffff00000000ULL) >> 32);
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
ath10k_warn(ar, "failed to send htt type stats request: %d",
|
|
ret);
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_send_frag_desc_bank_cfg_32(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
struct htt_frag_desc_bank_cfg32 *cfg;
|
|
int ret, size;
|
|
u8 info;
|
|
|
|
if (!ar->hw_params.continuous_frag_desc)
|
|
return 0;
|
|
|
|
if (!htt->frag_desc.paddr) {
|
|
ath10k_warn(ar, "invalid frag desc memory\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
size = sizeof(cmd->hdr) + sizeof(cmd->frag_desc_bank_cfg32);
|
|
skb = ath10k_htc_alloc_skb(ar, size);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, size);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_FRAG_DESC_BANK_CFG;
|
|
|
|
info = 0;
|
|
info |= SM(htt->tx_q_state.type,
|
|
HTT_FRAG_DESC_BANK_CFG_INFO_Q_STATE_DEPTH_TYPE);
|
|
|
|
if (test_bit(ATH10K_FW_FEATURE_PEER_FLOW_CONTROL,
|
|
ar->running_fw->fw_file.fw_features))
|
|
info |= HTT_FRAG_DESC_BANK_CFG_INFO_Q_STATE_VALID;
|
|
|
|
cfg = &cmd->frag_desc_bank_cfg32;
|
|
cfg->info = info;
|
|
cfg->num_banks = 1;
|
|
cfg->desc_size = sizeof(struct htt_msdu_ext_desc);
|
|
cfg->bank_base_addrs[0] = __cpu_to_le32(htt->frag_desc.paddr);
|
|
cfg->bank_id[0].bank_min_id = 0;
|
|
cfg->bank_id[0].bank_max_id = __cpu_to_le16(htt->max_num_pending_tx -
|
|
1);
|
|
|
|
cfg->q_state.paddr = cpu_to_le32(htt->tx_q_state.paddr);
|
|
cfg->q_state.num_peers = cpu_to_le16(htt->tx_q_state.num_peers);
|
|
cfg->q_state.num_tids = cpu_to_le16(htt->tx_q_state.num_tids);
|
|
cfg->q_state.record_size = HTT_TX_Q_STATE_ENTRY_SIZE;
|
|
cfg->q_state.record_multiplier = HTT_TX_Q_STATE_ENTRY_MULTIPLIER;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt frag desc bank cmd\n");
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
ath10k_warn(ar, "failed to send frag desc bank cfg request: %d\n",
|
|
ret);
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_send_frag_desc_bank_cfg_64(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
struct htt_frag_desc_bank_cfg64 *cfg;
|
|
int ret, size;
|
|
u8 info;
|
|
|
|
if (!ar->hw_params.continuous_frag_desc)
|
|
return 0;
|
|
|
|
if (!htt->frag_desc.paddr) {
|
|
ath10k_warn(ar, "invalid frag desc memory\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
size = sizeof(cmd->hdr) + sizeof(cmd->frag_desc_bank_cfg64);
|
|
skb = ath10k_htc_alloc_skb(ar, size);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, size);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_FRAG_DESC_BANK_CFG;
|
|
|
|
info = 0;
|
|
info |= SM(htt->tx_q_state.type,
|
|
HTT_FRAG_DESC_BANK_CFG_INFO_Q_STATE_DEPTH_TYPE);
|
|
|
|
if (test_bit(ATH10K_FW_FEATURE_PEER_FLOW_CONTROL,
|
|
ar->running_fw->fw_file.fw_features))
|
|
info |= HTT_FRAG_DESC_BANK_CFG_INFO_Q_STATE_VALID;
|
|
|
|
cfg = &cmd->frag_desc_bank_cfg64;
|
|
cfg->info = info;
|
|
cfg->num_banks = 1;
|
|
cfg->desc_size = sizeof(struct htt_msdu_ext_desc_64);
|
|
cfg->bank_base_addrs[0] = __cpu_to_le64(htt->frag_desc.paddr);
|
|
cfg->bank_id[0].bank_min_id = 0;
|
|
cfg->bank_id[0].bank_max_id = __cpu_to_le16(htt->max_num_pending_tx -
|
|
1);
|
|
|
|
cfg->q_state.paddr = cpu_to_le32(htt->tx_q_state.paddr);
|
|
cfg->q_state.num_peers = cpu_to_le16(htt->tx_q_state.num_peers);
|
|
cfg->q_state.num_tids = cpu_to_le16(htt->tx_q_state.num_tids);
|
|
cfg->q_state.record_size = HTT_TX_Q_STATE_ENTRY_SIZE;
|
|
cfg->q_state.record_multiplier = HTT_TX_Q_STATE_ENTRY_MULTIPLIER;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt frag desc bank cmd\n");
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
ath10k_warn(ar, "failed to send frag desc bank cfg request: %d\n",
|
|
ret);
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ath10k_htt_fill_rx_desc_offset_32(void *rx_ring)
|
|
{
|
|
struct htt_rx_ring_setup_ring32 *ring =
|
|
(struct htt_rx_ring_setup_ring32 *)rx_ring;
|
|
|
|
#define desc_offset(x) (offsetof(struct htt_rx_desc, x) / 4)
|
|
ring->mac80211_hdr_offset = __cpu_to_le16(desc_offset(rx_hdr_status));
|
|
ring->msdu_payload_offset = __cpu_to_le16(desc_offset(msdu_payload));
|
|
ring->ppdu_start_offset = __cpu_to_le16(desc_offset(ppdu_start));
|
|
ring->ppdu_end_offset = __cpu_to_le16(desc_offset(ppdu_end));
|
|
ring->mpdu_start_offset = __cpu_to_le16(desc_offset(mpdu_start));
|
|
ring->mpdu_end_offset = __cpu_to_le16(desc_offset(mpdu_end));
|
|
ring->msdu_start_offset = __cpu_to_le16(desc_offset(msdu_start));
|
|
ring->msdu_end_offset = __cpu_to_le16(desc_offset(msdu_end));
|
|
ring->rx_attention_offset = __cpu_to_le16(desc_offset(attention));
|
|
ring->frag_info_offset = __cpu_to_le16(desc_offset(frag_info));
|
|
#undef desc_offset
|
|
}
|
|
|
|
static void ath10k_htt_fill_rx_desc_offset_64(void *rx_ring)
|
|
{
|
|
struct htt_rx_ring_setup_ring64 *ring =
|
|
(struct htt_rx_ring_setup_ring64 *)rx_ring;
|
|
|
|
#define desc_offset(x) (offsetof(struct htt_rx_desc, x) / 4)
|
|
ring->mac80211_hdr_offset = __cpu_to_le16(desc_offset(rx_hdr_status));
|
|
ring->msdu_payload_offset = __cpu_to_le16(desc_offset(msdu_payload));
|
|
ring->ppdu_start_offset = __cpu_to_le16(desc_offset(ppdu_start));
|
|
ring->ppdu_end_offset = __cpu_to_le16(desc_offset(ppdu_end));
|
|
ring->mpdu_start_offset = __cpu_to_le16(desc_offset(mpdu_start));
|
|
ring->mpdu_end_offset = __cpu_to_le16(desc_offset(mpdu_end));
|
|
ring->msdu_start_offset = __cpu_to_le16(desc_offset(msdu_start));
|
|
ring->msdu_end_offset = __cpu_to_le16(desc_offset(msdu_end));
|
|
ring->rx_attention_offset = __cpu_to_le16(desc_offset(attention));
|
|
ring->frag_info_offset = __cpu_to_le16(desc_offset(frag_info));
|
|
#undef desc_offset
|
|
}
|
|
|
|
static int ath10k_htt_send_rx_ring_cfg_32(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
struct htt_rx_ring_setup_ring32 *ring;
|
|
const int num_rx_ring = 1;
|
|
u16 flags;
|
|
u32 fw_idx;
|
|
int len;
|
|
int ret;
|
|
|
|
/*
|
|
* the HW expects the buffer to be an integral number of 4-byte
|
|
* "words"
|
|
*/
|
|
BUILD_BUG_ON(!IS_ALIGNED(HTT_RX_BUF_SIZE, 4));
|
|
BUILD_BUG_ON((HTT_RX_BUF_SIZE & HTT_MAX_CACHE_LINE_SIZE_MASK) != 0);
|
|
|
|
len = sizeof(cmd->hdr) + sizeof(cmd->rx_setup_32.hdr)
|
|
+ (sizeof(*ring) * num_rx_ring);
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
ring = &cmd->rx_setup_32.rings[0];
|
|
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_RX_RING_CFG;
|
|
cmd->rx_setup_32.hdr.num_rings = 1;
|
|
|
|
/* FIXME: do we need all of this? */
|
|
flags = 0;
|
|
flags |= HTT_RX_RING_FLAGS_MAC80211_HDR;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_PAYLOAD;
|
|
flags |= HTT_RX_RING_FLAGS_PPDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_PPDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_MPDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_MPDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_RX_ATTENTION;
|
|
flags |= HTT_RX_RING_FLAGS_FRAG_INFO;
|
|
flags |= HTT_RX_RING_FLAGS_UNICAST_RX;
|
|
flags |= HTT_RX_RING_FLAGS_MULTICAST_RX;
|
|
flags |= HTT_RX_RING_FLAGS_CTRL_RX;
|
|
flags |= HTT_RX_RING_FLAGS_MGMT_RX;
|
|
flags |= HTT_RX_RING_FLAGS_NULL_RX;
|
|
flags |= HTT_RX_RING_FLAGS_PHY_DATA_RX;
|
|
|
|
fw_idx = __le32_to_cpu(*htt->rx_ring.alloc_idx.vaddr);
|
|
|
|
ring->fw_idx_shadow_reg_paddr =
|
|
__cpu_to_le32(htt->rx_ring.alloc_idx.paddr);
|
|
ring->rx_ring_base_paddr = __cpu_to_le32(htt->rx_ring.base_paddr);
|
|
ring->rx_ring_len = __cpu_to_le16(htt->rx_ring.size);
|
|
ring->rx_ring_bufsize = __cpu_to_le16(HTT_RX_BUF_SIZE);
|
|
ring->flags = __cpu_to_le16(flags);
|
|
ring->fw_idx_init_val = __cpu_to_le16(fw_idx);
|
|
|
|
ath10k_htt_fill_rx_desc_offset_32(ring);
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_send_rx_ring_cfg_64(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
struct htt_rx_ring_setup_ring64 *ring;
|
|
const int num_rx_ring = 1;
|
|
u16 flags;
|
|
u32 fw_idx;
|
|
int len;
|
|
int ret;
|
|
|
|
/* HW expects the buffer to be an integral number of 4-byte
|
|
* "words"
|
|
*/
|
|
BUILD_BUG_ON(!IS_ALIGNED(HTT_RX_BUF_SIZE, 4));
|
|
BUILD_BUG_ON((HTT_RX_BUF_SIZE & HTT_MAX_CACHE_LINE_SIZE_MASK) != 0);
|
|
|
|
len = sizeof(cmd->hdr) + sizeof(cmd->rx_setup_64.hdr)
|
|
+ (sizeof(*ring) * num_rx_ring);
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
ring = &cmd->rx_setup_64.rings[0];
|
|
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_RX_RING_CFG;
|
|
cmd->rx_setup_64.hdr.num_rings = 1;
|
|
|
|
flags = 0;
|
|
flags |= HTT_RX_RING_FLAGS_MAC80211_HDR;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_PAYLOAD;
|
|
flags |= HTT_RX_RING_FLAGS_PPDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_PPDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_MPDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_MPDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_START;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_END;
|
|
flags |= HTT_RX_RING_FLAGS_RX_ATTENTION;
|
|
flags |= HTT_RX_RING_FLAGS_FRAG_INFO;
|
|
flags |= HTT_RX_RING_FLAGS_UNICAST_RX;
|
|
flags |= HTT_RX_RING_FLAGS_MULTICAST_RX;
|
|
flags |= HTT_RX_RING_FLAGS_CTRL_RX;
|
|
flags |= HTT_RX_RING_FLAGS_MGMT_RX;
|
|
flags |= HTT_RX_RING_FLAGS_NULL_RX;
|
|
flags |= HTT_RX_RING_FLAGS_PHY_DATA_RX;
|
|
|
|
fw_idx = __le32_to_cpu(*htt->rx_ring.alloc_idx.vaddr);
|
|
|
|
ring->fw_idx_shadow_reg_paddr = __cpu_to_le64(htt->rx_ring.alloc_idx.paddr);
|
|
ring->rx_ring_base_paddr = __cpu_to_le64(htt->rx_ring.base_paddr);
|
|
ring->rx_ring_len = __cpu_to_le16(htt->rx_ring.size);
|
|
ring->rx_ring_bufsize = __cpu_to_le16(HTT_RX_BUF_SIZE);
|
|
ring->flags = __cpu_to_le16(flags);
|
|
ring->fw_idx_init_val = __cpu_to_le16(fw_idx);
|
|
|
|
ath10k_htt_fill_rx_desc_offset_64(ring);
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_send_rx_ring_cfg_hl(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
struct htt_rx_ring_setup_ring32 *ring;
|
|
const int num_rx_ring = 1;
|
|
u16 flags;
|
|
int len;
|
|
int ret;
|
|
|
|
/*
|
|
* the HW expects the buffer to be an integral number of 4-byte
|
|
* "words"
|
|
*/
|
|
BUILD_BUG_ON(!IS_ALIGNED(HTT_RX_BUF_SIZE, 4));
|
|
BUILD_BUG_ON((HTT_RX_BUF_SIZE & HTT_MAX_CACHE_LINE_SIZE_MASK) != 0);
|
|
|
|
len = sizeof(cmd->hdr) + sizeof(cmd->rx_setup_32.hdr)
|
|
+ (sizeof(*ring) * num_rx_ring);
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
ring = &cmd->rx_setup_32.rings[0];
|
|
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_RX_RING_CFG;
|
|
cmd->rx_setup_32.hdr.num_rings = 1;
|
|
|
|
flags = 0;
|
|
flags |= HTT_RX_RING_FLAGS_MSDU_PAYLOAD;
|
|
flags |= HTT_RX_RING_FLAGS_UNICAST_RX;
|
|
flags |= HTT_RX_RING_FLAGS_MULTICAST_RX;
|
|
|
|
memset(ring, 0, sizeof(*ring));
|
|
ring->rx_ring_len = __cpu_to_le16(HTT_RX_RING_SIZE_MIN);
|
|
ring->rx_ring_bufsize = __cpu_to_le16(HTT_RX_BUF_SIZE);
|
|
ring->flags = __cpu_to_le16(flags);
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_h2t_aggr_cfg_msg_32(struct ath10k_htt *htt,
|
|
u8 max_subfrms_ampdu,
|
|
u8 max_subfrms_amsdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct htt_aggr_conf *aggr_conf;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
int len;
|
|
int ret;
|
|
|
|
/* Firmware defaults are: amsdu = 3 and ampdu = 64 */
|
|
|
|
if (max_subfrms_ampdu == 0 || max_subfrms_ampdu > 64)
|
|
return -EINVAL;
|
|
|
|
if (max_subfrms_amsdu == 0 || max_subfrms_amsdu > 31)
|
|
return -EINVAL;
|
|
|
|
len = sizeof(cmd->hdr);
|
|
len += sizeof(cmd->aggr_conf);
|
|
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_AGGR_CFG;
|
|
|
|
aggr_conf = &cmd->aggr_conf;
|
|
aggr_conf->max_num_ampdu_subframes = max_subfrms_ampdu;
|
|
aggr_conf->max_num_amsdu_subframes = max_subfrms_amsdu;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt h2t aggr cfg msg amsdu %d ampdu %d",
|
|
aggr_conf->max_num_amsdu_subframes,
|
|
aggr_conf->max_num_ampdu_subframes);
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ath10k_htt_h2t_aggr_cfg_msg_v2(struct ath10k_htt *htt,
|
|
u8 max_subfrms_ampdu,
|
|
u8 max_subfrms_amsdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct htt_aggr_conf_v2 *aggr_conf;
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
int len;
|
|
int ret;
|
|
|
|
/* Firmware defaults are: amsdu = 3 and ampdu = 64 */
|
|
|
|
if (max_subfrms_ampdu == 0 || max_subfrms_ampdu > 64)
|
|
return -EINVAL;
|
|
|
|
if (max_subfrms_amsdu == 0 || max_subfrms_amsdu > 31)
|
|
return -EINVAL;
|
|
|
|
len = sizeof(cmd->hdr);
|
|
len += sizeof(cmd->aggr_conf_v2);
|
|
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_AGGR_CFG;
|
|
|
|
aggr_conf = &cmd->aggr_conf_v2;
|
|
aggr_conf->max_num_ampdu_subframes = max_subfrms_ampdu;
|
|
aggr_conf->max_num_amsdu_subframes = max_subfrms_amsdu;
|
|
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT, "htt h2t aggr cfg msg amsdu %d ampdu %d",
|
|
aggr_conf->max_num_amsdu_subframes,
|
|
aggr_conf->max_num_ampdu_subframes);
|
|
|
|
ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
|
|
if (ret) {
|
|
dev_kfree_skb_any(skb);
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ath10k_htt_tx_fetch_resp(struct ath10k *ar,
|
|
__le32 token,
|
|
__le16 fetch_seq_num,
|
|
struct htt_tx_fetch_record *records,
|
|
size_t num_records)
|
|
{
|
|
struct sk_buff *skb;
|
|
struct htt_cmd *cmd;
|
|
const u16 resp_id = 0;
|
|
int len = 0;
|
|
int ret;
|
|
|
|
/* Response IDs are echo-ed back only for host driver convienence
|
|
* purposes. They aren't used for anything in the driver yet so use 0.
|
|
*/
|
|
|
|
len += sizeof(cmd->hdr);
|
|
len += sizeof(cmd->tx_fetch_resp);
|
|
len += sizeof(cmd->tx_fetch_resp.records[0]) * num_records;
|
|
|
|
skb = ath10k_htc_alloc_skb(ar, len);
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
|
|
skb_put(skb, len);
|
|
cmd = (struct htt_cmd *)skb->data;
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FETCH_RESP;
|
|
cmd->tx_fetch_resp.resp_id = cpu_to_le16(resp_id);
|
|
cmd->tx_fetch_resp.fetch_seq_num = fetch_seq_num;
|
|
cmd->tx_fetch_resp.num_records = cpu_to_le16(num_records);
|
|
cmd->tx_fetch_resp.token = token;
|
|
|
|
memcpy(cmd->tx_fetch_resp.records, records,
|
|
sizeof(records[0]) * num_records);
|
|
|
|
ret = ath10k_htc_send(&ar->htc, ar->htt.eid, skb);
|
|
if (ret) {
|
|
ath10k_warn(ar, "failed to submit htc command: %d\n", ret);
|
|
goto err_free_skb;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_free_skb:
|
|
dev_kfree_skb_any(skb);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static u8 ath10k_htt_tx_get_vdev_id(struct ath10k *ar, struct sk_buff *skb)
|
|
{
|
|
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
|
|
struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
|
|
struct ath10k_vif *arvif;
|
|
|
|
if (info->flags & IEEE80211_TX_CTL_TX_OFFCHAN) {
|
|
return ar->scan.vdev_id;
|
|
} else if (cb->vif) {
|
|
arvif = (void *)cb->vif->drv_priv;
|
|
return arvif->vdev_id;
|
|
} else if (ar->monitor_started) {
|
|
return ar->monitor_vdev_id;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static u8 ath10k_htt_tx_get_tid(struct sk_buff *skb, bool is_eth)
|
|
{
|
|
struct ieee80211_hdr *hdr = (void *)skb->data;
|
|
struct ath10k_skb_cb *cb = ATH10K_SKB_CB(skb);
|
|
|
|
if (!is_eth && ieee80211_is_mgmt(hdr->frame_control))
|
|
return HTT_DATA_TX_EXT_TID_MGMT;
|
|
else if (cb->flags & ATH10K_SKB_F_QOS)
|
|
return skb->priority & IEEE80211_QOS_CTL_TID_MASK;
|
|
else
|
|
return HTT_DATA_TX_EXT_TID_NON_QOS_MCAST_BCAST;
|
|
}
|
|
|
|
int ath10k_htt_mgmt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct device *dev = ar->dev;
|
|
struct sk_buff *txdesc = NULL;
|
|
struct htt_cmd *cmd;
|
|
struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
|
|
u8 vdev_id = ath10k_htt_tx_get_vdev_id(ar, msdu);
|
|
int len = 0;
|
|
int msdu_id = -1;
|
|
int res;
|
|
const u8 *peer_addr;
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data;
|
|
|
|
len += sizeof(cmd->hdr);
|
|
len += sizeof(cmd->mgmt_tx);
|
|
|
|
res = ath10k_htt_tx_alloc_msdu_id(htt, msdu);
|
|
if (res < 0)
|
|
goto err;
|
|
|
|
msdu_id = res;
|
|
|
|
if ((ieee80211_is_action(hdr->frame_control) ||
|
|
ieee80211_is_deauth(hdr->frame_control) ||
|
|
ieee80211_is_disassoc(hdr->frame_control)) &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
peer_addr = hdr->addr1;
|
|
if (is_multicast_ether_addr(peer_addr)) {
|
|
skb_put(msdu, sizeof(struct ieee80211_mmie_16));
|
|
} else {
|
|
if (skb_cb->ucast_cipher == WLAN_CIPHER_SUITE_GCMP ||
|
|
skb_cb->ucast_cipher == WLAN_CIPHER_SUITE_GCMP_256)
|
|
skb_put(msdu, IEEE80211_GCMP_MIC_LEN);
|
|
else
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
}
|
|
}
|
|
|
|
txdesc = ath10k_htc_alloc_skb(ar, len);
|
|
if (!txdesc) {
|
|
res = -ENOMEM;
|
|
goto err_free_msdu_id;
|
|
}
|
|
|
|
skb_cb->paddr = dma_map_single(dev, msdu->data, msdu->len,
|
|
DMA_TO_DEVICE);
|
|
res = dma_mapping_error(dev, skb_cb->paddr);
|
|
if (res) {
|
|
res = -EIO;
|
|
goto err_free_txdesc;
|
|
}
|
|
|
|
skb_put(txdesc, len);
|
|
cmd = (struct htt_cmd *)txdesc->data;
|
|
memset(cmd, 0, len);
|
|
|
|
cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_MGMT_TX;
|
|
cmd->mgmt_tx.msdu_paddr = __cpu_to_le32(ATH10K_SKB_CB(msdu)->paddr);
|
|
cmd->mgmt_tx.len = __cpu_to_le32(msdu->len);
|
|
cmd->mgmt_tx.desc_id = __cpu_to_le32(msdu_id);
|
|
cmd->mgmt_tx.vdev_id = __cpu_to_le32(vdev_id);
|
|
memcpy(cmd->mgmt_tx.hdr, msdu->data,
|
|
min_t(int, msdu->len, HTT_MGMT_FRM_HDR_DOWNLOAD_LEN));
|
|
|
|
res = ath10k_htc_send(&htt->ar->htc, htt->eid, txdesc);
|
|
if (res)
|
|
goto err_unmap_msdu;
|
|
|
|
return 0;
|
|
|
|
err_unmap_msdu:
|
|
if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
|
|
dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE);
|
|
err_free_txdesc:
|
|
dev_kfree_skb_any(txdesc);
|
|
err_free_msdu_id:
|
|
spin_lock_bh(&htt->tx_lock);
|
|
ath10k_htt_tx_free_msdu_id(htt, msdu_id);
|
|
spin_unlock_bh(&htt->tx_lock);
|
|
err:
|
|
return res;
|
|
}
|
|
|
|
#define HTT_TX_HL_NEEDED_HEADROOM \
|
|
(unsigned int)(sizeof(struct htt_cmd_hdr) + \
|
|
sizeof(struct htt_data_tx_desc) + \
|
|
sizeof(struct ath10k_htc_hdr))
|
|
|
|
static int ath10k_htt_tx_hl(struct ath10k_htt *htt, enum ath10k_hw_txrx_mode txmode,
|
|
struct sk_buff *msdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
int res, data_len;
|
|
struct htt_cmd_hdr *cmd_hdr;
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data;
|
|
struct htt_data_tx_desc *tx_desc;
|
|
struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
|
|
struct sk_buff *tmp_skb;
|
|
bool is_eth = (txmode == ATH10K_HW_TXRX_ETHERNET);
|
|
u8 vdev_id = ath10k_htt_tx_get_vdev_id(ar, msdu);
|
|
u8 tid = ath10k_htt_tx_get_tid(msdu, is_eth);
|
|
u8 flags0 = 0;
|
|
u16 flags1 = 0;
|
|
u16 msdu_id = 0;
|
|
|
|
if ((ieee80211_is_action(hdr->frame_control) ||
|
|
ieee80211_is_deauth(hdr->frame_control) ||
|
|
ieee80211_is_disassoc(hdr->frame_control)) &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
}
|
|
|
|
data_len = msdu->len;
|
|
|
|
switch (txmode) {
|
|
case ATH10K_HW_TXRX_RAW:
|
|
case ATH10K_HW_TXRX_NATIVE_WIFI:
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
/* fall through */
|
|
case ATH10K_HW_TXRX_ETHERNET:
|
|
flags0 |= SM(txmode, HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
break;
|
|
case ATH10K_HW_TXRX_MGMT:
|
|
flags0 |= SM(ATH10K_HW_TXRX_MGMT,
|
|
HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
|
|
if (htt->disable_tx_comp)
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_TX_COMPLETE;
|
|
break;
|
|
}
|
|
|
|
if (skb_cb->flags & ATH10K_SKB_F_NO_HWCRYPT)
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_NO_ENCRYPT;
|
|
|
|
flags1 |= SM((u16)vdev_id, HTT_DATA_TX_DESC_FLAGS1_VDEV_ID);
|
|
flags1 |= SM((u16)tid, HTT_DATA_TX_DESC_FLAGS1_EXT_TID);
|
|
if (msdu->ip_summed == CHECKSUM_PARTIAL &&
|
|
!test_bit(ATH10K_FLAG_RAW_MODE, &ar->dev_flags)) {
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L3_OFFLOAD;
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L4_OFFLOAD;
|
|
}
|
|
|
|
/* Prepend the HTT header and TX desc struct to the data message
|
|
* and realloc the skb if it does not have enough headroom.
|
|
*/
|
|
if (skb_headroom(msdu) < HTT_TX_HL_NEEDED_HEADROOM) {
|
|
tmp_skb = msdu;
|
|
|
|
ath10k_dbg(htt->ar, ATH10K_DBG_HTT,
|
|
"Not enough headroom in skb. Current headroom: %u, needed: %u. Reallocating...\n",
|
|
skb_headroom(msdu), HTT_TX_HL_NEEDED_HEADROOM);
|
|
msdu = skb_realloc_headroom(msdu, HTT_TX_HL_NEEDED_HEADROOM);
|
|
kfree_skb(tmp_skb);
|
|
if (!msdu) {
|
|
ath10k_warn(htt->ar, "htt hl tx: Unable to realloc skb!\n");
|
|
res = -ENOMEM;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
if (ar->bus_param.hl_msdu_ids) {
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_POSTPONED;
|
|
res = ath10k_htt_tx_alloc_msdu_id(htt, msdu);
|
|
if (res < 0) {
|
|
ath10k_err(ar, "msdu_id allocation failed %d\n", res);
|
|
goto out;
|
|
}
|
|
msdu_id = res;
|
|
}
|
|
|
|
/* As msdu is freed by mac80211 (in ieee80211_tx_status()) and by
|
|
* ath10k (in ath10k_htt_htc_tx_complete()) we have to increase
|
|
* reference by one to avoid a use-after-free case and a double
|
|
* free.
|
|
*/
|
|
skb_get(msdu);
|
|
|
|
skb_push(msdu, sizeof(*cmd_hdr));
|
|
skb_push(msdu, sizeof(*tx_desc));
|
|
cmd_hdr = (struct htt_cmd_hdr *)msdu->data;
|
|
tx_desc = (struct htt_data_tx_desc *)(msdu->data + sizeof(*cmd_hdr));
|
|
|
|
cmd_hdr->msg_type = HTT_H2T_MSG_TYPE_TX_FRM;
|
|
tx_desc->flags0 = flags0;
|
|
tx_desc->flags1 = __cpu_to_le16(flags1);
|
|
tx_desc->len = __cpu_to_le16(data_len);
|
|
tx_desc->id = __cpu_to_le16(msdu_id);
|
|
tx_desc->frags_paddr = 0; /* always zero */
|
|
/* Initialize peer_id to INVALID_PEER because this is NOT
|
|
* Reinjection path
|
|
*/
|
|
tx_desc->peerid = __cpu_to_le32(HTT_INVALID_PEERID);
|
|
|
|
res = ath10k_htc_send_hl(&htt->ar->htc, htt->eid, msdu);
|
|
|
|
out:
|
|
return res;
|
|
}
|
|
|
|
static int ath10k_htt_tx_32(struct ath10k_htt *htt,
|
|
enum ath10k_hw_txrx_mode txmode,
|
|
struct sk_buff *msdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct device *dev = ar->dev;
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data;
|
|
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(msdu);
|
|
struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
|
|
struct ath10k_hif_sg_item sg_items[2];
|
|
struct ath10k_htt_txbuf_32 *txbuf;
|
|
struct htt_data_tx_desc_frag *frags;
|
|
bool is_eth = (txmode == ATH10K_HW_TXRX_ETHERNET);
|
|
u8 vdev_id = ath10k_htt_tx_get_vdev_id(ar, msdu);
|
|
u8 tid = ath10k_htt_tx_get_tid(msdu, is_eth);
|
|
int prefetch_len;
|
|
int res;
|
|
u8 flags0 = 0;
|
|
u16 msdu_id, flags1 = 0;
|
|
u16 freq = 0;
|
|
u32 frags_paddr = 0;
|
|
u32 txbuf_paddr;
|
|
struct htt_msdu_ext_desc *ext_desc = NULL;
|
|
struct htt_msdu_ext_desc *ext_desc_t = NULL;
|
|
|
|
res = ath10k_htt_tx_alloc_msdu_id(htt, msdu);
|
|
if (res < 0)
|
|
goto err;
|
|
|
|
msdu_id = res;
|
|
|
|
prefetch_len = min(htt->prefetch_len, msdu->len);
|
|
prefetch_len = roundup(prefetch_len, 4);
|
|
|
|
txbuf = htt->txbuf.vaddr_txbuff_32 + msdu_id;
|
|
txbuf_paddr = htt->txbuf.paddr +
|
|
(sizeof(struct ath10k_htt_txbuf_32) * msdu_id);
|
|
|
|
if ((ieee80211_is_action(hdr->frame_control) ||
|
|
ieee80211_is_deauth(hdr->frame_control) ||
|
|
ieee80211_is_disassoc(hdr->frame_control)) &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
} else if (!(skb_cb->flags & ATH10K_SKB_F_NO_HWCRYPT) &&
|
|
txmode == ATH10K_HW_TXRX_RAW &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
}
|
|
|
|
skb_cb->paddr = dma_map_single(dev, msdu->data, msdu->len,
|
|
DMA_TO_DEVICE);
|
|
res = dma_mapping_error(dev, skb_cb->paddr);
|
|
if (res) {
|
|
res = -EIO;
|
|
goto err_free_msdu_id;
|
|
}
|
|
|
|
if (unlikely(info->flags & IEEE80211_TX_CTL_TX_OFFCHAN))
|
|
freq = ar->scan.roc_freq;
|
|
|
|
switch (txmode) {
|
|
case ATH10K_HW_TXRX_RAW:
|
|
case ATH10K_HW_TXRX_NATIVE_WIFI:
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
/* fall through */
|
|
case ATH10K_HW_TXRX_ETHERNET:
|
|
if (ar->hw_params.continuous_frag_desc) {
|
|
ext_desc_t = htt->frag_desc.vaddr_desc_32;
|
|
memset(&ext_desc_t[msdu_id], 0,
|
|
sizeof(struct htt_msdu_ext_desc));
|
|
frags = (struct htt_data_tx_desc_frag *)
|
|
&ext_desc_t[msdu_id].frags;
|
|
ext_desc = &ext_desc_t[msdu_id];
|
|
frags[0].tword_addr.paddr_lo =
|
|
__cpu_to_le32(skb_cb->paddr);
|
|
frags[0].tword_addr.paddr_hi = 0;
|
|
frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len);
|
|
|
|
frags_paddr = htt->frag_desc.paddr +
|
|
(sizeof(struct htt_msdu_ext_desc) * msdu_id);
|
|
} else {
|
|
frags = txbuf->frags;
|
|
frags[0].dword_addr.paddr =
|
|
__cpu_to_le32(skb_cb->paddr);
|
|
frags[0].dword_addr.len = __cpu_to_le32(msdu->len);
|
|
frags[1].dword_addr.paddr = 0;
|
|
frags[1].dword_addr.len = 0;
|
|
|
|
frags_paddr = txbuf_paddr;
|
|
}
|
|
flags0 |= SM(txmode, HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
break;
|
|
case ATH10K_HW_TXRX_MGMT:
|
|
flags0 |= SM(ATH10K_HW_TXRX_MGMT,
|
|
HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
|
|
frags_paddr = skb_cb->paddr;
|
|
break;
|
|
}
|
|
|
|
/* Normally all commands go through HTC which manages tx credits for
|
|
* each endpoint and notifies when tx is completed.
|
|
*
|
|
* HTT endpoint is creditless so there's no need to care about HTC
|
|
* flags. In that case it is trivial to fill the HTC header here.
|
|
*
|
|
* MSDU transmission is considered completed upon HTT event. This
|
|
* implies no relevant resources can be freed until after the event is
|
|
* received. That's why HTC tx completion handler itself is ignored by
|
|
* setting NULL to transfer_context for all sg items.
|
|
*
|
|
* There is simply no point in pushing HTT TX_FRM through HTC tx path
|
|
* as it's a waste of resources. By bypassing HTC it is possible to
|
|
* avoid extra memory allocations, compress data structures and thus
|
|
* improve performance.
|
|
*/
|
|
|
|
txbuf->htc_hdr.eid = htt->eid;
|
|
txbuf->htc_hdr.len = __cpu_to_le16(sizeof(txbuf->cmd_hdr) +
|
|
sizeof(txbuf->cmd_tx) +
|
|
prefetch_len);
|
|
txbuf->htc_hdr.flags = 0;
|
|
|
|
if (skb_cb->flags & ATH10K_SKB_F_NO_HWCRYPT)
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_NO_ENCRYPT;
|
|
|
|
flags1 |= SM((u16)vdev_id, HTT_DATA_TX_DESC_FLAGS1_VDEV_ID);
|
|
flags1 |= SM((u16)tid, HTT_DATA_TX_DESC_FLAGS1_EXT_TID);
|
|
if (msdu->ip_summed == CHECKSUM_PARTIAL &&
|
|
!test_bit(ATH10K_FLAG_RAW_MODE, &ar->dev_flags)) {
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L3_OFFLOAD;
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L4_OFFLOAD;
|
|
if (ar->hw_params.continuous_frag_desc)
|
|
ext_desc->flags |= HTT_MSDU_CHECKSUM_ENABLE;
|
|
}
|
|
|
|
/* Prevent firmware from sending up tx inspection requests. There's
|
|
* nothing ath10k can do with frames requested for inspection so force
|
|
* it to simply rely a regular tx completion with discard status.
|
|
*/
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_POSTPONED;
|
|
|
|
txbuf->cmd_hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FRM;
|
|
txbuf->cmd_tx.flags0 = flags0;
|
|
txbuf->cmd_tx.flags1 = __cpu_to_le16(flags1);
|
|
txbuf->cmd_tx.len = __cpu_to_le16(msdu->len);
|
|
txbuf->cmd_tx.id = __cpu_to_le16(msdu_id);
|
|
txbuf->cmd_tx.frags_paddr = __cpu_to_le32(frags_paddr);
|
|
if (ath10k_mac_tx_frm_has_freq(ar)) {
|
|
txbuf->cmd_tx.offchan_tx.peerid =
|
|
__cpu_to_le16(HTT_INVALID_PEERID);
|
|
txbuf->cmd_tx.offchan_tx.freq =
|
|
__cpu_to_le16(freq);
|
|
} else {
|
|
txbuf->cmd_tx.peerid =
|
|
__cpu_to_le32(HTT_INVALID_PEERID);
|
|
}
|
|
|
|
trace_ath10k_htt_tx(ar, msdu_id, msdu->len, vdev_id, tid);
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT,
|
|
"htt tx flags0 %hhu flags1 %hu len %d id %hu frags_paddr %pad, msdu_paddr %pad vdev %hhu tid %hhu freq %hu\n",
|
|
flags0, flags1, msdu->len, msdu_id, &frags_paddr,
|
|
&skb_cb->paddr, vdev_id, tid, freq);
|
|
ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt tx msdu: ",
|
|
msdu->data, msdu->len);
|
|
trace_ath10k_tx_hdr(ar, msdu->data, msdu->len);
|
|
trace_ath10k_tx_payload(ar, msdu->data, msdu->len);
|
|
|
|
sg_items[0].transfer_id = 0;
|
|
sg_items[0].transfer_context = NULL;
|
|
sg_items[0].vaddr = &txbuf->htc_hdr;
|
|
sg_items[0].paddr = txbuf_paddr +
|
|
sizeof(txbuf->frags);
|
|
sg_items[0].len = sizeof(txbuf->htc_hdr) +
|
|
sizeof(txbuf->cmd_hdr) +
|
|
sizeof(txbuf->cmd_tx);
|
|
|
|
sg_items[1].transfer_id = 0;
|
|
sg_items[1].transfer_context = NULL;
|
|
sg_items[1].vaddr = msdu->data;
|
|
sg_items[1].paddr = skb_cb->paddr;
|
|
sg_items[1].len = prefetch_len;
|
|
|
|
res = ath10k_hif_tx_sg(htt->ar,
|
|
htt->ar->htc.endpoint[htt->eid].ul_pipe_id,
|
|
sg_items, ARRAY_SIZE(sg_items));
|
|
if (res)
|
|
goto err_unmap_msdu;
|
|
|
|
return 0;
|
|
|
|
err_unmap_msdu:
|
|
dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE);
|
|
err_free_msdu_id:
|
|
ath10k_htt_tx_free_msdu_id(htt, msdu_id);
|
|
err:
|
|
return res;
|
|
}
|
|
|
|
static int ath10k_htt_tx_64(struct ath10k_htt *htt,
|
|
enum ath10k_hw_txrx_mode txmode,
|
|
struct sk_buff *msdu)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
struct device *dev = ar->dev;
|
|
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data;
|
|
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(msdu);
|
|
struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
|
|
struct ath10k_hif_sg_item sg_items[2];
|
|
struct ath10k_htt_txbuf_64 *txbuf;
|
|
struct htt_data_tx_desc_frag *frags;
|
|
bool is_eth = (txmode == ATH10K_HW_TXRX_ETHERNET);
|
|
u8 vdev_id = ath10k_htt_tx_get_vdev_id(ar, msdu);
|
|
u8 tid = ath10k_htt_tx_get_tid(msdu, is_eth);
|
|
int prefetch_len;
|
|
int res;
|
|
u8 flags0 = 0;
|
|
u16 msdu_id, flags1 = 0;
|
|
u16 freq = 0;
|
|
dma_addr_t frags_paddr = 0;
|
|
dma_addr_t txbuf_paddr;
|
|
struct htt_msdu_ext_desc_64 *ext_desc = NULL;
|
|
struct htt_msdu_ext_desc_64 *ext_desc_t = NULL;
|
|
|
|
res = ath10k_htt_tx_alloc_msdu_id(htt, msdu);
|
|
if (res < 0)
|
|
goto err;
|
|
|
|
msdu_id = res;
|
|
|
|
prefetch_len = min(htt->prefetch_len, msdu->len);
|
|
prefetch_len = roundup(prefetch_len, 4);
|
|
|
|
txbuf = htt->txbuf.vaddr_txbuff_64 + msdu_id;
|
|
txbuf_paddr = htt->txbuf.paddr +
|
|
(sizeof(struct ath10k_htt_txbuf_64) * msdu_id);
|
|
|
|
if ((ieee80211_is_action(hdr->frame_control) ||
|
|
ieee80211_is_deauth(hdr->frame_control) ||
|
|
ieee80211_is_disassoc(hdr->frame_control)) &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
} else if (!(skb_cb->flags & ATH10K_SKB_F_NO_HWCRYPT) &&
|
|
txmode == ATH10K_HW_TXRX_RAW &&
|
|
ieee80211_has_protected(hdr->frame_control)) {
|
|
skb_put(msdu, IEEE80211_CCMP_MIC_LEN);
|
|
}
|
|
|
|
skb_cb->paddr = dma_map_single(dev, msdu->data, msdu->len,
|
|
DMA_TO_DEVICE);
|
|
res = dma_mapping_error(dev, skb_cb->paddr);
|
|
if (res) {
|
|
res = -EIO;
|
|
goto err_free_msdu_id;
|
|
}
|
|
|
|
if (unlikely(info->flags & IEEE80211_TX_CTL_TX_OFFCHAN))
|
|
freq = ar->scan.roc_freq;
|
|
|
|
switch (txmode) {
|
|
case ATH10K_HW_TXRX_RAW:
|
|
case ATH10K_HW_TXRX_NATIVE_WIFI:
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
/* fall through */
|
|
case ATH10K_HW_TXRX_ETHERNET:
|
|
if (ar->hw_params.continuous_frag_desc) {
|
|
ext_desc_t = htt->frag_desc.vaddr_desc_64;
|
|
memset(&ext_desc_t[msdu_id], 0,
|
|
sizeof(struct htt_msdu_ext_desc_64));
|
|
frags = (struct htt_data_tx_desc_frag *)
|
|
&ext_desc_t[msdu_id].frags;
|
|
ext_desc = &ext_desc_t[msdu_id];
|
|
frags[0].tword_addr.paddr_lo =
|
|
__cpu_to_le32(skb_cb->paddr);
|
|
frags[0].tword_addr.paddr_hi =
|
|
__cpu_to_le16(upper_32_bits(skb_cb->paddr));
|
|
frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len);
|
|
|
|
frags_paddr = htt->frag_desc.paddr +
|
|
(sizeof(struct htt_msdu_ext_desc_64) * msdu_id);
|
|
} else {
|
|
frags = txbuf->frags;
|
|
frags[0].tword_addr.paddr_lo =
|
|
__cpu_to_le32(skb_cb->paddr);
|
|
frags[0].tword_addr.paddr_hi =
|
|
__cpu_to_le16(upper_32_bits(skb_cb->paddr));
|
|
frags[0].tword_addr.len_16 = __cpu_to_le16(msdu->len);
|
|
frags[1].tword_addr.paddr_lo = 0;
|
|
frags[1].tword_addr.paddr_hi = 0;
|
|
frags[1].tword_addr.len_16 = 0;
|
|
}
|
|
flags0 |= SM(txmode, HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
break;
|
|
case ATH10K_HW_TXRX_MGMT:
|
|
flags0 |= SM(ATH10K_HW_TXRX_MGMT,
|
|
HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
|
|
|
|
frags_paddr = skb_cb->paddr;
|
|
break;
|
|
}
|
|
|
|
/* Normally all commands go through HTC which manages tx credits for
|
|
* each endpoint and notifies when tx is completed.
|
|
*
|
|
* HTT endpoint is creditless so there's no need to care about HTC
|
|
* flags. In that case it is trivial to fill the HTC header here.
|
|
*
|
|
* MSDU transmission is considered completed upon HTT event. This
|
|
* implies no relevant resources can be freed until after the event is
|
|
* received. That's why HTC tx completion handler itself is ignored by
|
|
* setting NULL to transfer_context for all sg items.
|
|
*
|
|
* There is simply no point in pushing HTT TX_FRM through HTC tx path
|
|
* as it's a waste of resources. By bypassing HTC it is possible to
|
|
* avoid extra memory allocations, compress data structures and thus
|
|
* improve performance.
|
|
*/
|
|
|
|
txbuf->htc_hdr.eid = htt->eid;
|
|
txbuf->htc_hdr.len = __cpu_to_le16(sizeof(txbuf->cmd_hdr) +
|
|
sizeof(txbuf->cmd_tx) +
|
|
prefetch_len);
|
|
txbuf->htc_hdr.flags = 0;
|
|
|
|
if (skb_cb->flags & ATH10K_SKB_F_NO_HWCRYPT)
|
|
flags0 |= HTT_DATA_TX_DESC_FLAGS0_NO_ENCRYPT;
|
|
|
|
flags1 |= SM((u16)vdev_id, HTT_DATA_TX_DESC_FLAGS1_VDEV_ID);
|
|
flags1 |= SM((u16)tid, HTT_DATA_TX_DESC_FLAGS1_EXT_TID);
|
|
if (msdu->ip_summed == CHECKSUM_PARTIAL &&
|
|
!test_bit(ATH10K_FLAG_RAW_MODE, &ar->dev_flags)) {
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L3_OFFLOAD;
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L4_OFFLOAD;
|
|
if (ar->hw_params.continuous_frag_desc) {
|
|
memset(ext_desc->tso_flag, 0, sizeof(ext_desc->tso_flag));
|
|
ext_desc->tso_flag[3] |=
|
|
__cpu_to_le32(HTT_MSDU_CHECKSUM_ENABLE_64);
|
|
}
|
|
}
|
|
|
|
/* Prevent firmware from sending up tx inspection requests. There's
|
|
* nothing ath10k can do with frames requested for inspection so force
|
|
* it to simply rely a regular tx completion with discard status.
|
|
*/
|
|
flags1 |= HTT_DATA_TX_DESC_FLAGS1_POSTPONED;
|
|
|
|
txbuf->cmd_hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FRM;
|
|
txbuf->cmd_tx.flags0 = flags0;
|
|
txbuf->cmd_tx.flags1 = __cpu_to_le16(flags1);
|
|
txbuf->cmd_tx.len = __cpu_to_le16(msdu->len);
|
|
txbuf->cmd_tx.id = __cpu_to_le16(msdu_id);
|
|
|
|
/* fill fragment descriptor */
|
|
txbuf->cmd_tx.frags_paddr = __cpu_to_le64(frags_paddr);
|
|
if (ath10k_mac_tx_frm_has_freq(ar)) {
|
|
txbuf->cmd_tx.offchan_tx.peerid =
|
|
__cpu_to_le16(HTT_INVALID_PEERID);
|
|
txbuf->cmd_tx.offchan_tx.freq =
|
|
__cpu_to_le16(freq);
|
|
} else {
|
|
txbuf->cmd_tx.peerid =
|
|
__cpu_to_le32(HTT_INVALID_PEERID);
|
|
}
|
|
|
|
trace_ath10k_htt_tx(ar, msdu_id, msdu->len, vdev_id, tid);
|
|
ath10k_dbg(ar, ATH10K_DBG_HTT,
|
|
"htt tx flags0 %hhu flags1 %hu len %d id %hu frags_paddr %pad, msdu_paddr %pad vdev %hhu tid %hhu freq %hu\n",
|
|
flags0, flags1, msdu->len, msdu_id, &frags_paddr,
|
|
&skb_cb->paddr, vdev_id, tid, freq);
|
|
ath10k_dbg_dump(ar, ATH10K_DBG_HTT_DUMP, NULL, "htt tx msdu: ",
|
|
msdu->data, msdu->len);
|
|
trace_ath10k_tx_hdr(ar, msdu->data, msdu->len);
|
|
trace_ath10k_tx_payload(ar, msdu->data, msdu->len);
|
|
|
|
sg_items[0].transfer_id = 0;
|
|
sg_items[0].transfer_context = NULL;
|
|
sg_items[0].vaddr = &txbuf->htc_hdr;
|
|
sg_items[0].paddr = txbuf_paddr +
|
|
sizeof(txbuf->frags);
|
|
sg_items[0].len = sizeof(txbuf->htc_hdr) +
|
|
sizeof(txbuf->cmd_hdr) +
|
|
sizeof(txbuf->cmd_tx);
|
|
|
|
sg_items[1].transfer_id = 0;
|
|
sg_items[1].transfer_context = NULL;
|
|
sg_items[1].vaddr = msdu->data;
|
|
sg_items[1].paddr = skb_cb->paddr;
|
|
sg_items[1].len = prefetch_len;
|
|
|
|
res = ath10k_hif_tx_sg(htt->ar,
|
|
htt->ar->htc.endpoint[htt->eid].ul_pipe_id,
|
|
sg_items, ARRAY_SIZE(sg_items));
|
|
if (res)
|
|
goto err_unmap_msdu;
|
|
|
|
return 0;
|
|
|
|
err_unmap_msdu:
|
|
dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE);
|
|
err_free_msdu_id:
|
|
ath10k_htt_tx_free_msdu_id(htt, msdu_id);
|
|
err:
|
|
return res;
|
|
}
|
|
|
|
static const struct ath10k_htt_tx_ops htt_tx_ops_32 = {
|
|
.htt_send_rx_ring_cfg = ath10k_htt_send_rx_ring_cfg_32,
|
|
.htt_send_frag_desc_bank_cfg = ath10k_htt_send_frag_desc_bank_cfg_32,
|
|
.htt_alloc_frag_desc = ath10k_htt_tx_alloc_cont_frag_desc_32,
|
|
.htt_free_frag_desc = ath10k_htt_tx_free_cont_frag_desc_32,
|
|
.htt_tx = ath10k_htt_tx_32,
|
|
.htt_alloc_txbuff = ath10k_htt_tx_alloc_cont_txbuf_32,
|
|
.htt_free_txbuff = ath10k_htt_tx_free_cont_txbuf_32,
|
|
.htt_h2t_aggr_cfg_msg = ath10k_htt_h2t_aggr_cfg_msg_32,
|
|
};
|
|
|
|
static const struct ath10k_htt_tx_ops htt_tx_ops_64 = {
|
|
.htt_send_rx_ring_cfg = ath10k_htt_send_rx_ring_cfg_64,
|
|
.htt_send_frag_desc_bank_cfg = ath10k_htt_send_frag_desc_bank_cfg_64,
|
|
.htt_alloc_frag_desc = ath10k_htt_tx_alloc_cont_frag_desc_64,
|
|
.htt_free_frag_desc = ath10k_htt_tx_free_cont_frag_desc_64,
|
|
.htt_tx = ath10k_htt_tx_64,
|
|
.htt_alloc_txbuff = ath10k_htt_tx_alloc_cont_txbuf_64,
|
|
.htt_free_txbuff = ath10k_htt_tx_free_cont_txbuf_64,
|
|
.htt_h2t_aggr_cfg_msg = ath10k_htt_h2t_aggr_cfg_msg_v2,
|
|
};
|
|
|
|
static const struct ath10k_htt_tx_ops htt_tx_ops_hl = {
|
|
.htt_send_rx_ring_cfg = ath10k_htt_send_rx_ring_cfg_hl,
|
|
.htt_send_frag_desc_bank_cfg = ath10k_htt_send_frag_desc_bank_cfg_32,
|
|
.htt_tx = ath10k_htt_tx_hl,
|
|
.htt_h2t_aggr_cfg_msg = ath10k_htt_h2t_aggr_cfg_msg_32,
|
|
.htt_flush_tx = ath10k_htt_flush_tx_queue,
|
|
};
|
|
|
|
void ath10k_htt_set_tx_ops(struct ath10k_htt *htt)
|
|
{
|
|
struct ath10k *ar = htt->ar;
|
|
|
|
if (ar->bus_param.dev_type == ATH10K_DEV_TYPE_HL)
|
|
htt->tx_ops = &htt_tx_ops_hl;
|
|
else if (ar->hw_params.target_64bit)
|
|
htt->tx_ops = &htt_tx_ops_64;
|
|
else
|
|
htt->tx_ops = &htt_tx_ops_32;
|
|
}
|