Linux Audio

Check our new training course

Linux debugging, profiling, tracing and performance analysis training

Apr 14-17, 2025
Register
Loading...
Note: File does not exist in v3.1.
  1// SPDX-License-Identifier: ISC
  2/*
  3 * Copyright (C) 2023 Lorenzo Bianconi <lorenzo@kernel.org>
  4 */
  5
  6#include "mt76.h"
  7#include "dma.h"
  8
  9void mt76_wed_release_rx_buf(struct mtk_wed_device *wed)
 10{
 11	struct mt76_dev *dev = container_of(wed, struct mt76_dev, mmio.wed);
 12	int i;
 13
 14	for (i = 0; i < dev->rx_token_size; i++) {
 15		struct mt76_txwi_cache *t;
 16
 17		t = mt76_rx_token_release(dev, i);
 18		if (!t || !t->ptr)
 19			continue;
 20
 21		mt76_put_page_pool_buf(t->ptr, false);
 22		t->ptr = NULL;
 23
 24		mt76_put_rxwi(dev, t);
 25	}
 26
 27	mt76_free_pending_rxwi(dev);
 28}
 29EXPORT_SYMBOL_GPL(mt76_wed_release_rx_buf);
 30
 31#ifdef CONFIG_NET_MEDIATEK_SOC_WED
 32u32 mt76_wed_init_rx_buf(struct mtk_wed_device *wed, int size)
 33{
 34	struct mt76_dev *dev = container_of(wed, struct mt76_dev, mmio.wed);
 35	struct mtk_wed_bm_desc *desc = wed->rx_buf_ring.desc;
 36	struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
 37	int i, len = SKB_WITH_OVERHEAD(q->buf_size);
 38	struct mt76_txwi_cache *t = NULL;
 39
 40	for (i = 0; i < size; i++) {
 41		enum dma_data_direction dir;
 42		dma_addr_t addr;
 43		u32 offset;
 44		int token;
 45		void *buf;
 46
 47		t = mt76_get_rxwi(dev);
 48		if (!t)
 49			goto unmap;
 50
 51		buf = mt76_get_page_pool_buf(q, &offset, q->buf_size);
 52		if (!buf)
 53			goto unmap;
 54
 55		addr = page_pool_get_dma_addr(virt_to_head_page(buf)) + offset;
 56		dir = page_pool_get_dma_dir(q->page_pool);
 57		dma_sync_single_for_device(dev->dma_dev, addr, len, dir);
 58
 59		desc->buf0 = cpu_to_le32(addr);
 60		token = mt76_rx_token_consume(dev, buf, t, addr);
 61		if (token < 0) {
 62			mt76_put_page_pool_buf(buf, false);
 63			goto unmap;
 64		}
 65
 66		token = FIELD_PREP(MT_DMA_CTL_TOKEN, token);
 67#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
 68		token |= FIELD_PREP(MT_DMA_CTL_SDP0_H, addr >> 32);
 69#endif
 70		desc->token |= cpu_to_le32(token);
 71		desc++;
 72	}
 73
 74	return 0;
 75
 76unmap:
 77	if (t)
 78		mt76_put_rxwi(dev, t);
 79	mt76_wed_release_rx_buf(wed);
 80
 81	return -ENOMEM;
 82}
 83EXPORT_SYMBOL_GPL(mt76_wed_init_rx_buf);
 84
 85int mt76_wed_offload_enable(struct mtk_wed_device *wed)
 86{
 87	struct mt76_dev *dev = container_of(wed, struct mt76_dev, mmio.wed);
 88
 89	spin_lock_bh(&dev->token_lock);
 90	dev->token_size = wed->wlan.token_start;
 91	spin_unlock_bh(&dev->token_lock);
 92
 93	return !wait_event_timeout(dev->tx_wait, !dev->wed_token_count, HZ);
 94}
 95EXPORT_SYMBOL_GPL(mt76_wed_offload_enable);
 96
 97int mt76_wed_dma_setup(struct mt76_dev *dev, struct mt76_queue *q, bool reset)
 98{
 99	int ret = 0, type, ring;
100	u16 flags;
101
102	if (!q || !q->ndesc)
103		return -EINVAL;
104
105	flags = q->flags;
106	if (!q->wed || !mtk_wed_device_active(q->wed))
107		q->flags &= ~MT_QFLAG_WED;
108
109	if (!(q->flags & MT_QFLAG_WED))
110		return 0;
111
112	type = FIELD_GET(MT_QFLAG_WED_TYPE, q->flags);
113	ring = FIELD_GET(MT_QFLAG_WED_RING, q->flags);
114
115	switch (type) {
116	case MT76_WED_Q_TX:
117		ret = mtk_wed_device_tx_ring_setup(q->wed, ring, q->regs,
118						   reset);
119		if (!ret)
120			q->wed_regs = q->wed->tx_ring[ring].reg_base;
121		break;
122	case MT76_WED_Q_TXFREE:
123		/* WED txfree queue needs ring to be initialized before setup */
124		q->flags = 0;
125		mt76_dma_queue_reset(dev, q);
126		mt76_dma_rx_fill(dev, q, false);
127
128		ret = mtk_wed_device_txfree_ring_setup(q->wed, q->regs);
129		if (!ret)
130			q->wed_regs = q->wed->txfree_ring.reg_base;
131		break;
132	case MT76_WED_Q_RX:
133		ret = mtk_wed_device_rx_ring_setup(q->wed, ring, q->regs,
134						   reset);
135		if (!ret)
136			q->wed_regs = q->wed->rx_ring[ring].reg_base;
137		break;
138	case MT76_WED_RRO_Q_DATA:
139		q->flags &= ~MT_QFLAG_WED;
140		__mt76_dma_queue_reset(dev, q, false);
141		mtk_wed_device_rro_rx_ring_setup(q->wed, ring, q->regs);
142		q->head = q->ndesc - 1;
143		q->queued = q->head;
144		break;
145	case MT76_WED_RRO_Q_MSDU_PG:
146		q->flags &= ~MT_QFLAG_WED;
147		__mt76_dma_queue_reset(dev, q, false);
148		mtk_wed_device_msdu_pg_rx_ring_setup(q->wed, ring, q->regs);
149		q->head = q->ndesc - 1;
150		q->queued = q->head;
151		break;
152	case MT76_WED_RRO_Q_IND:
153		q->flags &= ~MT_QFLAG_WED;
154		mt76_dma_queue_reset(dev, q);
155		mt76_dma_rx_fill(dev, q, false);
156		mtk_wed_device_ind_rx_ring_setup(q->wed, q->regs);
157		break;
158	default:
159		ret = -EINVAL;
160		break;
161	}
162	q->flags = flags;
163
164	return ret;
165}
166EXPORT_SYMBOL_GPL(mt76_wed_dma_setup);
167#endif /*CONFIG_NET_MEDIATEK_SOC_WED */
168
169void mt76_wed_offload_disable(struct mtk_wed_device *wed)
170{
171	struct mt76_dev *dev = container_of(wed, struct mt76_dev, mmio.wed);
172
173	spin_lock_bh(&dev->token_lock);
174	dev->token_size = dev->drv->token_size;
175	spin_unlock_bh(&dev->token_lock);
176}
177EXPORT_SYMBOL_GPL(mt76_wed_offload_disable);
178
179void mt76_wed_reset_complete(struct mtk_wed_device *wed)
180{
181	struct mt76_dev *dev = container_of(wed, struct mt76_dev, mmio.wed);
182
183	complete(&dev->mmio.wed_reset_complete);
184}
185EXPORT_SYMBOL_GPL(mt76_wed_reset_complete);
186
187int mt76_wed_net_setup_tc(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
188			  struct net_device *netdev, enum tc_setup_type type,
189			  void *type_data)
190{
191	struct mt76_phy *phy = hw->priv;
192	struct mtk_wed_device *wed = &phy->dev->mmio.wed;
193
194	if (!mtk_wed_device_active(wed))
195		return -EOPNOTSUPP;
196
197	return mtk_wed_device_setup_tc(wed, netdev, type, type_data);
198}
199EXPORT_SYMBOL_GPL(mt76_wed_net_setup_tc);
200
201void mt76_wed_dma_reset(struct mt76_dev *dev)
202{
203	struct mt76_mmio *mmio = &dev->mmio;
204
205	if (!test_bit(MT76_STATE_WED_RESET, &dev->phy.state))
206		return;
207
208	complete(&mmio->wed_reset);
209
210	if (!wait_for_completion_timeout(&mmio->wed_reset_complete, 3 * HZ))
211		dev_err(dev->dev, "wed reset complete timeout\n");
212}
213EXPORT_SYMBOL_GPL(mt76_wed_dma_reset);