Linux Audio

Check our new training course

Loading...
v6.8
  1// SPDX-License-Identifier: ISC
  2/* Copyright (C) 2020 MediaTek Inc.
  3 *
  4 * This file is written based on mt76/usb.c.
  5 *
  6 * Author: Felix Fietkau <nbd@nbd.name>
  7 *	   Lorenzo Bianconi <lorenzo@kernel.org>
  8 *	   Sean Wang <sean.wang@mediatek.com>
  9 */
 10
 11#include <linux/iopoll.h>
 12#include <linux/kernel.h>
 13#include <linux/module.h>
 14#include <linux/mmc/sdio_func.h>
 15#include <linux/mmc/card.h>
 16#include <linux/mmc/host.h>
 17#include <linux/sched.h>
 18#include <linux/kthread.h>
 19
 20#include "mt76.h"
 21#include "sdio.h"
 22
 23static u32 mt76s_read_whisr(struct mt76_dev *dev)
 24{
 25	return sdio_readl(dev->sdio.func, MCR_WHISR, NULL);
 26}
 27
 28u32 mt76s_read_pcr(struct mt76_dev *dev)
 29{
 30	struct mt76_sdio *sdio = &dev->sdio;
 31
 32	return sdio_readl(sdio->func, MCR_WHLPCR, NULL);
 33}
 34EXPORT_SYMBOL_GPL(mt76s_read_pcr);
 35
 36static u32 mt76s_read_mailbox(struct mt76_dev *dev, u32 offset)
 37{
 38	struct sdio_func *func = dev->sdio.func;
 39	u32 val = ~0, status;
 40	int err;
 41
 42	sdio_claim_host(func);
 43
 44	sdio_writel(func, offset, MCR_H2DSM0R, &err);
 45	if (err < 0) {
 46		dev_err(dev->dev, "failed setting address [err=%d]\n", err);
 47		goto out;
 48	}
 49
 50	sdio_writel(func, H2D_SW_INT_READ, MCR_WSICR, &err);
 51	if (err < 0) {
 52		dev_err(dev->dev, "failed setting read mode [err=%d]\n", err);
 53		goto out;
 54	}
 55
 56	err = readx_poll_timeout(mt76s_read_whisr, dev, status,
 57				 status & H2D_SW_INT_READ, 0, 1000000);
 58	if (err < 0) {
 59		dev_err(dev->dev, "query whisr timeout\n");
 60		goto out;
 61	}
 62
 63	sdio_writel(func, H2D_SW_INT_READ, MCR_WHISR, &err);
 64	if (err < 0) {
 65		dev_err(dev->dev, "failed setting read mode [err=%d]\n", err);
 66		goto out;
 67	}
 68
 69	val = sdio_readl(func, MCR_H2DSM0R, &err);
 70	if (err < 0) {
 71		dev_err(dev->dev, "failed reading h2dsm0r [err=%d]\n", err);
 72		goto out;
 73	}
 74
 75	if (val != offset) {
 76		dev_err(dev->dev, "register mismatch\n");
 77		val = ~0;
 78		goto out;
 79	}
 80
 81	val = sdio_readl(func, MCR_D2HRM1R, &err);
 82	if (err < 0)
 83		dev_err(dev->dev, "failed reading d2hrm1r [err=%d]\n", err);
 84
 85out:
 86	sdio_release_host(func);
 87
 88	return val;
 89}
 90
 91static void mt76s_write_mailbox(struct mt76_dev *dev, u32 offset, u32 val)
 92{
 93	struct sdio_func *func = dev->sdio.func;
 94	u32 status;
 95	int err;
 96
 97	sdio_claim_host(func);
 98
 99	sdio_writel(func, offset, MCR_H2DSM0R, &err);
100	if (err < 0) {
101		dev_err(dev->dev, "failed setting address [err=%d]\n", err);
102		goto out;
103	}
104
105	sdio_writel(func, val, MCR_H2DSM1R, &err);
106	if (err < 0) {
107		dev_err(dev->dev,
108			"failed setting write value [err=%d]\n", err);
109		goto out;
110	}
111
112	sdio_writel(func, H2D_SW_INT_WRITE, MCR_WSICR, &err);
113	if (err < 0) {
114		dev_err(dev->dev, "failed setting write mode [err=%d]\n", err);
115		goto out;
116	}
117
118	err = readx_poll_timeout(mt76s_read_whisr, dev, status,
119				 status & H2D_SW_INT_WRITE, 0, 1000000);
120	if (err < 0) {
121		dev_err(dev->dev, "query whisr timeout\n");
122		goto out;
123	}
124
125	sdio_writel(func, H2D_SW_INT_WRITE, MCR_WHISR, &err);
126	if (err < 0) {
127		dev_err(dev->dev, "failed setting write mode [err=%d]\n", err);
128		goto out;
129	}
130
131	val = sdio_readl(func, MCR_H2DSM0R, &err);
132	if (err < 0) {
133		dev_err(dev->dev, "failed reading h2dsm0r [err=%d]\n", err);
134		goto out;
135	}
136
137	if (val != offset)
138		dev_err(dev->dev, "register mismatch\n");
139
140out:
141	sdio_release_host(func);
142}
143
144u32 mt76s_rr(struct mt76_dev *dev, u32 offset)
145{
146	if (test_bit(MT76_STATE_MCU_RUNNING, &dev->phy.state))
147		return dev->mcu_ops->mcu_rr(dev, offset);
148	else
149		return mt76s_read_mailbox(dev, offset);
150}
151EXPORT_SYMBOL_GPL(mt76s_rr);
152
153void mt76s_wr(struct mt76_dev *dev, u32 offset, u32 val)
154{
155	if (test_bit(MT76_STATE_MCU_RUNNING, &dev->phy.state))
156		dev->mcu_ops->mcu_wr(dev, offset, val);
157	else
158		mt76s_write_mailbox(dev, offset, val);
159}
160EXPORT_SYMBOL_GPL(mt76s_wr);
161
162u32 mt76s_rmw(struct mt76_dev *dev, u32 offset, u32 mask, u32 val)
163{
164	val |= mt76s_rr(dev, offset) & ~mask;
165	mt76s_wr(dev, offset, val);
166
167	return val;
168}
169EXPORT_SYMBOL_GPL(mt76s_rmw);
170
171void mt76s_write_copy(struct mt76_dev *dev, u32 offset,
172		      const void *data, int len)
173{
174	const u32 *val = data;
175	int i;
176
177	for (i = 0; i < len / sizeof(u32); i++) {
178		mt76s_wr(dev, offset, val[i]);
179		offset += sizeof(u32);
180	}
181}
182EXPORT_SYMBOL_GPL(mt76s_write_copy);
183
184void mt76s_read_copy(struct mt76_dev *dev, u32 offset,
185		     void *data, int len)
186{
187	u32 *val = data;
188	int i;
189
190	for (i = 0; i < len / sizeof(u32); i++) {
191		val[i] = mt76s_rr(dev, offset);
192		offset += sizeof(u32);
193	}
194}
195EXPORT_SYMBOL_GPL(mt76s_read_copy);
196
197int mt76s_wr_rp(struct mt76_dev *dev, u32 base,
198		const struct mt76_reg_pair *data,
199		int len)
200{
201	int i;
202
203	for (i = 0; i < len; i++) {
204		mt76s_wr(dev, data->reg, data->value);
205		data++;
206	}
207
208	return 0;
209}
210EXPORT_SYMBOL_GPL(mt76s_wr_rp);
211
212int mt76s_rd_rp(struct mt76_dev *dev, u32 base,
213		struct mt76_reg_pair *data, int len)
214{
215	int i;
216
217	for (i = 0; i < len; i++) {
218		data->value = mt76s_rr(dev, data->reg);
219		data++;
220	}
221
222	return 0;
223}
224EXPORT_SYMBOL_GPL(mt76s_rd_rp);
225
226int mt76s_hw_init(struct mt76_dev *dev, struct sdio_func *func, int hw_ver)
227{
228	u32 status, ctrl;
229	int ret;
230
231	dev->sdio.hw_ver = hw_ver;
232
233	sdio_claim_host(func);
234
235	ret = sdio_enable_func(func);
236	if (ret < 0)
237		goto release;
238
239	/* Get ownership from the device */
240	sdio_writel(func, WHLPCR_INT_EN_CLR | WHLPCR_FW_OWN_REQ_CLR,
241		    MCR_WHLPCR, &ret);
242	if (ret < 0)
243		goto disable_func;
244
245	ret = readx_poll_timeout(mt76s_read_pcr, dev, status,
246				 status & WHLPCR_IS_DRIVER_OWN, 2000, 1000000);
247	if (ret < 0) {
248		dev_err(dev->dev, "Cannot get ownership from device");
249		goto disable_func;
250	}
251
252	ret = sdio_set_block_size(func, 512);
253	if (ret < 0)
254		goto disable_func;
255
256	/* Enable interrupt */
257	sdio_writel(func, WHLPCR_INT_EN_SET, MCR_WHLPCR, &ret);
258	if (ret < 0)
259		goto disable_func;
260
261	ctrl = WHIER_RX0_DONE_INT_EN | WHIER_TX_DONE_INT_EN;
262	if (hw_ver == MT76_CONNAC2_SDIO)
263		ctrl |= WHIER_RX1_DONE_INT_EN;
264	sdio_writel(func, ctrl, MCR_WHIER, &ret);
265	if (ret < 0)
266		goto disable_func;
267
268	switch (hw_ver) {
269	case MT76_CONNAC_SDIO:
270		/* set WHISR as read clear and Rx aggregation number as 16 */
271		ctrl = FIELD_PREP(MAX_HIF_RX_LEN_NUM, 16);
272		break;
273	default:
274		ctrl = sdio_readl(func, MCR_WHCR, &ret);
275		if (ret < 0)
276			goto disable_func;
277		ctrl &= ~MAX_HIF_RX_LEN_NUM_CONNAC2;
278		ctrl &= ~W_INT_CLR_CTRL; /* read clear */
279		ctrl |= FIELD_PREP(MAX_HIF_RX_LEN_NUM_CONNAC2, 0);
280		break;
281	}
282
283	sdio_writel(func, ctrl, MCR_WHCR, &ret);
284	if (ret < 0)
285		goto disable_func;
286
287	ret = sdio_claim_irq(func, mt76s_sdio_irq);
288	if (ret < 0)
289		goto disable_func;
290
291	sdio_release_host(func);
292
293	return 0;
294
295disable_func:
296	sdio_disable_func(func);
297release:
298	sdio_release_host(func);
299
300	return ret;
301}
302EXPORT_SYMBOL_GPL(mt76s_hw_init);
303
304int mt76s_alloc_rx_queue(struct mt76_dev *dev, enum mt76_rxq_id qid)
305{
306	struct mt76_queue *q = &dev->q_rx[qid];
307
308	spin_lock_init(&q->lock);
309	q->entry = devm_kcalloc(dev->dev,
310				MT76S_NUM_RX_ENTRIES, sizeof(*q->entry),
311				GFP_KERNEL);
312	if (!q->entry)
313		return -ENOMEM;
314
315	q->ndesc = MT76S_NUM_RX_ENTRIES;
316	q->head = q->tail = 0;
317	q->queued = 0;
318
319	return 0;
320}
321EXPORT_SYMBOL_GPL(mt76s_alloc_rx_queue);
322
323static struct mt76_queue *mt76s_alloc_tx_queue(struct mt76_dev *dev)
324{
325	struct mt76_queue *q;
 
326
327	q = devm_kzalloc(dev->dev, sizeof(*q), GFP_KERNEL);
328	if (!q)
329		return ERR_PTR(-ENOMEM);
330
331	spin_lock_init(&q->lock);
332	q->entry = devm_kcalloc(dev->dev,
333				MT76S_NUM_TX_ENTRIES, sizeof(*q->entry),
334				GFP_KERNEL);
335	if (!q->entry)
336		return ERR_PTR(-ENOMEM);
 
 
 
 
 
 
 
337
338	q->ndesc = MT76S_NUM_TX_ENTRIES;
 
339
340	return q;
341}
342
343int mt76s_alloc_tx(struct mt76_dev *dev)
344{
345	struct mt76_queue *q;
346	int i;
347
348	for (i = 0; i <= MT_TXQ_PSD; i++) {
349		q = mt76s_alloc_tx_queue(dev);
350		if (IS_ERR(q))
351			return PTR_ERR(q);
352
353		dev->phy.q_tx[i] = q;
354	}
 
355
356	q = mt76s_alloc_tx_queue(dev);
357	if (IS_ERR(q))
358		return PTR_ERR(q);
359
360	dev->q_mcu[MT_MCUQ_WM] = q;
 
 
361
362	return 0;
363}
364EXPORT_SYMBOL_GPL(mt76s_alloc_tx);
365
366static struct mt76_queue_entry *
367mt76s_get_next_rx_entry(struct mt76_queue *q)
368{
369	struct mt76_queue_entry *e = NULL;
370
371	spin_lock_bh(&q->lock);
372	if (q->queued > 0) {
373		e = &q->entry[q->tail];
374		q->tail = (q->tail + 1) % q->ndesc;
375		q->queued--;
376	}
377	spin_unlock_bh(&q->lock);
378
379	return e;
380}
381
382static int
383mt76s_process_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
384{
385	int qid = q - &dev->q_rx[MT_RXQ_MAIN];
386	int nframes = 0;
387
388	while (true) {
389		struct mt76_queue_entry *e;
390
391		if (!test_bit(MT76_STATE_INITIALIZED, &dev->phy.state))
392			break;
393
394		e = mt76s_get_next_rx_entry(q);
395		if (!e || !e->skb)
396			break;
397
398		dev->drv->rx_skb(dev, MT_RXQ_MAIN, e->skb, NULL);
399		e->skb = NULL;
400		nframes++;
401	}
402	if (qid == MT_RXQ_MAIN)
403		mt76_rx_poll_complete(dev, MT_RXQ_MAIN, NULL);
404
405	return nframes;
406}
407
408static void mt76s_net_worker(struct mt76_worker *w)
409{
410	struct mt76_sdio *sdio = container_of(w, struct mt76_sdio,
411					      net_worker);
412	struct mt76_dev *dev = container_of(sdio, struct mt76_dev, sdio);
413	int i, nframes;
414
415	do {
416		nframes = 0;
417
418		local_bh_disable();
419		rcu_read_lock();
420
421		mt76_for_each_q_rx(dev, i)
422			nframes += mt76s_process_rx_queue(dev, &dev->q_rx[i]);
423
424		rcu_read_unlock();
425		local_bh_enable();
426	} while (nframes > 0);
427}
428
429static int mt76s_process_tx_queue(struct mt76_dev *dev, struct mt76_queue *q)
430{
 
 
431	struct mt76_queue_entry entry;
432	int nframes = 0;
433	bool mcu;
434
435	if (!q)
436		return 0;
437
438	mcu = q == dev->q_mcu[MT_MCUQ_WM];
439	while (q->queued > 0) {
440		if (!q->entry[q->tail].done)
441			break;
442
443		entry = q->entry[q->tail];
444		q->entry[q->tail].done = false;
445
446		if (mcu) {
447			dev_kfree_skb(entry.skb);
448			entry.skb = NULL;
449		}
450
451		mt76_queue_tx_complete(dev, q, &entry);
452		nframes++;
 
 
 
 
 
 
 
453	}
454
455	if (!q->queued)
456		wake_up(&dev->tx_wait);
457
458	return nframes;
459}
460
461static void mt76s_status_worker(struct mt76_worker *w)
462{
463	struct mt76_sdio *sdio = container_of(w, struct mt76_sdio,
464					      status_worker);
465	struct mt76_dev *dev = container_of(sdio, struct mt76_dev, sdio);
466	bool resched = false;
467	int i, nframes;
468
469	do {
470		int ndata_frames = 0;
471
472		nframes = mt76s_process_tx_queue(dev, dev->q_mcu[MT_MCUQ_WM]);
 
473
474		for (i = 0; i <= MT_TXQ_PSD; i++)
475			ndata_frames += mt76s_process_tx_queue(dev,
476							       dev->phy.q_tx[i]);
477		nframes += ndata_frames;
478		if (ndata_frames > 0)
479			resched = true;
480
481		if (dev->drv->tx_status_data && ndata_frames > 0 &&
482		    !test_and_set_bit(MT76_READING_STATS, &dev->phy.state) &&
483		    !test_bit(MT76_STATE_SUSPEND, &dev->phy.state))
484			mt76_worker_schedule(&sdio->stat_worker);
485	} while (nframes > 0);
486
487	if (resched)
488		mt76_worker_schedule(&dev->tx_worker);
 
489}
490
491static void mt76s_tx_status_data(struct mt76_worker *worker)
492{
493	struct mt76_sdio *sdio;
494	struct mt76_dev *dev;
495	u8 update = 1;
496	u16 count = 0;
497
498	sdio = container_of(worker, struct mt76_sdio, stat_worker);
499	dev = container_of(sdio, struct mt76_dev, sdio);
500
501	while (true) {
502		if (test_bit(MT76_REMOVED, &dev->phy.state))
503			break;
504
505		if (!dev->drv->tx_status_data(dev, &update))
506			break;
507		count++;
508	}
509
510	if (count && test_bit(MT76_STATE_RUNNING, &dev->phy.state))
511		mt76_worker_schedule(&sdio->status_worker);
512	else
513		clear_bit(MT76_READING_STATS, &dev->phy.state);
514}
515
516static int
517mt76s_tx_queue_skb(struct mt76_dev *dev, struct mt76_queue *q,
518		   enum mt76_txq_id qid, struct sk_buff *skb,
519		   struct mt76_wcid *wcid, struct ieee80211_sta *sta)
520{
 
521	struct mt76_tx_info tx_info = {
522		.skb = skb,
523	};
524	int err, len = skb->len;
525	u16 idx = q->head;
526
527	if (q->queued == q->ndesc)
528		return -ENOSPC;
529
530	skb->prev = skb->next = NULL;
531	err = dev->drv->tx_prepare_skb(dev, NULL, qid, wcid, sta, &tx_info);
532	if (err < 0)
533		return err;
534
535	q->entry[q->head].skb = tx_info.skb;
536	q->entry[q->head].buf_sz = len;
537	q->entry[q->head].wcid = 0xffff;
538
539	smp_wmb();
540
541	q->head = (q->head + 1) % q->ndesc;
542	q->queued++;
543
544	return idx;
545}
546
547static int
548mt76s_tx_queue_skb_raw(struct mt76_dev *dev, struct mt76_queue *q,
549		       struct sk_buff *skb, u32 tx_info)
550{
551	int ret = -ENOSPC, len = skb->len, pad;
 
552
 
553	if (q->queued == q->ndesc)
554		goto error;
555
556	pad = round_up(skb->len, 4) - skb->len;
557	ret = mt76_skb_adjust_pad(skb, pad);
558	if (ret)
559		goto error;
560
561	spin_lock_bh(&q->lock);
562
563	q->entry[q->head].buf_sz = len;
564	q->entry[q->head].skb = skb;
565
566	/* ensure the entry fully updated before bus access */
567	smp_wmb();
568
569	q->head = (q->head + 1) % q->ndesc;
 
 
570	q->queued++;
571
 
572	spin_unlock_bh(&q->lock);
573
574	return 0;
575
576error:
577	dev_kfree_skb(skb);
578
579	return ret;
580}
581
582static void mt76s_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
583{
584	struct mt76_sdio *sdio = &dev->sdio;
585
586	mt76_worker_schedule(&sdio->txrx_worker);
587}
588
589static const struct mt76_queue_ops sdio_queue_ops = {
590	.tx_queue_skb = mt76s_tx_queue_skb,
591	.kick = mt76s_tx_kick,
592	.tx_queue_skb_raw = mt76s_tx_queue_skb_raw,
593};
594
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
595void mt76s_deinit(struct mt76_dev *dev)
596{
597	struct mt76_sdio *sdio = &dev->sdio;
598	int i;
599
600	mt76_worker_teardown(&sdio->txrx_worker);
601	mt76_worker_teardown(&sdio->status_worker);
602	mt76_worker_teardown(&sdio->net_worker);
603	mt76_worker_teardown(&sdio->stat_worker);
604
605	clear_bit(MT76_READING_STATS, &dev->phy.state);
606
607	mt76_tx_status_check(dev, true);
608
609	sdio_claim_host(sdio->func);
610	sdio_release_irq(sdio->func);
611	sdio_release_host(sdio->func);
612
613	mt76_for_each_q_rx(dev, i) {
614		struct mt76_queue *q = &dev->q_rx[i];
615		int j;
616
617		for (j = 0; j < q->ndesc; j++) {
618			struct mt76_queue_entry *e = &q->entry[j];
619
620			if (!e->skb)
621				continue;
622
623			dev_kfree_skb(e->skb);
624			e->skb = NULL;
625		}
626	}
627}
628EXPORT_SYMBOL_GPL(mt76s_deinit);
629
630int mt76s_init(struct mt76_dev *dev, struct sdio_func *func,
631	       const struct mt76_bus_ops *bus_ops)
632{
633	struct mt76_sdio *sdio = &dev->sdio;
634	u32 host_max_cap;
635	int err;
636
637	err = mt76_worker_setup(dev->hw, &sdio->status_worker,
638				mt76s_status_worker, "sdio-status");
639	if (err)
640		return err;
641
642	err = mt76_worker_setup(dev->hw, &sdio->net_worker, mt76s_net_worker,
643				"sdio-net");
644	if (err)
645		return err;
646
647	err = mt76_worker_setup(dev->hw, &sdio->stat_worker, mt76s_tx_status_data,
648				"sdio-sta");
649	if (err)
650		return err;
651
652	sched_set_fifo_low(sdio->status_worker.task);
653	sched_set_fifo_low(sdio->net_worker.task);
654	sched_set_fifo_low(sdio->stat_worker.task);
655
 
656	dev->queue_ops = &sdio_queue_ops;
657	dev->bus = bus_ops;
658	dev->sdio.func = func;
659
660	host_max_cap = min_t(u32, func->card->host->max_req_size,
661			     func->cur_blksize *
662			     func->card->host->max_blk_count);
663	dev->sdio.xmit_buf_sz = min_t(u32, host_max_cap, MT76S_XMIT_BUF_SZ);
664	dev->sdio.xmit_buf = devm_kmalloc(dev->dev, dev->sdio.xmit_buf_sz,
665					  GFP_KERNEL);
666	if (!dev->sdio.xmit_buf)
667		err = -ENOMEM;
668
669	return err;
670}
671EXPORT_SYMBOL_GPL(mt76s_init);
672
673MODULE_AUTHOR("Sean Wang <sean.wang@mediatek.com>");
674MODULE_AUTHOR("Lorenzo Bianconi <lorenzo@kernel.org>");
675MODULE_DESCRIPTION("MediaTek MT76x SDIO helpers");
676MODULE_LICENSE("Dual BSD/GPL");
v5.9
  1// SPDX-License-Identifier: ISC
  2/* Copyright (C) 2020 MediaTek Inc.
  3 *
  4 * This file is written based on mt76/usb.c.
  5 *
  6 * Author: Felix Fietkau <nbd@nbd.name>
  7 *	   Lorenzo Bianconi <lorenzo@kernel.org>
  8 *	   Sean Wang <sean.wang@mediatek.com>
  9 */
 10
 11#include <linux/iopoll.h>
 12#include <linux/kernel.h>
 13#include <linux/module.h>
 14#include <linux/mmc/sdio_func.h>
 
 
 15#include <linux/sched.h>
 16#include <linux/kthread.h>
 17
 18#include "mt76.h"
 
 19
 20static int
 21mt76s_alloc_rx_queue(struct mt76_dev *dev, enum mt76_rxq_id qid)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 22{
 23	struct mt76_queue *q = &dev->q_rx[qid];
 24
 25	spin_lock_init(&q->lock);
 26	q->entry = devm_kcalloc(dev->dev,
 27				MT_NUM_RX_ENTRIES, sizeof(*q->entry),
 28				GFP_KERNEL);
 29	if (!q->entry)
 30		return -ENOMEM;
 31
 32	q->ndesc = MT_NUM_RX_ENTRIES;
 33	q->head = q->tail = 0;
 34	q->queued = 0;
 35
 36	return 0;
 37}
 
 38
 39static int mt76s_alloc_tx(struct mt76_dev *dev)
 40{
 41	struct mt76_queue *q;
 42	int i;
 43
 44	for (i = 0; i < MT_TXQ_MCU_WA; i++) {
 45		INIT_LIST_HEAD(&dev->q_tx[i].swq);
 
 46
 47		q = devm_kzalloc(dev->dev, sizeof(*q), GFP_KERNEL);
 48		if (!q)
 49			return -ENOMEM;
 50
 51		spin_lock_init(&q->lock);
 52		q->hw_idx = i;
 53		dev->q_tx[i].q = q;
 54
 55		q->entry = devm_kcalloc(dev->dev,
 56					MT_NUM_TX_ENTRIES, sizeof(*q->entry),
 57					GFP_KERNEL);
 58		if (!q->entry)
 59			return -ENOMEM;
 60
 61		q->ndesc = MT_NUM_TX_ENTRIES;
 62	}
 63
 64	return 0;
 65}
 66
 67void mt76s_stop_txrx(struct mt76_dev *dev)
 68{
 69	struct mt76_sdio *sdio = &dev->sdio;
 
 70
 71	cancel_work_sync(&sdio->stat_work);
 72	clear_bit(MT76_READING_STATS, &dev->phy.state);
 
 
 73
 74	mt76_tx_status_check(dev, NULL, true);
 75}
 76EXPORT_SYMBOL_GPL(mt76s_stop_txrx);
 77
 78int mt76s_alloc_queues(struct mt76_dev *dev)
 79{
 80	int err;
 81
 82	err = mt76s_alloc_rx_queue(dev, MT_RXQ_MAIN);
 83	if (err < 0)
 84		return err;
 85
 86	return mt76s_alloc_tx(dev);
 87}
 88EXPORT_SYMBOL_GPL(mt76s_alloc_queues);
 89
 90static struct mt76_queue_entry *
 91mt76s_get_next_rx_entry(struct mt76_queue *q)
 92{
 93	struct mt76_queue_entry *e = NULL;
 94
 95	spin_lock_bh(&q->lock);
 96	if (q->queued > 0) {
 97		e = &q->entry[q->head];
 98		q->head = (q->head + 1) % q->ndesc;
 99		q->queued--;
100	}
101	spin_unlock_bh(&q->lock);
102
103	return e;
104}
105
106static int
107mt76s_process_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
108{
109	int qid = q - &dev->q_rx[MT_RXQ_MAIN];
110	int nframes = 0;
111
112	while (true) {
113		struct mt76_queue_entry *e;
114
115		if (!test_bit(MT76_STATE_INITIALIZED, &dev->phy.state))
116			break;
117
118		e = mt76s_get_next_rx_entry(q);
119		if (!e || !e->skb)
120			break;
121
122		dev->drv->rx_skb(dev, MT_RXQ_MAIN, e->skb);
123		e->skb = NULL;
124		nframes++;
125	}
126	if (qid == MT_RXQ_MAIN)
127		mt76_rx_poll_complete(dev, MT_RXQ_MAIN, NULL);
128
129	return nframes;
130}
131
132static int mt76s_process_tx_queue(struct mt76_dev *dev, enum mt76_txq_id qid)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133{
134	struct mt76_sw_queue *sq = &dev->q_tx[qid];
135	u32 n_dequeued = 0, n_sw_dequeued = 0;
136	struct mt76_queue_entry entry;
137	struct mt76_queue *q = sq->q;
138	bool wake;
 
 
 
139
140	while (q->queued > n_dequeued) {
141		if (!q->entry[q->head].done)
 
142			break;
143
144		if (q->entry[q->head].schedule) {
145			q->entry[q->head].schedule = false;
146			n_sw_dequeued++;
 
 
 
147		}
148
149		entry = q->entry[q->head];
150		q->entry[q->head].done = false;
151		q->head = (q->head + 1) % q->ndesc;
152		n_dequeued++;
153
154		if (qid == MT_TXQ_MCU)
155			dev_kfree_skb(entry.skb);
156		else
157			dev->drv->tx_complete_skb(dev, qid, &entry);
158	}
159
160	spin_lock_bh(&q->lock);
 
161
162	sq->swq_queued -= n_sw_dequeued;
163	q->queued -= n_dequeued;
164
165	wake = q->stopped && q->queued < q->ndesc - 8;
166	if (wake)
167		q->stopped = false;
168
169	if (!q->queued)
170		wake_up(&dev->tx_wait);
 
171
172	spin_unlock_bh(&q->lock);
 
173
174	if (qid == MT_TXQ_MCU)
175		goto out;
176
177	mt76_txq_schedule(&dev->phy, qid);
 
 
 
 
 
178
179	if (wake)
180		ieee80211_wake_queue(dev->hw, qid);
 
 
 
181
182	wake_up_process(dev->sdio.tx_kthread);
183out:
184	return n_dequeued;
185}
186
187static void mt76s_tx_status_data(struct work_struct *work)
188{
189	struct mt76_sdio *sdio;
190	struct mt76_dev *dev;
191	u8 update = 1;
192	u16 count = 0;
193
194	sdio = container_of(work, struct mt76_sdio, stat_work);
195	dev = container_of(sdio, struct mt76_dev, sdio);
196
197	while (true) {
198		if (test_bit(MT76_REMOVED, &dev->phy.state))
199			break;
200
201		if (!dev->drv->tx_status_data(dev, &update))
202			break;
203		count++;
204	}
205
206	if (count && test_bit(MT76_STATE_RUNNING, &dev->phy.state))
207		queue_work(dev->wq, &sdio->stat_work);
208	else
209		clear_bit(MT76_READING_STATS, &dev->phy.state);
210}
211
212static int
213mt76s_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
214		   struct sk_buff *skb, struct mt76_wcid *wcid,
215		   struct ieee80211_sta *sta)
216{
217	struct mt76_queue *q = dev->q_tx[qid].q;
218	struct mt76_tx_info tx_info = {
219		.skb = skb,
220	};
221	int err, len = skb->len;
222	u16 idx = q->tail;
223
224	if (q->queued == q->ndesc)
225		return -ENOSPC;
226
227	skb->prev = skb->next = NULL;
228	err = dev->drv->tx_prepare_skb(dev, NULL, qid, wcid, sta, &tx_info);
229	if (err < 0)
230		return err;
231
232	q->entry[q->tail].skb = tx_info.skb;
233	q->entry[q->tail].buf_sz = len;
234	q->tail = (q->tail + 1) % q->ndesc;
 
 
 
 
235	q->queued++;
236
237	return idx;
238}
239
240static int
241mt76s_tx_queue_skb_raw(struct mt76_dev *dev, enum mt76_txq_id qid,
242		       struct sk_buff *skb, u32 tx_info)
243{
244	struct mt76_queue *q = dev->q_tx[qid].q;
245	int ret = -ENOSPC, len = skb->len;
246
247	spin_lock_bh(&q->lock);
248	if (q->queued == q->ndesc)
249		goto out;
250
251	ret = mt76_skb_adjust_pad(skb);
 
252	if (ret)
253		goto out;
 
 
 
 
 
 
 
 
254
255	q->entry[q->tail].buf_sz = len;
256	q->entry[q->tail].skb = skb;
257	q->tail = (q->tail + 1) % q->ndesc;
258	q->queued++;
259
260out:
261	spin_unlock_bh(&q->lock);
262
 
 
 
 
 
263	return ret;
264}
265
266static void mt76s_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
267{
268	struct mt76_sdio *sdio = &dev->sdio;
269
270	wake_up_process(sdio->tx_kthread);
271}
272
273static const struct mt76_queue_ops sdio_queue_ops = {
274	.tx_queue_skb = mt76s_tx_queue_skb,
275	.kick = mt76s_tx_kick,
276	.tx_queue_skb_raw = mt76s_tx_queue_skb_raw,
277};
278
279static int mt76s_kthread_run(void *data)
280{
281	struct mt76_dev *dev = data;
282	struct mt76_phy *mphy = &dev->phy;
283
284	while (!kthread_should_stop()) {
285		int i, nframes = 0;
286
287		cond_resched();
288
289		/* rx processing */
290		local_bh_disable();
291		rcu_read_lock();
292
293		mt76_for_each_q_rx(dev, i)
294			nframes += mt76s_process_rx_queue(dev, &dev->q_rx[i]);
295
296		rcu_read_unlock();
297		local_bh_enable();
298
299		/* tx processing */
300		for (i = 0; i < MT_TXQ_MCU_WA; i++)
301			nframes += mt76s_process_tx_queue(dev, i);
302
303		if (dev->drv->tx_status_data &&
304		    !test_and_set_bit(MT76_READING_STATS, &mphy->state))
305			queue_work(dev->wq, &dev->sdio.stat_work);
306
307		if (!nframes || !test_bit(MT76_STATE_RUNNING, &mphy->state)) {
308			set_current_state(TASK_INTERRUPTIBLE);
309			schedule();
310		}
311	}
312
313	return 0;
314}
315
316void mt76s_deinit(struct mt76_dev *dev)
317{
318	struct mt76_sdio *sdio = &dev->sdio;
319	int i;
320
321	kthread_stop(sdio->kthread);
322	kthread_stop(sdio->tx_kthread);
323	mt76s_stop_txrx(dev);
 
 
 
 
 
324
325	sdio_claim_host(sdio->func);
326	sdio_release_irq(sdio->func);
327	sdio_release_host(sdio->func);
328
329	mt76_for_each_q_rx(dev, i) {
330		struct mt76_queue *q = &dev->q_rx[i];
331		int j;
332
333		for (j = 0; j < q->ndesc; j++) {
334			struct mt76_queue_entry *e = &q->entry[j];
335
336			if (!e->skb)
337				continue;
338
339			dev_kfree_skb(e->skb);
340			e->skb = NULL;
341		}
342	}
343}
344EXPORT_SYMBOL_GPL(mt76s_deinit);
345
346int mt76s_init(struct mt76_dev *dev, struct sdio_func *func,
347	       const struct mt76_bus_ops *bus_ops)
348{
349	struct mt76_sdio *sdio = &dev->sdio;
 
 
350
351	sdio->kthread = kthread_create(mt76s_kthread_run, dev, "mt76s");
352	if (IS_ERR(sdio->kthread))
353		return PTR_ERR(sdio->kthread);
 
354
355	INIT_WORK(&sdio->stat_work, mt76s_tx_status_data);
 
 
 
 
 
 
 
 
 
 
 
 
356
357	mutex_init(&sdio->sched.lock);
358	dev->queue_ops = &sdio_queue_ops;
359	dev->bus = bus_ops;
360	dev->sdio.func = func;
361
362	return 0;
 
 
 
 
 
 
 
 
 
363}
364EXPORT_SYMBOL_GPL(mt76s_init);
365
366MODULE_AUTHOR("Sean Wang <sean.wang@mediatek.com>");
367MODULE_AUTHOR("Lorenzo Bianconi <lorenzo@kernel.org>");
 
368MODULE_LICENSE("Dual BSD/GPL");