Linux Audio

Check our new training course

Loading...
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Queue between the tx operation and the bh workqueue.
  4 *
  5 * Copyright (c) 2017-2020, Silicon Laboratories, Inc.
  6 * Copyright (c) 2010, ST-Ericsson
  7 */
  8#include <linux/sched.h>
  9#include <net/mac80211.h>
 10
 11#include "queue.h"
 12#include "wfx.h"
 13#include "sta.h"
 14#include "data_tx.h"
 15#include "traces.h"
 16
 17void wfx_tx_lock(struct wfx_dev *wdev)
 18{
 19	atomic_inc(&wdev->tx_lock);
 20}
 21
 22void wfx_tx_unlock(struct wfx_dev *wdev)
 23{
 24	int tx_lock = atomic_dec_return(&wdev->tx_lock);
 25
 26	WARN(tx_lock < 0, "inconsistent tx_lock value");
 27	if (!tx_lock)
 28		wfx_bh_request_tx(wdev);
 29}
 30
 31void wfx_tx_flush(struct wfx_dev *wdev)
 32{
 33	int ret;
 34
 35	/* Do not wait for any reply if chip is frozen */
 36	if (wdev->chip_frozen)
 37		return;
 38
 39	wfx_tx_lock(wdev);
 40	mutex_lock(&wdev->hif_cmd.lock);
 41	ret = wait_event_timeout(wdev->hif.tx_buffers_empty, !wdev->hif.tx_buffers_used,
 42				 msecs_to_jiffies(3000));
 43	if (!ret) {
 44		dev_warn(wdev->dev, "cannot flush tx buffers (%d still busy)\n",
 45			 wdev->hif.tx_buffers_used);
 46		wfx_pending_dump_old_frames(wdev, 3000);
 47		/* FIXME: drop pending frames here */
 48		wdev->chip_frozen = true;
 49	}
 50	mutex_unlock(&wdev->hif_cmd.lock);
 51	wfx_tx_unlock(wdev);
 52}
 53
 54void wfx_tx_lock_flush(struct wfx_dev *wdev)
 55{
 56	wfx_tx_lock(wdev);
 57	wfx_tx_flush(wdev);
 58}
 59
 60void wfx_tx_queues_init(struct wfx_vif *wvif)
 61{
 62	/* The device is in charge to respect the details of the QoS parameters. The driver just
 63	 * ensure that it roughtly respect the priorities to avoid any shortage.
 64	 */
 65	const int priorities[IEEE80211_NUM_ACS] = { 1, 2, 64, 128 };
 66	int i;
 67
 68	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
 69		skb_queue_head_init(&wvif->tx_queue[i].normal);
 70		skb_queue_head_init(&wvif->tx_queue[i].cab);
 71		skb_queue_head_init(&wvif->tx_queue[i].offchan);
 72		wvif->tx_queue[i].priority = priorities[i];
 73	}
 74}
 75
 76bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue)
 77{
 78	return skb_queue_empty_lockless(&queue->normal) &&
 79	       skb_queue_empty_lockless(&queue->cab) &&
 80	       skb_queue_empty_lockless(&queue->offchan);
 81}
 82
 83void wfx_tx_queues_check_empty(struct wfx_vif *wvif)
 84{
 85	int i;
 86
 87	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
 88		WARN_ON(atomic_read(&wvif->tx_queue[i].pending_frames));
 89		WARN_ON(!wfx_tx_queue_empty(wvif, &wvif->tx_queue[i]));
 90	}
 91}
 92
 93static void __wfx_tx_queue_drop(struct wfx_vif *wvif,
 94				struct sk_buff_head *skb_queue, struct sk_buff_head *dropped)
 95{
 96	struct sk_buff *skb, *tmp;
 97
 98	spin_lock_bh(&skb_queue->lock);
 99	skb_queue_walk_safe(skb_queue, skb, tmp) {
100		__skb_unlink(skb, skb_queue);
101		skb_queue_head(dropped, skb);
102	}
103	spin_unlock_bh(&skb_queue->lock);
104}
105
106void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue,
107		       struct sk_buff_head *dropped)
108{
109	__wfx_tx_queue_drop(wvif, &queue->normal, dropped);
110	__wfx_tx_queue_drop(wvif, &queue->cab, dropped);
111	__wfx_tx_queue_drop(wvif, &queue->offchan, dropped);
112	wake_up(&wvif->wdev->tx_dequeue);
113}
114
115void wfx_tx_queues_put(struct wfx_vif *wvif, struct sk_buff *skb)
116{
117	struct wfx_queue *queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
118	struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
119
120	if (tx_info->flags & IEEE80211_TX_CTL_TX_OFFCHAN)
121		skb_queue_tail(&queue->offchan, skb);
122	else if (tx_info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM)
123		skb_queue_tail(&queue->cab, skb);
124	else
125		skb_queue_tail(&queue->normal, skb);
126}
127
128void wfx_pending_drop(struct wfx_dev *wdev, struct sk_buff_head *dropped)
129{
130	struct wfx_queue *queue;
131	struct wfx_vif *wvif;
 
132	struct sk_buff *skb;
133
134	WARN(!wdev->chip_frozen, "%s should only be used to recover a frozen device", __func__);
135	while ((skb = skb_dequeue(&wdev->tx_pending)) != NULL) {
136		wvif = wfx_skb_wvif(wdev, skb);
 
137		if (wvif) {
138			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
139			WARN_ON(skb_get_queue_mapping(skb) > 3);
140			WARN_ON(!atomic_read(&queue->pending_frames));
141			atomic_dec(&queue->pending_frames);
142		}
143		skb_queue_head(dropped, skb);
144	}
145}
146
147struct sk_buff *wfx_pending_get(struct wfx_dev *wdev, u32 packet_id)
148{
149	struct wfx_queue *queue;
150	struct wfx_hif_req_tx *req;
151	struct wfx_vif *wvif;
152	struct wfx_hif_msg *hif;
153	struct sk_buff *skb;
154
155	spin_lock_bh(&wdev->tx_pending.lock);
156	skb_queue_walk(&wdev->tx_pending, skb) {
157		hif = (struct wfx_hif_msg *)skb->data;
158		req = (struct wfx_hif_req_tx *)hif->body;
159		if (req->packet_id != packet_id)
160			continue;
161		spin_unlock_bh(&wdev->tx_pending.lock);
162		wvif = wfx_skb_wvif(wdev, skb);
163		if (wvif) {
164			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
165			WARN_ON(skb_get_queue_mapping(skb) > 3);
166			WARN_ON(!atomic_read(&queue->pending_frames));
167			atomic_dec(&queue->pending_frames);
168		}
169		skb_unlink(skb, &wdev->tx_pending);
170		return skb;
171	}
172	spin_unlock_bh(&wdev->tx_pending.lock);
173	WARN(1, "cannot find packet in pending queue");
174	return NULL;
175}
176
177void wfx_pending_dump_old_frames(struct wfx_dev *wdev, unsigned int limit_ms)
178{
179	ktime_t now = ktime_get();
180	struct wfx_tx_priv *tx_priv;
181	struct wfx_hif_req_tx *req;
182	struct sk_buff *skb;
183	bool first = true;
184
185	spin_lock_bh(&wdev->tx_pending.lock);
186	skb_queue_walk(&wdev->tx_pending, skb) {
187		tx_priv = wfx_skb_tx_priv(skb);
188		req = wfx_skb_txreq(skb);
189		if (ktime_after(now, ktime_add_ms(tx_priv->xmit_timestamp, limit_ms))) {
190			if (first) {
191				dev_info(wdev->dev, "frames stuck in firmware since %dms or more:\n",
192					 limit_ms);
193				first = false;
194			}
195			dev_info(wdev->dev, "   id %08x sent %lldms ago\n",
196				 req->packet_id, ktime_ms_delta(now, tx_priv->xmit_timestamp));
197		}
198	}
199	spin_unlock_bh(&wdev->tx_pending.lock);
200}
201
202unsigned int wfx_pending_get_pkt_us_delay(struct wfx_dev *wdev, struct sk_buff *skb)
203{
204	ktime_t now = ktime_get();
205	struct wfx_tx_priv *tx_priv = wfx_skb_tx_priv(skb);
206
207	return ktime_us_delta(now, tx_priv->xmit_timestamp);
208}
209
210bool wfx_tx_queues_has_cab(struct wfx_vif *wvif)
211{
212	struct ieee80211_vif *vif = wvif_to_vif(wvif);
213	int i;
214
215	if (vif->type != NL80211_IFTYPE_AP)
216		return false;
217	for (i = 0; i < IEEE80211_NUM_ACS; ++i)
218		/* Note: since only AP can have mcast frames in queue and only one vif can be AP,
219		 * all queued frames has same interface id
220		 */
221		if (!skb_queue_empty_lockless(&wvif->tx_queue[i].cab))
222			return true;
223	return false;
224}
225
226static int wfx_tx_queue_get_weight(struct wfx_queue *queue)
227{
228	return atomic_read(&queue->pending_frames) * queue->priority;
229}
230
231static struct sk_buff *wfx_tx_queues_get_skb(struct wfx_dev *wdev)
232{
233	struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)];
234	int i, j, num_queues = 0;
235	struct wfx_vif *wvif;
236	struct wfx_hif_msg *hif;
237	struct sk_buff *skb;
238
239	/* sort the queues */
240	wvif = NULL;
241	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
242		for (i = 0; i < IEEE80211_NUM_ACS; i++) {
243			WARN_ON(num_queues >= ARRAY_SIZE(queues));
244			queues[num_queues] = &wvif->tx_queue[i];
245			for (j = num_queues; j > 0; j--)
246				if (wfx_tx_queue_get_weight(queues[j]) <
247				    wfx_tx_queue_get_weight(queues[j - 1]))
248					swap(queues[j - 1], queues[j]);
249			num_queues++;
250		}
251	}
252
253	wvif = NULL;
254	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
255		for (i = 0; i < num_queues; i++) {
256			skb = skb_dequeue(&queues[i]->offchan);
257			if (!skb)
258				continue;
259			hif = (struct wfx_hif_msg *)skb->data;
260			/* Offchan frames are assigned to a special interface.
261			 * The only interface allowed to send data during scan.
262			 */
263			WARN_ON(hif->interface != 2);
264			atomic_inc(&queues[i]->pending_frames);
265			trace_queues_stats(wdev, queues[i]);
266			return skb;
267		}
268	}
269
270	if (mutex_is_locked(&wdev->scan_lock))
271		return NULL;
272
273	wvif = NULL;
274	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
275		if (!wvif->after_dtim_tx_allowed)
276			continue;
277		for (i = 0; i < num_queues; i++) {
278			skb = skb_dequeue(&queues[i]->cab);
279			if (!skb)
280				continue;
281			/* Note: since only AP can have mcast frames in queue and only one vif can
282			 * be AP, all queued frames has same interface id
283			 */
284			hif = (struct wfx_hif_msg *)skb->data;
285			WARN_ON(hif->interface != wvif->id);
286			WARN_ON(queues[i] != &wvif->tx_queue[skb_get_queue_mapping(skb)]);
287			atomic_inc(&queues[i]->pending_frames);
288			trace_queues_stats(wdev, queues[i]);
289			return skb;
290		}
291		/* No more multicast to sent */
292		wvif->after_dtim_tx_allowed = false;
293		schedule_work(&wvif->update_tim_work);
294	}
295
296	for (i = 0; i < num_queues; i++) {
297		skb = skb_dequeue(&queues[i]->normal);
298		if (skb) {
299			atomic_inc(&queues[i]->pending_frames);
300			trace_queues_stats(wdev, queues[i]);
301			return skb;
302		}
303	}
304	return NULL;
305}
306
307struct wfx_hif_msg *wfx_tx_queues_get(struct wfx_dev *wdev)
308{
309	struct wfx_tx_priv *tx_priv;
310	struct sk_buff *skb;
311
312	if (atomic_read(&wdev->tx_lock))
313		return NULL;
314	skb = wfx_tx_queues_get_skb(wdev);
315	if (!skb)
316		return NULL;
317	skb_queue_tail(&wdev->tx_pending, skb);
318	wake_up(&wdev->tx_dequeue);
319	tx_priv = wfx_skb_tx_priv(skb);
320	tx_priv->xmit_timestamp = ktime_get();
321	return (struct wfx_hif_msg *)skb->data;
322}
v6.2
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Queue between the tx operation and the bh workqueue.
  4 *
  5 * Copyright (c) 2017-2020, Silicon Laboratories, Inc.
  6 * Copyright (c) 2010, ST-Ericsson
  7 */
  8#include <linux/sched.h>
  9#include <net/mac80211.h>
 10
 11#include "queue.h"
 12#include "wfx.h"
 13#include "sta.h"
 14#include "data_tx.h"
 15#include "traces.h"
 16
 17void wfx_tx_lock(struct wfx_dev *wdev)
 18{
 19	atomic_inc(&wdev->tx_lock);
 20}
 21
 22void wfx_tx_unlock(struct wfx_dev *wdev)
 23{
 24	int tx_lock = atomic_dec_return(&wdev->tx_lock);
 25
 26	WARN(tx_lock < 0, "inconsistent tx_lock value");
 27	if (!tx_lock)
 28		wfx_bh_request_tx(wdev);
 29}
 30
 31void wfx_tx_flush(struct wfx_dev *wdev)
 32{
 33	int ret;
 34
 35	/* Do not wait for any reply if chip is frozen */
 36	if (wdev->chip_frozen)
 37		return;
 38
 39	wfx_tx_lock(wdev);
 40	mutex_lock(&wdev->hif_cmd.lock);
 41	ret = wait_event_timeout(wdev->hif.tx_buffers_empty, !wdev->hif.tx_buffers_used,
 42				 msecs_to_jiffies(3000));
 43	if (!ret) {
 44		dev_warn(wdev->dev, "cannot flush tx buffers (%d still busy)\n",
 45			 wdev->hif.tx_buffers_used);
 46		wfx_pending_dump_old_frames(wdev, 3000);
 47		/* FIXME: drop pending frames here */
 48		wdev->chip_frozen = true;
 49	}
 50	mutex_unlock(&wdev->hif_cmd.lock);
 51	wfx_tx_unlock(wdev);
 52}
 53
 54void wfx_tx_lock_flush(struct wfx_dev *wdev)
 55{
 56	wfx_tx_lock(wdev);
 57	wfx_tx_flush(wdev);
 58}
 59
 60void wfx_tx_queues_init(struct wfx_vif *wvif)
 61{
 62	/* The device is in charge to respect the details of the QoS parameters. The driver just
 63	 * ensure that it roughtly respect the priorities to avoid any shortage.
 64	 */
 65	const int priorities[IEEE80211_NUM_ACS] = { 1, 2, 64, 128 };
 66	int i;
 67
 68	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
 69		skb_queue_head_init(&wvif->tx_queue[i].normal);
 70		skb_queue_head_init(&wvif->tx_queue[i].cab);
 
 71		wvif->tx_queue[i].priority = priorities[i];
 72	}
 73}
 74
 75bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue)
 76{
 77	return skb_queue_empty_lockless(&queue->normal) && skb_queue_empty_lockless(&queue->cab);
 
 
 78}
 79
 80void wfx_tx_queues_check_empty(struct wfx_vif *wvif)
 81{
 82	int i;
 83
 84	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
 85		WARN_ON(atomic_read(&wvif->tx_queue[i].pending_frames));
 86		WARN_ON(!wfx_tx_queue_empty(wvif, &wvif->tx_queue[i]));
 87	}
 88}
 89
 90static void __wfx_tx_queue_drop(struct wfx_vif *wvif,
 91				struct sk_buff_head *skb_queue, struct sk_buff_head *dropped)
 92{
 93	struct sk_buff *skb, *tmp;
 94
 95	spin_lock_bh(&skb_queue->lock);
 96	skb_queue_walk_safe(skb_queue, skb, tmp) {
 97		__skb_unlink(skb, skb_queue);
 98		skb_queue_head(dropped, skb);
 99	}
100	spin_unlock_bh(&skb_queue->lock);
101}
102
103void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue,
104		       struct sk_buff_head *dropped)
105{
 
106	__wfx_tx_queue_drop(wvif, &queue->cab, dropped);
107	__wfx_tx_queue_drop(wvif, &queue->normal, dropped);
108	wake_up(&wvif->wdev->tx_dequeue);
109}
110
111void wfx_tx_queues_put(struct wfx_vif *wvif, struct sk_buff *skb)
112{
113	struct wfx_queue *queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
114	struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
115
116	if (tx_info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM)
 
 
117		skb_queue_tail(&queue->cab, skb);
118	else
119		skb_queue_tail(&queue->normal, skb);
120}
121
122void wfx_pending_drop(struct wfx_dev *wdev, struct sk_buff_head *dropped)
123{
124	struct wfx_queue *queue;
125	struct wfx_vif *wvif;
126	struct wfx_hif_msg *hif;
127	struct sk_buff *skb;
128
129	WARN(!wdev->chip_frozen, "%s should only be used to recover a frozen device", __func__);
130	while ((skb = skb_dequeue(&wdev->tx_pending)) != NULL) {
131		hif = (struct wfx_hif_msg *)skb->data;
132		wvif = wdev_to_wvif(wdev, hif->interface);
133		if (wvif) {
134			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
135			WARN_ON(skb_get_queue_mapping(skb) > 3);
136			WARN_ON(!atomic_read(&queue->pending_frames));
137			atomic_dec(&queue->pending_frames);
138		}
139		skb_queue_head(dropped, skb);
140	}
141}
142
143struct sk_buff *wfx_pending_get(struct wfx_dev *wdev, u32 packet_id)
144{
145	struct wfx_queue *queue;
146	struct wfx_hif_req_tx *req;
147	struct wfx_vif *wvif;
148	struct wfx_hif_msg *hif;
149	struct sk_buff *skb;
150
151	spin_lock_bh(&wdev->tx_pending.lock);
152	skb_queue_walk(&wdev->tx_pending, skb) {
153		hif = (struct wfx_hif_msg *)skb->data;
154		req = (struct wfx_hif_req_tx *)hif->body;
155		if (req->packet_id != packet_id)
156			continue;
157		spin_unlock_bh(&wdev->tx_pending.lock);
158		wvif = wdev_to_wvif(wdev, hif->interface);
159		if (wvif) {
160			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
161			WARN_ON(skb_get_queue_mapping(skb) > 3);
162			WARN_ON(!atomic_read(&queue->pending_frames));
163			atomic_dec(&queue->pending_frames);
164		}
165		skb_unlink(skb, &wdev->tx_pending);
166		return skb;
167	}
168	spin_unlock_bh(&wdev->tx_pending.lock);
169	WARN(1, "cannot find packet in pending queue");
170	return NULL;
171}
172
173void wfx_pending_dump_old_frames(struct wfx_dev *wdev, unsigned int limit_ms)
174{
175	ktime_t now = ktime_get();
176	struct wfx_tx_priv *tx_priv;
177	struct wfx_hif_req_tx *req;
178	struct sk_buff *skb;
179	bool first = true;
180
181	spin_lock_bh(&wdev->tx_pending.lock);
182	skb_queue_walk(&wdev->tx_pending, skb) {
183		tx_priv = wfx_skb_tx_priv(skb);
184		req = wfx_skb_txreq(skb);
185		if (ktime_after(now, ktime_add_ms(tx_priv->xmit_timestamp, limit_ms))) {
186			if (first) {
187				dev_info(wdev->dev, "frames stuck in firmware since %dms or more:\n",
188					 limit_ms);
189				first = false;
190			}
191			dev_info(wdev->dev, "   id %08x sent %lldms ago\n",
192				 req->packet_id, ktime_ms_delta(now, tx_priv->xmit_timestamp));
193		}
194	}
195	spin_unlock_bh(&wdev->tx_pending.lock);
196}
197
198unsigned int wfx_pending_get_pkt_us_delay(struct wfx_dev *wdev, struct sk_buff *skb)
199{
200	ktime_t now = ktime_get();
201	struct wfx_tx_priv *tx_priv = wfx_skb_tx_priv(skb);
202
203	return ktime_us_delta(now, tx_priv->xmit_timestamp);
204}
205
206bool wfx_tx_queues_has_cab(struct wfx_vif *wvif)
207{
208	struct ieee80211_vif *vif = wvif_to_vif(wvif);
209	int i;
210
211	if (vif->type != NL80211_IFTYPE_AP)
212		return false;
213	for (i = 0; i < IEEE80211_NUM_ACS; ++i)
214		/* Note: since only AP can have mcast frames in queue and only one vif can be AP,
215		 * all queued frames has same interface id
216		 */
217		if (!skb_queue_empty_lockless(&wvif->tx_queue[i].cab))
218			return true;
219	return false;
220}
221
222static int wfx_tx_queue_get_weight(struct wfx_queue *queue)
223{
224	return atomic_read(&queue->pending_frames) * queue->priority;
225}
226
227static struct sk_buff *wfx_tx_queues_get_skb(struct wfx_dev *wdev)
228{
229	struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)];
230	int i, j, num_queues = 0;
231	struct wfx_vif *wvif;
232	struct wfx_hif_msg *hif;
233	struct sk_buff *skb;
234
235	/* sort the queues */
236	wvif = NULL;
237	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
238		for (i = 0; i < IEEE80211_NUM_ACS; i++) {
239			WARN_ON(num_queues >= ARRAY_SIZE(queues));
240			queues[num_queues] = &wvif->tx_queue[i];
241			for (j = num_queues; j > 0; j--)
242				if (wfx_tx_queue_get_weight(queues[j]) <
243				    wfx_tx_queue_get_weight(queues[j - 1]))
244					swap(queues[j - 1], queues[j]);
245			num_queues++;
246		}
247	}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
248
249	wvif = NULL;
250	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
251		if (!wvif->after_dtim_tx_allowed)
252			continue;
253		for (i = 0; i < num_queues; i++) {
254			skb = skb_dequeue(&queues[i]->cab);
255			if (!skb)
256				continue;
257			/* Note: since only AP can have mcast frames in queue and only one vif can
258			 * be AP, all queued frames has same interface id
259			 */
260			hif = (struct wfx_hif_msg *)skb->data;
261			WARN_ON(hif->interface != wvif->id);
262			WARN_ON(queues[i] != &wvif->tx_queue[skb_get_queue_mapping(skb)]);
263			atomic_inc(&queues[i]->pending_frames);
264			trace_queues_stats(wdev, queues[i]);
265			return skb;
266		}
267		/* No more multicast to sent */
268		wvif->after_dtim_tx_allowed = false;
269		schedule_work(&wvif->update_tim_work);
270	}
271
272	for (i = 0; i < num_queues; i++) {
273		skb = skb_dequeue(&queues[i]->normal);
274		if (skb) {
275			atomic_inc(&queues[i]->pending_frames);
276			trace_queues_stats(wdev, queues[i]);
277			return skb;
278		}
279	}
280	return NULL;
281}
282
283struct wfx_hif_msg *wfx_tx_queues_get(struct wfx_dev *wdev)
284{
285	struct wfx_tx_priv *tx_priv;
286	struct sk_buff *skb;
287
288	if (atomic_read(&wdev->tx_lock))
289		return NULL;
290	skb = wfx_tx_queues_get_skb(wdev);
291	if (!skb)
292		return NULL;
293	skb_queue_tail(&wdev->tx_pending, skb);
294	wake_up(&wdev->tx_dequeue);
295	tx_priv = wfx_skb_tx_priv(skb);
296	tx_priv->xmit_timestamp = ktime_get();
297	return (struct wfx_hif_msg *)skb->data;
298}