Loading...
1// SPDX-License-Identifier: ISC
2/*
3 * Copyright (c) 2005-2011 Atheros Communications Inc.
4 * Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
5 * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
6 */
7
8#include "core.h"
9#include "hif.h"
10#include "debug.h"
11
12/********/
13/* Send */
14/********/
15
16static void ath10k_htc_control_tx_complete(struct ath10k *ar,
17 struct sk_buff *skb)
18{
19 kfree_skb(skb);
20}
21
22static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
23{
24 struct sk_buff *skb;
25 struct ath10k_skb_cb *skb_cb;
26
27 skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
28 if (!skb)
29 return NULL;
30
31 skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
32 WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
33
34 skb_cb = ATH10K_SKB_CB(skb);
35 memset(skb_cb, 0, sizeof(*skb_cb));
36
37 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
38 return skb;
39}
40
41static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
42 struct sk_buff *skb)
43{
44 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
45
46 if (htc->ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
47 dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
48 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
49}
50
51void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
52 struct sk_buff *skb)
53{
54 struct ath10k *ar = ep->htc->ar;
55 struct ath10k_htc_hdr *hdr;
56
57 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
58 ep->eid, skb);
59
60 /* A corner case where the copy completion is reaching to host but still
61 * copy engine is processing it due to which host unmaps corresponding
62 * memory and causes SMMU fault, hence as workaround adding delay
63 * the unmapping memory to avoid SMMU faults.
64 */
65 if (ar->hw_params.delay_unmap_buffer &&
66 ep->ul_pipe_id == 3)
67 mdelay(2);
68
69 hdr = (struct ath10k_htc_hdr *)skb->data;
70 ath10k_htc_restore_tx_skb(ep->htc, skb);
71
72 if (!ep->ep_ops.ep_tx_complete) {
73 ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
74 dev_kfree_skb_any(skb);
75 return;
76 }
77
78 if (hdr->flags & ATH10K_HTC_FLAG_SEND_BUNDLE) {
79 dev_kfree_skb_any(skb);
80 return;
81 }
82
83 ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
84}
85EXPORT_SYMBOL(ath10k_htc_notify_tx_completion);
86
87static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
88 struct sk_buff *skb)
89{
90 struct ath10k_htc_hdr *hdr;
91
92 hdr = (struct ath10k_htc_hdr *)skb->data;
93 memset(hdr, 0, sizeof(struct ath10k_htc_hdr));
94
95 hdr->eid = ep->eid;
96 hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
97 hdr->flags = 0;
98 if (ep->tx_credit_flow_enabled && !ep->bundle_tx)
99 hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
100
101 spin_lock_bh(&ep->htc->tx_lock);
102 hdr->seq_no = ep->seq_no++;
103 spin_unlock_bh(&ep->htc->tx_lock);
104}
105
106static int ath10k_htc_consume_credit(struct ath10k_htc_ep *ep,
107 unsigned int len,
108 bool consume)
109{
110 struct ath10k_htc *htc = ep->htc;
111 struct ath10k *ar = htc->ar;
112 enum ath10k_htc_ep_id eid = ep->eid;
113 int credits, ret = 0;
114
115 if (!ep->tx_credit_flow_enabled)
116 return 0;
117
118 credits = DIV_ROUND_UP(len, ep->tx_credit_size);
119 spin_lock_bh(&htc->tx_lock);
120
121 if (ep->tx_credits < credits) {
122 ath10k_dbg(ar, ATH10K_DBG_HTC,
123 "htc insufficient credits ep %d required %d available %d consume %d\n",
124 eid, credits, ep->tx_credits, consume);
125 ret = -EAGAIN;
126 goto unlock;
127 }
128
129 if (consume) {
130 ep->tx_credits -= credits;
131 ath10k_dbg(ar, ATH10K_DBG_HTC,
132 "htc ep %d consumed %d credits total %d\n",
133 eid, credits, ep->tx_credits);
134 }
135
136unlock:
137 spin_unlock_bh(&htc->tx_lock);
138 return ret;
139}
140
141static void ath10k_htc_release_credit(struct ath10k_htc_ep *ep, unsigned int len)
142{
143 struct ath10k_htc *htc = ep->htc;
144 struct ath10k *ar = htc->ar;
145 enum ath10k_htc_ep_id eid = ep->eid;
146 int credits;
147
148 if (!ep->tx_credit_flow_enabled)
149 return;
150
151 credits = DIV_ROUND_UP(len, ep->tx_credit_size);
152 spin_lock_bh(&htc->tx_lock);
153 ep->tx_credits += credits;
154 ath10k_dbg(ar, ATH10K_DBG_HTC,
155 "htc ep %d reverted %d credits back total %d\n",
156 eid, credits, ep->tx_credits);
157 spin_unlock_bh(&htc->tx_lock);
158
159 if (ep->ep_ops.ep_tx_credits)
160 ep->ep_ops.ep_tx_credits(htc->ar);
161}
162
163int ath10k_htc_send(struct ath10k_htc *htc,
164 enum ath10k_htc_ep_id eid,
165 struct sk_buff *skb)
166{
167 struct ath10k *ar = htc->ar;
168 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
169 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
170 struct ath10k_hif_sg_item sg_item;
171 struct device *dev = htc->ar->dev;
172 int ret;
173 unsigned int skb_len;
174
175 if (htc->ar->state == ATH10K_STATE_WEDGED)
176 return -ECOMM;
177
178 if (eid >= ATH10K_HTC_EP_COUNT) {
179 ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
180 return -ENOENT;
181 }
182
183 skb_push(skb, sizeof(struct ath10k_htc_hdr));
184
185 skb_len = skb->len;
186 ret = ath10k_htc_consume_credit(ep, skb_len, true);
187 if (ret)
188 goto err_pull;
189
190 ath10k_htc_prepare_tx_skb(ep, skb);
191
192 skb_cb->eid = eid;
193 if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL) {
194 skb_cb->paddr = dma_map_single(dev, skb->data, skb->len,
195 DMA_TO_DEVICE);
196 ret = dma_mapping_error(dev, skb_cb->paddr);
197 if (ret) {
198 ret = -EIO;
199 goto err_credits;
200 }
201 }
202
203 sg_item.transfer_id = ep->eid;
204 sg_item.transfer_context = skb;
205 sg_item.vaddr = skb->data;
206 sg_item.paddr = skb_cb->paddr;
207 sg_item.len = skb->len;
208
209 ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
210 if (ret)
211 goto err_unmap;
212
213 return 0;
214
215err_unmap:
216 if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
217 dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
218err_credits:
219 ath10k_htc_release_credit(ep, skb_len);
220err_pull:
221 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
222 return ret;
223}
224
225void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
226{
227 struct ath10k_htc *htc = &ar->htc;
228 struct ath10k_skb_cb *skb_cb;
229 struct ath10k_htc_ep *ep;
230
231 if (WARN_ON_ONCE(!skb))
232 return;
233
234 skb_cb = ATH10K_SKB_CB(skb);
235 ep = &htc->endpoint[skb_cb->eid];
236
237 ath10k_htc_notify_tx_completion(ep, skb);
238 /* the skb now belongs to the completion handler */
239}
240EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
241
242/***********/
243/* Receive */
244/***********/
245
246static void
247ath10k_htc_process_credit_report(struct ath10k_htc *htc,
248 const struct ath10k_htc_credit_report *report,
249 int len,
250 enum ath10k_htc_ep_id eid)
251{
252 struct ath10k *ar = htc->ar;
253 struct ath10k_htc_ep *ep;
254 int i, n_reports;
255
256 if (len % sizeof(*report))
257 ath10k_warn(ar, "Uneven credit report len %d", len);
258
259 n_reports = len / sizeof(*report);
260
261 spin_lock_bh(&htc->tx_lock);
262 for (i = 0; i < n_reports; i++, report++) {
263 if (report->eid >= ATH10K_HTC_EP_COUNT)
264 break;
265
266 ep = &htc->endpoint[report->eid];
267 ep->tx_credits += report->credits;
268
269 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
270 report->eid, report->credits, ep->tx_credits);
271
272 if (ep->ep_ops.ep_tx_credits) {
273 spin_unlock_bh(&htc->tx_lock);
274 ep->ep_ops.ep_tx_credits(htc->ar);
275 spin_lock_bh(&htc->tx_lock);
276 }
277 }
278 spin_unlock_bh(&htc->tx_lock);
279}
280
281static int
282ath10k_htc_process_lookahead(struct ath10k_htc *htc,
283 const struct ath10k_htc_lookahead_report *report,
284 int len,
285 enum ath10k_htc_ep_id eid,
286 void *next_lookaheads,
287 int *next_lookaheads_len)
288{
289 struct ath10k *ar = htc->ar;
290
291 /* Invalid lookahead flags are actually transmitted by
292 * the target in the HTC control message.
293 * Since this will happen at every boot we silently ignore
294 * the lookahead in this case
295 */
296 if (report->pre_valid != ((~report->post_valid) & 0xFF))
297 return 0;
298
299 if (next_lookaheads && next_lookaheads_len) {
300 ath10k_dbg(ar, ATH10K_DBG_HTC,
301 "htc rx lookahead found pre_valid 0x%x post_valid 0x%x\n",
302 report->pre_valid, report->post_valid);
303
304 /* look ahead bytes are valid, copy them over */
305 memcpy((u8 *)next_lookaheads, report->lookahead, 4);
306
307 *next_lookaheads_len = 1;
308 }
309
310 return 0;
311}
312
313static int
314ath10k_htc_process_lookahead_bundle(struct ath10k_htc *htc,
315 const struct ath10k_htc_lookahead_bundle *report,
316 int len,
317 enum ath10k_htc_ep_id eid,
318 void *next_lookaheads,
319 int *next_lookaheads_len)
320{
321 struct ath10k *ar = htc->ar;
322 int bundle_cnt = len / sizeof(*report);
323
324 if (!bundle_cnt || (bundle_cnt > htc->max_msgs_per_htc_bundle)) {
325 ath10k_warn(ar, "Invalid lookahead bundle count: %d\n",
326 bundle_cnt);
327 return -EINVAL;
328 }
329
330 if (next_lookaheads && next_lookaheads_len) {
331 int i;
332
333 for (i = 0; i < bundle_cnt; i++) {
334 memcpy(((u8 *)next_lookaheads) + 4 * i,
335 report->lookahead, 4);
336 report++;
337 }
338
339 *next_lookaheads_len = bundle_cnt;
340 }
341
342 return 0;
343}
344
345int ath10k_htc_process_trailer(struct ath10k_htc *htc,
346 u8 *buffer,
347 int length,
348 enum ath10k_htc_ep_id src_eid,
349 void *next_lookaheads,
350 int *next_lookaheads_len)
351{
352 struct ath10k_htc_lookahead_bundle *bundle;
353 struct ath10k *ar = htc->ar;
354 int status = 0;
355 struct ath10k_htc_record *record;
356 u8 *orig_buffer;
357 int orig_length;
358 size_t len;
359
360 orig_buffer = buffer;
361 orig_length = length;
362
363 while (length > 0) {
364 record = (struct ath10k_htc_record *)buffer;
365
366 if (length < sizeof(record->hdr)) {
367 status = -EINVAL;
368 break;
369 }
370
371 if (record->hdr.len > length) {
372 /* no room left in buffer for record */
373 ath10k_warn(ar, "Invalid record length: %d\n",
374 record->hdr.len);
375 status = -EINVAL;
376 break;
377 }
378
379 switch (record->hdr.id) {
380 case ATH10K_HTC_RECORD_CREDITS:
381 len = sizeof(struct ath10k_htc_credit_report);
382 if (record->hdr.len < len) {
383 ath10k_warn(ar, "Credit report too long\n");
384 status = -EINVAL;
385 break;
386 }
387 ath10k_htc_process_credit_report(htc,
388 record->credit_report,
389 record->hdr.len,
390 src_eid);
391 break;
392 case ATH10K_HTC_RECORD_LOOKAHEAD:
393 len = sizeof(struct ath10k_htc_lookahead_report);
394 if (record->hdr.len < len) {
395 ath10k_warn(ar, "Lookahead report too long\n");
396 status = -EINVAL;
397 break;
398 }
399 status = ath10k_htc_process_lookahead(htc,
400 record->lookahead_report,
401 record->hdr.len,
402 src_eid,
403 next_lookaheads,
404 next_lookaheads_len);
405 break;
406 case ATH10K_HTC_RECORD_LOOKAHEAD_BUNDLE:
407 bundle = record->lookahead_bundle;
408 status = ath10k_htc_process_lookahead_bundle(htc,
409 bundle,
410 record->hdr.len,
411 src_eid,
412 next_lookaheads,
413 next_lookaheads_len);
414 break;
415 default:
416 ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
417 record->hdr.id, record->hdr.len);
418 break;
419 }
420
421 if (status)
422 break;
423
424 /* multiple records may be present in a trailer */
425 buffer += sizeof(record->hdr) + record->hdr.len;
426 length -= sizeof(record->hdr) + record->hdr.len;
427 }
428
429 if (status)
430 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
431 orig_buffer, orig_length);
432
433 return status;
434}
435EXPORT_SYMBOL(ath10k_htc_process_trailer);
436
437void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
438{
439 int status = 0;
440 struct ath10k_htc *htc = &ar->htc;
441 struct ath10k_htc_hdr *hdr;
442 struct ath10k_htc_ep *ep;
443 u16 payload_len;
444 u32 trailer_len = 0;
445 size_t min_len;
446 u8 eid;
447 bool trailer_present;
448
449 hdr = (struct ath10k_htc_hdr *)skb->data;
450 skb_pull(skb, sizeof(*hdr));
451
452 eid = hdr->eid;
453
454 if (eid >= ATH10K_HTC_EP_COUNT) {
455 ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
456 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
457 hdr, sizeof(*hdr));
458 goto out;
459 }
460
461 ep = &htc->endpoint[eid];
462 if (ep->service_id == ATH10K_HTC_SVC_ID_UNUSED) {
463 ath10k_warn(ar, "htc rx endpoint %d is not connected\n", eid);
464 goto out;
465 }
466
467 payload_len = __le16_to_cpu(hdr->len);
468
469 if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
470 ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
471 payload_len + sizeof(*hdr));
472 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
473 hdr, sizeof(*hdr));
474 goto out;
475 }
476
477 if (skb->len < payload_len) {
478 ath10k_dbg(ar, ATH10K_DBG_HTC,
479 "HTC Rx: insufficient length, got %d, expected %d\n",
480 skb->len, payload_len);
481 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
482 "", hdr, sizeof(*hdr));
483 goto out;
484 }
485
486 /* get flags to check for trailer */
487 trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
488 if (trailer_present) {
489 u8 *trailer;
490
491 trailer_len = hdr->trailer_len;
492 min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
493
494 if ((trailer_len < min_len) ||
495 (trailer_len > payload_len)) {
496 ath10k_warn(ar, "Invalid trailer length: %d\n",
497 trailer_len);
498 goto out;
499 }
500
501 trailer = (u8 *)hdr;
502 trailer += sizeof(*hdr);
503 trailer += payload_len;
504 trailer -= trailer_len;
505 status = ath10k_htc_process_trailer(htc, trailer,
506 trailer_len, hdr->eid,
507 NULL, NULL);
508 if (status)
509 goto out;
510
511 skb_trim(skb, skb->len - trailer_len);
512 }
513
514 if (((int)payload_len - (int)trailer_len) <= 0)
515 /* zero length packet with trailer data, just drop these */
516 goto out;
517
518 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
519 eid, skb);
520 ep->ep_ops.ep_rx_complete(ar, skb);
521
522 /* skb is now owned by the rx completion handler */
523 skb = NULL;
524out:
525 kfree_skb(skb);
526}
527EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
528
529static void ath10k_htc_control_rx_complete(struct ath10k *ar,
530 struct sk_buff *skb)
531{
532 struct ath10k_htc *htc = &ar->htc;
533 struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
534
535 switch (__le16_to_cpu(msg->hdr.message_id)) {
536 case ATH10K_HTC_MSG_READY_ID:
537 case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
538 /* handle HTC control message */
539 if (completion_done(&htc->ctl_resp)) {
540 /* this is a fatal error, target should not be
541 * sending unsolicited messages on the ep 0
542 */
543 ath10k_warn(ar, "HTC rx ctrl still processing\n");
544 complete(&htc->ctl_resp);
545 goto out;
546 }
547
548 htc->control_resp_len =
549 min_t(int, skb->len,
550 ATH10K_HTC_MAX_CTRL_MSG_LEN);
551
552 memcpy(htc->control_resp_buffer, skb->data,
553 htc->control_resp_len);
554
555 complete(&htc->ctl_resp);
556 break;
557 case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
558 htc->htc_ops.target_send_suspend_complete(ar);
559 break;
560 default:
561 ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
562 break;
563 }
564
565out:
566 kfree_skb(skb);
567}
568
569/***************/
570/* Init/Deinit */
571/***************/
572
573static const char *htc_service_name(enum ath10k_htc_svc_id id)
574{
575 switch (id) {
576 case ATH10K_HTC_SVC_ID_RESERVED:
577 return "Reserved";
578 case ATH10K_HTC_SVC_ID_RSVD_CTRL:
579 return "Control";
580 case ATH10K_HTC_SVC_ID_WMI_CONTROL:
581 return "WMI";
582 case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
583 return "DATA BE";
584 case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
585 return "DATA BK";
586 case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
587 return "DATA VI";
588 case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
589 return "DATA VO";
590 case ATH10K_HTC_SVC_ID_NMI_CONTROL:
591 return "NMI Control";
592 case ATH10K_HTC_SVC_ID_NMI_DATA:
593 return "NMI Data";
594 case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
595 return "HTT Data";
596 case ATH10K_HTC_SVC_ID_HTT_DATA2_MSG:
597 return "HTT Data";
598 case ATH10K_HTC_SVC_ID_HTT_DATA3_MSG:
599 return "HTT Data";
600 case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
601 return "RAW";
602 case ATH10K_HTC_SVC_ID_HTT_LOG_MSG:
603 return "PKTLOG";
604 }
605
606 return "Unknown";
607}
608
609static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
610{
611 struct ath10k_htc_ep *ep;
612 int i;
613
614 for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
615 ep = &htc->endpoint[i];
616 ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
617 ep->max_ep_message_len = 0;
618 ep->max_tx_queue_depth = 0;
619 ep->eid = i;
620 ep->htc = htc;
621 ep->tx_credit_flow_enabled = true;
622 }
623}
624
625static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
626 u16 service_id)
627{
628 u8 allocation = 0;
629
630 /* The WMI control service is the only service with flow control.
631 * Let it have all transmit credits.
632 */
633 if (service_id == ATH10K_HTC_SVC_ID_WMI_CONTROL)
634 allocation = htc->total_transmit_credits;
635
636 return allocation;
637}
638
639static int ath10k_htc_send_bundle(struct ath10k_htc_ep *ep,
640 struct sk_buff *bundle_skb,
641 struct sk_buff_head *tx_save_head)
642{
643 struct ath10k_hif_sg_item sg_item;
644 struct ath10k_htc *htc = ep->htc;
645 struct ath10k *ar = htc->ar;
646 struct sk_buff *skb;
647 int ret, cn = 0;
648 unsigned int skb_len;
649
650 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle skb len %d\n", bundle_skb->len);
651 skb_len = bundle_skb->len;
652 ret = ath10k_htc_consume_credit(ep, skb_len, true);
653
654 if (!ret) {
655 sg_item.transfer_id = ep->eid;
656 sg_item.transfer_context = bundle_skb;
657 sg_item.vaddr = bundle_skb->data;
658 sg_item.len = bundle_skb->len;
659
660 ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
661 if (ret)
662 ath10k_htc_release_credit(ep, skb_len);
663 }
664
665 if (ret)
666 dev_kfree_skb_any(bundle_skb);
667
668 for (cn = 0; (skb = skb_dequeue_tail(tx_save_head)); cn++) {
669 if (ret) {
670 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
671 skb_queue_head(&ep->tx_req_head, skb);
672 } else {
673 skb_queue_tail(&ep->tx_complete_head, skb);
674 }
675 }
676
677 if (!ret)
678 queue_work(ar->workqueue_tx_complete, &ar->tx_complete_work);
679
680 ath10k_dbg(ar, ATH10K_DBG_HTC,
681 "bundle tx status %d eid %d req count %d count %d len %d\n",
682 ret, ep->eid, skb_queue_len(&ep->tx_req_head), cn, skb_len);
683 return ret;
684}
685
686static void ath10k_htc_send_one_skb(struct ath10k_htc_ep *ep, struct sk_buff *skb)
687{
688 struct ath10k_htc *htc = ep->htc;
689 struct ath10k *ar = htc->ar;
690 int ret;
691
692 ret = ath10k_htc_send(htc, ep->eid, skb);
693
694 if (ret)
695 skb_queue_head(&ep->tx_req_head, skb);
696
697 ath10k_dbg(ar, ATH10K_DBG_HTC, "tx one status %d eid %d len %d pending count %d\n",
698 ret, ep->eid, skb->len, skb_queue_len(&ep->tx_req_head));
699}
700
701static int ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep *ep)
702{
703 struct ath10k_htc *htc = ep->htc;
704 struct sk_buff *bundle_skb, *skb;
705 struct sk_buff_head tx_save_head;
706 struct ath10k_htc_hdr *hdr;
707 u8 *bundle_buf;
708 int ret = 0, credit_pad, credit_remainder, trans_len, bundles_left = 0;
709
710 if (htc->ar->state == ATH10K_STATE_WEDGED)
711 return -ECOMM;
712
713 if (ep->tx_credit_flow_enabled &&
714 ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE)
715 return 0;
716
717 bundles_left = ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
718 bundle_skb = dev_alloc_skb(bundles_left);
719
720 if (!bundle_skb)
721 return -ENOMEM;
722
723 bundle_buf = bundle_skb->data;
724 skb_queue_head_init(&tx_save_head);
725
726 while (true) {
727 skb = skb_dequeue(&ep->tx_req_head);
728 if (!skb)
729 break;
730
731 credit_pad = 0;
732 trans_len = skb->len + sizeof(*hdr);
733 credit_remainder = trans_len % ep->tx_credit_size;
734
735 if (credit_remainder != 0) {
736 credit_pad = ep->tx_credit_size - credit_remainder;
737 trans_len += credit_pad;
738 }
739
740 ret = ath10k_htc_consume_credit(ep,
741 bundle_buf + trans_len - bundle_skb->data,
742 false);
743 if (ret) {
744 skb_queue_head(&ep->tx_req_head, skb);
745 break;
746 }
747
748 if (bundles_left < trans_len) {
749 bundle_skb->len = bundle_buf - bundle_skb->data;
750 ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
751
752 if (ret) {
753 skb_queue_head(&ep->tx_req_head, skb);
754 return ret;
755 }
756
757 if (skb_queue_len(&ep->tx_req_head) == 0) {
758 ath10k_htc_send_one_skb(ep, skb);
759 return ret;
760 }
761
762 if (ep->tx_credit_flow_enabled &&
763 ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE) {
764 skb_queue_head(&ep->tx_req_head, skb);
765 return 0;
766 }
767
768 bundles_left =
769 ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
770 bundle_skb = dev_alloc_skb(bundles_left);
771
772 if (!bundle_skb) {
773 skb_queue_head(&ep->tx_req_head, skb);
774 return -ENOMEM;
775 }
776 bundle_buf = bundle_skb->data;
777 skb_queue_head_init(&tx_save_head);
778 }
779
780 skb_push(skb, sizeof(struct ath10k_htc_hdr));
781 ath10k_htc_prepare_tx_skb(ep, skb);
782
783 memcpy(bundle_buf, skb->data, skb->len);
784 hdr = (struct ath10k_htc_hdr *)bundle_buf;
785 hdr->flags |= ATH10K_HTC_FLAG_SEND_BUNDLE;
786 hdr->pad_len = __cpu_to_le16(credit_pad);
787 bundle_buf += trans_len;
788 bundles_left -= trans_len;
789 skb_queue_tail(&tx_save_head, skb);
790 }
791
792 if (bundle_buf != bundle_skb->data) {
793 bundle_skb->len = bundle_buf - bundle_skb->data;
794 ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
795 } else {
796 dev_kfree_skb_any(bundle_skb);
797 }
798
799 return ret;
800}
801
802static void ath10k_htc_bundle_tx_work(struct work_struct *work)
803{
804 struct ath10k *ar = container_of(work, struct ath10k, bundle_tx_work);
805 struct ath10k_htc_ep *ep;
806 struct sk_buff *skb;
807 int i;
808
809 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
810 ep = &ar->htc.endpoint[i];
811
812 if (!ep->bundle_tx)
813 continue;
814
815 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx work eid %d count %d\n",
816 ep->eid, skb_queue_len(&ep->tx_req_head));
817
818 if (skb_queue_len(&ep->tx_req_head) >=
819 ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE) {
820 ath10k_htc_send_bundle_skbs(ep);
821 } else {
822 skb = skb_dequeue(&ep->tx_req_head);
823
824 if (!skb)
825 continue;
826 ath10k_htc_send_one_skb(ep, skb);
827 }
828 }
829}
830
831static void ath10k_htc_tx_complete_work(struct work_struct *work)
832{
833 struct ath10k *ar = container_of(work, struct ath10k, tx_complete_work);
834 struct ath10k_htc_ep *ep;
835 enum ath10k_htc_ep_id eid;
836 struct sk_buff *skb;
837 int i;
838
839 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
840 ep = &ar->htc.endpoint[i];
841 eid = ep->eid;
842 if (ep->bundle_tx && eid == ar->htt.eid) {
843 ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx complete eid %d pending complete count%d\n",
844 ep->eid, skb_queue_len(&ep->tx_complete_head));
845
846 while (true) {
847 skb = skb_dequeue(&ep->tx_complete_head);
848 if (!skb)
849 break;
850 ath10k_htc_notify_tx_completion(ep, skb);
851 }
852 }
853 }
854}
855
856int ath10k_htc_send_hl(struct ath10k_htc *htc,
857 enum ath10k_htc_ep_id eid,
858 struct sk_buff *skb)
859{
860 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
861 struct ath10k *ar = htc->ar;
862
863 if (sizeof(struct ath10k_htc_hdr) + skb->len > ep->tx_credit_size) {
864 ath10k_dbg(ar, ATH10K_DBG_HTC, "tx exceed max len %d\n", skb->len);
865 return -ENOMEM;
866 }
867
868 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc send hl eid %d bundle %d tx count %d len %d\n",
869 eid, ep->bundle_tx, skb_queue_len(&ep->tx_req_head), skb->len);
870
871 if (ep->bundle_tx) {
872 skb_queue_tail(&ep->tx_req_head, skb);
873 queue_work(ar->workqueue, &ar->bundle_tx_work);
874 return 0;
875 } else {
876 return ath10k_htc_send(htc, eid, skb);
877 }
878}
879
880void ath10k_htc_setup_tx_req(struct ath10k_htc_ep *ep)
881{
882 if (ep->htc->max_msgs_per_htc_bundle >= ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE &&
883 !ep->bundle_tx) {
884 ep->bundle_tx = true;
885 skb_queue_head_init(&ep->tx_req_head);
886 skb_queue_head_init(&ep->tx_complete_head);
887 }
888}
889
890void ath10k_htc_stop_hl(struct ath10k *ar)
891{
892 struct ath10k_htc_ep *ep;
893 int i;
894
895 cancel_work_sync(&ar->bundle_tx_work);
896 cancel_work_sync(&ar->tx_complete_work);
897
898 for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
899 ep = &ar->htc.endpoint[i];
900
901 if (!ep->bundle_tx)
902 continue;
903
904 ath10k_dbg(ar, ATH10K_DBG_HTC, "stop tx work eid %d count %d\n",
905 ep->eid, skb_queue_len(&ep->tx_req_head));
906
907 skb_queue_purge(&ep->tx_req_head);
908 }
909}
910
911int ath10k_htc_wait_target(struct ath10k_htc *htc)
912{
913 struct ath10k *ar = htc->ar;
914 int i, status = 0;
915 unsigned long time_left;
916 struct ath10k_htc_msg *msg;
917 u16 message_id;
918
919 time_left = wait_for_completion_timeout(&htc->ctl_resp,
920 ATH10K_HTC_WAIT_TIMEOUT_HZ);
921 if (!time_left) {
922 /* Workaround: In some cases the PCI HIF doesn't
923 * receive interrupt for the control response message
924 * even if the buffer was completed. It is suspected
925 * iomap writes unmasking PCI CE irqs aren't propagated
926 * properly in KVM PCI-passthrough sometimes.
927 */
928 ath10k_warn(ar, "failed to receive control response completion, polling..\n");
929
930 for (i = 0; i < CE_COUNT; i++)
931 ath10k_hif_send_complete_check(htc->ar, i, 1);
932
933 time_left =
934 wait_for_completion_timeout(&htc->ctl_resp,
935 ATH10K_HTC_WAIT_TIMEOUT_HZ);
936
937 if (!time_left)
938 status = -ETIMEDOUT;
939 }
940
941 if (status < 0) {
942 ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
943 return status;
944 }
945
946 if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
947 ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
948 htc->control_resp_len);
949 return -ECOMM;
950 }
951
952 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
953 message_id = __le16_to_cpu(msg->hdr.message_id);
954
955 if (message_id != ATH10K_HTC_MSG_READY_ID) {
956 ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
957 return -ECOMM;
958 }
959
960 if (ar->hw_params.use_fw_tx_credits)
961 htc->total_transmit_credits = __le16_to_cpu(msg->ready.credit_count);
962 else
963 htc->total_transmit_credits = 1;
964
965 htc->target_credit_size = __le16_to_cpu(msg->ready.credit_size);
966
967 ath10k_dbg(ar, ATH10K_DBG_HTC,
968 "Target ready! transmit resources: %d size:%d actual credits:%d\n",
969 htc->total_transmit_credits,
970 htc->target_credit_size,
971 msg->ready.credit_count);
972
973 if ((htc->total_transmit_credits == 0) ||
974 (htc->target_credit_size == 0)) {
975 ath10k_err(ar, "Invalid credit size received\n");
976 return -ECOMM;
977 }
978
979 /* The only way to determine if the ready message is an extended
980 * message is from the size.
981 */
982 if (htc->control_resp_len >=
983 sizeof(msg->hdr) + sizeof(msg->ready_ext)) {
984 htc->alt_data_credit_size =
985 __le16_to_cpu(msg->ready_ext.reserved) &
986 ATH10K_HTC_MSG_READY_EXT_ALT_DATA_MASK;
987 htc->max_msgs_per_htc_bundle =
988 min_t(u8, msg->ready_ext.max_msgs_per_htc_bundle,
989 HTC_HOST_MAX_MSG_PER_RX_BUNDLE);
990 ath10k_dbg(ar, ATH10K_DBG_HTC,
991 "Extended ready message RX bundle size %d alt size %d\n",
992 htc->max_msgs_per_htc_bundle,
993 htc->alt_data_credit_size);
994 }
995
996 INIT_WORK(&ar->bundle_tx_work, ath10k_htc_bundle_tx_work);
997 INIT_WORK(&ar->tx_complete_work, ath10k_htc_tx_complete_work);
998
999 return 0;
1000}
1001
1002void ath10k_htc_change_tx_credit_flow(struct ath10k_htc *htc,
1003 enum ath10k_htc_ep_id eid,
1004 bool enable)
1005{
1006 struct ath10k *ar = htc->ar;
1007 struct ath10k_htc_ep *ep = &ar->htc.endpoint[eid];
1008
1009 ep->tx_credit_flow_enabled = enable;
1010}
1011
1012int ath10k_htc_connect_service(struct ath10k_htc *htc,
1013 struct ath10k_htc_svc_conn_req *conn_req,
1014 struct ath10k_htc_svc_conn_resp *conn_resp)
1015{
1016 struct ath10k *ar = htc->ar;
1017 struct ath10k_htc_msg *msg;
1018 struct ath10k_htc_conn_svc *req_msg;
1019 struct ath10k_htc_conn_svc_response resp_msg_dummy;
1020 struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
1021 enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
1022 struct ath10k_htc_ep *ep;
1023 struct sk_buff *skb;
1024 unsigned int max_msg_size = 0;
1025 int length, status;
1026 unsigned long time_left;
1027 bool disable_credit_flow_ctrl = false;
1028 u16 message_id, service_id, flags = 0;
1029 u8 tx_alloc = 0;
1030
1031 /* special case for HTC pseudo control service */
1032 if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
1033 disable_credit_flow_ctrl = true;
1034 assigned_eid = ATH10K_HTC_EP_0;
1035 max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
1036 memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
1037 goto setup;
1038 }
1039
1040 tx_alloc = ath10k_htc_get_credit_allocation(htc,
1041 conn_req->service_id);
1042 if (!tx_alloc)
1043 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1044 "boot htc service %s does not allocate target credits\n",
1045 htc_service_name(conn_req->service_id));
1046
1047 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1048 if (!skb) {
1049 ath10k_err(ar, "Failed to allocate HTC packet\n");
1050 return -ENOMEM;
1051 }
1052
1053 length = sizeof(msg->hdr) + sizeof(msg->connect_service);
1054 skb_put(skb, length);
1055 memset(skb->data, 0, length);
1056
1057 msg = (struct ath10k_htc_msg *)skb->data;
1058 msg->hdr.message_id =
1059 __cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
1060
1061 flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
1062
1063 /* Only enable credit flow control for WMI ctrl service */
1064 if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
1065 flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
1066 disable_credit_flow_ctrl = true;
1067 }
1068
1069 req_msg = &msg->connect_service;
1070 req_msg->flags = __cpu_to_le16(flags);
1071 req_msg->service_id = __cpu_to_le16(conn_req->service_id);
1072
1073 reinit_completion(&htc->ctl_resp);
1074
1075 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1076 if (status) {
1077 kfree_skb(skb);
1078 return status;
1079 }
1080
1081 /* wait for response */
1082 time_left = wait_for_completion_timeout(&htc->ctl_resp,
1083 ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
1084 if (!time_left) {
1085 ath10k_err(ar, "Service connect timeout\n");
1086 return -ETIMEDOUT;
1087 }
1088
1089 /* we controlled the buffer creation, it's aligned */
1090 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
1091 resp_msg = &msg->connect_service_response;
1092 message_id = __le16_to_cpu(msg->hdr.message_id);
1093 service_id = __le16_to_cpu(resp_msg->service_id);
1094
1095 if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
1096 (htc->control_resp_len < sizeof(msg->hdr) +
1097 sizeof(msg->connect_service_response))) {
1098 ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
1099 return -EPROTO;
1100 }
1101
1102 ath10k_dbg(ar, ATH10K_DBG_HTC,
1103 "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
1104 htc_service_name(service_id),
1105 resp_msg->status, resp_msg->eid);
1106
1107 conn_resp->connect_resp_code = resp_msg->status;
1108
1109 /* check response status */
1110 if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
1111 ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
1112 htc_service_name(service_id),
1113 resp_msg->status);
1114 return -EPROTO;
1115 }
1116
1117 assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
1118 max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
1119
1120setup:
1121
1122 if (assigned_eid >= ATH10K_HTC_EP_COUNT)
1123 return -EPROTO;
1124
1125 if (max_msg_size == 0)
1126 return -EPROTO;
1127
1128 ep = &htc->endpoint[assigned_eid];
1129 ep->eid = assigned_eid;
1130
1131 if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
1132 return -EPROTO;
1133
1134 /* return assigned endpoint to caller */
1135 conn_resp->eid = assigned_eid;
1136 conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
1137
1138 /* setup the endpoint */
1139 ep->service_id = conn_req->service_id;
1140 ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
1141 ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
1142 ep->tx_credits = tx_alloc;
1143 ep->tx_credit_size = htc->target_credit_size;
1144
1145 if (conn_req->service_id == ATH10K_HTC_SVC_ID_HTT_DATA_MSG &&
1146 htc->alt_data_credit_size != 0)
1147 ep->tx_credit_size = htc->alt_data_credit_size;
1148
1149 /* copy all the callbacks */
1150 ep->ep_ops = conn_req->ep_ops;
1151
1152 status = ath10k_hif_map_service_to_pipe(htc->ar,
1153 ep->service_id,
1154 &ep->ul_pipe_id,
1155 &ep->dl_pipe_id);
1156 if (status) {
1157 ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC service id: %d\n",
1158 ep->service_id);
1159 return status;
1160 }
1161
1162 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1163 "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
1164 htc_service_name(ep->service_id), ep->ul_pipe_id,
1165 ep->dl_pipe_id, ep->eid);
1166
1167 if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
1168 ep->tx_credit_flow_enabled = false;
1169 ath10k_dbg(ar, ATH10K_DBG_BOOT,
1170 "boot htc service '%s' eid %d TX flow control disabled\n",
1171 htc_service_name(ep->service_id), assigned_eid);
1172 }
1173
1174 return status;
1175}
1176
1177struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
1178{
1179 struct sk_buff *skb;
1180
1181 skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
1182 if (!skb)
1183 return NULL;
1184
1185 skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
1186
1187 /* FW/HTC requires 4-byte aligned streams */
1188 if (!IS_ALIGNED((unsigned long)skb->data, 4))
1189 ath10k_warn(ar, "Unaligned HTC tx skb\n");
1190
1191 return skb;
1192}
1193
1194static void ath10k_htc_pktlog_process_rx(struct ath10k *ar, struct sk_buff *skb)
1195{
1196 trace_ath10k_htt_pktlog(ar, skb->data, skb->len);
1197 dev_kfree_skb_any(skb);
1198}
1199
1200static int ath10k_htc_pktlog_connect(struct ath10k *ar)
1201{
1202 struct ath10k_htc_svc_conn_resp conn_resp;
1203 struct ath10k_htc_svc_conn_req conn_req;
1204 int status;
1205
1206 memset(&conn_req, 0, sizeof(conn_req));
1207 memset(&conn_resp, 0, sizeof(conn_resp));
1208
1209 conn_req.ep_ops.ep_tx_complete = NULL;
1210 conn_req.ep_ops.ep_rx_complete = ath10k_htc_pktlog_process_rx;
1211 conn_req.ep_ops.ep_tx_credits = NULL;
1212
1213 /* connect to control service */
1214 conn_req.service_id = ATH10K_HTC_SVC_ID_HTT_LOG_MSG;
1215 status = ath10k_htc_connect_service(&ar->htc, &conn_req, &conn_resp);
1216 if (status) {
1217 ath10k_warn(ar, "failed to connect to PKTLOG service: %d\n",
1218 status);
1219 return status;
1220 }
1221
1222 return 0;
1223}
1224
1225static bool ath10k_htc_pktlog_svc_supported(struct ath10k *ar)
1226{
1227 u8 ul_pipe_id;
1228 u8 dl_pipe_id;
1229 int status;
1230
1231 status = ath10k_hif_map_service_to_pipe(ar, ATH10K_HTC_SVC_ID_HTT_LOG_MSG,
1232 &ul_pipe_id,
1233 &dl_pipe_id);
1234 if (status) {
1235 ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC pktlog service id: %d\n",
1236 ATH10K_HTC_SVC_ID_HTT_LOG_MSG);
1237
1238 return false;
1239 }
1240
1241 return true;
1242}
1243
1244int ath10k_htc_start(struct ath10k_htc *htc)
1245{
1246 struct ath10k *ar = htc->ar;
1247 struct sk_buff *skb;
1248 int status = 0;
1249 struct ath10k_htc_msg *msg;
1250
1251 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1252 if (!skb)
1253 return -ENOMEM;
1254
1255 skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
1256 memset(skb->data, 0, skb->len);
1257
1258 msg = (struct ath10k_htc_msg *)skb->data;
1259 msg->hdr.message_id =
1260 __cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
1261
1262 if (ar->hif.bus == ATH10K_BUS_SDIO) {
1263 /* Extra setup params used by SDIO */
1264 msg->setup_complete_ext.flags =
1265 __cpu_to_le32(ATH10K_HTC_SETUP_COMPLETE_FLAGS_RX_BNDL_EN);
1266 msg->setup_complete_ext.max_msgs_per_bundled_recv =
1267 htc->max_msgs_per_htc_bundle;
1268 }
1269 ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
1270
1271 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1272 if (status) {
1273 kfree_skb(skb);
1274 return status;
1275 }
1276
1277 if (ath10k_htc_pktlog_svc_supported(ar)) {
1278 status = ath10k_htc_pktlog_connect(ar);
1279 if (status) {
1280 ath10k_err(ar, "failed to connect to pktlog: %d\n", status);
1281 return status;
1282 }
1283 }
1284
1285 return 0;
1286}
1287
1288/* registered target arrival callback from the HIF layer */
1289int ath10k_htc_init(struct ath10k *ar)
1290{
1291 int status;
1292 struct ath10k_htc *htc = &ar->htc;
1293 struct ath10k_htc_svc_conn_req conn_req;
1294 struct ath10k_htc_svc_conn_resp conn_resp;
1295
1296 spin_lock_init(&htc->tx_lock);
1297
1298 ath10k_htc_reset_endpoint_states(htc);
1299
1300 htc->ar = ar;
1301
1302 /* setup our pseudo HTC control endpoint connection */
1303 memset(&conn_req, 0, sizeof(conn_req));
1304 memset(&conn_resp, 0, sizeof(conn_resp));
1305 conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
1306 conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
1307 conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
1308 conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
1309
1310 /* connect fake service */
1311 status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
1312 if (status) {
1313 ath10k_err(ar, "could not connect to htc service (%d)\n",
1314 status);
1315 return status;
1316 }
1317
1318 init_completion(&htc->ctl_resp);
1319
1320 return 0;
1321}
1/*
2 * Copyright (c) 2005-2011 Atheros Communications Inc.
3 * Copyright (c) 2011-2013 Qualcomm Atheros, Inc.
4 *
5 * Permission to use, copy, modify, and/or distribute this software for any
6 * purpose with or without fee is hereby granted, provided that the above
7 * copyright notice and this permission notice appear in all copies.
8 *
9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
16 */
17
18#include "core.h"
19#include "hif.h"
20#include "debug.h"
21
22/********/
23/* Send */
24/********/
25
26static void ath10k_htc_control_tx_complete(struct ath10k *ar,
27 struct sk_buff *skb)
28{
29 kfree_skb(skb);
30}
31
32static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
33{
34 struct sk_buff *skb;
35 struct ath10k_skb_cb *skb_cb;
36
37 skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
38 if (!skb)
39 return NULL;
40
41 skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
42 WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
43
44 skb_cb = ATH10K_SKB_CB(skb);
45 memset(skb_cb, 0, sizeof(*skb_cb));
46
47 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
48 return skb;
49}
50
51static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
52 struct sk_buff *skb)
53{
54 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
55
56 dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
57 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
58}
59
60static void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
61 struct sk_buff *skb)
62{
63 struct ath10k *ar = ep->htc->ar;
64
65 ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
66 ep->eid, skb);
67
68 ath10k_htc_restore_tx_skb(ep->htc, skb);
69
70 if (!ep->ep_ops.ep_tx_complete) {
71 ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
72 dev_kfree_skb_any(skb);
73 return;
74 }
75
76 ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
77}
78
79static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
80 struct sk_buff *skb)
81{
82 struct ath10k_htc_hdr *hdr;
83
84 hdr = (struct ath10k_htc_hdr *)skb->data;
85
86 hdr->eid = ep->eid;
87 hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
88 hdr->flags = 0;
89 hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
90
91 spin_lock_bh(&ep->htc->tx_lock);
92 hdr->seq_no = ep->seq_no++;
93 spin_unlock_bh(&ep->htc->tx_lock);
94}
95
96int ath10k_htc_send(struct ath10k_htc *htc,
97 enum ath10k_htc_ep_id eid,
98 struct sk_buff *skb)
99{
100 struct ath10k *ar = htc->ar;
101 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
102 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
103 struct ath10k_hif_sg_item sg_item;
104 struct device *dev = htc->ar->dev;
105 int credits = 0;
106 int ret;
107
108 if (htc->ar->state == ATH10K_STATE_WEDGED)
109 return -ECOMM;
110
111 if (eid >= ATH10K_HTC_EP_COUNT) {
112 ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
113 return -ENOENT;
114 }
115
116 skb_push(skb, sizeof(struct ath10k_htc_hdr));
117
118 if (ep->tx_credit_flow_enabled) {
119 credits = DIV_ROUND_UP(skb->len, htc->target_credit_size);
120 spin_lock_bh(&htc->tx_lock);
121 if (ep->tx_credits < credits) {
122 spin_unlock_bh(&htc->tx_lock);
123 ret = -EAGAIN;
124 goto err_pull;
125 }
126 ep->tx_credits -= credits;
127 ath10k_dbg(ar, ATH10K_DBG_HTC,
128 "htc ep %d consumed %d credits (total %d)\n",
129 eid, credits, ep->tx_credits);
130 spin_unlock_bh(&htc->tx_lock);
131 }
132
133 ath10k_htc_prepare_tx_skb(ep, skb);
134
135 skb_cb->eid = eid;
136 skb_cb->paddr = dma_map_single(dev, skb->data, skb->len, DMA_TO_DEVICE);
137 ret = dma_mapping_error(dev, skb_cb->paddr);
138 if (ret) {
139 ret = -EIO;
140 goto err_credits;
141 }
142
143 sg_item.transfer_id = ep->eid;
144 sg_item.transfer_context = skb;
145 sg_item.vaddr = skb->data;
146 sg_item.paddr = skb_cb->paddr;
147 sg_item.len = skb->len;
148
149 ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
150 if (ret)
151 goto err_unmap;
152
153 return 0;
154
155err_unmap:
156 dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
157err_credits:
158 if (ep->tx_credit_flow_enabled) {
159 spin_lock_bh(&htc->tx_lock);
160 ep->tx_credits += credits;
161 ath10k_dbg(ar, ATH10K_DBG_HTC,
162 "htc ep %d reverted %d credits back (total %d)\n",
163 eid, credits, ep->tx_credits);
164 spin_unlock_bh(&htc->tx_lock);
165
166 if (ep->ep_ops.ep_tx_credits)
167 ep->ep_ops.ep_tx_credits(htc->ar);
168 }
169err_pull:
170 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
171 return ret;
172}
173
174void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
175{
176 struct ath10k_htc *htc = &ar->htc;
177 struct ath10k_skb_cb *skb_cb;
178 struct ath10k_htc_ep *ep;
179
180 if (WARN_ON_ONCE(!skb))
181 return;
182
183 skb_cb = ATH10K_SKB_CB(skb);
184 ep = &htc->endpoint[skb_cb->eid];
185
186 ath10k_htc_notify_tx_completion(ep, skb);
187 /* the skb now belongs to the completion handler */
188}
189EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
190
191/***********/
192/* Receive */
193/***********/
194
195static void
196ath10k_htc_process_credit_report(struct ath10k_htc *htc,
197 const struct ath10k_htc_credit_report *report,
198 int len,
199 enum ath10k_htc_ep_id eid)
200{
201 struct ath10k *ar = htc->ar;
202 struct ath10k_htc_ep *ep;
203 int i, n_reports;
204
205 if (len % sizeof(*report))
206 ath10k_warn(ar, "Uneven credit report len %d", len);
207
208 n_reports = len / sizeof(*report);
209
210 spin_lock_bh(&htc->tx_lock);
211 for (i = 0; i < n_reports; i++, report++) {
212 if (report->eid >= ATH10K_HTC_EP_COUNT)
213 break;
214
215 ep = &htc->endpoint[report->eid];
216 ep->tx_credits += report->credits;
217
218 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
219 report->eid, report->credits, ep->tx_credits);
220
221 if (ep->ep_ops.ep_tx_credits) {
222 spin_unlock_bh(&htc->tx_lock);
223 ep->ep_ops.ep_tx_credits(htc->ar);
224 spin_lock_bh(&htc->tx_lock);
225 }
226 }
227 spin_unlock_bh(&htc->tx_lock);
228}
229
230static int ath10k_htc_process_trailer(struct ath10k_htc *htc,
231 u8 *buffer,
232 int length,
233 enum ath10k_htc_ep_id src_eid)
234{
235 struct ath10k *ar = htc->ar;
236 int status = 0;
237 struct ath10k_htc_record *record;
238 u8 *orig_buffer;
239 int orig_length;
240 size_t len;
241
242 orig_buffer = buffer;
243 orig_length = length;
244
245 while (length > 0) {
246 record = (struct ath10k_htc_record *)buffer;
247
248 if (length < sizeof(record->hdr)) {
249 status = -EINVAL;
250 break;
251 }
252
253 if (record->hdr.len > length) {
254 /* no room left in buffer for record */
255 ath10k_warn(ar, "Invalid record length: %d\n",
256 record->hdr.len);
257 status = -EINVAL;
258 break;
259 }
260
261 switch (record->hdr.id) {
262 case ATH10K_HTC_RECORD_CREDITS:
263 len = sizeof(struct ath10k_htc_credit_report);
264 if (record->hdr.len < len) {
265 ath10k_warn(ar, "Credit report too long\n");
266 status = -EINVAL;
267 break;
268 }
269 ath10k_htc_process_credit_report(htc,
270 record->credit_report,
271 record->hdr.len,
272 src_eid);
273 break;
274 default:
275 ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
276 record->hdr.id, record->hdr.len);
277 break;
278 }
279
280 if (status)
281 break;
282
283 /* multiple records may be present in a trailer */
284 buffer += sizeof(record->hdr) + record->hdr.len;
285 length -= sizeof(record->hdr) + record->hdr.len;
286 }
287
288 if (status)
289 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
290 orig_buffer, orig_length);
291
292 return status;
293}
294
295void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
296{
297 int status = 0;
298 struct ath10k_htc *htc = &ar->htc;
299 struct ath10k_htc_hdr *hdr;
300 struct ath10k_htc_ep *ep;
301 u16 payload_len;
302 u32 trailer_len = 0;
303 size_t min_len;
304 u8 eid;
305 bool trailer_present;
306
307 hdr = (struct ath10k_htc_hdr *)skb->data;
308 skb_pull(skb, sizeof(*hdr));
309
310 eid = hdr->eid;
311
312 if (eid >= ATH10K_HTC_EP_COUNT) {
313 ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
314 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
315 hdr, sizeof(*hdr));
316 goto out;
317 }
318
319 ep = &htc->endpoint[eid];
320
321 payload_len = __le16_to_cpu(hdr->len);
322
323 if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
324 ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
325 payload_len + sizeof(*hdr));
326 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
327 hdr, sizeof(*hdr));
328 goto out;
329 }
330
331 if (skb->len < payload_len) {
332 ath10k_dbg(ar, ATH10K_DBG_HTC,
333 "HTC Rx: insufficient length, got %d, expected %d\n",
334 skb->len, payload_len);
335 ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
336 "", hdr, sizeof(*hdr));
337 goto out;
338 }
339
340 /* get flags to check for trailer */
341 trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
342 if (trailer_present) {
343 u8 *trailer;
344
345 trailer_len = hdr->trailer_len;
346 min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
347
348 if ((trailer_len < min_len) ||
349 (trailer_len > payload_len)) {
350 ath10k_warn(ar, "Invalid trailer length: %d\n",
351 trailer_len);
352 goto out;
353 }
354
355 trailer = (u8 *)hdr;
356 trailer += sizeof(*hdr);
357 trailer += payload_len;
358 trailer -= trailer_len;
359 status = ath10k_htc_process_trailer(htc, trailer,
360 trailer_len, hdr->eid);
361 if (status)
362 goto out;
363
364 skb_trim(skb, skb->len - trailer_len);
365 }
366
367 if (((int)payload_len - (int)trailer_len) <= 0)
368 /* zero length packet with trailer data, just drop these */
369 goto out;
370
371 if (eid == ATH10K_HTC_EP_0) {
372 struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
373
374 switch (__le16_to_cpu(msg->hdr.message_id)) {
375 case ATH10K_HTC_MSG_READY_ID:
376 case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
377 /* handle HTC control message */
378 if (completion_done(&htc->ctl_resp)) {
379 /*
380 * this is a fatal error, target should not be
381 * sending unsolicited messages on the ep 0
382 */
383 ath10k_warn(ar, "HTC rx ctrl still processing\n");
384 complete(&htc->ctl_resp);
385 goto out;
386 }
387
388 htc->control_resp_len =
389 min_t(int, skb->len,
390 ATH10K_HTC_MAX_CTRL_MSG_LEN);
391
392 memcpy(htc->control_resp_buffer, skb->data,
393 htc->control_resp_len);
394
395 complete(&htc->ctl_resp);
396 break;
397 case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
398 htc->htc_ops.target_send_suspend_complete(ar);
399 break;
400 default:
401 ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
402 break;
403 }
404 goto out;
405 }
406
407 ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
408 eid, skb);
409 ep->ep_ops.ep_rx_complete(ar, skb);
410
411 /* skb is now owned by the rx completion handler */
412 skb = NULL;
413out:
414 kfree_skb(skb);
415}
416EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
417
418static void ath10k_htc_control_rx_complete(struct ath10k *ar,
419 struct sk_buff *skb)
420{
421 /* This is unexpected. FW is not supposed to send regular rx on this
422 * endpoint. */
423 ath10k_warn(ar, "unexpected htc rx\n");
424 kfree_skb(skb);
425}
426
427/***************/
428/* Init/Deinit */
429/***************/
430
431static const char *htc_service_name(enum ath10k_htc_svc_id id)
432{
433 switch (id) {
434 case ATH10K_HTC_SVC_ID_RESERVED:
435 return "Reserved";
436 case ATH10K_HTC_SVC_ID_RSVD_CTRL:
437 return "Control";
438 case ATH10K_HTC_SVC_ID_WMI_CONTROL:
439 return "WMI";
440 case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
441 return "DATA BE";
442 case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
443 return "DATA BK";
444 case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
445 return "DATA VI";
446 case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
447 return "DATA VO";
448 case ATH10K_HTC_SVC_ID_NMI_CONTROL:
449 return "NMI Control";
450 case ATH10K_HTC_SVC_ID_NMI_DATA:
451 return "NMI Data";
452 case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
453 return "HTT Data";
454 case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
455 return "RAW";
456 }
457
458 return "Unknown";
459}
460
461static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
462{
463 struct ath10k_htc_ep *ep;
464 int i;
465
466 for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
467 ep = &htc->endpoint[i];
468 ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
469 ep->max_ep_message_len = 0;
470 ep->max_tx_queue_depth = 0;
471 ep->eid = i;
472 ep->htc = htc;
473 ep->tx_credit_flow_enabled = true;
474 }
475}
476
477static void ath10k_htc_setup_target_buffer_assignments(struct ath10k_htc *htc)
478{
479 struct ath10k_htc_svc_tx_credits *entry;
480
481 entry = &htc->service_tx_alloc[0];
482
483 /*
484 * for PCIE allocate all credists/HTC buffers to WMI.
485 * no buffers are used/required for data. data always
486 * remains on host.
487 */
488 entry++;
489 entry->service_id = ATH10K_HTC_SVC_ID_WMI_CONTROL;
490 entry->credit_allocation = htc->total_transmit_credits;
491}
492
493static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
494 u16 service_id)
495{
496 u8 allocation = 0;
497 int i;
498
499 for (i = 0; i < ATH10K_HTC_EP_COUNT; i++) {
500 if (htc->service_tx_alloc[i].service_id == service_id)
501 allocation =
502 htc->service_tx_alloc[i].credit_allocation;
503 }
504
505 return allocation;
506}
507
508int ath10k_htc_wait_target(struct ath10k_htc *htc)
509{
510 struct ath10k *ar = htc->ar;
511 int i, status = 0;
512 unsigned long time_left;
513 struct ath10k_htc_svc_conn_req conn_req;
514 struct ath10k_htc_svc_conn_resp conn_resp;
515 struct ath10k_htc_msg *msg;
516 u16 message_id;
517 u16 credit_count;
518 u16 credit_size;
519
520 time_left = wait_for_completion_timeout(&htc->ctl_resp,
521 ATH10K_HTC_WAIT_TIMEOUT_HZ);
522 if (!time_left) {
523 /* Workaround: In some cases the PCI HIF doesn't
524 * receive interrupt for the control response message
525 * even if the buffer was completed. It is suspected
526 * iomap writes unmasking PCI CE irqs aren't propagated
527 * properly in KVM PCI-passthrough sometimes.
528 */
529 ath10k_warn(ar, "failed to receive control response completion, polling..\n");
530
531 for (i = 0; i < CE_COUNT; i++)
532 ath10k_hif_send_complete_check(htc->ar, i, 1);
533
534 time_left =
535 wait_for_completion_timeout(&htc->ctl_resp,
536 ATH10K_HTC_WAIT_TIMEOUT_HZ);
537
538 if (!time_left)
539 status = -ETIMEDOUT;
540 }
541
542 if (status < 0) {
543 ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
544 return status;
545 }
546
547 if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
548 ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
549 htc->control_resp_len);
550 return -ECOMM;
551 }
552
553 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
554 message_id = __le16_to_cpu(msg->hdr.message_id);
555 credit_count = __le16_to_cpu(msg->ready.credit_count);
556 credit_size = __le16_to_cpu(msg->ready.credit_size);
557
558 if (message_id != ATH10K_HTC_MSG_READY_ID) {
559 ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
560 return -ECOMM;
561 }
562
563 htc->total_transmit_credits = credit_count;
564 htc->target_credit_size = credit_size;
565
566 ath10k_dbg(ar, ATH10K_DBG_HTC,
567 "Target ready! transmit resources: %d size:%d\n",
568 htc->total_transmit_credits,
569 htc->target_credit_size);
570
571 if ((htc->total_transmit_credits == 0) ||
572 (htc->target_credit_size == 0)) {
573 ath10k_err(ar, "Invalid credit size received\n");
574 return -ECOMM;
575 }
576
577 ath10k_htc_setup_target_buffer_assignments(htc);
578
579 /* setup our pseudo HTC control endpoint connection */
580 memset(&conn_req, 0, sizeof(conn_req));
581 memset(&conn_resp, 0, sizeof(conn_resp));
582 conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
583 conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
584 conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
585 conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
586
587 /* connect fake service */
588 status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
589 if (status) {
590 ath10k_err(ar, "could not connect to htc service (%d)\n",
591 status);
592 return status;
593 }
594
595 return 0;
596}
597
598int ath10k_htc_connect_service(struct ath10k_htc *htc,
599 struct ath10k_htc_svc_conn_req *conn_req,
600 struct ath10k_htc_svc_conn_resp *conn_resp)
601{
602 struct ath10k *ar = htc->ar;
603 struct ath10k_htc_msg *msg;
604 struct ath10k_htc_conn_svc *req_msg;
605 struct ath10k_htc_conn_svc_response resp_msg_dummy;
606 struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
607 enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
608 struct ath10k_htc_ep *ep;
609 struct sk_buff *skb;
610 unsigned int max_msg_size = 0;
611 int length, status;
612 unsigned long time_left;
613 bool disable_credit_flow_ctrl = false;
614 u16 message_id, service_id, flags = 0;
615 u8 tx_alloc = 0;
616
617 /* special case for HTC pseudo control service */
618 if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
619 disable_credit_flow_ctrl = true;
620 assigned_eid = ATH10K_HTC_EP_0;
621 max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
622 memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
623 goto setup;
624 }
625
626 tx_alloc = ath10k_htc_get_credit_allocation(htc,
627 conn_req->service_id);
628 if (!tx_alloc)
629 ath10k_dbg(ar, ATH10K_DBG_BOOT,
630 "boot htc service %s does not allocate target credits\n",
631 htc_service_name(conn_req->service_id));
632
633 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
634 if (!skb) {
635 ath10k_err(ar, "Failed to allocate HTC packet\n");
636 return -ENOMEM;
637 }
638
639 length = sizeof(msg->hdr) + sizeof(msg->connect_service);
640 skb_put(skb, length);
641 memset(skb->data, 0, length);
642
643 msg = (struct ath10k_htc_msg *)skb->data;
644 msg->hdr.message_id =
645 __cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
646
647 flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
648
649 /* Only enable credit flow control for WMI ctrl service */
650 if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
651 flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
652 disable_credit_flow_ctrl = true;
653 }
654
655 req_msg = &msg->connect_service;
656 req_msg->flags = __cpu_to_le16(flags);
657 req_msg->service_id = __cpu_to_le16(conn_req->service_id);
658
659 reinit_completion(&htc->ctl_resp);
660
661 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
662 if (status) {
663 kfree_skb(skb);
664 return status;
665 }
666
667 /* wait for response */
668 time_left = wait_for_completion_timeout(&htc->ctl_resp,
669 ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
670 if (!time_left) {
671 ath10k_err(ar, "Service connect timeout\n");
672 return -ETIMEDOUT;
673 }
674
675 /* we controlled the buffer creation, it's aligned */
676 msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
677 resp_msg = &msg->connect_service_response;
678 message_id = __le16_to_cpu(msg->hdr.message_id);
679 service_id = __le16_to_cpu(resp_msg->service_id);
680
681 if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
682 (htc->control_resp_len < sizeof(msg->hdr) +
683 sizeof(msg->connect_service_response))) {
684 ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
685 return -EPROTO;
686 }
687
688 ath10k_dbg(ar, ATH10K_DBG_HTC,
689 "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
690 htc_service_name(service_id),
691 resp_msg->status, resp_msg->eid);
692
693 conn_resp->connect_resp_code = resp_msg->status;
694
695 /* check response status */
696 if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
697 ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
698 htc_service_name(service_id),
699 resp_msg->status);
700 return -EPROTO;
701 }
702
703 assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
704 max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
705
706setup:
707
708 if (assigned_eid >= ATH10K_HTC_EP_COUNT)
709 return -EPROTO;
710
711 if (max_msg_size == 0)
712 return -EPROTO;
713
714 ep = &htc->endpoint[assigned_eid];
715 ep->eid = assigned_eid;
716
717 if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
718 return -EPROTO;
719
720 /* return assigned endpoint to caller */
721 conn_resp->eid = assigned_eid;
722 conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
723
724 /* setup the endpoint */
725 ep->service_id = conn_req->service_id;
726 ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
727 ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
728 ep->tx_credits = tx_alloc;
729 ep->tx_credit_size = htc->target_credit_size;
730 ep->tx_credits_per_max_message = ep->max_ep_message_len /
731 htc->target_credit_size;
732
733 if (ep->max_ep_message_len % htc->target_credit_size)
734 ep->tx_credits_per_max_message++;
735
736 /* copy all the callbacks */
737 ep->ep_ops = conn_req->ep_ops;
738
739 status = ath10k_hif_map_service_to_pipe(htc->ar,
740 ep->service_id,
741 &ep->ul_pipe_id,
742 &ep->dl_pipe_id);
743 if (status)
744 return status;
745
746 ath10k_dbg(ar, ATH10K_DBG_BOOT,
747 "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
748 htc_service_name(ep->service_id), ep->ul_pipe_id,
749 ep->dl_pipe_id, ep->eid);
750
751 if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
752 ep->tx_credit_flow_enabled = false;
753 ath10k_dbg(ar, ATH10K_DBG_BOOT,
754 "boot htc service '%s' eid %d TX flow control disabled\n",
755 htc_service_name(ep->service_id), assigned_eid);
756 }
757
758 return status;
759}
760
761struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
762{
763 struct sk_buff *skb;
764
765 skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
766 if (!skb)
767 return NULL;
768
769 skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
770
771 /* FW/HTC requires 4-byte aligned streams */
772 if (!IS_ALIGNED((unsigned long)skb->data, 4))
773 ath10k_warn(ar, "Unaligned HTC tx skb\n");
774
775 return skb;
776}
777
778int ath10k_htc_start(struct ath10k_htc *htc)
779{
780 struct ath10k *ar = htc->ar;
781 struct sk_buff *skb;
782 int status = 0;
783 struct ath10k_htc_msg *msg;
784
785 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
786 if (!skb)
787 return -ENOMEM;
788
789 skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
790 memset(skb->data, 0, skb->len);
791
792 msg = (struct ath10k_htc_msg *)skb->data;
793 msg->hdr.message_id =
794 __cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
795
796 ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
797
798 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
799 if (status) {
800 kfree_skb(skb);
801 return status;
802 }
803
804 return 0;
805}
806
807/* registered target arrival callback from the HIF layer */
808int ath10k_htc_init(struct ath10k *ar)
809{
810 struct ath10k_htc_ep *ep = NULL;
811 struct ath10k_htc *htc = &ar->htc;
812
813 spin_lock_init(&htc->tx_lock);
814
815 ath10k_htc_reset_endpoint_states(htc);
816
817 htc->ar = ar;
818
819 /* Get HIF default pipe for HTC message exchange */
820 ep = &htc->endpoint[ATH10K_HTC_EP_0];
821
822 ath10k_hif_get_default_pipe(ar, &ep->ul_pipe_id, &ep->dl_pipe_id);
823
824 init_completion(&htc->ctl_resp);
825
826 return 0;
827}