Loading...
1/*
2 * Copyright (c) 2007-2011 Atheros Communications Inc.
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#include "core.h"
18#include "debug.h"
19#include "hif-ops.h"
20
21#define HTC_PACKET_CONTAINER_ALLOCATION 32
22#define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
23
24static int ath6kl_htc_pipe_tx(struct htc_target *handle,
25 struct htc_packet *packet);
26static void ath6kl_htc_pipe_cleanup(struct htc_target *handle);
27
28/* htc pipe tx path */
29static inline void restore_tx_packet(struct htc_packet *packet)
30{
31 if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
32 skb_pull(packet->skb, sizeof(struct htc_frame_hdr));
33 packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
34 }
35}
36
37static void do_send_completion(struct htc_endpoint *ep,
38 struct list_head *queue_to_indicate)
39{
40 struct htc_packet *packet;
41
42 if (list_empty(queue_to_indicate)) {
43 /* nothing to indicate */
44 return;
45 }
46
47 if (ep->ep_cb.tx_comp_multi != NULL) {
48 ath6kl_dbg(ATH6KL_DBG_HTC,
49 "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
50 __func__, ep->eid,
51 get_queue_depth(queue_to_indicate));
52 /*
53 * a multiple send complete handler is being used,
54 * pass the queue to the handler
55 */
56 ep->ep_cb.tx_comp_multi(ep->target, queue_to_indicate);
57 /*
58 * all packets are now owned by the callback,
59 * reset queue to be safe
60 */
61 INIT_LIST_HEAD(queue_to_indicate);
62 } else {
63 /* using legacy EpTxComplete */
64 do {
65 packet = list_first_entry(queue_to_indicate,
66 struct htc_packet, list);
67
68 list_del(&packet->list);
69 ath6kl_dbg(ATH6KL_DBG_HTC,
70 "%s: calling ep %d send complete callback on packet 0x%p\n",
71 __func__, ep->eid, packet);
72 ep->ep_cb.tx_complete(ep->target, packet);
73 } while (!list_empty(queue_to_indicate));
74 }
75}
76
77static void send_packet_completion(struct htc_target *target,
78 struct htc_packet *packet)
79{
80 struct htc_endpoint *ep = &target->endpoint[packet->endpoint];
81 struct list_head container;
82
83 restore_tx_packet(packet);
84 INIT_LIST_HEAD(&container);
85 list_add_tail(&packet->list, &container);
86
87 /* do completion */
88 do_send_completion(ep, &container);
89}
90
91static void get_htc_packet_credit_based(struct htc_target *target,
92 struct htc_endpoint *ep,
93 struct list_head *queue)
94{
95 int credits_required;
96 int remainder;
97 u8 send_flags;
98 struct htc_packet *packet;
99 unsigned int transfer_len;
100
101 /* NOTE : the TX lock is held when this function is called */
102
103 /* loop until we can grab as many packets out of the queue as we can */
104 while (true) {
105 send_flags = 0;
106 if (list_empty(&ep->txq))
107 break;
108
109 /* get packet at head, but don't remove it */
110 packet = list_first_entry(&ep->txq, struct htc_packet, list);
111
112 ath6kl_dbg(ATH6KL_DBG_HTC,
113 "%s: got head packet:0x%p , queue depth: %d\n",
114 __func__, packet, get_queue_depth(&ep->txq));
115
116 transfer_len = packet->act_len + HTC_HDR_LENGTH;
117
118 if (transfer_len <= target->tgt_cred_sz) {
119 credits_required = 1;
120 } else {
121 /* figure out how many credits this message requires */
122 credits_required = transfer_len / target->tgt_cred_sz;
123 remainder = transfer_len % target->tgt_cred_sz;
124
125 if (remainder)
126 credits_required++;
127 }
128
129 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: creds required:%d got:%d\n",
130 __func__, credits_required, ep->cred_dist.credits);
131
132 if (ep->eid == ENDPOINT_0) {
133 /*
134 * endpoint 0 is special, it always has a credit and
135 * does not require credit based flow control
136 */
137 credits_required = 0;
138
139 } else {
140 if (ep->cred_dist.credits < credits_required)
141 break;
142
143 ep->cred_dist.credits -= credits_required;
144 ep->ep_st.cred_cosumd += credits_required;
145
146 /* check if we need credits back from the target */
147 if (ep->cred_dist.credits <
148 ep->cred_dist.cred_per_msg) {
149 /* tell the target we need credits ASAP! */
150 send_flags |= HTC_FLAGS_NEED_CREDIT_UPDATE;
151 ep->ep_st.cred_low_indicate += 1;
152 ath6kl_dbg(ATH6KL_DBG_HTC,
153 "%s: host needs credits\n",
154 __func__);
155 }
156 }
157
158 /* now we can fully dequeue */
159 packet = list_first_entry(&ep->txq, struct htc_packet, list);
160
161 list_del(&packet->list);
162 /* save the number of credits this packet consumed */
163 packet->info.tx.cred_used = credits_required;
164 /* save send flags */
165 packet->info.tx.flags = send_flags;
166 packet->info.tx.seqno = ep->seqno;
167 ep->seqno++;
168 /* queue this packet into the caller's queue */
169 list_add_tail(&packet->list, queue);
170 }
171}
172
173static void get_htc_packet(struct htc_target *target,
174 struct htc_endpoint *ep,
175 struct list_head *queue, int resources)
176{
177 struct htc_packet *packet;
178
179 /* NOTE : the TX lock is held when this function is called */
180
181 /* loop until we can grab as many packets out of the queue as we can */
182 while (resources) {
183 if (list_empty(&ep->txq))
184 break;
185
186 packet = list_first_entry(&ep->txq, struct htc_packet, list);
187 list_del(&packet->list);
188
189 ath6kl_dbg(ATH6KL_DBG_HTC,
190 "%s: got packet:0x%p , new queue depth: %d\n",
191 __func__, packet, get_queue_depth(&ep->txq));
192 packet->info.tx.seqno = ep->seqno;
193 packet->info.tx.flags = 0;
194 packet->info.tx.cred_used = 0;
195 ep->seqno++;
196
197 /* queue this packet into the caller's queue */
198 list_add_tail(&packet->list, queue);
199 resources--;
200 }
201}
202
203static int htc_issue_packets(struct htc_target *target,
204 struct htc_endpoint *ep,
205 struct list_head *pkt_queue)
206{
207 int status = 0;
208 u16 payload_len;
209 struct sk_buff *skb;
210 struct htc_frame_hdr *htc_hdr;
211 struct htc_packet *packet;
212
213 ath6kl_dbg(ATH6KL_DBG_HTC,
214 "%s: queue: 0x%p, pkts %d\n", __func__,
215 pkt_queue, get_queue_depth(pkt_queue));
216
217 while (!list_empty(pkt_queue)) {
218 packet = list_first_entry(pkt_queue, struct htc_packet, list);
219 list_del(&packet->list);
220
221 skb = packet->skb;
222 if (!skb) {
223 WARN_ON_ONCE(1);
224 status = -EINVAL;
225 break;
226 }
227
228 payload_len = packet->act_len;
229
230 /* setup HTC frame header */
231 htc_hdr = skb_push(skb, sizeof(*htc_hdr));
232 if (!htc_hdr) {
233 WARN_ON_ONCE(1);
234 status = -EINVAL;
235 break;
236 }
237
238 packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
239
240 /* Endianess? */
241 put_unaligned((u16) payload_len, &htc_hdr->payld_len);
242 htc_hdr->flags = packet->info.tx.flags;
243 htc_hdr->eid = (u8) packet->endpoint;
244 htc_hdr->ctrl[0] = 0;
245 htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
246
247 spin_lock_bh(&target->tx_lock);
248
249 /* store in look up queue to match completions */
250 list_add_tail(&packet->list, &ep->pipe.tx_lookup_queue);
251 ep->ep_st.tx_issued += 1;
252 spin_unlock_bh(&target->tx_lock);
253
254 status = ath6kl_hif_pipe_send(target->dev->ar,
255 ep->pipe.pipeid_ul, NULL, skb);
256
257 if (status != 0) {
258 if (status != -ENOMEM) {
259 /* TODO: if more than 1 endpoint maps to the
260 * same PipeID, it is possible to run out of
261 * resources in the HIF layer.
262 * Don't emit the error
263 */
264 ath6kl_dbg(ATH6KL_DBG_HTC,
265 "%s: failed status:%d\n",
266 __func__, status);
267 }
268 spin_lock_bh(&target->tx_lock);
269 list_del(&packet->list);
270
271 /* reclaim credits */
272 ep->cred_dist.credits += packet->info.tx.cred_used;
273 spin_unlock_bh(&target->tx_lock);
274
275 /* put it back into the callers queue */
276 list_add(&packet->list, pkt_queue);
277 break;
278 }
279 }
280
281 if (status != 0) {
282 while (!list_empty(pkt_queue)) {
283 if (status != -ENOMEM) {
284 ath6kl_dbg(ATH6KL_DBG_HTC,
285 "%s: failed pkt:0x%p status:%d\n",
286 __func__, packet, status);
287 }
288
289 packet = list_first_entry(pkt_queue,
290 struct htc_packet, list);
291 list_del(&packet->list);
292 packet->status = status;
293 send_packet_completion(target, packet);
294 }
295 }
296
297 return status;
298}
299
300static enum htc_send_queue_result htc_try_send(struct htc_target *target,
301 struct htc_endpoint *ep,
302 struct list_head *txq)
303{
304 struct list_head send_queue; /* temp queue to hold packets */
305 struct htc_packet *packet, *tmp_pkt;
306 struct ath6kl *ar = target->dev->ar;
307 enum htc_send_full_action action;
308 int tx_resources, overflow, txqueue_depth, i, good_pkts;
309 u8 pipeid;
310
311 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: (queue:0x%p depth:%d)\n",
312 __func__, txq,
313 (txq == NULL) ? 0 : get_queue_depth(txq));
314
315 /* init the local send queue */
316 INIT_LIST_HEAD(&send_queue);
317
318 /*
319 * txq equals to NULL means
320 * caller didn't provide a queue, just wants us to
321 * check queues and send
322 */
323 if (txq != NULL) {
324 if (list_empty(txq)) {
325 /* empty queue */
326 return HTC_SEND_QUEUE_DROP;
327 }
328
329 spin_lock_bh(&target->tx_lock);
330 txqueue_depth = get_queue_depth(&ep->txq);
331 spin_unlock_bh(&target->tx_lock);
332
333 if (txqueue_depth >= ep->max_txq_depth) {
334 /* we've already overflowed */
335 overflow = get_queue_depth(txq);
336 } else {
337 /* get how much we will overflow by */
338 overflow = txqueue_depth;
339 overflow += get_queue_depth(txq);
340 /* get how much we will overflow the TX queue by */
341 overflow -= ep->max_txq_depth;
342 }
343
344 /* if overflow is negative or zero, we are okay */
345 if (overflow > 0) {
346 ath6kl_dbg(ATH6KL_DBG_HTC,
347 "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
348 __func__, ep->eid, overflow, txqueue_depth,
349 ep->max_txq_depth);
350 }
351 if ((overflow <= 0) ||
352 (ep->ep_cb.tx_full == NULL)) {
353 /*
354 * all packets will fit or caller did not provide send
355 * full indication handler -- just move all of them
356 * to the local send_queue object
357 */
358 list_splice_tail_init(txq, &send_queue);
359 } else {
360 good_pkts = get_queue_depth(txq) - overflow;
361 if (good_pkts < 0) {
362 WARN_ON_ONCE(1);
363 return HTC_SEND_QUEUE_DROP;
364 }
365
366 /* we have overflowed, and a callback is provided */
367 /* dequeue all non-overflow packets to the sendqueue */
368 for (i = 0; i < good_pkts; i++) {
369 /* pop off caller's queue */
370 packet = list_first_entry(txq,
371 struct htc_packet,
372 list);
373 /* move to local queue */
374 list_move_tail(&packet->list, &send_queue);
375 }
376
377 /*
378 * the caller's queue has all the packets that won't fit
379 * walk through the caller's queue and indicate each to
380 * the send full handler
381 */
382 list_for_each_entry_safe(packet, tmp_pkt,
383 txq, list) {
384 ath6kl_dbg(ATH6KL_DBG_HTC,
385 "%s: Indicate overflowed TX pkts: %p\n",
386 __func__, packet);
387 action = ep->ep_cb.tx_full(ep->target, packet);
388 if (action == HTC_SEND_FULL_DROP) {
389 /* callback wants the packet dropped */
390 ep->ep_st.tx_dropped += 1;
391
392 /* leave this one in the caller's queue
393 * for cleanup */
394 } else {
395 /* callback wants to keep this packet,
396 * move from caller's queue to the send
397 * queue */
398 list_move_tail(&packet->list,
399 &send_queue);
400 }
401 }
402
403 if (list_empty(&send_queue)) {
404 /* no packets made it in, caller will cleanup */
405 return HTC_SEND_QUEUE_DROP;
406 }
407 }
408 }
409
410 if (!ep->pipe.tx_credit_flow_enabled) {
411 tx_resources =
412 ath6kl_hif_pipe_get_free_queue_number(ar,
413 ep->pipe.pipeid_ul);
414 } else {
415 tx_resources = 0;
416 }
417
418 spin_lock_bh(&target->tx_lock);
419 if (!list_empty(&send_queue)) {
420 /* transfer packets to tail */
421 list_splice_tail_init(&send_queue, &ep->txq);
422 if (!list_empty(&send_queue)) {
423 WARN_ON_ONCE(1);
424 spin_unlock_bh(&target->tx_lock);
425 return HTC_SEND_QUEUE_DROP;
426 }
427 INIT_LIST_HEAD(&send_queue);
428 }
429
430 /* increment tx processing count on entry */
431 ep->tx_proc_cnt++;
432
433 if (ep->tx_proc_cnt > 1) {
434 /*
435 * Another thread or task is draining the TX queues on this
436 * endpoint that thread will reset the tx processing count
437 * when the queue is drained.
438 */
439 ep->tx_proc_cnt--;
440 spin_unlock_bh(&target->tx_lock);
441 return HTC_SEND_QUEUE_OK;
442 }
443
444 /***** beyond this point only 1 thread may enter ******/
445
446 /*
447 * Now drain the endpoint TX queue for transmission as long as we have
448 * enough transmit resources.
449 */
450 while (true) {
451 if (get_queue_depth(&ep->txq) == 0)
452 break;
453
454 if (ep->pipe.tx_credit_flow_enabled) {
455 /*
456 * Credit based mechanism provides flow control
457 * based on target transmit resource availability,
458 * we assume that the HIF layer will always have
459 * bus resources greater than target transmit
460 * resources.
461 */
462 get_htc_packet_credit_based(target, ep, &send_queue);
463 } else {
464 /*
465 * Get all packets for this endpoint that we can
466 * for this pass.
467 */
468 get_htc_packet(target, ep, &send_queue, tx_resources);
469 }
470
471 if (get_queue_depth(&send_queue) == 0) {
472 /*
473 * Didn't get packets due to out of resources or TX
474 * queue was drained.
475 */
476 break;
477 }
478
479 spin_unlock_bh(&target->tx_lock);
480
481 /* send what we can */
482 htc_issue_packets(target, ep, &send_queue);
483
484 if (!ep->pipe.tx_credit_flow_enabled) {
485 pipeid = ep->pipe.pipeid_ul;
486 tx_resources =
487 ath6kl_hif_pipe_get_free_queue_number(ar, pipeid);
488 }
489
490 spin_lock_bh(&target->tx_lock);
491 }
492
493 /* done with this endpoint, we can clear the count */
494 ep->tx_proc_cnt = 0;
495 spin_unlock_bh(&target->tx_lock);
496
497 return HTC_SEND_QUEUE_OK;
498}
499
500/* htc control packet manipulation */
501static void destroy_htc_txctrl_packet(struct htc_packet *packet)
502{
503 struct sk_buff *skb;
504 skb = packet->skb;
505 dev_kfree_skb(skb);
506 kfree(packet);
507}
508
509static struct htc_packet *build_htc_txctrl_packet(void)
510{
511 struct htc_packet *packet = NULL;
512 struct sk_buff *skb;
513
514 packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
515 if (packet == NULL)
516 return NULL;
517
518 skb = __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE, GFP_KERNEL);
519
520 if (skb == NULL) {
521 kfree(packet);
522 return NULL;
523 }
524 packet->skb = skb;
525
526 return packet;
527}
528
529static void htc_free_txctrl_packet(struct htc_target *target,
530 struct htc_packet *packet)
531{
532 destroy_htc_txctrl_packet(packet);
533}
534
535static struct htc_packet *htc_alloc_txctrl_packet(struct htc_target *target)
536{
537 return build_htc_txctrl_packet();
538}
539
540static void htc_txctrl_complete(struct htc_target *target,
541 struct htc_packet *packet)
542{
543 htc_free_txctrl_packet(target, packet);
544}
545
546#define MAX_MESSAGE_SIZE 1536
547
548static int htc_setup_target_buffer_assignments(struct htc_target *target)
549{
550 int status, credits, credit_per_maxmsg, i;
551 struct htc_pipe_txcredit_alloc *entry;
552 unsigned int hif_usbaudioclass = 0;
553
554 credit_per_maxmsg = MAX_MESSAGE_SIZE / target->tgt_cred_sz;
555 if (MAX_MESSAGE_SIZE % target->tgt_cred_sz)
556 credit_per_maxmsg++;
557
558 /* TODO, this should be configured by the caller! */
559
560 credits = target->tgt_creds;
561 entry = &target->pipe.txcredit_alloc[0];
562
563 status = -ENOMEM;
564
565 /* FIXME: hif_usbaudioclass is always zero */
566 if (hif_usbaudioclass) {
567 ath6kl_dbg(ATH6KL_DBG_HTC,
568 "%s: For USB Audio Class- Total:%d\n",
569 __func__, credits);
570 entry++;
571 entry++;
572 /* Setup VO Service To have Max Credits */
573 entry->service_id = WMI_DATA_VO_SVC;
574 entry->credit_alloc = (credits - 6);
575 if (entry->credit_alloc == 0)
576 entry->credit_alloc++;
577
578 credits -= (int) entry->credit_alloc;
579 if (credits <= 0)
580 return status;
581
582 entry++;
583 entry->service_id = WMI_CONTROL_SVC;
584 entry->credit_alloc = credit_per_maxmsg;
585 credits -= (int) entry->credit_alloc;
586 if (credits <= 0)
587 return status;
588
589 /* leftovers go to best effort */
590 entry++;
591 entry++;
592 entry->service_id = WMI_DATA_BE_SVC;
593 entry->credit_alloc = (u8) credits;
594 status = 0;
595 } else {
596 entry++;
597 entry->service_id = WMI_DATA_VI_SVC;
598 entry->credit_alloc = credits / 4;
599 if (entry->credit_alloc == 0)
600 entry->credit_alloc++;
601
602 credits -= (int) entry->credit_alloc;
603 if (credits <= 0)
604 return status;
605
606 entry++;
607 entry->service_id = WMI_DATA_VO_SVC;
608 entry->credit_alloc = credits / 4;
609 if (entry->credit_alloc == 0)
610 entry->credit_alloc++;
611
612 credits -= (int) entry->credit_alloc;
613 if (credits <= 0)
614 return status;
615
616 entry++;
617 entry->service_id = WMI_CONTROL_SVC;
618 entry->credit_alloc = credit_per_maxmsg;
619 credits -= (int) entry->credit_alloc;
620 if (credits <= 0)
621 return status;
622
623 entry++;
624 entry->service_id = WMI_DATA_BK_SVC;
625 entry->credit_alloc = credit_per_maxmsg;
626 credits -= (int) entry->credit_alloc;
627 if (credits <= 0)
628 return status;
629
630 /* leftovers go to best effort */
631 entry++;
632 entry->service_id = WMI_DATA_BE_SVC;
633 entry->credit_alloc = (u8) credits;
634 status = 0;
635 }
636
637 if (status == 0) {
638 for (i = 0; i < ENDPOINT_MAX; i++) {
639 if (target->pipe.txcredit_alloc[i].service_id != 0) {
640 ath6kl_dbg(ATH6KL_DBG_HTC,
641 "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
642 i,
643 target->pipe.txcredit_alloc[i].
644 service_id,
645 target->pipe.txcredit_alloc[i].
646 credit_alloc);
647 }
648 }
649 }
650 return status;
651}
652
653/* process credit reports and call distribution function */
654static void htc_process_credit_report(struct htc_target *target,
655 struct htc_credit_report *rpt,
656 int num_entries,
657 enum htc_endpoint_id from_ep)
658{
659 int total_credits = 0, i;
660 struct htc_endpoint *ep;
661
662 /* lock out TX while we update credits */
663 spin_lock_bh(&target->tx_lock);
664
665 for (i = 0; i < num_entries; i++, rpt++) {
666 if (rpt->eid >= ENDPOINT_MAX) {
667 WARN_ON_ONCE(1);
668 spin_unlock_bh(&target->tx_lock);
669 return;
670 }
671
672 ep = &target->endpoint[rpt->eid];
673 ep->cred_dist.credits += rpt->credits;
674
675 if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) {
676 spin_unlock_bh(&target->tx_lock);
677 htc_try_send(target, ep, NULL);
678 spin_lock_bh(&target->tx_lock);
679 }
680
681 total_credits += rpt->credits;
682 }
683 ath6kl_dbg(ATH6KL_DBG_HTC,
684 "Report indicated %d credits to distribute\n",
685 total_credits);
686
687 spin_unlock_bh(&target->tx_lock);
688}
689
690/* flush endpoint TX queue */
691static void htc_flush_tx_endpoint(struct htc_target *target,
692 struct htc_endpoint *ep, u16 tag)
693{
694 struct htc_packet *packet;
695
696 spin_lock_bh(&target->tx_lock);
697 while (get_queue_depth(&ep->txq)) {
698 packet = list_first_entry(&ep->txq, struct htc_packet, list);
699 list_del(&packet->list);
700 packet->status = 0;
701 send_packet_completion(target, packet);
702 }
703 spin_unlock_bh(&target->tx_lock);
704}
705
706/*
707 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
708 * since upper layers expects struct htc_packet containers we use the completed
709 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
710 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
711 * HTC.
712 */
713static struct htc_packet *htc_lookup_tx_packet(struct htc_target *target,
714 struct htc_endpoint *ep,
715 struct sk_buff *skb)
716{
717 struct htc_packet *packet, *tmp_pkt, *found_packet = NULL;
718
719 spin_lock_bh(&target->tx_lock);
720
721 /*
722 * interate from the front of tx lookup queue
723 * this lookup should be fast since lower layers completes in-order and
724 * so the completed packet should be at the head of the list generally
725 */
726 list_for_each_entry_safe(packet, tmp_pkt, &ep->pipe.tx_lookup_queue,
727 list) {
728 /* check for removal */
729 if (skb == packet->skb) {
730 /* found it */
731 list_del(&packet->list);
732 found_packet = packet;
733 break;
734 }
735 }
736
737 spin_unlock_bh(&target->tx_lock);
738
739 return found_packet;
740}
741
742static int ath6kl_htc_pipe_tx_complete(struct ath6kl *ar, struct sk_buff *skb)
743{
744 struct htc_target *target = ar->htc_target;
745 struct htc_frame_hdr *htc_hdr;
746 struct htc_endpoint *ep;
747 struct htc_packet *packet;
748 u8 ep_id, *netdata;
749
750 netdata = skb->data;
751
752 htc_hdr = (struct htc_frame_hdr *) netdata;
753
754 ep_id = htc_hdr->eid;
755 ep = &target->endpoint[ep_id];
756
757 packet = htc_lookup_tx_packet(target, ep, skb);
758 if (packet == NULL) {
759 /* may have already been flushed and freed */
760 ath6kl_err("HTC TX lookup failed!\n");
761 } else {
762 /* will be giving this buffer back to upper layers */
763 packet->status = 0;
764 send_packet_completion(target, packet);
765 }
766 skb = NULL;
767
768 if (!ep->pipe.tx_credit_flow_enabled) {
769 /*
770 * note: when using TX credit flow, the re-checking of queues
771 * happens when credits flow back from the target. in the
772 * non-TX credit case, we recheck after the packet completes
773 */
774 htc_try_send(target, ep, NULL);
775 }
776
777 return 0;
778}
779
780static int htc_send_packets_multiple(struct htc_target *target,
781 struct list_head *pkt_queue)
782{
783 struct htc_endpoint *ep;
784 struct htc_packet *packet, *tmp_pkt;
785
786 if (list_empty(pkt_queue))
787 return -EINVAL;
788
789 /* get first packet to find out which ep the packets will go into */
790 packet = list_first_entry(pkt_queue, struct htc_packet, list);
791
792 if (packet->endpoint >= ENDPOINT_MAX) {
793 WARN_ON_ONCE(1);
794 return -EINVAL;
795 }
796 ep = &target->endpoint[packet->endpoint];
797
798 htc_try_send(target, ep, pkt_queue);
799
800 /* do completion on any packets that couldn't get in */
801 if (!list_empty(pkt_queue)) {
802 list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
803 packet->status = -ENOMEM;
804 }
805
806 do_send_completion(ep, pkt_queue);
807 }
808
809 return 0;
810}
811
812/* htc pipe rx path */
813static struct htc_packet *alloc_htc_packet_container(struct htc_target *target)
814{
815 struct htc_packet *packet;
816 spin_lock_bh(&target->rx_lock);
817
818 if (target->pipe.htc_packet_pool == NULL) {
819 spin_unlock_bh(&target->rx_lock);
820 return NULL;
821 }
822
823 packet = target->pipe.htc_packet_pool;
824 target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
825
826 spin_unlock_bh(&target->rx_lock);
827
828 packet->list.next = NULL;
829 return packet;
830}
831
832static void free_htc_packet_container(struct htc_target *target,
833 struct htc_packet *packet)
834{
835 struct list_head *lh;
836
837 spin_lock_bh(&target->rx_lock);
838
839 if (target->pipe.htc_packet_pool == NULL) {
840 target->pipe.htc_packet_pool = packet;
841 packet->list.next = NULL;
842 } else {
843 lh = (struct list_head *) target->pipe.htc_packet_pool;
844 packet->list.next = lh;
845 target->pipe.htc_packet_pool = packet;
846 }
847
848 spin_unlock_bh(&target->rx_lock);
849}
850
851static int htc_process_trailer(struct htc_target *target, u8 *buffer,
852 int len, enum htc_endpoint_id from_ep)
853{
854 struct htc_credit_report *report;
855 struct htc_record_hdr *record;
856 u8 *record_buf;
857 int status = 0;
858
859 while (len > 0) {
860 if (len < sizeof(struct htc_record_hdr)) {
861 status = -EINVAL;
862 break;
863 }
864
865 /* these are byte aligned structs */
866 record = (struct htc_record_hdr *) buffer;
867 len -= sizeof(struct htc_record_hdr);
868 buffer += sizeof(struct htc_record_hdr);
869
870 if (record->len > len) {
871 /* no room left in buffer for record */
872 ath6kl_dbg(ATH6KL_DBG_HTC,
873 "invalid length: %d (id:%d) buffer has: %d bytes left\n",
874 record->len, record->rec_id, len);
875 status = -EINVAL;
876 break;
877 }
878
879 /* start of record follows the header */
880 record_buf = buffer;
881
882 switch (record->rec_id) {
883 case HTC_RECORD_CREDITS:
884 if (record->len < sizeof(struct htc_credit_report)) {
885 WARN_ON_ONCE(1);
886 return -EINVAL;
887 }
888
889 report = (struct htc_credit_report *) record_buf;
890 htc_process_credit_report(target, report,
891 record->len / sizeof(*report),
892 from_ep);
893 break;
894 default:
895 ath6kl_dbg(ATH6KL_DBG_HTC,
896 "unhandled record: id:%d length:%d\n",
897 record->rec_id, record->len);
898 break;
899 }
900
901 /* advance buffer past this record for next time around */
902 buffer += record->len;
903 len -= record->len;
904 }
905
906 return status;
907}
908
909static void do_recv_completion(struct htc_endpoint *ep,
910 struct list_head *queue_to_indicate)
911{
912 struct htc_packet *packet;
913
914 if (list_empty(queue_to_indicate)) {
915 /* nothing to indicate */
916 return;
917 }
918
919 /* using legacy EpRecv */
920 while (!list_empty(queue_to_indicate)) {
921 packet = list_first_entry(queue_to_indicate,
922 struct htc_packet, list);
923 list_del(&packet->list);
924 ep->ep_cb.rx(ep->target, packet);
925 }
926
927 return;
928}
929
930static void recv_packet_completion(struct htc_target *target,
931 struct htc_endpoint *ep,
932 struct htc_packet *packet)
933{
934 struct list_head container;
935 INIT_LIST_HEAD(&container);
936 list_add_tail(&packet->list, &container);
937
938 /* do completion */
939 do_recv_completion(ep, &container);
940}
941
942static int ath6kl_htc_pipe_rx_complete(struct ath6kl *ar, struct sk_buff *skb,
943 u8 pipeid)
944{
945 struct htc_target *target = ar->htc_target;
946 u8 *netdata, *trailer, hdr_info;
947 struct htc_frame_hdr *htc_hdr;
948 u32 netlen, trailerlen = 0;
949 struct htc_packet *packet;
950 struct htc_endpoint *ep;
951 u16 payload_len;
952 int status = 0;
953
954 /*
955 * ar->htc_target can be NULL due to a race condition that can occur
956 * during driver initialization(we do 'ath6kl_hif_power_on' before
957 * initializing 'ar->htc_target' via 'ath6kl_htc_create').
958 * 'ath6kl_hif_power_on' assigns 'ath6kl_recv_complete' as
959 * usb_complete_t/callback function for 'usb_fill_bulk_urb'.
960 * Thus the possibility of ar->htc_target being NULL
961 * via ath6kl_recv_complete -> ath6kl_usb_io_comp_work.
962 */
963 if (!target) {
964 ath6kl_dbg(ATH6KL_DBG_HTC, "Target not yet initialized\n");
965 status = -EINVAL;
966 goto free_skb;
967 }
968
969
970 netdata = skb->data;
971 netlen = skb->len;
972
973 htc_hdr = (struct htc_frame_hdr *) netdata;
974
975 if (htc_hdr->eid >= ENDPOINT_MAX) {
976 ath6kl_dbg(ATH6KL_DBG_HTC,
977 "HTC Rx: invalid EndpointID=%d\n",
978 htc_hdr->eid);
979 status = -EINVAL;
980 goto free_skb;
981 }
982 ep = &target->endpoint[htc_hdr->eid];
983
984 payload_len = le16_to_cpu(get_unaligned(&htc_hdr->payld_len));
985
986 if (netlen < (payload_len + HTC_HDR_LENGTH)) {
987 ath6kl_dbg(ATH6KL_DBG_HTC,
988 "HTC Rx: insufficient length, got:%d expected =%zu\n",
989 netlen, payload_len + HTC_HDR_LENGTH);
990 status = -EINVAL;
991 goto free_skb;
992 }
993
994 /* get flags to check for trailer */
995 hdr_info = htc_hdr->flags;
996 if (hdr_info & HTC_FLG_RX_TRAILER) {
997 /* extract the trailer length */
998 hdr_info = htc_hdr->ctrl[0];
999 if ((hdr_info < sizeof(struct htc_record_hdr)) ||
1000 (hdr_info > payload_len)) {
1001 ath6kl_dbg(ATH6KL_DBG_HTC,
1002 "invalid header: payloadlen should be %d, CB[0]: %d\n",
1003 payload_len, hdr_info);
1004 status = -EINVAL;
1005 goto free_skb;
1006 }
1007
1008 trailerlen = hdr_info;
1009 /* process trailer after hdr/apps payload */
1010 trailer = (u8 *) htc_hdr + HTC_HDR_LENGTH +
1011 payload_len - hdr_info;
1012 status = htc_process_trailer(target, trailer, hdr_info,
1013 htc_hdr->eid);
1014 if (status != 0)
1015 goto free_skb;
1016 }
1017
1018 if (((int) payload_len - (int) trailerlen) <= 0) {
1019 /* zero length packet with trailer, just drop these */
1020 goto free_skb;
1021 }
1022
1023 if (htc_hdr->eid == ENDPOINT_0) {
1024 /* handle HTC control message */
1025 if (target->htc_flags & HTC_OP_STATE_SETUP_COMPLETE) {
1026 /*
1027 * fatal: target should not send unsolicited
1028 * messageson the endpoint 0
1029 */
1030 ath6kl_dbg(ATH6KL_DBG_HTC,
1031 "HTC ignores Rx Ctrl after setup complete\n");
1032 status = -EINVAL;
1033 goto free_skb;
1034 }
1035
1036 /* remove HTC header */
1037 skb_pull(skb, HTC_HDR_LENGTH);
1038
1039 netdata = skb->data;
1040 netlen = skb->len;
1041
1042 spin_lock_bh(&target->rx_lock);
1043
1044 target->pipe.ctrl_response_valid = true;
1045 target->pipe.ctrl_response_len = min_t(int, netlen,
1046 HTC_MAX_CTRL_MSG_LEN);
1047 memcpy(target->pipe.ctrl_response_buf, netdata,
1048 target->pipe.ctrl_response_len);
1049
1050 spin_unlock_bh(&target->rx_lock);
1051
1052 dev_kfree_skb(skb);
1053 skb = NULL;
1054
1055 goto free_skb;
1056 }
1057
1058 /*
1059 * TODO: the message based HIF architecture allocates net bufs
1060 * for recv packets since it bridges that HIF to upper layers,
1061 * which expects HTC packets, we form the packets here
1062 */
1063 packet = alloc_htc_packet_container(target);
1064 if (packet == NULL) {
1065 status = -ENOMEM;
1066 goto free_skb;
1067 }
1068
1069 packet->status = 0;
1070 packet->endpoint = htc_hdr->eid;
1071 packet->pkt_cntxt = skb;
1072
1073 /* TODO: for backwards compatibility */
1074 packet->buf = skb_push(skb, 0) + HTC_HDR_LENGTH;
1075 packet->act_len = netlen - HTC_HDR_LENGTH - trailerlen;
1076
1077 /*
1078 * TODO: this is a hack because the driver layer will set the
1079 * actual len of the skb again which will just double the len
1080 */
1081 skb_trim(skb, 0);
1082
1083 recv_packet_completion(target, ep, packet);
1084
1085 /* recover the packet container */
1086 free_htc_packet_container(target, packet);
1087 skb = NULL;
1088
1089free_skb:
1090 dev_kfree_skb(skb);
1091
1092 return status;
1093}
1094
1095static void htc_flush_rx_queue(struct htc_target *target,
1096 struct htc_endpoint *ep)
1097{
1098 struct list_head container;
1099 struct htc_packet *packet;
1100
1101 spin_lock_bh(&target->rx_lock);
1102
1103 while (1) {
1104 if (list_empty(&ep->rx_bufq))
1105 break;
1106
1107 packet = list_first_entry(&ep->rx_bufq,
1108 struct htc_packet, list);
1109 list_del(&packet->list);
1110
1111 spin_unlock_bh(&target->rx_lock);
1112 packet->status = -ECANCELED;
1113 packet->act_len = 0;
1114
1115 ath6kl_dbg(ATH6KL_DBG_HTC,
1116 "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1117 packet, packet->buf_len,
1118 packet->endpoint);
1119
1120 INIT_LIST_HEAD(&container);
1121 list_add_tail(&packet->list, &container);
1122
1123 /* give the packet back */
1124 do_recv_completion(ep, &container);
1125 spin_lock_bh(&target->rx_lock);
1126 }
1127
1128 spin_unlock_bh(&target->rx_lock);
1129}
1130
1131/* polling routine to wait for a control packet to be received */
1132static int htc_wait_recv_ctrl_message(struct htc_target *target)
1133{
1134 int count = HTC_TARGET_RESPONSE_POLL_COUNT;
1135
1136 while (count > 0) {
1137 spin_lock_bh(&target->rx_lock);
1138
1139 if (target->pipe.ctrl_response_valid) {
1140 target->pipe.ctrl_response_valid = false;
1141 spin_unlock_bh(&target->rx_lock);
1142 break;
1143 }
1144
1145 spin_unlock_bh(&target->rx_lock);
1146
1147 count--;
1148
1149 msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT);
1150 }
1151
1152 if (count <= 0) {
1153 ath6kl_warn("htc pipe control receive timeout!\n");
1154 return -ETIMEDOUT;
1155 }
1156
1157 return 0;
1158}
1159
1160static void htc_rxctrl_complete(struct htc_target *context,
1161 struct htc_packet *packet)
1162{
1163 struct sk_buff *skb = packet->skb;
1164
1165 if (packet->endpoint == ENDPOINT_0 &&
1166 packet->status == -ECANCELED &&
1167 skb != NULL)
1168 dev_kfree_skb(skb);
1169}
1170
1171/* htc pipe initialization */
1172static void reset_endpoint_states(struct htc_target *target)
1173{
1174 struct htc_endpoint *ep;
1175 int i;
1176
1177 for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1178 ep = &target->endpoint[i];
1179 ep->svc_id = 0;
1180 ep->len_max = 0;
1181 ep->max_txq_depth = 0;
1182 ep->eid = i;
1183 INIT_LIST_HEAD(&ep->txq);
1184 INIT_LIST_HEAD(&ep->pipe.tx_lookup_queue);
1185 INIT_LIST_HEAD(&ep->rx_bufq);
1186 ep->target = target;
1187 ep->pipe.tx_credit_flow_enabled = true;
1188 }
1189}
1190
1191/* start HTC, this is called after all services are connected */
1192static int htc_config_target_hif_pipe(struct htc_target *target)
1193{
1194 return 0;
1195}
1196
1197/* htc service functions */
1198static u8 htc_get_credit_alloc(struct htc_target *target, u16 service_id)
1199{
1200 u8 allocation = 0;
1201 int i;
1202
1203 for (i = 0; i < ENDPOINT_MAX; i++) {
1204 if (target->pipe.txcredit_alloc[i].service_id == service_id)
1205 allocation =
1206 target->pipe.txcredit_alloc[i].credit_alloc;
1207 }
1208
1209 if (allocation == 0) {
1210 ath6kl_dbg(ATH6KL_DBG_HTC,
1211 "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1212 service_id);
1213 }
1214
1215 return allocation;
1216}
1217
1218static int ath6kl_htc_pipe_conn_service(struct htc_target *target,
1219 struct htc_service_connect_req *conn_req,
1220 struct htc_service_connect_resp *conn_resp)
1221{
1222 struct ath6kl *ar = target->dev->ar;
1223 struct htc_packet *packet = NULL;
1224 struct htc_conn_service_resp *resp_msg;
1225 struct htc_conn_service_msg *conn_msg;
1226 enum htc_endpoint_id assigned_epid = ENDPOINT_MAX;
1227 bool disable_credit_flowctrl = false;
1228 unsigned int max_msg_size = 0;
1229 struct htc_endpoint *ep;
1230 int length, status = 0;
1231 struct sk_buff *skb;
1232 u8 tx_alloc;
1233 u16 flags;
1234
1235 if (conn_req->svc_id == 0) {
1236 WARN_ON_ONCE(1);
1237 status = -EINVAL;
1238 goto free_packet;
1239 }
1240
1241 if (conn_req->svc_id == HTC_CTRL_RSVD_SVC) {
1242 /* special case for pseudo control service */
1243 assigned_epid = ENDPOINT_0;
1244 max_msg_size = HTC_MAX_CTRL_MSG_LEN;
1245 tx_alloc = 0;
1246
1247 } else {
1248 tx_alloc = htc_get_credit_alloc(target, conn_req->svc_id);
1249 if (tx_alloc == 0) {
1250 status = -ENOMEM;
1251 goto free_packet;
1252 }
1253
1254 /* allocate a packet to send to the target */
1255 packet = htc_alloc_txctrl_packet(target);
1256
1257 if (packet == NULL) {
1258 WARN_ON_ONCE(1);
1259 status = -ENOMEM;
1260 goto free_packet;
1261 }
1262
1263 skb = packet->skb;
1264 length = sizeof(struct htc_conn_service_msg);
1265
1266 /* assemble connect service message */
1267 conn_msg = skb_put(skb, length);
1268 if (conn_msg == NULL) {
1269 WARN_ON_ONCE(1);
1270 status = -EINVAL;
1271 goto free_packet;
1272 }
1273
1274 memset(conn_msg, 0,
1275 sizeof(struct htc_conn_service_msg));
1276 conn_msg->msg_id = cpu_to_le16(HTC_MSG_CONN_SVC_ID);
1277 conn_msg->svc_id = cpu_to_le16(conn_req->svc_id);
1278 conn_msg->conn_flags = cpu_to_le16(conn_req->conn_flags &
1279 ~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK);
1280
1281 /* tell target desired recv alloc for this ep */
1282 flags = tx_alloc << HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT;
1283 conn_msg->conn_flags |= cpu_to_le16(flags);
1284
1285 if (conn_req->conn_flags &
1286 HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL) {
1287 disable_credit_flowctrl = true;
1288 }
1289
1290 set_htc_pkt_info(packet, NULL, (u8 *) conn_msg,
1291 length,
1292 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1293
1294 status = ath6kl_htc_pipe_tx(target, packet);
1295
1296 /* we don't own it anymore */
1297 packet = NULL;
1298 if (status != 0)
1299 goto free_packet;
1300
1301 /* wait for response */
1302 status = htc_wait_recv_ctrl_message(target);
1303 if (status != 0)
1304 goto free_packet;
1305
1306 /* we controlled the buffer creation so it has to be
1307 * properly aligned
1308 */
1309 resp_msg = (struct htc_conn_service_resp *)
1310 target->pipe.ctrl_response_buf;
1311
1312 if (resp_msg->msg_id != cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID) ||
1313 (target->pipe.ctrl_response_len < sizeof(*resp_msg))) {
1314 /* this message is not valid */
1315 WARN_ON_ONCE(1);
1316 status = -EINVAL;
1317 goto free_packet;
1318 }
1319
1320 ath6kl_dbg(ATH6KL_DBG_TRC,
1321 "%s: service 0x%X conn resp: status: %d ep: %d\n",
1322 __func__, resp_msg->svc_id, resp_msg->status,
1323 resp_msg->eid);
1324
1325 conn_resp->resp_code = resp_msg->status;
1326 /* check response status */
1327 if (resp_msg->status != HTC_SERVICE_SUCCESS) {
1328 ath6kl_dbg(ATH6KL_DBG_HTC,
1329 "Target failed service 0x%X connect request (status:%d)\n",
1330 resp_msg->svc_id, resp_msg->status);
1331 status = -EINVAL;
1332 goto free_packet;
1333 }
1334
1335 assigned_epid = (enum htc_endpoint_id) resp_msg->eid;
1336 max_msg_size = le16_to_cpu(resp_msg->max_msg_sz);
1337 }
1338
1339 /* the rest are parameter checks so set the error status */
1340 status = -EINVAL;
1341
1342 if (assigned_epid >= ENDPOINT_MAX) {
1343 WARN_ON_ONCE(1);
1344 goto free_packet;
1345 }
1346
1347 if (max_msg_size == 0) {
1348 WARN_ON_ONCE(1);
1349 goto free_packet;
1350 }
1351
1352 ep = &target->endpoint[assigned_epid];
1353 ep->eid = assigned_epid;
1354 if (ep->svc_id != 0) {
1355 /* endpoint already in use! */
1356 WARN_ON_ONCE(1);
1357 goto free_packet;
1358 }
1359
1360 /* return assigned endpoint to caller */
1361 conn_resp->endpoint = assigned_epid;
1362 conn_resp->len_max = max_msg_size;
1363
1364 /* setup the endpoint */
1365 ep->svc_id = conn_req->svc_id; /* this marks ep in use */
1366 ep->max_txq_depth = conn_req->max_txq_depth;
1367 ep->len_max = max_msg_size;
1368 ep->cred_dist.credits = tx_alloc;
1369 ep->cred_dist.cred_sz = target->tgt_cred_sz;
1370 ep->cred_dist.cred_per_msg = max_msg_size / target->tgt_cred_sz;
1371 if (max_msg_size % target->tgt_cred_sz)
1372 ep->cred_dist.cred_per_msg++;
1373
1374 /* copy all the callbacks */
1375 ep->ep_cb = conn_req->ep_cb;
1376
1377 /* initialize tx_drop_packet_threshold */
1378 ep->tx_drop_packet_threshold = MAX_HI_COOKIE_NUM;
1379
1380 status = ath6kl_hif_pipe_map_service(ar, ep->svc_id,
1381 &ep->pipe.pipeid_ul,
1382 &ep->pipe.pipeid_dl);
1383 if (status != 0)
1384 goto free_packet;
1385
1386 ath6kl_dbg(ATH6KL_DBG_HTC,
1387 "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1388 ep->svc_id, ep->pipe.pipeid_ul,
1389 ep->pipe.pipeid_dl, ep->eid);
1390
1391 if (disable_credit_flowctrl && ep->pipe.tx_credit_flow_enabled) {
1392 ep->pipe.tx_credit_flow_enabled = false;
1393 ath6kl_dbg(ATH6KL_DBG_HTC,
1394 "SVC: 0x%4.4X ep:%d TX flow control off\n",
1395 ep->svc_id, assigned_epid);
1396 }
1397
1398free_packet:
1399 if (packet != NULL)
1400 htc_free_txctrl_packet(target, packet);
1401 return status;
1402}
1403
1404/* htc export functions */
1405static void *ath6kl_htc_pipe_create(struct ath6kl *ar)
1406{
1407 int status = 0;
1408 struct htc_endpoint *ep = NULL;
1409 struct htc_target *target = NULL;
1410 struct htc_packet *packet;
1411 int i;
1412
1413 target = kzalloc(sizeof(struct htc_target), GFP_KERNEL);
1414 if (target == NULL) {
1415 ath6kl_err("htc create unable to allocate memory\n");
1416 status = -ENOMEM;
1417 goto fail_htc_create;
1418 }
1419
1420 spin_lock_init(&target->htc_lock);
1421 spin_lock_init(&target->rx_lock);
1422 spin_lock_init(&target->tx_lock);
1423
1424 reset_endpoint_states(target);
1425
1426 for (i = 0; i < HTC_PACKET_CONTAINER_ALLOCATION; i++) {
1427 packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
1428
1429 if (packet != NULL)
1430 free_htc_packet_container(target, packet);
1431 }
1432
1433 target->dev = kzalloc(sizeof(*target->dev), GFP_KERNEL);
1434 if (!target->dev) {
1435 ath6kl_err("unable to allocate memory\n");
1436 status = -ENOMEM;
1437 goto fail_htc_create;
1438 }
1439 target->dev->ar = ar;
1440 target->dev->htc_cnxt = target;
1441
1442 /* Get HIF default pipe for HTC message exchange */
1443 ep = &target->endpoint[ENDPOINT_0];
1444
1445 ath6kl_hif_pipe_get_default(ar, &ep->pipe.pipeid_ul,
1446 &ep->pipe.pipeid_dl);
1447
1448 return target;
1449
1450fail_htc_create:
1451 if (status != 0) {
1452 if (target != NULL)
1453 ath6kl_htc_pipe_cleanup(target);
1454
1455 target = NULL;
1456 }
1457 return target;
1458}
1459
1460/* cleanup the HTC instance */
1461static void ath6kl_htc_pipe_cleanup(struct htc_target *target)
1462{
1463 struct htc_packet *packet;
1464
1465 while (true) {
1466 packet = alloc_htc_packet_container(target);
1467 if (packet == NULL)
1468 break;
1469 kfree(packet);
1470 }
1471
1472 kfree(target->dev);
1473
1474 /* kfree our instance */
1475 kfree(target);
1476}
1477
1478static int ath6kl_htc_pipe_start(struct htc_target *target)
1479{
1480 struct sk_buff *skb;
1481 struct htc_setup_comp_ext_msg *setup;
1482 struct htc_packet *packet;
1483
1484 htc_config_target_hif_pipe(target);
1485
1486 /* allocate a buffer to send */
1487 packet = htc_alloc_txctrl_packet(target);
1488 if (packet == NULL) {
1489 WARN_ON_ONCE(1);
1490 return -ENOMEM;
1491 }
1492
1493 skb = packet->skb;
1494
1495 /* assemble setup complete message */
1496 setup = skb_put(skb, sizeof(*setup));
1497 memset(setup, 0, sizeof(struct htc_setup_comp_ext_msg));
1498 setup->msg_id = cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID);
1499
1500 ath6kl_dbg(ATH6KL_DBG_HTC, "HTC using TX credit flow control\n");
1501
1502 set_htc_pkt_info(packet, NULL, (u8 *) setup,
1503 sizeof(struct htc_setup_comp_ext_msg),
1504 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1505
1506 target->htc_flags |= HTC_OP_STATE_SETUP_COMPLETE;
1507
1508 return ath6kl_htc_pipe_tx(target, packet);
1509}
1510
1511static void ath6kl_htc_pipe_stop(struct htc_target *target)
1512{
1513 int i;
1514 struct htc_endpoint *ep;
1515
1516 /* cleanup endpoints */
1517 for (i = 0; i < ENDPOINT_MAX; i++) {
1518 ep = &target->endpoint[i];
1519 htc_flush_rx_queue(target, ep);
1520 htc_flush_tx_endpoint(target, ep, HTC_TX_PACKET_TAG_ALL);
1521 }
1522
1523 reset_endpoint_states(target);
1524 target->htc_flags &= ~HTC_OP_STATE_SETUP_COMPLETE;
1525}
1526
1527static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target *target,
1528 enum htc_endpoint_id endpoint)
1529{
1530 int num;
1531
1532 spin_lock_bh(&target->rx_lock);
1533 num = get_queue_depth(&(target->endpoint[endpoint].rx_bufq));
1534 spin_unlock_bh(&target->rx_lock);
1535
1536 return num;
1537}
1538
1539static int ath6kl_htc_pipe_tx(struct htc_target *target,
1540 struct htc_packet *packet)
1541{
1542 struct list_head queue;
1543
1544 ath6kl_dbg(ATH6KL_DBG_HTC,
1545 "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1546 __func__, packet->endpoint, packet->buf,
1547 packet->act_len);
1548
1549 INIT_LIST_HEAD(&queue);
1550 list_add_tail(&packet->list, &queue);
1551
1552 return htc_send_packets_multiple(target, &queue);
1553}
1554
1555static int ath6kl_htc_pipe_wait_target(struct htc_target *target)
1556{
1557 struct htc_ready_ext_msg *ready_msg;
1558 struct htc_service_connect_req connect;
1559 struct htc_service_connect_resp resp;
1560 int status = 0;
1561
1562 status = htc_wait_recv_ctrl_message(target);
1563
1564 if (status != 0)
1565 return status;
1566
1567 if (target->pipe.ctrl_response_len < sizeof(*ready_msg)) {
1568 ath6kl_warn("invalid htc pipe ready msg len: %d\n",
1569 target->pipe.ctrl_response_len);
1570 return -ECOMM;
1571 }
1572
1573 ready_msg = (struct htc_ready_ext_msg *) target->pipe.ctrl_response_buf;
1574
1575 if (ready_msg->ver2_0_info.msg_id != cpu_to_le16(HTC_MSG_READY_ID)) {
1576 ath6kl_warn("invalid htc pipe ready msg: 0x%x\n",
1577 ready_msg->ver2_0_info.msg_id);
1578 return -ECOMM;
1579 }
1580
1581 ath6kl_dbg(ATH6KL_DBG_HTC,
1582 "Target Ready! : transmit resources : %d size:%d\n",
1583 ready_msg->ver2_0_info.cred_cnt,
1584 ready_msg->ver2_0_info.cred_sz);
1585
1586 target->tgt_creds = le16_to_cpu(ready_msg->ver2_0_info.cred_cnt);
1587 target->tgt_cred_sz = le16_to_cpu(ready_msg->ver2_0_info.cred_sz);
1588
1589 if ((target->tgt_creds == 0) || (target->tgt_cred_sz == 0))
1590 return -ECOMM;
1591
1592 htc_setup_target_buffer_assignments(target);
1593
1594 /* setup our pseudo HTC control endpoint connection */
1595 memset(&connect, 0, sizeof(connect));
1596 memset(&resp, 0, sizeof(resp));
1597 connect.ep_cb.tx_complete = htc_txctrl_complete;
1598 connect.ep_cb.rx = htc_rxctrl_complete;
1599 connect.max_txq_depth = NUM_CONTROL_TX_BUFFERS;
1600 connect.svc_id = HTC_CTRL_RSVD_SVC;
1601
1602 /* connect fake service */
1603 status = ath6kl_htc_pipe_conn_service(target, &connect, &resp);
1604
1605 return status;
1606}
1607
1608static void ath6kl_htc_pipe_flush_txep(struct htc_target *target,
1609 enum htc_endpoint_id endpoint, u16 tag)
1610{
1611 struct htc_endpoint *ep = &target->endpoint[endpoint];
1612
1613 if (ep->svc_id == 0) {
1614 WARN_ON_ONCE(1);
1615 /* not in use.. */
1616 return;
1617 }
1618
1619 htc_flush_tx_endpoint(target, ep, tag);
1620}
1621
1622static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target *target,
1623 struct list_head *pkt_queue)
1624{
1625 struct htc_packet *packet, *tmp_pkt, *first;
1626 struct htc_endpoint *ep;
1627 int status = 0;
1628
1629 if (list_empty(pkt_queue))
1630 return -EINVAL;
1631
1632 first = list_first_entry(pkt_queue, struct htc_packet, list);
1633
1634 if (first->endpoint >= ENDPOINT_MAX) {
1635 WARN_ON_ONCE(1);
1636 return -EINVAL;
1637 }
1638
1639 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: epid: %d, cnt:%d, len: %d\n",
1640 __func__, first->endpoint, get_queue_depth(pkt_queue),
1641 first->buf_len);
1642
1643 ep = &target->endpoint[first->endpoint];
1644
1645 spin_lock_bh(&target->rx_lock);
1646
1647 /* store receive packets */
1648 list_splice_tail_init(pkt_queue, &ep->rx_bufq);
1649
1650 spin_unlock_bh(&target->rx_lock);
1651
1652 if (status != 0) {
1653 /* walk through queue and mark each one canceled */
1654 list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
1655 packet->status = -ECANCELED;
1656 }
1657
1658 do_recv_completion(ep, pkt_queue);
1659 }
1660
1661 return status;
1662}
1663
1664static void ath6kl_htc_pipe_activity_changed(struct htc_target *target,
1665 enum htc_endpoint_id ep,
1666 bool active)
1667{
1668 /* TODO */
1669}
1670
1671static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target *target)
1672{
1673 struct htc_endpoint *endpoint;
1674 struct htc_packet *packet, *tmp_pkt;
1675 int i;
1676
1677 for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1678 endpoint = &target->endpoint[i];
1679
1680 spin_lock_bh(&target->rx_lock);
1681
1682 list_for_each_entry_safe(packet, tmp_pkt,
1683 &endpoint->rx_bufq, list) {
1684 list_del(&packet->list);
1685 spin_unlock_bh(&target->rx_lock);
1686 ath6kl_dbg(ATH6KL_DBG_HTC,
1687 "htc rx flush pkt 0x%p len %d ep %d\n",
1688 packet, packet->buf_len,
1689 packet->endpoint);
1690 dev_kfree_skb(packet->pkt_cntxt);
1691 spin_lock_bh(&target->rx_lock);
1692 }
1693
1694 spin_unlock_bh(&target->rx_lock);
1695 }
1696}
1697
1698static int ath6kl_htc_pipe_credit_setup(struct htc_target *target,
1699 struct ath6kl_htc_credit_info *info)
1700{
1701 return 0;
1702}
1703
1704static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops = {
1705 .create = ath6kl_htc_pipe_create,
1706 .wait_target = ath6kl_htc_pipe_wait_target,
1707 .start = ath6kl_htc_pipe_start,
1708 .conn_service = ath6kl_htc_pipe_conn_service,
1709 .tx = ath6kl_htc_pipe_tx,
1710 .stop = ath6kl_htc_pipe_stop,
1711 .cleanup = ath6kl_htc_pipe_cleanup,
1712 .flush_txep = ath6kl_htc_pipe_flush_txep,
1713 .flush_rx_buf = ath6kl_htc_pipe_flush_rx_buf,
1714 .activity_changed = ath6kl_htc_pipe_activity_changed,
1715 .get_rxbuf_num = ath6kl_htc_pipe_get_rxbuf_num,
1716 .add_rxbuf_multiple = ath6kl_htc_pipe_add_rxbuf_multiple,
1717 .credit_setup = ath6kl_htc_pipe_credit_setup,
1718 .tx_complete = ath6kl_htc_pipe_tx_complete,
1719 .rx_complete = ath6kl_htc_pipe_rx_complete,
1720};
1721
1722void ath6kl_htc_pipe_attach(struct ath6kl *ar)
1723{
1724 ar->htc_ops = &ath6kl_htc_pipe_ops;
1725}
1/*
2 * Copyright (c) 2007-2011 Atheros Communications Inc.
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#include "core.h"
18#include "debug.h"
19#include "hif-ops.h"
20
21#define HTC_PACKET_CONTAINER_ALLOCATION 32
22#define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
23
24static int ath6kl_htc_pipe_tx(struct htc_target *handle,
25 struct htc_packet *packet);
26static void ath6kl_htc_pipe_cleanup(struct htc_target *handle);
27
28/* htc pipe tx path */
29static inline void restore_tx_packet(struct htc_packet *packet)
30{
31 if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
32 skb_pull(packet->skb, sizeof(struct htc_frame_hdr));
33 packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
34 }
35}
36
37static void do_send_completion(struct htc_endpoint *ep,
38 struct list_head *queue_to_indicate)
39{
40 struct htc_packet *packet;
41
42 if (list_empty(queue_to_indicate)) {
43 /* nothing to indicate */
44 return;
45 }
46
47 if (ep->ep_cb.tx_comp_multi != NULL) {
48 ath6kl_dbg(ATH6KL_DBG_HTC,
49 "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
50 __func__, ep->eid,
51 get_queue_depth(queue_to_indicate));
52 /*
53 * a multiple send complete handler is being used,
54 * pass the queue to the handler
55 */
56 ep->ep_cb.tx_comp_multi(ep->target, queue_to_indicate);
57 /*
58 * all packets are now owned by the callback,
59 * reset queue to be safe
60 */
61 INIT_LIST_HEAD(queue_to_indicate);
62 } else {
63 /* using legacy EpTxComplete */
64 do {
65 packet = list_first_entry(queue_to_indicate,
66 struct htc_packet, list);
67
68 list_del(&packet->list);
69 ath6kl_dbg(ATH6KL_DBG_HTC,
70 "%s: calling ep %d send complete callback on packet 0x%p\n",
71 __func__, ep->eid, packet);
72 ep->ep_cb.tx_complete(ep->target, packet);
73 } while (!list_empty(queue_to_indicate));
74 }
75}
76
77static void send_packet_completion(struct htc_target *target,
78 struct htc_packet *packet)
79{
80 struct htc_endpoint *ep = &target->endpoint[packet->endpoint];
81 struct list_head container;
82
83 restore_tx_packet(packet);
84 INIT_LIST_HEAD(&container);
85 list_add_tail(&packet->list, &container);
86
87 /* do completion */
88 do_send_completion(ep, &container);
89}
90
91static void get_htc_packet_credit_based(struct htc_target *target,
92 struct htc_endpoint *ep,
93 struct list_head *queue)
94{
95 int credits_required;
96 int remainder;
97 u8 send_flags;
98 struct htc_packet *packet;
99 unsigned int transfer_len;
100
101 /* NOTE : the TX lock is held when this function is called */
102
103 /* loop until we can grab as many packets out of the queue as we can */
104 while (true) {
105 send_flags = 0;
106 if (list_empty(&ep->txq))
107 break;
108
109 /* get packet at head, but don't remove it */
110 packet = list_first_entry(&ep->txq, struct htc_packet, list);
111
112 ath6kl_dbg(ATH6KL_DBG_HTC,
113 "%s: got head packet:0x%p , queue depth: %d\n",
114 __func__, packet, get_queue_depth(&ep->txq));
115
116 transfer_len = packet->act_len + HTC_HDR_LENGTH;
117
118 if (transfer_len <= target->tgt_cred_sz) {
119 credits_required = 1;
120 } else {
121 /* figure out how many credits this message requires */
122 credits_required = transfer_len / target->tgt_cred_sz;
123 remainder = transfer_len % target->tgt_cred_sz;
124
125 if (remainder)
126 credits_required++;
127 }
128
129 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: creds required:%d got:%d\n",
130 __func__, credits_required, ep->cred_dist.credits);
131
132 if (ep->eid == ENDPOINT_0) {
133 /*
134 * endpoint 0 is special, it always has a credit and
135 * does not require credit based flow control
136 */
137 credits_required = 0;
138
139 } else {
140
141 if (ep->cred_dist.credits < credits_required)
142 break;
143
144 ep->cred_dist.credits -= credits_required;
145 ep->ep_st.cred_cosumd += credits_required;
146
147 /* check if we need credits back from the target */
148 if (ep->cred_dist.credits <
149 ep->cred_dist.cred_per_msg) {
150 /* tell the target we need credits ASAP! */
151 send_flags |= HTC_FLAGS_NEED_CREDIT_UPDATE;
152 ep->ep_st.cred_low_indicate += 1;
153 ath6kl_dbg(ATH6KL_DBG_HTC,
154 "%s: host needs credits\n",
155 __func__);
156 }
157 }
158
159 /* now we can fully dequeue */
160 packet = list_first_entry(&ep->txq, struct htc_packet, list);
161
162 list_del(&packet->list);
163 /* save the number of credits this packet consumed */
164 packet->info.tx.cred_used = credits_required;
165 /* save send flags */
166 packet->info.tx.flags = send_flags;
167 packet->info.tx.seqno = ep->seqno;
168 ep->seqno++;
169 /* queue this packet into the caller's queue */
170 list_add_tail(&packet->list, queue);
171 }
172
173}
174
175static void get_htc_packet(struct htc_target *target,
176 struct htc_endpoint *ep,
177 struct list_head *queue, int resources)
178{
179 struct htc_packet *packet;
180
181 /* NOTE : the TX lock is held when this function is called */
182
183 /* loop until we can grab as many packets out of the queue as we can */
184 while (resources) {
185 if (list_empty(&ep->txq))
186 break;
187
188 packet = list_first_entry(&ep->txq, struct htc_packet, list);
189 list_del(&packet->list);
190
191 ath6kl_dbg(ATH6KL_DBG_HTC,
192 "%s: got packet:0x%p , new queue depth: %d\n",
193 __func__, packet, get_queue_depth(&ep->txq));
194 packet->info.tx.seqno = ep->seqno;
195 packet->info.tx.flags = 0;
196 packet->info.tx.cred_used = 0;
197 ep->seqno++;
198
199 /* queue this packet into the caller's queue */
200 list_add_tail(&packet->list, queue);
201 resources--;
202 }
203}
204
205static int htc_issue_packets(struct htc_target *target,
206 struct htc_endpoint *ep,
207 struct list_head *pkt_queue)
208{
209 int status = 0;
210 u16 payload_len;
211 struct sk_buff *skb;
212 struct htc_frame_hdr *htc_hdr;
213 struct htc_packet *packet;
214
215 ath6kl_dbg(ATH6KL_DBG_HTC,
216 "%s: queue: 0x%p, pkts %d\n", __func__,
217 pkt_queue, get_queue_depth(pkt_queue));
218
219 while (!list_empty(pkt_queue)) {
220 packet = list_first_entry(pkt_queue, struct htc_packet, list);
221 list_del(&packet->list);
222
223 skb = packet->skb;
224 if (!skb) {
225 WARN_ON_ONCE(1);
226 status = -EINVAL;
227 break;
228 }
229
230 payload_len = packet->act_len;
231
232 /* setup HTC frame header */
233 htc_hdr = (struct htc_frame_hdr *) skb_push(skb,
234 sizeof(*htc_hdr));
235 if (!htc_hdr) {
236 WARN_ON_ONCE(1);
237 status = -EINVAL;
238 break;
239 }
240
241 packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
242
243 /* Endianess? */
244 put_unaligned((u16) payload_len, &htc_hdr->payld_len);
245 htc_hdr->flags = packet->info.tx.flags;
246 htc_hdr->eid = (u8) packet->endpoint;
247 htc_hdr->ctrl[0] = 0;
248 htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
249
250 spin_lock_bh(&target->tx_lock);
251
252 /* store in look up queue to match completions */
253 list_add_tail(&packet->list, &ep->pipe.tx_lookup_queue);
254 ep->ep_st.tx_issued += 1;
255 spin_unlock_bh(&target->tx_lock);
256
257 status = ath6kl_hif_pipe_send(target->dev->ar,
258 ep->pipe.pipeid_ul, NULL, skb);
259
260 if (status != 0) {
261 if (status != -ENOMEM) {
262 /* TODO: if more than 1 endpoint maps to the
263 * same PipeID, it is possible to run out of
264 * resources in the HIF layer.
265 * Don't emit the error
266 */
267 ath6kl_dbg(ATH6KL_DBG_HTC,
268 "%s: failed status:%d\n",
269 __func__, status);
270 }
271 spin_lock_bh(&target->tx_lock);
272 list_del(&packet->list);
273
274 /* reclaim credits */
275 ep->cred_dist.credits += packet->info.tx.cred_used;
276 spin_unlock_bh(&target->tx_lock);
277
278 /* put it back into the callers queue */
279 list_add(&packet->list, pkt_queue);
280 break;
281 }
282
283 }
284
285 if (status != 0) {
286 while (!list_empty(pkt_queue)) {
287 if (status != -ENOMEM) {
288 ath6kl_dbg(ATH6KL_DBG_HTC,
289 "%s: failed pkt:0x%p status:%d\n",
290 __func__, packet, status);
291 }
292
293 packet = list_first_entry(pkt_queue,
294 struct htc_packet, list);
295 list_del(&packet->list);
296 packet->status = status;
297 send_packet_completion(target, packet);
298 }
299 }
300
301 return status;
302}
303
304static enum htc_send_queue_result htc_try_send(struct htc_target *target,
305 struct htc_endpoint *ep,
306 struct list_head *txq)
307{
308 struct list_head send_queue; /* temp queue to hold packets */
309 struct htc_packet *packet, *tmp_pkt;
310 struct ath6kl *ar = target->dev->ar;
311 enum htc_send_full_action action;
312 int tx_resources, overflow, txqueue_depth, i, good_pkts;
313 u8 pipeid;
314
315 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: (queue:0x%p depth:%d)\n",
316 __func__, txq,
317 (txq == NULL) ? 0 : get_queue_depth(txq));
318
319 /* init the local send queue */
320 INIT_LIST_HEAD(&send_queue);
321
322 /*
323 * txq equals to NULL means
324 * caller didn't provide a queue, just wants us to
325 * check queues and send
326 */
327 if (txq != NULL) {
328 if (list_empty(txq)) {
329 /* empty queue */
330 return HTC_SEND_QUEUE_DROP;
331 }
332
333 spin_lock_bh(&target->tx_lock);
334 txqueue_depth = get_queue_depth(&ep->txq);
335 spin_unlock_bh(&target->tx_lock);
336
337 if (txqueue_depth >= ep->max_txq_depth) {
338 /* we've already overflowed */
339 overflow = get_queue_depth(txq);
340 } else {
341 /* get how much we will overflow by */
342 overflow = txqueue_depth;
343 overflow += get_queue_depth(txq);
344 /* get how much we will overflow the TX queue by */
345 overflow -= ep->max_txq_depth;
346 }
347
348 /* if overflow is negative or zero, we are okay */
349 if (overflow > 0) {
350 ath6kl_dbg(ATH6KL_DBG_HTC,
351 "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
352 __func__, ep->eid, overflow, txqueue_depth,
353 ep->max_txq_depth);
354 }
355 if ((overflow <= 0) ||
356 (ep->ep_cb.tx_full == NULL)) {
357 /*
358 * all packets will fit or caller did not provide send
359 * full indication handler -- just move all of them
360 * to the local send_queue object
361 */
362 list_splice_tail_init(txq, &send_queue);
363 } else {
364 good_pkts = get_queue_depth(txq) - overflow;
365 if (good_pkts < 0) {
366 WARN_ON_ONCE(1);
367 return HTC_SEND_QUEUE_DROP;
368 }
369
370 /* we have overflowed, and a callback is provided */
371 /* dequeue all non-overflow packets to the sendqueue */
372 for (i = 0; i < good_pkts; i++) {
373 /* pop off caller's queue */
374 packet = list_first_entry(txq,
375 struct htc_packet,
376 list);
377 list_del(&packet->list);
378 /* insert into local queue */
379 list_add_tail(&packet->list, &send_queue);
380 }
381
382 /*
383 * the caller's queue has all the packets that won't fit
384 * walk through the caller's queue and indicate each to
385 * the send full handler
386 */
387 list_for_each_entry_safe(packet, tmp_pkt,
388 txq, list) {
389
390 ath6kl_dbg(ATH6KL_DBG_HTC,
391 "%s: Indicat overflowed TX pkts: %p\n",
392 __func__, packet);
393 action = ep->ep_cb.tx_full(ep->target, packet);
394 if (action == HTC_SEND_FULL_DROP) {
395 /* callback wants the packet dropped */
396 ep->ep_st.tx_dropped += 1;
397
398 /* leave this one in the caller's queue
399 * for cleanup */
400 } else {
401 /* callback wants to keep this packet,
402 * remove from caller's queue */
403 list_del(&packet->list);
404 /* put it in the send queue */
405 list_add_tail(&packet->list,
406 &send_queue);
407 }
408
409 }
410
411 if (list_empty(&send_queue)) {
412 /* no packets made it in, caller will cleanup */
413 return HTC_SEND_QUEUE_DROP;
414 }
415 }
416 }
417
418 if (!ep->pipe.tx_credit_flow_enabled) {
419 tx_resources =
420 ath6kl_hif_pipe_get_free_queue_number(ar,
421 ep->pipe.pipeid_ul);
422 } else {
423 tx_resources = 0;
424 }
425
426 spin_lock_bh(&target->tx_lock);
427 if (!list_empty(&send_queue)) {
428 /* transfer packets to tail */
429 list_splice_tail_init(&send_queue, &ep->txq);
430 if (!list_empty(&send_queue)) {
431 WARN_ON_ONCE(1);
432 spin_unlock_bh(&target->tx_lock);
433 return HTC_SEND_QUEUE_DROP;
434 }
435 INIT_LIST_HEAD(&send_queue);
436 }
437
438 /* increment tx processing count on entry */
439 ep->tx_proc_cnt++;
440
441 if (ep->tx_proc_cnt > 1) {
442 /*
443 * Another thread or task is draining the TX queues on this
444 * endpoint that thread will reset the tx processing count
445 * when the queue is drained.
446 */
447 ep->tx_proc_cnt--;
448 spin_unlock_bh(&target->tx_lock);
449 return HTC_SEND_QUEUE_OK;
450 }
451
452 /***** beyond this point only 1 thread may enter ******/
453
454 /*
455 * Now drain the endpoint TX queue for transmission as long as we have
456 * enough transmit resources.
457 */
458 while (true) {
459
460 if (get_queue_depth(&ep->txq) == 0)
461 break;
462
463 if (ep->pipe.tx_credit_flow_enabled) {
464 /*
465 * Credit based mechanism provides flow control
466 * based on target transmit resource availability,
467 * we assume that the HIF layer will always have
468 * bus resources greater than target transmit
469 * resources.
470 */
471 get_htc_packet_credit_based(target, ep, &send_queue);
472 } else {
473 /*
474 * Get all packets for this endpoint that we can
475 * for this pass.
476 */
477 get_htc_packet(target, ep, &send_queue, tx_resources);
478 }
479
480 if (get_queue_depth(&send_queue) == 0) {
481 /*
482 * Didn't get packets due to out of resources or TX
483 * queue was drained.
484 */
485 break;
486 }
487
488 spin_unlock_bh(&target->tx_lock);
489
490 /* send what we can */
491 htc_issue_packets(target, ep, &send_queue);
492
493 if (!ep->pipe.tx_credit_flow_enabled) {
494 pipeid = ep->pipe.pipeid_ul;
495 tx_resources =
496 ath6kl_hif_pipe_get_free_queue_number(ar, pipeid);
497 }
498
499 spin_lock_bh(&target->tx_lock);
500
501 }
502 /* done with this endpoint, we can clear the count */
503 ep->tx_proc_cnt = 0;
504 spin_unlock_bh(&target->tx_lock);
505
506 return HTC_SEND_QUEUE_OK;
507}
508
509/* htc control packet manipulation */
510static void destroy_htc_txctrl_packet(struct htc_packet *packet)
511{
512 struct sk_buff *skb;
513 skb = packet->skb;
514 if (skb != NULL)
515 dev_kfree_skb(skb);
516
517 kfree(packet);
518}
519
520static struct htc_packet *build_htc_txctrl_packet(void)
521{
522 struct htc_packet *packet = NULL;
523 struct sk_buff *skb;
524
525 packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
526 if (packet == NULL)
527 return NULL;
528
529 skb = __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE, GFP_KERNEL);
530
531 if (skb == NULL) {
532 kfree(packet);
533 return NULL;
534 }
535 packet->skb = skb;
536
537 return packet;
538}
539
540static void htc_free_txctrl_packet(struct htc_target *target,
541 struct htc_packet *packet)
542{
543 destroy_htc_txctrl_packet(packet);
544}
545
546static struct htc_packet *htc_alloc_txctrl_packet(struct htc_target *target)
547{
548 return build_htc_txctrl_packet();
549}
550
551static void htc_txctrl_complete(struct htc_target *target,
552 struct htc_packet *packet)
553{
554 htc_free_txctrl_packet(target, packet);
555}
556
557#define MAX_MESSAGE_SIZE 1536
558
559static int htc_setup_target_buffer_assignments(struct htc_target *target)
560{
561 int status, credits, credit_per_maxmsg, i;
562 struct htc_pipe_txcredit_alloc *entry;
563 unsigned int hif_usbaudioclass = 0;
564
565 credit_per_maxmsg = MAX_MESSAGE_SIZE / target->tgt_cred_sz;
566 if (MAX_MESSAGE_SIZE % target->tgt_cred_sz)
567 credit_per_maxmsg++;
568
569 /* TODO, this should be configured by the caller! */
570
571 credits = target->tgt_creds;
572 entry = &target->pipe.txcredit_alloc[0];
573
574 status = -ENOMEM;
575
576 /* FIXME: hif_usbaudioclass is always zero */
577 if (hif_usbaudioclass) {
578 ath6kl_dbg(ATH6KL_DBG_HTC,
579 "%s: For USB Audio Class- Total:%d\n",
580 __func__, credits);
581 entry++;
582 entry++;
583 /* Setup VO Service To have Max Credits */
584 entry->service_id = WMI_DATA_VO_SVC;
585 entry->credit_alloc = (credits - 6);
586 if (entry->credit_alloc == 0)
587 entry->credit_alloc++;
588
589 credits -= (int) entry->credit_alloc;
590 if (credits <= 0)
591 return status;
592
593 entry++;
594 entry->service_id = WMI_CONTROL_SVC;
595 entry->credit_alloc = credit_per_maxmsg;
596 credits -= (int) entry->credit_alloc;
597 if (credits <= 0)
598 return status;
599
600 /* leftovers go to best effort */
601 entry++;
602 entry++;
603 entry->service_id = WMI_DATA_BE_SVC;
604 entry->credit_alloc = (u8) credits;
605 status = 0;
606 } else {
607 entry++;
608 entry->service_id = WMI_DATA_VI_SVC;
609 entry->credit_alloc = credits / 4;
610 if (entry->credit_alloc == 0)
611 entry->credit_alloc++;
612
613 credits -= (int) entry->credit_alloc;
614 if (credits <= 0)
615 return status;
616
617 entry++;
618 entry->service_id = WMI_DATA_VO_SVC;
619 entry->credit_alloc = credits / 4;
620 if (entry->credit_alloc == 0)
621 entry->credit_alloc++;
622
623 credits -= (int) entry->credit_alloc;
624 if (credits <= 0)
625 return status;
626
627 entry++;
628 entry->service_id = WMI_CONTROL_SVC;
629 entry->credit_alloc = credit_per_maxmsg;
630 credits -= (int) entry->credit_alloc;
631 if (credits <= 0)
632 return status;
633
634 entry++;
635 entry->service_id = WMI_DATA_BK_SVC;
636 entry->credit_alloc = credit_per_maxmsg;
637 credits -= (int) entry->credit_alloc;
638 if (credits <= 0)
639 return status;
640
641 /* leftovers go to best effort */
642 entry++;
643 entry->service_id = WMI_DATA_BE_SVC;
644 entry->credit_alloc = (u8) credits;
645 status = 0;
646 }
647
648 if (status == 0) {
649 for (i = 0; i < ENDPOINT_MAX; i++) {
650 if (target->pipe.txcredit_alloc[i].service_id != 0) {
651 ath6kl_dbg(ATH6KL_DBG_HTC,
652 "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
653 i,
654 target->pipe.txcredit_alloc[i].
655 service_id,
656 target->pipe.txcredit_alloc[i].
657 credit_alloc);
658 }
659 }
660 }
661 return status;
662}
663
664/* process credit reports and call distribution function */
665static void htc_process_credit_report(struct htc_target *target,
666 struct htc_credit_report *rpt,
667 int num_entries,
668 enum htc_endpoint_id from_ep)
669{
670 int total_credits = 0, i;
671 struct htc_endpoint *ep;
672
673 /* lock out TX while we update credits */
674 spin_lock_bh(&target->tx_lock);
675
676 for (i = 0; i < num_entries; i++, rpt++) {
677 if (rpt->eid >= ENDPOINT_MAX) {
678 WARN_ON_ONCE(1);
679 spin_unlock_bh(&target->tx_lock);
680 return;
681 }
682
683 ep = &target->endpoint[rpt->eid];
684 ep->cred_dist.credits += rpt->credits;
685
686 if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) {
687 spin_unlock_bh(&target->tx_lock);
688 htc_try_send(target, ep, NULL);
689 spin_lock_bh(&target->tx_lock);
690 }
691
692 total_credits += rpt->credits;
693 }
694 ath6kl_dbg(ATH6KL_DBG_HTC,
695 "Report indicated %d credits to distribute\n",
696 total_credits);
697
698 spin_unlock_bh(&target->tx_lock);
699}
700
701/* flush endpoint TX queue */
702static void htc_flush_tx_endpoint(struct htc_target *target,
703 struct htc_endpoint *ep, u16 tag)
704{
705 struct htc_packet *packet;
706
707 spin_lock_bh(&target->tx_lock);
708 while (get_queue_depth(&ep->txq)) {
709 packet = list_first_entry(&ep->txq, struct htc_packet, list);
710 list_del(&packet->list);
711 packet->status = 0;
712 send_packet_completion(target, packet);
713 }
714 spin_unlock_bh(&target->tx_lock);
715}
716
717/*
718 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
719 * since upper layers expects struct htc_packet containers we use the completed
720 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
721 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
722 * HTC.
723 */
724static struct htc_packet *htc_lookup_tx_packet(struct htc_target *target,
725 struct htc_endpoint *ep,
726 struct sk_buff *skb)
727{
728 struct htc_packet *packet, *tmp_pkt, *found_packet = NULL;
729
730 spin_lock_bh(&target->tx_lock);
731
732 /*
733 * interate from the front of tx lookup queue
734 * this lookup should be fast since lower layers completes in-order and
735 * so the completed packet should be at the head of the list generally
736 */
737 list_for_each_entry_safe(packet, tmp_pkt, &ep->pipe.tx_lookup_queue,
738 list) {
739 /* check for removal */
740 if (skb == packet->skb) {
741 /* found it */
742 list_del(&packet->list);
743 found_packet = packet;
744 break;
745 }
746 }
747
748 spin_unlock_bh(&target->tx_lock);
749
750 return found_packet;
751}
752
753static int ath6kl_htc_pipe_tx_complete(struct ath6kl *ar, struct sk_buff *skb)
754{
755 struct htc_target *target = ar->htc_target;
756 struct htc_frame_hdr *htc_hdr;
757 struct htc_endpoint *ep;
758 struct htc_packet *packet;
759 u8 ep_id, *netdata;
760 u32 netlen;
761
762 netdata = skb->data;
763 netlen = skb->len;
764
765 htc_hdr = (struct htc_frame_hdr *) netdata;
766
767 ep_id = htc_hdr->eid;
768 ep = &target->endpoint[ep_id];
769
770 packet = htc_lookup_tx_packet(target, ep, skb);
771 if (packet == NULL) {
772 /* may have already been flushed and freed */
773 ath6kl_err("HTC TX lookup failed!\n");
774 } else {
775 /* will be giving this buffer back to upper layers */
776 packet->status = 0;
777 send_packet_completion(target, packet);
778 }
779 skb = NULL;
780
781 if (!ep->pipe.tx_credit_flow_enabled) {
782 /*
783 * note: when using TX credit flow, the re-checking of queues
784 * happens when credits flow back from the target. in the
785 * non-TX credit case, we recheck after the packet completes
786 */
787 htc_try_send(target, ep, NULL);
788 }
789
790 return 0;
791}
792
793static int htc_send_packets_multiple(struct htc_target *target,
794 struct list_head *pkt_queue)
795{
796 struct htc_endpoint *ep;
797 struct htc_packet *packet, *tmp_pkt;
798
799 if (list_empty(pkt_queue))
800 return -EINVAL;
801
802 /* get first packet to find out which ep the packets will go into */
803 packet = list_first_entry(pkt_queue, struct htc_packet, list);
804
805 if (packet->endpoint >= ENDPOINT_MAX) {
806 WARN_ON_ONCE(1);
807 return -EINVAL;
808 }
809 ep = &target->endpoint[packet->endpoint];
810
811 htc_try_send(target, ep, pkt_queue);
812
813 /* do completion on any packets that couldn't get in */
814 if (!list_empty(pkt_queue)) {
815 list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
816 packet->status = -ENOMEM;
817 }
818
819 do_send_completion(ep, pkt_queue);
820 }
821
822 return 0;
823}
824
825/* htc pipe rx path */
826static struct htc_packet *alloc_htc_packet_container(struct htc_target *target)
827{
828 struct htc_packet *packet;
829 spin_lock_bh(&target->rx_lock);
830
831 if (target->pipe.htc_packet_pool == NULL) {
832 spin_unlock_bh(&target->rx_lock);
833 return NULL;
834 }
835
836 packet = target->pipe.htc_packet_pool;
837 target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
838
839 spin_unlock_bh(&target->rx_lock);
840
841 packet->list.next = NULL;
842 return packet;
843}
844
845static void free_htc_packet_container(struct htc_target *target,
846 struct htc_packet *packet)
847{
848 struct list_head *lh;
849
850 spin_lock_bh(&target->rx_lock);
851
852 if (target->pipe.htc_packet_pool == NULL) {
853 target->pipe.htc_packet_pool = packet;
854 packet->list.next = NULL;
855 } else {
856 lh = (struct list_head *) target->pipe.htc_packet_pool;
857 packet->list.next = lh;
858 target->pipe.htc_packet_pool = packet;
859 }
860
861 spin_unlock_bh(&target->rx_lock);
862}
863
864static int htc_process_trailer(struct htc_target *target, u8 *buffer,
865 int len, enum htc_endpoint_id from_ep)
866{
867 struct htc_credit_report *report;
868 struct htc_record_hdr *record;
869 u8 *record_buf, *orig_buf;
870 int orig_len, status;
871
872 orig_buf = buffer;
873 orig_len = len;
874 status = 0;
875
876 while (len > 0) {
877 if (len < sizeof(struct htc_record_hdr)) {
878 status = -EINVAL;
879 break;
880 }
881
882 /* these are byte aligned structs */
883 record = (struct htc_record_hdr *) buffer;
884 len -= sizeof(struct htc_record_hdr);
885 buffer += sizeof(struct htc_record_hdr);
886
887 if (record->len > len) {
888 /* no room left in buffer for record */
889 ath6kl_dbg(ATH6KL_DBG_HTC,
890 "invalid length: %d (id:%d) buffer has: %d bytes left\n",
891 record->len, record->rec_id, len);
892 status = -EINVAL;
893 break;
894 }
895
896 /* start of record follows the header */
897 record_buf = buffer;
898
899 switch (record->rec_id) {
900 case HTC_RECORD_CREDITS:
901 if (record->len < sizeof(struct htc_credit_report)) {
902 WARN_ON_ONCE(1);
903 return -EINVAL;
904 }
905
906 report = (struct htc_credit_report *) record_buf;
907 htc_process_credit_report(target, report,
908 record->len / sizeof(*report),
909 from_ep);
910 break;
911 default:
912 ath6kl_dbg(ATH6KL_DBG_HTC,
913 "unhandled record: id:%d length:%d\n",
914 record->rec_id, record->len);
915 break;
916 }
917
918 if (status != 0)
919 break;
920
921 /* advance buffer past this record for next time around */
922 buffer += record->len;
923 len -= record->len;
924 }
925
926 return status;
927}
928
929static void do_recv_completion(struct htc_endpoint *ep,
930 struct list_head *queue_to_indicate)
931{
932 struct htc_packet *packet;
933
934 if (list_empty(queue_to_indicate)) {
935 /* nothing to indicate */
936 return;
937 }
938
939 /* using legacy EpRecv */
940 while (!list_empty(queue_to_indicate)) {
941 packet = list_first_entry(queue_to_indicate,
942 struct htc_packet, list);
943 list_del(&packet->list);
944 ep->ep_cb.rx(ep->target, packet);
945 }
946
947 return;
948}
949
950static void recv_packet_completion(struct htc_target *target,
951 struct htc_endpoint *ep,
952 struct htc_packet *packet)
953{
954 struct list_head container;
955 INIT_LIST_HEAD(&container);
956 list_add_tail(&packet->list, &container);
957
958 /* do completion */
959 do_recv_completion(ep, &container);
960}
961
962static int ath6kl_htc_pipe_rx_complete(struct ath6kl *ar, struct sk_buff *skb,
963 u8 pipeid)
964{
965 struct htc_target *target = ar->htc_target;
966 u8 *netdata, *trailer, hdr_info;
967 struct htc_frame_hdr *htc_hdr;
968 u32 netlen, trailerlen = 0;
969 struct htc_packet *packet;
970 struct htc_endpoint *ep;
971 u16 payload_len;
972 int status = 0;
973
974 netdata = skb->data;
975 netlen = skb->len;
976
977 htc_hdr = (struct htc_frame_hdr *) netdata;
978
979 ep = &target->endpoint[htc_hdr->eid];
980
981 if (htc_hdr->eid >= ENDPOINT_MAX) {
982 ath6kl_dbg(ATH6KL_DBG_HTC,
983 "HTC Rx: invalid EndpointID=%d\n",
984 htc_hdr->eid);
985 status = -EINVAL;
986 goto free_skb;
987 }
988
989 payload_len = le16_to_cpu(get_unaligned(&htc_hdr->payld_len));
990
991 if (netlen < (payload_len + HTC_HDR_LENGTH)) {
992 ath6kl_dbg(ATH6KL_DBG_HTC,
993 "HTC Rx: insufficient length, got:%d expected =%u\n",
994 netlen, payload_len + HTC_HDR_LENGTH);
995 status = -EINVAL;
996 goto free_skb;
997 }
998
999 /* get flags to check for trailer */
1000 hdr_info = htc_hdr->flags;
1001 if (hdr_info & HTC_FLG_RX_TRAILER) {
1002 /* extract the trailer length */
1003 hdr_info = htc_hdr->ctrl[0];
1004 if ((hdr_info < sizeof(struct htc_record_hdr)) ||
1005 (hdr_info > payload_len)) {
1006 ath6kl_dbg(ATH6KL_DBG_HTC,
1007 "invalid header: payloadlen should be %d, CB[0]: %d\n",
1008 payload_len, hdr_info);
1009 status = -EINVAL;
1010 goto free_skb;
1011 }
1012
1013 trailerlen = hdr_info;
1014 /* process trailer after hdr/apps payload */
1015 trailer = (u8 *) htc_hdr + HTC_HDR_LENGTH +
1016 payload_len - hdr_info;
1017 status = htc_process_trailer(target, trailer, hdr_info,
1018 htc_hdr->eid);
1019 if (status != 0)
1020 goto free_skb;
1021 }
1022
1023 if (((int) payload_len - (int) trailerlen) <= 0) {
1024 /* zero length packet with trailer, just drop these */
1025 goto free_skb;
1026 }
1027
1028 if (htc_hdr->eid == ENDPOINT_0) {
1029 /* handle HTC control message */
1030 if (target->htc_flags & HTC_OP_STATE_SETUP_COMPLETE) {
1031 /*
1032 * fatal: target should not send unsolicited
1033 * messageson the endpoint 0
1034 */
1035 ath6kl_dbg(ATH6KL_DBG_HTC,
1036 "HTC ignores Rx Ctrl after setup complete\n");
1037 status = -EINVAL;
1038 goto free_skb;
1039 }
1040
1041 /* remove HTC header */
1042 skb_pull(skb, HTC_HDR_LENGTH);
1043
1044 netdata = skb->data;
1045 netlen = skb->len;
1046
1047 spin_lock_bh(&target->rx_lock);
1048
1049 target->pipe.ctrl_response_valid = true;
1050 target->pipe.ctrl_response_len = min_t(int, netlen,
1051 HTC_MAX_CTRL_MSG_LEN);
1052 memcpy(target->pipe.ctrl_response_buf, netdata,
1053 target->pipe.ctrl_response_len);
1054
1055 spin_unlock_bh(&target->rx_lock);
1056
1057 dev_kfree_skb(skb);
1058 skb = NULL;
1059 goto free_skb;
1060 }
1061
1062 /*
1063 * TODO: the message based HIF architecture allocates net bufs
1064 * for recv packets since it bridges that HIF to upper layers,
1065 * which expects HTC packets, we form the packets here
1066 */
1067 packet = alloc_htc_packet_container(target);
1068 if (packet == NULL) {
1069 status = -ENOMEM;
1070 goto free_skb;
1071 }
1072
1073 packet->status = 0;
1074 packet->endpoint = htc_hdr->eid;
1075 packet->pkt_cntxt = skb;
1076
1077 /* TODO: for backwards compatibility */
1078 packet->buf = skb_push(skb, 0) + HTC_HDR_LENGTH;
1079 packet->act_len = netlen - HTC_HDR_LENGTH - trailerlen;
1080
1081 /*
1082 * TODO: this is a hack because the driver layer will set the
1083 * actual len of the skb again which will just double the len
1084 */
1085 skb_trim(skb, 0);
1086
1087 recv_packet_completion(target, ep, packet);
1088
1089 /* recover the packet container */
1090 free_htc_packet_container(target, packet);
1091 skb = NULL;
1092
1093free_skb:
1094 if (skb != NULL)
1095 dev_kfree_skb(skb);
1096
1097 return status;
1098
1099}
1100
1101static void htc_flush_rx_queue(struct htc_target *target,
1102 struct htc_endpoint *ep)
1103{
1104 struct list_head container;
1105 struct htc_packet *packet;
1106
1107 spin_lock_bh(&target->rx_lock);
1108
1109 while (1) {
1110 if (list_empty(&ep->rx_bufq))
1111 break;
1112
1113 packet = list_first_entry(&ep->rx_bufq,
1114 struct htc_packet, list);
1115 list_del(&packet->list);
1116
1117 spin_unlock_bh(&target->rx_lock);
1118 packet->status = -ECANCELED;
1119 packet->act_len = 0;
1120
1121 ath6kl_dbg(ATH6KL_DBG_HTC,
1122 "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1123 packet, packet->buf_len,
1124 packet->endpoint);
1125
1126 INIT_LIST_HEAD(&container);
1127 list_add_tail(&packet->list, &container);
1128
1129 /* give the packet back */
1130 do_recv_completion(ep, &container);
1131 spin_lock_bh(&target->rx_lock);
1132 }
1133
1134 spin_unlock_bh(&target->rx_lock);
1135}
1136
1137/* polling routine to wait for a control packet to be received */
1138static int htc_wait_recv_ctrl_message(struct htc_target *target)
1139{
1140 int count = HTC_TARGET_RESPONSE_POLL_COUNT;
1141
1142 while (count > 0) {
1143 spin_lock_bh(&target->rx_lock);
1144
1145 if (target->pipe.ctrl_response_valid) {
1146 target->pipe.ctrl_response_valid = false;
1147 spin_unlock_bh(&target->rx_lock);
1148 break;
1149 }
1150
1151 spin_unlock_bh(&target->rx_lock);
1152
1153 count--;
1154
1155 msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT);
1156 }
1157
1158 if (count <= 0) {
1159 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: Timeout!\n", __func__);
1160 return -ECOMM;
1161 }
1162
1163 return 0;
1164}
1165
1166static void htc_rxctrl_complete(struct htc_target *context,
1167 struct htc_packet *packet)
1168{
1169 /* TODO, can't really receive HTC control messages yet.... */
1170 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: invalid call function\n", __func__);
1171}
1172
1173/* htc pipe initialization */
1174static void reset_endpoint_states(struct htc_target *target)
1175{
1176 struct htc_endpoint *ep;
1177 int i;
1178
1179 for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1180 ep = &target->endpoint[i];
1181 ep->svc_id = 0;
1182 ep->len_max = 0;
1183 ep->max_txq_depth = 0;
1184 ep->eid = i;
1185 INIT_LIST_HEAD(&ep->txq);
1186 INIT_LIST_HEAD(&ep->pipe.tx_lookup_queue);
1187 INIT_LIST_HEAD(&ep->rx_bufq);
1188 ep->target = target;
1189 ep->pipe.tx_credit_flow_enabled = (bool) 1; /* FIXME */
1190 }
1191}
1192
1193/* start HTC, this is called after all services are connected */
1194static int htc_config_target_hif_pipe(struct htc_target *target)
1195{
1196 return 0;
1197}
1198
1199/* htc service functions */
1200static u8 htc_get_credit_alloc(struct htc_target *target, u16 service_id)
1201{
1202 u8 allocation = 0;
1203 int i;
1204
1205 for (i = 0; i < ENDPOINT_MAX; i++) {
1206 if (target->pipe.txcredit_alloc[i].service_id == service_id)
1207 allocation =
1208 target->pipe.txcredit_alloc[i].credit_alloc;
1209 }
1210
1211 if (allocation == 0) {
1212 ath6kl_dbg(ATH6KL_DBG_HTC,
1213 "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1214 service_id);
1215 }
1216
1217 return allocation;
1218}
1219
1220static int ath6kl_htc_pipe_conn_service(struct htc_target *target,
1221 struct htc_service_connect_req *conn_req,
1222 struct htc_service_connect_resp *conn_resp)
1223{
1224 struct ath6kl *ar = target->dev->ar;
1225 struct htc_packet *packet = NULL;
1226 struct htc_conn_service_resp *resp_msg;
1227 struct htc_conn_service_msg *conn_msg;
1228 enum htc_endpoint_id assigned_epid = ENDPOINT_MAX;
1229 bool disable_credit_flowctrl = false;
1230 unsigned int max_msg_size = 0;
1231 struct htc_endpoint *ep;
1232 int length, status = 0;
1233 struct sk_buff *skb;
1234 u8 tx_alloc;
1235 u16 flags;
1236
1237 if (conn_req->svc_id == 0) {
1238 WARN_ON_ONCE(1);
1239 status = -EINVAL;
1240 goto free_packet;
1241 }
1242
1243 if (conn_req->svc_id == HTC_CTRL_RSVD_SVC) {
1244 /* special case for pseudo control service */
1245 assigned_epid = ENDPOINT_0;
1246 max_msg_size = HTC_MAX_CTRL_MSG_LEN;
1247 tx_alloc = 0;
1248
1249 } else {
1250
1251 tx_alloc = htc_get_credit_alloc(target, conn_req->svc_id);
1252 if (tx_alloc == 0) {
1253 status = -ENOMEM;
1254 goto free_packet;
1255 }
1256
1257 /* allocate a packet to send to the target */
1258 packet = htc_alloc_txctrl_packet(target);
1259
1260 if (packet == NULL) {
1261 WARN_ON_ONCE(1);
1262 status = -ENOMEM;
1263 goto free_packet;
1264 }
1265
1266 skb = packet->skb;
1267 length = sizeof(struct htc_conn_service_msg);
1268
1269 /* assemble connect service message */
1270 conn_msg = (struct htc_conn_service_msg *) skb_put(skb,
1271 length);
1272 if (conn_msg == NULL) {
1273 WARN_ON_ONCE(1);
1274 status = -EINVAL;
1275 goto free_packet;
1276 }
1277
1278 memset(conn_msg, 0,
1279 sizeof(struct htc_conn_service_msg));
1280 conn_msg->msg_id = cpu_to_le16(HTC_MSG_CONN_SVC_ID);
1281 conn_msg->svc_id = cpu_to_le16(conn_req->svc_id);
1282 conn_msg->conn_flags = cpu_to_le16(conn_req->conn_flags &
1283 ~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK);
1284
1285 /* tell target desired recv alloc for this ep */
1286 flags = tx_alloc << HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT;
1287 conn_msg->conn_flags |= cpu_to_le16(flags);
1288
1289 if (conn_req->conn_flags &
1290 HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL) {
1291 disable_credit_flowctrl = true;
1292 }
1293
1294 set_htc_pkt_info(packet, NULL, (u8 *) conn_msg,
1295 length,
1296 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1297
1298 status = ath6kl_htc_pipe_tx(target, packet);
1299
1300 /* we don't own it anymore */
1301 packet = NULL;
1302 if (status != 0)
1303 goto free_packet;
1304
1305 /* wait for response */
1306 status = htc_wait_recv_ctrl_message(target);
1307 if (status != 0)
1308 goto free_packet;
1309
1310 /* we controlled the buffer creation so it has to be
1311 * properly aligned
1312 */
1313 resp_msg = (struct htc_conn_service_resp *)
1314 target->pipe.ctrl_response_buf;
1315
1316 if (resp_msg->msg_id != cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID) ||
1317 (target->pipe.ctrl_response_len < sizeof(*resp_msg))) {
1318 /* this message is not valid */
1319 WARN_ON_ONCE(1);
1320 status = -EINVAL;
1321 goto free_packet;
1322 }
1323
1324 ath6kl_dbg(ATH6KL_DBG_TRC,
1325 "%s: service 0x%X conn resp: status: %d ep: %d\n",
1326 __func__, resp_msg->svc_id, resp_msg->status,
1327 resp_msg->eid);
1328
1329 conn_resp->resp_code = resp_msg->status;
1330 /* check response status */
1331 if (resp_msg->status != HTC_SERVICE_SUCCESS) {
1332 ath6kl_dbg(ATH6KL_DBG_HTC,
1333 "Target failed service 0x%X connect request (status:%d)\n",
1334 resp_msg->svc_id, resp_msg->status);
1335 status = -EINVAL;
1336 goto free_packet;
1337 }
1338
1339 assigned_epid = (enum htc_endpoint_id) resp_msg->eid;
1340 max_msg_size = le16_to_cpu(resp_msg->max_msg_sz);
1341 }
1342
1343 /* the rest are parameter checks so set the error status */
1344 status = -EINVAL;
1345
1346 if (assigned_epid >= ENDPOINT_MAX) {
1347 WARN_ON_ONCE(1);
1348 goto free_packet;
1349 }
1350
1351 if (max_msg_size == 0) {
1352 WARN_ON_ONCE(1);
1353 goto free_packet;
1354 }
1355
1356 ep = &target->endpoint[assigned_epid];
1357 ep->eid = assigned_epid;
1358 if (ep->svc_id != 0) {
1359 /* endpoint already in use! */
1360 WARN_ON_ONCE(1);
1361 goto free_packet;
1362 }
1363
1364 /* return assigned endpoint to caller */
1365 conn_resp->endpoint = assigned_epid;
1366 conn_resp->len_max = max_msg_size;
1367
1368 /* setup the endpoint */
1369 ep->svc_id = conn_req->svc_id; /* this marks ep in use */
1370 ep->max_txq_depth = conn_req->max_txq_depth;
1371 ep->len_max = max_msg_size;
1372 ep->cred_dist.credits = tx_alloc;
1373 ep->cred_dist.cred_sz = target->tgt_cred_sz;
1374 ep->cred_dist.cred_per_msg = max_msg_size / target->tgt_cred_sz;
1375 if (max_msg_size % target->tgt_cred_sz)
1376 ep->cred_dist.cred_per_msg++;
1377
1378 /* copy all the callbacks */
1379 ep->ep_cb = conn_req->ep_cb;
1380
1381 /* initialize tx_drop_packet_threshold */
1382 ep->tx_drop_packet_threshold = MAX_HI_COOKIE_NUM;
1383
1384 status = ath6kl_hif_pipe_map_service(ar, ep->svc_id,
1385 &ep->pipe.pipeid_ul,
1386 &ep->pipe.pipeid_dl);
1387 if (status != 0)
1388 goto free_packet;
1389
1390 ath6kl_dbg(ATH6KL_DBG_HTC,
1391 "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1392 ep->svc_id, ep->pipe.pipeid_ul,
1393 ep->pipe.pipeid_dl, ep->eid);
1394
1395 if (disable_credit_flowctrl && ep->pipe.tx_credit_flow_enabled) {
1396 ep->pipe.tx_credit_flow_enabled = false;
1397 ath6kl_dbg(ATH6KL_DBG_HTC,
1398 "SVC: 0x%4.4X ep:%d TX flow control off\n",
1399 ep->svc_id, assigned_epid);
1400 }
1401
1402free_packet:
1403 if (packet != NULL)
1404 htc_free_txctrl_packet(target, packet);
1405 return status;
1406}
1407
1408/* htc export functions */
1409static void *ath6kl_htc_pipe_create(struct ath6kl *ar)
1410{
1411 int status = 0;
1412 struct htc_endpoint *ep = NULL;
1413 struct htc_target *target = NULL;
1414 struct htc_packet *packet;
1415 int i;
1416
1417 target = kzalloc(sizeof(struct htc_target), GFP_KERNEL);
1418 if (target == NULL) {
1419 ath6kl_err("htc create unable to allocate memory\n");
1420 status = -ENOMEM;
1421 goto fail_htc_create;
1422 }
1423
1424 spin_lock_init(&target->htc_lock);
1425 spin_lock_init(&target->rx_lock);
1426 spin_lock_init(&target->tx_lock);
1427
1428 reset_endpoint_states(target);
1429
1430 for (i = 0; i < HTC_PACKET_CONTAINER_ALLOCATION; i++) {
1431 packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
1432
1433 if (packet != NULL)
1434 free_htc_packet_container(target, packet);
1435 }
1436
1437 target->dev = kzalloc(sizeof(*target->dev), GFP_KERNEL);
1438 if (!target->dev) {
1439 ath6kl_err("unable to allocate memory\n");
1440 status = -ENOMEM;
1441 goto fail_htc_create;
1442 }
1443 target->dev->ar = ar;
1444 target->dev->htc_cnxt = target;
1445
1446 /* Get HIF default pipe for HTC message exchange */
1447 ep = &target->endpoint[ENDPOINT_0];
1448
1449 ath6kl_hif_pipe_get_default(ar, &ep->pipe.pipeid_ul,
1450 &ep->pipe.pipeid_dl);
1451
1452 return target;
1453
1454fail_htc_create:
1455 if (status != 0) {
1456 if (target != NULL)
1457 ath6kl_htc_pipe_cleanup(target);
1458
1459 target = NULL;
1460 }
1461 return target;
1462}
1463
1464/* cleanup the HTC instance */
1465static void ath6kl_htc_pipe_cleanup(struct htc_target *target)
1466{
1467 struct htc_packet *packet;
1468
1469 while (true) {
1470 packet = alloc_htc_packet_container(target);
1471 if (packet == NULL)
1472 break;
1473 kfree(packet);
1474 }
1475
1476 kfree(target->dev);
1477
1478 /* kfree our instance */
1479 kfree(target);
1480}
1481
1482static int ath6kl_htc_pipe_start(struct htc_target *target)
1483{
1484 struct sk_buff *skb;
1485 struct htc_setup_comp_ext_msg *setup;
1486 struct htc_packet *packet;
1487
1488 htc_config_target_hif_pipe(target);
1489
1490 /* allocate a buffer to send */
1491 packet = htc_alloc_txctrl_packet(target);
1492 if (packet == NULL) {
1493 WARN_ON_ONCE(1);
1494 return -ENOMEM;
1495 }
1496
1497 skb = packet->skb;
1498
1499 /* assemble setup complete message */
1500 setup = (struct htc_setup_comp_ext_msg *) skb_put(skb,
1501 sizeof(*setup));
1502 memset(setup, 0, sizeof(struct htc_setup_comp_ext_msg));
1503 setup->msg_id = cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID);
1504
1505 ath6kl_dbg(ATH6KL_DBG_HTC, "HTC using TX credit flow control\n");
1506
1507 set_htc_pkt_info(packet, NULL, (u8 *) setup,
1508 sizeof(struct htc_setup_comp_ext_msg),
1509 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1510
1511 target->htc_flags |= HTC_OP_STATE_SETUP_COMPLETE;
1512
1513 return ath6kl_htc_pipe_tx(target, packet);
1514}
1515
1516static void ath6kl_htc_pipe_stop(struct htc_target *target)
1517{
1518 int i;
1519 struct htc_endpoint *ep;
1520
1521 /* cleanup endpoints */
1522 for (i = 0; i < ENDPOINT_MAX; i++) {
1523 ep = &target->endpoint[i];
1524 htc_flush_rx_queue(target, ep);
1525 htc_flush_tx_endpoint(target, ep, HTC_TX_PACKET_TAG_ALL);
1526 }
1527
1528 reset_endpoint_states(target);
1529 target->htc_flags &= ~HTC_OP_STATE_SETUP_COMPLETE;
1530}
1531
1532static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target *target,
1533 enum htc_endpoint_id endpoint)
1534{
1535 int num;
1536
1537 spin_lock_bh(&target->rx_lock);
1538 num = get_queue_depth(&(target->endpoint[endpoint].rx_bufq));
1539 spin_unlock_bh(&target->rx_lock);
1540
1541 return num;
1542}
1543
1544static int ath6kl_htc_pipe_tx(struct htc_target *target,
1545 struct htc_packet *packet)
1546{
1547 struct list_head queue;
1548
1549 ath6kl_dbg(ATH6KL_DBG_HTC,
1550 "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1551 __func__, packet->endpoint, packet->buf,
1552 packet->act_len);
1553
1554 INIT_LIST_HEAD(&queue);
1555 list_add_tail(&packet->list, &queue);
1556
1557 return htc_send_packets_multiple(target, &queue);
1558}
1559
1560static int ath6kl_htc_pipe_wait_target(struct htc_target *target)
1561{
1562 struct htc_ready_ext_msg *ready_msg;
1563 struct htc_service_connect_req connect;
1564 struct htc_service_connect_resp resp;
1565 int status = 0;
1566
1567 status = htc_wait_recv_ctrl_message(target);
1568
1569 if (status != 0)
1570 return status;
1571
1572 if (target->pipe.ctrl_response_len < sizeof(*ready_msg)) {
1573 ath6kl_dbg(ATH6KL_DBG_HTC, "invalid htc ready msg len:%d!\n",
1574 target->pipe.ctrl_response_len);
1575 return -ECOMM;
1576 }
1577
1578 ready_msg = (struct htc_ready_ext_msg *) target->pipe.ctrl_response_buf;
1579
1580 if (ready_msg->ver2_0_info.msg_id != cpu_to_le16(HTC_MSG_READY_ID)) {
1581 ath6kl_dbg(ATH6KL_DBG_HTC, "invalid htc ready msg : 0x%X !\n",
1582 ready_msg->ver2_0_info.msg_id);
1583 return -ECOMM;
1584 }
1585
1586 ath6kl_dbg(ATH6KL_DBG_HTC,
1587 "Target Ready! : transmit resources : %d size:%d\n",
1588 ready_msg->ver2_0_info.cred_cnt,
1589 ready_msg->ver2_0_info.cred_sz);
1590
1591 target->tgt_creds = le16_to_cpu(ready_msg->ver2_0_info.cred_cnt);
1592 target->tgt_cred_sz = le16_to_cpu(ready_msg->ver2_0_info.cred_sz);
1593
1594 if ((target->tgt_creds == 0) || (target->tgt_cred_sz == 0))
1595 return -ECOMM;
1596
1597 htc_setup_target_buffer_assignments(target);
1598
1599 /* setup our pseudo HTC control endpoint connection */
1600 memset(&connect, 0, sizeof(connect));
1601 memset(&resp, 0, sizeof(resp));
1602 connect.ep_cb.tx_complete = htc_txctrl_complete;
1603 connect.ep_cb.rx = htc_rxctrl_complete;
1604 connect.max_txq_depth = NUM_CONTROL_TX_BUFFERS;
1605 connect.svc_id = HTC_CTRL_RSVD_SVC;
1606
1607 /* connect fake service */
1608 status = ath6kl_htc_pipe_conn_service(target, &connect, &resp);
1609
1610 return status;
1611}
1612
1613static void ath6kl_htc_pipe_flush_txep(struct htc_target *target,
1614 enum htc_endpoint_id endpoint, u16 tag)
1615{
1616 struct htc_endpoint *ep = &target->endpoint[endpoint];
1617
1618 if (ep->svc_id == 0) {
1619 WARN_ON_ONCE(1);
1620 /* not in use.. */
1621 return;
1622 }
1623
1624 htc_flush_tx_endpoint(target, ep, tag);
1625}
1626
1627static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target *target,
1628 struct list_head *pkt_queue)
1629{
1630 struct htc_packet *packet, *tmp_pkt, *first;
1631 struct htc_endpoint *ep;
1632 int status = 0;
1633
1634 if (list_empty(pkt_queue))
1635 return -EINVAL;
1636
1637 first = list_first_entry(pkt_queue, struct htc_packet, list);
1638
1639 if (first->endpoint >= ENDPOINT_MAX) {
1640 WARN_ON_ONCE(1);
1641 return -EINVAL;
1642 }
1643
1644 ath6kl_dbg(ATH6KL_DBG_HTC, "%s: epid: %d, cnt:%d, len: %d\n",
1645 __func__, first->endpoint, get_queue_depth(pkt_queue),
1646 first->buf_len);
1647
1648 ep = &target->endpoint[first->endpoint];
1649
1650 spin_lock_bh(&target->rx_lock);
1651
1652 /* store receive packets */
1653 list_splice_tail_init(pkt_queue, &ep->rx_bufq);
1654
1655 spin_unlock_bh(&target->rx_lock);
1656
1657 if (status != 0) {
1658 /* walk through queue and mark each one canceled */
1659 list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
1660 packet->status = -ECANCELED;
1661 }
1662
1663 do_recv_completion(ep, pkt_queue);
1664 }
1665
1666 return status;
1667}
1668
1669static void ath6kl_htc_pipe_activity_changed(struct htc_target *target,
1670 enum htc_endpoint_id ep,
1671 bool active)
1672{
1673 /* TODO */
1674}
1675
1676static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target *target)
1677{
1678 /* TODO */
1679}
1680
1681static int ath6kl_htc_pipe_credit_setup(struct htc_target *target,
1682 struct ath6kl_htc_credit_info *info)
1683{
1684 return 0;
1685}
1686
1687static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops = {
1688 .create = ath6kl_htc_pipe_create,
1689 .wait_target = ath6kl_htc_pipe_wait_target,
1690 .start = ath6kl_htc_pipe_start,
1691 .conn_service = ath6kl_htc_pipe_conn_service,
1692 .tx = ath6kl_htc_pipe_tx,
1693 .stop = ath6kl_htc_pipe_stop,
1694 .cleanup = ath6kl_htc_pipe_cleanup,
1695 .flush_txep = ath6kl_htc_pipe_flush_txep,
1696 .flush_rx_buf = ath6kl_htc_pipe_flush_rx_buf,
1697 .activity_changed = ath6kl_htc_pipe_activity_changed,
1698 .get_rxbuf_num = ath6kl_htc_pipe_get_rxbuf_num,
1699 .add_rxbuf_multiple = ath6kl_htc_pipe_add_rxbuf_multiple,
1700 .credit_setup = ath6kl_htc_pipe_credit_setup,
1701 .tx_complete = ath6kl_htc_pipe_tx_complete,
1702 .rx_complete = ath6kl_htc_pipe_rx_complete,
1703};
1704
1705void ath6kl_htc_pipe_attach(struct ath6kl *ar)
1706{
1707 ar->htc_ops = &ath6kl_htc_pipe_ops;
1708}