Linux Audio

Check our new training course

Loading...
v6.8
   1/*
   2 * Copyright (c) 2007-2011 Atheros Communications Inc.
   3 *
   4 * Permission to use, copy, modify, and/or distribute this software for any
   5 * purpose with or without fee is hereby granted, provided that the above
   6 * copyright notice and this permission notice appear in all copies.
   7 *
   8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
   9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
  10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
  11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
  12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
  13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
  14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
  15 */
  16
  17#include "core.h"
  18#include "debug.h"
  19#include "hif-ops.h"
  20
  21#define HTC_PACKET_CONTAINER_ALLOCATION 32
  22#define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
  23
  24static int ath6kl_htc_pipe_tx(struct htc_target *handle,
  25			      struct htc_packet *packet);
  26static void ath6kl_htc_pipe_cleanup(struct htc_target *handle);
  27
  28/* htc pipe tx path */
  29static inline void restore_tx_packet(struct htc_packet *packet)
  30{
  31	if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
  32		skb_pull(packet->skb, sizeof(struct htc_frame_hdr));
  33		packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
  34	}
  35}
  36
  37static void do_send_completion(struct htc_endpoint *ep,
  38			       struct list_head *queue_to_indicate)
  39{
  40	struct htc_packet *packet;
  41
  42	if (list_empty(queue_to_indicate)) {
  43		/* nothing to indicate */
  44		return;
  45	}
  46
  47	if (ep->ep_cb.tx_comp_multi != NULL) {
  48		ath6kl_dbg(ATH6KL_DBG_HTC,
  49			   "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
  50			   __func__, ep->eid,
  51			   get_queue_depth(queue_to_indicate));
  52		/*
  53		 * a multiple send complete handler is being used,
  54		 * pass the queue to the handler
  55		 */
  56		ep->ep_cb.tx_comp_multi(ep->target, queue_to_indicate);
  57		/*
  58		 * all packets are now owned by the callback,
  59		 * reset queue to be safe
  60		 */
  61		INIT_LIST_HEAD(queue_to_indicate);
  62	} else {
  63		/* using legacy EpTxComplete */
  64		do {
  65			packet = list_first_entry(queue_to_indicate,
  66						  struct htc_packet, list);
  67
  68			list_del(&packet->list);
  69			ath6kl_dbg(ATH6KL_DBG_HTC,
  70				   "%s: calling ep %d send complete callback on packet 0x%p\n",
  71				   __func__, ep->eid, packet);
  72			ep->ep_cb.tx_complete(ep->target, packet);
  73		} while (!list_empty(queue_to_indicate));
  74	}
  75}
  76
  77static void send_packet_completion(struct htc_target *target,
  78				   struct htc_packet *packet)
  79{
  80	struct htc_endpoint *ep = &target->endpoint[packet->endpoint];
  81	struct list_head container;
  82
  83	restore_tx_packet(packet);
  84	INIT_LIST_HEAD(&container);
  85	list_add_tail(&packet->list, &container);
  86
  87	/* do completion */
  88	do_send_completion(ep, &container);
  89}
  90
  91static void get_htc_packet_credit_based(struct htc_target *target,
  92					struct htc_endpoint *ep,
  93					struct list_head *queue)
  94{
  95	int credits_required;
  96	int remainder;
  97	u8 send_flags;
  98	struct htc_packet *packet;
  99	unsigned int transfer_len;
 100
 101	/* NOTE : the TX lock is held when this function is called */
 102
 103	/* loop until we can grab as many packets out of the queue as we can */
 104	while (true) {
 105		send_flags = 0;
 106		if (list_empty(&ep->txq))
 107			break;
 108
 109		/* get packet at head, but don't remove it */
 110		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 111
 112		ath6kl_dbg(ATH6KL_DBG_HTC,
 113			   "%s: got head packet:0x%p , queue depth: %d\n",
 114			   __func__, packet, get_queue_depth(&ep->txq));
 115
 116		transfer_len = packet->act_len + HTC_HDR_LENGTH;
 117
 118		if (transfer_len <= target->tgt_cred_sz) {
 119			credits_required = 1;
 120		} else {
 121			/* figure out how many credits this message requires */
 122			credits_required = transfer_len / target->tgt_cred_sz;
 123			remainder = transfer_len % target->tgt_cred_sz;
 124
 125			if (remainder)
 126				credits_required++;
 127		}
 128
 129		ath6kl_dbg(ATH6KL_DBG_HTC, "%s: creds required:%d got:%d\n",
 130			   __func__, credits_required, ep->cred_dist.credits);
 131
 132		if (ep->eid == ENDPOINT_0) {
 133			/*
 134			 * endpoint 0 is special, it always has a credit and
 135			 * does not require credit based flow control
 136			 */
 137			credits_required = 0;
 138
 139		} else {
 
 140			if (ep->cred_dist.credits < credits_required)
 141				break;
 142
 143			ep->cred_dist.credits -= credits_required;
 144			ep->ep_st.cred_cosumd += credits_required;
 145
 146			/* check if we need credits back from the target */
 147			if (ep->cred_dist.credits <
 148					ep->cred_dist.cred_per_msg) {
 149				/* tell the target we need credits ASAP! */
 150				send_flags |= HTC_FLAGS_NEED_CREDIT_UPDATE;
 151				ep->ep_st.cred_low_indicate += 1;
 152				ath6kl_dbg(ATH6KL_DBG_HTC,
 153					   "%s: host needs credits\n",
 154					   __func__);
 155			}
 156		}
 157
 158		/* now we can fully dequeue */
 159		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 160
 161		list_del(&packet->list);
 162		/* save the number of credits this packet consumed */
 163		packet->info.tx.cred_used = credits_required;
 164		/* save send flags */
 165		packet->info.tx.flags = send_flags;
 166		packet->info.tx.seqno = ep->seqno;
 167		ep->seqno++;
 168		/* queue this packet into the caller's queue */
 169		list_add_tail(&packet->list, queue);
 170	}
 
 171}
 172
 173static void get_htc_packet(struct htc_target *target,
 174			   struct htc_endpoint *ep,
 175			   struct list_head *queue, int resources)
 176{
 177	struct htc_packet *packet;
 178
 179	/* NOTE : the TX lock is held when this function is called */
 180
 181	/* loop until we can grab as many packets out of the queue as we can */
 182	while (resources) {
 183		if (list_empty(&ep->txq))
 184			break;
 185
 186		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 187		list_del(&packet->list);
 188
 189		ath6kl_dbg(ATH6KL_DBG_HTC,
 190			   "%s: got packet:0x%p , new queue depth: %d\n",
 191			   __func__, packet, get_queue_depth(&ep->txq));
 192		packet->info.tx.seqno = ep->seqno;
 193		packet->info.tx.flags = 0;
 194		packet->info.tx.cred_used = 0;
 195		ep->seqno++;
 196
 197		/* queue this packet into the caller's queue */
 198		list_add_tail(&packet->list, queue);
 199		resources--;
 200	}
 201}
 202
 203static int htc_issue_packets(struct htc_target *target,
 204			     struct htc_endpoint *ep,
 205			     struct list_head *pkt_queue)
 206{
 207	int status = 0;
 208	u16 payload_len;
 209	struct sk_buff *skb;
 210	struct htc_frame_hdr *htc_hdr;
 211	struct htc_packet *packet;
 212
 213	ath6kl_dbg(ATH6KL_DBG_HTC,
 214		   "%s: queue: 0x%p, pkts %d\n", __func__,
 215		   pkt_queue, get_queue_depth(pkt_queue));
 216
 217	while (!list_empty(pkt_queue)) {
 218		packet = list_first_entry(pkt_queue, struct htc_packet, list);
 219		list_del(&packet->list);
 220
 221		skb = packet->skb;
 222		if (!skb) {
 223			WARN_ON_ONCE(1);
 224			status = -EINVAL;
 225			break;
 226		}
 227
 228		payload_len = packet->act_len;
 229
 230		/* setup HTC frame header */
 231		htc_hdr = skb_push(skb, sizeof(*htc_hdr));
 
 232		if (!htc_hdr) {
 233			WARN_ON_ONCE(1);
 234			status = -EINVAL;
 235			break;
 236		}
 237
 238		packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
 239
 240		/* Endianess? */
 241		put_unaligned((u16) payload_len, &htc_hdr->payld_len);
 242		htc_hdr->flags = packet->info.tx.flags;
 243		htc_hdr->eid = (u8) packet->endpoint;
 244		htc_hdr->ctrl[0] = 0;
 245		htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
 246
 247		spin_lock_bh(&target->tx_lock);
 248
 249		/* store in look up queue to match completions */
 250		list_add_tail(&packet->list, &ep->pipe.tx_lookup_queue);
 251		ep->ep_st.tx_issued += 1;
 252		spin_unlock_bh(&target->tx_lock);
 253
 254		status = ath6kl_hif_pipe_send(target->dev->ar,
 255					      ep->pipe.pipeid_ul, NULL, skb);
 256
 257		if (status != 0) {
 258			if (status != -ENOMEM) {
 259				/* TODO: if more than 1 endpoint maps to the
 260				 * same PipeID, it is possible to run out of
 261				 * resources in the HIF layer.
 262				 * Don't emit the error
 263				 */
 264				ath6kl_dbg(ATH6KL_DBG_HTC,
 265					   "%s: failed status:%d\n",
 266					   __func__, status);
 267			}
 268			spin_lock_bh(&target->tx_lock);
 269			list_del(&packet->list);
 270
 271			/* reclaim credits */
 272			ep->cred_dist.credits += packet->info.tx.cred_used;
 273			spin_unlock_bh(&target->tx_lock);
 274
 275			/* put it back into the callers queue */
 276			list_add(&packet->list, pkt_queue);
 277			break;
 278		}
 
 279	}
 280
 281	if (status != 0) {
 282		while (!list_empty(pkt_queue)) {
 283			if (status != -ENOMEM) {
 284				ath6kl_dbg(ATH6KL_DBG_HTC,
 285					   "%s: failed pkt:0x%p status:%d\n",
 286					   __func__, packet, status);
 287			}
 288
 289			packet = list_first_entry(pkt_queue,
 290						  struct htc_packet, list);
 291			list_del(&packet->list);
 292			packet->status = status;
 293			send_packet_completion(target, packet);
 294		}
 295	}
 296
 297	return status;
 298}
 299
 300static enum htc_send_queue_result htc_try_send(struct htc_target *target,
 301					       struct htc_endpoint *ep,
 302					       struct list_head *txq)
 303{
 304	struct list_head send_queue;	/* temp queue to hold packets */
 305	struct htc_packet *packet, *tmp_pkt;
 306	struct ath6kl *ar = target->dev->ar;
 307	enum htc_send_full_action action;
 308	int tx_resources, overflow, txqueue_depth, i, good_pkts;
 309	u8 pipeid;
 310
 311	ath6kl_dbg(ATH6KL_DBG_HTC, "%s: (queue:0x%p depth:%d)\n",
 312		   __func__, txq,
 313		   (txq == NULL) ? 0 : get_queue_depth(txq));
 314
 315	/* init the local send queue */
 316	INIT_LIST_HEAD(&send_queue);
 317
 318	/*
 319	 * txq equals to NULL means
 320	 * caller didn't provide a queue, just wants us to
 321	 * check queues and send
 322	 */
 323	if (txq != NULL) {
 324		if (list_empty(txq)) {
 325			/* empty queue */
 326			return HTC_SEND_QUEUE_DROP;
 327		}
 328
 329		spin_lock_bh(&target->tx_lock);
 330		txqueue_depth = get_queue_depth(&ep->txq);
 331		spin_unlock_bh(&target->tx_lock);
 332
 333		if (txqueue_depth >= ep->max_txq_depth) {
 334			/* we've already overflowed */
 335			overflow = get_queue_depth(txq);
 336		} else {
 337			/* get how much we will overflow by */
 338			overflow = txqueue_depth;
 339			overflow += get_queue_depth(txq);
 340			/* get how much we will overflow the TX queue by */
 341			overflow -= ep->max_txq_depth;
 342		}
 343
 344		/* if overflow is negative or zero, we are okay */
 345		if (overflow > 0) {
 346			ath6kl_dbg(ATH6KL_DBG_HTC,
 347				   "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
 348				   __func__, ep->eid, overflow, txqueue_depth,
 349				   ep->max_txq_depth);
 350		}
 351		if ((overflow <= 0) ||
 352		    (ep->ep_cb.tx_full == NULL)) {
 353			/*
 354			 * all packets will fit or caller did not provide send
 355			 * full indication handler -- just move all of them
 356			 * to the local send_queue object
 357			 */
 358			list_splice_tail_init(txq, &send_queue);
 359		} else {
 360			good_pkts = get_queue_depth(txq) - overflow;
 361			if (good_pkts < 0) {
 362				WARN_ON_ONCE(1);
 363				return HTC_SEND_QUEUE_DROP;
 364			}
 365
 366			/* we have overflowed, and a callback is provided */
 367			/* dequeue all non-overflow packets to the sendqueue */
 368			for (i = 0; i < good_pkts; i++) {
 369				/* pop off caller's queue */
 370				packet = list_first_entry(txq,
 371							  struct htc_packet,
 372							  list);
 373				/* move to local queue */
 374				list_move_tail(&packet->list, &send_queue);
 
 375			}
 376
 377			/*
 378			 * the caller's queue has all the packets that won't fit
 379			 * walk through the caller's queue and indicate each to
 380			 * the send full handler
 381			 */
 382			list_for_each_entry_safe(packet, tmp_pkt,
 383						 txq, list) {
 
 384				ath6kl_dbg(ATH6KL_DBG_HTC,
 385					   "%s: Indicate overflowed TX pkts: %p\n",
 386					   __func__, packet);
 387				action = ep->ep_cb.tx_full(ep->target, packet);
 388				if (action == HTC_SEND_FULL_DROP) {
 389					/* callback wants the packet dropped */
 390					ep->ep_st.tx_dropped += 1;
 391
 392					/* leave this one in the caller's queue
 393					 * for cleanup */
 394				} else {
 395					/* callback wants to keep this packet,
 396					 * move from caller's queue to the send
 397					 * queue */
 398					list_move_tail(&packet->list,
 399						       &send_queue);
 
 400				}
 
 401			}
 402
 403			if (list_empty(&send_queue)) {
 404				/* no packets made it in, caller will cleanup */
 405				return HTC_SEND_QUEUE_DROP;
 406			}
 407		}
 408	}
 409
 410	if (!ep->pipe.tx_credit_flow_enabled) {
 411		tx_resources =
 412		    ath6kl_hif_pipe_get_free_queue_number(ar,
 413							  ep->pipe.pipeid_ul);
 414	} else {
 415		tx_resources = 0;
 416	}
 417
 418	spin_lock_bh(&target->tx_lock);
 419	if (!list_empty(&send_queue)) {
 420		/* transfer packets to tail */
 421		list_splice_tail_init(&send_queue, &ep->txq);
 422		if (!list_empty(&send_queue)) {
 423			WARN_ON_ONCE(1);
 424			spin_unlock_bh(&target->tx_lock);
 425			return HTC_SEND_QUEUE_DROP;
 426		}
 427		INIT_LIST_HEAD(&send_queue);
 428	}
 429
 430	/* increment tx processing count on entry */
 431	ep->tx_proc_cnt++;
 432
 433	if (ep->tx_proc_cnt > 1) {
 434		/*
 435		 * Another thread or task is draining the TX queues on this
 436		 * endpoint that thread will reset the tx processing count
 437		 * when the queue is drained.
 438		 */
 439		ep->tx_proc_cnt--;
 440		spin_unlock_bh(&target->tx_lock);
 441		return HTC_SEND_QUEUE_OK;
 442	}
 443
 444	/***** beyond this point only 1 thread may enter ******/
 445
 446	/*
 447	 * Now drain the endpoint TX queue for transmission as long as we have
 448	 * enough transmit resources.
 449	 */
 450	while (true) {
 
 451		if (get_queue_depth(&ep->txq) == 0)
 452			break;
 453
 454		if (ep->pipe.tx_credit_flow_enabled) {
 455			/*
 456			 * Credit based mechanism provides flow control
 457			 * based on target transmit resource availability,
 458			 * we assume that the HIF layer will always have
 459			 * bus resources greater than target transmit
 460			 * resources.
 461			 */
 462			get_htc_packet_credit_based(target, ep, &send_queue);
 463		} else {
 464			/*
 465			 * Get all packets for this endpoint that we can
 466			 * for this pass.
 467			 */
 468			get_htc_packet(target, ep, &send_queue, tx_resources);
 469		}
 470
 471		if (get_queue_depth(&send_queue) == 0) {
 472			/*
 473			 * Didn't get packets due to out of resources or TX
 474			 * queue was drained.
 475			 */
 476			break;
 477		}
 478
 479		spin_unlock_bh(&target->tx_lock);
 480
 481		/* send what we can */
 482		htc_issue_packets(target, ep, &send_queue);
 483
 484		if (!ep->pipe.tx_credit_flow_enabled) {
 485			pipeid = ep->pipe.pipeid_ul;
 486			tx_resources =
 487			    ath6kl_hif_pipe_get_free_queue_number(ar, pipeid);
 488		}
 489
 490		spin_lock_bh(&target->tx_lock);
 491	}
 492
 
 493	/* done with this endpoint, we can clear the count */
 494	ep->tx_proc_cnt = 0;
 495	spin_unlock_bh(&target->tx_lock);
 496
 497	return HTC_SEND_QUEUE_OK;
 498}
 499
 500/* htc control packet manipulation */
 501static void destroy_htc_txctrl_packet(struct htc_packet *packet)
 502{
 503	struct sk_buff *skb;
 504	skb = packet->skb;
 505	dev_kfree_skb(skb);
 
 
 506	kfree(packet);
 507}
 508
 509static struct htc_packet *build_htc_txctrl_packet(void)
 510{
 511	struct htc_packet *packet = NULL;
 512	struct sk_buff *skb;
 513
 514	packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
 515	if (packet == NULL)
 516		return NULL;
 517
 518	skb = __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE, GFP_KERNEL);
 519
 520	if (skb == NULL) {
 521		kfree(packet);
 522		return NULL;
 523	}
 524	packet->skb = skb;
 525
 526	return packet;
 527}
 528
 529static void htc_free_txctrl_packet(struct htc_target *target,
 530				   struct htc_packet *packet)
 531{
 532	destroy_htc_txctrl_packet(packet);
 533}
 534
 535static struct htc_packet *htc_alloc_txctrl_packet(struct htc_target *target)
 536{
 537	return build_htc_txctrl_packet();
 538}
 539
 540static void htc_txctrl_complete(struct htc_target *target,
 541				struct htc_packet *packet)
 542{
 543	htc_free_txctrl_packet(target, packet);
 544}
 545
 546#define MAX_MESSAGE_SIZE 1536
 547
 548static int htc_setup_target_buffer_assignments(struct htc_target *target)
 549{
 550	int status, credits, credit_per_maxmsg, i;
 551	struct htc_pipe_txcredit_alloc *entry;
 552	unsigned int hif_usbaudioclass = 0;
 553
 554	credit_per_maxmsg = MAX_MESSAGE_SIZE / target->tgt_cred_sz;
 555	if (MAX_MESSAGE_SIZE % target->tgt_cred_sz)
 556		credit_per_maxmsg++;
 557
 558	/* TODO, this should be configured by the caller! */
 559
 560	credits = target->tgt_creds;
 561	entry = &target->pipe.txcredit_alloc[0];
 562
 563	status = -ENOMEM;
 564
 565	/* FIXME: hif_usbaudioclass is always zero */
 566	if (hif_usbaudioclass) {
 567		ath6kl_dbg(ATH6KL_DBG_HTC,
 568			   "%s: For USB Audio Class- Total:%d\n",
 569			   __func__, credits);
 570		entry++;
 571		entry++;
 572		/* Setup VO Service To have Max Credits */
 573		entry->service_id = WMI_DATA_VO_SVC;
 574		entry->credit_alloc = (credits - 6);
 575		if (entry->credit_alloc == 0)
 576			entry->credit_alloc++;
 577
 578		credits -= (int) entry->credit_alloc;
 579		if (credits <= 0)
 580			return status;
 581
 582		entry++;
 583		entry->service_id = WMI_CONTROL_SVC;
 584		entry->credit_alloc = credit_per_maxmsg;
 585		credits -= (int) entry->credit_alloc;
 586		if (credits <= 0)
 587			return status;
 588
 589		/* leftovers go to best effort */
 590		entry++;
 591		entry++;
 592		entry->service_id = WMI_DATA_BE_SVC;
 593		entry->credit_alloc = (u8) credits;
 594		status = 0;
 595	} else {
 596		entry++;
 597		entry->service_id = WMI_DATA_VI_SVC;
 598		entry->credit_alloc = credits / 4;
 599		if (entry->credit_alloc == 0)
 600			entry->credit_alloc++;
 601
 602		credits -= (int) entry->credit_alloc;
 603		if (credits <= 0)
 604			return status;
 605
 606		entry++;
 607		entry->service_id = WMI_DATA_VO_SVC;
 608		entry->credit_alloc = credits / 4;
 609		if (entry->credit_alloc == 0)
 610			entry->credit_alloc++;
 611
 612		credits -= (int) entry->credit_alloc;
 613		if (credits <= 0)
 614			return status;
 615
 616		entry++;
 617		entry->service_id = WMI_CONTROL_SVC;
 618		entry->credit_alloc = credit_per_maxmsg;
 619		credits -= (int) entry->credit_alloc;
 620		if (credits <= 0)
 621			return status;
 622
 623		entry++;
 624		entry->service_id = WMI_DATA_BK_SVC;
 625		entry->credit_alloc = credit_per_maxmsg;
 626		credits -= (int) entry->credit_alloc;
 627		if (credits <= 0)
 628			return status;
 629
 630		/* leftovers go to best effort */
 631		entry++;
 632		entry->service_id = WMI_DATA_BE_SVC;
 633		entry->credit_alloc = (u8) credits;
 634		status = 0;
 635	}
 636
 637	if (status == 0) {
 638		for (i = 0; i < ENDPOINT_MAX; i++) {
 639			if (target->pipe.txcredit_alloc[i].service_id != 0) {
 640				ath6kl_dbg(ATH6KL_DBG_HTC,
 641					   "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
 642					   i,
 643					   target->pipe.txcredit_alloc[i].
 644					   service_id,
 645					   target->pipe.txcredit_alloc[i].
 646					   credit_alloc);
 647			}
 648		}
 649	}
 650	return status;
 651}
 652
 653/* process credit reports and call distribution function */
 654static void htc_process_credit_report(struct htc_target *target,
 655				      struct htc_credit_report *rpt,
 656				      int num_entries,
 657				      enum htc_endpoint_id from_ep)
 658{
 659	int total_credits = 0, i;
 660	struct htc_endpoint *ep;
 661
 662	/* lock out TX while we update credits */
 663	spin_lock_bh(&target->tx_lock);
 664
 665	for (i = 0; i < num_entries; i++, rpt++) {
 666		if (rpt->eid >= ENDPOINT_MAX) {
 667			WARN_ON_ONCE(1);
 668			spin_unlock_bh(&target->tx_lock);
 669			return;
 670		}
 671
 672		ep = &target->endpoint[rpt->eid];
 673		ep->cred_dist.credits += rpt->credits;
 674
 675		if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) {
 676			spin_unlock_bh(&target->tx_lock);
 677			htc_try_send(target, ep, NULL);
 678			spin_lock_bh(&target->tx_lock);
 679		}
 680
 681		total_credits += rpt->credits;
 682	}
 683	ath6kl_dbg(ATH6KL_DBG_HTC,
 684		   "Report indicated %d credits to distribute\n",
 685		   total_credits);
 686
 687	spin_unlock_bh(&target->tx_lock);
 688}
 689
 690/* flush endpoint TX queue */
 691static void htc_flush_tx_endpoint(struct htc_target *target,
 692				  struct htc_endpoint *ep, u16 tag)
 693{
 694	struct htc_packet *packet;
 695
 696	spin_lock_bh(&target->tx_lock);
 697	while (get_queue_depth(&ep->txq)) {
 698		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 699		list_del(&packet->list);
 700		packet->status = 0;
 701		send_packet_completion(target, packet);
 702	}
 703	spin_unlock_bh(&target->tx_lock);
 704}
 705
 706/*
 707 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
 708 * since upper layers expects struct htc_packet containers we use the completed
 709 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
 710 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
 711 * HTC.
 712 */
 713static struct htc_packet *htc_lookup_tx_packet(struct htc_target *target,
 714					       struct htc_endpoint *ep,
 715					       struct sk_buff *skb)
 716{
 717	struct htc_packet *packet, *tmp_pkt, *found_packet = NULL;
 718
 719	spin_lock_bh(&target->tx_lock);
 720
 721	/*
 722	 * interate from the front of tx lookup queue
 723	 * this lookup should be fast since lower layers completes in-order and
 724	 * so the completed packet should be at the head of the list generally
 725	 */
 726	list_for_each_entry_safe(packet, tmp_pkt, &ep->pipe.tx_lookup_queue,
 727				 list) {
 728		/* check for removal */
 729		if (skb == packet->skb) {
 730			/* found it */
 731			list_del(&packet->list);
 732			found_packet = packet;
 733			break;
 734		}
 735	}
 736
 737	spin_unlock_bh(&target->tx_lock);
 738
 739	return found_packet;
 740}
 741
 742static int ath6kl_htc_pipe_tx_complete(struct ath6kl *ar, struct sk_buff *skb)
 743{
 744	struct htc_target *target = ar->htc_target;
 745	struct htc_frame_hdr *htc_hdr;
 746	struct htc_endpoint *ep;
 747	struct htc_packet *packet;
 748	u8 ep_id, *netdata;
 
 749
 750	netdata = skb->data;
 
 751
 752	htc_hdr = (struct htc_frame_hdr *) netdata;
 753
 754	ep_id = htc_hdr->eid;
 755	ep = &target->endpoint[ep_id];
 756
 757	packet = htc_lookup_tx_packet(target, ep, skb);
 758	if (packet == NULL) {
 759		/* may have already been flushed and freed */
 760		ath6kl_err("HTC TX lookup failed!\n");
 761	} else {
 762		/* will be giving this buffer back to upper layers */
 763		packet->status = 0;
 764		send_packet_completion(target, packet);
 765	}
 766	skb = NULL;
 767
 768	if (!ep->pipe.tx_credit_flow_enabled) {
 769		/*
 770		 * note: when using TX credit flow, the re-checking of queues
 771		 * happens when credits flow back from the target. in the
 772		 * non-TX credit case, we recheck after the packet completes
 773		 */
 774		htc_try_send(target, ep, NULL);
 775	}
 776
 777	return 0;
 778}
 779
 780static int htc_send_packets_multiple(struct htc_target *target,
 781				     struct list_head *pkt_queue)
 782{
 783	struct htc_endpoint *ep;
 784	struct htc_packet *packet, *tmp_pkt;
 785
 786	if (list_empty(pkt_queue))
 787		return -EINVAL;
 788
 789	/* get first packet to find out which ep the packets will go into */
 790	packet = list_first_entry(pkt_queue, struct htc_packet, list);
 791
 792	if (packet->endpoint >= ENDPOINT_MAX) {
 793		WARN_ON_ONCE(1);
 794		return -EINVAL;
 795	}
 796	ep = &target->endpoint[packet->endpoint];
 797
 798	htc_try_send(target, ep, pkt_queue);
 799
 800	/* do completion on any packets that couldn't get in */
 801	if (!list_empty(pkt_queue)) {
 802		list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
 803			packet->status = -ENOMEM;
 804		}
 805
 806		do_send_completion(ep, pkt_queue);
 807	}
 808
 809	return 0;
 810}
 811
 812/* htc pipe rx path */
 813static struct htc_packet *alloc_htc_packet_container(struct htc_target *target)
 814{
 815	struct htc_packet *packet;
 816	spin_lock_bh(&target->rx_lock);
 817
 818	if (target->pipe.htc_packet_pool == NULL) {
 819		spin_unlock_bh(&target->rx_lock);
 820		return NULL;
 821	}
 822
 823	packet = target->pipe.htc_packet_pool;
 824	target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
 825
 826	spin_unlock_bh(&target->rx_lock);
 827
 828	packet->list.next = NULL;
 829	return packet;
 830}
 831
 832static void free_htc_packet_container(struct htc_target *target,
 833				      struct htc_packet *packet)
 834{
 835	struct list_head *lh;
 836
 837	spin_lock_bh(&target->rx_lock);
 838
 839	if (target->pipe.htc_packet_pool == NULL) {
 840		target->pipe.htc_packet_pool = packet;
 841		packet->list.next = NULL;
 842	} else {
 843		lh = (struct list_head *) target->pipe.htc_packet_pool;
 844		packet->list.next = lh;
 845		target->pipe.htc_packet_pool = packet;
 846	}
 847
 848	spin_unlock_bh(&target->rx_lock);
 849}
 850
 851static int htc_process_trailer(struct htc_target *target, u8 *buffer,
 852			       int len, enum htc_endpoint_id from_ep)
 853{
 854	struct htc_credit_report *report;
 855	struct htc_record_hdr *record;
 856	u8 *record_buf;
 857	int status = 0;
 
 
 
 
 858
 859	while (len > 0) {
 860		if (len < sizeof(struct htc_record_hdr)) {
 861			status = -EINVAL;
 862			break;
 863		}
 864
 865		/* these are byte aligned structs */
 866		record = (struct htc_record_hdr *) buffer;
 867		len -= sizeof(struct htc_record_hdr);
 868		buffer += sizeof(struct htc_record_hdr);
 869
 870		if (record->len > len) {
 871			/* no room left in buffer for record */
 872			ath6kl_dbg(ATH6KL_DBG_HTC,
 873				   "invalid length: %d (id:%d) buffer has: %d bytes left\n",
 874				   record->len, record->rec_id, len);
 875			status = -EINVAL;
 876			break;
 877		}
 878
 879		/* start of record follows the header */
 880		record_buf = buffer;
 881
 882		switch (record->rec_id) {
 883		case HTC_RECORD_CREDITS:
 884			if (record->len < sizeof(struct htc_credit_report)) {
 885				WARN_ON_ONCE(1);
 886				return -EINVAL;
 887			}
 888
 889			report = (struct htc_credit_report *) record_buf;
 890			htc_process_credit_report(target, report,
 891						  record->len / sizeof(*report),
 892						  from_ep);
 893			break;
 894		default:
 895			ath6kl_dbg(ATH6KL_DBG_HTC,
 896				   "unhandled record: id:%d length:%d\n",
 897				   record->rec_id, record->len);
 898			break;
 899		}
 900
 
 
 
 901		/* advance buffer past this record for next time around */
 902		buffer += record->len;
 903		len -= record->len;
 904	}
 905
 906	return status;
 907}
 908
 909static void do_recv_completion(struct htc_endpoint *ep,
 910			       struct list_head *queue_to_indicate)
 911{
 912	struct htc_packet *packet;
 913
 914	if (list_empty(queue_to_indicate)) {
 915		/* nothing to indicate */
 916		return;
 917	}
 918
 919	/* using legacy EpRecv */
 920	while (!list_empty(queue_to_indicate)) {
 921		packet = list_first_entry(queue_to_indicate,
 922					  struct htc_packet, list);
 923		list_del(&packet->list);
 924		ep->ep_cb.rx(ep->target, packet);
 925	}
 926
 927	return;
 928}
 929
 930static void recv_packet_completion(struct htc_target *target,
 931				   struct htc_endpoint *ep,
 932				   struct htc_packet *packet)
 933{
 934	struct list_head container;
 935	INIT_LIST_HEAD(&container);
 936	list_add_tail(&packet->list, &container);
 937
 938	/* do completion */
 939	do_recv_completion(ep, &container);
 940}
 941
 942static int ath6kl_htc_pipe_rx_complete(struct ath6kl *ar, struct sk_buff *skb,
 943				       u8 pipeid)
 944{
 945	struct htc_target *target = ar->htc_target;
 946	u8 *netdata, *trailer, hdr_info;
 947	struct htc_frame_hdr *htc_hdr;
 948	u32 netlen, trailerlen = 0;
 949	struct htc_packet *packet;
 950	struct htc_endpoint *ep;
 951	u16 payload_len;
 952	int status = 0;
 953
 954	/*
 955	 * ar->htc_target can be NULL due to a race condition that can occur
 956	 * during driver initialization(we do 'ath6kl_hif_power_on' before
 957	 * initializing 'ar->htc_target' via 'ath6kl_htc_create').
 958	 * 'ath6kl_hif_power_on' assigns 'ath6kl_recv_complete' as
 959	 * usb_complete_t/callback function for 'usb_fill_bulk_urb'.
 960	 * Thus the possibility of ar->htc_target being NULL
 961	 * via ath6kl_recv_complete -> ath6kl_usb_io_comp_work.
 962	 */
 963	if (!target) {
 964		ath6kl_dbg(ATH6KL_DBG_HTC, "Target not yet initialized\n");
 965		status = -EINVAL;
 966		goto free_skb;
 967	}
 968
 969
 970	netdata = skb->data;
 971	netlen = skb->len;
 972
 973	htc_hdr = (struct htc_frame_hdr *) netdata;
 974
 
 
 975	if (htc_hdr->eid >= ENDPOINT_MAX) {
 976		ath6kl_dbg(ATH6KL_DBG_HTC,
 977			   "HTC Rx: invalid EndpointID=%d\n",
 978			   htc_hdr->eid);
 979		status = -EINVAL;
 980		goto free_skb;
 981	}
 982	ep = &target->endpoint[htc_hdr->eid];
 983
 984	payload_len = le16_to_cpu(get_unaligned(&htc_hdr->payld_len));
 985
 986	if (netlen < (payload_len + HTC_HDR_LENGTH)) {
 987		ath6kl_dbg(ATH6KL_DBG_HTC,
 988			   "HTC Rx: insufficient length, got:%d expected =%zu\n",
 989			   netlen, payload_len + HTC_HDR_LENGTH);
 990		status = -EINVAL;
 991		goto free_skb;
 992	}
 993
 994	/* get flags to check for trailer */
 995	hdr_info = htc_hdr->flags;
 996	if (hdr_info & HTC_FLG_RX_TRAILER) {
 997		/* extract the trailer length */
 998		hdr_info = htc_hdr->ctrl[0];
 999		if ((hdr_info < sizeof(struct htc_record_hdr)) ||
1000		    (hdr_info > payload_len)) {
1001			ath6kl_dbg(ATH6KL_DBG_HTC,
1002				   "invalid header: payloadlen should be %d, CB[0]: %d\n",
1003				   payload_len, hdr_info);
1004			status = -EINVAL;
1005			goto free_skb;
1006		}
1007
1008		trailerlen = hdr_info;
1009		/* process trailer after hdr/apps payload */
1010		trailer = (u8 *) htc_hdr + HTC_HDR_LENGTH +
1011			payload_len - hdr_info;
1012		status = htc_process_trailer(target, trailer, hdr_info,
1013					     htc_hdr->eid);
1014		if (status != 0)
1015			goto free_skb;
1016	}
1017
1018	if (((int) payload_len - (int) trailerlen) <= 0) {
1019		/* zero length packet with trailer, just drop these */
1020		goto free_skb;
1021	}
1022
1023	if (htc_hdr->eid == ENDPOINT_0) {
1024		/* handle HTC control message */
1025		if (target->htc_flags & HTC_OP_STATE_SETUP_COMPLETE) {
1026			/*
1027			 * fatal: target should not send unsolicited
1028			 * messageson the endpoint 0
1029			 */
1030			ath6kl_dbg(ATH6KL_DBG_HTC,
1031				   "HTC ignores Rx Ctrl after setup complete\n");
1032			status = -EINVAL;
1033			goto free_skb;
1034		}
1035
1036		/* remove HTC header */
1037		skb_pull(skb, HTC_HDR_LENGTH);
1038
1039		netdata = skb->data;
1040		netlen = skb->len;
1041
1042		spin_lock_bh(&target->rx_lock);
1043
1044		target->pipe.ctrl_response_valid = true;
1045		target->pipe.ctrl_response_len = min_t(int, netlen,
1046						       HTC_MAX_CTRL_MSG_LEN);
1047		memcpy(target->pipe.ctrl_response_buf, netdata,
1048		       target->pipe.ctrl_response_len);
1049
1050		spin_unlock_bh(&target->rx_lock);
1051
1052		dev_kfree_skb(skb);
1053		skb = NULL;
1054
1055		goto free_skb;
1056	}
1057
1058	/*
1059	 * TODO: the message based HIF architecture allocates net bufs
1060	 * for recv packets since it bridges that HIF to upper layers,
1061	 * which expects HTC packets, we form the packets here
1062	 */
1063	packet = alloc_htc_packet_container(target);
1064	if (packet == NULL) {
1065		status = -ENOMEM;
1066		goto free_skb;
1067	}
1068
1069	packet->status = 0;
1070	packet->endpoint = htc_hdr->eid;
1071	packet->pkt_cntxt = skb;
1072
1073	/* TODO: for backwards compatibility */
1074	packet->buf = skb_push(skb, 0) + HTC_HDR_LENGTH;
1075	packet->act_len = netlen - HTC_HDR_LENGTH - trailerlen;
1076
1077	/*
1078	 * TODO: this is a hack because the driver layer will set the
1079	 * actual len of the skb again which will just double the len
1080	 */
1081	skb_trim(skb, 0);
1082
1083	recv_packet_completion(target, ep, packet);
1084
1085	/* recover the packet container */
1086	free_htc_packet_container(target, packet);
1087	skb = NULL;
1088
1089free_skb:
1090	dev_kfree_skb(skb);
 
1091
1092	return status;
 
1093}
1094
1095static void htc_flush_rx_queue(struct htc_target *target,
1096			       struct htc_endpoint *ep)
1097{
1098	struct list_head container;
1099	struct htc_packet *packet;
1100
1101	spin_lock_bh(&target->rx_lock);
1102
1103	while (1) {
1104		if (list_empty(&ep->rx_bufq))
1105			break;
1106
1107		packet = list_first_entry(&ep->rx_bufq,
1108					  struct htc_packet, list);
1109		list_del(&packet->list);
1110
1111		spin_unlock_bh(&target->rx_lock);
1112		packet->status = -ECANCELED;
1113		packet->act_len = 0;
1114
1115		ath6kl_dbg(ATH6KL_DBG_HTC,
1116			   "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1117			   packet, packet->buf_len,
1118			   packet->endpoint);
1119
1120		INIT_LIST_HEAD(&container);
1121		list_add_tail(&packet->list, &container);
1122
1123		/* give the packet back */
1124		do_recv_completion(ep, &container);
1125		spin_lock_bh(&target->rx_lock);
1126	}
1127
1128	spin_unlock_bh(&target->rx_lock);
1129}
1130
1131/* polling routine to wait for a control packet to be received */
1132static int htc_wait_recv_ctrl_message(struct htc_target *target)
1133{
1134	int count = HTC_TARGET_RESPONSE_POLL_COUNT;
1135
1136	while (count > 0) {
1137		spin_lock_bh(&target->rx_lock);
1138
1139		if (target->pipe.ctrl_response_valid) {
1140			target->pipe.ctrl_response_valid = false;
1141			spin_unlock_bh(&target->rx_lock);
1142			break;
1143		}
1144
1145		spin_unlock_bh(&target->rx_lock);
1146
1147		count--;
1148
1149		msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT);
1150	}
1151
1152	if (count <= 0) {
1153		ath6kl_warn("htc pipe control receive timeout!\n");
1154		return -ETIMEDOUT;
1155	}
1156
1157	return 0;
1158}
1159
1160static void htc_rxctrl_complete(struct htc_target *context,
1161				struct htc_packet *packet)
1162{
1163	struct sk_buff *skb = packet->skb;
1164
1165	if (packet->endpoint == ENDPOINT_0 &&
1166	    packet->status == -ECANCELED &&
1167	    skb != NULL)
1168		dev_kfree_skb(skb);
1169}
1170
1171/* htc pipe initialization */
1172static void reset_endpoint_states(struct htc_target *target)
1173{
1174	struct htc_endpoint *ep;
1175	int i;
1176
1177	for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1178		ep = &target->endpoint[i];
1179		ep->svc_id = 0;
1180		ep->len_max = 0;
1181		ep->max_txq_depth = 0;
1182		ep->eid = i;
1183		INIT_LIST_HEAD(&ep->txq);
1184		INIT_LIST_HEAD(&ep->pipe.tx_lookup_queue);
1185		INIT_LIST_HEAD(&ep->rx_bufq);
1186		ep->target = target;
1187		ep->pipe.tx_credit_flow_enabled = true;
1188	}
1189}
1190
1191/* start HTC, this is called after all services are connected */
1192static int htc_config_target_hif_pipe(struct htc_target *target)
1193{
1194	return 0;
1195}
1196
1197/* htc service functions */
1198static u8 htc_get_credit_alloc(struct htc_target *target, u16 service_id)
1199{
1200	u8 allocation = 0;
1201	int i;
1202
1203	for (i = 0; i < ENDPOINT_MAX; i++) {
1204		if (target->pipe.txcredit_alloc[i].service_id == service_id)
1205			allocation =
1206				target->pipe.txcredit_alloc[i].credit_alloc;
1207	}
1208
1209	if (allocation == 0) {
1210		ath6kl_dbg(ATH6KL_DBG_HTC,
1211			   "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1212			   service_id);
1213	}
1214
1215	return allocation;
1216}
1217
1218static int ath6kl_htc_pipe_conn_service(struct htc_target *target,
1219		     struct htc_service_connect_req *conn_req,
1220		     struct htc_service_connect_resp *conn_resp)
1221{
1222	struct ath6kl *ar = target->dev->ar;
1223	struct htc_packet *packet = NULL;
1224	struct htc_conn_service_resp *resp_msg;
1225	struct htc_conn_service_msg *conn_msg;
1226	enum htc_endpoint_id assigned_epid = ENDPOINT_MAX;
1227	bool disable_credit_flowctrl = false;
1228	unsigned int max_msg_size = 0;
1229	struct htc_endpoint *ep;
1230	int length, status = 0;
1231	struct sk_buff *skb;
1232	u8 tx_alloc;
1233	u16 flags;
1234
1235	if (conn_req->svc_id == 0) {
1236		WARN_ON_ONCE(1);
1237		status = -EINVAL;
1238		goto free_packet;
1239	}
1240
1241	if (conn_req->svc_id == HTC_CTRL_RSVD_SVC) {
1242		/* special case for pseudo control service */
1243		assigned_epid = ENDPOINT_0;
1244		max_msg_size = HTC_MAX_CTRL_MSG_LEN;
1245		tx_alloc = 0;
1246
1247	} else {
 
1248		tx_alloc = htc_get_credit_alloc(target, conn_req->svc_id);
1249		if (tx_alloc == 0) {
1250			status = -ENOMEM;
1251			goto free_packet;
1252		}
1253
1254		/* allocate a packet to send to the target */
1255		packet = htc_alloc_txctrl_packet(target);
1256
1257		if (packet == NULL) {
1258			WARN_ON_ONCE(1);
1259			status = -ENOMEM;
1260			goto free_packet;
1261		}
1262
1263		skb = packet->skb;
1264		length = sizeof(struct htc_conn_service_msg);
1265
1266		/* assemble connect service message */
1267		conn_msg = skb_put(skb, length);
 
1268		if (conn_msg == NULL) {
1269			WARN_ON_ONCE(1);
1270			status = -EINVAL;
1271			goto free_packet;
1272		}
1273
1274		memset(conn_msg, 0,
1275		       sizeof(struct htc_conn_service_msg));
1276		conn_msg->msg_id = cpu_to_le16(HTC_MSG_CONN_SVC_ID);
1277		conn_msg->svc_id = cpu_to_le16(conn_req->svc_id);
1278		conn_msg->conn_flags = cpu_to_le16(conn_req->conn_flags &
1279					~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK);
1280
1281		/* tell target desired recv alloc for this ep */
1282		flags = tx_alloc << HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT;
1283		conn_msg->conn_flags |= cpu_to_le16(flags);
1284
1285		if (conn_req->conn_flags &
1286		    HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL) {
1287			disable_credit_flowctrl = true;
1288		}
1289
1290		set_htc_pkt_info(packet, NULL, (u8 *) conn_msg,
1291				 length,
1292				 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1293
1294		status = ath6kl_htc_pipe_tx(target, packet);
1295
1296		/* we don't own it anymore */
1297		packet = NULL;
1298		if (status != 0)
1299			goto free_packet;
1300
1301		/* wait for response */
1302		status = htc_wait_recv_ctrl_message(target);
1303		if (status != 0)
1304			goto free_packet;
1305
1306		/* we controlled the buffer creation so it has to be
1307		 * properly aligned
1308		 */
1309		resp_msg = (struct htc_conn_service_resp *)
1310		    target->pipe.ctrl_response_buf;
1311
1312		if (resp_msg->msg_id != cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID) ||
1313		    (target->pipe.ctrl_response_len < sizeof(*resp_msg))) {
1314			/* this message is not valid */
1315			WARN_ON_ONCE(1);
1316			status = -EINVAL;
1317			goto free_packet;
1318		}
1319
1320		ath6kl_dbg(ATH6KL_DBG_TRC,
1321			   "%s: service 0x%X conn resp: status: %d ep: %d\n",
1322			   __func__, resp_msg->svc_id, resp_msg->status,
1323			   resp_msg->eid);
1324
1325		conn_resp->resp_code = resp_msg->status;
1326		/* check response status */
1327		if (resp_msg->status != HTC_SERVICE_SUCCESS) {
1328			ath6kl_dbg(ATH6KL_DBG_HTC,
1329				   "Target failed service 0x%X connect request (status:%d)\n",
1330				   resp_msg->svc_id, resp_msg->status);
1331			status = -EINVAL;
1332			goto free_packet;
1333		}
1334
1335		assigned_epid = (enum htc_endpoint_id) resp_msg->eid;
1336		max_msg_size = le16_to_cpu(resp_msg->max_msg_sz);
1337	}
1338
1339	/* the rest are parameter checks so set the error status */
1340	status = -EINVAL;
1341
1342	if (assigned_epid >= ENDPOINT_MAX) {
1343		WARN_ON_ONCE(1);
1344		goto free_packet;
1345	}
1346
1347	if (max_msg_size == 0) {
1348		WARN_ON_ONCE(1);
1349		goto free_packet;
1350	}
1351
1352	ep = &target->endpoint[assigned_epid];
1353	ep->eid = assigned_epid;
1354	if (ep->svc_id != 0) {
1355		/* endpoint already in use! */
1356		WARN_ON_ONCE(1);
1357		goto free_packet;
1358	}
1359
1360	/* return assigned endpoint to caller */
1361	conn_resp->endpoint = assigned_epid;
1362	conn_resp->len_max = max_msg_size;
1363
1364	/* setup the endpoint */
1365	ep->svc_id = conn_req->svc_id; /* this marks ep in use */
1366	ep->max_txq_depth = conn_req->max_txq_depth;
1367	ep->len_max = max_msg_size;
1368	ep->cred_dist.credits = tx_alloc;
1369	ep->cred_dist.cred_sz = target->tgt_cred_sz;
1370	ep->cred_dist.cred_per_msg = max_msg_size / target->tgt_cred_sz;
1371	if (max_msg_size % target->tgt_cred_sz)
1372		ep->cred_dist.cred_per_msg++;
1373
1374	/* copy all the callbacks */
1375	ep->ep_cb = conn_req->ep_cb;
1376
1377	/* initialize tx_drop_packet_threshold */
1378	ep->tx_drop_packet_threshold = MAX_HI_COOKIE_NUM;
1379
1380	status = ath6kl_hif_pipe_map_service(ar, ep->svc_id,
1381					     &ep->pipe.pipeid_ul,
1382					     &ep->pipe.pipeid_dl);
1383	if (status != 0)
1384		goto free_packet;
1385
1386	ath6kl_dbg(ATH6KL_DBG_HTC,
1387		   "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1388		   ep->svc_id, ep->pipe.pipeid_ul,
1389		   ep->pipe.pipeid_dl, ep->eid);
1390
1391	if (disable_credit_flowctrl && ep->pipe.tx_credit_flow_enabled) {
1392		ep->pipe.tx_credit_flow_enabled = false;
1393		ath6kl_dbg(ATH6KL_DBG_HTC,
1394			   "SVC: 0x%4.4X ep:%d TX flow control off\n",
1395			   ep->svc_id, assigned_epid);
1396	}
1397
1398free_packet:
1399	if (packet != NULL)
1400		htc_free_txctrl_packet(target, packet);
1401	return status;
1402}
1403
1404/* htc export functions */
1405static void *ath6kl_htc_pipe_create(struct ath6kl *ar)
1406{
1407	int status = 0;
1408	struct htc_endpoint *ep = NULL;
1409	struct htc_target *target = NULL;
1410	struct htc_packet *packet;
1411	int i;
1412
1413	target = kzalloc(sizeof(struct htc_target), GFP_KERNEL);
1414	if (target == NULL) {
1415		ath6kl_err("htc create unable to allocate memory\n");
1416		status = -ENOMEM;
1417		goto fail_htc_create;
1418	}
1419
1420	spin_lock_init(&target->htc_lock);
1421	spin_lock_init(&target->rx_lock);
1422	spin_lock_init(&target->tx_lock);
1423
1424	reset_endpoint_states(target);
1425
1426	for (i = 0; i < HTC_PACKET_CONTAINER_ALLOCATION; i++) {
1427		packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
1428
1429		if (packet != NULL)
1430			free_htc_packet_container(target, packet);
1431	}
1432
1433	target->dev = kzalloc(sizeof(*target->dev), GFP_KERNEL);
1434	if (!target->dev) {
1435		ath6kl_err("unable to allocate memory\n");
1436		status = -ENOMEM;
1437		goto fail_htc_create;
1438	}
1439	target->dev->ar = ar;
1440	target->dev->htc_cnxt = target;
1441
1442	/* Get HIF default pipe for HTC message exchange */
1443	ep = &target->endpoint[ENDPOINT_0];
1444
1445	ath6kl_hif_pipe_get_default(ar, &ep->pipe.pipeid_ul,
1446				    &ep->pipe.pipeid_dl);
1447
1448	return target;
1449
1450fail_htc_create:
1451	if (status != 0) {
1452		if (target != NULL)
1453			ath6kl_htc_pipe_cleanup(target);
1454
1455		target = NULL;
1456	}
1457	return target;
1458}
1459
1460/* cleanup the HTC instance */
1461static void ath6kl_htc_pipe_cleanup(struct htc_target *target)
1462{
1463	struct htc_packet *packet;
1464
1465	while (true) {
1466		packet = alloc_htc_packet_container(target);
1467		if (packet == NULL)
1468			break;
1469		kfree(packet);
1470	}
1471
1472	kfree(target->dev);
1473
1474	/* kfree our instance */
1475	kfree(target);
1476}
1477
1478static int ath6kl_htc_pipe_start(struct htc_target *target)
1479{
1480	struct sk_buff *skb;
1481	struct htc_setup_comp_ext_msg *setup;
1482	struct htc_packet *packet;
1483
1484	htc_config_target_hif_pipe(target);
1485
1486	/* allocate a buffer to send */
1487	packet = htc_alloc_txctrl_packet(target);
1488	if (packet == NULL) {
1489		WARN_ON_ONCE(1);
1490		return -ENOMEM;
1491	}
1492
1493	skb = packet->skb;
1494
1495	/* assemble setup complete message */
1496	setup = skb_put(skb, sizeof(*setup));
 
1497	memset(setup, 0, sizeof(struct htc_setup_comp_ext_msg));
1498	setup->msg_id = cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID);
1499
1500	ath6kl_dbg(ATH6KL_DBG_HTC, "HTC using TX credit flow control\n");
1501
1502	set_htc_pkt_info(packet, NULL, (u8 *) setup,
1503			 sizeof(struct htc_setup_comp_ext_msg),
1504			 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1505
1506	target->htc_flags |= HTC_OP_STATE_SETUP_COMPLETE;
1507
1508	return ath6kl_htc_pipe_tx(target, packet);
1509}
1510
1511static void ath6kl_htc_pipe_stop(struct htc_target *target)
1512{
1513	int i;
1514	struct htc_endpoint *ep;
1515
1516	/* cleanup endpoints */
1517	for (i = 0; i < ENDPOINT_MAX; i++) {
1518		ep = &target->endpoint[i];
1519		htc_flush_rx_queue(target, ep);
1520		htc_flush_tx_endpoint(target, ep, HTC_TX_PACKET_TAG_ALL);
1521	}
1522
1523	reset_endpoint_states(target);
1524	target->htc_flags &= ~HTC_OP_STATE_SETUP_COMPLETE;
1525}
1526
1527static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target *target,
1528					 enum htc_endpoint_id endpoint)
1529{
1530	int num;
1531
1532	spin_lock_bh(&target->rx_lock);
1533	num = get_queue_depth(&(target->endpoint[endpoint].rx_bufq));
1534	spin_unlock_bh(&target->rx_lock);
1535
1536	return num;
1537}
1538
1539static int ath6kl_htc_pipe_tx(struct htc_target *target,
1540			      struct htc_packet *packet)
1541{
1542	struct list_head queue;
1543
1544	ath6kl_dbg(ATH6KL_DBG_HTC,
1545		   "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1546		   __func__, packet->endpoint, packet->buf,
1547		   packet->act_len);
1548
1549	INIT_LIST_HEAD(&queue);
1550	list_add_tail(&packet->list, &queue);
1551
1552	return htc_send_packets_multiple(target, &queue);
1553}
1554
1555static int ath6kl_htc_pipe_wait_target(struct htc_target *target)
1556{
1557	struct htc_ready_ext_msg *ready_msg;
1558	struct htc_service_connect_req connect;
1559	struct htc_service_connect_resp resp;
1560	int status = 0;
1561
1562	status = htc_wait_recv_ctrl_message(target);
1563
1564	if (status != 0)
1565		return status;
1566
1567	if (target->pipe.ctrl_response_len < sizeof(*ready_msg)) {
1568		ath6kl_warn("invalid htc pipe ready msg len: %d\n",
1569			    target->pipe.ctrl_response_len);
1570		return -ECOMM;
1571	}
1572
1573	ready_msg = (struct htc_ready_ext_msg *) target->pipe.ctrl_response_buf;
1574
1575	if (ready_msg->ver2_0_info.msg_id != cpu_to_le16(HTC_MSG_READY_ID)) {
1576		ath6kl_warn("invalid htc pipe ready msg: 0x%x\n",
1577			    ready_msg->ver2_0_info.msg_id);
1578		return -ECOMM;
1579	}
1580
1581	ath6kl_dbg(ATH6KL_DBG_HTC,
1582		   "Target Ready! : transmit resources : %d size:%d\n",
1583		   ready_msg->ver2_0_info.cred_cnt,
1584		   ready_msg->ver2_0_info.cred_sz);
1585
1586	target->tgt_creds = le16_to_cpu(ready_msg->ver2_0_info.cred_cnt);
1587	target->tgt_cred_sz = le16_to_cpu(ready_msg->ver2_0_info.cred_sz);
1588
1589	if ((target->tgt_creds == 0) || (target->tgt_cred_sz == 0))
1590		return -ECOMM;
1591
1592	htc_setup_target_buffer_assignments(target);
1593
1594	/* setup our pseudo HTC control endpoint connection */
1595	memset(&connect, 0, sizeof(connect));
1596	memset(&resp, 0, sizeof(resp));
1597	connect.ep_cb.tx_complete = htc_txctrl_complete;
1598	connect.ep_cb.rx = htc_rxctrl_complete;
1599	connect.max_txq_depth = NUM_CONTROL_TX_BUFFERS;
1600	connect.svc_id = HTC_CTRL_RSVD_SVC;
1601
1602	/* connect fake service */
1603	status = ath6kl_htc_pipe_conn_service(target, &connect, &resp);
1604
1605	return status;
1606}
1607
1608static void ath6kl_htc_pipe_flush_txep(struct htc_target *target,
1609				       enum htc_endpoint_id endpoint, u16 tag)
1610{
1611	struct htc_endpoint *ep = &target->endpoint[endpoint];
1612
1613	if (ep->svc_id == 0) {
1614		WARN_ON_ONCE(1);
1615		/* not in use.. */
1616		return;
1617	}
1618
1619	htc_flush_tx_endpoint(target, ep, tag);
1620}
1621
1622static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target *target,
1623					      struct list_head *pkt_queue)
1624{
1625	struct htc_packet *packet, *tmp_pkt, *first;
1626	struct htc_endpoint *ep;
1627	int status = 0;
1628
1629	if (list_empty(pkt_queue))
1630		return -EINVAL;
1631
1632	first = list_first_entry(pkt_queue, struct htc_packet, list);
1633
1634	if (first->endpoint >= ENDPOINT_MAX) {
1635		WARN_ON_ONCE(1);
1636		return -EINVAL;
1637	}
1638
1639	ath6kl_dbg(ATH6KL_DBG_HTC, "%s: epid: %d, cnt:%d, len: %d\n",
1640		   __func__, first->endpoint, get_queue_depth(pkt_queue),
1641		   first->buf_len);
1642
1643	ep = &target->endpoint[first->endpoint];
1644
1645	spin_lock_bh(&target->rx_lock);
1646
1647	/* store receive packets */
1648	list_splice_tail_init(pkt_queue, &ep->rx_bufq);
1649
1650	spin_unlock_bh(&target->rx_lock);
1651
1652	if (status != 0) {
1653		/* walk through queue and mark each one canceled */
1654		list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
1655			packet->status = -ECANCELED;
1656		}
1657
1658		do_recv_completion(ep, pkt_queue);
1659	}
1660
1661	return status;
1662}
1663
1664static void ath6kl_htc_pipe_activity_changed(struct htc_target *target,
1665					     enum htc_endpoint_id ep,
1666					     bool active)
1667{
1668	/* TODO */
1669}
1670
1671static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target *target)
1672{
1673	struct htc_endpoint *endpoint;
1674	struct htc_packet *packet, *tmp_pkt;
1675	int i;
1676
1677	for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1678		endpoint = &target->endpoint[i];
1679
1680		spin_lock_bh(&target->rx_lock);
1681
1682		list_for_each_entry_safe(packet, tmp_pkt,
1683					 &endpoint->rx_bufq, list) {
1684			list_del(&packet->list);
1685			spin_unlock_bh(&target->rx_lock);
1686			ath6kl_dbg(ATH6KL_DBG_HTC,
1687				   "htc rx flush pkt 0x%p len %d ep %d\n",
1688				   packet, packet->buf_len,
1689				   packet->endpoint);
1690			dev_kfree_skb(packet->pkt_cntxt);
1691			spin_lock_bh(&target->rx_lock);
1692		}
1693
1694		spin_unlock_bh(&target->rx_lock);
1695	}
1696}
1697
1698static int ath6kl_htc_pipe_credit_setup(struct htc_target *target,
1699					struct ath6kl_htc_credit_info *info)
1700{
1701	return 0;
1702}
1703
1704static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops = {
1705	.create = ath6kl_htc_pipe_create,
1706	.wait_target = ath6kl_htc_pipe_wait_target,
1707	.start = ath6kl_htc_pipe_start,
1708	.conn_service = ath6kl_htc_pipe_conn_service,
1709	.tx = ath6kl_htc_pipe_tx,
1710	.stop = ath6kl_htc_pipe_stop,
1711	.cleanup = ath6kl_htc_pipe_cleanup,
1712	.flush_txep = ath6kl_htc_pipe_flush_txep,
1713	.flush_rx_buf = ath6kl_htc_pipe_flush_rx_buf,
1714	.activity_changed = ath6kl_htc_pipe_activity_changed,
1715	.get_rxbuf_num = ath6kl_htc_pipe_get_rxbuf_num,
1716	.add_rxbuf_multiple = ath6kl_htc_pipe_add_rxbuf_multiple,
1717	.credit_setup = ath6kl_htc_pipe_credit_setup,
1718	.tx_complete = ath6kl_htc_pipe_tx_complete,
1719	.rx_complete = ath6kl_htc_pipe_rx_complete,
1720};
1721
1722void ath6kl_htc_pipe_attach(struct ath6kl *ar)
1723{
1724	ar->htc_ops = &ath6kl_htc_pipe_ops;
1725}
v3.5.6
   1/*
   2 * Copyright (c) 2007-2011 Atheros Communications Inc.
   3 *
   4 * Permission to use, copy, modify, and/or distribute this software for any
   5 * purpose with or without fee is hereby granted, provided that the above
   6 * copyright notice and this permission notice appear in all copies.
   7 *
   8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
   9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
  10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
  11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
  12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
  13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
  14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
  15 */
  16
  17#include "core.h"
  18#include "debug.h"
  19#include "hif-ops.h"
  20
  21#define HTC_PACKET_CONTAINER_ALLOCATION 32
  22#define HTC_CONTROL_BUFFER_SIZE (HTC_MAX_CTRL_MSG_LEN + HTC_HDR_LENGTH)
  23
  24static int ath6kl_htc_pipe_tx(struct htc_target *handle,
  25			      struct htc_packet *packet);
  26static void ath6kl_htc_pipe_cleanup(struct htc_target *handle);
  27
  28/* htc pipe tx path */
  29static inline void restore_tx_packet(struct htc_packet *packet)
  30{
  31	if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
  32		skb_pull(packet->skb, sizeof(struct htc_frame_hdr));
  33		packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
  34	}
  35}
  36
  37static void do_send_completion(struct htc_endpoint *ep,
  38			       struct list_head *queue_to_indicate)
  39{
  40	struct htc_packet *packet;
  41
  42	if (list_empty(queue_to_indicate)) {
  43		/* nothing to indicate */
  44		return;
  45	}
  46
  47	if (ep->ep_cb.tx_comp_multi != NULL) {
  48		ath6kl_dbg(ATH6KL_DBG_HTC,
  49			   "%s: calling ep %d, send complete multiple callback (%d pkts)\n",
  50			   __func__, ep->eid,
  51			   get_queue_depth(queue_to_indicate));
  52		/*
  53		 * a multiple send complete handler is being used,
  54		 * pass the queue to the handler
  55		 */
  56		ep->ep_cb.tx_comp_multi(ep->target, queue_to_indicate);
  57		/*
  58		 * all packets are now owned by the callback,
  59		 * reset queue to be safe
  60		 */
  61		INIT_LIST_HEAD(queue_to_indicate);
  62	} else {
  63		/* using legacy EpTxComplete */
  64		do {
  65			packet = list_first_entry(queue_to_indicate,
  66						  struct htc_packet, list);
  67
  68			list_del(&packet->list);
  69			ath6kl_dbg(ATH6KL_DBG_HTC,
  70				   "%s: calling ep %d send complete callback on packet 0x%p\n",
  71				   __func__, ep->eid, packet);
  72			ep->ep_cb.tx_complete(ep->target, packet);
  73		} while (!list_empty(queue_to_indicate));
  74	}
  75}
  76
  77static void send_packet_completion(struct htc_target *target,
  78				   struct htc_packet *packet)
  79{
  80	struct htc_endpoint *ep = &target->endpoint[packet->endpoint];
  81	struct list_head container;
  82
  83	restore_tx_packet(packet);
  84	INIT_LIST_HEAD(&container);
  85	list_add_tail(&packet->list, &container);
  86
  87	/* do completion */
  88	do_send_completion(ep, &container);
  89}
  90
  91static void get_htc_packet_credit_based(struct htc_target *target,
  92					struct htc_endpoint *ep,
  93					struct list_head *queue)
  94{
  95	int credits_required;
  96	int remainder;
  97	u8 send_flags;
  98	struct htc_packet *packet;
  99	unsigned int transfer_len;
 100
 101	/* NOTE : the TX lock is held when this function is called */
 102
 103	/* loop until we can grab as many packets out of the queue as we can */
 104	while (true) {
 105		send_flags = 0;
 106		if (list_empty(&ep->txq))
 107			break;
 108
 109		/* get packet at head, but don't remove it */
 110		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 111
 112		ath6kl_dbg(ATH6KL_DBG_HTC,
 113			   "%s: got head packet:0x%p , queue depth: %d\n",
 114			   __func__, packet, get_queue_depth(&ep->txq));
 115
 116		transfer_len = packet->act_len + HTC_HDR_LENGTH;
 117
 118		if (transfer_len <= target->tgt_cred_sz) {
 119			credits_required = 1;
 120		} else {
 121			/* figure out how many credits this message requires */
 122			credits_required = transfer_len / target->tgt_cred_sz;
 123			remainder = transfer_len % target->tgt_cred_sz;
 124
 125			if (remainder)
 126				credits_required++;
 127		}
 128
 129		ath6kl_dbg(ATH6KL_DBG_HTC, "%s: creds required:%d got:%d\n",
 130			   __func__, credits_required, ep->cred_dist.credits);
 131
 132		if (ep->eid == ENDPOINT_0) {
 133			/*
 134			 * endpoint 0 is special, it always has a credit and
 135			 * does not require credit based flow control
 136			 */
 137			credits_required = 0;
 138
 139		} else {
 140
 141			if (ep->cred_dist.credits < credits_required)
 142				break;
 143
 144			ep->cred_dist.credits -= credits_required;
 145			ep->ep_st.cred_cosumd += credits_required;
 146
 147			/* check if we need credits back from the target */
 148			if (ep->cred_dist.credits <
 149					ep->cred_dist.cred_per_msg) {
 150				/* tell the target we need credits ASAP! */
 151				send_flags |= HTC_FLAGS_NEED_CREDIT_UPDATE;
 152				ep->ep_st.cred_low_indicate += 1;
 153				ath6kl_dbg(ATH6KL_DBG_HTC,
 154					   "%s: host needs credits\n",
 155					   __func__);
 156			}
 157		}
 158
 159		/* now we can fully dequeue */
 160		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 161
 162		list_del(&packet->list);
 163		/* save the number of credits this packet consumed */
 164		packet->info.tx.cred_used = credits_required;
 165		/* save send flags */
 166		packet->info.tx.flags = send_flags;
 167		packet->info.tx.seqno = ep->seqno;
 168		ep->seqno++;
 169		/* queue this packet into the caller's queue */
 170		list_add_tail(&packet->list, queue);
 171	}
 172
 173}
 174
 175static void get_htc_packet(struct htc_target *target,
 176			   struct htc_endpoint *ep,
 177			   struct list_head *queue, int resources)
 178{
 179	struct htc_packet *packet;
 180
 181	/* NOTE : the TX lock is held when this function is called */
 182
 183	/* loop until we can grab as many packets out of the queue as we can */
 184	while (resources) {
 185		if (list_empty(&ep->txq))
 186			break;
 187
 188		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 189		list_del(&packet->list);
 190
 191		ath6kl_dbg(ATH6KL_DBG_HTC,
 192			   "%s: got packet:0x%p , new queue depth: %d\n",
 193			   __func__, packet, get_queue_depth(&ep->txq));
 194		packet->info.tx.seqno = ep->seqno;
 195		packet->info.tx.flags = 0;
 196		packet->info.tx.cred_used = 0;
 197		ep->seqno++;
 198
 199		/* queue this packet into the caller's queue */
 200		list_add_tail(&packet->list, queue);
 201		resources--;
 202	}
 203}
 204
 205static int htc_issue_packets(struct htc_target *target,
 206			     struct htc_endpoint *ep,
 207			     struct list_head *pkt_queue)
 208{
 209	int status = 0;
 210	u16 payload_len;
 211	struct sk_buff *skb;
 212	struct htc_frame_hdr *htc_hdr;
 213	struct htc_packet *packet;
 214
 215	ath6kl_dbg(ATH6KL_DBG_HTC,
 216		   "%s: queue: 0x%p, pkts %d\n", __func__,
 217		   pkt_queue, get_queue_depth(pkt_queue));
 218
 219	while (!list_empty(pkt_queue)) {
 220		packet = list_first_entry(pkt_queue, struct htc_packet, list);
 221		list_del(&packet->list);
 222
 223		skb = packet->skb;
 224		if (!skb) {
 225			WARN_ON_ONCE(1);
 226			status = -EINVAL;
 227			break;
 228		}
 229
 230		payload_len = packet->act_len;
 231
 232		/* setup HTC frame header */
 233		htc_hdr = (struct htc_frame_hdr *) skb_push(skb,
 234							    sizeof(*htc_hdr));
 235		if (!htc_hdr) {
 236			WARN_ON_ONCE(1);
 237			status = -EINVAL;
 238			break;
 239		}
 240
 241		packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
 242
 243		/* Endianess? */
 244		put_unaligned((u16) payload_len, &htc_hdr->payld_len);
 245		htc_hdr->flags = packet->info.tx.flags;
 246		htc_hdr->eid = (u8) packet->endpoint;
 247		htc_hdr->ctrl[0] = 0;
 248		htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
 249
 250		spin_lock_bh(&target->tx_lock);
 251
 252		/* store in look up queue to match completions */
 253		list_add_tail(&packet->list, &ep->pipe.tx_lookup_queue);
 254		ep->ep_st.tx_issued += 1;
 255		spin_unlock_bh(&target->tx_lock);
 256
 257		status = ath6kl_hif_pipe_send(target->dev->ar,
 258					      ep->pipe.pipeid_ul, NULL, skb);
 259
 260		if (status != 0) {
 261			if (status != -ENOMEM) {
 262				/* TODO: if more than 1 endpoint maps to the
 263				 * same PipeID, it is possible to run out of
 264				 * resources in the HIF layer.
 265				 * Don't emit the error
 266				 */
 267				ath6kl_dbg(ATH6KL_DBG_HTC,
 268					   "%s: failed status:%d\n",
 269					   __func__, status);
 270			}
 271			spin_lock_bh(&target->tx_lock);
 272			list_del(&packet->list);
 273
 274			/* reclaim credits */
 275			ep->cred_dist.credits += packet->info.tx.cred_used;
 276			spin_unlock_bh(&target->tx_lock);
 277
 278			/* put it back into the callers queue */
 279			list_add(&packet->list, pkt_queue);
 280			break;
 281		}
 282
 283	}
 284
 285	if (status != 0) {
 286		while (!list_empty(pkt_queue)) {
 287			if (status != -ENOMEM) {
 288				ath6kl_dbg(ATH6KL_DBG_HTC,
 289					   "%s: failed pkt:0x%p status:%d\n",
 290					   __func__, packet, status);
 291			}
 292
 293			packet = list_first_entry(pkt_queue,
 294						  struct htc_packet, list);
 295			list_del(&packet->list);
 296			packet->status = status;
 297			send_packet_completion(target, packet);
 298		}
 299	}
 300
 301	return status;
 302}
 303
 304static enum htc_send_queue_result htc_try_send(struct htc_target *target,
 305					       struct htc_endpoint *ep,
 306					       struct list_head *txq)
 307{
 308	struct list_head send_queue;	/* temp queue to hold packets */
 309	struct htc_packet *packet, *tmp_pkt;
 310	struct ath6kl *ar = target->dev->ar;
 311	enum htc_send_full_action action;
 312	int tx_resources, overflow, txqueue_depth, i, good_pkts;
 313	u8 pipeid;
 314
 315	ath6kl_dbg(ATH6KL_DBG_HTC, "%s: (queue:0x%p depth:%d)\n",
 316		   __func__, txq,
 317		   (txq == NULL) ? 0 : get_queue_depth(txq));
 318
 319	/* init the local send queue */
 320	INIT_LIST_HEAD(&send_queue);
 321
 322	/*
 323	 * txq equals to NULL means
 324	 * caller didn't provide a queue, just wants us to
 325	 * check queues and send
 326	 */
 327	if (txq != NULL) {
 328		if (list_empty(txq)) {
 329			/* empty queue */
 330			return HTC_SEND_QUEUE_DROP;
 331		}
 332
 333		spin_lock_bh(&target->tx_lock);
 334		txqueue_depth = get_queue_depth(&ep->txq);
 335		spin_unlock_bh(&target->tx_lock);
 336
 337		if (txqueue_depth >= ep->max_txq_depth) {
 338			/* we've already overflowed */
 339			overflow = get_queue_depth(txq);
 340		} else {
 341			/* get how much we will overflow by */
 342			overflow = txqueue_depth;
 343			overflow += get_queue_depth(txq);
 344			/* get how much we will overflow the TX queue by */
 345			overflow -= ep->max_txq_depth;
 346		}
 347
 348		/* if overflow is negative or zero, we are okay */
 349		if (overflow > 0) {
 350			ath6kl_dbg(ATH6KL_DBG_HTC,
 351				   "%s: Endpoint %d, TX queue will overflow :%d, Tx Depth:%d, Max:%d\n",
 352				   __func__, ep->eid, overflow, txqueue_depth,
 353				   ep->max_txq_depth);
 354		}
 355		if ((overflow <= 0) ||
 356		    (ep->ep_cb.tx_full == NULL)) {
 357			/*
 358			 * all packets will fit or caller did not provide send
 359			 * full indication handler -- just move all of them
 360			 * to the local send_queue object
 361			 */
 362			list_splice_tail_init(txq, &send_queue);
 363		} else {
 364			good_pkts = get_queue_depth(txq) - overflow;
 365			if (good_pkts < 0) {
 366				WARN_ON_ONCE(1);
 367				return HTC_SEND_QUEUE_DROP;
 368			}
 369
 370			/* we have overflowed, and a callback is provided */
 371			/* dequeue all non-overflow packets to the sendqueue */
 372			for (i = 0; i < good_pkts; i++) {
 373				/* pop off caller's queue */
 374				packet = list_first_entry(txq,
 375							  struct htc_packet,
 376							  list);
 377				list_del(&packet->list);
 378				/* insert into local queue */
 379				list_add_tail(&packet->list, &send_queue);
 380			}
 381
 382			/*
 383			 * the caller's queue has all the packets that won't fit
 384			 * walk through the caller's queue and indicate each to
 385			 * the send full handler
 386			 */
 387			list_for_each_entry_safe(packet, tmp_pkt,
 388						 txq, list) {
 389
 390				ath6kl_dbg(ATH6KL_DBG_HTC,
 391					   "%s: Indicat overflowed TX pkts: %p\n",
 392					   __func__, packet);
 393				action = ep->ep_cb.tx_full(ep->target, packet);
 394				if (action == HTC_SEND_FULL_DROP) {
 395					/* callback wants the packet dropped */
 396					ep->ep_st.tx_dropped += 1;
 397
 398					/* leave this one in the caller's queue
 399					 * for cleanup */
 400				} else {
 401					/* callback wants to keep this packet,
 402					 * remove from caller's queue */
 403					list_del(&packet->list);
 404					/* put it in the send queue */
 405					list_add_tail(&packet->list,
 406						      &send_queue);
 407				}
 408
 409			}
 410
 411			if (list_empty(&send_queue)) {
 412				/* no packets made it in, caller will cleanup */
 413				return HTC_SEND_QUEUE_DROP;
 414			}
 415		}
 416	}
 417
 418	if (!ep->pipe.tx_credit_flow_enabled) {
 419		tx_resources =
 420		    ath6kl_hif_pipe_get_free_queue_number(ar,
 421							  ep->pipe.pipeid_ul);
 422	} else {
 423		tx_resources = 0;
 424	}
 425
 426	spin_lock_bh(&target->tx_lock);
 427	if (!list_empty(&send_queue)) {
 428		/* transfer packets to tail */
 429		list_splice_tail_init(&send_queue, &ep->txq);
 430		if (!list_empty(&send_queue)) {
 431			WARN_ON_ONCE(1);
 432			spin_unlock_bh(&target->tx_lock);
 433			return HTC_SEND_QUEUE_DROP;
 434		}
 435		INIT_LIST_HEAD(&send_queue);
 436	}
 437
 438	/* increment tx processing count on entry */
 439	ep->tx_proc_cnt++;
 440
 441	if (ep->tx_proc_cnt > 1) {
 442		/*
 443		 * Another thread or task is draining the TX queues on this
 444		 * endpoint that thread will reset the tx processing count
 445		 * when the queue is drained.
 446		 */
 447		ep->tx_proc_cnt--;
 448		spin_unlock_bh(&target->tx_lock);
 449		return HTC_SEND_QUEUE_OK;
 450	}
 451
 452	/***** beyond this point only 1 thread may enter ******/
 453
 454	/*
 455	 * Now drain the endpoint TX queue for transmission as long as we have
 456	 * enough transmit resources.
 457	 */
 458	while (true) {
 459
 460		if (get_queue_depth(&ep->txq) == 0)
 461			break;
 462
 463		if (ep->pipe.tx_credit_flow_enabled) {
 464			/*
 465			 * Credit based mechanism provides flow control
 466			 * based on target transmit resource availability,
 467			 * we assume that the HIF layer will always have
 468			 * bus resources greater than target transmit
 469			 * resources.
 470			 */
 471			get_htc_packet_credit_based(target, ep, &send_queue);
 472		} else {
 473			/*
 474			 * Get all packets for this endpoint that we can
 475			 * for this pass.
 476			 */
 477			get_htc_packet(target, ep, &send_queue, tx_resources);
 478		}
 479
 480		if (get_queue_depth(&send_queue) == 0) {
 481			/*
 482			 * Didn't get packets due to out of resources or TX
 483			 * queue was drained.
 484			 */
 485			break;
 486		}
 487
 488		spin_unlock_bh(&target->tx_lock);
 489
 490		/* send what we can */
 491		htc_issue_packets(target, ep, &send_queue);
 492
 493		if (!ep->pipe.tx_credit_flow_enabled) {
 494			pipeid = ep->pipe.pipeid_ul;
 495			tx_resources =
 496			    ath6kl_hif_pipe_get_free_queue_number(ar, pipeid);
 497		}
 498
 499		spin_lock_bh(&target->tx_lock);
 
 500
 501	}
 502	/* done with this endpoint, we can clear the count */
 503	ep->tx_proc_cnt = 0;
 504	spin_unlock_bh(&target->tx_lock);
 505
 506	return HTC_SEND_QUEUE_OK;
 507}
 508
 509/* htc control packet manipulation */
 510static void destroy_htc_txctrl_packet(struct htc_packet *packet)
 511{
 512	struct sk_buff *skb;
 513	skb = packet->skb;
 514	if (skb != NULL)
 515		dev_kfree_skb(skb);
 516
 517	kfree(packet);
 518}
 519
 520static struct htc_packet *build_htc_txctrl_packet(void)
 521{
 522	struct htc_packet *packet = NULL;
 523	struct sk_buff *skb;
 524
 525	packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
 526	if (packet == NULL)
 527		return NULL;
 528
 529	skb = __dev_alloc_skb(HTC_CONTROL_BUFFER_SIZE, GFP_KERNEL);
 530
 531	if (skb == NULL) {
 532		kfree(packet);
 533		return NULL;
 534	}
 535	packet->skb = skb;
 536
 537	return packet;
 538}
 539
 540static void htc_free_txctrl_packet(struct htc_target *target,
 541				   struct htc_packet *packet)
 542{
 543	destroy_htc_txctrl_packet(packet);
 544}
 545
 546static struct htc_packet *htc_alloc_txctrl_packet(struct htc_target *target)
 547{
 548	return build_htc_txctrl_packet();
 549}
 550
 551static void htc_txctrl_complete(struct htc_target *target,
 552				struct htc_packet *packet)
 553{
 554	htc_free_txctrl_packet(target, packet);
 555}
 556
 557#define MAX_MESSAGE_SIZE 1536
 558
 559static int htc_setup_target_buffer_assignments(struct htc_target *target)
 560{
 561	int status, credits, credit_per_maxmsg, i;
 562	struct htc_pipe_txcredit_alloc *entry;
 563	unsigned int hif_usbaudioclass = 0;
 564
 565	credit_per_maxmsg = MAX_MESSAGE_SIZE / target->tgt_cred_sz;
 566	if (MAX_MESSAGE_SIZE % target->tgt_cred_sz)
 567		credit_per_maxmsg++;
 568
 569	/* TODO, this should be configured by the caller! */
 570
 571	credits = target->tgt_creds;
 572	entry = &target->pipe.txcredit_alloc[0];
 573
 574	status = -ENOMEM;
 575
 576	/* FIXME: hif_usbaudioclass is always zero */
 577	if (hif_usbaudioclass) {
 578		ath6kl_dbg(ATH6KL_DBG_HTC,
 579			   "%s: For USB Audio Class- Total:%d\n",
 580			   __func__, credits);
 581		entry++;
 582		entry++;
 583		/* Setup VO Service To have Max Credits */
 584		entry->service_id = WMI_DATA_VO_SVC;
 585		entry->credit_alloc = (credits - 6);
 586		if (entry->credit_alloc == 0)
 587			entry->credit_alloc++;
 588
 589		credits -= (int) entry->credit_alloc;
 590		if (credits <= 0)
 591			return status;
 592
 593		entry++;
 594		entry->service_id = WMI_CONTROL_SVC;
 595		entry->credit_alloc = credit_per_maxmsg;
 596		credits -= (int) entry->credit_alloc;
 597		if (credits <= 0)
 598			return status;
 599
 600		/* leftovers go to best effort */
 601		entry++;
 602		entry++;
 603		entry->service_id = WMI_DATA_BE_SVC;
 604		entry->credit_alloc = (u8) credits;
 605		status = 0;
 606	} else {
 607		entry++;
 608		entry->service_id = WMI_DATA_VI_SVC;
 609		entry->credit_alloc = credits / 4;
 610		if (entry->credit_alloc == 0)
 611			entry->credit_alloc++;
 612
 613		credits -= (int) entry->credit_alloc;
 614		if (credits <= 0)
 615			return status;
 616
 617		entry++;
 618		entry->service_id = WMI_DATA_VO_SVC;
 619		entry->credit_alloc = credits / 4;
 620		if (entry->credit_alloc == 0)
 621			entry->credit_alloc++;
 622
 623		credits -= (int) entry->credit_alloc;
 624		if (credits <= 0)
 625			return status;
 626
 627		entry++;
 628		entry->service_id = WMI_CONTROL_SVC;
 629		entry->credit_alloc = credit_per_maxmsg;
 630		credits -= (int) entry->credit_alloc;
 631		if (credits <= 0)
 632			return status;
 633
 634		entry++;
 635		entry->service_id = WMI_DATA_BK_SVC;
 636		entry->credit_alloc = credit_per_maxmsg;
 637		credits -= (int) entry->credit_alloc;
 638		if (credits <= 0)
 639			return status;
 640
 641		/* leftovers go to best effort */
 642		entry++;
 643		entry->service_id = WMI_DATA_BE_SVC;
 644		entry->credit_alloc = (u8) credits;
 645		status = 0;
 646	}
 647
 648	if (status == 0) {
 649		for (i = 0; i < ENDPOINT_MAX; i++) {
 650			if (target->pipe.txcredit_alloc[i].service_id != 0) {
 651				ath6kl_dbg(ATH6KL_DBG_HTC,
 652					   "HTC Service Index : %d TX : 0x%2.2X : alloc:%d\n",
 653					   i,
 654					   target->pipe.txcredit_alloc[i].
 655					   service_id,
 656					   target->pipe.txcredit_alloc[i].
 657					   credit_alloc);
 658			}
 659		}
 660	}
 661	return status;
 662}
 663
 664/* process credit reports and call distribution function */
 665static void htc_process_credit_report(struct htc_target *target,
 666				      struct htc_credit_report *rpt,
 667				      int num_entries,
 668				      enum htc_endpoint_id from_ep)
 669{
 670	int total_credits = 0, i;
 671	struct htc_endpoint *ep;
 672
 673	/* lock out TX while we update credits */
 674	spin_lock_bh(&target->tx_lock);
 675
 676	for (i = 0; i < num_entries; i++, rpt++) {
 677		if (rpt->eid >= ENDPOINT_MAX) {
 678			WARN_ON_ONCE(1);
 679			spin_unlock_bh(&target->tx_lock);
 680			return;
 681		}
 682
 683		ep = &target->endpoint[rpt->eid];
 684		ep->cred_dist.credits += rpt->credits;
 685
 686		if (ep->cred_dist.credits && get_queue_depth(&ep->txq)) {
 687			spin_unlock_bh(&target->tx_lock);
 688			htc_try_send(target, ep, NULL);
 689			spin_lock_bh(&target->tx_lock);
 690		}
 691
 692		total_credits += rpt->credits;
 693	}
 694	ath6kl_dbg(ATH6KL_DBG_HTC,
 695		   "Report indicated %d credits to distribute\n",
 696		   total_credits);
 697
 698	spin_unlock_bh(&target->tx_lock);
 699}
 700
 701/* flush endpoint TX queue */
 702static void htc_flush_tx_endpoint(struct htc_target *target,
 703				  struct htc_endpoint *ep, u16 tag)
 704{
 705	struct htc_packet *packet;
 706
 707	spin_lock_bh(&target->tx_lock);
 708	while (get_queue_depth(&ep->txq)) {
 709		packet = list_first_entry(&ep->txq, struct htc_packet, list);
 710		list_del(&packet->list);
 711		packet->status = 0;
 712		send_packet_completion(target, packet);
 713	}
 714	spin_unlock_bh(&target->tx_lock);
 715}
 716
 717/*
 718 * In the adapted HIF layer, struct sk_buff * are passed between HIF and HTC,
 719 * since upper layers expects struct htc_packet containers we use the completed
 720 * skb and lookup it's corresponding HTC packet buffer from a lookup list.
 721 * This is extra overhead that can be fixed by re-aligning HIF interfaces with
 722 * HTC.
 723 */
 724static struct htc_packet *htc_lookup_tx_packet(struct htc_target *target,
 725					       struct htc_endpoint *ep,
 726					       struct sk_buff *skb)
 727{
 728	struct htc_packet *packet, *tmp_pkt, *found_packet = NULL;
 729
 730	spin_lock_bh(&target->tx_lock);
 731
 732	/*
 733	 * interate from the front of tx lookup queue
 734	 * this lookup should be fast since lower layers completes in-order and
 735	 * so the completed packet should be at the head of the list generally
 736	 */
 737	list_for_each_entry_safe(packet, tmp_pkt, &ep->pipe.tx_lookup_queue,
 738				 list) {
 739		/* check for removal */
 740		if (skb == packet->skb) {
 741			/* found it */
 742			list_del(&packet->list);
 743			found_packet = packet;
 744			break;
 745		}
 746	}
 747
 748	spin_unlock_bh(&target->tx_lock);
 749
 750	return found_packet;
 751}
 752
 753static int ath6kl_htc_pipe_tx_complete(struct ath6kl *ar, struct sk_buff *skb)
 754{
 755	struct htc_target *target = ar->htc_target;
 756	struct htc_frame_hdr *htc_hdr;
 757	struct htc_endpoint *ep;
 758	struct htc_packet *packet;
 759	u8 ep_id, *netdata;
 760	u32 netlen;
 761
 762	netdata = skb->data;
 763	netlen = skb->len;
 764
 765	htc_hdr = (struct htc_frame_hdr *) netdata;
 766
 767	ep_id = htc_hdr->eid;
 768	ep = &target->endpoint[ep_id];
 769
 770	packet = htc_lookup_tx_packet(target, ep, skb);
 771	if (packet == NULL) {
 772		/* may have already been flushed and freed */
 773		ath6kl_err("HTC TX lookup failed!\n");
 774	} else {
 775		/* will be giving this buffer back to upper layers */
 776		packet->status = 0;
 777		send_packet_completion(target, packet);
 778	}
 779	skb = NULL;
 780
 781	if (!ep->pipe.tx_credit_flow_enabled) {
 782		/*
 783		 * note: when using TX credit flow, the re-checking of queues
 784		 * happens when credits flow back from the target. in the
 785		 * non-TX credit case, we recheck after the packet completes
 786		 */
 787		htc_try_send(target, ep, NULL);
 788	}
 789
 790	return 0;
 791}
 792
 793static int htc_send_packets_multiple(struct htc_target *target,
 794				     struct list_head *pkt_queue)
 795{
 796	struct htc_endpoint *ep;
 797	struct htc_packet *packet, *tmp_pkt;
 798
 799	if (list_empty(pkt_queue))
 800		return -EINVAL;
 801
 802	/* get first packet to find out which ep the packets will go into */
 803	packet = list_first_entry(pkt_queue, struct htc_packet, list);
 804
 805	if (packet->endpoint >= ENDPOINT_MAX) {
 806		WARN_ON_ONCE(1);
 807		return -EINVAL;
 808	}
 809	ep = &target->endpoint[packet->endpoint];
 810
 811	htc_try_send(target, ep, pkt_queue);
 812
 813	/* do completion on any packets that couldn't get in */
 814	if (!list_empty(pkt_queue)) {
 815		list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
 816			packet->status = -ENOMEM;
 817		}
 818
 819		do_send_completion(ep, pkt_queue);
 820	}
 821
 822	return 0;
 823}
 824
 825/* htc pipe rx path */
 826static struct htc_packet *alloc_htc_packet_container(struct htc_target *target)
 827{
 828	struct htc_packet *packet;
 829	spin_lock_bh(&target->rx_lock);
 830
 831	if (target->pipe.htc_packet_pool == NULL) {
 832		spin_unlock_bh(&target->rx_lock);
 833		return NULL;
 834	}
 835
 836	packet = target->pipe.htc_packet_pool;
 837	target->pipe.htc_packet_pool = (struct htc_packet *) packet->list.next;
 838
 839	spin_unlock_bh(&target->rx_lock);
 840
 841	packet->list.next = NULL;
 842	return packet;
 843}
 844
 845static void free_htc_packet_container(struct htc_target *target,
 846				      struct htc_packet *packet)
 847{
 848	struct list_head *lh;
 849
 850	spin_lock_bh(&target->rx_lock);
 851
 852	if (target->pipe.htc_packet_pool == NULL) {
 853		target->pipe.htc_packet_pool = packet;
 854		packet->list.next = NULL;
 855	} else {
 856		lh = (struct list_head *) target->pipe.htc_packet_pool;
 857		packet->list.next = lh;
 858		target->pipe.htc_packet_pool = packet;
 859	}
 860
 861	spin_unlock_bh(&target->rx_lock);
 862}
 863
 864static int htc_process_trailer(struct htc_target *target, u8 *buffer,
 865			       int len, enum htc_endpoint_id from_ep)
 866{
 867	struct htc_credit_report *report;
 868	struct htc_record_hdr *record;
 869	u8 *record_buf, *orig_buf;
 870	int orig_len, status;
 871
 872	orig_buf = buffer;
 873	orig_len = len;
 874	status = 0;
 875
 876	while (len > 0) {
 877		if (len < sizeof(struct htc_record_hdr)) {
 878			status = -EINVAL;
 879			break;
 880		}
 881
 882		/* these are byte aligned structs */
 883		record = (struct htc_record_hdr *) buffer;
 884		len -= sizeof(struct htc_record_hdr);
 885		buffer += sizeof(struct htc_record_hdr);
 886
 887		if (record->len > len) {
 888			/* no room left in buffer for record */
 889			ath6kl_dbg(ATH6KL_DBG_HTC,
 890				   "invalid length: %d (id:%d) buffer has: %d bytes left\n",
 891				   record->len, record->rec_id, len);
 892			status = -EINVAL;
 893			break;
 894		}
 895
 896		/* start of record follows the header */
 897		record_buf = buffer;
 898
 899		switch (record->rec_id) {
 900		case HTC_RECORD_CREDITS:
 901			if (record->len < sizeof(struct htc_credit_report)) {
 902				WARN_ON_ONCE(1);
 903				return -EINVAL;
 904			}
 905
 906			report = (struct htc_credit_report *) record_buf;
 907			htc_process_credit_report(target, report,
 908						  record->len / sizeof(*report),
 909						  from_ep);
 910			break;
 911		default:
 912			ath6kl_dbg(ATH6KL_DBG_HTC,
 913				   "unhandled record: id:%d length:%d\n",
 914				   record->rec_id, record->len);
 915			break;
 916		}
 917
 918		if (status != 0)
 919			break;
 920
 921		/* advance buffer past this record for next time around */
 922		buffer += record->len;
 923		len -= record->len;
 924	}
 925
 926	return status;
 927}
 928
 929static void do_recv_completion(struct htc_endpoint *ep,
 930			       struct list_head *queue_to_indicate)
 931{
 932	struct htc_packet *packet;
 933
 934	if (list_empty(queue_to_indicate)) {
 935		/* nothing to indicate */
 936		return;
 937	}
 938
 939	/* using legacy EpRecv */
 940	while (!list_empty(queue_to_indicate)) {
 941		packet = list_first_entry(queue_to_indicate,
 942					  struct htc_packet, list);
 943		list_del(&packet->list);
 944		ep->ep_cb.rx(ep->target, packet);
 945	}
 946
 947	return;
 948}
 949
 950static void recv_packet_completion(struct htc_target *target,
 951				   struct htc_endpoint *ep,
 952				   struct htc_packet *packet)
 953{
 954	struct list_head container;
 955	INIT_LIST_HEAD(&container);
 956	list_add_tail(&packet->list, &container);
 957
 958	/* do completion */
 959	do_recv_completion(ep, &container);
 960}
 961
 962static int ath6kl_htc_pipe_rx_complete(struct ath6kl *ar, struct sk_buff *skb,
 963				       u8 pipeid)
 964{
 965	struct htc_target *target = ar->htc_target;
 966	u8 *netdata, *trailer, hdr_info;
 967	struct htc_frame_hdr *htc_hdr;
 968	u32 netlen, trailerlen = 0;
 969	struct htc_packet *packet;
 970	struct htc_endpoint *ep;
 971	u16 payload_len;
 972	int status = 0;
 973
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 974	netdata = skb->data;
 975	netlen = skb->len;
 976
 977	htc_hdr = (struct htc_frame_hdr *) netdata;
 978
 979	ep = &target->endpoint[htc_hdr->eid];
 980
 981	if (htc_hdr->eid >= ENDPOINT_MAX) {
 982		ath6kl_dbg(ATH6KL_DBG_HTC,
 983			   "HTC Rx: invalid EndpointID=%d\n",
 984			   htc_hdr->eid);
 985		status = -EINVAL;
 986		goto free_skb;
 987	}
 
 988
 989	payload_len = le16_to_cpu(get_unaligned(&htc_hdr->payld_len));
 990
 991	if (netlen < (payload_len + HTC_HDR_LENGTH)) {
 992		ath6kl_dbg(ATH6KL_DBG_HTC,
 993			   "HTC Rx: insufficient length, got:%d expected =%u\n",
 994			   netlen, payload_len + HTC_HDR_LENGTH);
 995		status = -EINVAL;
 996		goto free_skb;
 997	}
 998
 999	/* get flags to check for trailer */
1000	hdr_info = htc_hdr->flags;
1001	if (hdr_info & HTC_FLG_RX_TRAILER) {
1002		/* extract the trailer length */
1003		hdr_info = htc_hdr->ctrl[0];
1004		if ((hdr_info < sizeof(struct htc_record_hdr)) ||
1005		    (hdr_info > payload_len)) {
1006			ath6kl_dbg(ATH6KL_DBG_HTC,
1007				   "invalid header: payloadlen should be %d, CB[0]: %d\n",
1008				   payload_len, hdr_info);
1009			status = -EINVAL;
1010			goto free_skb;
1011		}
1012
1013		trailerlen = hdr_info;
1014		/* process trailer after hdr/apps payload */
1015		trailer = (u8 *) htc_hdr + HTC_HDR_LENGTH +
1016			payload_len - hdr_info;
1017		status = htc_process_trailer(target, trailer, hdr_info,
1018					     htc_hdr->eid);
1019		if (status != 0)
1020			goto free_skb;
1021	}
1022
1023	if (((int) payload_len - (int) trailerlen) <= 0) {
1024		/* zero length packet with trailer, just drop these */
1025		goto free_skb;
1026	}
1027
1028	if (htc_hdr->eid == ENDPOINT_0) {
1029		/* handle HTC control message */
1030		if (target->htc_flags & HTC_OP_STATE_SETUP_COMPLETE) {
1031			/*
1032			 * fatal: target should not send unsolicited
1033			 * messageson the endpoint 0
1034			 */
1035			ath6kl_dbg(ATH6KL_DBG_HTC,
1036				   "HTC ignores Rx Ctrl after setup complete\n");
1037			status = -EINVAL;
1038			goto free_skb;
1039		}
1040
1041		/* remove HTC header */
1042		skb_pull(skb, HTC_HDR_LENGTH);
1043
1044		netdata = skb->data;
1045		netlen = skb->len;
1046
1047		spin_lock_bh(&target->rx_lock);
1048
1049		target->pipe.ctrl_response_valid = true;
1050		target->pipe.ctrl_response_len = min_t(int, netlen,
1051						       HTC_MAX_CTRL_MSG_LEN);
1052		memcpy(target->pipe.ctrl_response_buf, netdata,
1053		       target->pipe.ctrl_response_len);
1054
1055		spin_unlock_bh(&target->rx_lock);
1056
1057		dev_kfree_skb(skb);
1058		skb = NULL;
 
1059		goto free_skb;
1060	}
1061
1062	/*
1063	 * TODO: the message based HIF architecture allocates net bufs
1064	 * for recv packets since it bridges that HIF to upper layers,
1065	 * which expects HTC packets, we form the packets here
1066	 */
1067	packet = alloc_htc_packet_container(target);
1068	if (packet == NULL) {
1069		status = -ENOMEM;
1070		goto free_skb;
1071	}
1072
1073	packet->status = 0;
1074	packet->endpoint = htc_hdr->eid;
1075	packet->pkt_cntxt = skb;
1076
1077	/* TODO: for backwards compatibility */
1078	packet->buf = skb_push(skb, 0) + HTC_HDR_LENGTH;
1079	packet->act_len = netlen - HTC_HDR_LENGTH - trailerlen;
1080
1081	/*
1082	 * TODO: this is a hack because the driver layer will set the
1083	 * actual len of the skb again which will just double the len
1084	 */
1085	skb_trim(skb, 0);
1086
1087	recv_packet_completion(target, ep, packet);
1088
1089	/* recover the packet container */
1090	free_htc_packet_container(target, packet);
1091	skb = NULL;
1092
1093free_skb:
1094	if (skb != NULL)
1095		dev_kfree_skb(skb);
1096
1097	return status;
1098
1099}
1100
1101static void htc_flush_rx_queue(struct htc_target *target,
1102			       struct htc_endpoint *ep)
1103{
1104	struct list_head container;
1105	struct htc_packet *packet;
1106
1107	spin_lock_bh(&target->rx_lock);
1108
1109	while (1) {
1110		if (list_empty(&ep->rx_bufq))
1111			break;
1112
1113		packet = list_first_entry(&ep->rx_bufq,
1114					  struct htc_packet, list);
1115		list_del(&packet->list);
1116
1117		spin_unlock_bh(&target->rx_lock);
1118		packet->status = -ECANCELED;
1119		packet->act_len = 0;
1120
1121		ath6kl_dbg(ATH6KL_DBG_HTC,
1122			   "Flushing RX packet:0x%p, length:%d, ep:%d\n",
1123			   packet, packet->buf_len,
1124			   packet->endpoint);
1125
1126		INIT_LIST_HEAD(&container);
1127		list_add_tail(&packet->list, &container);
1128
1129		/* give the packet back */
1130		do_recv_completion(ep, &container);
1131		spin_lock_bh(&target->rx_lock);
1132	}
1133
1134	spin_unlock_bh(&target->rx_lock);
1135}
1136
1137/* polling routine to wait for a control packet to be received */
1138static int htc_wait_recv_ctrl_message(struct htc_target *target)
1139{
1140	int count = HTC_TARGET_RESPONSE_POLL_COUNT;
1141
1142	while (count > 0) {
1143		spin_lock_bh(&target->rx_lock);
1144
1145		if (target->pipe.ctrl_response_valid) {
1146			target->pipe.ctrl_response_valid = false;
1147			spin_unlock_bh(&target->rx_lock);
1148			break;
1149		}
1150
1151		spin_unlock_bh(&target->rx_lock);
1152
1153		count--;
1154
1155		msleep_interruptible(HTC_TARGET_RESPONSE_POLL_WAIT);
1156	}
1157
1158	if (count <= 0) {
1159		ath6kl_dbg(ATH6KL_DBG_HTC, "%s: Timeout!\n", __func__);
1160		return -ECOMM;
1161	}
1162
1163	return 0;
1164}
1165
1166static void htc_rxctrl_complete(struct htc_target *context,
1167				struct htc_packet *packet)
1168{
1169	/* TODO, can't really receive HTC control messages yet.... */
1170	ath6kl_dbg(ATH6KL_DBG_HTC, "%s: invalid call function\n", __func__);
 
 
 
 
1171}
1172
1173/* htc pipe initialization */
1174static void reset_endpoint_states(struct htc_target *target)
1175{
1176	struct htc_endpoint *ep;
1177	int i;
1178
1179	for (i = ENDPOINT_0; i < ENDPOINT_MAX; i++) {
1180		ep = &target->endpoint[i];
1181		ep->svc_id = 0;
1182		ep->len_max = 0;
1183		ep->max_txq_depth = 0;
1184		ep->eid = i;
1185		INIT_LIST_HEAD(&ep->txq);
1186		INIT_LIST_HEAD(&ep->pipe.tx_lookup_queue);
1187		INIT_LIST_HEAD(&ep->rx_bufq);
1188		ep->target = target;
1189		ep->pipe.tx_credit_flow_enabled = (bool) 1; /* FIXME */
1190	}
1191}
1192
1193/* start HTC, this is called after all services are connected */
1194static int htc_config_target_hif_pipe(struct htc_target *target)
1195{
1196	return 0;
1197}
1198
1199/* htc service functions */
1200static u8 htc_get_credit_alloc(struct htc_target *target, u16 service_id)
1201{
1202	u8 allocation = 0;
1203	int i;
1204
1205	for (i = 0; i < ENDPOINT_MAX; i++) {
1206		if (target->pipe.txcredit_alloc[i].service_id == service_id)
1207			allocation =
1208				target->pipe.txcredit_alloc[i].credit_alloc;
1209	}
1210
1211	if (allocation == 0) {
1212		ath6kl_dbg(ATH6KL_DBG_HTC,
1213			   "HTC Service TX : 0x%2.2X : allocation is zero!\n",
1214			   service_id);
1215	}
1216
1217	return allocation;
1218}
1219
1220static int ath6kl_htc_pipe_conn_service(struct htc_target *target,
1221		     struct htc_service_connect_req *conn_req,
1222		     struct htc_service_connect_resp *conn_resp)
1223{
1224	struct ath6kl *ar = target->dev->ar;
1225	struct htc_packet *packet = NULL;
1226	struct htc_conn_service_resp *resp_msg;
1227	struct htc_conn_service_msg *conn_msg;
1228	enum htc_endpoint_id assigned_epid = ENDPOINT_MAX;
1229	bool disable_credit_flowctrl = false;
1230	unsigned int max_msg_size = 0;
1231	struct htc_endpoint *ep;
1232	int length, status = 0;
1233	struct sk_buff *skb;
1234	u8 tx_alloc;
1235	u16 flags;
1236
1237	if (conn_req->svc_id == 0) {
1238		WARN_ON_ONCE(1);
1239		status = -EINVAL;
1240		goto free_packet;
1241	}
1242
1243	if (conn_req->svc_id == HTC_CTRL_RSVD_SVC) {
1244		/* special case for pseudo control service */
1245		assigned_epid = ENDPOINT_0;
1246		max_msg_size = HTC_MAX_CTRL_MSG_LEN;
1247		tx_alloc = 0;
1248
1249	} else {
1250
1251		tx_alloc = htc_get_credit_alloc(target, conn_req->svc_id);
1252		if (tx_alloc == 0) {
1253			status = -ENOMEM;
1254			goto free_packet;
1255		}
1256
1257		/* allocate a packet to send to the target */
1258		packet = htc_alloc_txctrl_packet(target);
1259
1260		if (packet == NULL) {
1261			WARN_ON_ONCE(1);
1262			status = -ENOMEM;
1263			goto free_packet;
1264		}
1265
1266		skb = packet->skb;
1267		length = sizeof(struct htc_conn_service_msg);
1268
1269		/* assemble connect service message */
1270		conn_msg = (struct htc_conn_service_msg *) skb_put(skb,
1271								   length);
1272		if (conn_msg == NULL) {
1273			WARN_ON_ONCE(1);
1274			status = -EINVAL;
1275			goto free_packet;
1276		}
1277
1278		memset(conn_msg, 0,
1279		       sizeof(struct htc_conn_service_msg));
1280		conn_msg->msg_id = cpu_to_le16(HTC_MSG_CONN_SVC_ID);
1281		conn_msg->svc_id = cpu_to_le16(conn_req->svc_id);
1282		conn_msg->conn_flags = cpu_to_le16(conn_req->conn_flags &
1283					~HTC_CONN_FLGS_SET_RECV_ALLOC_MASK);
1284
1285		/* tell target desired recv alloc for this ep */
1286		flags = tx_alloc << HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT;
1287		conn_msg->conn_flags |= cpu_to_le16(flags);
1288
1289		if (conn_req->conn_flags &
1290		    HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL) {
1291			disable_credit_flowctrl = true;
1292		}
1293
1294		set_htc_pkt_info(packet, NULL, (u8 *) conn_msg,
1295				 length,
1296				 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1297
1298		status = ath6kl_htc_pipe_tx(target, packet);
1299
1300		/* we don't own it anymore */
1301		packet = NULL;
1302		if (status != 0)
1303			goto free_packet;
1304
1305		/* wait for response */
1306		status = htc_wait_recv_ctrl_message(target);
1307		if (status != 0)
1308			goto free_packet;
1309
1310		/* we controlled the buffer creation so it has to be
1311		 * properly aligned
1312		 */
1313		resp_msg = (struct htc_conn_service_resp *)
1314		    target->pipe.ctrl_response_buf;
1315
1316		if (resp_msg->msg_id != cpu_to_le16(HTC_MSG_CONN_SVC_RESP_ID) ||
1317		    (target->pipe.ctrl_response_len < sizeof(*resp_msg))) {
1318			/* this message is not valid */
1319			WARN_ON_ONCE(1);
1320			status = -EINVAL;
1321			goto free_packet;
1322		}
1323
1324		ath6kl_dbg(ATH6KL_DBG_TRC,
1325			   "%s: service 0x%X conn resp: status: %d ep: %d\n",
1326			   __func__, resp_msg->svc_id, resp_msg->status,
1327			   resp_msg->eid);
1328
1329		conn_resp->resp_code = resp_msg->status;
1330		/* check response status */
1331		if (resp_msg->status != HTC_SERVICE_SUCCESS) {
1332			ath6kl_dbg(ATH6KL_DBG_HTC,
1333				   "Target failed service 0x%X connect request (status:%d)\n",
1334				   resp_msg->svc_id, resp_msg->status);
1335			status = -EINVAL;
1336			goto free_packet;
1337		}
1338
1339		assigned_epid = (enum htc_endpoint_id) resp_msg->eid;
1340		max_msg_size = le16_to_cpu(resp_msg->max_msg_sz);
1341	}
1342
1343	/* the rest are parameter checks so set the error status */
1344	status = -EINVAL;
1345
1346	if (assigned_epid >= ENDPOINT_MAX) {
1347		WARN_ON_ONCE(1);
1348		goto free_packet;
1349	}
1350
1351	if (max_msg_size == 0) {
1352		WARN_ON_ONCE(1);
1353		goto free_packet;
1354	}
1355
1356	ep = &target->endpoint[assigned_epid];
1357	ep->eid = assigned_epid;
1358	if (ep->svc_id != 0) {
1359		/* endpoint already in use! */
1360		WARN_ON_ONCE(1);
1361		goto free_packet;
1362	}
1363
1364	/* return assigned endpoint to caller */
1365	conn_resp->endpoint = assigned_epid;
1366	conn_resp->len_max = max_msg_size;
1367
1368	/* setup the endpoint */
1369	ep->svc_id = conn_req->svc_id; /* this marks ep in use */
1370	ep->max_txq_depth = conn_req->max_txq_depth;
1371	ep->len_max = max_msg_size;
1372	ep->cred_dist.credits = tx_alloc;
1373	ep->cred_dist.cred_sz = target->tgt_cred_sz;
1374	ep->cred_dist.cred_per_msg = max_msg_size / target->tgt_cred_sz;
1375	if (max_msg_size % target->tgt_cred_sz)
1376		ep->cred_dist.cred_per_msg++;
1377
1378	/* copy all the callbacks */
1379	ep->ep_cb = conn_req->ep_cb;
1380
1381	/* initialize tx_drop_packet_threshold */
1382	ep->tx_drop_packet_threshold = MAX_HI_COOKIE_NUM;
1383
1384	status = ath6kl_hif_pipe_map_service(ar, ep->svc_id,
1385					     &ep->pipe.pipeid_ul,
1386					     &ep->pipe.pipeid_dl);
1387	if (status != 0)
1388		goto free_packet;
1389
1390	ath6kl_dbg(ATH6KL_DBG_HTC,
1391		   "SVC Ready: 0x%4.4X: ULpipe:%d DLpipe:%d id:%d\n",
1392		   ep->svc_id, ep->pipe.pipeid_ul,
1393		   ep->pipe.pipeid_dl, ep->eid);
1394
1395	if (disable_credit_flowctrl && ep->pipe.tx_credit_flow_enabled) {
1396		ep->pipe.tx_credit_flow_enabled = false;
1397		ath6kl_dbg(ATH6KL_DBG_HTC,
1398			   "SVC: 0x%4.4X ep:%d TX flow control off\n",
1399			   ep->svc_id, assigned_epid);
1400	}
1401
1402free_packet:
1403	if (packet != NULL)
1404		htc_free_txctrl_packet(target, packet);
1405	return status;
1406}
1407
1408/* htc export functions */
1409static void *ath6kl_htc_pipe_create(struct ath6kl *ar)
1410{
1411	int status = 0;
1412	struct htc_endpoint *ep = NULL;
1413	struct htc_target *target = NULL;
1414	struct htc_packet *packet;
1415	int i;
1416
1417	target = kzalloc(sizeof(struct htc_target), GFP_KERNEL);
1418	if (target == NULL) {
1419		ath6kl_err("htc create unable to allocate memory\n");
1420		status = -ENOMEM;
1421		goto fail_htc_create;
1422	}
1423
1424	spin_lock_init(&target->htc_lock);
1425	spin_lock_init(&target->rx_lock);
1426	spin_lock_init(&target->tx_lock);
1427
1428	reset_endpoint_states(target);
1429
1430	for (i = 0; i < HTC_PACKET_CONTAINER_ALLOCATION; i++) {
1431		packet = kzalloc(sizeof(struct htc_packet), GFP_KERNEL);
1432
1433		if (packet != NULL)
1434			free_htc_packet_container(target, packet);
1435	}
1436
1437	target->dev = kzalloc(sizeof(*target->dev), GFP_KERNEL);
1438	if (!target->dev) {
1439		ath6kl_err("unable to allocate memory\n");
1440		status = -ENOMEM;
1441		goto fail_htc_create;
1442	}
1443	target->dev->ar = ar;
1444	target->dev->htc_cnxt = target;
1445
1446	/* Get HIF default pipe for HTC message exchange */
1447	ep = &target->endpoint[ENDPOINT_0];
1448
1449	ath6kl_hif_pipe_get_default(ar, &ep->pipe.pipeid_ul,
1450				    &ep->pipe.pipeid_dl);
1451
1452	return target;
1453
1454fail_htc_create:
1455	if (status != 0) {
1456		if (target != NULL)
1457			ath6kl_htc_pipe_cleanup(target);
1458
1459		target = NULL;
1460	}
1461	return target;
1462}
1463
1464/* cleanup the HTC instance */
1465static void ath6kl_htc_pipe_cleanup(struct htc_target *target)
1466{
1467	struct htc_packet *packet;
1468
1469	while (true) {
1470		packet = alloc_htc_packet_container(target);
1471		if (packet == NULL)
1472			break;
1473		kfree(packet);
1474	}
1475
1476	kfree(target->dev);
1477
1478	/* kfree our instance */
1479	kfree(target);
1480}
1481
1482static int ath6kl_htc_pipe_start(struct htc_target *target)
1483{
1484	struct sk_buff *skb;
1485	struct htc_setup_comp_ext_msg *setup;
1486	struct htc_packet *packet;
1487
1488	htc_config_target_hif_pipe(target);
1489
1490	/* allocate a buffer to send */
1491	packet = htc_alloc_txctrl_packet(target);
1492	if (packet == NULL) {
1493		WARN_ON_ONCE(1);
1494		return -ENOMEM;
1495	}
1496
1497	skb = packet->skb;
1498
1499	/* assemble setup complete message */
1500	setup = (struct htc_setup_comp_ext_msg *) skb_put(skb,
1501							  sizeof(*setup));
1502	memset(setup, 0, sizeof(struct htc_setup_comp_ext_msg));
1503	setup->msg_id = cpu_to_le16(HTC_MSG_SETUP_COMPLETE_EX_ID);
1504
1505	ath6kl_dbg(ATH6KL_DBG_HTC, "HTC using TX credit flow control\n");
1506
1507	set_htc_pkt_info(packet, NULL, (u8 *) setup,
1508			 sizeof(struct htc_setup_comp_ext_msg),
1509			 ENDPOINT_0, HTC_SERVICE_TX_PACKET_TAG);
1510
1511	target->htc_flags |= HTC_OP_STATE_SETUP_COMPLETE;
1512
1513	return ath6kl_htc_pipe_tx(target, packet);
1514}
1515
1516static void ath6kl_htc_pipe_stop(struct htc_target *target)
1517{
1518	int i;
1519	struct htc_endpoint *ep;
1520
1521	/* cleanup endpoints */
1522	for (i = 0; i < ENDPOINT_MAX; i++) {
1523		ep = &target->endpoint[i];
1524		htc_flush_rx_queue(target, ep);
1525		htc_flush_tx_endpoint(target, ep, HTC_TX_PACKET_TAG_ALL);
1526	}
1527
1528	reset_endpoint_states(target);
1529	target->htc_flags &= ~HTC_OP_STATE_SETUP_COMPLETE;
1530}
1531
1532static int ath6kl_htc_pipe_get_rxbuf_num(struct htc_target *target,
1533					 enum htc_endpoint_id endpoint)
1534{
1535	int num;
1536
1537	spin_lock_bh(&target->rx_lock);
1538	num = get_queue_depth(&(target->endpoint[endpoint].rx_bufq));
1539	spin_unlock_bh(&target->rx_lock);
1540
1541	return num;
1542}
1543
1544static int ath6kl_htc_pipe_tx(struct htc_target *target,
1545			      struct htc_packet *packet)
1546{
1547	struct list_head queue;
1548
1549	ath6kl_dbg(ATH6KL_DBG_HTC,
1550		   "%s: endPointId: %d, buffer: 0x%p, length: %d\n",
1551		   __func__, packet->endpoint, packet->buf,
1552		   packet->act_len);
1553
1554	INIT_LIST_HEAD(&queue);
1555	list_add_tail(&packet->list, &queue);
1556
1557	return htc_send_packets_multiple(target, &queue);
1558}
1559
1560static int ath6kl_htc_pipe_wait_target(struct htc_target *target)
1561{
1562	struct htc_ready_ext_msg *ready_msg;
1563	struct htc_service_connect_req connect;
1564	struct htc_service_connect_resp resp;
1565	int status = 0;
1566
1567	status = htc_wait_recv_ctrl_message(target);
1568
1569	if (status != 0)
1570		return status;
1571
1572	if (target->pipe.ctrl_response_len < sizeof(*ready_msg)) {
1573		ath6kl_dbg(ATH6KL_DBG_HTC, "invalid htc ready msg len:%d!\n",
1574			   target->pipe.ctrl_response_len);
1575		return -ECOMM;
1576	}
1577
1578	ready_msg = (struct htc_ready_ext_msg *) target->pipe.ctrl_response_buf;
1579
1580	if (ready_msg->ver2_0_info.msg_id != cpu_to_le16(HTC_MSG_READY_ID)) {
1581		ath6kl_dbg(ATH6KL_DBG_HTC, "invalid htc ready msg : 0x%X !\n",
1582			   ready_msg->ver2_0_info.msg_id);
1583		return -ECOMM;
1584	}
1585
1586	ath6kl_dbg(ATH6KL_DBG_HTC,
1587		   "Target Ready! : transmit resources : %d size:%d\n",
1588		   ready_msg->ver2_0_info.cred_cnt,
1589		   ready_msg->ver2_0_info.cred_sz);
1590
1591	target->tgt_creds = le16_to_cpu(ready_msg->ver2_0_info.cred_cnt);
1592	target->tgt_cred_sz = le16_to_cpu(ready_msg->ver2_0_info.cred_sz);
1593
1594	if ((target->tgt_creds == 0) || (target->tgt_cred_sz == 0))
1595		return -ECOMM;
1596
1597	htc_setup_target_buffer_assignments(target);
1598
1599	/* setup our pseudo HTC control endpoint connection */
1600	memset(&connect, 0, sizeof(connect));
1601	memset(&resp, 0, sizeof(resp));
1602	connect.ep_cb.tx_complete = htc_txctrl_complete;
1603	connect.ep_cb.rx = htc_rxctrl_complete;
1604	connect.max_txq_depth = NUM_CONTROL_TX_BUFFERS;
1605	connect.svc_id = HTC_CTRL_RSVD_SVC;
1606
1607	/* connect fake service */
1608	status = ath6kl_htc_pipe_conn_service(target, &connect, &resp);
1609
1610	return status;
1611}
1612
1613static void ath6kl_htc_pipe_flush_txep(struct htc_target *target,
1614				       enum htc_endpoint_id endpoint, u16 tag)
1615{
1616	struct htc_endpoint *ep = &target->endpoint[endpoint];
1617
1618	if (ep->svc_id == 0) {
1619		WARN_ON_ONCE(1);
1620		/* not in use.. */
1621		return;
1622	}
1623
1624	htc_flush_tx_endpoint(target, ep, tag);
1625}
1626
1627static int ath6kl_htc_pipe_add_rxbuf_multiple(struct htc_target *target,
1628					      struct list_head *pkt_queue)
1629{
1630	struct htc_packet *packet, *tmp_pkt, *first;
1631	struct htc_endpoint *ep;
1632	int status = 0;
1633
1634	if (list_empty(pkt_queue))
1635		return -EINVAL;
1636
1637	first = list_first_entry(pkt_queue, struct htc_packet, list);
1638
1639	if (first->endpoint >= ENDPOINT_MAX) {
1640		WARN_ON_ONCE(1);
1641		return -EINVAL;
1642	}
1643
1644	ath6kl_dbg(ATH6KL_DBG_HTC, "%s: epid: %d, cnt:%d, len: %d\n",
1645		   __func__, first->endpoint, get_queue_depth(pkt_queue),
1646		   first->buf_len);
1647
1648	ep = &target->endpoint[first->endpoint];
1649
1650	spin_lock_bh(&target->rx_lock);
1651
1652	/* store receive packets */
1653	list_splice_tail_init(pkt_queue, &ep->rx_bufq);
1654
1655	spin_unlock_bh(&target->rx_lock);
1656
1657	if (status != 0) {
1658		/* walk through queue and mark each one canceled */
1659		list_for_each_entry_safe(packet, tmp_pkt, pkt_queue, list) {
1660			packet->status = -ECANCELED;
1661		}
1662
1663		do_recv_completion(ep, pkt_queue);
1664	}
1665
1666	return status;
1667}
1668
1669static void ath6kl_htc_pipe_activity_changed(struct htc_target *target,
1670					     enum htc_endpoint_id ep,
1671					     bool active)
1672{
1673	/* TODO */
1674}
1675
1676static void ath6kl_htc_pipe_flush_rx_buf(struct htc_target *target)
1677{
1678	/* TODO */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1679}
1680
1681static int ath6kl_htc_pipe_credit_setup(struct htc_target *target,
1682					struct ath6kl_htc_credit_info *info)
1683{
1684	return 0;
1685}
1686
1687static const struct ath6kl_htc_ops ath6kl_htc_pipe_ops = {
1688	.create = ath6kl_htc_pipe_create,
1689	.wait_target = ath6kl_htc_pipe_wait_target,
1690	.start = ath6kl_htc_pipe_start,
1691	.conn_service = ath6kl_htc_pipe_conn_service,
1692	.tx = ath6kl_htc_pipe_tx,
1693	.stop = ath6kl_htc_pipe_stop,
1694	.cleanup = ath6kl_htc_pipe_cleanup,
1695	.flush_txep = ath6kl_htc_pipe_flush_txep,
1696	.flush_rx_buf = ath6kl_htc_pipe_flush_rx_buf,
1697	.activity_changed = ath6kl_htc_pipe_activity_changed,
1698	.get_rxbuf_num = ath6kl_htc_pipe_get_rxbuf_num,
1699	.add_rxbuf_multiple = ath6kl_htc_pipe_add_rxbuf_multiple,
1700	.credit_setup = ath6kl_htc_pipe_credit_setup,
1701	.tx_complete = ath6kl_htc_pipe_tx_complete,
1702	.rx_complete = ath6kl_htc_pipe_rx_complete,
1703};
1704
1705void ath6kl_htc_pipe_attach(struct ath6kl *ar)
1706{
1707	ar->htc_ops = &ath6kl_htc_pipe_ops;
1708}