Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
   1/*
   2 * Copyright (c) 2005-2011 Atheros Communications Inc.
   3 * Copyright (c) 2011-2013 Qualcomm Atheros, Inc.
   4 *
   5 * Permission to use, copy, modify, and/or distribute this software for any
   6 * purpose with or without fee is hereby granted, provided that the above
   7 * copyright notice and this permission notice appear in all copies.
   8 *
   9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
  10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
  11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
  12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
  13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
  14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
  15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
  16 */
  17
  18#include "hif.h"
  19#include "pci.h"
  20#include "ce.h"
  21#include "debug.h"
  22
  23/*
  24 * Support for Copy Engine hardware, which is mainly used for
  25 * communication between Host and Target over a PCIe interconnect.
  26 */
  27
  28/*
  29 * A single CopyEngine (CE) comprises two "rings":
  30 *   a source ring
  31 *   a destination ring
  32 *
  33 * Each ring consists of a number of descriptors which specify
  34 * an address, length, and meta-data.
  35 *
  36 * Typically, one side of the PCIe interconnect (Host or Target)
  37 * controls one ring and the other side controls the other ring.
  38 * The source side chooses when to initiate a transfer and it
  39 * chooses what to send (buffer address, length). The destination
  40 * side keeps a supply of "anonymous receive buffers" available and
  41 * it handles incoming data as it arrives (when the destination
  42 * recieves an interrupt).
  43 *
  44 * The sender may send a simple buffer (address/length) or it may
  45 * send a small list of buffers.  When a small list is sent, hardware
  46 * "gathers" these and they end up in a single destination buffer
  47 * with a single interrupt.
  48 *
  49 * There are several "contexts" managed by this layer -- more, it
  50 * may seem -- than should be needed. These are provided mainly for
  51 * maximum flexibility and especially to facilitate a simpler HIF
  52 * implementation. There are per-CopyEngine recv, send, and watermark
  53 * contexts. These are supplied by the caller when a recv, send,
  54 * or watermark handler is established and they are echoed back to
  55 * the caller when the respective callbacks are invoked. There is
  56 * also a per-transfer context supplied by the caller when a buffer
  57 * (or sendlist) is sent and when a buffer is enqueued for recv.
  58 * These per-transfer contexts are echoed back to the caller when
  59 * the buffer is sent/received.
  60 */
  61
  62static inline void ath10k_ce_dest_ring_write_index_set(struct ath10k *ar,
  63						       u32 ce_ctrl_addr,
  64						       unsigned int n)
  65{
  66	ath10k_pci_write32(ar, ce_ctrl_addr + DST_WR_INDEX_ADDRESS, n);
  67}
  68
  69static inline u32 ath10k_ce_dest_ring_write_index_get(struct ath10k *ar,
  70						      u32 ce_ctrl_addr)
  71{
  72	return ath10k_pci_read32(ar, ce_ctrl_addr + DST_WR_INDEX_ADDRESS);
  73}
  74
  75static inline void ath10k_ce_src_ring_write_index_set(struct ath10k *ar,
  76						      u32 ce_ctrl_addr,
  77						      unsigned int n)
  78{
  79	ath10k_pci_write32(ar, ce_ctrl_addr + SR_WR_INDEX_ADDRESS, n);
  80}
  81
  82static inline u32 ath10k_ce_src_ring_write_index_get(struct ath10k *ar,
  83						     u32 ce_ctrl_addr)
  84{
  85	return ath10k_pci_read32(ar, ce_ctrl_addr + SR_WR_INDEX_ADDRESS);
  86}
  87
  88static inline u32 ath10k_ce_src_ring_read_index_get(struct ath10k *ar,
  89						    u32 ce_ctrl_addr)
  90{
  91	return ath10k_pci_read32(ar, ce_ctrl_addr + CURRENT_SRRI_ADDRESS);
  92}
  93
  94static inline void ath10k_ce_src_ring_base_addr_set(struct ath10k *ar,
  95						    u32 ce_ctrl_addr,
  96						    unsigned int addr)
  97{
  98	ath10k_pci_write32(ar, ce_ctrl_addr + SR_BA_ADDRESS, addr);
  99}
 100
 101static inline void ath10k_ce_src_ring_size_set(struct ath10k *ar,
 102					       u32 ce_ctrl_addr,
 103					       unsigned int n)
 104{
 105	ath10k_pci_write32(ar, ce_ctrl_addr + SR_SIZE_ADDRESS, n);
 106}
 107
 108static inline void ath10k_ce_src_ring_dmax_set(struct ath10k *ar,
 109					       u32 ce_ctrl_addr,
 110					       unsigned int n)
 111{
 112	u32 ctrl1_addr = ath10k_pci_read32((ar),
 113					   (ce_ctrl_addr) + CE_CTRL1_ADDRESS);
 114
 115	ath10k_pci_write32(ar, ce_ctrl_addr + CE_CTRL1_ADDRESS,
 116			   (ctrl1_addr &  ~CE_CTRL1_DMAX_LENGTH_MASK) |
 117			   CE_CTRL1_DMAX_LENGTH_SET(n));
 118}
 119
 120static inline void ath10k_ce_src_ring_byte_swap_set(struct ath10k *ar,
 121						    u32 ce_ctrl_addr,
 122						    unsigned int n)
 123{
 124	u32 ctrl1_addr = ath10k_pci_read32(ar, ce_ctrl_addr + CE_CTRL1_ADDRESS);
 125
 126	ath10k_pci_write32(ar, ce_ctrl_addr + CE_CTRL1_ADDRESS,
 127			   (ctrl1_addr & ~CE_CTRL1_SRC_RING_BYTE_SWAP_EN_MASK) |
 128			   CE_CTRL1_SRC_RING_BYTE_SWAP_EN_SET(n));
 129}
 130
 131static inline void ath10k_ce_dest_ring_byte_swap_set(struct ath10k *ar,
 132						     u32 ce_ctrl_addr,
 133						     unsigned int n)
 134{
 135	u32 ctrl1_addr = ath10k_pci_read32(ar, ce_ctrl_addr + CE_CTRL1_ADDRESS);
 136
 137	ath10k_pci_write32(ar, ce_ctrl_addr + CE_CTRL1_ADDRESS,
 138			   (ctrl1_addr & ~CE_CTRL1_DST_RING_BYTE_SWAP_EN_MASK) |
 139			   CE_CTRL1_DST_RING_BYTE_SWAP_EN_SET(n));
 140}
 141
 142static inline u32 ath10k_ce_dest_ring_read_index_get(struct ath10k *ar,
 143						     u32 ce_ctrl_addr)
 144{
 145	return ath10k_pci_read32(ar, ce_ctrl_addr + CURRENT_DRRI_ADDRESS);
 146}
 147
 148static inline void ath10k_ce_dest_ring_base_addr_set(struct ath10k *ar,
 149						     u32 ce_ctrl_addr,
 150						     u32 addr)
 151{
 152	ath10k_pci_write32(ar, ce_ctrl_addr + DR_BA_ADDRESS, addr);
 153}
 154
 155static inline void ath10k_ce_dest_ring_size_set(struct ath10k *ar,
 156						u32 ce_ctrl_addr,
 157						unsigned int n)
 158{
 159	ath10k_pci_write32(ar, ce_ctrl_addr + DR_SIZE_ADDRESS, n);
 160}
 161
 162static inline void ath10k_ce_src_ring_highmark_set(struct ath10k *ar,
 163						   u32 ce_ctrl_addr,
 164						   unsigned int n)
 165{
 166	u32 addr = ath10k_pci_read32(ar, ce_ctrl_addr + SRC_WATERMARK_ADDRESS);
 167
 168	ath10k_pci_write32(ar, ce_ctrl_addr + SRC_WATERMARK_ADDRESS,
 169			   (addr & ~SRC_WATERMARK_HIGH_MASK) |
 170			   SRC_WATERMARK_HIGH_SET(n));
 171}
 172
 173static inline void ath10k_ce_src_ring_lowmark_set(struct ath10k *ar,
 174						  u32 ce_ctrl_addr,
 175						  unsigned int n)
 176{
 177	u32 addr = ath10k_pci_read32(ar, ce_ctrl_addr + SRC_WATERMARK_ADDRESS);
 178
 179	ath10k_pci_write32(ar, ce_ctrl_addr + SRC_WATERMARK_ADDRESS,
 180			   (addr & ~SRC_WATERMARK_LOW_MASK) |
 181			   SRC_WATERMARK_LOW_SET(n));
 182}
 183
 184static inline void ath10k_ce_dest_ring_highmark_set(struct ath10k *ar,
 185						    u32 ce_ctrl_addr,
 186						    unsigned int n)
 187{
 188	u32 addr = ath10k_pci_read32(ar, ce_ctrl_addr + DST_WATERMARK_ADDRESS);
 189
 190	ath10k_pci_write32(ar, ce_ctrl_addr + DST_WATERMARK_ADDRESS,
 191			   (addr & ~DST_WATERMARK_HIGH_MASK) |
 192			   DST_WATERMARK_HIGH_SET(n));
 193}
 194
 195static inline void ath10k_ce_dest_ring_lowmark_set(struct ath10k *ar,
 196						   u32 ce_ctrl_addr,
 197						   unsigned int n)
 198{
 199	u32 addr = ath10k_pci_read32(ar, ce_ctrl_addr + DST_WATERMARK_ADDRESS);
 200
 201	ath10k_pci_write32(ar, ce_ctrl_addr + DST_WATERMARK_ADDRESS,
 202			   (addr & ~DST_WATERMARK_LOW_MASK) |
 203			   DST_WATERMARK_LOW_SET(n));
 204}
 205
 206static inline void ath10k_ce_copy_complete_inter_enable(struct ath10k *ar,
 207							u32 ce_ctrl_addr)
 208{
 209	u32 host_ie_addr = ath10k_pci_read32(ar,
 210					     ce_ctrl_addr + HOST_IE_ADDRESS);
 211
 212	ath10k_pci_write32(ar, ce_ctrl_addr + HOST_IE_ADDRESS,
 213			   host_ie_addr | HOST_IE_COPY_COMPLETE_MASK);
 214}
 215
 216static inline void ath10k_ce_copy_complete_intr_disable(struct ath10k *ar,
 217							u32 ce_ctrl_addr)
 218{
 219	u32 host_ie_addr = ath10k_pci_read32(ar,
 220					     ce_ctrl_addr + HOST_IE_ADDRESS);
 221
 222	ath10k_pci_write32(ar, ce_ctrl_addr + HOST_IE_ADDRESS,
 223			   host_ie_addr & ~HOST_IE_COPY_COMPLETE_MASK);
 224}
 225
 226static inline void ath10k_ce_watermark_intr_disable(struct ath10k *ar,
 227						    u32 ce_ctrl_addr)
 228{
 229	u32 host_ie_addr = ath10k_pci_read32(ar,
 230					     ce_ctrl_addr + HOST_IE_ADDRESS);
 231
 232	ath10k_pci_write32(ar, ce_ctrl_addr + HOST_IE_ADDRESS,
 233			   host_ie_addr & ~CE_WATERMARK_MASK);
 234}
 235
 236static inline void ath10k_ce_error_intr_enable(struct ath10k *ar,
 237					       u32 ce_ctrl_addr)
 238{
 239	u32 misc_ie_addr = ath10k_pci_read32(ar,
 240					     ce_ctrl_addr + MISC_IE_ADDRESS);
 241
 242	ath10k_pci_write32(ar, ce_ctrl_addr + MISC_IE_ADDRESS,
 243			   misc_ie_addr | CE_ERROR_MASK);
 244}
 245
 246static inline void ath10k_ce_error_intr_disable(struct ath10k *ar,
 247						u32 ce_ctrl_addr)
 248{
 249	u32 misc_ie_addr = ath10k_pci_read32(ar,
 250					     ce_ctrl_addr + MISC_IE_ADDRESS);
 251
 252	ath10k_pci_write32(ar, ce_ctrl_addr + MISC_IE_ADDRESS,
 253			   misc_ie_addr & ~CE_ERROR_MASK);
 254}
 255
 256static inline void ath10k_ce_engine_int_status_clear(struct ath10k *ar,
 257						     u32 ce_ctrl_addr,
 258						     unsigned int mask)
 259{
 260	ath10k_pci_write32(ar, ce_ctrl_addr + HOST_IS_ADDRESS, mask);
 261}
 262
 263
 264/*
 265 * Guts of ath10k_ce_send, used by both ath10k_ce_send and
 266 * ath10k_ce_sendlist_send.
 267 * The caller takes responsibility for any needed locking.
 268 */
 269int ath10k_ce_send_nolock(struct ath10k_ce_pipe *ce_state,
 270			  void *per_transfer_context,
 271			  u32 buffer,
 272			  unsigned int nbytes,
 273			  unsigned int transfer_id,
 274			  unsigned int flags)
 275{
 276	struct ath10k *ar = ce_state->ar;
 277	struct ath10k_ce_ring *src_ring = ce_state->src_ring;
 278	struct ce_desc *desc, *sdesc;
 279	unsigned int nentries_mask = src_ring->nentries_mask;
 280	unsigned int sw_index = src_ring->sw_index;
 281	unsigned int write_index = src_ring->write_index;
 282	u32 ctrl_addr = ce_state->ctrl_addr;
 283	u32 desc_flags = 0;
 284	int ret = 0;
 285
 286	if (nbytes > ce_state->src_sz_max)
 287		ath10k_warn("%s: send more we can (nbytes: %d, max: %d)\n",
 288			    __func__, nbytes, ce_state->src_sz_max);
 289
 290	ret = ath10k_pci_wake(ar);
 291	if (ret)
 292		return ret;
 293
 294	if (unlikely(CE_RING_DELTA(nentries_mask,
 295				   write_index, sw_index - 1) <= 0)) {
 296		ret = -ENOSR;
 297		goto exit;
 298	}
 299
 300	desc = CE_SRC_RING_TO_DESC(src_ring->base_addr_owner_space,
 301				   write_index);
 302	sdesc = CE_SRC_RING_TO_DESC(src_ring->shadow_base, write_index);
 303
 304	desc_flags |= SM(transfer_id, CE_DESC_FLAGS_META_DATA);
 305
 306	if (flags & CE_SEND_FLAG_GATHER)
 307		desc_flags |= CE_DESC_FLAGS_GATHER;
 308	if (flags & CE_SEND_FLAG_BYTE_SWAP)
 309		desc_flags |= CE_DESC_FLAGS_BYTE_SWAP;
 310
 311	sdesc->addr   = __cpu_to_le32(buffer);
 312	sdesc->nbytes = __cpu_to_le16(nbytes);
 313	sdesc->flags  = __cpu_to_le16(desc_flags);
 314
 315	*desc = *sdesc;
 316
 317	src_ring->per_transfer_context[write_index] = per_transfer_context;
 318
 319	/* Update Source Ring Write Index */
 320	write_index = CE_RING_IDX_INCR(nentries_mask, write_index);
 321
 322	/* WORKAROUND */
 323	if (!(flags & CE_SEND_FLAG_GATHER))
 324		ath10k_ce_src_ring_write_index_set(ar, ctrl_addr, write_index);
 325
 326	src_ring->write_index = write_index;
 327exit:
 328	ath10k_pci_sleep(ar);
 329	return ret;
 330}
 331
 332int ath10k_ce_send(struct ath10k_ce_pipe *ce_state,
 333		   void *per_transfer_context,
 334		   u32 buffer,
 335		   unsigned int nbytes,
 336		   unsigned int transfer_id,
 337		   unsigned int flags)
 338{
 339	struct ath10k *ar = ce_state->ar;
 340	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 341	int ret;
 342
 343	spin_lock_bh(&ar_pci->ce_lock);
 344	ret = ath10k_ce_send_nolock(ce_state, per_transfer_context,
 345				    buffer, nbytes, transfer_id, flags);
 346	spin_unlock_bh(&ar_pci->ce_lock);
 347
 348	return ret;
 349}
 350
 351int ath10k_ce_num_free_src_entries(struct ath10k_ce_pipe *pipe)
 352{
 353	struct ath10k *ar = pipe->ar;
 354	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 355	int delta;
 356
 357	spin_lock_bh(&ar_pci->ce_lock);
 358	delta = CE_RING_DELTA(pipe->src_ring->nentries_mask,
 359			      pipe->src_ring->write_index,
 360			      pipe->src_ring->sw_index - 1);
 361	spin_unlock_bh(&ar_pci->ce_lock);
 362
 363	return delta;
 364}
 365
 366int ath10k_ce_recv_buf_enqueue(struct ath10k_ce_pipe *ce_state,
 367			       void *per_recv_context,
 368			       u32 buffer)
 369{
 370	struct ath10k_ce_ring *dest_ring = ce_state->dest_ring;
 371	u32 ctrl_addr = ce_state->ctrl_addr;
 372	struct ath10k *ar = ce_state->ar;
 373	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 374	unsigned int nentries_mask = dest_ring->nentries_mask;
 375	unsigned int write_index;
 376	unsigned int sw_index;
 377	int ret;
 378
 379	spin_lock_bh(&ar_pci->ce_lock);
 380	write_index = dest_ring->write_index;
 381	sw_index = dest_ring->sw_index;
 382
 383	ret = ath10k_pci_wake(ar);
 384	if (ret)
 385		goto out;
 386
 387	if (CE_RING_DELTA(nentries_mask, write_index, sw_index - 1) > 0) {
 388		struct ce_desc *base = dest_ring->base_addr_owner_space;
 389		struct ce_desc *desc = CE_DEST_RING_TO_DESC(base, write_index);
 390
 391		/* Update destination descriptor */
 392		desc->addr    = __cpu_to_le32(buffer);
 393		desc->nbytes = 0;
 394
 395		dest_ring->per_transfer_context[write_index] =
 396							per_recv_context;
 397
 398		/* Update Destination Ring Write Index */
 399		write_index = CE_RING_IDX_INCR(nentries_mask, write_index);
 400		ath10k_ce_dest_ring_write_index_set(ar, ctrl_addr, write_index);
 401		dest_ring->write_index = write_index;
 402		ret = 0;
 403	} else {
 404		ret = -EIO;
 405	}
 406	ath10k_pci_sleep(ar);
 407
 408out:
 409	spin_unlock_bh(&ar_pci->ce_lock);
 410
 411	return ret;
 412}
 413
 414/*
 415 * Guts of ath10k_ce_completed_recv_next.
 416 * The caller takes responsibility for any necessary locking.
 417 */
 418static int ath10k_ce_completed_recv_next_nolock(struct ath10k_ce_pipe *ce_state,
 419						void **per_transfer_contextp,
 420						u32 *bufferp,
 421						unsigned int *nbytesp,
 422						unsigned int *transfer_idp,
 423						unsigned int *flagsp)
 424{
 425	struct ath10k_ce_ring *dest_ring = ce_state->dest_ring;
 426	unsigned int nentries_mask = dest_ring->nentries_mask;
 427	unsigned int sw_index = dest_ring->sw_index;
 428
 429	struct ce_desc *base = dest_ring->base_addr_owner_space;
 430	struct ce_desc *desc = CE_DEST_RING_TO_DESC(base, sw_index);
 431	struct ce_desc sdesc;
 432	u16 nbytes;
 433
 434	/* Copy in one go for performance reasons */
 435	sdesc = *desc;
 436
 437	nbytes = __le16_to_cpu(sdesc.nbytes);
 438	if (nbytes == 0) {
 439		/*
 440		 * This closes a relatively unusual race where the Host
 441		 * sees the updated DRRI before the update to the
 442		 * corresponding descriptor has completed. We treat this
 443		 * as a descriptor that is not yet done.
 444		 */
 445		return -EIO;
 446	}
 447
 448	desc->nbytes = 0;
 449
 450	/* Return data from completed destination descriptor */
 451	*bufferp = __le32_to_cpu(sdesc.addr);
 452	*nbytesp = nbytes;
 453	*transfer_idp = MS(__le16_to_cpu(sdesc.flags), CE_DESC_FLAGS_META_DATA);
 454
 455	if (__le16_to_cpu(sdesc.flags) & CE_DESC_FLAGS_BYTE_SWAP)
 456		*flagsp = CE_RECV_FLAG_SWAPPED;
 457	else
 458		*flagsp = 0;
 459
 460	if (per_transfer_contextp)
 461		*per_transfer_contextp =
 462			dest_ring->per_transfer_context[sw_index];
 463
 464	/* sanity */
 465	dest_ring->per_transfer_context[sw_index] = NULL;
 466
 467	/* Update sw_index */
 468	sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
 469	dest_ring->sw_index = sw_index;
 470
 471	return 0;
 472}
 473
 474int ath10k_ce_completed_recv_next(struct ath10k_ce_pipe *ce_state,
 475				  void **per_transfer_contextp,
 476				  u32 *bufferp,
 477				  unsigned int *nbytesp,
 478				  unsigned int *transfer_idp,
 479				  unsigned int *flagsp)
 480{
 481	struct ath10k *ar = ce_state->ar;
 482	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 483	int ret;
 484
 485	spin_lock_bh(&ar_pci->ce_lock);
 486	ret = ath10k_ce_completed_recv_next_nolock(ce_state,
 487						   per_transfer_contextp,
 488						   bufferp, nbytesp,
 489						   transfer_idp, flagsp);
 490	spin_unlock_bh(&ar_pci->ce_lock);
 491
 492	return ret;
 493}
 494
 495int ath10k_ce_revoke_recv_next(struct ath10k_ce_pipe *ce_state,
 496			       void **per_transfer_contextp,
 497			       u32 *bufferp)
 498{
 499	struct ath10k_ce_ring *dest_ring;
 500	unsigned int nentries_mask;
 501	unsigned int sw_index;
 502	unsigned int write_index;
 503	int ret;
 504	struct ath10k *ar;
 505	struct ath10k_pci *ar_pci;
 506
 507	dest_ring = ce_state->dest_ring;
 508
 509	if (!dest_ring)
 510		return -EIO;
 511
 512	ar = ce_state->ar;
 513	ar_pci = ath10k_pci_priv(ar);
 514
 515	spin_lock_bh(&ar_pci->ce_lock);
 516
 517	nentries_mask = dest_ring->nentries_mask;
 518	sw_index = dest_ring->sw_index;
 519	write_index = dest_ring->write_index;
 520	if (write_index != sw_index) {
 521		struct ce_desc *base = dest_ring->base_addr_owner_space;
 522		struct ce_desc *desc = CE_DEST_RING_TO_DESC(base, sw_index);
 523
 524		/* Return data from completed destination descriptor */
 525		*bufferp = __le32_to_cpu(desc->addr);
 526
 527		if (per_transfer_contextp)
 528			*per_transfer_contextp =
 529				dest_ring->per_transfer_context[sw_index];
 530
 531		/* sanity */
 532		dest_ring->per_transfer_context[sw_index] = NULL;
 533
 534		/* Update sw_index */
 535		sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
 536		dest_ring->sw_index = sw_index;
 537		ret = 0;
 538	} else {
 539		ret = -EIO;
 540	}
 541
 542	spin_unlock_bh(&ar_pci->ce_lock);
 543
 544	return ret;
 545}
 546
 547/*
 548 * Guts of ath10k_ce_completed_send_next.
 549 * The caller takes responsibility for any necessary locking.
 550 */
 551static int ath10k_ce_completed_send_next_nolock(struct ath10k_ce_pipe *ce_state,
 552						void **per_transfer_contextp,
 553						u32 *bufferp,
 554						unsigned int *nbytesp,
 555						unsigned int *transfer_idp)
 556{
 557	struct ath10k_ce_ring *src_ring = ce_state->src_ring;
 558	u32 ctrl_addr = ce_state->ctrl_addr;
 559	struct ath10k *ar = ce_state->ar;
 560	unsigned int nentries_mask = src_ring->nentries_mask;
 561	unsigned int sw_index = src_ring->sw_index;
 562	struct ce_desc *sdesc, *sbase;
 563	unsigned int read_index;
 564	int ret;
 565
 566	if (src_ring->hw_index == sw_index) {
 567		/*
 568		 * The SW completion index has caught up with the cached
 569		 * version of the HW completion index.
 570		 * Update the cached HW completion index to see whether
 571		 * the SW has really caught up to the HW, or if the cached
 572		 * value of the HW index has become stale.
 573		 */
 574
 575		ret = ath10k_pci_wake(ar);
 576		if (ret)
 577			return ret;
 578
 579		src_ring->hw_index =
 580			ath10k_ce_src_ring_read_index_get(ar, ctrl_addr);
 581		src_ring->hw_index &= nentries_mask;
 582
 583		ath10k_pci_sleep(ar);
 584	}
 585
 586	read_index = src_ring->hw_index;
 587
 588	if ((read_index == sw_index) || (read_index == 0xffffffff))
 589		return -EIO;
 590
 591	sbase = src_ring->shadow_base;
 592	sdesc = CE_SRC_RING_TO_DESC(sbase, sw_index);
 593
 594	/* Return data from completed source descriptor */
 595	*bufferp = __le32_to_cpu(sdesc->addr);
 596	*nbytesp = __le16_to_cpu(sdesc->nbytes);
 597	*transfer_idp = MS(__le16_to_cpu(sdesc->flags),
 598			   CE_DESC_FLAGS_META_DATA);
 599
 600	if (per_transfer_contextp)
 601		*per_transfer_contextp =
 602			src_ring->per_transfer_context[sw_index];
 603
 604	/* sanity */
 605	src_ring->per_transfer_context[sw_index] = NULL;
 606
 607	/* Update sw_index */
 608	sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
 609	src_ring->sw_index = sw_index;
 610
 611	return 0;
 612}
 613
 614/* NB: Modeled after ath10k_ce_completed_send_next */
 615int ath10k_ce_cancel_send_next(struct ath10k_ce_pipe *ce_state,
 616			       void **per_transfer_contextp,
 617			       u32 *bufferp,
 618			       unsigned int *nbytesp,
 619			       unsigned int *transfer_idp)
 620{
 621	struct ath10k_ce_ring *src_ring;
 622	unsigned int nentries_mask;
 623	unsigned int sw_index;
 624	unsigned int write_index;
 625	int ret;
 626	struct ath10k *ar;
 627	struct ath10k_pci *ar_pci;
 628
 629	src_ring = ce_state->src_ring;
 630
 631	if (!src_ring)
 632		return -EIO;
 633
 634	ar = ce_state->ar;
 635	ar_pci = ath10k_pci_priv(ar);
 636
 637	spin_lock_bh(&ar_pci->ce_lock);
 638
 639	nentries_mask = src_ring->nentries_mask;
 640	sw_index = src_ring->sw_index;
 641	write_index = src_ring->write_index;
 642
 643	if (write_index != sw_index) {
 644		struct ce_desc *base = src_ring->base_addr_owner_space;
 645		struct ce_desc *desc = CE_SRC_RING_TO_DESC(base, sw_index);
 646
 647		/* Return data from completed source descriptor */
 648		*bufferp = __le32_to_cpu(desc->addr);
 649		*nbytesp = __le16_to_cpu(desc->nbytes);
 650		*transfer_idp = MS(__le16_to_cpu(desc->flags),
 651						CE_DESC_FLAGS_META_DATA);
 652
 653		if (per_transfer_contextp)
 654			*per_transfer_contextp =
 655				src_ring->per_transfer_context[sw_index];
 656
 657		/* sanity */
 658		src_ring->per_transfer_context[sw_index] = NULL;
 659
 660		/* Update sw_index */
 661		sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
 662		src_ring->sw_index = sw_index;
 663		ret = 0;
 664	} else {
 665		ret = -EIO;
 666	}
 667
 668	spin_unlock_bh(&ar_pci->ce_lock);
 669
 670	return ret;
 671}
 672
 673int ath10k_ce_completed_send_next(struct ath10k_ce_pipe *ce_state,
 674				  void **per_transfer_contextp,
 675				  u32 *bufferp,
 676				  unsigned int *nbytesp,
 677				  unsigned int *transfer_idp)
 678{
 679	struct ath10k *ar = ce_state->ar;
 680	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 681	int ret;
 682
 683	spin_lock_bh(&ar_pci->ce_lock);
 684	ret = ath10k_ce_completed_send_next_nolock(ce_state,
 685						   per_transfer_contextp,
 686						   bufferp, nbytesp,
 687						   transfer_idp);
 688	spin_unlock_bh(&ar_pci->ce_lock);
 689
 690	return ret;
 691}
 692
 693/*
 694 * Guts of interrupt handler for per-engine interrupts on a particular CE.
 695 *
 696 * Invokes registered callbacks for recv_complete,
 697 * send_complete, and watermarks.
 698 */
 699void ath10k_ce_per_engine_service(struct ath10k *ar, unsigned int ce_id)
 700{
 701	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 702	struct ath10k_ce_pipe *ce_state = &ar_pci->ce_states[ce_id];
 703	u32 ctrl_addr = ce_state->ctrl_addr;
 704	int ret;
 705
 706	ret = ath10k_pci_wake(ar);
 707	if (ret)
 708		return;
 709
 710	spin_lock_bh(&ar_pci->ce_lock);
 711
 712	/* Clear the copy-complete interrupts that will be handled here. */
 713	ath10k_ce_engine_int_status_clear(ar, ctrl_addr,
 714					  HOST_IS_COPY_COMPLETE_MASK);
 715
 716	spin_unlock_bh(&ar_pci->ce_lock);
 717
 718	if (ce_state->recv_cb)
 719		ce_state->recv_cb(ce_state);
 720
 721	if (ce_state->send_cb)
 722		ce_state->send_cb(ce_state);
 723
 724	spin_lock_bh(&ar_pci->ce_lock);
 725
 726	/*
 727	 * Misc CE interrupts are not being handled, but still need
 728	 * to be cleared.
 729	 */
 730	ath10k_ce_engine_int_status_clear(ar, ctrl_addr, CE_WATERMARK_MASK);
 731
 732	spin_unlock_bh(&ar_pci->ce_lock);
 733	ath10k_pci_sleep(ar);
 734}
 735
 736/*
 737 * Handler for per-engine interrupts on ALL active CEs.
 738 * This is used in cases where the system is sharing a
 739 * single interrput for all CEs
 740 */
 741
 742void ath10k_ce_per_engine_service_any(struct ath10k *ar)
 743{
 744	int ce_id, ret;
 745	u32 intr_summary;
 746
 747	ret = ath10k_pci_wake(ar);
 748	if (ret)
 749		return;
 750
 751	intr_summary = CE_INTERRUPT_SUMMARY(ar);
 752
 753	for (ce_id = 0; intr_summary && (ce_id < CE_COUNT); ce_id++) {
 754		if (intr_summary & (1 << ce_id))
 755			intr_summary &= ~(1 << ce_id);
 756		else
 757			/* no intr pending on this CE */
 758			continue;
 759
 760		ath10k_ce_per_engine_service(ar, ce_id);
 761	}
 762
 763	ath10k_pci_sleep(ar);
 764}
 765
 766/*
 767 * Adjust interrupts for the copy complete handler.
 768 * If it's needed for either send or recv, then unmask
 769 * this interrupt; otherwise, mask it.
 770 *
 771 * Called with ce_lock held.
 772 */
 773static void ath10k_ce_per_engine_handler_adjust(struct ath10k_ce_pipe *ce_state,
 774						int disable_copy_compl_intr)
 775{
 776	u32 ctrl_addr = ce_state->ctrl_addr;
 777	struct ath10k *ar = ce_state->ar;
 778	int ret;
 779
 780	ret = ath10k_pci_wake(ar);
 781	if (ret)
 782		return;
 783
 784	if ((!disable_copy_compl_intr) &&
 785	    (ce_state->send_cb || ce_state->recv_cb))
 786		ath10k_ce_copy_complete_inter_enable(ar, ctrl_addr);
 787	else
 788		ath10k_ce_copy_complete_intr_disable(ar, ctrl_addr);
 789
 790	ath10k_ce_watermark_intr_disable(ar, ctrl_addr);
 791
 792	ath10k_pci_sleep(ar);
 793}
 794
 795int ath10k_ce_disable_interrupts(struct ath10k *ar)
 796{
 797	int ce_id, ret;
 798
 799	ret = ath10k_pci_wake(ar);
 800	if (ret)
 801		return ret;
 802
 803	for (ce_id = 0; ce_id < CE_COUNT; ce_id++) {
 804		u32 ctrl_addr = ath10k_ce_base_address(ce_id);
 805
 806		ath10k_ce_copy_complete_intr_disable(ar, ctrl_addr);
 807		ath10k_ce_error_intr_disable(ar, ctrl_addr);
 808		ath10k_ce_watermark_intr_disable(ar, ctrl_addr);
 809	}
 810
 811	ath10k_pci_sleep(ar);
 812
 813	return 0;
 814}
 815
 816void ath10k_ce_send_cb_register(struct ath10k_ce_pipe *ce_state,
 817				void (*send_cb)(struct ath10k_ce_pipe *),
 818				int disable_interrupts)
 819{
 820	struct ath10k *ar = ce_state->ar;
 821	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 822
 823	spin_lock_bh(&ar_pci->ce_lock);
 824	ce_state->send_cb = send_cb;
 825	ath10k_ce_per_engine_handler_adjust(ce_state, disable_interrupts);
 826	spin_unlock_bh(&ar_pci->ce_lock);
 827}
 828
 829void ath10k_ce_recv_cb_register(struct ath10k_ce_pipe *ce_state,
 830				void (*recv_cb)(struct ath10k_ce_pipe *))
 831{
 832	struct ath10k *ar = ce_state->ar;
 833	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 834
 835	spin_lock_bh(&ar_pci->ce_lock);
 836	ce_state->recv_cb = recv_cb;
 837	ath10k_ce_per_engine_handler_adjust(ce_state, 0);
 838	spin_unlock_bh(&ar_pci->ce_lock);
 839}
 840
 841static int ath10k_ce_init_src_ring(struct ath10k *ar,
 842				   unsigned int ce_id,
 843				   struct ath10k_ce_pipe *ce_state,
 844				   const struct ce_attr *attr)
 845{
 846	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 847	struct ath10k_ce_ring *src_ring;
 848	unsigned int nentries = attr->src_nentries;
 849	unsigned int ce_nbytes;
 850	u32 ctrl_addr = ath10k_ce_base_address(ce_id);
 851	dma_addr_t base_addr;
 852	char *ptr;
 853
 854	nentries = roundup_pow_of_two(nentries);
 855
 856	if (ce_state->src_ring) {
 857		WARN_ON(ce_state->src_ring->nentries != nentries);
 858		return 0;
 859	}
 860
 861	ce_nbytes = sizeof(struct ath10k_ce_ring) + (nentries * sizeof(void *));
 862	ptr = kzalloc(ce_nbytes, GFP_KERNEL);
 863	if (ptr == NULL)
 864		return -ENOMEM;
 865
 866	ce_state->src_ring = (struct ath10k_ce_ring *)ptr;
 867	src_ring = ce_state->src_ring;
 868
 869	ptr += sizeof(struct ath10k_ce_ring);
 870	src_ring->nentries = nentries;
 871	src_ring->nentries_mask = nentries - 1;
 872
 873	src_ring->sw_index = ath10k_ce_src_ring_read_index_get(ar, ctrl_addr);
 874	src_ring->sw_index &= src_ring->nentries_mask;
 875	src_ring->hw_index = src_ring->sw_index;
 876
 877	src_ring->write_index =
 878		ath10k_ce_src_ring_write_index_get(ar, ctrl_addr);
 879	src_ring->write_index &= src_ring->nentries_mask;
 880
 881	src_ring->per_transfer_context = (void **)ptr;
 882
 883	/*
 884	 * Legacy platforms that do not support cache
 885	 * coherent DMA are unsupported
 886	 */
 887	src_ring->base_addr_owner_space_unaligned =
 888		pci_alloc_consistent(ar_pci->pdev,
 889				     (nentries * sizeof(struct ce_desc) +
 890				      CE_DESC_RING_ALIGN),
 891				     &base_addr);
 892	if (!src_ring->base_addr_owner_space_unaligned) {
 893		kfree(ce_state->src_ring);
 894		ce_state->src_ring = NULL;
 895		return -ENOMEM;
 896	}
 897
 898	src_ring->base_addr_ce_space_unaligned = base_addr;
 899
 900	src_ring->base_addr_owner_space = PTR_ALIGN(
 901			src_ring->base_addr_owner_space_unaligned,
 902			CE_DESC_RING_ALIGN);
 903	src_ring->base_addr_ce_space = ALIGN(
 904			src_ring->base_addr_ce_space_unaligned,
 905			CE_DESC_RING_ALIGN);
 906
 907	/*
 908	 * Also allocate a shadow src ring in regular
 909	 * mem to use for faster access.
 910	 */
 911	src_ring->shadow_base_unaligned =
 912		kmalloc((nentries * sizeof(struct ce_desc) +
 913			 CE_DESC_RING_ALIGN), GFP_KERNEL);
 914	if (!src_ring->shadow_base_unaligned) {
 915		pci_free_consistent(ar_pci->pdev,
 916				    (nentries * sizeof(struct ce_desc) +
 917				     CE_DESC_RING_ALIGN),
 918				    src_ring->base_addr_owner_space,
 919				    src_ring->base_addr_ce_space);
 920		kfree(ce_state->src_ring);
 921		ce_state->src_ring = NULL;
 922		return -ENOMEM;
 923	}
 924
 925	src_ring->shadow_base = PTR_ALIGN(
 926			src_ring->shadow_base_unaligned,
 927			CE_DESC_RING_ALIGN);
 928
 929	ath10k_ce_src_ring_base_addr_set(ar, ctrl_addr,
 930					 src_ring->base_addr_ce_space);
 931	ath10k_ce_src_ring_size_set(ar, ctrl_addr, nentries);
 932	ath10k_ce_src_ring_dmax_set(ar, ctrl_addr, attr->src_sz_max);
 933	ath10k_ce_src_ring_byte_swap_set(ar, ctrl_addr, 0);
 934	ath10k_ce_src_ring_lowmark_set(ar, ctrl_addr, 0);
 935	ath10k_ce_src_ring_highmark_set(ar, ctrl_addr, nentries);
 936
 937	ath10k_dbg(ATH10K_DBG_BOOT,
 938		   "boot ce src ring id %d entries %d base_addr %p\n",
 939		   ce_id, nentries, src_ring->base_addr_owner_space);
 940
 941	return 0;
 942}
 943
 944static int ath10k_ce_init_dest_ring(struct ath10k *ar,
 945				    unsigned int ce_id,
 946				    struct ath10k_ce_pipe *ce_state,
 947				    const struct ce_attr *attr)
 948{
 949	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
 950	struct ath10k_ce_ring *dest_ring;
 951	unsigned int nentries = attr->dest_nentries;
 952	unsigned int ce_nbytes;
 953	u32 ctrl_addr = ath10k_ce_base_address(ce_id);
 954	dma_addr_t base_addr;
 955	char *ptr;
 956
 957	nentries = roundup_pow_of_two(nentries);
 958
 959	if (ce_state->dest_ring) {
 960		WARN_ON(ce_state->dest_ring->nentries != nentries);
 961		return 0;
 962	}
 963
 964	ce_nbytes = sizeof(struct ath10k_ce_ring) + (nentries * sizeof(void *));
 965	ptr = kzalloc(ce_nbytes, GFP_KERNEL);
 966	if (ptr == NULL)
 967		return -ENOMEM;
 968
 969	ce_state->dest_ring = (struct ath10k_ce_ring *)ptr;
 970	dest_ring = ce_state->dest_ring;
 971
 972	ptr += sizeof(struct ath10k_ce_ring);
 973	dest_ring->nentries = nentries;
 974	dest_ring->nentries_mask = nentries - 1;
 975
 976	dest_ring->sw_index = ath10k_ce_dest_ring_read_index_get(ar, ctrl_addr);
 977	dest_ring->sw_index &= dest_ring->nentries_mask;
 978	dest_ring->write_index =
 979		ath10k_ce_dest_ring_write_index_get(ar, ctrl_addr);
 980	dest_ring->write_index &= dest_ring->nentries_mask;
 981
 982	dest_ring->per_transfer_context = (void **)ptr;
 983
 984	/*
 985	 * Legacy platforms that do not support cache
 986	 * coherent DMA are unsupported
 987	 */
 988	dest_ring->base_addr_owner_space_unaligned =
 989		pci_alloc_consistent(ar_pci->pdev,
 990				     (nentries * sizeof(struct ce_desc) +
 991				      CE_DESC_RING_ALIGN),
 992				     &base_addr);
 993	if (!dest_ring->base_addr_owner_space_unaligned) {
 994		kfree(ce_state->dest_ring);
 995		ce_state->dest_ring = NULL;
 996		return -ENOMEM;
 997	}
 998
 999	dest_ring->base_addr_ce_space_unaligned = base_addr;
1000
1001	/*
1002	 * Correctly initialize memory to 0 to prevent garbage
1003	 * data crashing system when download firmware
1004	 */
1005	memset(dest_ring->base_addr_owner_space_unaligned, 0,
1006	       nentries * sizeof(struct ce_desc) + CE_DESC_RING_ALIGN);
1007
1008	dest_ring->base_addr_owner_space = PTR_ALIGN(
1009			dest_ring->base_addr_owner_space_unaligned,
1010			CE_DESC_RING_ALIGN);
1011	dest_ring->base_addr_ce_space = ALIGN(
1012			dest_ring->base_addr_ce_space_unaligned,
1013			CE_DESC_RING_ALIGN);
1014
1015	ath10k_ce_dest_ring_base_addr_set(ar, ctrl_addr,
1016					  dest_ring->base_addr_ce_space);
1017	ath10k_ce_dest_ring_size_set(ar, ctrl_addr, nentries);
1018	ath10k_ce_dest_ring_byte_swap_set(ar, ctrl_addr, 0);
1019	ath10k_ce_dest_ring_lowmark_set(ar, ctrl_addr, 0);
1020	ath10k_ce_dest_ring_highmark_set(ar, ctrl_addr, nentries);
1021
1022	ath10k_dbg(ATH10K_DBG_BOOT,
1023		   "boot ce dest ring id %d entries %d base_addr %p\n",
1024		   ce_id, nentries, dest_ring->base_addr_owner_space);
1025
1026	return 0;
1027}
1028
1029static struct ath10k_ce_pipe *ath10k_ce_init_state(struct ath10k *ar,
1030					     unsigned int ce_id,
1031					     const struct ce_attr *attr)
1032{
1033	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1034	struct ath10k_ce_pipe *ce_state = &ar_pci->ce_states[ce_id];
1035	u32 ctrl_addr = ath10k_ce_base_address(ce_id);
1036
1037	spin_lock_bh(&ar_pci->ce_lock);
1038
1039	ce_state->ar = ar;
1040	ce_state->id = ce_id;
1041	ce_state->ctrl_addr = ctrl_addr;
1042	ce_state->attr_flags = attr->flags;
1043	ce_state->src_sz_max = attr->src_sz_max;
1044
1045	spin_unlock_bh(&ar_pci->ce_lock);
1046
1047	return ce_state;
1048}
1049
1050/*
1051 * Initialize a Copy Engine based on caller-supplied attributes.
1052 * This may be called once to initialize both source and destination
1053 * rings or it may be called twice for separate source and destination
1054 * initialization. It may be that only one side or the other is
1055 * initialized by software/firmware.
1056 */
1057struct ath10k_ce_pipe *ath10k_ce_init(struct ath10k *ar,
1058				unsigned int ce_id,
1059				const struct ce_attr *attr)
1060{
1061	struct ath10k_ce_pipe *ce_state;
1062	int ret;
1063
1064	/*
1065	 * Make sure there's enough CE ringbuffer entries for HTT TX to avoid
1066	 * additional TX locking checks.
1067	 *
1068	 * For the lack of a better place do the check here.
1069	 */
1070	BUILD_BUG_ON(2*TARGET_NUM_MSDU_DESC >
1071		     (CE_HTT_H2T_MSG_SRC_NENTRIES - 1));
1072	BUILD_BUG_ON(2*TARGET_10X_NUM_MSDU_DESC >
1073		     (CE_HTT_H2T_MSG_SRC_NENTRIES - 1));
1074
1075	ret = ath10k_pci_wake(ar);
1076	if (ret)
1077		return NULL;
1078
1079	ce_state = ath10k_ce_init_state(ar, ce_id, attr);
1080	if (!ce_state) {
1081		ath10k_err("Failed to initialize CE state for ID: %d\n", ce_id);
1082		goto out;
1083	}
1084
1085	if (attr->src_nentries) {
1086		ret = ath10k_ce_init_src_ring(ar, ce_id, ce_state, attr);
1087		if (ret) {
1088			ath10k_err("Failed to initialize CE src ring for ID: %d (%d)\n",
1089				   ce_id, ret);
1090			ath10k_ce_deinit(ce_state);
1091			ce_state = NULL;
1092			goto out;
1093		}
1094	}
1095
1096	if (attr->dest_nentries) {
1097		ret = ath10k_ce_init_dest_ring(ar, ce_id, ce_state, attr);
1098		if (ret) {
1099			ath10k_err("Failed to initialize CE dest ring for ID: %d (%d)\n",
1100				   ce_id, ret);
1101			ath10k_ce_deinit(ce_state);
1102			ce_state = NULL;
1103			goto out;
1104		}
1105	}
1106
1107out:
1108	ath10k_pci_sleep(ar);
1109	return ce_state;
1110}
1111
1112void ath10k_ce_deinit(struct ath10k_ce_pipe *ce_state)
1113{
1114	struct ath10k *ar = ce_state->ar;
1115	struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1116
1117	if (ce_state->src_ring) {
1118		kfree(ce_state->src_ring->shadow_base_unaligned);
1119		pci_free_consistent(ar_pci->pdev,
1120				    (ce_state->src_ring->nentries *
1121				     sizeof(struct ce_desc) +
1122				     CE_DESC_RING_ALIGN),
1123				    ce_state->src_ring->base_addr_owner_space,
1124				    ce_state->src_ring->base_addr_ce_space);
1125		kfree(ce_state->src_ring);
1126	}
1127
1128	if (ce_state->dest_ring) {
1129		pci_free_consistent(ar_pci->pdev,
1130				    (ce_state->dest_ring->nentries *
1131				     sizeof(struct ce_desc) +
1132				     CE_DESC_RING_ALIGN),
1133				    ce_state->dest_ring->base_addr_owner_space,
1134				    ce_state->dest_ring->base_addr_ce_space);
1135		kfree(ce_state->dest_ring);
1136	}
1137
1138	ce_state->src_ring = NULL;
1139	ce_state->dest_ring = NULL;
1140}