Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.15.
   1// SPDX-License-Identifier: GPL-2.0-only
   2/*
   3 * STM32 DMA3 controller driver
   4 *
   5 * Copyright (C) STMicroelectronics 2024
   6 * Author(s): Amelie Delaunay <amelie.delaunay@foss.st.com>
   7 */
   8
   9#include <linux/bitfield.h>
  10#include <linux/clk.h>
  11#include <linux/dma-mapping.h>
  12#include <linux/dmaengine.h>
  13#include <linux/dmapool.h>
  14#include <linux/init.h>
  15#include <linux/iopoll.h>
  16#include <linux/list.h>
  17#include <linux/module.h>
  18#include <linux/of_dma.h>
  19#include <linux/platform_device.h>
  20#include <linux/pm_runtime.h>
  21#include <linux/reset.h>
  22#include <linux/slab.h>
  23
  24#include "../virt-dma.h"
  25
  26#define STM32_DMA3_SECCFGR		0x00
  27#define STM32_DMA3_PRIVCFGR		0x04
  28#define STM32_DMA3_RCFGLOCKR		0x08
  29#define STM32_DMA3_MISR			0x0c
  30#define STM32_DMA3_SMISR		0x10
  31
  32#define STM32_DMA3_CLBAR(x)		(0x50 + 0x80 * (x))
  33#define STM32_DMA3_CCIDCFGR(x)		(0x54 + 0x80 * (x))
  34#define STM32_DMA3_CSEMCR(x)		(0x58 + 0x80 * (x))
  35#define STM32_DMA3_CFCR(x)		(0x5c + 0x80 * (x))
  36#define STM32_DMA3_CSR(x)		(0x60 + 0x80 * (x))
  37#define STM32_DMA3_CCR(x)		(0x64 + 0x80 * (x))
  38#define STM32_DMA3_CTR1(x)		(0x90 + 0x80 * (x))
  39#define STM32_DMA3_CTR2(x)		(0x94 + 0x80 * (x))
  40#define STM32_DMA3_CBR1(x)		(0x98 + 0x80 * (x))
  41#define STM32_DMA3_CSAR(x)		(0x9c + 0x80 * (x))
  42#define STM32_DMA3_CDAR(x)		(0xa0 + 0x80 * (x))
  43#define STM32_DMA3_CLLR(x)		(0xcc + 0x80 * (x))
  44
  45#define STM32_DMA3_HWCFGR13		0xfc0 /* G_PER_CTRL(X) x=8..15 */
  46#define STM32_DMA3_HWCFGR12		0xfc4 /* G_PER_CTRL(X) x=0..7 */
  47#define STM32_DMA3_HWCFGR4		0xfe4 /* G_FIFO_SIZE(X) x=8..15 */
  48#define STM32_DMA3_HWCFGR3		0xfe8 /* G_FIFO_SIZE(X) x=0..7 */
  49#define STM32_DMA3_HWCFGR2		0xfec /* G_MAX_REQ_ID */
  50#define STM32_DMA3_HWCFGR1		0xff0 /* G_MASTER_PORTS, G_NUM_CHANNELS, G_Mx_DATA_WIDTH */
  51#define STM32_DMA3_VERR			0xff4
  52
  53/* SECCFGR DMA secure configuration register */
  54#define SECCFGR_SEC(x)			BIT(x)
  55
  56/* MISR DMA non-secure/secure masked interrupt status register */
  57#define MISR_MIS(x)			BIT(x)
  58
  59/* CxLBAR DMA channel x linked_list base address register */
  60#define CLBAR_LBA			GENMASK(31, 16)
  61
  62/* CxCIDCFGR DMA channel x CID register */
  63#define CCIDCFGR_CFEN			BIT(0)
  64#define CCIDCFGR_SEM_EN			BIT(1)
  65#define CCIDCFGR_SCID			GENMASK(5, 4)
  66#define CCIDCFGR_SEM_WLIST_CID0		BIT(16)
  67#define CCIDCFGR_SEM_WLIST_CID1		BIT(17)
  68#define CCIDCFGR_SEM_WLIST_CID2		BIT(18)
  69
  70enum ccidcfgr_cid {
  71	CCIDCFGR_CID0,
  72	CCIDCFGR_CID1,
  73	CCIDCFGR_CID2,
  74};
  75
  76/* CxSEMCR DMA channel x semaphore control register */
  77#define CSEMCR_SEM_MUTEX		BIT(0)
  78#define CSEMCR_SEM_CCID			GENMASK(5, 4)
  79
  80/* CxFCR DMA channel x flag clear register */
  81#define CFCR_TCF			BIT(8)
  82#define CFCR_HTF			BIT(9)
  83#define CFCR_DTEF			BIT(10)
  84#define CFCR_ULEF			BIT(11)
  85#define CFCR_USEF			BIT(12)
  86#define CFCR_SUSPF			BIT(13)
  87
  88/* CxSR DMA channel x status register */
  89#define CSR_IDLEF			BIT(0)
  90#define CSR_TCF				BIT(8)
  91#define CSR_HTF				BIT(9)
  92#define CSR_DTEF			BIT(10)
  93#define CSR_ULEF			BIT(11)
  94#define CSR_USEF			BIT(12)
  95#define CSR_SUSPF			BIT(13)
  96#define CSR_ALL_F			GENMASK(13, 8)
  97#define CSR_FIFOL			GENMASK(24, 16)
  98
  99/* CxCR DMA channel x control register */
 100#define CCR_EN				BIT(0)
 101#define CCR_RESET			BIT(1)
 102#define CCR_SUSP			BIT(2)
 103#define CCR_TCIE			BIT(8)
 104#define CCR_HTIE			BIT(9)
 105#define CCR_DTEIE			BIT(10)
 106#define CCR_ULEIE			BIT(11)
 107#define CCR_USEIE			BIT(12)
 108#define CCR_SUSPIE			BIT(13)
 109#define CCR_ALLIE			GENMASK(13, 8)
 110#define CCR_LSM				BIT(16)
 111#define CCR_LAP				BIT(17)
 112#define CCR_PRIO			GENMASK(23, 22)
 113
 114enum ccr_prio {
 115	CCR_PRIO_LOW,
 116	CCR_PRIO_MID,
 117	CCR_PRIO_HIGH,
 118	CCR_PRIO_VERY_HIGH,
 119};
 120
 121/* CxTR1 DMA channel x transfer register 1 */
 122#define CTR1_SINC			BIT(3)
 123#define CTR1_SBL_1			GENMASK(9, 4)
 124#define CTR1_DINC			BIT(19)
 125#define CTR1_DBL_1			GENMASK(25, 20)
 126#define CTR1_SDW_LOG2			GENMASK(1, 0)
 127#define CTR1_PAM			GENMASK(12, 11)
 128#define CTR1_SAP			BIT(14)
 129#define CTR1_DDW_LOG2			GENMASK(17, 16)
 130#define CTR1_DAP			BIT(30)
 131
 132enum ctr1_dw {
 133	CTR1_DW_BYTE,
 134	CTR1_DW_HWORD,
 135	CTR1_DW_WORD,
 136	CTR1_DW_DWORD, /* Depends on HWCFGR1.G_M0_DATA_WIDTH_ENC and .G_M1_DATA_WIDTH_ENC */
 137};
 138
 139enum ctr1_pam {
 140	CTR1_PAM_0S_LT,		/* if DDW > SDW, padded with 0s else left-truncated */
 141	CTR1_PAM_SE_RT,		/* if DDW > SDW, sign extended else right-truncated */
 142	CTR1_PAM_PACK_UNPACK,	/* FIFO queued */
 143};
 144
 145/* CxTR2 DMA channel x transfer register 2 */
 146#define CTR2_REQSEL			GENMASK(7, 0)
 147#define CTR2_SWREQ			BIT(9)
 148#define CTR2_DREQ			BIT(10)
 149#define CTR2_BREQ			BIT(11)
 150#define CTR2_PFREQ			BIT(12)
 151#define CTR2_TCEM			GENMASK(31, 30)
 152
 153enum ctr2_tcem {
 154	CTR2_TCEM_BLOCK,
 155	CTR2_TCEM_REPEAT_BLOCK,
 156	CTR2_TCEM_LLI,
 157	CTR2_TCEM_CHANNEL,
 158};
 159
 160/* CxBR1 DMA channel x block register 1 */
 161#define CBR1_BNDT			GENMASK(15, 0)
 162
 163/* CxLLR DMA channel x linked-list address register */
 164#define CLLR_LA				GENMASK(15, 2)
 165#define CLLR_ULL			BIT(16)
 166#define CLLR_UDA			BIT(27)
 167#define CLLR_USA			BIT(28)
 168#define CLLR_UB1			BIT(29)
 169#define CLLR_UT2			BIT(30)
 170#define CLLR_UT1			BIT(31)
 171
 172/* HWCFGR13 DMA hardware configuration register 13 x=8..15 */
 173/* HWCFGR12 DMA hardware configuration register 12 x=0..7 */
 174#define G_PER_CTRL(x)			(ULL(0x1) << (4 * (x)))
 175
 176/* HWCFGR4 DMA hardware configuration register 4 x=8..15 */
 177/* HWCFGR3 DMA hardware configuration register 3 x=0..7 */
 178#define G_FIFO_SIZE(x)			(ULL(0x7) << (4 * (x)))
 179
 180#define get_chan_hwcfg(x, mask, reg)	(((reg) & (mask)) >> (4 * (x)))
 181
 182/* HWCFGR2 DMA hardware configuration register 2 */
 183#define G_MAX_REQ_ID			GENMASK(7, 0)
 184
 185/* HWCFGR1 DMA hardware configuration register 1 */
 186#define G_MASTER_PORTS			GENMASK(2, 0)
 187#define G_NUM_CHANNELS			GENMASK(12, 8)
 188#define G_M0_DATA_WIDTH_ENC		GENMASK(25, 24)
 189#define G_M1_DATA_WIDTH_ENC		GENMASK(29, 28)
 190
 191enum stm32_dma3_master_ports {
 192	AXI64,		/* 1x AXI: 64-bit port 0 */
 193	AHB32,		/* 1x AHB: 32-bit port 0 */
 194	AHB32_AHB32,	/* 2x AHB: 32-bit port 0 and 32-bit port 1 */
 195	AXI64_AHB32,	/* 1x AXI 64-bit port 0 and 1x AHB 32-bit port 1 */
 196	AXI64_AXI64,	/* 2x AXI: 64-bit port 0 and 64-bit port 1 */
 197	AXI128_AHB32,	/* 1x AXI 128-bit port 0 and 1x AHB 32-bit port 1 */
 198};
 199
 200enum stm32_dma3_port_data_width {
 201	DW_32,		/* 32-bit, for AHB */
 202	DW_64,		/* 64-bit, for AXI */
 203	DW_128,		/* 128-bit, for AXI */
 204	DW_INVALID,
 205};
 206
 207/* VERR DMA version register */
 208#define VERR_MINREV			GENMASK(3, 0)
 209#define VERR_MAJREV			GENMASK(7, 4)
 210
 211/* Device tree */
 212/* struct stm32_dma3_dt_conf */
 213/* .ch_conf */
 214#define STM32_DMA3_DT_PRIO		GENMASK(1, 0) /* CCR_PRIO */
 215#define STM32_DMA3_DT_FIFO		GENMASK(7, 4)
 216/* .tr_conf */
 217#define STM32_DMA3_DT_SINC		BIT(0) /* CTR1_SINC */
 218#define STM32_DMA3_DT_SAP		BIT(1) /* CTR1_SAP */
 219#define STM32_DMA3_DT_DINC		BIT(4) /* CTR1_DINC */
 220#define STM32_DMA3_DT_DAP		BIT(5) /* CTR1_DAP */
 221#define STM32_DMA3_DT_BREQ		BIT(8) /* CTR2_BREQ */
 222#define STM32_DMA3_DT_PFREQ		BIT(9) /* CTR2_PFREQ */
 223#define STM32_DMA3_DT_TCEM		GENMASK(13, 12) /* CTR2_TCEM */
 224#define STM32_DMA3_DT_NOPACK		BIT(16) /* CTR1_PAM */
 225#define STM32_DMA3_DT_NOREFACT		BIT(17)
 226
 227/* struct stm32_dma3_chan .config_set bitfield */
 228#define STM32_DMA3_CFG_SET_DT		BIT(0)
 229#define STM32_DMA3_CFG_SET_DMA		BIT(1)
 230#define STM32_DMA3_CFG_SET_BOTH		(STM32_DMA3_CFG_SET_DT | STM32_DMA3_CFG_SET_DMA)
 231
 232#define STM32_DMA3_MAX_BLOCK_SIZE	ALIGN_DOWN(CBR1_BNDT, 64)
 233#define STM32_DMA3_MAX_BURST_LEN	(1 + min_t(u32, FIELD_MAX(CTR1_SBL_1), \
 234							FIELD_MAX(CTR1_DBL_1)))
 235#define port_is_ahb(maxdw)		({ typeof(maxdw) (_maxdw) = (maxdw); \
 236					   ((_maxdw) != DW_INVALID) && ((_maxdw) == DW_32); })
 237#define port_is_axi(maxdw)		({ typeof(maxdw) (_maxdw) = (maxdw); \
 238					   ((_maxdw) != DW_INVALID) && ((_maxdw) != DW_32); })
 239#define get_chan_max_dw(maxdw, maxburst)((port_is_ahb(maxdw) ||			     \
 240					  (maxburst) < DMA_SLAVE_BUSWIDTH_8_BYTES) ? \
 241					 DMA_SLAVE_BUSWIDTH_4_BYTES : DMA_SLAVE_BUSWIDTH_8_BYTES)
 242
 243/* Static linked-list data structure (depends on update bits UT1/UT2/UB1/USA/UDA/ULL) */
 244struct stm32_dma3_hwdesc {
 245	u32 ctr1;
 246	u32 ctr2;
 247	u32 cbr1;
 248	u32 csar;
 249	u32 cdar;
 250	u32 cllr;
 251} __packed __aligned(32);
 252
 253/*
 254 * CLLR_LA / sizeof(struct stm32_dma3_hwdesc) represents the number of hdwdesc that can be addressed
 255 * by the pointer to the next linked-list data structure. The __aligned forces the 32-byte
 256 * alignment. So use hardcoded 32. Multiplied by the max block size of each item, it represents
 257 * the sg size limitation.
 258 */
 259#define STM32_DMA3_MAX_SEG_SIZE		((CLLR_LA / 32) * STM32_DMA3_MAX_BLOCK_SIZE)
 260
 261/*
 262 * Linked-list items
 263 */
 264struct stm32_dma3_lli {
 265	struct stm32_dma3_hwdesc *hwdesc;
 266	dma_addr_t hwdesc_addr;
 267};
 268
 269struct stm32_dma3_swdesc {
 270	struct virt_dma_desc vdesc;
 271	u32 ccr;
 272	bool cyclic;
 273	u32 lli_size;
 274	struct stm32_dma3_lli lli[] __counted_by(lli_size);
 275};
 276
 277struct stm32_dma3_dt_conf {
 278	u32 ch_id;
 279	u32 req_line;
 280	u32 ch_conf;
 281	u32 tr_conf;
 282};
 283
 284struct stm32_dma3_chan {
 285	struct virt_dma_chan vchan;
 286	u32 id;
 287	int irq;
 288	u32 fifo_size;
 289	u32 max_burst;
 290	bool semaphore_mode;
 291	struct stm32_dma3_dt_conf dt_config;
 292	struct dma_slave_config dma_config;
 293	u8 config_set;
 294	struct dma_pool *lli_pool;
 295	struct stm32_dma3_swdesc *swdesc;
 296	enum ctr2_tcem tcem;
 297	u32 dma_status;
 298};
 299
 300struct stm32_dma3_pdata {
 301	u32 axi_max_burst_len;
 302};
 303
 304struct stm32_dma3_ddata {
 305	struct dma_device dma_dev;
 306	void __iomem *base;
 307	struct clk *clk;
 308	struct stm32_dma3_chan *chans;
 309	u32 dma_channels;
 310	u32 dma_requests;
 311	enum stm32_dma3_port_data_width ports_max_dw[2];
 312	u32 axi_max_burst_len;
 313};
 314
 315static inline struct stm32_dma3_ddata *to_stm32_dma3_ddata(struct stm32_dma3_chan *chan)
 316{
 317	return container_of(chan->vchan.chan.device, struct stm32_dma3_ddata, dma_dev);
 318}
 319
 320static inline struct stm32_dma3_chan *to_stm32_dma3_chan(struct dma_chan *c)
 321{
 322	return container_of(c, struct stm32_dma3_chan, vchan.chan);
 323}
 324
 325static inline struct stm32_dma3_swdesc *to_stm32_dma3_swdesc(struct virt_dma_desc *vdesc)
 326{
 327	return container_of(vdesc, struct stm32_dma3_swdesc, vdesc);
 328}
 329
 330static struct device *chan2dev(struct stm32_dma3_chan *chan)
 331{
 332	return &chan->vchan.chan.dev->device;
 333}
 334
 335static void stm32_dma3_chan_dump_reg(struct stm32_dma3_chan *chan)
 336{
 337	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 338	struct device *dev = chan2dev(chan);
 339	u32 id = chan->id, offset;
 340
 341	offset = STM32_DMA3_SECCFGR;
 342	dev_dbg(dev, "SECCFGR(0x%03x): %08x\n", offset, readl_relaxed(ddata->base + offset));
 343	offset = STM32_DMA3_PRIVCFGR;
 344	dev_dbg(dev, "PRIVCFGR(0x%03x): %08x\n", offset, readl_relaxed(ddata->base + offset));
 345	offset = STM32_DMA3_CCIDCFGR(id);
 346	dev_dbg(dev, "C%dCIDCFGR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 347	offset = STM32_DMA3_CSEMCR(id);
 348	dev_dbg(dev, "C%dSEMCR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 349	offset = STM32_DMA3_CSR(id);
 350	dev_dbg(dev, "C%dSR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 351	offset = STM32_DMA3_CCR(id);
 352	dev_dbg(dev, "C%dCR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 353	offset = STM32_DMA3_CTR1(id);
 354	dev_dbg(dev, "C%dTR1(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 355	offset = STM32_DMA3_CTR2(id);
 356	dev_dbg(dev, "C%dTR2(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 357	offset = STM32_DMA3_CBR1(id);
 358	dev_dbg(dev, "C%dBR1(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 359	offset = STM32_DMA3_CSAR(id);
 360	dev_dbg(dev, "C%dSAR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 361	offset = STM32_DMA3_CDAR(id);
 362	dev_dbg(dev, "C%dDAR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 363	offset = STM32_DMA3_CLLR(id);
 364	dev_dbg(dev, "C%dLLR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 365	offset = STM32_DMA3_CLBAR(id);
 366	dev_dbg(dev, "C%dLBAR(0x%03x): %08x\n", id, offset, readl_relaxed(ddata->base + offset));
 367}
 368
 369static void stm32_dma3_chan_dump_hwdesc(struct stm32_dma3_chan *chan,
 370					struct stm32_dma3_swdesc *swdesc)
 371{
 372	struct stm32_dma3_hwdesc *hwdesc;
 373	int i;
 374
 375	for (i = 0; i < swdesc->lli_size; i++) {
 376		hwdesc = swdesc->lli[i].hwdesc;
 377		if (i)
 378			dev_dbg(chan2dev(chan), "V\n");
 379		dev_dbg(chan2dev(chan), "[%d]@%pad\n", i, &swdesc->lli[i].hwdesc_addr);
 380		dev_dbg(chan2dev(chan), "| C%dTR1: %08x\n", chan->id, hwdesc->ctr1);
 381		dev_dbg(chan2dev(chan), "| C%dTR2: %08x\n", chan->id, hwdesc->ctr2);
 382		dev_dbg(chan2dev(chan), "| C%dBR1: %08x\n", chan->id, hwdesc->cbr1);
 383		dev_dbg(chan2dev(chan), "| C%dSAR: %08x\n", chan->id, hwdesc->csar);
 384		dev_dbg(chan2dev(chan), "| C%dDAR: %08x\n", chan->id, hwdesc->cdar);
 385		dev_dbg(chan2dev(chan), "| C%dLLR: %08x\n", chan->id, hwdesc->cllr);
 386	}
 387
 388	if (swdesc->cyclic) {
 389		dev_dbg(chan2dev(chan), "|\n");
 390		dev_dbg(chan2dev(chan), "-->[0]@%pad\n", &swdesc->lli[0].hwdesc_addr);
 391	} else {
 392		dev_dbg(chan2dev(chan), "X\n");
 393	}
 394}
 395
 396static struct stm32_dma3_swdesc *stm32_dma3_chan_desc_alloc(struct stm32_dma3_chan *chan, u32 count)
 397{
 398	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 399	struct stm32_dma3_swdesc *swdesc;
 400	int i;
 401
 402	/*
 403	 * If the memory to be allocated for the number of hwdesc (6 u32 members but 32-bytes
 404	 * aligned) is greater than the maximum address of CLLR_LA, then the last items can't be
 405	 * addressed, so abort the allocation.
 406	 */
 407	if ((count * 32) > CLLR_LA) {
 408		dev_err(chan2dev(chan), "Transfer is too big (> %luB)\n", STM32_DMA3_MAX_SEG_SIZE);
 409		return NULL;
 410	}
 411
 412	swdesc = kzalloc(struct_size(swdesc, lli, count), GFP_NOWAIT);
 413	if (!swdesc)
 414		return NULL;
 415	swdesc->lli_size = count;
 416
 417	for (i = 0; i < count; i++) {
 418		swdesc->lli[i].hwdesc = dma_pool_zalloc(chan->lli_pool, GFP_NOWAIT,
 419							&swdesc->lli[i].hwdesc_addr);
 420		if (!swdesc->lli[i].hwdesc)
 421			goto err_pool_free;
 422	}
 423	swdesc->ccr = 0;
 424
 425	/* Set LL base address */
 426	writel_relaxed(swdesc->lli[0].hwdesc_addr & CLBAR_LBA,
 427		       ddata->base + STM32_DMA3_CLBAR(chan->id));
 428
 429	/* Set LL allocated port */
 430	swdesc->ccr &= ~CCR_LAP;
 431
 432	return swdesc;
 433
 434err_pool_free:
 435	dev_err(chan2dev(chan), "Failed to alloc descriptors\n");
 436	while (--i >= 0)
 437		dma_pool_free(chan->lli_pool, swdesc->lli[i].hwdesc, swdesc->lli[i].hwdesc_addr);
 438	kfree(swdesc);
 439
 440	return NULL;
 441}
 442
 443static void stm32_dma3_chan_desc_free(struct stm32_dma3_chan *chan,
 444				      struct stm32_dma3_swdesc *swdesc)
 445{
 446	int i;
 447
 448	for (i = 0; i < swdesc->lli_size; i++)
 449		dma_pool_free(chan->lli_pool, swdesc->lli[i].hwdesc, swdesc->lli[i].hwdesc_addr);
 450
 451	kfree(swdesc);
 452}
 453
 454static void stm32_dma3_chan_vdesc_free(struct virt_dma_desc *vdesc)
 455{
 456	struct stm32_dma3_swdesc *swdesc = to_stm32_dma3_swdesc(vdesc);
 457	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(vdesc->tx.chan);
 458
 459	stm32_dma3_chan_desc_free(chan, swdesc);
 460}
 461
 462static void stm32_dma3_check_user_setting(struct stm32_dma3_chan *chan)
 463{
 464	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 465	struct device *dev = chan2dev(chan);
 466	u32 ctr1 = readl_relaxed(ddata->base + STM32_DMA3_CTR1(chan->id));
 467	u32 cbr1 = readl_relaxed(ddata->base + STM32_DMA3_CBR1(chan->id));
 468	u32 csar = readl_relaxed(ddata->base + STM32_DMA3_CSAR(chan->id));
 469	u32 cdar = readl_relaxed(ddata->base + STM32_DMA3_CDAR(chan->id));
 470	u32 cllr = readl_relaxed(ddata->base + STM32_DMA3_CLLR(chan->id));
 471	u32 bndt = FIELD_GET(CBR1_BNDT, cbr1);
 472	u32 sdw = 1 << FIELD_GET(CTR1_SDW_LOG2, ctr1);
 473	u32 ddw = 1 << FIELD_GET(CTR1_DDW_LOG2, ctr1);
 474	u32 sap = FIELD_GET(CTR1_SAP, ctr1);
 475	u32 dap = FIELD_GET(CTR1_DAP, ctr1);
 476
 477	if (!bndt && !FIELD_GET(CLLR_UB1, cllr))
 478		dev_err(dev, "null source block size and no update of this value\n");
 479	if (bndt % sdw)
 480		dev_err(dev, "source block size not multiple of src data width\n");
 481	if (FIELD_GET(CTR1_PAM, ctr1) == CTR1_PAM_PACK_UNPACK && bndt % ddw)
 482		dev_err(dev, "(un)packing mode w/ src block size not multiple of dst data width\n");
 483	if (csar % sdw)
 484		dev_err(dev, "unaligned source address not multiple of src data width\n");
 485	if (cdar % ddw)
 486		dev_err(dev, "unaligned destination address not multiple of dst data width\n");
 487	if (sdw == DMA_SLAVE_BUSWIDTH_8_BYTES && port_is_ahb(ddata->ports_max_dw[sap]))
 488		dev_err(dev, "double-word source data width not supported on port %u\n", sap);
 489	if (ddw == DMA_SLAVE_BUSWIDTH_8_BYTES && port_is_ahb(ddata->ports_max_dw[dap]))
 490		dev_err(dev, "double-word destination data width not supported on port %u\n", dap);
 491}
 492
 493static void stm32_dma3_chan_prep_hwdesc(struct stm32_dma3_chan *chan,
 494					struct stm32_dma3_swdesc *swdesc,
 495					u32 curr, dma_addr_t src, dma_addr_t dst, u32 len,
 496					u32 ctr1, u32 ctr2, bool is_last, bool is_cyclic)
 497{
 498	struct stm32_dma3_hwdesc *hwdesc;
 499	dma_addr_t next_lli;
 500	u32 next = curr + 1;
 501
 502	hwdesc = swdesc->lli[curr].hwdesc;
 503	hwdesc->ctr1 = ctr1;
 504	hwdesc->ctr2 = ctr2;
 505	hwdesc->cbr1 = FIELD_PREP(CBR1_BNDT, len);
 506	hwdesc->csar = src;
 507	hwdesc->cdar = dst;
 508
 509	if (is_last) {
 510		if (is_cyclic)
 511			next_lli = swdesc->lli[0].hwdesc_addr;
 512		else
 513			next_lli = 0;
 514	} else {
 515		next_lli = swdesc->lli[next].hwdesc_addr;
 516	}
 517
 518	hwdesc->cllr = 0;
 519	if (next_lli) {
 520		hwdesc->cllr |= CLLR_UT1 | CLLR_UT2 | CLLR_UB1;
 521		hwdesc->cllr |= CLLR_USA | CLLR_UDA | CLLR_ULL;
 522		hwdesc->cllr |= (next_lli & CLLR_LA);
 523	}
 524
 525	/*
 526	 * Make sure to flush the CPU's write buffers so that the descriptors are ready to be read
 527	 * by DMA3. By explicitly using a write memory barrier here, instead of doing it with writel
 528	 * to enable the channel, we avoid an unnecessary barrier in the case where the descriptors
 529	 * are reused (DMA_CTRL_REUSE).
 530	 */
 531	if (is_last)
 532		dma_wmb();
 533}
 534
 535static enum dma_slave_buswidth stm32_dma3_get_max_dw(u32 chan_max_burst,
 536						     enum stm32_dma3_port_data_width port_max_dw,
 537						     u32 len, dma_addr_t addr)
 538{
 539	enum dma_slave_buswidth max_dw = get_chan_max_dw(port_max_dw, chan_max_burst);
 540
 541	/* len and addr must be a multiple of dw */
 542	return 1 << __ffs(len | addr | max_dw);
 543}
 544
 545static u32 stm32_dma3_get_max_burst(u32 len, enum dma_slave_buswidth dw,
 546				    u32 chan_max_burst, u32 bus_max_burst)
 547{
 548	u32 max_burst = chan_max_burst ? chan_max_burst / dw : 1;
 549
 550	/* len is a multiple of dw, so if len is < chan_max_burst, shorten burst */
 551	if (len < chan_max_burst)
 552		max_burst = len / dw;
 553
 554	/*
 555	 * HW doesn't modify the burst if burst size <= half of the fifo size.
 556	 * If len is not a multiple of burst size, last burst is shortened by HW.
 557	 * Take care of maximum burst supported on interconnect bus.
 558	 */
 559	return min_t(u32, max_burst, bus_max_burst);
 560}
 561
 562static int stm32_dma3_chan_prep_hw(struct stm32_dma3_chan *chan, enum dma_transfer_direction dir,
 563				   u32 *ccr, u32 *ctr1, u32 *ctr2,
 564				   dma_addr_t src_addr, dma_addr_t dst_addr, u32 len)
 565{
 566	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 567	struct dma_device dma_device = ddata->dma_dev;
 568	u32 src_max_burst = STM32_DMA3_MAX_BURST_LEN, dst_max_burst = STM32_DMA3_MAX_BURST_LEN;
 569	u32 sdw, ddw, sbl_max, dbl_max, tcem, init_dw, init_bl_max;
 570	u32 _ctr1 = 0, _ctr2 = 0;
 571	u32 ch_conf = chan->dt_config.ch_conf;
 572	u32 tr_conf = chan->dt_config.tr_conf;
 573	u32 sap = FIELD_GET(STM32_DMA3_DT_SAP, tr_conf), sap_max_dw;
 574	u32 dap = FIELD_GET(STM32_DMA3_DT_DAP, tr_conf), dap_max_dw;
 575
 576	dev_dbg(chan2dev(chan), "%s from %pad to %pad\n",
 577		dmaengine_get_direction_text(dir), &src_addr, &dst_addr);
 578
 579	sdw = chan->dma_config.src_addr_width ? : get_chan_max_dw(sap, chan->max_burst);
 580	ddw = chan->dma_config.dst_addr_width ? : get_chan_max_dw(dap, chan->max_burst);
 581	sbl_max = chan->dma_config.src_maxburst ? : 1;
 582	dbl_max = chan->dma_config.dst_maxburst ? : 1;
 583
 584	/* Following conditions would raise User Setting Error interrupt */
 585	if (!(dma_device.src_addr_widths & BIT(sdw)) || !(dma_device.dst_addr_widths & BIT(ddw))) {
 586		dev_err(chan2dev(chan), "Bus width (src=%u, dst=%u) not supported\n", sdw, ddw);
 587		return -EINVAL;
 588	}
 589
 590	if (ddata->ports_max_dw[1] == DW_INVALID && (sap || dap)) {
 591		dev_err(chan2dev(chan), "Only one master port, port 1 is not supported\n");
 592		return -EINVAL;
 593	}
 594
 595	sap_max_dw = ddata->ports_max_dw[sap];
 596	dap_max_dw = ddata->ports_max_dw[dap];
 597	if ((port_is_ahb(sap_max_dw) && sdw == DMA_SLAVE_BUSWIDTH_8_BYTES) ||
 598	    (port_is_ahb(dap_max_dw) && ddw == DMA_SLAVE_BUSWIDTH_8_BYTES)) {
 599		dev_err(chan2dev(chan),
 600			"8 bytes buswidth (src=%u, dst=%u) not supported on port (sap=%u, dap=%u\n",
 601			sdw, ddw, sap, dap);
 602		return -EINVAL;
 603	}
 604
 605	if (FIELD_GET(STM32_DMA3_DT_SINC, tr_conf))
 606		_ctr1 |= CTR1_SINC;
 607	if (sap)
 608		_ctr1 |= CTR1_SAP;
 609	if (port_is_axi(sap_max_dw)) /* AXI - apply axi maximum burst limitation */
 610		src_max_burst = ddata->axi_max_burst_len;
 611	if (FIELD_GET(STM32_DMA3_DT_DINC, tr_conf))
 612		_ctr1 |= CTR1_DINC;
 613	if (dap)
 614		_ctr1 |= CTR1_DAP;
 615	if (port_is_axi(dap_max_dw)) /* AXI - apply axi maximum burst limitation */
 616		dst_max_burst = ddata->axi_max_burst_len;
 617
 618	_ctr2 |= FIELD_PREP(CTR2_REQSEL, chan->dt_config.req_line) & ~CTR2_SWREQ;
 619	if (FIELD_GET(STM32_DMA3_DT_BREQ, tr_conf))
 620		_ctr2 |= CTR2_BREQ;
 621	if (dir == DMA_DEV_TO_MEM && FIELD_GET(STM32_DMA3_DT_PFREQ, tr_conf))
 622		_ctr2 |= CTR2_PFREQ;
 623	tcem = FIELD_GET(STM32_DMA3_DT_TCEM, tr_conf);
 624	_ctr2 |= FIELD_PREP(CTR2_TCEM, tcem);
 625
 626	/* Store TCEM to know on which event TC flag occurred */
 627	chan->tcem = tcem;
 628	/* Store direction for residue computation */
 629	chan->dma_config.direction = dir;
 630
 631	switch (dir) {
 632	case DMA_MEM_TO_DEV:
 633		/* Set destination (device) data width and burst */
 634		ddw = min_t(u32, ddw, stm32_dma3_get_max_dw(chan->max_burst, dap_max_dw,
 635							    len, dst_addr));
 636		dbl_max = min_t(u32, dbl_max, stm32_dma3_get_max_burst(len, ddw, chan->max_burst,
 637								       dst_max_burst));
 638
 639		/* Set source (memory) data width and burst */
 640		sdw = stm32_dma3_get_max_dw(chan->max_burst, sap_max_dw, len, src_addr);
 641		sbl_max = stm32_dma3_get_max_burst(len, sdw, chan->max_burst, src_max_burst);
 642		if (!!FIELD_GET(STM32_DMA3_DT_NOPACK, tr_conf)) {
 643			sdw = ddw;
 644			sbl_max = dbl_max;
 645		}
 646
 647		_ctr1 |= FIELD_PREP(CTR1_SDW_LOG2, ilog2(sdw));
 648		_ctr1 |= FIELD_PREP(CTR1_SBL_1, sbl_max - 1);
 649		_ctr1 |= FIELD_PREP(CTR1_DDW_LOG2, ilog2(ddw));
 650		_ctr1 |= FIELD_PREP(CTR1_DBL_1, dbl_max - 1);
 651
 652		if (ddw != sdw) {
 653			_ctr1 |= FIELD_PREP(CTR1_PAM, CTR1_PAM_PACK_UNPACK);
 654			/* Should never reach this case as ddw is clamped down */
 655			if (len & (ddw - 1)) {
 656				dev_err(chan2dev(chan),
 657					"Packing mode is enabled and len is not multiple of ddw");
 658				return -EINVAL;
 659			}
 660		}
 661
 662		/* dst = dev */
 663		_ctr2 |= CTR2_DREQ;
 664
 665		break;
 666
 667	case DMA_DEV_TO_MEM:
 668		/* Set source (device) data width and burst */
 669		sdw = min_t(u32, sdw, stm32_dma3_get_max_dw(chan->max_burst, sap_max_dw,
 670							    len, src_addr));
 671		sbl_max = min_t(u32, sbl_max, stm32_dma3_get_max_burst(len, sdw, chan->max_burst,
 672								       src_max_burst));
 673
 674		/* Set destination (memory) data width and burst */
 675		ddw = stm32_dma3_get_max_dw(chan->max_burst, dap_max_dw, len, dst_addr);
 676		dbl_max = stm32_dma3_get_max_burst(len, ddw, chan->max_burst, dst_max_burst);
 677		if (!!FIELD_GET(STM32_DMA3_DT_NOPACK, tr_conf) ||
 678		    ((_ctr2 & CTR2_PFREQ) && ddw > sdw)) { /* Packing to wider ddw not supported */
 679			ddw = sdw;
 680			dbl_max = sbl_max;
 681		}
 682
 683		_ctr1 |= FIELD_PREP(CTR1_SDW_LOG2, ilog2(sdw));
 684		_ctr1 |= FIELD_PREP(CTR1_SBL_1, sbl_max - 1);
 685		_ctr1 |= FIELD_PREP(CTR1_DDW_LOG2, ilog2(ddw));
 686		_ctr1 |= FIELD_PREP(CTR1_DBL_1, dbl_max - 1);
 687
 688		if (ddw != sdw) {
 689			_ctr1 |= FIELD_PREP(CTR1_PAM, CTR1_PAM_PACK_UNPACK);
 690			/* Should never reach this case as ddw is clamped down */
 691			if (len & (ddw - 1)) {
 692				dev_err(chan2dev(chan),
 693					"Packing mode is enabled and len is not multiple of ddw\n");
 694				return -EINVAL;
 695			}
 696		}
 697
 698		/* dst = mem */
 699		_ctr2 &= ~CTR2_DREQ;
 700
 701		break;
 702
 703	case DMA_MEM_TO_MEM:
 704		/* Set source (memory) data width and burst */
 705		init_dw = sdw;
 706		init_bl_max = sbl_max;
 707		sdw = stm32_dma3_get_max_dw(chan->max_burst, sap_max_dw, len, src_addr);
 708		sbl_max = stm32_dma3_get_max_burst(len, sdw, chan->max_burst, src_max_burst);
 709		if (chan->config_set & STM32_DMA3_CFG_SET_DMA) {
 710			sdw = min_t(u32, init_dw, sdw);
 711			sbl_max = min_t(u32, init_bl_max, stm32_dma3_get_max_burst(len, sdw,
 712										   chan->max_burst,
 713										   src_max_burst));
 714		}
 715
 716		/* Set destination (memory) data width and burst */
 717		init_dw = ddw;
 718		init_bl_max = dbl_max;
 719		ddw = stm32_dma3_get_max_dw(chan->max_burst, dap_max_dw, len, dst_addr);
 720		dbl_max = stm32_dma3_get_max_burst(len, ddw, chan->max_burst, dst_max_burst);
 721		if (chan->config_set & STM32_DMA3_CFG_SET_DMA) {
 722			ddw = min_t(u32, init_dw, ddw);
 723			dbl_max = min_t(u32, init_bl_max, stm32_dma3_get_max_burst(len, ddw,
 724										   chan->max_burst,
 725										   dst_max_burst));
 726		}
 727
 728		_ctr1 |= FIELD_PREP(CTR1_SDW_LOG2, ilog2(sdw));
 729		_ctr1 |= FIELD_PREP(CTR1_SBL_1, sbl_max - 1);
 730		_ctr1 |= FIELD_PREP(CTR1_DDW_LOG2, ilog2(ddw));
 731		_ctr1 |= FIELD_PREP(CTR1_DBL_1, dbl_max - 1);
 732
 733		if (ddw != sdw) {
 734			_ctr1 |= FIELD_PREP(CTR1_PAM, CTR1_PAM_PACK_UNPACK);
 735			/* Should never reach this case as ddw is clamped down */
 736			if (len & (ddw - 1)) {
 737				dev_err(chan2dev(chan),
 738					"Packing mode is enabled and len is not multiple of ddw");
 739				return -EINVAL;
 740			}
 741		}
 742
 743		/* CTR2_REQSEL/DREQ/BREQ/PFREQ are ignored with CTR2_SWREQ=1 */
 744		_ctr2 |= CTR2_SWREQ;
 745
 746		break;
 747
 748	default:
 749		dev_err(chan2dev(chan), "Direction %s not supported\n",
 750			dmaengine_get_direction_text(dir));
 751		return -EINVAL;
 752	}
 753
 754	*ccr |= FIELD_PREP(CCR_PRIO, FIELD_GET(STM32_DMA3_DT_PRIO, ch_conf));
 755	*ctr1 = _ctr1;
 756	*ctr2 = _ctr2;
 757
 758	dev_dbg(chan2dev(chan), "%s: sdw=%u bytes sbl=%u beats ddw=%u bytes dbl=%u beats\n",
 759		__func__, sdw, sbl_max, ddw, dbl_max);
 760
 761	return 0;
 762}
 763
 764static void stm32_dma3_chan_start(struct stm32_dma3_chan *chan)
 765{
 766	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 767	struct virt_dma_desc *vdesc;
 768	struct stm32_dma3_hwdesc *hwdesc;
 769	u32 id = chan->id;
 770	u32 csr, ccr;
 771
 772	vdesc = vchan_next_desc(&chan->vchan);
 773	if (!vdesc) {
 774		chan->swdesc = NULL;
 775		return;
 776	}
 777	list_del(&vdesc->node);
 778
 779	chan->swdesc = to_stm32_dma3_swdesc(vdesc);
 780	hwdesc = chan->swdesc->lli[0].hwdesc;
 781
 782	stm32_dma3_chan_dump_hwdesc(chan, chan->swdesc);
 783
 784	writel_relaxed(chan->swdesc->ccr, ddata->base + STM32_DMA3_CCR(id));
 785	writel_relaxed(hwdesc->ctr1, ddata->base + STM32_DMA3_CTR1(id));
 786	writel_relaxed(hwdesc->ctr2, ddata->base + STM32_DMA3_CTR2(id));
 787	writel_relaxed(hwdesc->cbr1, ddata->base + STM32_DMA3_CBR1(id));
 788	writel_relaxed(hwdesc->csar, ddata->base + STM32_DMA3_CSAR(id));
 789	writel_relaxed(hwdesc->cdar, ddata->base + STM32_DMA3_CDAR(id));
 790	writel_relaxed(hwdesc->cllr, ddata->base + STM32_DMA3_CLLR(id));
 791
 792	/* Clear any pending interrupts */
 793	csr = readl_relaxed(ddata->base + STM32_DMA3_CSR(id));
 794	if (csr & CSR_ALL_F)
 795		writel_relaxed(csr, ddata->base + STM32_DMA3_CFCR(id));
 796
 797	stm32_dma3_chan_dump_reg(chan);
 798
 799	ccr = readl_relaxed(ddata->base + STM32_DMA3_CCR(id));
 800	writel_relaxed(ccr | CCR_EN, ddata->base + STM32_DMA3_CCR(id));
 801
 802	chan->dma_status = DMA_IN_PROGRESS;
 803
 804	dev_dbg(chan2dev(chan), "vchan %pK: started\n", &chan->vchan);
 805}
 806
 807static int stm32_dma3_chan_suspend(struct stm32_dma3_chan *chan, bool susp)
 808{
 809	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 810	u32 csr, ccr = readl_relaxed(ddata->base + STM32_DMA3_CCR(chan->id)) & ~CCR_EN;
 811	int ret = 0;
 812
 813	if (susp)
 814		ccr |= CCR_SUSP;
 815	else
 816		ccr &= ~CCR_SUSP;
 817
 818	writel_relaxed(ccr, ddata->base + STM32_DMA3_CCR(chan->id));
 819
 820	if (susp) {
 821		ret = readl_relaxed_poll_timeout_atomic(ddata->base + STM32_DMA3_CSR(chan->id), csr,
 822							csr & CSR_SUSPF, 1, 10);
 823		if (!ret)
 824			writel_relaxed(CFCR_SUSPF, ddata->base + STM32_DMA3_CFCR(chan->id));
 825
 826		stm32_dma3_chan_dump_reg(chan);
 827	}
 828
 829	return ret;
 830}
 831
 832static void stm32_dma3_chan_reset(struct stm32_dma3_chan *chan)
 833{
 834	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 835	u32 ccr = readl_relaxed(ddata->base + STM32_DMA3_CCR(chan->id)) & ~CCR_EN;
 836
 837	writel_relaxed(ccr |= CCR_RESET, ddata->base + STM32_DMA3_CCR(chan->id));
 838}
 839
 840static int stm32_dma3_chan_get_curr_hwdesc(struct stm32_dma3_swdesc *swdesc, u32 cllr, u32 *residue)
 841{
 842	u32 i, lli_offset, next_lli_offset = cllr & CLLR_LA;
 843
 844	/* If cllr is null, it means it is either the last or single item */
 845	if (!cllr)
 846		return swdesc->lli_size - 1;
 847
 848	/* In cyclic mode, go fast and first check we are not on the last item */
 849	if (swdesc->cyclic && next_lli_offset == (swdesc->lli[0].hwdesc_addr & CLLR_LA))
 850		return swdesc->lli_size - 1;
 851
 852	/* As transfer is in progress, look backward from the last item */
 853	for (i = swdesc->lli_size - 1; i > 0; i--) {
 854		*residue += FIELD_GET(CBR1_BNDT, swdesc->lli[i].hwdesc->cbr1);
 855		lli_offset = swdesc->lli[i].hwdesc_addr & CLLR_LA;
 856		if (lli_offset == next_lli_offset)
 857			return i - 1;
 858	}
 859
 860	return -EINVAL;
 861}
 862
 863static void stm32_dma3_chan_set_residue(struct stm32_dma3_chan *chan,
 864					struct stm32_dma3_swdesc *swdesc,
 865					struct dma_tx_state *txstate)
 866{
 867	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 868	struct device *dev = chan2dev(chan);
 869	struct stm32_dma3_hwdesc *hwdesc;
 870	u32 residue, curr_lli, csr, cdar, cbr1, cllr, bndt, fifol;
 871	bool pack_unpack;
 872	int ret;
 873
 874	csr = readl_relaxed(ddata->base + STM32_DMA3_CSR(chan->id));
 875	if (!(csr & CSR_IDLEF) && chan->dma_status != DMA_PAUSED) {
 876		/* Suspend current transfer to read registers for a snapshot */
 877		writel_relaxed(swdesc->ccr | CCR_SUSP, ddata->base + STM32_DMA3_CCR(chan->id));
 878		ret = readl_relaxed_poll_timeout_atomic(ddata->base + STM32_DMA3_CSR(chan->id), csr,
 879							csr & (CSR_SUSPF | CSR_IDLEF), 1, 10);
 880
 881		if (ret || ((csr & CSR_TCF) && (csr & CSR_IDLEF))) {
 882			writel_relaxed(CFCR_SUSPF, ddata->base + STM32_DMA3_CFCR(chan->id));
 883			writel_relaxed(swdesc->ccr, ddata->base + STM32_DMA3_CCR(chan->id));
 884			if (ret)
 885				dev_err(dev, "Channel suspension timeout, csr=%08x\n", csr);
 886		}
 887	}
 888
 889	/* If channel is still active (CSR_IDLEF is not set), can't get a reliable residue */
 890	if (!(csr & CSR_IDLEF))
 891		dev_warn(dev, "Can't get residue: channel still active, csr=%08x\n", csr);
 892
 893	/*
 894	 * If channel is not suspended, but Idle and Transfer Complete are set,
 895	 * linked-list is over, no residue
 896	 */
 897	if (!(csr & CSR_SUSPF) && (csr & CSR_TCF) && (csr & CSR_IDLEF))
 898		return;
 899
 900	/* Read registers to have a snapshot */
 901	cllr = readl_relaxed(ddata->base + STM32_DMA3_CLLR(chan->id));
 902	cbr1 = readl_relaxed(ddata->base + STM32_DMA3_CBR1(chan->id));
 903	cdar = readl_relaxed(ddata->base + STM32_DMA3_CDAR(chan->id));
 904
 905	/* Resume current transfer */
 906	if (csr & CSR_SUSPF) {
 907		writel_relaxed(CFCR_SUSPF, ddata->base + STM32_DMA3_CFCR(chan->id));
 908		writel_relaxed(swdesc->ccr, ddata->base + STM32_DMA3_CCR(chan->id));
 909	}
 910
 911	/* Add current BNDT */
 912	bndt = FIELD_GET(CBR1_BNDT, cbr1);
 913	residue = bndt;
 914
 915	/* Get current hwdesc and cumulate residue of pending hwdesc BNDT */
 916	ret = stm32_dma3_chan_get_curr_hwdesc(swdesc, cllr, &residue);
 917	if (ret < 0) {
 918		dev_err(chan2dev(chan), "Can't get residue: current hwdesc not found\n");
 919		return;
 920	}
 921	curr_lli = ret;
 922
 923	/* Read current FIFO level - in units of programmed destination data width */
 924	hwdesc = swdesc->lli[curr_lli].hwdesc;
 925	fifol = FIELD_GET(CSR_FIFOL, csr) * (1 << FIELD_GET(CTR1_DDW_LOG2, hwdesc->ctr1));
 926	/* If the FIFO contains as many bytes as its size, it can't contain more */
 927	if (fifol == (1 << (chan->fifo_size + 1)))
 928		goto skip_fifol_update;
 929
 930	/*
 931	 * In case of PACKING (Destination burst length > Source burst length) or UNPACKING
 932	 * (Source burst length > Destination burst length), bytes could be pending in the FIFO
 933	 * (to be packed up to Destination burst length or unpacked into Destination burst length
 934	 * chunks).
 935	 * BNDT is not reliable, as it reflects the number of bytes read from the source but not the
 936	 * number of bytes written to the destination.
 937	 * FIFOL is also not sufficient, because it reflects the number of available write beats in
 938	 * units of Destination data width but not the bytes not yet packed or unpacked.
 939	 * In case of Destination increment DINC, it is possible to compute the number of bytes in
 940	 * the FIFO:
 941	 * fifol_in_bytes = bytes_read - bytes_written.
 942	 */
 943	pack_unpack = !!(FIELD_GET(CTR1_PAM, hwdesc->ctr1) == CTR1_PAM_PACK_UNPACK);
 944	if (pack_unpack && (hwdesc->ctr1 & CTR1_DINC)) {
 945		int bytes_read = FIELD_GET(CBR1_BNDT, hwdesc->cbr1) - bndt;
 946		int bytes_written = cdar - hwdesc->cdar;
 947
 948		if (bytes_read > 0)
 949			fifol = bytes_read - bytes_written;
 950	}
 951
 952skip_fifol_update:
 953	if (fifol) {
 954		dev_dbg(chan2dev(chan), "%u byte(s) in the FIFO\n", fifol);
 955		dma_set_in_flight_bytes(txstate, fifol);
 956		/*
 957		 * Residue is already accurate for DMA_MEM_TO_DEV as BNDT reflects data read from
 958		 * the source memory buffer, so just need to add fifol to residue in case of
 959		 * DMA_DEV_TO_MEM transfer because these bytes are not yet written in destination
 960		 * memory buffer.
 961		 */
 962		if (chan->dma_config.direction == DMA_DEV_TO_MEM)
 963			residue += fifol;
 964	}
 965	dma_set_residue(txstate, residue);
 966}
 967
 968static int stm32_dma3_chan_stop(struct stm32_dma3_chan *chan)
 969{
 970	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
 971	u32 ccr;
 972	int ret = 0;
 973
 974	chan->dma_status = DMA_COMPLETE;
 975
 976	/* Disable interrupts */
 977	ccr = readl_relaxed(ddata->base + STM32_DMA3_CCR(chan->id));
 978	writel_relaxed(ccr & ~(CCR_ALLIE | CCR_EN), ddata->base + STM32_DMA3_CCR(chan->id));
 979
 980	if (!(ccr & CCR_SUSP) && (ccr & CCR_EN)) {
 981		/* Suspend the channel */
 982		ret = stm32_dma3_chan_suspend(chan, true);
 983		if (ret)
 984			dev_warn(chan2dev(chan), "%s: timeout, data might be lost\n", __func__);
 985	}
 986
 987	/*
 988	 * Reset the channel: this causes the reset of the FIFO and the reset of the channel
 989	 * internal state, the reset of CCR_EN and CCR_SUSP bits.
 990	 */
 991	stm32_dma3_chan_reset(chan);
 992
 993	return ret;
 994}
 995
 996static void stm32_dma3_chan_complete(struct stm32_dma3_chan *chan)
 997{
 998	if (!chan->swdesc)
 999		return;
1000
1001	vchan_cookie_complete(&chan->swdesc->vdesc);
1002	chan->swdesc = NULL;
1003	stm32_dma3_chan_start(chan);
1004}
1005
1006static irqreturn_t stm32_dma3_chan_irq(int irq, void *devid)
1007{
1008	struct stm32_dma3_chan *chan = devid;
1009	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
1010	u32 misr, csr, ccr;
1011
1012	spin_lock(&chan->vchan.lock);
1013
1014	misr = readl_relaxed(ddata->base + STM32_DMA3_MISR);
1015	if (!(misr & MISR_MIS(chan->id))) {
1016		spin_unlock(&chan->vchan.lock);
1017		return IRQ_NONE;
1018	}
1019
1020	csr = readl_relaxed(ddata->base + STM32_DMA3_CSR(chan->id));
1021	ccr = readl_relaxed(ddata->base + STM32_DMA3_CCR(chan->id)) & CCR_ALLIE;
1022
1023	if (csr & CSR_TCF && ccr & CCR_TCIE) {
1024		if (chan->swdesc->cyclic)
1025			vchan_cyclic_callback(&chan->swdesc->vdesc);
1026		else
1027			stm32_dma3_chan_complete(chan);
1028	}
1029
1030	if (csr & CSR_USEF && ccr & CCR_USEIE) {
1031		dev_err(chan2dev(chan), "User setting error\n");
1032		chan->dma_status = DMA_ERROR;
1033		/* CCR.EN automatically cleared by HW */
1034		stm32_dma3_check_user_setting(chan);
1035		stm32_dma3_chan_reset(chan);
1036	}
1037
1038	if (csr & CSR_ULEF && ccr & CCR_ULEIE) {
1039		dev_err(chan2dev(chan), "Update link transfer error\n");
1040		chan->dma_status = DMA_ERROR;
1041		/* CCR.EN automatically cleared by HW */
1042		stm32_dma3_chan_reset(chan);
1043	}
1044
1045	if (csr & CSR_DTEF && ccr & CCR_DTEIE) {
1046		dev_err(chan2dev(chan), "Data transfer error\n");
1047		chan->dma_status = DMA_ERROR;
1048		/* CCR.EN automatically cleared by HW */
1049		stm32_dma3_chan_reset(chan);
1050	}
1051
1052	/*
1053	 * Half Transfer Interrupt may be disabled but Half Transfer Flag can be set,
1054	 * ensure HTF flag to be cleared, with other flags.
1055	 */
1056	csr &= (ccr | CCR_HTIE);
1057
1058	if (csr)
1059		writel_relaxed(csr, ddata->base + STM32_DMA3_CFCR(chan->id));
1060
1061	spin_unlock(&chan->vchan.lock);
1062
1063	return IRQ_HANDLED;
1064}
1065
1066static int stm32_dma3_alloc_chan_resources(struct dma_chan *c)
1067{
1068	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1069	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
1070	u32 id = chan->id, csemcr, ccid;
1071	int ret;
1072
1073	ret = pm_runtime_resume_and_get(ddata->dma_dev.dev);
1074	if (ret < 0)
1075		return ret;
1076
1077	/* Ensure the channel is free */
1078	if (chan->semaphore_mode &&
1079	    readl_relaxed(ddata->base + STM32_DMA3_CSEMCR(chan->id)) & CSEMCR_SEM_MUTEX) {
1080		ret = -EBUSY;
1081		goto err_put_sync;
1082	}
1083
1084	chan->lli_pool = dmam_pool_create(dev_name(&c->dev->device), c->device->dev,
1085					  sizeof(struct stm32_dma3_hwdesc),
1086					  __alignof__(struct stm32_dma3_hwdesc), SZ_64K);
1087	if (!chan->lli_pool) {
1088		dev_err(chan2dev(chan), "Failed to create LLI pool\n");
1089		ret = -ENOMEM;
1090		goto err_put_sync;
1091	}
1092
1093	/* Take the channel semaphore */
1094	if (chan->semaphore_mode) {
1095		writel_relaxed(CSEMCR_SEM_MUTEX, ddata->base + STM32_DMA3_CSEMCR(id));
1096		csemcr = readl_relaxed(ddata->base + STM32_DMA3_CSEMCR(id));
1097		ccid = FIELD_GET(CSEMCR_SEM_CCID, csemcr);
1098		/* Check that the channel is well taken */
1099		if (ccid != CCIDCFGR_CID1) {
1100			dev_err(chan2dev(chan), "Not under CID1 control (in-use by CID%d)\n", ccid);
1101			ret = -EPERM;
1102			goto err_pool_destroy;
1103		}
1104		dev_dbg(chan2dev(chan), "Under CID1 control (semcr=0x%08x)\n", csemcr);
1105	}
1106
1107	return 0;
1108
1109err_pool_destroy:
1110	dmam_pool_destroy(chan->lli_pool);
1111	chan->lli_pool = NULL;
1112
1113err_put_sync:
1114	pm_runtime_put_sync(ddata->dma_dev.dev);
1115
1116	return ret;
1117}
1118
1119static void stm32_dma3_free_chan_resources(struct dma_chan *c)
1120{
1121	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1122	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
1123	unsigned long flags;
1124
1125	/* Ensure channel is in idle state */
1126	spin_lock_irqsave(&chan->vchan.lock, flags);
1127	stm32_dma3_chan_stop(chan);
1128	chan->swdesc = NULL;
1129	spin_unlock_irqrestore(&chan->vchan.lock, flags);
1130
1131	vchan_free_chan_resources(to_virt_chan(c));
1132
1133	dmam_pool_destroy(chan->lli_pool);
1134	chan->lli_pool = NULL;
1135
1136	/* Release the channel semaphore */
1137	if (chan->semaphore_mode)
1138		writel_relaxed(0, ddata->base + STM32_DMA3_CSEMCR(chan->id));
1139
1140	pm_runtime_put_sync(ddata->dma_dev.dev);
1141
1142	/* Reset configuration */
1143	memset(&chan->dt_config, 0, sizeof(chan->dt_config));
1144	memset(&chan->dma_config, 0, sizeof(chan->dma_config));
1145	chan->config_set = 0;
1146}
1147
1148static u32 stm32_dma3_get_ll_count(struct stm32_dma3_chan *chan, size_t len, bool prevent_refactor)
1149{
1150	u32 count;
1151
1152	if (prevent_refactor)
1153		return DIV_ROUND_UP(len, STM32_DMA3_MAX_BLOCK_SIZE);
1154
1155	count = len / STM32_DMA3_MAX_BLOCK_SIZE;
1156	len -= (len / STM32_DMA3_MAX_BLOCK_SIZE) * STM32_DMA3_MAX_BLOCK_SIZE;
1157
1158	if (len >= chan->max_burst) {
1159		count += 1; /* len < STM32_DMA3_MAX_BLOCK_SIZE here, so it fits in one item */
1160		len -= (len / chan->max_burst) * chan->max_burst;
1161	}
1162
1163	/* Unaligned remainder fits in one extra item */
1164	if (len > 0)
1165		count += 1;
1166
1167	return count;
1168}
1169
1170static void stm32_dma3_init_chan_config_for_memcpy(struct stm32_dma3_chan *chan,
1171						   dma_addr_t dst, dma_addr_t src)
1172{
1173	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
1174	u32 dw = get_chan_max_dw(ddata->ports_max_dw[0], chan->max_burst); /* port 0 by default */
1175	u32 burst = chan->max_burst / dw;
1176
1177	/* Initialize dt_config if channel not pre-configured through DT */
1178	if (!(chan->config_set & STM32_DMA3_CFG_SET_DT)) {
1179		chan->dt_config.ch_conf = FIELD_PREP(STM32_DMA3_DT_PRIO, CCR_PRIO_VERY_HIGH);
1180		chan->dt_config.ch_conf |= FIELD_PREP(STM32_DMA3_DT_FIFO, chan->fifo_size);
1181		chan->dt_config.tr_conf = STM32_DMA3_DT_SINC | STM32_DMA3_DT_DINC;
1182		chan->dt_config.tr_conf |= FIELD_PREP(STM32_DMA3_DT_TCEM, CTR2_TCEM_CHANNEL);
1183	}
1184
1185	/* Initialize dma_config if dmaengine_slave_config() not used */
1186	if (!(chan->config_set & STM32_DMA3_CFG_SET_DMA)) {
1187		chan->dma_config.src_addr_width = dw;
1188		chan->dma_config.dst_addr_width = dw;
1189		chan->dma_config.src_maxburst = burst;
1190		chan->dma_config.dst_maxburst = burst;
1191		chan->dma_config.src_addr = src;
1192		chan->dma_config.dst_addr = dst;
1193	}
1194}
1195
1196static struct dma_async_tx_descriptor *stm32_dma3_prep_dma_memcpy(struct dma_chan *c,
1197								  dma_addr_t dst, dma_addr_t src,
1198								  size_t len, unsigned long flags)
1199{
1200	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1201	struct stm32_dma3_swdesc *swdesc;
1202	size_t next_size, offset;
1203	u32 count, i, ctr1, ctr2;
1204	bool prevent_refactor = !!FIELD_GET(STM32_DMA3_DT_NOPACK, chan->dt_config.tr_conf) ||
1205				!!FIELD_GET(STM32_DMA3_DT_NOREFACT, chan->dt_config.tr_conf);
1206
1207	count = stm32_dma3_get_ll_count(chan, len, prevent_refactor);
1208
1209	swdesc = stm32_dma3_chan_desc_alloc(chan, count);
1210	if (!swdesc)
1211		return NULL;
1212
1213	if (chan->config_set != STM32_DMA3_CFG_SET_BOTH)
1214		stm32_dma3_init_chan_config_for_memcpy(chan, dst, src);
1215
1216	for (i = 0, offset = 0; offset < len; i++, offset += next_size) {
1217		size_t remaining;
1218		int ret;
1219
1220		remaining = len - offset;
1221		next_size = min_t(size_t, remaining, STM32_DMA3_MAX_BLOCK_SIZE);
1222
1223		if (!prevent_refactor &&
1224		    (next_size < STM32_DMA3_MAX_BLOCK_SIZE && next_size >= chan->max_burst))
1225			next_size = chan->max_burst * (remaining / chan->max_burst);
1226
1227		ret = stm32_dma3_chan_prep_hw(chan, DMA_MEM_TO_MEM, &swdesc->ccr, &ctr1, &ctr2,
1228					      src + offset, dst + offset, next_size);
1229		if (ret)
1230			goto err_desc_free;
1231
1232		stm32_dma3_chan_prep_hwdesc(chan, swdesc, i, src + offset, dst + offset, next_size,
1233					    ctr1, ctr2, next_size == remaining, false);
1234	}
1235
1236	/* Enable Errors interrupts */
1237	swdesc->ccr |= CCR_USEIE | CCR_ULEIE | CCR_DTEIE;
1238	/* Enable Transfer state interrupts */
1239	swdesc->ccr |= CCR_TCIE;
1240
1241	swdesc->cyclic = false;
1242
1243	return vchan_tx_prep(&chan->vchan, &swdesc->vdesc, flags);
1244
1245err_desc_free:
1246	stm32_dma3_chan_desc_free(chan, swdesc);
1247
1248	return NULL;
1249}
1250
1251static struct dma_async_tx_descriptor *stm32_dma3_prep_slave_sg(struct dma_chan *c,
1252								struct scatterlist *sgl,
1253								unsigned int sg_len,
1254								enum dma_transfer_direction dir,
1255								unsigned long flags, void *context)
1256{
1257	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1258	struct stm32_dma3_swdesc *swdesc;
1259	struct scatterlist *sg;
1260	size_t len;
1261	dma_addr_t sg_addr, dev_addr, src, dst;
1262	u32 i, j, count, ctr1, ctr2;
1263	bool prevent_refactor = !!FIELD_GET(STM32_DMA3_DT_NOPACK, chan->dt_config.tr_conf) ||
1264				!!FIELD_GET(STM32_DMA3_DT_NOREFACT, chan->dt_config.tr_conf);
1265	int ret;
1266
1267	count = 0;
1268	for_each_sg(sgl, sg, sg_len, i)
1269		count += stm32_dma3_get_ll_count(chan, sg_dma_len(sg), prevent_refactor);
1270
1271	swdesc = stm32_dma3_chan_desc_alloc(chan, count);
1272	if (!swdesc)
1273		return NULL;
1274
1275	/* sg_len and i correspond to the initial sgl; count and j correspond to the hwdesc LL */
1276	j = 0;
1277	for_each_sg(sgl, sg, sg_len, i) {
1278		sg_addr = sg_dma_address(sg);
1279		dev_addr = (dir == DMA_MEM_TO_DEV) ? chan->dma_config.dst_addr :
1280						     chan->dma_config.src_addr;
1281		len = sg_dma_len(sg);
1282
1283		do {
1284			size_t chunk = min_t(size_t, len, STM32_DMA3_MAX_BLOCK_SIZE);
1285
1286			if (!prevent_refactor &&
1287			    (chunk < STM32_DMA3_MAX_BLOCK_SIZE && chunk >= chan->max_burst))
1288				chunk = chan->max_burst * (len / chan->max_burst);
1289
1290			if (dir == DMA_MEM_TO_DEV) {
1291				src = sg_addr;
1292				dst = dev_addr;
1293
1294				ret = stm32_dma3_chan_prep_hw(chan, dir, &swdesc->ccr, &ctr1, &ctr2,
1295							      src, dst, chunk);
1296
1297				if (FIELD_GET(CTR1_DINC, ctr1))
1298					dev_addr += chunk;
1299			} else { /* (dir == DMA_DEV_TO_MEM || dir == DMA_MEM_TO_MEM) */
1300				src = dev_addr;
1301				dst = sg_addr;
1302
1303				ret = stm32_dma3_chan_prep_hw(chan, dir, &swdesc->ccr, &ctr1, &ctr2,
1304							      src, dst, chunk);
1305
1306				if (FIELD_GET(CTR1_SINC, ctr1))
1307					dev_addr += chunk;
1308			}
1309
1310			if (ret)
1311				goto err_desc_free;
1312
1313			stm32_dma3_chan_prep_hwdesc(chan, swdesc, j, src, dst, chunk,
1314						    ctr1, ctr2, j == (count - 1), false);
1315
1316			sg_addr += chunk;
1317			len -= chunk;
1318			j++;
1319		} while (len);
1320	}
1321
1322	if (count != sg_len && chan->tcem != CTR2_TCEM_CHANNEL)
1323		dev_warn(chan2dev(chan), "Linked-list refactored, %d items instead of %d\n",
1324			 count, sg_len);
1325
1326	/* Enable Error interrupts */
1327	swdesc->ccr |= CCR_USEIE | CCR_ULEIE | CCR_DTEIE;
1328	/* Enable Transfer state interrupts */
1329	swdesc->ccr |= CCR_TCIE;
1330
1331	swdesc->cyclic = false;
1332
1333	return vchan_tx_prep(&chan->vchan, &swdesc->vdesc, flags);
1334
1335err_desc_free:
1336	stm32_dma3_chan_desc_free(chan, swdesc);
1337
1338	return NULL;
1339}
1340
1341static struct dma_async_tx_descriptor *stm32_dma3_prep_dma_cyclic(struct dma_chan *c,
1342								  dma_addr_t buf_addr,
1343								  size_t buf_len, size_t period_len,
1344								  enum dma_transfer_direction dir,
1345								  unsigned long flags)
1346{
1347	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1348	struct stm32_dma3_swdesc *swdesc;
1349	dma_addr_t src, dst;
1350	u32 count, i, ctr1, ctr2;
1351	int ret;
1352
1353	if (!buf_len || !period_len || period_len > STM32_DMA3_MAX_BLOCK_SIZE) {
1354		dev_err(chan2dev(chan), "Invalid buffer/period length\n");
1355		return NULL;
1356	}
1357
1358	if (buf_len % period_len) {
1359		dev_err(chan2dev(chan), "Buffer length not multiple of period length\n");
1360		return NULL;
1361	}
1362
1363	count = buf_len / period_len;
1364	swdesc = stm32_dma3_chan_desc_alloc(chan, count);
1365	if (!swdesc)
1366		return NULL;
1367
1368	if (dir == DMA_MEM_TO_DEV) {
1369		src = buf_addr;
1370		dst = chan->dma_config.dst_addr;
1371
1372		ret = stm32_dma3_chan_prep_hw(chan, DMA_MEM_TO_DEV, &swdesc->ccr, &ctr1, &ctr2,
1373					      src, dst, period_len);
1374	} else if (dir == DMA_DEV_TO_MEM) {
1375		src = chan->dma_config.src_addr;
1376		dst = buf_addr;
1377
1378		ret = stm32_dma3_chan_prep_hw(chan, DMA_DEV_TO_MEM, &swdesc->ccr, &ctr1, &ctr2,
1379					      src, dst, period_len);
1380	} else {
1381		dev_err(chan2dev(chan), "Invalid direction\n");
1382		ret = -EINVAL;
1383	}
1384
1385	if (ret)
1386		goto err_desc_free;
1387
1388	for (i = 0; i < count; i++) {
1389		if (dir == DMA_MEM_TO_DEV) {
1390			src = buf_addr + i * period_len;
1391			dst = chan->dma_config.dst_addr;
1392		} else { /* (dir == DMA_DEV_TO_MEM) */
1393			src = chan->dma_config.src_addr;
1394			dst = buf_addr + i * period_len;
1395		}
1396
1397		stm32_dma3_chan_prep_hwdesc(chan, swdesc, i, src, dst, period_len,
1398					    ctr1, ctr2, i == (count - 1), true);
1399	}
1400
1401	/* Enable Error interrupts */
1402	swdesc->ccr |= CCR_USEIE | CCR_ULEIE | CCR_DTEIE;
1403	/* Enable Transfer state interrupts */
1404	swdesc->ccr |= CCR_TCIE;
1405
1406	swdesc->cyclic = true;
1407
1408	return vchan_tx_prep(&chan->vchan, &swdesc->vdesc, flags);
1409
1410err_desc_free:
1411	stm32_dma3_chan_desc_free(chan, swdesc);
1412
1413	return NULL;
1414}
1415
1416static void stm32_dma3_caps(struct dma_chan *c, struct dma_slave_caps *caps)
1417{
1418	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1419
1420	if (!chan->fifo_size) {
1421		caps->max_burst = 0;
1422		caps->src_addr_widths &= ~BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1423		caps->dst_addr_widths &= ~BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1424	} else {
1425		/* Burst transfer should not exceed half of the fifo size */
1426		caps->max_burst = chan->max_burst;
1427		if (caps->max_burst < DMA_SLAVE_BUSWIDTH_8_BYTES) {
1428			caps->src_addr_widths &= ~BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1429			caps->dst_addr_widths &= ~BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1430		}
1431	}
1432}
1433
1434static int stm32_dma3_config(struct dma_chan *c, struct dma_slave_config *config)
1435{
1436	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1437
1438	memcpy(&chan->dma_config, config, sizeof(*config));
1439	chan->config_set |= STM32_DMA3_CFG_SET_DMA;
1440
1441	return 0;
1442}
1443
1444static int stm32_dma3_pause(struct dma_chan *c)
1445{
1446	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1447	int ret;
1448
1449	ret = stm32_dma3_chan_suspend(chan, true);
1450	if (ret)
1451		return ret;
1452
1453	chan->dma_status = DMA_PAUSED;
1454
1455	dev_dbg(chan2dev(chan), "vchan %pK: paused\n", &chan->vchan);
1456
1457	return 0;
1458}
1459
1460static int stm32_dma3_resume(struct dma_chan *c)
1461{
1462	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1463
1464	stm32_dma3_chan_suspend(chan, false);
1465
1466	chan->dma_status = DMA_IN_PROGRESS;
1467
1468	dev_dbg(chan2dev(chan), "vchan %pK: resumed\n", &chan->vchan);
1469
1470	return 0;
1471}
1472
1473static int stm32_dma3_terminate_all(struct dma_chan *c)
1474{
1475	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1476	unsigned long flags;
1477	LIST_HEAD(head);
1478
1479	spin_lock_irqsave(&chan->vchan.lock, flags);
1480
1481	if (chan->swdesc) {
1482		vchan_terminate_vdesc(&chan->swdesc->vdesc);
1483		chan->swdesc = NULL;
1484	}
1485
1486	stm32_dma3_chan_stop(chan);
1487
1488	vchan_get_all_descriptors(&chan->vchan, &head);
1489
1490	spin_unlock_irqrestore(&chan->vchan.lock, flags);
1491	vchan_dma_desc_free_list(&chan->vchan, &head);
1492
1493	dev_dbg(chan2dev(chan), "vchan %pK: terminated\n", &chan->vchan);
1494
1495	return 0;
1496}
1497
1498static void stm32_dma3_synchronize(struct dma_chan *c)
1499{
1500	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1501
1502	vchan_synchronize(&chan->vchan);
1503}
1504
1505static enum dma_status stm32_dma3_tx_status(struct dma_chan *c, dma_cookie_t cookie,
1506					    struct dma_tx_state *txstate)
1507{
1508	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1509	struct stm32_dma3_swdesc *swdesc = NULL;
1510	enum dma_status status;
1511	unsigned long flags;
1512	struct virt_dma_desc *vd;
1513
1514	status = dma_cookie_status(c, cookie, txstate);
1515	if (status == DMA_COMPLETE)
1516		return status;
1517
1518	if (!txstate)
1519		return chan->dma_status;
1520
1521	spin_lock_irqsave(&chan->vchan.lock, flags);
1522
1523	vd = vchan_find_desc(&chan->vchan, cookie);
1524	if (vd)
1525		swdesc = to_stm32_dma3_swdesc(vd);
1526	else if (chan->swdesc && chan->swdesc->vdesc.tx.cookie == cookie)
1527		swdesc = chan->swdesc;
1528
1529	/* Get residue/in_flight_bytes only if a transfer is currently running (swdesc != NULL) */
1530	if (swdesc)
1531		stm32_dma3_chan_set_residue(chan, swdesc, txstate);
1532
1533	spin_unlock_irqrestore(&chan->vchan.lock, flags);
1534
1535	return chan->dma_status;
1536}
1537
1538static void stm32_dma3_issue_pending(struct dma_chan *c)
1539{
1540	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1541	unsigned long flags;
1542
1543	spin_lock_irqsave(&chan->vchan.lock, flags);
1544
1545	if (vchan_issue_pending(&chan->vchan) && !chan->swdesc) {
1546		dev_dbg(chan2dev(chan), "vchan %pK: issued\n", &chan->vchan);
1547		stm32_dma3_chan_start(chan);
1548	}
1549
1550	spin_unlock_irqrestore(&chan->vchan.lock, flags);
1551}
1552
1553static bool stm32_dma3_filter_fn(struct dma_chan *c, void *fn_param)
1554{
1555	struct stm32_dma3_chan *chan = to_stm32_dma3_chan(c);
1556	struct stm32_dma3_ddata *ddata = to_stm32_dma3_ddata(chan);
1557	struct stm32_dma3_dt_conf *conf = fn_param;
1558	u32 mask, semcr;
1559	int ret;
1560
1561	dev_dbg(c->device->dev, "%s(%s): req_line=%d ch_conf=%08x tr_conf=%08x\n",
1562		__func__, dma_chan_name(c), conf->req_line, conf->ch_conf, conf->tr_conf);
1563
1564	if (!of_property_read_u32(c->device->dev->of_node, "dma-channel-mask", &mask))
1565		if (!(mask & BIT(chan->id)))
1566			return false;
1567
1568	ret = pm_runtime_resume_and_get(ddata->dma_dev.dev);
1569	if (ret < 0)
1570		return false;
1571	semcr = readl_relaxed(ddata->base + STM32_DMA3_CSEMCR(chan->id));
1572	pm_runtime_put_sync(ddata->dma_dev.dev);
1573
1574	/* Check if chan is free */
1575	if (semcr & CSEMCR_SEM_MUTEX)
1576		return false;
1577
1578	/* Check if chan fifo fits well */
1579	if (FIELD_GET(STM32_DMA3_DT_FIFO, conf->ch_conf) != chan->fifo_size)
1580		return false;
1581
1582	return true;
1583}
1584
1585static struct dma_chan *stm32_dma3_of_xlate(struct of_phandle_args *dma_spec, struct of_dma *ofdma)
1586{
1587	struct stm32_dma3_ddata *ddata = ofdma->of_dma_data;
1588	dma_cap_mask_t mask = ddata->dma_dev.cap_mask;
1589	struct stm32_dma3_dt_conf conf;
1590	struct stm32_dma3_chan *chan;
1591	struct dma_chan *c;
1592
1593	if (dma_spec->args_count < 3) {
1594		dev_err(ddata->dma_dev.dev, "Invalid args count\n");
1595		return NULL;
1596	}
1597
1598	conf.req_line = dma_spec->args[0];
1599	conf.ch_conf = dma_spec->args[1];
1600	conf.tr_conf = dma_spec->args[2];
1601
1602	if (conf.req_line >= ddata->dma_requests) {
1603		dev_err(ddata->dma_dev.dev, "Invalid request line\n");
1604		return NULL;
1605	}
1606
1607	/* Request dma channel among the generic dma controller list */
1608	c = dma_request_channel(mask, stm32_dma3_filter_fn, &conf);
1609	if (!c) {
1610		dev_err(ddata->dma_dev.dev, "No suitable channel found\n");
1611		return NULL;
1612	}
1613
1614	chan = to_stm32_dma3_chan(c);
1615	chan->dt_config = conf;
1616	chan->config_set |= STM32_DMA3_CFG_SET_DT;
1617
1618	return c;
1619}
1620
1621static u32 stm32_dma3_check_rif(struct stm32_dma3_ddata *ddata)
1622{
1623	u32 chan_reserved, mask = 0, i, ccidcfgr, invalid_cid = 0;
1624
1625	/* Reserve Secure channels */
1626	chan_reserved = readl_relaxed(ddata->base + STM32_DMA3_SECCFGR);
1627
1628	/*
1629	 * CID filtering must be configured to ensure that the DMA3 channel will inherit the CID of
1630	 * the processor which is configuring and using the given channel.
1631	 * In case CID filtering is not configured, dma-channel-mask property can be used to
1632	 * specify available DMA channels to the kernel.
1633	 */
1634	of_property_read_u32(ddata->dma_dev.dev->of_node, "dma-channel-mask", &mask);
1635
1636	/* Reserve !CID-filtered not in dma-channel-mask, static CID != CID1, CID1 not allowed */
1637	for (i = 0; i < ddata->dma_channels; i++) {
1638		ccidcfgr = readl_relaxed(ddata->base + STM32_DMA3_CCIDCFGR(i));
1639
1640		if (!(ccidcfgr & CCIDCFGR_CFEN)) { /* !CID-filtered */
1641			invalid_cid |= BIT(i);
1642			if (!(mask & BIT(i))) /* Not in dma-channel-mask */
1643				chan_reserved |= BIT(i);
1644		} else { /* CID-filtered */
1645			if (!(ccidcfgr & CCIDCFGR_SEM_EN)) { /* Static CID mode */
1646				if (FIELD_GET(CCIDCFGR_SCID, ccidcfgr) != CCIDCFGR_CID1)
1647					chan_reserved |= BIT(i);
1648			} else { /* Semaphore mode */
1649				if (!FIELD_GET(CCIDCFGR_SEM_WLIST_CID1, ccidcfgr))
1650					chan_reserved |= BIT(i);
1651				ddata->chans[i].semaphore_mode = true;
1652			}
1653		}
1654		dev_dbg(ddata->dma_dev.dev, "chan%d: %s mode, %s\n", i,
1655			!(ccidcfgr & CCIDCFGR_CFEN) ? "!CID-filtered" :
1656			ddata->chans[i].semaphore_mode ? "Semaphore" : "Static CID",
1657			(chan_reserved & BIT(i)) ? "denied" :
1658			mask & BIT(i) ? "force allowed" : "allowed");
1659	}
1660
1661	if (invalid_cid)
1662		dev_warn(ddata->dma_dev.dev, "chan%*pbl have invalid CID configuration\n",
1663			 ddata->dma_channels, &invalid_cid);
1664
1665	return chan_reserved;
1666}
1667
1668static struct stm32_dma3_pdata stm32mp25_pdata = {
1669	.axi_max_burst_len = 16,
1670};
1671
1672static const struct of_device_id stm32_dma3_of_match[] = {
1673	{ .compatible = "st,stm32mp25-dma3", .data = &stm32mp25_pdata, },
1674	{ /* sentinel */ },
1675};
1676MODULE_DEVICE_TABLE(of, stm32_dma3_of_match);
1677
1678static int stm32_dma3_probe(struct platform_device *pdev)
1679{
1680	struct device_node *np = pdev->dev.of_node;
1681	const struct stm32_dma3_pdata *pdata;
1682	struct stm32_dma3_ddata *ddata;
1683	struct reset_control *reset;
1684	struct stm32_dma3_chan *chan;
1685	struct dma_device *dma_dev;
1686	u32 master_ports, chan_reserved, i, verr;
1687	u64 hwcfgr;
1688	int ret;
1689
1690	ddata = devm_kzalloc(&pdev->dev, sizeof(*ddata), GFP_KERNEL);
1691	if (!ddata)
1692		return -ENOMEM;
1693	platform_set_drvdata(pdev, ddata);
1694
1695	dma_dev = &ddata->dma_dev;
1696
1697	ddata->base = devm_platform_ioremap_resource(pdev, 0);
1698	if (IS_ERR(ddata->base))
1699		return PTR_ERR(ddata->base);
1700
1701	ddata->clk = devm_clk_get(&pdev->dev, NULL);
1702	if (IS_ERR(ddata->clk))
1703		return dev_err_probe(&pdev->dev, PTR_ERR(ddata->clk), "Failed to get clk\n");
1704
1705	reset = devm_reset_control_get_optional(&pdev->dev, NULL);
1706	if (IS_ERR(reset))
1707		return dev_err_probe(&pdev->dev, PTR_ERR(reset), "Failed to get reset\n");
1708
1709	ret = clk_prepare_enable(ddata->clk);
1710	if (ret)
1711		return dev_err_probe(&pdev->dev, ret, "Failed to enable clk\n");
1712
1713	reset_control_reset(reset);
1714
1715	INIT_LIST_HEAD(&dma_dev->channels);
1716
1717	dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
1718	dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask);
1719	dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask);
1720	dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
1721	dma_dev->dev = &pdev->dev;
1722	/*
1723	 * This controller supports up to 8-byte buswidth depending on the port used and the
1724	 * channel, and can only access address at even boundaries, multiple of the buswidth.
1725	 */
1726	dma_dev->copy_align = DMAENGINE_ALIGN_8_BYTES;
1727	dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) |
1728				   BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) |
1729				   BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) |
1730				   BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1731	dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) |
1732				   BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) |
1733				   BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) |
1734				   BIT(DMA_SLAVE_BUSWIDTH_8_BYTES);
1735	dma_dev->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV) | BIT(DMA_MEM_TO_MEM);
1736
1737	dma_dev->descriptor_reuse = true;
1738	dma_dev->max_sg_burst = STM32_DMA3_MAX_SEG_SIZE;
1739	dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
1740	dma_dev->device_alloc_chan_resources = stm32_dma3_alloc_chan_resources;
1741	dma_dev->device_free_chan_resources = stm32_dma3_free_chan_resources;
1742	dma_dev->device_prep_dma_memcpy = stm32_dma3_prep_dma_memcpy;
1743	dma_dev->device_prep_slave_sg = stm32_dma3_prep_slave_sg;
1744	dma_dev->device_prep_dma_cyclic = stm32_dma3_prep_dma_cyclic;
1745	dma_dev->device_caps = stm32_dma3_caps;
1746	dma_dev->device_config = stm32_dma3_config;
1747	dma_dev->device_pause = stm32_dma3_pause;
1748	dma_dev->device_resume = stm32_dma3_resume;
1749	dma_dev->device_terminate_all = stm32_dma3_terminate_all;
1750	dma_dev->device_synchronize = stm32_dma3_synchronize;
1751	dma_dev->device_tx_status = stm32_dma3_tx_status;
1752	dma_dev->device_issue_pending = stm32_dma3_issue_pending;
1753
1754	/* if dma_channels is not modified, get it from hwcfgr1 */
1755	if (of_property_read_u32(np, "dma-channels", &ddata->dma_channels)) {
1756		hwcfgr = readl_relaxed(ddata->base + STM32_DMA3_HWCFGR1);
1757		ddata->dma_channels = FIELD_GET(G_NUM_CHANNELS, hwcfgr);
1758	}
1759
1760	/* if dma_requests is not modified, get it from hwcfgr2 */
1761	if (of_property_read_u32(np, "dma-requests", &ddata->dma_requests)) {
1762		hwcfgr = readl_relaxed(ddata->base + STM32_DMA3_HWCFGR2);
1763		ddata->dma_requests = FIELD_GET(G_MAX_REQ_ID, hwcfgr) + 1;
1764	}
1765
1766	/* G_MASTER_PORTS, G_M0_DATA_WIDTH_ENC, G_M1_DATA_WIDTH_ENC in HWCFGR1 */
1767	hwcfgr = readl_relaxed(ddata->base + STM32_DMA3_HWCFGR1);
1768	master_ports = FIELD_GET(G_MASTER_PORTS, hwcfgr);
1769
1770	ddata->ports_max_dw[0] = FIELD_GET(G_M0_DATA_WIDTH_ENC, hwcfgr);
1771	if (master_ports == AXI64 || master_ports == AHB32) /* Single master port */
1772		ddata->ports_max_dw[1] = DW_INVALID;
1773	else /* Dual master ports */
1774		ddata->ports_max_dw[1] = FIELD_GET(G_M1_DATA_WIDTH_ENC, hwcfgr);
1775
1776	/* axi_max_burst_len is optional, if not defined, use STM32_DMA3_MAX_BURST_LEN  */
1777	ddata->axi_max_burst_len = STM32_DMA3_MAX_BURST_LEN;
1778	pdata = device_get_match_data(&pdev->dev);
1779	if (pdata && pdata->axi_max_burst_len) {
1780		ddata->axi_max_burst_len = min_t(u32, pdata->axi_max_burst_len,
1781						 STM32_DMA3_MAX_BURST_LEN);
1782		dev_dbg(&pdev->dev, "Burst is limited to %u beats through AXI port\n",
1783			ddata->axi_max_burst_len);
1784	}
1785
1786	ddata->chans = devm_kcalloc(&pdev->dev, ddata->dma_channels, sizeof(*ddata->chans),
1787				    GFP_KERNEL);
1788	if (!ddata->chans) {
1789		ret = -ENOMEM;
1790		goto err_clk_disable;
1791	}
1792
1793	chan_reserved = stm32_dma3_check_rif(ddata);
1794
1795	if (chan_reserved == GENMASK(ddata->dma_channels - 1, 0)) {
1796		ret = -ENODEV;
1797		dev_err_probe(&pdev->dev, ret, "No channel available, abort registration\n");
1798		goto err_clk_disable;
1799	}
1800
1801	/* G_FIFO_SIZE x=0..7 in HWCFGR3 and G_FIFO_SIZE x=8..15 in HWCFGR4 */
1802	hwcfgr = readl_relaxed(ddata->base + STM32_DMA3_HWCFGR3);
1803	hwcfgr |= ((u64)readl_relaxed(ddata->base + STM32_DMA3_HWCFGR4)) << 32;
1804
1805	for (i = 0; i < ddata->dma_channels; i++) {
1806		if (chan_reserved & BIT(i))
1807			continue;
1808
1809		chan = &ddata->chans[i];
1810		chan->id = i;
1811		chan->fifo_size = get_chan_hwcfg(i, G_FIFO_SIZE(i), hwcfgr);
1812		/* If chan->fifo_size > 0 then half of the fifo size, else no burst when no FIFO */
1813		chan->max_burst = (chan->fifo_size) ? (1 << (chan->fifo_size + 1)) / 2 : 0;
1814	}
1815
1816	ret = dmaenginem_async_device_register(dma_dev);
1817	if (ret)
1818		goto err_clk_disable;
1819
1820	for (i = 0; i < ddata->dma_channels; i++) {
1821		char name[12];
1822
1823		if (chan_reserved & BIT(i))
1824			continue;
1825
1826		chan = &ddata->chans[i];
1827		snprintf(name, sizeof(name), "dma%dchan%d", ddata->dma_dev.dev_id, chan->id);
1828
1829		chan->vchan.desc_free = stm32_dma3_chan_vdesc_free;
1830		vchan_init(&chan->vchan, dma_dev);
1831
1832		ret = dma_async_device_channel_register(&ddata->dma_dev, &chan->vchan.chan, name);
1833		if (ret) {
1834			dev_err_probe(&pdev->dev, ret, "Failed to register channel %s\n", name);
1835			goto err_clk_disable;
1836		}
1837
1838		ret = platform_get_irq(pdev, i);
1839		if (ret < 0)
1840			goto err_clk_disable;
1841		chan->irq = ret;
1842
1843		ret = devm_request_irq(&pdev->dev, chan->irq, stm32_dma3_chan_irq, 0,
1844				       dev_name(chan2dev(chan)), chan);
1845		if (ret) {
1846			dev_err_probe(&pdev->dev, ret, "Failed to request channel %s IRQ\n",
1847				      dev_name(chan2dev(chan)));
1848			goto err_clk_disable;
1849		}
1850	}
1851
1852	ret = of_dma_controller_register(np, stm32_dma3_of_xlate, ddata);
1853	if (ret) {
1854		dev_err_probe(&pdev->dev, ret, "Failed to register controller\n");
1855		goto err_clk_disable;
1856	}
1857
1858	verr = readl_relaxed(ddata->base + STM32_DMA3_VERR);
1859
1860	pm_runtime_set_active(&pdev->dev);
1861	pm_runtime_enable(&pdev->dev);
1862	pm_runtime_get_noresume(&pdev->dev);
1863	pm_runtime_put(&pdev->dev);
1864
1865	dev_info(&pdev->dev, "STM32 DMA3 registered rev:%lu.%lu\n",
1866		 FIELD_GET(VERR_MAJREV, verr), FIELD_GET(VERR_MINREV, verr));
1867
1868	return 0;
1869
1870err_clk_disable:
1871	clk_disable_unprepare(ddata->clk);
1872
1873	return ret;
1874}
1875
1876static void stm32_dma3_remove(struct platform_device *pdev)
1877{
1878	pm_runtime_disable(&pdev->dev);
1879}
1880
1881static int stm32_dma3_runtime_suspend(struct device *dev)
1882{
1883	struct stm32_dma3_ddata *ddata = dev_get_drvdata(dev);
1884
1885	clk_disable_unprepare(ddata->clk);
1886
1887	return 0;
1888}
1889
1890static int stm32_dma3_runtime_resume(struct device *dev)
1891{
1892	struct stm32_dma3_ddata *ddata = dev_get_drvdata(dev);
1893	int ret;
1894
1895	ret = clk_prepare_enable(ddata->clk);
1896	if (ret)
1897		dev_err(dev, "Failed to enable clk: %d\n", ret);
1898
1899	return ret;
1900}
1901
1902static const struct dev_pm_ops stm32_dma3_pm_ops = {
1903	SYSTEM_SLEEP_PM_OPS(pm_runtime_force_suspend, pm_runtime_force_resume)
1904	RUNTIME_PM_OPS(stm32_dma3_runtime_suspend, stm32_dma3_runtime_resume, NULL)
1905};
1906
1907static struct platform_driver stm32_dma3_driver = {
1908	.probe = stm32_dma3_probe,
1909	.remove = stm32_dma3_remove,
1910	.driver = {
1911		.name = "stm32-dma3",
1912		.of_match_table = stm32_dma3_of_match,
1913		.pm = pm_ptr(&stm32_dma3_pm_ops),
1914	},
1915};
1916
1917static int __init stm32_dma3_init(void)
1918{
1919	return platform_driver_register(&stm32_dma3_driver);
1920}
1921
1922subsys_initcall(stm32_dma3_init);
1923
1924MODULE_DESCRIPTION("STM32 DMA3 controller driver");
1925MODULE_AUTHOR("Amelie Delaunay <amelie.delaunay@foss.st.com>");
1926MODULE_LICENSE("GPL");