Linux Audio

Check our new training course

Real-Time Linux with PREEMPT_RT training

Feb 18-20, 2025
Register
Loading...
v6.2
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
   4 *
   5 * Author: Vitor Soares <vitor.soares@synopsys.com>
   6 */
   7
   8#include <linux/bitops.h>
   9#include <linux/clk.h>
  10#include <linux/completion.h>
  11#include <linux/err.h>
  12#include <linux/errno.h>
  13#include <linux/i3c/master.h>
  14#include <linux/interrupt.h>
  15#include <linux/ioport.h>
  16#include <linux/iopoll.h>
  17#include <linux/list.h>
  18#include <linux/module.h>
  19#include <linux/of.h>
  20#include <linux/platform_device.h>
  21#include <linux/reset.h>
  22#include <linux/slab.h>
  23
 
 
  24#define DEVICE_CTRL			0x0
  25#define DEV_CTRL_ENABLE			BIT(31)
  26#define DEV_CTRL_RESUME			BIT(30)
  27#define DEV_CTRL_HOT_JOIN_NACK		BIT(8)
  28#define DEV_CTRL_I2C_SLAVE_PRESENT	BIT(7)
  29
  30#define DEVICE_ADDR			0x4
  31#define DEV_ADDR_DYNAMIC_ADDR_VALID	BIT(31)
  32#define DEV_ADDR_DYNAMIC(x)		(((x) << 16) & GENMASK(22, 16))
  33
  34#define HW_CAPABILITY			0x8
  35#define COMMAND_QUEUE_PORT		0xc
  36#define COMMAND_PORT_TOC		BIT(30)
  37#define COMMAND_PORT_READ_TRANSFER	BIT(28)
  38#define COMMAND_PORT_SDAP		BIT(27)
  39#define COMMAND_PORT_ROC		BIT(26)
  40#define COMMAND_PORT_SPEED(x)		(((x) << 21) & GENMASK(23, 21))
  41#define COMMAND_PORT_DEV_INDEX(x)	(((x) << 16) & GENMASK(20, 16))
  42#define COMMAND_PORT_CP			BIT(15)
  43#define COMMAND_PORT_CMD(x)		(((x) << 7) & GENMASK(14, 7))
  44#define COMMAND_PORT_TID(x)		(((x) << 3) & GENMASK(6, 3))
  45
  46#define COMMAND_PORT_ARG_DATA_LEN(x)	(((x) << 16) & GENMASK(31, 16))
  47#define COMMAND_PORT_ARG_DATA_LEN_MAX	65536
  48#define COMMAND_PORT_TRANSFER_ARG	0x01
  49
  50#define COMMAND_PORT_SDA_DATA_BYTE_3(x)	(((x) << 24) & GENMASK(31, 24))
  51#define COMMAND_PORT_SDA_DATA_BYTE_2(x)	(((x) << 16) & GENMASK(23, 16))
  52#define COMMAND_PORT_SDA_DATA_BYTE_1(x)	(((x) << 8) & GENMASK(15, 8))
  53#define COMMAND_PORT_SDA_BYTE_STRB_3	BIT(5)
  54#define COMMAND_PORT_SDA_BYTE_STRB_2	BIT(4)
  55#define COMMAND_PORT_SDA_BYTE_STRB_1	BIT(3)
  56#define COMMAND_PORT_SHORT_DATA_ARG	0x02
  57
  58#define COMMAND_PORT_DEV_COUNT(x)	(((x) << 21) & GENMASK(25, 21))
  59#define COMMAND_PORT_ADDR_ASSGN_CMD	0x03
  60
  61#define RESPONSE_QUEUE_PORT		0x10
  62#define RESPONSE_PORT_ERR_STATUS(x)	(((x) & GENMASK(31, 28)) >> 28)
  63#define RESPONSE_NO_ERROR		0
  64#define RESPONSE_ERROR_CRC		1
  65#define RESPONSE_ERROR_PARITY		2
  66#define RESPONSE_ERROR_FRAME		3
  67#define RESPONSE_ERROR_IBA_NACK		4
  68#define RESPONSE_ERROR_ADDRESS_NACK	5
  69#define RESPONSE_ERROR_OVER_UNDER_FLOW	6
  70#define RESPONSE_ERROR_TRANSF_ABORT	8
  71#define RESPONSE_ERROR_I2C_W_NACK_ERR	9
  72#define RESPONSE_PORT_TID(x)		(((x) & GENMASK(27, 24)) >> 24)
  73#define RESPONSE_PORT_DATA_LEN(x)	((x) & GENMASK(15, 0))
  74
  75#define RX_TX_DATA_PORT			0x14
  76#define IBI_QUEUE_STATUS		0x18
 
 
 
 
 
 
 
 
 
 
 
  77#define QUEUE_THLD_CTRL			0x1c
 
 
 
 
  78#define QUEUE_THLD_CTRL_RESP_BUF_MASK	GENMASK(15, 8)
  79#define QUEUE_THLD_CTRL_RESP_BUF(x)	(((x) - 1) << 8)
  80
  81#define DATA_BUFFER_THLD_CTRL		0x20
  82#define DATA_BUFFER_THLD_CTRL_RX_BUF	GENMASK(11, 8)
  83
  84#define IBI_QUEUE_CTRL			0x24
  85#define IBI_MR_REQ_REJECT		0x2C
  86#define IBI_SIR_REQ_REJECT		0x30
  87#define IBI_REQ_REJECT_ALL		GENMASK(31, 0)
  88
  89#define RESET_CTRL			0x34
  90#define RESET_CTRL_IBI_QUEUE		BIT(5)
  91#define RESET_CTRL_RX_FIFO		BIT(4)
  92#define RESET_CTRL_TX_FIFO		BIT(3)
  93#define RESET_CTRL_RESP_QUEUE		BIT(2)
  94#define RESET_CTRL_CMD_QUEUE		BIT(1)
  95#define RESET_CTRL_SOFT			BIT(0)
  96
  97#define SLV_EVENT_CTRL			0x38
  98#define INTR_STATUS			0x3c
  99#define INTR_STATUS_EN			0x40
 100#define INTR_SIGNAL_EN			0x44
 101#define INTR_FORCE			0x48
 102#define INTR_BUSOWNER_UPDATE_STAT	BIT(13)
 103#define INTR_IBI_UPDATED_STAT		BIT(12)
 104#define INTR_READ_REQ_RECV_STAT		BIT(11)
 105#define INTR_DEFSLV_STAT		BIT(10)
 106#define INTR_TRANSFER_ERR_STAT		BIT(9)
 107#define INTR_DYN_ADDR_ASSGN_STAT	BIT(8)
 108#define INTR_CCC_UPDATED_STAT		BIT(6)
 109#define INTR_TRANSFER_ABORT_STAT	BIT(5)
 110#define INTR_RESP_READY_STAT		BIT(4)
 111#define INTR_CMD_QUEUE_READY_STAT	BIT(3)
 112#define INTR_IBI_THLD_STAT		BIT(2)
 113#define INTR_RX_THLD_STAT		BIT(1)
 114#define INTR_TX_THLD_STAT		BIT(0)
 115#define INTR_ALL			(INTR_BUSOWNER_UPDATE_STAT |	\
 116					INTR_IBI_UPDATED_STAT |		\
 117					INTR_READ_REQ_RECV_STAT |	\
 118					INTR_DEFSLV_STAT |		\
 119					INTR_TRANSFER_ERR_STAT |	\
 120					INTR_DYN_ADDR_ASSGN_STAT |	\
 121					INTR_CCC_UPDATED_STAT |		\
 122					INTR_TRANSFER_ABORT_STAT |	\
 123					INTR_RESP_READY_STAT |		\
 124					INTR_CMD_QUEUE_READY_STAT |	\
 125					INTR_IBI_THLD_STAT |		\
 126					INTR_TX_THLD_STAT |		\
 127					INTR_RX_THLD_STAT)
 128
 129#define INTR_MASTER_MASK		(INTR_TRANSFER_ERR_STAT |	\
 130					 INTR_RESP_READY_STAT)
 131
 132#define QUEUE_STATUS_LEVEL		0x4c
 133#define QUEUE_STATUS_IBI_STATUS_CNT(x)	(((x) & GENMASK(28, 24)) >> 24)
 134#define QUEUE_STATUS_IBI_BUF_BLR(x)	(((x) & GENMASK(23, 16)) >> 16)
 135#define QUEUE_STATUS_LEVEL_RESP(x)	(((x) & GENMASK(15, 8)) >> 8)
 136#define QUEUE_STATUS_LEVEL_CMD(x)	((x) & GENMASK(7, 0))
 137
 138#define DATA_BUFFER_STATUS_LEVEL	0x50
 139#define DATA_BUFFER_STATUS_LEVEL_TX(x)	((x) & GENMASK(7, 0))
 140
 141#define PRESENT_STATE			0x54
 142#define CCC_DEVICE_STATUS		0x58
 143#define DEVICE_ADDR_TABLE_POINTER	0x5c
 144#define DEVICE_ADDR_TABLE_DEPTH(x)	(((x) & GENMASK(31, 16)) >> 16)
 145#define DEVICE_ADDR_TABLE_ADDR(x)	((x) & GENMASK(7, 0))
 146
 147#define DEV_CHAR_TABLE_POINTER		0x60
 148#define VENDOR_SPECIFIC_REG_POINTER	0x6c
 149#define SLV_PID_VALUE			0x74
 150#define SLV_CHAR_CTRL			0x78
 151#define SLV_MAX_LEN			0x7c
 152#define MAX_READ_TURNAROUND		0x80
 153#define MAX_DATA_SPEED			0x84
 154#define SLV_DEBUG_STATUS		0x88
 155#define SLV_INTR_REQ			0x8c
 156#define DEVICE_CTRL_EXTENDED		0xb0
 157#define SCL_I3C_OD_TIMING		0xb4
 158#define SCL_I3C_PP_TIMING		0xb8
 159#define SCL_I3C_TIMING_HCNT(x)		(((x) << 16) & GENMASK(23, 16))
 160#define SCL_I3C_TIMING_LCNT(x)		((x) & GENMASK(7, 0))
 161#define SCL_I3C_TIMING_CNT_MIN		5
 162
 163#define SCL_I2C_FM_TIMING		0xbc
 164#define SCL_I2C_FM_TIMING_HCNT(x)	(((x) << 16) & GENMASK(31, 16))
 165#define SCL_I2C_FM_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 166
 167#define SCL_I2C_FMP_TIMING		0xc0
 168#define SCL_I2C_FMP_TIMING_HCNT(x)	(((x) << 16) & GENMASK(23, 16))
 169#define SCL_I2C_FMP_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 170
 171#define SCL_EXT_LCNT_TIMING		0xc8
 172#define SCL_EXT_LCNT_4(x)		(((x) << 24) & GENMASK(31, 24))
 173#define SCL_EXT_LCNT_3(x)		(((x) << 16) & GENMASK(23, 16))
 174#define SCL_EXT_LCNT_2(x)		(((x) << 8) & GENMASK(15, 8))
 175#define SCL_EXT_LCNT_1(x)		((x) & GENMASK(7, 0))
 176
 177#define SCL_EXT_TERMN_LCNT_TIMING	0xcc
 178#define BUS_FREE_TIMING			0xd4
 179#define BUS_I3C_MST_FREE(x)		((x) & GENMASK(15, 0))
 180
 181#define BUS_IDLE_TIMING			0xd8
 182#define I3C_VER_ID			0xe0
 183#define I3C_VER_TYPE			0xe4
 184#define EXTENDED_CAPABILITY		0xe8
 185#define SLAVE_CONFIG			0xec
 186
 
 
 187#define DEV_ADDR_TABLE_LEGACY_I2C_DEV	BIT(31)
 188#define DEV_ADDR_TABLE_DYNAMIC_ADDR(x)	(((x) << 16) & GENMASK(23, 16))
 189#define DEV_ADDR_TABLE_STATIC_ADDR(x)	((x) & GENMASK(6, 0))
 190#define DEV_ADDR_TABLE_LOC(start, idx)	((start) + ((idx) << 2))
 191
 192#define MAX_DEVS 32
 193
 194#define I3C_BUS_SDR1_SCL_RATE		8000000
 195#define I3C_BUS_SDR2_SCL_RATE		6000000
 196#define I3C_BUS_SDR3_SCL_RATE		4000000
 197#define I3C_BUS_SDR4_SCL_RATE		2000000
 198#define I3C_BUS_I2C_FM_TLOW_MIN_NS	1300
 199#define I3C_BUS_I2C_FMP_TLOW_MIN_NS	500
 200#define I3C_BUS_THIGH_MAX_NS		41
 201
 202#define XFER_TIMEOUT (msecs_to_jiffies(1000))
 203
 204struct dw_i3c_master_caps {
 205	u8 cmdfifodepth;
 206	u8 datafifodepth;
 207};
 208
 209struct dw_i3c_cmd {
 210	u32 cmd_lo;
 211	u32 cmd_hi;
 212	u16 tx_len;
 213	const void *tx_buf;
 214	u16 rx_len;
 215	void *rx_buf;
 216	u8 error;
 217};
 218
 219struct dw_i3c_xfer {
 220	struct list_head node;
 221	struct completion comp;
 222	int ret;
 223	unsigned int ncmds;
 224	struct dw_i3c_cmd cmds[];
 225};
 226
 227struct dw_i3c_master {
 228	struct i3c_master_controller base;
 229	u16 maxdevs;
 230	u16 datstartaddr;
 231	u32 free_pos;
 232	struct {
 233		struct list_head list;
 234		struct dw_i3c_xfer *cur;
 235		spinlock_t lock;
 236	} xferqueue;
 237	struct dw_i3c_master_caps caps;
 238	void __iomem *regs;
 239	struct reset_control *core_rst;
 240	struct clk *core_clk;
 241	char version[5];
 242	char type[5];
 243	u8 addrs[MAX_DEVS];
 244};
 245
 246struct dw_i3c_i2c_dev_data {
 247	u8 index;
 
 248};
 249
 250static u8 even_parity(u8 p)
 251{
 252	p ^= p >> 4;
 253	p &= 0xf;
 254
 255	return (0x9669 >> p) & 1;
 256}
 257
 258static bool dw_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m,
 259					   const struct i3c_ccc_cmd *cmd)
 260{
 261	if (cmd->ndests > 1)
 262		return false;
 263
 264	switch (cmd->id) {
 265	case I3C_CCC_ENEC(true):
 266	case I3C_CCC_ENEC(false):
 267	case I3C_CCC_DISEC(true):
 268	case I3C_CCC_DISEC(false):
 269	case I3C_CCC_ENTAS(0, true):
 270	case I3C_CCC_ENTAS(0, false):
 271	case I3C_CCC_RSTDAA(true):
 272	case I3C_CCC_RSTDAA(false):
 273	case I3C_CCC_ENTDAA:
 274	case I3C_CCC_SETMWL(true):
 275	case I3C_CCC_SETMWL(false):
 276	case I3C_CCC_SETMRL(true):
 277	case I3C_CCC_SETMRL(false):
 278	case I3C_CCC_ENTHDR(0):
 279	case I3C_CCC_SETDASA:
 280	case I3C_CCC_SETNEWDA:
 281	case I3C_CCC_GETMWL:
 282	case I3C_CCC_GETMRL:
 283	case I3C_CCC_GETPID:
 284	case I3C_CCC_GETBCR:
 285	case I3C_CCC_GETDCR:
 286	case I3C_CCC_GETSTATUS:
 287	case I3C_CCC_GETMXDS:
 288	case I3C_CCC_GETHDRCAP:
 289		return true;
 290	default:
 291		return false;
 292	}
 293}
 294
 295static inline struct dw_i3c_master *
 296to_dw_i3c_master(struct i3c_master_controller *master)
 297{
 298	return container_of(master, struct dw_i3c_master, base);
 299}
 300
 301static void dw_i3c_master_disable(struct dw_i3c_master *master)
 302{
 303	writel(readl(master->regs + DEVICE_CTRL) & ~DEV_CTRL_ENABLE,
 304	       master->regs + DEVICE_CTRL);
 305}
 306
 307static void dw_i3c_master_enable(struct dw_i3c_master *master)
 308{
 309	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_ENABLE,
 310	       master->regs + DEVICE_CTRL);
 311}
 312
 313static int dw_i3c_master_get_addr_pos(struct dw_i3c_master *master, u8 addr)
 314{
 315	int pos;
 316
 317	for (pos = 0; pos < master->maxdevs; pos++) {
 318		if (addr == master->addrs[pos])
 319			return pos;
 320	}
 321
 322	return -EINVAL;
 323}
 324
 325static int dw_i3c_master_get_free_pos(struct dw_i3c_master *master)
 326{
 327	if (!(master->free_pos & GENMASK(master->maxdevs - 1, 0)))
 328		return -ENOSPC;
 329
 330	return ffs(master->free_pos) - 1;
 331}
 332
 333static void dw_i3c_master_wr_tx_fifo(struct dw_i3c_master *master,
 334				     const u8 *bytes, int nbytes)
 335{
 336	writesl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 337	if (nbytes & 3) {
 338		u32 tmp = 0;
 339
 340		memcpy(&tmp, bytes + (nbytes & ~3), nbytes & 3);
 341		writesl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 342	}
 343}
 344
 345static void dw_i3c_master_read_rx_fifo(struct dw_i3c_master *master,
 346				       u8 *bytes, int nbytes)
 347{
 348	readsl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 349	if (nbytes & 3) {
 350		u32 tmp;
 351
 352		readsl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 353		memcpy(bytes + (nbytes & ~3), &tmp, nbytes & 3);
 354	}
 355}
 356
 
 
 
 
 
 
 
 
 
 
 
 
 357static struct dw_i3c_xfer *
 358dw_i3c_master_alloc_xfer(struct dw_i3c_master *master, unsigned int ncmds)
 359{
 360	struct dw_i3c_xfer *xfer;
 361
 362	xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL);
 363	if (!xfer)
 364		return NULL;
 365
 366	INIT_LIST_HEAD(&xfer->node);
 367	xfer->ncmds = ncmds;
 368	xfer->ret = -ETIMEDOUT;
 369
 370	return xfer;
 371}
 372
 373static void dw_i3c_master_free_xfer(struct dw_i3c_xfer *xfer)
 374{
 375	kfree(xfer);
 376}
 377
 378static void dw_i3c_master_start_xfer_locked(struct dw_i3c_master *master)
 379{
 380	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 381	unsigned int i;
 382	u32 thld_ctrl;
 383
 384	if (!xfer)
 385		return;
 386
 387	for (i = 0; i < xfer->ncmds; i++) {
 388		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 389
 390		dw_i3c_master_wr_tx_fifo(master, cmd->tx_buf, cmd->tx_len);
 391	}
 392
 393	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 394	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 395	thld_ctrl |= QUEUE_THLD_CTRL_RESP_BUF(xfer->ncmds);
 396	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 397
 398	for (i = 0; i < xfer->ncmds; i++) {
 399		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 400
 401		writel(cmd->cmd_hi, master->regs + COMMAND_QUEUE_PORT);
 402		writel(cmd->cmd_lo, master->regs + COMMAND_QUEUE_PORT);
 403	}
 404}
 405
 406static void dw_i3c_master_enqueue_xfer(struct dw_i3c_master *master,
 407				       struct dw_i3c_xfer *xfer)
 408{
 409	unsigned long flags;
 410
 411	init_completion(&xfer->comp);
 412	spin_lock_irqsave(&master->xferqueue.lock, flags);
 413	if (master->xferqueue.cur) {
 414		list_add_tail(&xfer->node, &master->xferqueue.list);
 415	} else {
 416		master->xferqueue.cur = xfer;
 417		dw_i3c_master_start_xfer_locked(master);
 418	}
 419	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 420}
 421
 422static void dw_i3c_master_dequeue_xfer_locked(struct dw_i3c_master *master,
 423					      struct dw_i3c_xfer *xfer)
 424{
 425	if (master->xferqueue.cur == xfer) {
 426		u32 status;
 427
 428		master->xferqueue.cur = NULL;
 429
 430		writel(RESET_CTRL_RX_FIFO | RESET_CTRL_TX_FIFO |
 431		       RESET_CTRL_RESP_QUEUE | RESET_CTRL_CMD_QUEUE,
 432		       master->regs + RESET_CTRL);
 433
 434		readl_poll_timeout_atomic(master->regs + RESET_CTRL, status,
 435					  !status, 10, 1000000);
 436	} else {
 437		list_del_init(&xfer->node);
 438	}
 439}
 440
 441static void dw_i3c_master_dequeue_xfer(struct dw_i3c_master *master,
 442				       struct dw_i3c_xfer *xfer)
 443{
 444	unsigned long flags;
 445
 446	spin_lock_irqsave(&master->xferqueue.lock, flags);
 447	dw_i3c_master_dequeue_xfer_locked(master, xfer);
 448	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 449}
 450
 451static void dw_i3c_master_end_xfer_locked(struct dw_i3c_master *master, u32 isr)
 452{
 453	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 454	int i, ret = 0;
 455	u32 nresp;
 456
 457	if (!xfer)
 458		return;
 459
 460	nresp = readl(master->regs + QUEUE_STATUS_LEVEL);
 461	nresp = QUEUE_STATUS_LEVEL_RESP(nresp);
 462
 463	for (i = 0; i < nresp; i++) {
 464		struct dw_i3c_cmd *cmd;
 465		u32 resp;
 466
 467		resp = readl(master->regs + RESPONSE_QUEUE_PORT);
 468
 469		cmd = &xfer->cmds[RESPONSE_PORT_TID(resp)];
 470		cmd->rx_len = RESPONSE_PORT_DATA_LEN(resp);
 471		cmd->error = RESPONSE_PORT_ERR_STATUS(resp);
 472		if (cmd->rx_len && !cmd->error)
 473			dw_i3c_master_read_rx_fifo(master, cmd->rx_buf,
 474						   cmd->rx_len);
 475	}
 476
 477	for (i = 0; i < nresp; i++) {
 478		switch (xfer->cmds[i].error) {
 479		case RESPONSE_NO_ERROR:
 480			break;
 481		case RESPONSE_ERROR_PARITY:
 482		case RESPONSE_ERROR_IBA_NACK:
 483		case RESPONSE_ERROR_TRANSF_ABORT:
 484		case RESPONSE_ERROR_CRC:
 485		case RESPONSE_ERROR_FRAME:
 486			ret = -EIO;
 487			break;
 488		case RESPONSE_ERROR_OVER_UNDER_FLOW:
 489			ret = -ENOSPC;
 490			break;
 491		case RESPONSE_ERROR_I2C_W_NACK_ERR:
 492		case RESPONSE_ERROR_ADDRESS_NACK:
 493		default:
 494			ret = -EINVAL;
 495			break;
 496		}
 497	}
 498
 499	xfer->ret = ret;
 500	complete(&xfer->comp);
 501
 502	if (ret < 0) {
 503		dw_i3c_master_dequeue_xfer_locked(master, xfer);
 504		writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_RESUME,
 505		       master->regs + DEVICE_CTRL);
 506	}
 507
 508	xfer = list_first_entry_or_null(&master->xferqueue.list,
 509					struct dw_i3c_xfer,
 510					node);
 511	if (xfer)
 512		list_del_init(&xfer->node);
 513
 514	master->xferqueue.cur = xfer;
 515	dw_i3c_master_start_xfer_locked(master);
 516}
 517
 518static int dw_i3c_clk_cfg(struct dw_i3c_master *master)
 519{
 520	unsigned long core_rate, core_period;
 521	u32 scl_timing;
 522	u8 hcnt, lcnt;
 523
 524	core_rate = clk_get_rate(master->core_clk);
 525	if (!core_rate)
 526		return -EINVAL;
 527
 528	core_period = DIV_ROUND_UP(1000000000, core_rate);
 529
 530	hcnt = DIV_ROUND_UP(I3C_BUS_THIGH_MAX_NS, core_period) - 1;
 531	if (hcnt < SCL_I3C_TIMING_CNT_MIN)
 532		hcnt = SCL_I3C_TIMING_CNT_MIN;
 533
 534	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_TYP_I3C_SCL_RATE) - hcnt;
 535	if (lcnt < SCL_I3C_TIMING_CNT_MIN)
 536		lcnt = SCL_I3C_TIMING_CNT_MIN;
 537
 538	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 539	writel(scl_timing, master->regs + SCL_I3C_PP_TIMING);
 540
 541	if (!(readl(master->regs + DEVICE_CTRL) & DEV_CTRL_I2C_SLAVE_PRESENT))
 
 
 
 
 542		writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 543
 544	lcnt = DIV_ROUND_UP(I3C_BUS_TLOW_OD_MIN_NS, core_period);
 
 545	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 546	writel(scl_timing, master->regs + SCL_I3C_OD_TIMING);
 547
 548	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR1_SCL_RATE) - hcnt;
 549	scl_timing = SCL_EXT_LCNT_1(lcnt);
 550	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR2_SCL_RATE) - hcnt;
 551	scl_timing |= SCL_EXT_LCNT_2(lcnt);
 552	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR3_SCL_RATE) - hcnt;
 553	scl_timing |= SCL_EXT_LCNT_3(lcnt);
 554	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR4_SCL_RATE) - hcnt;
 555	scl_timing |= SCL_EXT_LCNT_4(lcnt);
 556	writel(scl_timing, master->regs + SCL_EXT_LCNT_TIMING);
 557
 558	return 0;
 559}
 560
 561static int dw_i2c_clk_cfg(struct dw_i3c_master *master)
 562{
 563	unsigned long core_rate, core_period;
 564	u16 hcnt, lcnt;
 565	u32 scl_timing;
 566
 567	core_rate = clk_get_rate(master->core_clk);
 568	if (!core_rate)
 569		return -EINVAL;
 570
 571	core_period = DIV_ROUND_UP(1000000000, core_rate);
 572
 573	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FMP_TLOW_MIN_NS, core_period);
 574	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_PLUS_SCL_RATE) - lcnt;
 575	scl_timing = SCL_I2C_FMP_TIMING_HCNT(hcnt) |
 576		     SCL_I2C_FMP_TIMING_LCNT(lcnt);
 577	writel(scl_timing, master->regs + SCL_I2C_FMP_TIMING);
 578
 579	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FM_TLOW_MIN_NS, core_period);
 580	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_SCL_RATE) - lcnt;
 581	scl_timing = SCL_I2C_FM_TIMING_HCNT(hcnt) |
 582		     SCL_I2C_FM_TIMING_LCNT(lcnt);
 583	writel(scl_timing, master->regs + SCL_I2C_FM_TIMING);
 584
 585	writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 586	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_I2C_SLAVE_PRESENT,
 587	       master->regs + DEVICE_CTRL);
 588
 589	return 0;
 590}
 591
 592static int dw_i3c_master_bus_init(struct i3c_master_controller *m)
 593{
 594	struct dw_i3c_master *master = to_dw_i3c_master(m);
 595	struct i3c_bus *bus = i3c_master_get_bus(m);
 596	struct i3c_device_info info = { };
 597	u32 thld_ctrl;
 598	int ret;
 599
 
 
 
 
 600	switch (bus->mode) {
 601	case I3C_BUS_MODE_MIXED_FAST:
 602	case I3C_BUS_MODE_MIXED_LIMITED:
 603		ret = dw_i2c_clk_cfg(master);
 604		if (ret)
 605			return ret;
 606		fallthrough;
 607	case I3C_BUS_MODE_PURE:
 608		ret = dw_i3c_clk_cfg(master);
 609		if (ret)
 610			return ret;
 611		break;
 612	default:
 613		return -EINVAL;
 614	}
 615
 616	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 617	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 
 
 
 
 618	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 619
 620	thld_ctrl = readl(master->regs + DATA_BUFFER_THLD_CTRL);
 621	thld_ctrl &= ~DATA_BUFFER_THLD_CTRL_RX_BUF;
 622	writel(thld_ctrl, master->regs + DATA_BUFFER_THLD_CTRL);
 623
 624	writel(INTR_ALL, master->regs + INTR_STATUS);
 625	writel(INTR_MASTER_MASK, master->regs + INTR_STATUS_EN);
 626	writel(INTR_MASTER_MASK, master->regs + INTR_SIGNAL_EN);
 627
 628	ret = i3c_master_get_free_addr(m, 0);
 629	if (ret < 0)
 630		return ret;
 631
 632	writel(DEV_ADDR_DYNAMIC_ADDR_VALID | DEV_ADDR_DYNAMIC(ret),
 633	       master->regs + DEVICE_ADDR);
 634
 635	memset(&info, 0, sizeof(info));
 636	info.dyn_addr = ret;
 637
 638	ret = i3c_master_set_info(&master->base, &info);
 639	if (ret)
 640		return ret;
 641
 642	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_SIR_REQ_REJECT);
 643	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_MR_REQ_REJECT);
 644
 645	/* For now don't support Hot-Join */
 646	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_HOT_JOIN_NACK,
 647	       master->regs + DEVICE_CTRL);
 648
 649	dw_i3c_master_enable(master);
 650
 651	return 0;
 652}
 653
 654static void dw_i3c_master_bus_cleanup(struct i3c_master_controller *m)
 655{
 656	struct dw_i3c_master *master = to_dw_i3c_master(m);
 657
 658	dw_i3c_master_disable(master);
 659}
 660
 661static int dw_i3c_ccc_set(struct dw_i3c_master *master,
 662			  struct i3c_ccc_cmd *ccc)
 663{
 664	struct dw_i3c_xfer *xfer;
 665	struct dw_i3c_cmd *cmd;
 666	int ret, pos = 0;
 667
 668	if (ccc->id & I3C_CCC_DIRECT) {
 669		pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 670		if (pos < 0)
 671			return pos;
 672	}
 673
 674	xfer = dw_i3c_master_alloc_xfer(master, 1);
 675	if (!xfer)
 676		return -ENOMEM;
 677
 678	cmd = xfer->cmds;
 679	cmd->tx_buf = ccc->dests[0].payload.data;
 680	cmd->tx_len = ccc->dests[0].payload.len;
 681
 682	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 683		      COMMAND_PORT_TRANSFER_ARG;
 684
 685	cmd->cmd_lo = COMMAND_PORT_CP |
 686		      COMMAND_PORT_DEV_INDEX(pos) |
 687		      COMMAND_PORT_CMD(ccc->id) |
 688		      COMMAND_PORT_TOC |
 689		      COMMAND_PORT_ROC;
 690
 691	dw_i3c_master_enqueue_xfer(master, xfer);
 692	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 693		dw_i3c_master_dequeue_xfer(master, xfer);
 694
 695	ret = xfer->ret;
 696	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 697		ccc->err = I3C_ERROR_M2;
 698
 699	dw_i3c_master_free_xfer(xfer);
 700
 701	return ret;
 702}
 703
 704static int dw_i3c_ccc_get(struct dw_i3c_master *master, struct i3c_ccc_cmd *ccc)
 705{
 706	struct dw_i3c_xfer *xfer;
 707	struct dw_i3c_cmd *cmd;
 708	int ret, pos;
 709
 710	pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 711	if (pos < 0)
 712		return pos;
 713
 714	xfer = dw_i3c_master_alloc_xfer(master, 1);
 715	if (!xfer)
 716		return -ENOMEM;
 717
 718	cmd = xfer->cmds;
 719	cmd->rx_buf = ccc->dests[0].payload.data;
 720	cmd->rx_len = ccc->dests[0].payload.len;
 721
 722	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 723		      COMMAND_PORT_TRANSFER_ARG;
 724
 725	cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 726		      COMMAND_PORT_CP |
 727		      COMMAND_PORT_DEV_INDEX(pos) |
 728		      COMMAND_PORT_CMD(ccc->id) |
 729		      COMMAND_PORT_TOC |
 730		      COMMAND_PORT_ROC;
 731
 732	dw_i3c_master_enqueue_xfer(master, xfer);
 733	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 734		dw_i3c_master_dequeue_xfer(master, xfer);
 735
 736	ret = xfer->ret;
 737	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 738		ccc->err = I3C_ERROR_M2;
 739	dw_i3c_master_free_xfer(xfer);
 740
 741	return ret;
 742}
 743
 744static int dw_i3c_master_send_ccc_cmd(struct i3c_master_controller *m,
 745				      struct i3c_ccc_cmd *ccc)
 746{
 747	struct dw_i3c_master *master = to_dw_i3c_master(m);
 748	int ret = 0;
 749
 750	if (ccc->id == I3C_CCC_ENTDAA)
 751		return -EINVAL;
 752
 753	if (ccc->rnw)
 754		ret = dw_i3c_ccc_get(master, ccc);
 755	else
 756		ret = dw_i3c_ccc_set(master, ccc);
 757
 758	return ret;
 759}
 760
 761static int dw_i3c_master_daa(struct i3c_master_controller *m)
 762{
 763	struct dw_i3c_master *master = to_dw_i3c_master(m);
 764	struct dw_i3c_xfer *xfer;
 765	struct dw_i3c_cmd *cmd;
 766	u32 olddevs, newdevs;
 767	u8 p, last_addr = 0;
 768	int ret, pos;
 769
 770	olddevs = ~(master->free_pos);
 771
 772	/* Prepare DAT before launching DAA. */
 773	for (pos = 0; pos < master->maxdevs; pos++) {
 774		if (olddevs & BIT(pos))
 775			continue;
 776
 777		ret = i3c_master_get_free_addr(m, last_addr + 1);
 778		if (ret < 0)
 779			return -ENOSPC;
 780
 781		master->addrs[pos] = ret;
 782		p = even_parity(ret);
 783		last_addr = ret;
 784		ret |= (p << 7);
 785
 786		writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(ret),
 787		       master->regs +
 788		       DEV_ADDR_TABLE_LOC(master->datstartaddr, pos));
 789	}
 790
 791	xfer = dw_i3c_master_alloc_xfer(master, 1);
 792	if (!xfer)
 793		return -ENOMEM;
 794
 795	pos = dw_i3c_master_get_free_pos(master);
 796	if (pos < 0) {
 797		dw_i3c_master_free_xfer(xfer);
 798		return pos;
 799	}
 800	cmd = &xfer->cmds[0];
 801	cmd->cmd_hi = 0x1;
 802	cmd->cmd_lo = COMMAND_PORT_DEV_COUNT(master->maxdevs - pos) |
 803		      COMMAND_PORT_DEV_INDEX(pos) |
 804		      COMMAND_PORT_CMD(I3C_CCC_ENTDAA) |
 805		      COMMAND_PORT_ADDR_ASSGN_CMD |
 806		      COMMAND_PORT_TOC |
 807		      COMMAND_PORT_ROC;
 808
 809	dw_i3c_master_enqueue_xfer(master, xfer);
 810	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 811		dw_i3c_master_dequeue_xfer(master, xfer);
 812
 813	newdevs = GENMASK(master->maxdevs - cmd->rx_len - 1, 0);
 814	newdevs &= ~olddevs;
 815
 816	for (pos = 0; pos < master->maxdevs; pos++) {
 817		if (newdevs & BIT(pos))
 818			i3c_master_add_i3c_dev_locked(m, master->addrs[pos]);
 819	}
 820
 821	dw_i3c_master_free_xfer(xfer);
 822
 823	return 0;
 824}
 825
 826static int dw_i3c_master_priv_xfers(struct i3c_dev_desc *dev,
 827				    struct i3c_priv_xfer *i3c_xfers,
 828				    int i3c_nxfers)
 829{
 830	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 831	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 832	struct dw_i3c_master *master = to_dw_i3c_master(m);
 833	unsigned int nrxwords = 0, ntxwords = 0;
 834	struct dw_i3c_xfer *xfer;
 835	int i, ret = 0;
 836
 837	if (!i3c_nxfers)
 838		return 0;
 839
 840	if (i3c_nxfers > master->caps.cmdfifodepth)
 841		return -ENOTSUPP;
 842
 843	for (i = 0; i < i3c_nxfers; i++) {
 844		if (i3c_xfers[i].rnw)
 845			nrxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 846		else
 847			ntxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 848	}
 849
 850	if (ntxwords > master->caps.datafifodepth ||
 851	    nrxwords > master->caps.datafifodepth)
 852		return -ENOTSUPP;
 853
 854	xfer = dw_i3c_master_alloc_xfer(master, i3c_nxfers);
 855	if (!xfer)
 856		return -ENOMEM;
 857
 858	for (i = 0; i < i3c_nxfers; i++) {
 859		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 860
 861		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i3c_xfers[i].len) |
 862			COMMAND_PORT_TRANSFER_ARG;
 863
 864		if (i3c_xfers[i].rnw) {
 865			cmd->rx_buf = i3c_xfers[i].data.in;
 866			cmd->rx_len = i3c_xfers[i].len;
 867			cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 868				      COMMAND_PORT_SPEED(dev->info.max_read_ds);
 869
 870		} else {
 871			cmd->tx_buf = i3c_xfers[i].data.out;
 872			cmd->tx_len = i3c_xfers[i].len;
 873			cmd->cmd_lo =
 874				COMMAND_PORT_SPEED(dev->info.max_write_ds);
 875		}
 876
 877		cmd->cmd_lo |= COMMAND_PORT_TID(i) |
 878			       COMMAND_PORT_DEV_INDEX(data->index) |
 879			       COMMAND_PORT_ROC;
 880
 881		if (i == (i3c_nxfers - 1))
 882			cmd->cmd_lo |= COMMAND_PORT_TOC;
 883	}
 884
 885	dw_i3c_master_enqueue_xfer(master, xfer);
 886	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 887		dw_i3c_master_dequeue_xfer(master, xfer);
 888
 
 
 
 
 
 
 
 889	ret = xfer->ret;
 890	dw_i3c_master_free_xfer(xfer);
 891
 892	return ret;
 893}
 894
 895static int dw_i3c_master_reattach_i3c_dev(struct i3c_dev_desc *dev,
 896					  u8 old_dyn_addr)
 897{
 898	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 899	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 900	struct dw_i3c_master *master = to_dw_i3c_master(m);
 901	int pos;
 902
 903	pos = dw_i3c_master_get_free_pos(master);
 904
 905	if (data->index > pos && pos > 0) {
 906		writel(0,
 907		       master->regs +
 908		       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 909
 910		master->addrs[data->index] = 0;
 911		master->free_pos |= BIT(data->index);
 912
 913		data->index = pos;
 914		master->addrs[pos] = dev->info.dyn_addr;
 915		master->free_pos &= ~BIT(pos);
 916	}
 917
 918	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(dev->info.dyn_addr),
 919	       master->regs +
 920	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 921
 922	master->addrs[data->index] = dev->info.dyn_addr;
 923
 924	return 0;
 925}
 926
 927static int dw_i3c_master_attach_i3c_dev(struct i3c_dev_desc *dev)
 928{
 929	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 930	struct dw_i3c_master *master = to_dw_i3c_master(m);
 931	struct dw_i3c_i2c_dev_data *data;
 932	int pos;
 933
 934	pos = dw_i3c_master_get_free_pos(master);
 935	if (pos < 0)
 936		return pos;
 937
 938	data = kzalloc(sizeof(*data), GFP_KERNEL);
 939	if (!data)
 940		return -ENOMEM;
 941
 942	data->index = pos;
 943	master->addrs[pos] = dev->info.dyn_addr ? : dev->info.static_addr;
 944	master->free_pos &= ~BIT(pos);
 945	i3c_dev_set_master_data(dev, data);
 946
 947	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(master->addrs[pos]),
 948	       master->regs +
 949	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 950
 951	return 0;
 952}
 953
 954static void dw_i3c_master_detach_i3c_dev(struct i3c_dev_desc *dev)
 955{
 956	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 957	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 958	struct dw_i3c_master *master = to_dw_i3c_master(m);
 959
 960	writel(0,
 961	       master->regs +
 962	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 963
 964	i3c_dev_set_master_data(dev, NULL);
 965	master->addrs[data->index] = 0;
 966	master->free_pos |= BIT(data->index);
 967	kfree(data);
 968}
 969
 970static int dw_i3c_master_i2c_xfers(struct i2c_dev_desc *dev,
 971				   const struct i2c_msg *i2c_xfers,
 972				   int i2c_nxfers)
 973{
 974	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
 975	struct i3c_master_controller *m = i2c_dev_get_master(dev);
 976	struct dw_i3c_master *master = to_dw_i3c_master(m);
 977	unsigned int nrxwords = 0, ntxwords = 0;
 978	struct dw_i3c_xfer *xfer;
 979	int i, ret = 0;
 980
 981	if (!i2c_nxfers)
 982		return 0;
 983
 984	if (i2c_nxfers > master->caps.cmdfifodepth)
 985		return -ENOTSUPP;
 986
 987	for (i = 0; i < i2c_nxfers; i++) {
 988		if (i2c_xfers[i].flags & I2C_M_RD)
 989			nrxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
 990		else
 991			ntxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
 992	}
 993
 994	if (ntxwords > master->caps.datafifodepth ||
 995	    nrxwords > master->caps.datafifodepth)
 996		return -ENOTSUPP;
 997
 998	xfer = dw_i3c_master_alloc_xfer(master, i2c_nxfers);
 999	if (!xfer)
1000		return -ENOMEM;
1001
1002	for (i = 0; i < i2c_nxfers; i++) {
1003		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
1004
1005		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i2c_xfers[i].len) |
1006			COMMAND_PORT_TRANSFER_ARG;
1007
1008		cmd->cmd_lo = COMMAND_PORT_TID(i) |
1009			      COMMAND_PORT_DEV_INDEX(data->index) |
1010			      COMMAND_PORT_ROC;
1011
1012		if (i2c_xfers[i].flags & I2C_M_RD) {
1013			cmd->cmd_lo |= COMMAND_PORT_READ_TRANSFER;
1014			cmd->rx_buf = i2c_xfers[i].buf;
1015			cmd->rx_len = i2c_xfers[i].len;
1016		} else {
1017			cmd->tx_buf = i2c_xfers[i].buf;
1018			cmd->tx_len = i2c_xfers[i].len;
1019		}
1020
1021		if (i == (i2c_nxfers - 1))
1022			cmd->cmd_lo |= COMMAND_PORT_TOC;
1023	}
1024
1025	dw_i3c_master_enqueue_xfer(master, xfer);
1026	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
1027		dw_i3c_master_dequeue_xfer(master, xfer);
1028
1029	ret = xfer->ret;
1030	dw_i3c_master_free_xfer(xfer);
1031
1032	return ret;
1033}
1034
1035static int dw_i3c_master_attach_i2c_dev(struct i2c_dev_desc *dev)
1036{
1037	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1038	struct dw_i3c_master *master = to_dw_i3c_master(m);
1039	struct dw_i3c_i2c_dev_data *data;
1040	int pos;
1041
1042	pos = dw_i3c_master_get_free_pos(master);
1043	if (pos < 0)
1044		return pos;
1045
1046	data = kzalloc(sizeof(*data), GFP_KERNEL);
1047	if (!data)
1048		return -ENOMEM;
1049
1050	data->index = pos;
1051	master->addrs[pos] = dev->addr;
1052	master->free_pos &= ~BIT(pos);
1053	i2c_dev_set_master_data(dev, data);
1054
1055	writel(DEV_ADDR_TABLE_LEGACY_I2C_DEV |
1056	       DEV_ADDR_TABLE_STATIC_ADDR(dev->addr),
1057	       master->regs +
1058	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1059
1060	return 0;
1061}
1062
1063static void dw_i3c_master_detach_i2c_dev(struct i2c_dev_desc *dev)
1064{
1065	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1066	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1067	struct dw_i3c_master *master = to_dw_i3c_master(m);
1068
1069	writel(0,
1070	       master->regs +
1071	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1072
1073	i2c_dev_set_master_data(dev, NULL);
1074	master->addrs[data->index] = 0;
1075	master->free_pos |= BIT(data->index);
1076	kfree(data);
1077}
1078
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1079static irqreturn_t dw_i3c_master_irq_handler(int irq, void *dev_id)
1080{
1081	struct dw_i3c_master *master = dev_id;
1082	u32 status;
1083
1084	status = readl(master->regs + INTR_STATUS);
1085
1086	if (!(status & readl(master->regs + INTR_STATUS_EN))) {
1087		writel(INTR_ALL, master->regs + INTR_STATUS);
1088		return IRQ_NONE;
1089	}
1090
1091	spin_lock(&master->xferqueue.lock);
1092	dw_i3c_master_end_xfer_locked(master, status);
1093	if (status & INTR_TRANSFER_ERR_STAT)
1094		writel(INTR_TRANSFER_ERR_STAT, master->regs + INTR_STATUS);
1095	spin_unlock(&master->xferqueue.lock);
1096
 
 
 
1097	return IRQ_HANDLED;
1098}
1099
1100static const struct i3c_master_controller_ops dw_mipi_i3c_ops = {
1101	.bus_init = dw_i3c_master_bus_init,
1102	.bus_cleanup = dw_i3c_master_bus_cleanup,
1103	.attach_i3c_dev = dw_i3c_master_attach_i3c_dev,
1104	.reattach_i3c_dev = dw_i3c_master_reattach_i3c_dev,
1105	.detach_i3c_dev = dw_i3c_master_detach_i3c_dev,
1106	.do_daa = dw_i3c_master_daa,
1107	.supports_ccc_cmd = dw_i3c_master_supports_ccc_cmd,
1108	.send_ccc_cmd = dw_i3c_master_send_ccc_cmd,
1109	.priv_xfers = dw_i3c_master_priv_xfers,
1110	.attach_i2c_dev = dw_i3c_master_attach_i2c_dev,
1111	.detach_i2c_dev = dw_i3c_master_detach_i2c_dev,
1112	.i2c_xfers = dw_i3c_master_i2c_xfers,
1113};
1114
1115static int dw_i3c_probe(struct platform_device *pdev)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1116{
1117	struct dw_i3c_master *master;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1118	int ret, irq;
1119
1120	master = devm_kzalloc(&pdev->dev, sizeof(*master), GFP_KERNEL);
1121	if (!master)
1122		return -ENOMEM;
1123
1124	master->regs = devm_platform_ioremap_resource(pdev, 0);
1125	if (IS_ERR(master->regs))
1126		return PTR_ERR(master->regs);
1127
1128	master->core_clk = devm_clk_get(&pdev->dev, NULL);
1129	if (IS_ERR(master->core_clk))
1130		return PTR_ERR(master->core_clk);
1131
1132	master->core_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
1133								    "core_rst");
1134	if (IS_ERR(master->core_rst))
1135		return PTR_ERR(master->core_rst);
1136
1137	ret = clk_prepare_enable(master->core_clk);
1138	if (ret)
1139		goto err_disable_core_clk;
1140
1141	reset_control_deassert(master->core_rst);
1142
1143	spin_lock_init(&master->xferqueue.lock);
1144	INIT_LIST_HEAD(&master->xferqueue.list);
1145
1146	writel(INTR_ALL, master->regs + INTR_STATUS);
1147	irq = platform_get_irq(pdev, 0);
1148	ret = devm_request_irq(&pdev->dev, irq,
1149			       dw_i3c_master_irq_handler, 0,
1150			       dev_name(&pdev->dev), master);
1151	if (ret)
1152		goto err_assert_rst;
1153
1154	platform_set_drvdata(pdev, master);
1155
1156	/* Information regarding the FIFOs/QUEUEs depth */
1157	ret = readl(master->regs + QUEUE_STATUS_LEVEL);
1158	master->caps.cmdfifodepth = QUEUE_STATUS_LEVEL_CMD(ret);
1159
1160	ret = readl(master->regs + DATA_BUFFER_STATUS_LEVEL);
1161	master->caps.datafifodepth = DATA_BUFFER_STATUS_LEVEL_TX(ret);
1162
1163	ret = readl(master->regs + DEVICE_ADDR_TABLE_POINTER);
1164	master->datstartaddr = ret;
1165	master->maxdevs = ret >> 16;
1166	master->free_pos = GENMASK(master->maxdevs - 1, 0);
1167
1168	ret = i3c_master_register(&master->base, &pdev->dev,
1169				  &dw_mipi_i3c_ops, false);
 
 
 
1170	if (ret)
1171		goto err_assert_rst;
1172
1173	return 0;
1174
1175err_assert_rst:
1176	reset_control_assert(master->core_rst);
1177
1178err_disable_core_clk:
1179	clk_disable_unprepare(master->core_clk);
1180
1181	return ret;
1182}
 
1183
1184static int dw_i3c_remove(struct platform_device *pdev)
1185{
1186	struct dw_i3c_master *master = platform_get_drvdata(pdev);
1187	int ret;
1188
1189	ret = i3c_master_unregister(&master->base);
1190	if (ret)
1191		return ret;
1192
1193	reset_control_assert(master->core_rst);
1194
1195	clk_disable_unprepare(master->core_clk);
 
 
1196
1197	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1198}
1199
1200static const struct of_device_id dw_i3c_master_of_match[] = {
1201	{ .compatible = "snps,dw-i3c-master-1.00a", },
1202	{},
1203};
1204MODULE_DEVICE_TABLE(of, dw_i3c_master_of_match);
1205
1206static struct platform_driver dw_i3c_driver = {
1207	.probe = dw_i3c_probe,
1208	.remove = dw_i3c_remove,
1209	.driver = {
1210		.name = "dw-i3c-master",
1211		.of_match_table = of_match_ptr(dw_i3c_master_of_match),
1212	},
1213};
1214module_platform_driver(dw_i3c_driver);
1215
1216MODULE_AUTHOR("Vitor Soares <vitor.soares@synopsys.com>");
1217MODULE_DESCRIPTION("DesignWare MIPI I3C driver");
1218MODULE_LICENSE("GPL v2");
v6.8
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
   4 *
   5 * Author: Vitor Soares <vitor.soares@synopsys.com>
   6 */
   7
   8#include <linux/bitops.h>
   9#include <linux/clk.h>
  10#include <linux/completion.h>
  11#include <linux/err.h>
  12#include <linux/errno.h>
  13#include <linux/i3c/master.h>
  14#include <linux/interrupt.h>
  15#include <linux/ioport.h>
  16#include <linux/iopoll.h>
  17#include <linux/list.h>
  18#include <linux/module.h>
  19#include <linux/of.h>
  20#include <linux/platform_device.h>
  21#include <linux/reset.h>
  22#include <linux/slab.h>
  23
  24#include "dw-i3c-master.h"
  25
  26#define DEVICE_CTRL			0x0
  27#define DEV_CTRL_ENABLE			BIT(31)
  28#define DEV_CTRL_RESUME			BIT(30)
  29#define DEV_CTRL_HOT_JOIN_NACK		BIT(8)
  30#define DEV_CTRL_I2C_SLAVE_PRESENT	BIT(7)
  31
  32#define DEVICE_ADDR			0x4
  33#define DEV_ADDR_DYNAMIC_ADDR_VALID	BIT(31)
  34#define DEV_ADDR_DYNAMIC(x)		(((x) << 16) & GENMASK(22, 16))
  35
  36#define HW_CAPABILITY			0x8
  37#define COMMAND_QUEUE_PORT		0xc
  38#define COMMAND_PORT_TOC		BIT(30)
  39#define COMMAND_PORT_READ_TRANSFER	BIT(28)
  40#define COMMAND_PORT_SDAP		BIT(27)
  41#define COMMAND_PORT_ROC		BIT(26)
  42#define COMMAND_PORT_SPEED(x)		(((x) << 21) & GENMASK(23, 21))
  43#define COMMAND_PORT_DEV_INDEX(x)	(((x) << 16) & GENMASK(20, 16))
  44#define COMMAND_PORT_CP			BIT(15)
  45#define COMMAND_PORT_CMD(x)		(((x) << 7) & GENMASK(14, 7))
  46#define COMMAND_PORT_TID(x)		(((x) << 3) & GENMASK(6, 3))
  47
  48#define COMMAND_PORT_ARG_DATA_LEN(x)	(((x) << 16) & GENMASK(31, 16))
  49#define COMMAND_PORT_ARG_DATA_LEN_MAX	65536
  50#define COMMAND_PORT_TRANSFER_ARG	0x01
  51
  52#define COMMAND_PORT_SDA_DATA_BYTE_3(x)	(((x) << 24) & GENMASK(31, 24))
  53#define COMMAND_PORT_SDA_DATA_BYTE_2(x)	(((x) << 16) & GENMASK(23, 16))
  54#define COMMAND_PORT_SDA_DATA_BYTE_1(x)	(((x) << 8) & GENMASK(15, 8))
  55#define COMMAND_PORT_SDA_BYTE_STRB_3	BIT(5)
  56#define COMMAND_PORT_SDA_BYTE_STRB_2	BIT(4)
  57#define COMMAND_PORT_SDA_BYTE_STRB_1	BIT(3)
  58#define COMMAND_PORT_SHORT_DATA_ARG	0x02
  59
  60#define COMMAND_PORT_DEV_COUNT(x)	(((x) << 21) & GENMASK(25, 21))
  61#define COMMAND_PORT_ADDR_ASSGN_CMD	0x03
  62
  63#define RESPONSE_QUEUE_PORT		0x10
  64#define RESPONSE_PORT_ERR_STATUS(x)	(((x) & GENMASK(31, 28)) >> 28)
  65#define RESPONSE_NO_ERROR		0
  66#define RESPONSE_ERROR_CRC		1
  67#define RESPONSE_ERROR_PARITY		2
  68#define RESPONSE_ERROR_FRAME		3
  69#define RESPONSE_ERROR_IBA_NACK		4
  70#define RESPONSE_ERROR_ADDRESS_NACK	5
  71#define RESPONSE_ERROR_OVER_UNDER_FLOW	6
  72#define RESPONSE_ERROR_TRANSF_ABORT	8
  73#define RESPONSE_ERROR_I2C_W_NACK_ERR	9
  74#define RESPONSE_PORT_TID(x)		(((x) & GENMASK(27, 24)) >> 24)
  75#define RESPONSE_PORT_DATA_LEN(x)	((x) & GENMASK(15, 0))
  76
  77#define RX_TX_DATA_PORT			0x14
  78#define IBI_QUEUE_STATUS		0x18
  79#define IBI_QUEUE_STATUS_IBI_ID(x)	(((x) & GENMASK(15, 8)) >> 8)
  80#define IBI_QUEUE_STATUS_DATA_LEN(x)	((x) & GENMASK(7, 0))
  81#define IBI_QUEUE_IBI_ADDR(x)		(IBI_QUEUE_STATUS_IBI_ID(x) >> 1)
  82#define IBI_QUEUE_IBI_RNW(x)		(IBI_QUEUE_STATUS_IBI_ID(x) & BIT(0))
  83#define IBI_TYPE_MR(x)                                                         \
  84	((IBI_QUEUE_IBI_ADDR(x) != I3C_HOT_JOIN_ADDR) && !IBI_QUEUE_IBI_RNW(x))
  85#define IBI_TYPE_HJ(x)                                                         \
  86	((IBI_QUEUE_IBI_ADDR(x) == I3C_HOT_JOIN_ADDR) && !IBI_QUEUE_IBI_RNW(x))
  87#define IBI_TYPE_SIRQ(x)                                                        \
  88	((IBI_QUEUE_IBI_ADDR(x) != I3C_HOT_JOIN_ADDR) && IBI_QUEUE_IBI_RNW(x))
  89
  90#define QUEUE_THLD_CTRL			0x1c
  91#define QUEUE_THLD_CTRL_IBI_STAT_MASK	GENMASK(31, 24)
  92#define QUEUE_THLD_CTRL_IBI_STAT(x)	(((x) - 1) << 24)
  93#define QUEUE_THLD_CTRL_IBI_DATA_MASK	GENMASK(20, 16)
  94#define QUEUE_THLD_CTRL_IBI_DATA(x)	((x) << 16)
  95#define QUEUE_THLD_CTRL_RESP_BUF_MASK	GENMASK(15, 8)
  96#define QUEUE_THLD_CTRL_RESP_BUF(x)	(((x) - 1) << 8)
  97
  98#define DATA_BUFFER_THLD_CTRL		0x20
  99#define DATA_BUFFER_THLD_CTRL_RX_BUF	GENMASK(11, 8)
 100
 101#define IBI_QUEUE_CTRL			0x24
 102#define IBI_MR_REQ_REJECT		0x2C
 103#define IBI_SIR_REQ_REJECT		0x30
 104#define IBI_REQ_REJECT_ALL		GENMASK(31, 0)
 105
 106#define RESET_CTRL			0x34
 107#define RESET_CTRL_IBI_QUEUE		BIT(5)
 108#define RESET_CTRL_RX_FIFO		BIT(4)
 109#define RESET_CTRL_TX_FIFO		BIT(3)
 110#define RESET_CTRL_RESP_QUEUE		BIT(2)
 111#define RESET_CTRL_CMD_QUEUE		BIT(1)
 112#define RESET_CTRL_SOFT			BIT(0)
 113
 114#define SLV_EVENT_CTRL			0x38
 115#define INTR_STATUS			0x3c
 116#define INTR_STATUS_EN			0x40
 117#define INTR_SIGNAL_EN			0x44
 118#define INTR_FORCE			0x48
 119#define INTR_BUSOWNER_UPDATE_STAT	BIT(13)
 120#define INTR_IBI_UPDATED_STAT		BIT(12)
 121#define INTR_READ_REQ_RECV_STAT		BIT(11)
 122#define INTR_DEFSLV_STAT		BIT(10)
 123#define INTR_TRANSFER_ERR_STAT		BIT(9)
 124#define INTR_DYN_ADDR_ASSGN_STAT	BIT(8)
 125#define INTR_CCC_UPDATED_STAT		BIT(6)
 126#define INTR_TRANSFER_ABORT_STAT	BIT(5)
 127#define INTR_RESP_READY_STAT		BIT(4)
 128#define INTR_CMD_QUEUE_READY_STAT	BIT(3)
 129#define INTR_IBI_THLD_STAT		BIT(2)
 130#define INTR_RX_THLD_STAT		BIT(1)
 131#define INTR_TX_THLD_STAT		BIT(0)
 132#define INTR_ALL			(INTR_BUSOWNER_UPDATE_STAT |	\
 133					INTR_IBI_UPDATED_STAT |		\
 134					INTR_READ_REQ_RECV_STAT |	\
 135					INTR_DEFSLV_STAT |		\
 136					INTR_TRANSFER_ERR_STAT |	\
 137					INTR_DYN_ADDR_ASSGN_STAT |	\
 138					INTR_CCC_UPDATED_STAT |		\
 139					INTR_TRANSFER_ABORT_STAT |	\
 140					INTR_RESP_READY_STAT |		\
 141					INTR_CMD_QUEUE_READY_STAT |	\
 142					INTR_IBI_THLD_STAT |		\
 143					INTR_TX_THLD_STAT |		\
 144					INTR_RX_THLD_STAT)
 145
 146#define INTR_MASTER_MASK		(INTR_TRANSFER_ERR_STAT |	\
 147					 INTR_RESP_READY_STAT)
 148
 149#define QUEUE_STATUS_LEVEL		0x4c
 150#define QUEUE_STATUS_IBI_STATUS_CNT(x)	(((x) & GENMASK(28, 24)) >> 24)
 151#define QUEUE_STATUS_IBI_BUF_BLR(x)	(((x) & GENMASK(23, 16)) >> 16)
 152#define QUEUE_STATUS_LEVEL_RESP(x)	(((x) & GENMASK(15, 8)) >> 8)
 153#define QUEUE_STATUS_LEVEL_CMD(x)	((x) & GENMASK(7, 0))
 154
 155#define DATA_BUFFER_STATUS_LEVEL	0x50
 156#define DATA_BUFFER_STATUS_LEVEL_TX(x)	((x) & GENMASK(7, 0))
 157
 158#define PRESENT_STATE			0x54
 159#define CCC_DEVICE_STATUS		0x58
 160#define DEVICE_ADDR_TABLE_POINTER	0x5c
 161#define DEVICE_ADDR_TABLE_DEPTH(x)	(((x) & GENMASK(31, 16)) >> 16)
 162#define DEVICE_ADDR_TABLE_ADDR(x)	((x) & GENMASK(7, 0))
 163
 164#define DEV_CHAR_TABLE_POINTER		0x60
 165#define VENDOR_SPECIFIC_REG_POINTER	0x6c
 166#define SLV_PID_VALUE			0x74
 167#define SLV_CHAR_CTRL			0x78
 168#define SLV_MAX_LEN			0x7c
 169#define MAX_READ_TURNAROUND		0x80
 170#define MAX_DATA_SPEED			0x84
 171#define SLV_DEBUG_STATUS		0x88
 172#define SLV_INTR_REQ			0x8c
 173#define DEVICE_CTRL_EXTENDED		0xb0
 174#define SCL_I3C_OD_TIMING		0xb4
 175#define SCL_I3C_PP_TIMING		0xb8
 176#define SCL_I3C_TIMING_HCNT(x)		(((x) << 16) & GENMASK(23, 16))
 177#define SCL_I3C_TIMING_LCNT(x)		((x) & GENMASK(7, 0))
 178#define SCL_I3C_TIMING_CNT_MIN		5
 179
 180#define SCL_I2C_FM_TIMING		0xbc
 181#define SCL_I2C_FM_TIMING_HCNT(x)	(((x) << 16) & GENMASK(31, 16))
 182#define SCL_I2C_FM_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 183
 184#define SCL_I2C_FMP_TIMING		0xc0
 185#define SCL_I2C_FMP_TIMING_HCNT(x)	(((x) << 16) & GENMASK(23, 16))
 186#define SCL_I2C_FMP_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 187
 188#define SCL_EXT_LCNT_TIMING		0xc8
 189#define SCL_EXT_LCNT_4(x)		(((x) << 24) & GENMASK(31, 24))
 190#define SCL_EXT_LCNT_3(x)		(((x) << 16) & GENMASK(23, 16))
 191#define SCL_EXT_LCNT_2(x)		(((x) << 8) & GENMASK(15, 8))
 192#define SCL_EXT_LCNT_1(x)		((x) & GENMASK(7, 0))
 193
 194#define SCL_EXT_TERMN_LCNT_TIMING	0xcc
 195#define BUS_FREE_TIMING			0xd4
 196#define BUS_I3C_MST_FREE(x)		((x) & GENMASK(15, 0))
 197
 198#define BUS_IDLE_TIMING			0xd8
 199#define I3C_VER_ID			0xe0
 200#define I3C_VER_TYPE			0xe4
 201#define EXTENDED_CAPABILITY		0xe8
 202#define SLAVE_CONFIG			0xec
 203
 204#define DEV_ADDR_TABLE_IBI_MDB		BIT(12)
 205#define DEV_ADDR_TABLE_SIR_REJECT	BIT(13)
 206#define DEV_ADDR_TABLE_LEGACY_I2C_DEV	BIT(31)
 207#define DEV_ADDR_TABLE_DYNAMIC_ADDR(x)	(((x) << 16) & GENMASK(23, 16))
 208#define DEV_ADDR_TABLE_STATIC_ADDR(x)	((x) & GENMASK(6, 0))
 209#define DEV_ADDR_TABLE_LOC(start, idx)	((start) + ((idx) << 2))
 210
 
 
 211#define I3C_BUS_SDR1_SCL_RATE		8000000
 212#define I3C_BUS_SDR2_SCL_RATE		6000000
 213#define I3C_BUS_SDR3_SCL_RATE		4000000
 214#define I3C_BUS_SDR4_SCL_RATE		2000000
 215#define I3C_BUS_I2C_FM_TLOW_MIN_NS	1300
 216#define I3C_BUS_I2C_FMP_TLOW_MIN_NS	500
 217#define I3C_BUS_THIGH_MAX_NS		41
 218
 219#define XFER_TIMEOUT (msecs_to_jiffies(1000))
 220
 
 
 
 
 
 221struct dw_i3c_cmd {
 222	u32 cmd_lo;
 223	u32 cmd_hi;
 224	u16 tx_len;
 225	const void *tx_buf;
 226	u16 rx_len;
 227	void *rx_buf;
 228	u8 error;
 229};
 230
 231struct dw_i3c_xfer {
 232	struct list_head node;
 233	struct completion comp;
 234	int ret;
 235	unsigned int ncmds;
 236	struct dw_i3c_cmd cmds[] __counted_by(ncmds);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 237};
 238
 239struct dw_i3c_i2c_dev_data {
 240	u8 index;
 241	struct i3c_generic_ibi_pool *ibi_pool;
 242};
 243
 244static u8 even_parity(u8 p)
 245{
 246	p ^= p >> 4;
 247	p &= 0xf;
 248
 249	return (0x9669 >> p) & 1;
 250}
 251
 252static bool dw_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m,
 253					   const struct i3c_ccc_cmd *cmd)
 254{
 255	if (cmd->ndests > 1)
 256		return false;
 257
 258	switch (cmd->id) {
 259	case I3C_CCC_ENEC(true):
 260	case I3C_CCC_ENEC(false):
 261	case I3C_CCC_DISEC(true):
 262	case I3C_CCC_DISEC(false):
 263	case I3C_CCC_ENTAS(0, true):
 264	case I3C_CCC_ENTAS(0, false):
 265	case I3C_CCC_RSTDAA(true):
 266	case I3C_CCC_RSTDAA(false):
 267	case I3C_CCC_ENTDAA:
 268	case I3C_CCC_SETMWL(true):
 269	case I3C_CCC_SETMWL(false):
 270	case I3C_CCC_SETMRL(true):
 271	case I3C_CCC_SETMRL(false):
 272	case I3C_CCC_ENTHDR(0):
 273	case I3C_CCC_SETDASA:
 274	case I3C_CCC_SETNEWDA:
 275	case I3C_CCC_GETMWL:
 276	case I3C_CCC_GETMRL:
 277	case I3C_CCC_GETPID:
 278	case I3C_CCC_GETBCR:
 279	case I3C_CCC_GETDCR:
 280	case I3C_CCC_GETSTATUS:
 281	case I3C_CCC_GETMXDS:
 282	case I3C_CCC_GETHDRCAP:
 283		return true;
 284	default:
 285		return false;
 286	}
 287}
 288
 289static inline struct dw_i3c_master *
 290to_dw_i3c_master(struct i3c_master_controller *master)
 291{
 292	return container_of(master, struct dw_i3c_master, base);
 293}
 294
 295static void dw_i3c_master_disable(struct dw_i3c_master *master)
 296{
 297	writel(readl(master->regs + DEVICE_CTRL) & ~DEV_CTRL_ENABLE,
 298	       master->regs + DEVICE_CTRL);
 299}
 300
 301static void dw_i3c_master_enable(struct dw_i3c_master *master)
 302{
 303	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_ENABLE,
 304	       master->regs + DEVICE_CTRL);
 305}
 306
 307static int dw_i3c_master_get_addr_pos(struct dw_i3c_master *master, u8 addr)
 308{
 309	int pos;
 310
 311	for (pos = 0; pos < master->maxdevs; pos++) {
 312		if (addr == master->devs[pos].addr)
 313			return pos;
 314	}
 315
 316	return -EINVAL;
 317}
 318
 319static int dw_i3c_master_get_free_pos(struct dw_i3c_master *master)
 320{
 321	if (!(master->free_pos & GENMASK(master->maxdevs - 1, 0)))
 322		return -ENOSPC;
 323
 324	return ffs(master->free_pos) - 1;
 325}
 326
 327static void dw_i3c_master_wr_tx_fifo(struct dw_i3c_master *master,
 328				     const u8 *bytes, int nbytes)
 329{
 330	writesl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 331	if (nbytes & 3) {
 332		u32 tmp = 0;
 333
 334		memcpy(&tmp, bytes + (nbytes & ~3), nbytes & 3);
 335		writesl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 336	}
 337}
 338
 339static void dw_i3c_master_read_fifo(struct dw_i3c_master *master,
 340				    int reg,  u8 *bytes, int nbytes)
 341{
 342	readsl(master->regs + reg, bytes, nbytes / 4);
 343	if (nbytes & 3) {
 344		u32 tmp;
 345
 346		readsl(master->regs + reg, &tmp, 1);
 347		memcpy(bytes + (nbytes & ~3), &tmp, nbytes & 3);
 348	}
 349}
 350
 351static void dw_i3c_master_read_rx_fifo(struct dw_i3c_master *master,
 352				       u8 *bytes, int nbytes)
 353{
 354	return dw_i3c_master_read_fifo(master, RX_TX_DATA_PORT, bytes, nbytes);
 355}
 356
 357static void dw_i3c_master_read_ibi_fifo(struct dw_i3c_master *master,
 358					u8 *bytes, int nbytes)
 359{
 360	return dw_i3c_master_read_fifo(master, IBI_QUEUE_STATUS, bytes, nbytes);
 361}
 362
 363static struct dw_i3c_xfer *
 364dw_i3c_master_alloc_xfer(struct dw_i3c_master *master, unsigned int ncmds)
 365{
 366	struct dw_i3c_xfer *xfer;
 367
 368	xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL);
 369	if (!xfer)
 370		return NULL;
 371
 372	INIT_LIST_HEAD(&xfer->node);
 373	xfer->ncmds = ncmds;
 374	xfer->ret = -ETIMEDOUT;
 375
 376	return xfer;
 377}
 378
 379static void dw_i3c_master_free_xfer(struct dw_i3c_xfer *xfer)
 380{
 381	kfree(xfer);
 382}
 383
 384static void dw_i3c_master_start_xfer_locked(struct dw_i3c_master *master)
 385{
 386	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 387	unsigned int i;
 388	u32 thld_ctrl;
 389
 390	if (!xfer)
 391		return;
 392
 393	for (i = 0; i < xfer->ncmds; i++) {
 394		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 395
 396		dw_i3c_master_wr_tx_fifo(master, cmd->tx_buf, cmd->tx_len);
 397	}
 398
 399	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 400	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 401	thld_ctrl |= QUEUE_THLD_CTRL_RESP_BUF(xfer->ncmds);
 402	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 403
 404	for (i = 0; i < xfer->ncmds; i++) {
 405		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 406
 407		writel(cmd->cmd_hi, master->regs + COMMAND_QUEUE_PORT);
 408		writel(cmd->cmd_lo, master->regs + COMMAND_QUEUE_PORT);
 409	}
 410}
 411
 412static void dw_i3c_master_enqueue_xfer(struct dw_i3c_master *master,
 413				       struct dw_i3c_xfer *xfer)
 414{
 415	unsigned long flags;
 416
 417	init_completion(&xfer->comp);
 418	spin_lock_irqsave(&master->xferqueue.lock, flags);
 419	if (master->xferqueue.cur) {
 420		list_add_tail(&xfer->node, &master->xferqueue.list);
 421	} else {
 422		master->xferqueue.cur = xfer;
 423		dw_i3c_master_start_xfer_locked(master);
 424	}
 425	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 426}
 427
 428static void dw_i3c_master_dequeue_xfer_locked(struct dw_i3c_master *master,
 429					      struct dw_i3c_xfer *xfer)
 430{
 431	if (master->xferqueue.cur == xfer) {
 432		u32 status;
 433
 434		master->xferqueue.cur = NULL;
 435
 436		writel(RESET_CTRL_RX_FIFO | RESET_CTRL_TX_FIFO |
 437		       RESET_CTRL_RESP_QUEUE | RESET_CTRL_CMD_QUEUE,
 438		       master->regs + RESET_CTRL);
 439
 440		readl_poll_timeout_atomic(master->regs + RESET_CTRL, status,
 441					  !status, 10, 1000000);
 442	} else {
 443		list_del_init(&xfer->node);
 444	}
 445}
 446
 447static void dw_i3c_master_dequeue_xfer(struct dw_i3c_master *master,
 448				       struct dw_i3c_xfer *xfer)
 449{
 450	unsigned long flags;
 451
 452	spin_lock_irqsave(&master->xferqueue.lock, flags);
 453	dw_i3c_master_dequeue_xfer_locked(master, xfer);
 454	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 455}
 456
 457static void dw_i3c_master_end_xfer_locked(struct dw_i3c_master *master, u32 isr)
 458{
 459	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 460	int i, ret = 0;
 461	u32 nresp;
 462
 463	if (!xfer)
 464		return;
 465
 466	nresp = readl(master->regs + QUEUE_STATUS_LEVEL);
 467	nresp = QUEUE_STATUS_LEVEL_RESP(nresp);
 468
 469	for (i = 0; i < nresp; i++) {
 470		struct dw_i3c_cmd *cmd;
 471		u32 resp;
 472
 473		resp = readl(master->regs + RESPONSE_QUEUE_PORT);
 474
 475		cmd = &xfer->cmds[RESPONSE_PORT_TID(resp)];
 476		cmd->rx_len = RESPONSE_PORT_DATA_LEN(resp);
 477		cmd->error = RESPONSE_PORT_ERR_STATUS(resp);
 478		if (cmd->rx_len && !cmd->error)
 479			dw_i3c_master_read_rx_fifo(master, cmd->rx_buf,
 480						   cmd->rx_len);
 481	}
 482
 483	for (i = 0; i < nresp; i++) {
 484		switch (xfer->cmds[i].error) {
 485		case RESPONSE_NO_ERROR:
 486			break;
 487		case RESPONSE_ERROR_PARITY:
 488		case RESPONSE_ERROR_IBA_NACK:
 489		case RESPONSE_ERROR_TRANSF_ABORT:
 490		case RESPONSE_ERROR_CRC:
 491		case RESPONSE_ERROR_FRAME:
 492			ret = -EIO;
 493			break;
 494		case RESPONSE_ERROR_OVER_UNDER_FLOW:
 495			ret = -ENOSPC;
 496			break;
 497		case RESPONSE_ERROR_I2C_W_NACK_ERR:
 498		case RESPONSE_ERROR_ADDRESS_NACK:
 499		default:
 500			ret = -EINVAL;
 501			break;
 502		}
 503	}
 504
 505	xfer->ret = ret;
 506	complete(&xfer->comp);
 507
 508	if (ret < 0) {
 509		dw_i3c_master_dequeue_xfer_locked(master, xfer);
 510		writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_RESUME,
 511		       master->regs + DEVICE_CTRL);
 512	}
 513
 514	xfer = list_first_entry_or_null(&master->xferqueue.list,
 515					struct dw_i3c_xfer,
 516					node);
 517	if (xfer)
 518		list_del_init(&xfer->node);
 519
 520	master->xferqueue.cur = xfer;
 521	dw_i3c_master_start_xfer_locked(master);
 522}
 523
 524static int dw_i3c_clk_cfg(struct dw_i3c_master *master)
 525{
 526	unsigned long core_rate, core_period;
 527	u32 scl_timing;
 528	u8 hcnt, lcnt;
 529
 530	core_rate = clk_get_rate(master->core_clk);
 531	if (!core_rate)
 532		return -EINVAL;
 533
 534	core_period = DIV_ROUND_UP(1000000000, core_rate);
 535
 536	hcnt = DIV_ROUND_UP(I3C_BUS_THIGH_MAX_NS, core_period) - 1;
 537	if (hcnt < SCL_I3C_TIMING_CNT_MIN)
 538		hcnt = SCL_I3C_TIMING_CNT_MIN;
 539
 540	lcnt = DIV_ROUND_UP(core_rate, master->base.bus.scl_rate.i3c) - hcnt;
 541	if (lcnt < SCL_I3C_TIMING_CNT_MIN)
 542		lcnt = SCL_I3C_TIMING_CNT_MIN;
 543
 544	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 545	writel(scl_timing, master->regs + SCL_I3C_PP_TIMING);
 546
 547	/*
 548	 * In pure i3c mode, MST_FREE represents tCAS. In shared mode, this
 549	 * will be set up by dw_i2c_clk_cfg as tLOW.
 550	 */
 551	if (master->base.bus.mode == I3C_BUS_MODE_PURE)
 552		writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 553
 554	lcnt = max_t(u8,
 555		     DIV_ROUND_UP(I3C_BUS_TLOW_OD_MIN_NS, core_period), lcnt);
 556	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 557	writel(scl_timing, master->regs + SCL_I3C_OD_TIMING);
 558
 559	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR1_SCL_RATE) - hcnt;
 560	scl_timing = SCL_EXT_LCNT_1(lcnt);
 561	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR2_SCL_RATE) - hcnt;
 562	scl_timing |= SCL_EXT_LCNT_2(lcnt);
 563	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR3_SCL_RATE) - hcnt;
 564	scl_timing |= SCL_EXT_LCNT_3(lcnt);
 565	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR4_SCL_RATE) - hcnt;
 566	scl_timing |= SCL_EXT_LCNT_4(lcnt);
 567	writel(scl_timing, master->regs + SCL_EXT_LCNT_TIMING);
 568
 569	return 0;
 570}
 571
 572static int dw_i2c_clk_cfg(struct dw_i3c_master *master)
 573{
 574	unsigned long core_rate, core_period;
 575	u16 hcnt, lcnt;
 576	u32 scl_timing;
 577
 578	core_rate = clk_get_rate(master->core_clk);
 579	if (!core_rate)
 580		return -EINVAL;
 581
 582	core_period = DIV_ROUND_UP(1000000000, core_rate);
 583
 584	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FMP_TLOW_MIN_NS, core_period);
 585	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_PLUS_SCL_RATE) - lcnt;
 586	scl_timing = SCL_I2C_FMP_TIMING_HCNT(hcnt) |
 587		     SCL_I2C_FMP_TIMING_LCNT(lcnt);
 588	writel(scl_timing, master->regs + SCL_I2C_FMP_TIMING);
 589
 590	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FM_TLOW_MIN_NS, core_period);
 591	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_SCL_RATE) - lcnt;
 592	scl_timing = SCL_I2C_FM_TIMING_HCNT(hcnt) |
 593		     SCL_I2C_FM_TIMING_LCNT(lcnt);
 594	writel(scl_timing, master->regs + SCL_I2C_FM_TIMING);
 595
 596	writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 597	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_I2C_SLAVE_PRESENT,
 598	       master->regs + DEVICE_CTRL);
 599
 600	return 0;
 601}
 602
 603static int dw_i3c_master_bus_init(struct i3c_master_controller *m)
 604{
 605	struct dw_i3c_master *master = to_dw_i3c_master(m);
 606	struct i3c_bus *bus = i3c_master_get_bus(m);
 607	struct i3c_device_info info = { };
 608	u32 thld_ctrl;
 609	int ret;
 610
 611	ret = master->platform_ops->init(master);
 612	if (ret)
 613		return ret;
 614
 615	switch (bus->mode) {
 616	case I3C_BUS_MODE_MIXED_FAST:
 617	case I3C_BUS_MODE_MIXED_LIMITED:
 618		ret = dw_i2c_clk_cfg(master);
 619		if (ret)
 620			return ret;
 621		fallthrough;
 622	case I3C_BUS_MODE_PURE:
 623		ret = dw_i3c_clk_cfg(master);
 624		if (ret)
 625			return ret;
 626		break;
 627	default:
 628		return -EINVAL;
 629	}
 630
 631	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 632	thld_ctrl &= ~(QUEUE_THLD_CTRL_RESP_BUF_MASK |
 633		       QUEUE_THLD_CTRL_IBI_STAT_MASK |
 634		       QUEUE_THLD_CTRL_IBI_STAT_MASK);
 635	thld_ctrl |= QUEUE_THLD_CTRL_IBI_STAT(1) |
 636		QUEUE_THLD_CTRL_IBI_DATA(31);
 637	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 638
 639	thld_ctrl = readl(master->regs + DATA_BUFFER_THLD_CTRL);
 640	thld_ctrl &= ~DATA_BUFFER_THLD_CTRL_RX_BUF;
 641	writel(thld_ctrl, master->regs + DATA_BUFFER_THLD_CTRL);
 642
 643	writel(INTR_ALL, master->regs + INTR_STATUS);
 644	writel(INTR_MASTER_MASK, master->regs + INTR_STATUS_EN);
 645	writel(INTR_MASTER_MASK, master->regs + INTR_SIGNAL_EN);
 646
 647	ret = i3c_master_get_free_addr(m, 0);
 648	if (ret < 0)
 649		return ret;
 650
 651	writel(DEV_ADDR_DYNAMIC_ADDR_VALID | DEV_ADDR_DYNAMIC(ret),
 652	       master->regs + DEVICE_ADDR);
 653
 654	memset(&info, 0, sizeof(info));
 655	info.dyn_addr = ret;
 656
 657	ret = i3c_master_set_info(&master->base, &info);
 658	if (ret)
 659		return ret;
 660
 661	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_SIR_REQ_REJECT);
 662	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_MR_REQ_REJECT);
 663
 664	/* For now don't support Hot-Join */
 665	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_HOT_JOIN_NACK,
 666	       master->regs + DEVICE_CTRL);
 667
 668	dw_i3c_master_enable(master);
 669
 670	return 0;
 671}
 672
 673static void dw_i3c_master_bus_cleanup(struct i3c_master_controller *m)
 674{
 675	struct dw_i3c_master *master = to_dw_i3c_master(m);
 676
 677	dw_i3c_master_disable(master);
 678}
 679
 680static int dw_i3c_ccc_set(struct dw_i3c_master *master,
 681			  struct i3c_ccc_cmd *ccc)
 682{
 683	struct dw_i3c_xfer *xfer;
 684	struct dw_i3c_cmd *cmd;
 685	int ret, pos = 0;
 686
 687	if (ccc->id & I3C_CCC_DIRECT) {
 688		pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 689		if (pos < 0)
 690			return pos;
 691	}
 692
 693	xfer = dw_i3c_master_alloc_xfer(master, 1);
 694	if (!xfer)
 695		return -ENOMEM;
 696
 697	cmd = xfer->cmds;
 698	cmd->tx_buf = ccc->dests[0].payload.data;
 699	cmd->tx_len = ccc->dests[0].payload.len;
 700
 701	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 702		      COMMAND_PORT_TRANSFER_ARG;
 703
 704	cmd->cmd_lo = COMMAND_PORT_CP |
 705		      COMMAND_PORT_DEV_INDEX(pos) |
 706		      COMMAND_PORT_CMD(ccc->id) |
 707		      COMMAND_PORT_TOC |
 708		      COMMAND_PORT_ROC;
 709
 710	dw_i3c_master_enqueue_xfer(master, xfer);
 711	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 712		dw_i3c_master_dequeue_xfer(master, xfer);
 713
 714	ret = xfer->ret;
 715	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 716		ccc->err = I3C_ERROR_M2;
 717
 718	dw_i3c_master_free_xfer(xfer);
 719
 720	return ret;
 721}
 722
 723static int dw_i3c_ccc_get(struct dw_i3c_master *master, struct i3c_ccc_cmd *ccc)
 724{
 725	struct dw_i3c_xfer *xfer;
 726	struct dw_i3c_cmd *cmd;
 727	int ret, pos;
 728
 729	pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 730	if (pos < 0)
 731		return pos;
 732
 733	xfer = dw_i3c_master_alloc_xfer(master, 1);
 734	if (!xfer)
 735		return -ENOMEM;
 736
 737	cmd = xfer->cmds;
 738	cmd->rx_buf = ccc->dests[0].payload.data;
 739	cmd->rx_len = ccc->dests[0].payload.len;
 740
 741	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 742		      COMMAND_PORT_TRANSFER_ARG;
 743
 744	cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 745		      COMMAND_PORT_CP |
 746		      COMMAND_PORT_DEV_INDEX(pos) |
 747		      COMMAND_PORT_CMD(ccc->id) |
 748		      COMMAND_PORT_TOC |
 749		      COMMAND_PORT_ROC;
 750
 751	dw_i3c_master_enqueue_xfer(master, xfer);
 752	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 753		dw_i3c_master_dequeue_xfer(master, xfer);
 754
 755	ret = xfer->ret;
 756	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 757		ccc->err = I3C_ERROR_M2;
 758	dw_i3c_master_free_xfer(xfer);
 759
 760	return ret;
 761}
 762
 763static int dw_i3c_master_send_ccc_cmd(struct i3c_master_controller *m,
 764				      struct i3c_ccc_cmd *ccc)
 765{
 766	struct dw_i3c_master *master = to_dw_i3c_master(m);
 767	int ret = 0;
 768
 769	if (ccc->id == I3C_CCC_ENTDAA)
 770		return -EINVAL;
 771
 772	if (ccc->rnw)
 773		ret = dw_i3c_ccc_get(master, ccc);
 774	else
 775		ret = dw_i3c_ccc_set(master, ccc);
 776
 777	return ret;
 778}
 779
 780static int dw_i3c_master_daa(struct i3c_master_controller *m)
 781{
 782	struct dw_i3c_master *master = to_dw_i3c_master(m);
 783	struct dw_i3c_xfer *xfer;
 784	struct dw_i3c_cmd *cmd;
 785	u32 olddevs, newdevs;
 786	u8 p, last_addr = 0;
 787	int ret, pos;
 788
 789	olddevs = ~(master->free_pos);
 790
 791	/* Prepare DAT before launching DAA. */
 792	for (pos = 0; pos < master->maxdevs; pos++) {
 793		if (olddevs & BIT(pos))
 794			continue;
 795
 796		ret = i3c_master_get_free_addr(m, last_addr + 1);
 797		if (ret < 0)
 798			return -ENOSPC;
 799
 800		master->devs[pos].addr = ret;
 801		p = even_parity(ret);
 802		last_addr = ret;
 803		ret |= (p << 7);
 804
 805		writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(ret),
 806		       master->regs +
 807		       DEV_ADDR_TABLE_LOC(master->datstartaddr, pos));
 808	}
 809
 810	xfer = dw_i3c_master_alloc_xfer(master, 1);
 811	if (!xfer)
 812		return -ENOMEM;
 813
 814	pos = dw_i3c_master_get_free_pos(master);
 815	if (pos < 0) {
 816		dw_i3c_master_free_xfer(xfer);
 817		return pos;
 818	}
 819	cmd = &xfer->cmds[0];
 820	cmd->cmd_hi = 0x1;
 821	cmd->cmd_lo = COMMAND_PORT_DEV_COUNT(master->maxdevs - pos) |
 822		      COMMAND_PORT_DEV_INDEX(pos) |
 823		      COMMAND_PORT_CMD(I3C_CCC_ENTDAA) |
 824		      COMMAND_PORT_ADDR_ASSGN_CMD |
 825		      COMMAND_PORT_TOC |
 826		      COMMAND_PORT_ROC;
 827
 828	dw_i3c_master_enqueue_xfer(master, xfer);
 829	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 830		dw_i3c_master_dequeue_xfer(master, xfer);
 831
 832	newdevs = GENMASK(master->maxdevs - cmd->rx_len - 1, 0);
 833	newdevs &= ~olddevs;
 834
 835	for (pos = 0; pos < master->maxdevs; pos++) {
 836		if (newdevs & BIT(pos))
 837			i3c_master_add_i3c_dev_locked(m, master->devs[pos].addr);
 838	}
 839
 840	dw_i3c_master_free_xfer(xfer);
 841
 842	return 0;
 843}
 844
 845static int dw_i3c_master_priv_xfers(struct i3c_dev_desc *dev,
 846				    struct i3c_priv_xfer *i3c_xfers,
 847				    int i3c_nxfers)
 848{
 849	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 850	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 851	struct dw_i3c_master *master = to_dw_i3c_master(m);
 852	unsigned int nrxwords = 0, ntxwords = 0;
 853	struct dw_i3c_xfer *xfer;
 854	int i, ret = 0;
 855
 856	if (!i3c_nxfers)
 857		return 0;
 858
 859	if (i3c_nxfers > master->caps.cmdfifodepth)
 860		return -ENOTSUPP;
 861
 862	for (i = 0; i < i3c_nxfers; i++) {
 863		if (i3c_xfers[i].rnw)
 864			nrxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 865		else
 866			ntxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 867	}
 868
 869	if (ntxwords > master->caps.datafifodepth ||
 870	    nrxwords > master->caps.datafifodepth)
 871		return -ENOTSUPP;
 872
 873	xfer = dw_i3c_master_alloc_xfer(master, i3c_nxfers);
 874	if (!xfer)
 875		return -ENOMEM;
 876
 877	for (i = 0; i < i3c_nxfers; i++) {
 878		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 879
 880		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i3c_xfers[i].len) |
 881			COMMAND_PORT_TRANSFER_ARG;
 882
 883		if (i3c_xfers[i].rnw) {
 884			cmd->rx_buf = i3c_xfers[i].data.in;
 885			cmd->rx_len = i3c_xfers[i].len;
 886			cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 887				      COMMAND_PORT_SPEED(dev->info.max_read_ds);
 888
 889		} else {
 890			cmd->tx_buf = i3c_xfers[i].data.out;
 891			cmd->tx_len = i3c_xfers[i].len;
 892			cmd->cmd_lo =
 893				COMMAND_PORT_SPEED(dev->info.max_write_ds);
 894		}
 895
 896		cmd->cmd_lo |= COMMAND_PORT_TID(i) |
 897			       COMMAND_PORT_DEV_INDEX(data->index) |
 898			       COMMAND_PORT_ROC;
 899
 900		if (i == (i3c_nxfers - 1))
 901			cmd->cmd_lo |= COMMAND_PORT_TOC;
 902	}
 903
 904	dw_i3c_master_enqueue_xfer(master, xfer);
 905	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 906		dw_i3c_master_dequeue_xfer(master, xfer);
 907
 908	for (i = 0; i < i3c_nxfers; i++) {
 909		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 910
 911		if (i3c_xfers[i].rnw)
 912			i3c_xfers[i].len = cmd->rx_len;
 913	}
 914
 915	ret = xfer->ret;
 916	dw_i3c_master_free_xfer(xfer);
 917
 918	return ret;
 919}
 920
 921static int dw_i3c_master_reattach_i3c_dev(struct i3c_dev_desc *dev,
 922					  u8 old_dyn_addr)
 923{
 924	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 925	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 926	struct dw_i3c_master *master = to_dw_i3c_master(m);
 927	int pos;
 928
 929	pos = dw_i3c_master_get_free_pos(master);
 930
 931	if (data->index > pos && pos > 0) {
 932		writel(0,
 933		       master->regs +
 934		       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 935
 936		master->devs[data->index].addr = 0;
 937		master->free_pos |= BIT(data->index);
 938
 939		data->index = pos;
 940		master->devs[pos].addr = dev->info.dyn_addr;
 941		master->free_pos &= ~BIT(pos);
 942	}
 943
 944	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(dev->info.dyn_addr),
 945	       master->regs +
 946	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 947
 948	master->devs[data->index].addr = dev->info.dyn_addr;
 949
 950	return 0;
 951}
 952
 953static int dw_i3c_master_attach_i3c_dev(struct i3c_dev_desc *dev)
 954{
 955	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 956	struct dw_i3c_master *master = to_dw_i3c_master(m);
 957	struct dw_i3c_i2c_dev_data *data;
 958	int pos;
 959
 960	pos = dw_i3c_master_get_free_pos(master);
 961	if (pos < 0)
 962		return pos;
 963
 964	data = kzalloc(sizeof(*data), GFP_KERNEL);
 965	if (!data)
 966		return -ENOMEM;
 967
 968	data->index = pos;
 969	master->devs[pos].addr = dev->info.dyn_addr ? : dev->info.static_addr;
 970	master->free_pos &= ~BIT(pos);
 971	i3c_dev_set_master_data(dev, data);
 972
 973	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(master->devs[pos].addr),
 974	       master->regs +
 975	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 976
 977	return 0;
 978}
 979
 980static void dw_i3c_master_detach_i3c_dev(struct i3c_dev_desc *dev)
 981{
 982	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 983	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 984	struct dw_i3c_master *master = to_dw_i3c_master(m);
 985
 986	writel(0,
 987	       master->regs +
 988	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 989
 990	i3c_dev_set_master_data(dev, NULL);
 991	master->devs[data->index].addr = 0;
 992	master->free_pos |= BIT(data->index);
 993	kfree(data);
 994}
 995
 996static int dw_i3c_master_i2c_xfers(struct i2c_dev_desc *dev,
 997				   const struct i2c_msg *i2c_xfers,
 998				   int i2c_nxfers)
 999{
1000	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1001	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1002	struct dw_i3c_master *master = to_dw_i3c_master(m);
1003	unsigned int nrxwords = 0, ntxwords = 0;
1004	struct dw_i3c_xfer *xfer;
1005	int i, ret = 0;
1006
1007	if (!i2c_nxfers)
1008		return 0;
1009
1010	if (i2c_nxfers > master->caps.cmdfifodepth)
1011		return -ENOTSUPP;
1012
1013	for (i = 0; i < i2c_nxfers; i++) {
1014		if (i2c_xfers[i].flags & I2C_M_RD)
1015			nrxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
1016		else
1017			ntxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
1018	}
1019
1020	if (ntxwords > master->caps.datafifodepth ||
1021	    nrxwords > master->caps.datafifodepth)
1022		return -ENOTSUPP;
1023
1024	xfer = dw_i3c_master_alloc_xfer(master, i2c_nxfers);
1025	if (!xfer)
1026		return -ENOMEM;
1027
1028	for (i = 0; i < i2c_nxfers; i++) {
1029		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
1030
1031		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i2c_xfers[i].len) |
1032			COMMAND_PORT_TRANSFER_ARG;
1033
1034		cmd->cmd_lo = COMMAND_PORT_TID(i) |
1035			      COMMAND_PORT_DEV_INDEX(data->index) |
1036			      COMMAND_PORT_ROC;
1037
1038		if (i2c_xfers[i].flags & I2C_M_RD) {
1039			cmd->cmd_lo |= COMMAND_PORT_READ_TRANSFER;
1040			cmd->rx_buf = i2c_xfers[i].buf;
1041			cmd->rx_len = i2c_xfers[i].len;
1042		} else {
1043			cmd->tx_buf = i2c_xfers[i].buf;
1044			cmd->tx_len = i2c_xfers[i].len;
1045		}
1046
1047		if (i == (i2c_nxfers - 1))
1048			cmd->cmd_lo |= COMMAND_PORT_TOC;
1049	}
1050
1051	dw_i3c_master_enqueue_xfer(master, xfer);
1052	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
1053		dw_i3c_master_dequeue_xfer(master, xfer);
1054
1055	ret = xfer->ret;
1056	dw_i3c_master_free_xfer(xfer);
1057
1058	return ret;
1059}
1060
1061static int dw_i3c_master_attach_i2c_dev(struct i2c_dev_desc *dev)
1062{
1063	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1064	struct dw_i3c_master *master = to_dw_i3c_master(m);
1065	struct dw_i3c_i2c_dev_data *data;
1066	int pos;
1067
1068	pos = dw_i3c_master_get_free_pos(master);
1069	if (pos < 0)
1070		return pos;
1071
1072	data = kzalloc(sizeof(*data), GFP_KERNEL);
1073	if (!data)
1074		return -ENOMEM;
1075
1076	data->index = pos;
1077	master->devs[pos].addr = dev->addr;
1078	master->free_pos &= ~BIT(pos);
1079	i2c_dev_set_master_data(dev, data);
1080
1081	writel(DEV_ADDR_TABLE_LEGACY_I2C_DEV |
1082	       DEV_ADDR_TABLE_STATIC_ADDR(dev->addr),
1083	       master->regs +
1084	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1085
1086	return 0;
1087}
1088
1089static void dw_i3c_master_detach_i2c_dev(struct i2c_dev_desc *dev)
1090{
1091	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1092	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1093	struct dw_i3c_master *master = to_dw_i3c_master(m);
1094
1095	writel(0,
1096	       master->regs +
1097	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1098
1099	i2c_dev_set_master_data(dev, NULL);
1100	master->devs[data->index].addr = 0;
1101	master->free_pos |= BIT(data->index);
1102	kfree(data);
1103}
1104
1105static int dw_i3c_master_request_ibi(struct i3c_dev_desc *dev,
1106				     const struct i3c_ibi_setup *req)
1107{
1108	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1109	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1110	struct dw_i3c_master *master = to_dw_i3c_master(m);
1111	unsigned long flags;
1112
1113	data->ibi_pool = i3c_generic_ibi_alloc_pool(dev, req);
1114	if (IS_ERR(data->ibi_pool))
1115		return PTR_ERR(data->ibi_pool);
1116
1117	spin_lock_irqsave(&master->devs_lock, flags);
1118	master->devs[data->index].ibi_dev = dev;
1119	spin_unlock_irqrestore(&master->devs_lock, flags);
1120
1121	return 0;
1122}
1123
1124static void dw_i3c_master_free_ibi(struct i3c_dev_desc *dev)
1125{
1126	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1127	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1128	struct dw_i3c_master *master = to_dw_i3c_master(m);
1129	unsigned long flags;
1130
1131	spin_lock_irqsave(&master->devs_lock, flags);
1132	master->devs[data->index].ibi_dev = NULL;
1133	spin_unlock_irqrestore(&master->devs_lock, flags);
1134
1135	i3c_generic_ibi_free_pool(data->ibi_pool);
1136	data->ibi_pool = NULL;
1137}
1138
1139static void dw_i3c_master_set_sir_enabled(struct dw_i3c_master *master,
1140					  struct i3c_dev_desc *dev,
1141					  u8 idx, bool enable)
1142{
1143	unsigned long flags;
1144	u32 dat_entry, reg;
1145	bool global;
1146
1147	dat_entry = DEV_ADDR_TABLE_LOC(master->datstartaddr, idx);
1148
1149	spin_lock_irqsave(&master->devs_lock, flags);
1150	reg = readl(master->regs + dat_entry);
1151	if (enable) {
1152		reg &= ~DEV_ADDR_TABLE_SIR_REJECT;
1153		if (dev->info.bcr & I3C_BCR_IBI_PAYLOAD)
1154			reg |= DEV_ADDR_TABLE_IBI_MDB;
1155	} else {
1156		reg |= DEV_ADDR_TABLE_SIR_REJECT;
1157	}
1158	master->platform_ops->set_dat_ibi(master, dev, enable, &reg);
1159	writel(reg, master->regs + dat_entry);
1160
1161	reg = readl(master->regs + IBI_SIR_REQ_REJECT);
1162	if (enable) {
1163		global = reg == 0xffffffff;
1164		reg &= ~BIT(idx);
1165	} else {
1166		global = reg == 0;
1167		reg |= BIT(idx);
1168	}
1169	writel(reg, master->regs + IBI_SIR_REQ_REJECT);
1170
1171	if (global) {
1172		reg = readl(master->regs + INTR_STATUS_EN);
1173		reg &= ~INTR_IBI_THLD_STAT;
1174		if (enable)
1175			reg |= INTR_IBI_THLD_STAT;
1176		writel(reg, master->regs + INTR_STATUS_EN);
1177
1178		reg = readl(master->regs + INTR_SIGNAL_EN);
1179		reg &= ~INTR_IBI_THLD_STAT;
1180		if (enable)
1181			reg |= INTR_IBI_THLD_STAT;
1182		writel(reg, master->regs + INTR_SIGNAL_EN);
1183	}
1184
1185	spin_unlock_irqrestore(&master->devs_lock, flags);
1186}
1187
1188static int dw_i3c_master_enable_ibi(struct i3c_dev_desc *dev)
1189{
1190	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1191	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1192	struct dw_i3c_master *master = to_dw_i3c_master(m);
1193	int rc;
1194
1195	dw_i3c_master_set_sir_enabled(master, dev, data->index, true);
1196
1197	rc = i3c_master_enec_locked(m, dev->info.dyn_addr, I3C_CCC_EVENT_SIR);
1198
1199	if (rc)
1200		dw_i3c_master_set_sir_enabled(master, dev, data->index, false);
1201
1202	return rc;
1203}
1204
1205static int dw_i3c_master_disable_ibi(struct i3c_dev_desc *dev)
1206{
1207	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1208	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1209	struct dw_i3c_master *master = to_dw_i3c_master(m);
1210	int rc;
1211
1212	rc = i3c_master_disec_locked(m, dev->info.dyn_addr, I3C_CCC_EVENT_SIR);
1213	if (rc)
1214		return rc;
1215
1216	dw_i3c_master_set_sir_enabled(master, dev, data->index, false);
1217
1218	return 0;
1219}
1220
1221static void dw_i3c_master_recycle_ibi_slot(struct i3c_dev_desc *dev,
1222					   struct i3c_ibi_slot *slot)
1223{
1224	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1225
1226	i3c_generic_ibi_recycle_slot(data->ibi_pool, slot);
1227}
1228
1229static void dw_i3c_master_drain_ibi_queue(struct dw_i3c_master *master,
1230					  int len)
1231{
1232	int i;
1233
1234	for (i = 0; i < DIV_ROUND_UP(len, 4); i++)
1235		readl(master->regs + IBI_QUEUE_STATUS);
1236}
1237
1238static void dw_i3c_master_handle_ibi_sir(struct dw_i3c_master *master,
1239					 u32 status)
1240{
1241	struct dw_i3c_i2c_dev_data *data;
1242	struct i3c_ibi_slot *slot;
1243	struct i3c_dev_desc *dev;
1244	unsigned long flags;
1245	u8 addr, len;
1246	int idx;
1247
1248	addr = IBI_QUEUE_IBI_ADDR(status);
1249	len = IBI_QUEUE_STATUS_DATA_LEN(status);
1250
1251	/*
1252	 * We be tempted to check the error status in bit 30; however, due
1253	 * to the PEC errata workaround on some platform implementations (see
1254	 * ast2600_i3c_set_dat_ibi()), those will almost always have a PEC
1255	 * error on IBI payload data, as well as losing the last byte of
1256	 * payload.
1257	 *
1258	 * If we implement error status checking on that bit, we may need
1259	 * a new platform op to validate it.
1260	 */
1261
1262	spin_lock_irqsave(&master->devs_lock, flags);
1263	idx = dw_i3c_master_get_addr_pos(master, addr);
1264	if (idx < 0) {
1265		dev_dbg_ratelimited(&master->base.dev,
1266			 "IBI from unknown addr 0x%x\n", addr);
1267		goto err_drain;
1268	}
1269
1270	dev = master->devs[idx].ibi_dev;
1271	if (!dev || !dev->ibi) {
1272		dev_dbg_ratelimited(&master->base.dev,
1273			 "IBI from non-requested dev idx %d\n", idx);
1274		goto err_drain;
1275	}
1276
1277	data = i3c_dev_get_master_data(dev);
1278	slot = i3c_generic_ibi_get_free_slot(data->ibi_pool);
1279	if (!slot) {
1280		dev_dbg_ratelimited(&master->base.dev,
1281				    "No IBI slots available\n");
1282		goto err_drain;
1283	}
1284
1285	if (dev->ibi->max_payload_len < len) {
1286		dev_dbg_ratelimited(&master->base.dev,
1287				    "IBI payload len %d greater than max %d\n",
1288				    len, dev->ibi->max_payload_len);
1289		goto err_drain;
1290	}
1291
1292	if (len) {
1293		dw_i3c_master_read_ibi_fifo(master, slot->data, len);
1294		slot->len = len;
1295	}
1296	i3c_master_queue_ibi(dev, slot);
1297
1298	spin_unlock_irqrestore(&master->devs_lock, flags);
1299
1300	return;
1301
1302err_drain:
1303	dw_i3c_master_drain_ibi_queue(master, len);
1304
1305	spin_unlock_irqrestore(&master->devs_lock, flags);
1306}
1307
1308/* "ibis": referring to In-Band Interrupts, and not
1309 * https://en.wikipedia.org/wiki/Australian_white_ibis. The latter should
1310 * not be handled.
1311 */
1312static void dw_i3c_master_irq_handle_ibis(struct dw_i3c_master *master)
1313{
1314	unsigned int i, len, n_ibis;
1315	u32 reg;
1316
1317	reg = readl(master->regs + QUEUE_STATUS_LEVEL);
1318	n_ibis = QUEUE_STATUS_IBI_STATUS_CNT(reg);
1319	if (!n_ibis)
1320		return;
1321
1322	for (i = 0; i < n_ibis; i++) {
1323		reg = readl(master->regs + IBI_QUEUE_STATUS);
1324
1325		if (IBI_TYPE_SIRQ(reg)) {
1326			dw_i3c_master_handle_ibi_sir(master, reg);
1327		} else {
1328			len = IBI_QUEUE_STATUS_DATA_LEN(reg);
1329			dev_info(&master->base.dev,
1330				 "unsupported IBI type 0x%lx len %d\n",
1331				 IBI_QUEUE_STATUS_IBI_ID(reg), len);
1332			dw_i3c_master_drain_ibi_queue(master, len);
1333		}
1334	}
1335}
1336
1337static irqreturn_t dw_i3c_master_irq_handler(int irq, void *dev_id)
1338{
1339	struct dw_i3c_master *master = dev_id;
1340	u32 status;
1341
1342	status = readl(master->regs + INTR_STATUS);
1343
1344	if (!(status & readl(master->regs + INTR_STATUS_EN))) {
1345		writel(INTR_ALL, master->regs + INTR_STATUS);
1346		return IRQ_NONE;
1347	}
1348
1349	spin_lock(&master->xferqueue.lock);
1350	dw_i3c_master_end_xfer_locked(master, status);
1351	if (status & INTR_TRANSFER_ERR_STAT)
1352		writel(INTR_TRANSFER_ERR_STAT, master->regs + INTR_STATUS);
1353	spin_unlock(&master->xferqueue.lock);
1354
1355	if (status & INTR_IBI_THLD_STAT)
1356		dw_i3c_master_irq_handle_ibis(master);
1357
1358	return IRQ_HANDLED;
1359}
1360
1361static const struct i3c_master_controller_ops dw_mipi_i3c_ops = {
1362	.bus_init = dw_i3c_master_bus_init,
1363	.bus_cleanup = dw_i3c_master_bus_cleanup,
1364	.attach_i3c_dev = dw_i3c_master_attach_i3c_dev,
1365	.reattach_i3c_dev = dw_i3c_master_reattach_i3c_dev,
1366	.detach_i3c_dev = dw_i3c_master_detach_i3c_dev,
1367	.do_daa = dw_i3c_master_daa,
1368	.supports_ccc_cmd = dw_i3c_master_supports_ccc_cmd,
1369	.send_ccc_cmd = dw_i3c_master_send_ccc_cmd,
1370	.priv_xfers = dw_i3c_master_priv_xfers,
1371	.attach_i2c_dev = dw_i3c_master_attach_i2c_dev,
1372	.detach_i2c_dev = dw_i3c_master_detach_i2c_dev,
1373	.i2c_xfers = dw_i3c_master_i2c_xfers,
1374};
1375
1376static const struct i3c_master_controller_ops dw_mipi_i3c_ibi_ops = {
1377	.bus_init = dw_i3c_master_bus_init,
1378	.bus_cleanup = dw_i3c_master_bus_cleanup,
1379	.attach_i3c_dev = dw_i3c_master_attach_i3c_dev,
1380	.reattach_i3c_dev = dw_i3c_master_reattach_i3c_dev,
1381	.detach_i3c_dev = dw_i3c_master_detach_i3c_dev,
1382	.do_daa = dw_i3c_master_daa,
1383	.supports_ccc_cmd = dw_i3c_master_supports_ccc_cmd,
1384	.send_ccc_cmd = dw_i3c_master_send_ccc_cmd,
1385	.priv_xfers = dw_i3c_master_priv_xfers,
1386	.attach_i2c_dev = dw_i3c_master_attach_i2c_dev,
1387	.detach_i2c_dev = dw_i3c_master_detach_i2c_dev,
1388	.i2c_xfers = dw_i3c_master_i2c_xfers,
1389	.request_ibi = dw_i3c_master_request_ibi,
1390	.free_ibi = dw_i3c_master_free_ibi,
1391	.enable_ibi = dw_i3c_master_enable_ibi,
1392	.disable_ibi = dw_i3c_master_disable_ibi,
1393	.recycle_ibi_slot = dw_i3c_master_recycle_ibi_slot,
1394};
1395
1396/* default platform ops implementations */
1397static int dw_i3c_platform_init_nop(struct dw_i3c_master *i3c)
1398{
1399	return 0;
1400}
1401
1402static void dw_i3c_platform_set_dat_ibi_nop(struct dw_i3c_master *i3c,
1403					struct i3c_dev_desc *dev,
1404					bool enable, u32 *dat)
1405{
1406}
1407
1408static const struct dw_i3c_platform_ops dw_i3c_platform_ops_default = {
1409	.init = dw_i3c_platform_init_nop,
1410	.set_dat_ibi = dw_i3c_platform_set_dat_ibi_nop,
1411};
1412
1413int dw_i3c_common_probe(struct dw_i3c_master *master,
1414			struct platform_device *pdev)
1415{
1416	const struct i3c_master_controller_ops *ops;
1417	int ret, irq;
1418
1419	if (!master->platform_ops)
1420		master->platform_ops = &dw_i3c_platform_ops_default;
 
1421
1422	master->regs = devm_platform_ioremap_resource(pdev, 0);
1423	if (IS_ERR(master->regs))
1424		return PTR_ERR(master->regs);
1425
1426	master->core_clk = devm_clk_get(&pdev->dev, NULL);
1427	if (IS_ERR(master->core_clk))
1428		return PTR_ERR(master->core_clk);
1429
1430	master->core_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
1431								    "core_rst");
1432	if (IS_ERR(master->core_rst))
1433		return PTR_ERR(master->core_rst);
1434
1435	ret = clk_prepare_enable(master->core_clk);
1436	if (ret)
1437		goto err_disable_core_clk;
1438
1439	reset_control_deassert(master->core_rst);
1440
1441	spin_lock_init(&master->xferqueue.lock);
1442	INIT_LIST_HEAD(&master->xferqueue.list);
1443
1444	writel(INTR_ALL, master->regs + INTR_STATUS);
1445	irq = platform_get_irq(pdev, 0);
1446	ret = devm_request_irq(&pdev->dev, irq,
1447			       dw_i3c_master_irq_handler, 0,
1448			       dev_name(&pdev->dev), master);
1449	if (ret)
1450		goto err_assert_rst;
1451
1452	platform_set_drvdata(pdev, master);
1453
1454	/* Information regarding the FIFOs/QUEUEs depth */
1455	ret = readl(master->regs + QUEUE_STATUS_LEVEL);
1456	master->caps.cmdfifodepth = QUEUE_STATUS_LEVEL_CMD(ret);
1457
1458	ret = readl(master->regs + DATA_BUFFER_STATUS_LEVEL);
1459	master->caps.datafifodepth = DATA_BUFFER_STATUS_LEVEL_TX(ret);
1460
1461	ret = readl(master->regs + DEVICE_ADDR_TABLE_POINTER);
1462	master->datstartaddr = ret;
1463	master->maxdevs = ret >> 16;
1464	master->free_pos = GENMASK(master->maxdevs - 1, 0);
1465
1466	ops = &dw_mipi_i3c_ops;
1467	if (master->ibi_capable)
1468		ops = &dw_mipi_i3c_ibi_ops;
1469
1470	ret = i3c_master_register(&master->base, &pdev->dev, ops, false);
1471	if (ret)
1472		goto err_assert_rst;
1473
1474	return 0;
1475
1476err_assert_rst:
1477	reset_control_assert(master->core_rst);
1478
1479err_disable_core_clk:
1480	clk_disable_unprepare(master->core_clk);
1481
1482	return ret;
1483}
1484EXPORT_SYMBOL_GPL(dw_i3c_common_probe);
1485
1486void dw_i3c_common_remove(struct dw_i3c_master *master)
1487{
1488	i3c_master_unregister(&master->base);
 
 
 
 
 
1489
1490	reset_control_assert(master->core_rst);
1491
1492	clk_disable_unprepare(master->core_clk);
1493}
1494EXPORT_SYMBOL_GPL(dw_i3c_common_remove);
1495
1496/* base platform implementation */
1497
1498static int dw_i3c_probe(struct platform_device *pdev)
1499{
1500	struct dw_i3c_master *master;
1501
1502	master = devm_kzalloc(&pdev->dev, sizeof(*master), GFP_KERNEL);
1503	if (!master)
1504		return -ENOMEM;
1505
1506	return dw_i3c_common_probe(master, pdev);
1507}
1508
1509static void dw_i3c_remove(struct platform_device *pdev)
1510{
1511	struct dw_i3c_master *master = platform_get_drvdata(pdev);
1512
1513	dw_i3c_common_remove(master);
1514}
1515
1516static const struct of_device_id dw_i3c_master_of_match[] = {
1517	{ .compatible = "snps,dw-i3c-master-1.00a", },
1518	{},
1519};
1520MODULE_DEVICE_TABLE(of, dw_i3c_master_of_match);
1521
1522static struct platform_driver dw_i3c_driver = {
1523	.probe = dw_i3c_probe,
1524	.remove_new = dw_i3c_remove,
1525	.driver = {
1526		.name = "dw-i3c-master",
1527		.of_match_table = dw_i3c_master_of_match,
1528	},
1529};
1530module_platform_driver(dw_i3c_driver);
1531
1532MODULE_AUTHOR("Vitor Soares <vitor.soares@synopsys.com>");
1533MODULE_DESCRIPTION("DesignWare MIPI I3C driver");
1534MODULE_LICENSE("GPL v2");