Linux Audio

Check our new training course

Loading...
v6.2
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
   4 *
   5 * Author: Vitor Soares <vitor.soares@synopsys.com>
   6 */
   7
   8#include <linux/bitops.h>
   9#include <linux/clk.h>
  10#include <linux/completion.h>
  11#include <linux/err.h>
  12#include <linux/errno.h>
  13#include <linux/i3c/master.h>
  14#include <linux/interrupt.h>
  15#include <linux/ioport.h>
  16#include <linux/iopoll.h>
  17#include <linux/list.h>
  18#include <linux/module.h>
  19#include <linux/of.h>
 
  20#include <linux/platform_device.h>
 
  21#include <linux/reset.h>
  22#include <linux/slab.h>
  23
 
 
  24#define DEVICE_CTRL			0x0
  25#define DEV_CTRL_ENABLE			BIT(31)
  26#define DEV_CTRL_RESUME			BIT(30)
  27#define DEV_CTRL_HOT_JOIN_NACK		BIT(8)
  28#define DEV_CTRL_I2C_SLAVE_PRESENT	BIT(7)
  29
  30#define DEVICE_ADDR			0x4
  31#define DEV_ADDR_DYNAMIC_ADDR_VALID	BIT(31)
  32#define DEV_ADDR_DYNAMIC(x)		(((x) << 16) & GENMASK(22, 16))
  33
  34#define HW_CAPABILITY			0x8
  35#define COMMAND_QUEUE_PORT		0xc
  36#define COMMAND_PORT_TOC		BIT(30)
  37#define COMMAND_PORT_READ_TRANSFER	BIT(28)
  38#define COMMAND_PORT_SDAP		BIT(27)
  39#define COMMAND_PORT_ROC		BIT(26)
  40#define COMMAND_PORT_SPEED(x)		(((x) << 21) & GENMASK(23, 21))
  41#define COMMAND_PORT_DEV_INDEX(x)	(((x) << 16) & GENMASK(20, 16))
  42#define COMMAND_PORT_CP			BIT(15)
  43#define COMMAND_PORT_CMD(x)		(((x) << 7) & GENMASK(14, 7))
  44#define COMMAND_PORT_TID(x)		(((x) << 3) & GENMASK(6, 3))
  45
  46#define COMMAND_PORT_ARG_DATA_LEN(x)	(((x) << 16) & GENMASK(31, 16))
  47#define COMMAND_PORT_ARG_DATA_LEN_MAX	65536
  48#define COMMAND_PORT_TRANSFER_ARG	0x01
  49
  50#define COMMAND_PORT_SDA_DATA_BYTE_3(x)	(((x) << 24) & GENMASK(31, 24))
  51#define COMMAND_PORT_SDA_DATA_BYTE_2(x)	(((x) << 16) & GENMASK(23, 16))
  52#define COMMAND_PORT_SDA_DATA_BYTE_1(x)	(((x) << 8) & GENMASK(15, 8))
  53#define COMMAND_PORT_SDA_BYTE_STRB_3	BIT(5)
  54#define COMMAND_PORT_SDA_BYTE_STRB_2	BIT(4)
  55#define COMMAND_PORT_SDA_BYTE_STRB_1	BIT(3)
  56#define COMMAND_PORT_SHORT_DATA_ARG	0x02
  57
  58#define COMMAND_PORT_DEV_COUNT(x)	(((x) << 21) & GENMASK(25, 21))
  59#define COMMAND_PORT_ADDR_ASSGN_CMD	0x03
  60
  61#define RESPONSE_QUEUE_PORT		0x10
  62#define RESPONSE_PORT_ERR_STATUS(x)	(((x) & GENMASK(31, 28)) >> 28)
  63#define RESPONSE_NO_ERROR		0
  64#define RESPONSE_ERROR_CRC		1
  65#define RESPONSE_ERROR_PARITY		2
  66#define RESPONSE_ERROR_FRAME		3
  67#define RESPONSE_ERROR_IBA_NACK		4
  68#define RESPONSE_ERROR_ADDRESS_NACK	5
  69#define RESPONSE_ERROR_OVER_UNDER_FLOW	6
  70#define RESPONSE_ERROR_TRANSF_ABORT	8
  71#define RESPONSE_ERROR_I2C_W_NACK_ERR	9
  72#define RESPONSE_PORT_TID(x)		(((x) & GENMASK(27, 24)) >> 24)
  73#define RESPONSE_PORT_DATA_LEN(x)	((x) & GENMASK(15, 0))
  74
  75#define RX_TX_DATA_PORT			0x14
  76#define IBI_QUEUE_STATUS		0x18
 
 
 
 
 
 
 
 
 
 
 
  77#define QUEUE_THLD_CTRL			0x1c
 
 
 
 
  78#define QUEUE_THLD_CTRL_RESP_BUF_MASK	GENMASK(15, 8)
  79#define QUEUE_THLD_CTRL_RESP_BUF(x)	(((x) - 1) << 8)
  80
  81#define DATA_BUFFER_THLD_CTRL		0x20
  82#define DATA_BUFFER_THLD_CTRL_RX_BUF	GENMASK(11, 8)
  83
  84#define IBI_QUEUE_CTRL			0x24
  85#define IBI_MR_REQ_REJECT		0x2C
  86#define IBI_SIR_REQ_REJECT		0x30
  87#define IBI_REQ_REJECT_ALL		GENMASK(31, 0)
  88
  89#define RESET_CTRL			0x34
  90#define RESET_CTRL_IBI_QUEUE		BIT(5)
  91#define RESET_CTRL_RX_FIFO		BIT(4)
  92#define RESET_CTRL_TX_FIFO		BIT(3)
  93#define RESET_CTRL_RESP_QUEUE		BIT(2)
  94#define RESET_CTRL_CMD_QUEUE		BIT(1)
  95#define RESET_CTRL_SOFT			BIT(0)
  96
  97#define SLV_EVENT_CTRL			0x38
  98#define INTR_STATUS			0x3c
  99#define INTR_STATUS_EN			0x40
 100#define INTR_SIGNAL_EN			0x44
 101#define INTR_FORCE			0x48
 102#define INTR_BUSOWNER_UPDATE_STAT	BIT(13)
 103#define INTR_IBI_UPDATED_STAT		BIT(12)
 104#define INTR_READ_REQ_RECV_STAT		BIT(11)
 105#define INTR_DEFSLV_STAT		BIT(10)
 106#define INTR_TRANSFER_ERR_STAT		BIT(9)
 107#define INTR_DYN_ADDR_ASSGN_STAT	BIT(8)
 108#define INTR_CCC_UPDATED_STAT		BIT(6)
 109#define INTR_TRANSFER_ABORT_STAT	BIT(5)
 110#define INTR_RESP_READY_STAT		BIT(4)
 111#define INTR_CMD_QUEUE_READY_STAT	BIT(3)
 112#define INTR_IBI_THLD_STAT		BIT(2)
 113#define INTR_RX_THLD_STAT		BIT(1)
 114#define INTR_TX_THLD_STAT		BIT(0)
 115#define INTR_ALL			(INTR_BUSOWNER_UPDATE_STAT |	\
 116					INTR_IBI_UPDATED_STAT |		\
 117					INTR_READ_REQ_RECV_STAT |	\
 118					INTR_DEFSLV_STAT |		\
 119					INTR_TRANSFER_ERR_STAT |	\
 120					INTR_DYN_ADDR_ASSGN_STAT |	\
 121					INTR_CCC_UPDATED_STAT |		\
 122					INTR_TRANSFER_ABORT_STAT |	\
 123					INTR_RESP_READY_STAT |		\
 124					INTR_CMD_QUEUE_READY_STAT |	\
 125					INTR_IBI_THLD_STAT |		\
 126					INTR_TX_THLD_STAT |		\
 127					INTR_RX_THLD_STAT)
 128
 129#define INTR_MASTER_MASK		(INTR_TRANSFER_ERR_STAT |	\
 130					 INTR_RESP_READY_STAT)
 131
 132#define QUEUE_STATUS_LEVEL		0x4c
 133#define QUEUE_STATUS_IBI_STATUS_CNT(x)	(((x) & GENMASK(28, 24)) >> 24)
 134#define QUEUE_STATUS_IBI_BUF_BLR(x)	(((x) & GENMASK(23, 16)) >> 16)
 135#define QUEUE_STATUS_LEVEL_RESP(x)	(((x) & GENMASK(15, 8)) >> 8)
 136#define QUEUE_STATUS_LEVEL_CMD(x)	((x) & GENMASK(7, 0))
 137
 138#define DATA_BUFFER_STATUS_LEVEL	0x50
 139#define DATA_BUFFER_STATUS_LEVEL_TX(x)	((x) & GENMASK(7, 0))
 140
 141#define PRESENT_STATE			0x54
 142#define CCC_DEVICE_STATUS		0x58
 143#define DEVICE_ADDR_TABLE_POINTER	0x5c
 144#define DEVICE_ADDR_TABLE_DEPTH(x)	(((x) & GENMASK(31, 16)) >> 16)
 145#define DEVICE_ADDR_TABLE_ADDR(x)	((x) & GENMASK(7, 0))
 146
 147#define DEV_CHAR_TABLE_POINTER		0x60
 148#define VENDOR_SPECIFIC_REG_POINTER	0x6c
 149#define SLV_PID_VALUE			0x74
 150#define SLV_CHAR_CTRL			0x78
 151#define SLV_MAX_LEN			0x7c
 152#define MAX_READ_TURNAROUND		0x80
 153#define MAX_DATA_SPEED			0x84
 154#define SLV_DEBUG_STATUS		0x88
 155#define SLV_INTR_REQ			0x8c
 156#define DEVICE_CTRL_EXTENDED		0xb0
 157#define SCL_I3C_OD_TIMING		0xb4
 158#define SCL_I3C_PP_TIMING		0xb8
 159#define SCL_I3C_TIMING_HCNT(x)		(((x) << 16) & GENMASK(23, 16))
 160#define SCL_I3C_TIMING_LCNT(x)		((x) & GENMASK(7, 0))
 161#define SCL_I3C_TIMING_CNT_MIN		5
 162
 163#define SCL_I2C_FM_TIMING		0xbc
 164#define SCL_I2C_FM_TIMING_HCNT(x)	(((x) << 16) & GENMASK(31, 16))
 165#define SCL_I2C_FM_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 166
 167#define SCL_I2C_FMP_TIMING		0xc0
 168#define SCL_I2C_FMP_TIMING_HCNT(x)	(((x) << 16) & GENMASK(23, 16))
 169#define SCL_I2C_FMP_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 170
 171#define SCL_EXT_LCNT_TIMING		0xc8
 172#define SCL_EXT_LCNT_4(x)		(((x) << 24) & GENMASK(31, 24))
 173#define SCL_EXT_LCNT_3(x)		(((x) << 16) & GENMASK(23, 16))
 174#define SCL_EXT_LCNT_2(x)		(((x) << 8) & GENMASK(15, 8))
 175#define SCL_EXT_LCNT_1(x)		((x) & GENMASK(7, 0))
 176
 177#define SCL_EXT_TERMN_LCNT_TIMING	0xcc
 178#define BUS_FREE_TIMING			0xd4
 179#define BUS_I3C_MST_FREE(x)		((x) & GENMASK(15, 0))
 180
 181#define BUS_IDLE_TIMING			0xd8
 182#define I3C_VER_ID			0xe0
 183#define I3C_VER_TYPE			0xe4
 184#define EXTENDED_CAPABILITY		0xe8
 185#define SLAVE_CONFIG			0xec
 186
 
 
 187#define DEV_ADDR_TABLE_LEGACY_I2C_DEV	BIT(31)
 188#define DEV_ADDR_TABLE_DYNAMIC_ADDR(x)	(((x) << 16) & GENMASK(23, 16))
 189#define DEV_ADDR_TABLE_STATIC_ADDR(x)	((x) & GENMASK(6, 0))
 190#define DEV_ADDR_TABLE_LOC(start, idx)	((start) + ((idx) << 2))
 191
 192#define MAX_DEVS 32
 193
 194#define I3C_BUS_SDR1_SCL_RATE		8000000
 195#define I3C_BUS_SDR2_SCL_RATE		6000000
 196#define I3C_BUS_SDR3_SCL_RATE		4000000
 197#define I3C_BUS_SDR4_SCL_RATE		2000000
 198#define I3C_BUS_I2C_FM_TLOW_MIN_NS	1300
 199#define I3C_BUS_I2C_FMP_TLOW_MIN_NS	500
 200#define I3C_BUS_THIGH_MAX_NS		41
 201
 202#define XFER_TIMEOUT (msecs_to_jiffies(1000))
 
 203
 204struct dw_i3c_master_caps {
 205	u8 cmdfifodepth;
 206	u8 datafifodepth;
 207};
 
 
 208
 209struct dw_i3c_cmd {
 210	u32 cmd_lo;
 211	u32 cmd_hi;
 212	u16 tx_len;
 213	const void *tx_buf;
 214	u16 rx_len;
 215	void *rx_buf;
 216	u8 error;
 217};
 218
 219struct dw_i3c_xfer {
 220	struct list_head node;
 221	struct completion comp;
 222	int ret;
 223	unsigned int ncmds;
 224	struct dw_i3c_cmd cmds[];
 225};
 226
 227struct dw_i3c_master {
 228	struct i3c_master_controller base;
 229	u16 maxdevs;
 230	u16 datstartaddr;
 231	u32 free_pos;
 232	struct {
 233		struct list_head list;
 234		struct dw_i3c_xfer *cur;
 235		spinlock_t lock;
 236	} xferqueue;
 237	struct dw_i3c_master_caps caps;
 238	void __iomem *regs;
 239	struct reset_control *core_rst;
 240	struct clk *core_clk;
 241	char version[5];
 242	char type[5];
 243	u8 addrs[MAX_DEVS];
 244};
 245
 246struct dw_i3c_i2c_dev_data {
 247	u8 index;
 
 248};
 249
 250static u8 even_parity(u8 p)
 251{
 252	p ^= p >> 4;
 253	p &= 0xf;
 254
 255	return (0x9669 >> p) & 1;
 256}
 257
 258static bool dw_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m,
 259					   const struct i3c_ccc_cmd *cmd)
 260{
 261	if (cmd->ndests > 1)
 262		return false;
 263
 264	switch (cmd->id) {
 265	case I3C_CCC_ENEC(true):
 266	case I3C_CCC_ENEC(false):
 267	case I3C_CCC_DISEC(true):
 268	case I3C_CCC_DISEC(false):
 269	case I3C_CCC_ENTAS(0, true):
 270	case I3C_CCC_ENTAS(0, false):
 271	case I3C_CCC_RSTDAA(true):
 272	case I3C_CCC_RSTDAA(false):
 273	case I3C_CCC_ENTDAA:
 274	case I3C_CCC_SETMWL(true):
 275	case I3C_CCC_SETMWL(false):
 276	case I3C_CCC_SETMRL(true):
 277	case I3C_CCC_SETMRL(false):
 278	case I3C_CCC_ENTHDR(0):
 279	case I3C_CCC_SETDASA:
 280	case I3C_CCC_SETNEWDA:
 281	case I3C_CCC_GETMWL:
 282	case I3C_CCC_GETMRL:
 283	case I3C_CCC_GETPID:
 284	case I3C_CCC_GETBCR:
 285	case I3C_CCC_GETDCR:
 286	case I3C_CCC_GETSTATUS:
 287	case I3C_CCC_GETMXDS:
 288	case I3C_CCC_GETHDRCAP:
 289		return true;
 290	default:
 291		return false;
 292	}
 293}
 294
 295static inline struct dw_i3c_master *
 296to_dw_i3c_master(struct i3c_master_controller *master)
 297{
 298	return container_of(master, struct dw_i3c_master, base);
 299}
 300
 301static void dw_i3c_master_disable(struct dw_i3c_master *master)
 302{
 303	writel(readl(master->regs + DEVICE_CTRL) & ~DEV_CTRL_ENABLE,
 304	       master->regs + DEVICE_CTRL);
 305}
 306
 307static void dw_i3c_master_enable(struct dw_i3c_master *master)
 308{
 309	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_ENABLE,
 
 
 
 
 
 
 
 310	       master->regs + DEVICE_CTRL);
 311}
 312
 313static int dw_i3c_master_get_addr_pos(struct dw_i3c_master *master, u8 addr)
 314{
 315	int pos;
 316
 317	for (pos = 0; pos < master->maxdevs; pos++) {
 318		if (addr == master->addrs[pos])
 319			return pos;
 320	}
 321
 322	return -EINVAL;
 323}
 324
 325static int dw_i3c_master_get_free_pos(struct dw_i3c_master *master)
 326{
 327	if (!(master->free_pos & GENMASK(master->maxdevs - 1, 0)))
 328		return -ENOSPC;
 329
 330	return ffs(master->free_pos) - 1;
 331}
 332
 333static void dw_i3c_master_wr_tx_fifo(struct dw_i3c_master *master,
 334				     const u8 *bytes, int nbytes)
 335{
 336	writesl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 337	if (nbytes & 3) {
 338		u32 tmp = 0;
 339
 340		memcpy(&tmp, bytes + (nbytes & ~3), nbytes & 3);
 341		writesl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 342	}
 343}
 344
 345static void dw_i3c_master_read_rx_fifo(struct dw_i3c_master *master,
 346				       u8 *bytes, int nbytes)
 347{
 348	readsl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 349	if (nbytes & 3) {
 350		u32 tmp;
 351
 352		readsl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 353		memcpy(bytes + (nbytes & ~3), &tmp, nbytes & 3);
 354	}
 355}
 356
 
 
 
 
 
 
 
 
 
 
 
 
 357static struct dw_i3c_xfer *
 358dw_i3c_master_alloc_xfer(struct dw_i3c_master *master, unsigned int ncmds)
 359{
 360	struct dw_i3c_xfer *xfer;
 361
 362	xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL);
 363	if (!xfer)
 364		return NULL;
 365
 366	INIT_LIST_HEAD(&xfer->node);
 367	xfer->ncmds = ncmds;
 368	xfer->ret = -ETIMEDOUT;
 369
 370	return xfer;
 371}
 372
 373static void dw_i3c_master_free_xfer(struct dw_i3c_xfer *xfer)
 374{
 375	kfree(xfer);
 376}
 377
 378static void dw_i3c_master_start_xfer_locked(struct dw_i3c_master *master)
 379{
 380	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 381	unsigned int i;
 382	u32 thld_ctrl;
 383
 384	if (!xfer)
 385		return;
 386
 387	for (i = 0; i < xfer->ncmds; i++) {
 388		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 389
 390		dw_i3c_master_wr_tx_fifo(master, cmd->tx_buf, cmd->tx_len);
 391	}
 392
 393	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 394	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 395	thld_ctrl |= QUEUE_THLD_CTRL_RESP_BUF(xfer->ncmds);
 396	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 397
 398	for (i = 0; i < xfer->ncmds; i++) {
 399		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 400
 401		writel(cmd->cmd_hi, master->regs + COMMAND_QUEUE_PORT);
 402		writel(cmd->cmd_lo, master->regs + COMMAND_QUEUE_PORT);
 403	}
 404}
 405
 406static void dw_i3c_master_enqueue_xfer(struct dw_i3c_master *master,
 407				       struct dw_i3c_xfer *xfer)
 408{
 409	unsigned long flags;
 410
 411	init_completion(&xfer->comp);
 412	spin_lock_irqsave(&master->xferqueue.lock, flags);
 413	if (master->xferqueue.cur) {
 414		list_add_tail(&xfer->node, &master->xferqueue.list);
 415	} else {
 416		master->xferqueue.cur = xfer;
 417		dw_i3c_master_start_xfer_locked(master);
 418	}
 419	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 420}
 421
 422static void dw_i3c_master_dequeue_xfer_locked(struct dw_i3c_master *master,
 423					      struct dw_i3c_xfer *xfer)
 424{
 425	if (master->xferqueue.cur == xfer) {
 426		u32 status;
 427
 428		master->xferqueue.cur = NULL;
 429
 430		writel(RESET_CTRL_RX_FIFO | RESET_CTRL_TX_FIFO |
 431		       RESET_CTRL_RESP_QUEUE | RESET_CTRL_CMD_QUEUE,
 432		       master->regs + RESET_CTRL);
 433
 434		readl_poll_timeout_atomic(master->regs + RESET_CTRL, status,
 435					  !status, 10, 1000000);
 436	} else {
 437		list_del_init(&xfer->node);
 438	}
 439}
 440
 441static void dw_i3c_master_dequeue_xfer(struct dw_i3c_master *master,
 442				       struct dw_i3c_xfer *xfer)
 443{
 444	unsigned long flags;
 445
 446	spin_lock_irqsave(&master->xferqueue.lock, flags);
 447	dw_i3c_master_dequeue_xfer_locked(master, xfer);
 448	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 449}
 450
 451static void dw_i3c_master_end_xfer_locked(struct dw_i3c_master *master, u32 isr)
 452{
 453	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 454	int i, ret = 0;
 455	u32 nresp;
 456
 457	if (!xfer)
 458		return;
 459
 460	nresp = readl(master->regs + QUEUE_STATUS_LEVEL);
 461	nresp = QUEUE_STATUS_LEVEL_RESP(nresp);
 462
 463	for (i = 0; i < nresp; i++) {
 464		struct dw_i3c_cmd *cmd;
 465		u32 resp;
 466
 467		resp = readl(master->regs + RESPONSE_QUEUE_PORT);
 468
 469		cmd = &xfer->cmds[RESPONSE_PORT_TID(resp)];
 470		cmd->rx_len = RESPONSE_PORT_DATA_LEN(resp);
 471		cmd->error = RESPONSE_PORT_ERR_STATUS(resp);
 472		if (cmd->rx_len && !cmd->error)
 473			dw_i3c_master_read_rx_fifo(master, cmd->rx_buf,
 474						   cmd->rx_len);
 475	}
 476
 477	for (i = 0; i < nresp; i++) {
 478		switch (xfer->cmds[i].error) {
 479		case RESPONSE_NO_ERROR:
 480			break;
 481		case RESPONSE_ERROR_PARITY:
 482		case RESPONSE_ERROR_IBA_NACK:
 483		case RESPONSE_ERROR_TRANSF_ABORT:
 484		case RESPONSE_ERROR_CRC:
 485		case RESPONSE_ERROR_FRAME:
 486			ret = -EIO;
 487			break;
 488		case RESPONSE_ERROR_OVER_UNDER_FLOW:
 489			ret = -ENOSPC;
 490			break;
 491		case RESPONSE_ERROR_I2C_W_NACK_ERR:
 492		case RESPONSE_ERROR_ADDRESS_NACK:
 493		default:
 494			ret = -EINVAL;
 495			break;
 496		}
 497	}
 498
 499	xfer->ret = ret;
 500	complete(&xfer->comp);
 501
 502	if (ret < 0) {
 503		dw_i3c_master_dequeue_xfer_locked(master, xfer);
 504		writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_RESUME,
 505		       master->regs + DEVICE_CTRL);
 506	}
 507
 508	xfer = list_first_entry_or_null(&master->xferqueue.list,
 509					struct dw_i3c_xfer,
 510					node);
 511	if (xfer)
 512		list_del_init(&xfer->node);
 513
 514	master->xferqueue.cur = xfer;
 515	dw_i3c_master_start_xfer_locked(master);
 516}
 517
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 518static int dw_i3c_clk_cfg(struct dw_i3c_master *master)
 519{
 520	unsigned long core_rate, core_period;
 521	u32 scl_timing;
 522	u8 hcnt, lcnt;
 523
 524	core_rate = clk_get_rate(master->core_clk);
 525	if (!core_rate)
 526		return -EINVAL;
 527
 528	core_period = DIV_ROUND_UP(1000000000, core_rate);
 529
 530	hcnt = DIV_ROUND_UP(I3C_BUS_THIGH_MAX_NS, core_period) - 1;
 531	if (hcnt < SCL_I3C_TIMING_CNT_MIN)
 532		hcnt = SCL_I3C_TIMING_CNT_MIN;
 533
 534	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_TYP_I3C_SCL_RATE) - hcnt;
 535	if (lcnt < SCL_I3C_TIMING_CNT_MIN)
 536		lcnt = SCL_I3C_TIMING_CNT_MIN;
 537
 538	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 539	writel(scl_timing, master->regs + SCL_I3C_PP_TIMING);
 
 540
 541	if (!(readl(master->regs + DEVICE_CTRL) & DEV_CTRL_I2C_SLAVE_PRESENT))
 
 
 
 
 542		writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 
 
 543
 544	lcnt = DIV_ROUND_UP(I3C_BUS_TLOW_OD_MIN_NS, core_period);
 
 545	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 546	writel(scl_timing, master->regs + SCL_I3C_OD_TIMING);
 
 547
 548	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR1_SCL_RATE) - hcnt;
 549	scl_timing = SCL_EXT_LCNT_1(lcnt);
 550	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR2_SCL_RATE) - hcnt;
 551	scl_timing |= SCL_EXT_LCNT_2(lcnt);
 552	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR3_SCL_RATE) - hcnt;
 553	scl_timing |= SCL_EXT_LCNT_3(lcnt);
 554	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR4_SCL_RATE) - hcnt;
 555	scl_timing |= SCL_EXT_LCNT_4(lcnt);
 556	writel(scl_timing, master->regs + SCL_EXT_LCNT_TIMING);
 
 557
 558	return 0;
 559}
 560
 561static int dw_i2c_clk_cfg(struct dw_i3c_master *master)
 562{
 563	unsigned long core_rate, core_period;
 564	u16 hcnt, lcnt;
 565	u32 scl_timing;
 566
 567	core_rate = clk_get_rate(master->core_clk);
 568	if (!core_rate)
 569		return -EINVAL;
 570
 571	core_period = DIV_ROUND_UP(1000000000, core_rate);
 572
 573	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FMP_TLOW_MIN_NS, core_period);
 574	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_PLUS_SCL_RATE) - lcnt;
 575	scl_timing = SCL_I2C_FMP_TIMING_HCNT(hcnt) |
 576		     SCL_I2C_FMP_TIMING_LCNT(lcnt);
 577	writel(scl_timing, master->regs + SCL_I2C_FMP_TIMING);
 
 578
 579	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FM_TLOW_MIN_NS, core_period);
 580	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_SCL_RATE) - lcnt;
 581	scl_timing = SCL_I2C_FM_TIMING_HCNT(hcnt) |
 582		     SCL_I2C_FM_TIMING_LCNT(lcnt);
 583	writel(scl_timing, master->regs + SCL_I2C_FM_TIMING);
 
 584
 585	writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 
 
 586	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_I2C_SLAVE_PRESENT,
 587	       master->regs + DEVICE_CTRL);
 
 588
 589	return 0;
 590}
 591
 592static int dw_i3c_master_bus_init(struct i3c_master_controller *m)
 593{
 594	struct dw_i3c_master *master = to_dw_i3c_master(m);
 595	struct i3c_bus *bus = i3c_master_get_bus(m);
 596	struct i3c_device_info info = { };
 597	u32 thld_ctrl;
 598	int ret;
 599
 
 
 
 
 
 
 
 
 
 
 
 
 600	switch (bus->mode) {
 601	case I3C_BUS_MODE_MIXED_FAST:
 602	case I3C_BUS_MODE_MIXED_LIMITED:
 603		ret = dw_i2c_clk_cfg(master);
 604		if (ret)
 605			return ret;
 606		fallthrough;
 607	case I3C_BUS_MODE_PURE:
 608		ret = dw_i3c_clk_cfg(master);
 609		if (ret)
 610			return ret;
 611		break;
 612	default:
 613		return -EINVAL;
 
 614	}
 615
 616	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 617	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 618	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 619
 620	thld_ctrl = readl(master->regs + DATA_BUFFER_THLD_CTRL);
 621	thld_ctrl &= ~DATA_BUFFER_THLD_CTRL_RX_BUF;
 622	writel(thld_ctrl, master->regs + DATA_BUFFER_THLD_CTRL);
 623
 624	writel(INTR_ALL, master->regs + INTR_STATUS);
 625	writel(INTR_MASTER_MASK, master->regs + INTR_STATUS_EN);
 626	writel(INTR_MASTER_MASK, master->regs + INTR_SIGNAL_EN);
 627
 628	ret = i3c_master_get_free_addr(m, 0);
 629	if (ret < 0)
 630		return ret;
 631
 632	writel(DEV_ADDR_DYNAMIC_ADDR_VALID | DEV_ADDR_DYNAMIC(ret),
 633	       master->regs + DEVICE_ADDR);
 634
 635	memset(&info, 0, sizeof(info));
 636	info.dyn_addr = ret;
 637
 638	ret = i3c_master_set_info(&master->base, &info);
 639	if (ret)
 640		return ret;
 641
 642	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_SIR_REQ_REJECT);
 643	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_MR_REQ_REJECT);
 644
 645	/* For now don't support Hot-Join */
 646	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_HOT_JOIN_NACK,
 647	       master->regs + DEVICE_CTRL);
 648
 
 649	dw_i3c_master_enable(master);
 650
 651	return 0;
 
 
 
 652}
 653
 654static void dw_i3c_master_bus_cleanup(struct i3c_master_controller *m)
 655{
 656	struct dw_i3c_master *master = to_dw_i3c_master(m);
 657
 658	dw_i3c_master_disable(master);
 659}
 660
 661static int dw_i3c_ccc_set(struct dw_i3c_master *master,
 662			  struct i3c_ccc_cmd *ccc)
 663{
 664	struct dw_i3c_xfer *xfer;
 665	struct dw_i3c_cmd *cmd;
 666	int ret, pos = 0;
 667
 668	if (ccc->id & I3C_CCC_DIRECT) {
 669		pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 670		if (pos < 0)
 671			return pos;
 672	}
 673
 674	xfer = dw_i3c_master_alloc_xfer(master, 1);
 675	if (!xfer)
 676		return -ENOMEM;
 677
 678	cmd = xfer->cmds;
 679	cmd->tx_buf = ccc->dests[0].payload.data;
 680	cmd->tx_len = ccc->dests[0].payload.len;
 681
 682	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 683		      COMMAND_PORT_TRANSFER_ARG;
 684
 685	cmd->cmd_lo = COMMAND_PORT_CP |
 686		      COMMAND_PORT_DEV_INDEX(pos) |
 687		      COMMAND_PORT_CMD(ccc->id) |
 688		      COMMAND_PORT_TOC |
 689		      COMMAND_PORT_ROC;
 690
 691	dw_i3c_master_enqueue_xfer(master, xfer);
 692	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 693		dw_i3c_master_dequeue_xfer(master, xfer);
 694
 695	ret = xfer->ret;
 696	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 697		ccc->err = I3C_ERROR_M2;
 698
 699	dw_i3c_master_free_xfer(xfer);
 700
 701	return ret;
 702}
 703
 704static int dw_i3c_ccc_get(struct dw_i3c_master *master, struct i3c_ccc_cmd *ccc)
 705{
 706	struct dw_i3c_xfer *xfer;
 707	struct dw_i3c_cmd *cmd;
 708	int ret, pos;
 709
 710	pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 711	if (pos < 0)
 712		return pos;
 713
 714	xfer = dw_i3c_master_alloc_xfer(master, 1);
 715	if (!xfer)
 716		return -ENOMEM;
 717
 718	cmd = xfer->cmds;
 719	cmd->rx_buf = ccc->dests[0].payload.data;
 720	cmd->rx_len = ccc->dests[0].payload.len;
 721
 722	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 723		      COMMAND_PORT_TRANSFER_ARG;
 724
 725	cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 726		      COMMAND_PORT_CP |
 727		      COMMAND_PORT_DEV_INDEX(pos) |
 728		      COMMAND_PORT_CMD(ccc->id) |
 729		      COMMAND_PORT_TOC |
 730		      COMMAND_PORT_ROC;
 731
 732	dw_i3c_master_enqueue_xfer(master, xfer);
 733	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 734		dw_i3c_master_dequeue_xfer(master, xfer);
 735
 736	ret = xfer->ret;
 737	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 738		ccc->err = I3C_ERROR_M2;
 739	dw_i3c_master_free_xfer(xfer);
 740
 741	return ret;
 742}
 743
 
 
 
 
 
 
 744static int dw_i3c_master_send_ccc_cmd(struct i3c_master_controller *m,
 745				      struct i3c_ccc_cmd *ccc)
 746{
 747	struct dw_i3c_master *master = to_dw_i3c_master(m);
 748	int ret = 0;
 749
 750	if (ccc->id == I3C_CCC_ENTDAA)
 751		return -EINVAL;
 752
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 753	if (ccc->rnw)
 754		ret = dw_i3c_ccc_get(master, ccc);
 755	else
 756		ret = dw_i3c_ccc_set(master, ccc);
 757
 
 
 758	return ret;
 759}
 760
 761static int dw_i3c_master_daa(struct i3c_master_controller *m)
 762{
 763	struct dw_i3c_master *master = to_dw_i3c_master(m);
 764	struct dw_i3c_xfer *xfer;
 765	struct dw_i3c_cmd *cmd;
 766	u32 olddevs, newdevs;
 767	u8 p, last_addr = 0;
 768	int ret, pos;
 769
 
 
 
 
 
 
 
 
 770	olddevs = ~(master->free_pos);
 771
 772	/* Prepare DAT before launching DAA. */
 773	for (pos = 0; pos < master->maxdevs; pos++) {
 774		if (olddevs & BIT(pos))
 775			continue;
 776
 777		ret = i3c_master_get_free_addr(m, last_addr + 1);
 778		if (ret < 0)
 779			return -ENOSPC;
 
 
 780
 781		master->addrs[pos] = ret;
 782		p = even_parity(ret);
 783		last_addr = ret;
 784		ret |= (p << 7);
 785
 786		writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(ret),
 787		       master->regs +
 788		       DEV_ADDR_TABLE_LOC(master->datstartaddr, pos));
 
 
 789	}
 790
 791	xfer = dw_i3c_master_alloc_xfer(master, 1);
 792	if (!xfer)
 793		return -ENOMEM;
 
 
 794
 795	pos = dw_i3c_master_get_free_pos(master);
 796	if (pos < 0) {
 797		dw_i3c_master_free_xfer(xfer);
 798		return pos;
 
 799	}
 800	cmd = &xfer->cmds[0];
 801	cmd->cmd_hi = 0x1;
 802	cmd->cmd_lo = COMMAND_PORT_DEV_COUNT(master->maxdevs - pos) |
 803		      COMMAND_PORT_DEV_INDEX(pos) |
 804		      COMMAND_PORT_CMD(I3C_CCC_ENTDAA) |
 805		      COMMAND_PORT_ADDR_ASSGN_CMD |
 806		      COMMAND_PORT_TOC |
 807		      COMMAND_PORT_ROC;
 808
 809	dw_i3c_master_enqueue_xfer(master, xfer);
 810	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 811		dw_i3c_master_dequeue_xfer(master, xfer);
 812
 813	newdevs = GENMASK(master->maxdevs - cmd->rx_len - 1, 0);
 814	newdevs &= ~olddevs;
 815
 816	for (pos = 0; pos < master->maxdevs; pos++) {
 817		if (newdevs & BIT(pos))
 818			i3c_master_add_i3c_dev_locked(m, master->addrs[pos]);
 819	}
 820
 821	dw_i3c_master_free_xfer(xfer);
 822
 823	return 0;
 
 
 
 824}
 825
 826static int dw_i3c_master_priv_xfers(struct i3c_dev_desc *dev,
 827				    struct i3c_priv_xfer *i3c_xfers,
 828				    int i3c_nxfers)
 829{
 830	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 831	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 832	struct dw_i3c_master *master = to_dw_i3c_master(m);
 833	unsigned int nrxwords = 0, ntxwords = 0;
 834	struct dw_i3c_xfer *xfer;
 835	int i, ret = 0;
 836
 837	if (!i3c_nxfers)
 838		return 0;
 839
 840	if (i3c_nxfers > master->caps.cmdfifodepth)
 841		return -ENOTSUPP;
 842
 843	for (i = 0; i < i3c_nxfers; i++) {
 844		if (i3c_xfers[i].rnw)
 845			nrxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 846		else
 847			ntxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 848	}
 849
 850	if (ntxwords > master->caps.datafifodepth ||
 851	    nrxwords > master->caps.datafifodepth)
 852		return -ENOTSUPP;
 853
 854	xfer = dw_i3c_master_alloc_xfer(master, i3c_nxfers);
 855	if (!xfer)
 856		return -ENOMEM;
 857
 
 
 
 
 
 
 
 
 858	for (i = 0; i < i3c_nxfers; i++) {
 859		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 860
 861		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i3c_xfers[i].len) |
 862			COMMAND_PORT_TRANSFER_ARG;
 863
 864		if (i3c_xfers[i].rnw) {
 865			cmd->rx_buf = i3c_xfers[i].data.in;
 866			cmd->rx_len = i3c_xfers[i].len;
 867			cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 868				      COMMAND_PORT_SPEED(dev->info.max_read_ds);
 869
 870		} else {
 871			cmd->tx_buf = i3c_xfers[i].data.out;
 872			cmd->tx_len = i3c_xfers[i].len;
 873			cmd->cmd_lo =
 874				COMMAND_PORT_SPEED(dev->info.max_write_ds);
 875		}
 876
 877		cmd->cmd_lo |= COMMAND_PORT_TID(i) |
 878			       COMMAND_PORT_DEV_INDEX(data->index) |
 879			       COMMAND_PORT_ROC;
 880
 881		if (i == (i3c_nxfers - 1))
 882			cmd->cmd_lo |= COMMAND_PORT_TOC;
 883	}
 884
 885	dw_i3c_master_enqueue_xfer(master, xfer);
 886	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 887		dw_i3c_master_dequeue_xfer(master, xfer);
 888
 
 
 
 
 
 
 
 889	ret = xfer->ret;
 890	dw_i3c_master_free_xfer(xfer);
 891
 
 
 892	return ret;
 893}
 894
 895static int dw_i3c_master_reattach_i3c_dev(struct i3c_dev_desc *dev,
 896					  u8 old_dyn_addr)
 897{
 898	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 899	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 900	struct dw_i3c_master *master = to_dw_i3c_master(m);
 901	int pos;
 902
 903	pos = dw_i3c_master_get_free_pos(master);
 904
 905	if (data->index > pos && pos > 0) {
 906		writel(0,
 907		       master->regs +
 908		       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 909
 910		master->addrs[data->index] = 0;
 911		master->free_pos |= BIT(data->index);
 912
 913		data->index = pos;
 914		master->addrs[pos] = dev->info.dyn_addr;
 915		master->free_pos &= ~BIT(pos);
 916	}
 917
 918	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(dev->info.dyn_addr),
 919	       master->regs +
 920	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 921
 922	master->addrs[data->index] = dev->info.dyn_addr;
 923
 924	return 0;
 925}
 926
 927static int dw_i3c_master_attach_i3c_dev(struct i3c_dev_desc *dev)
 928{
 929	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 930	struct dw_i3c_master *master = to_dw_i3c_master(m);
 931	struct dw_i3c_i2c_dev_data *data;
 932	int pos;
 933
 934	pos = dw_i3c_master_get_free_pos(master);
 935	if (pos < 0)
 936		return pos;
 937
 938	data = kzalloc(sizeof(*data), GFP_KERNEL);
 939	if (!data)
 940		return -ENOMEM;
 941
 942	data->index = pos;
 943	master->addrs[pos] = dev->info.dyn_addr ? : dev->info.static_addr;
 944	master->free_pos &= ~BIT(pos);
 945	i3c_dev_set_master_data(dev, data);
 946
 947	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(master->addrs[pos]),
 948	       master->regs +
 949	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 950
 951	return 0;
 952}
 953
 954static void dw_i3c_master_detach_i3c_dev(struct i3c_dev_desc *dev)
 955{
 956	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 957	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 958	struct dw_i3c_master *master = to_dw_i3c_master(m);
 959
 960	writel(0,
 961	       master->regs +
 962	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
 963
 964	i3c_dev_set_master_data(dev, NULL);
 965	master->addrs[data->index] = 0;
 966	master->free_pos |= BIT(data->index);
 967	kfree(data);
 968}
 969
 970static int dw_i3c_master_i2c_xfers(struct i2c_dev_desc *dev,
 971				   const struct i2c_msg *i2c_xfers,
 972				   int i2c_nxfers)
 973{
 974	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
 975	struct i3c_master_controller *m = i2c_dev_get_master(dev);
 976	struct dw_i3c_master *master = to_dw_i3c_master(m);
 977	unsigned int nrxwords = 0, ntxwords = 0;
 978	struct dw_i3c_xfer *xfer;
 979	int i, ret = 0;
 980
 981	if (!i2c_nxfers)
 982		return 0;
 983
 984	if (i2c_nxfers > master->caps.cmdfifodepth)
 985		return -ENOTSUPP;
 986
 987	for (i = 0; i < i2c_nxfers; i++) {
 988		if (i2c_xfers[i].flags & I2C_M_RD)
 989			nrxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
 990		else
 991			ntxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
 992	}
 993
 994	if (ntxwords > master->caps.datafifodepth ||
 995	    nrxwords > master->caps.datafifodepth)
 996		return -ENOTSUPP;
 997
 998	xfer = dw_i3c_master_alloc_xfer(master, i2c_nxfers);
 999	if (!xfer)
1000		return -ENOMEM;
1001
 
 
 
 
 
 
 
 
1002	for (i = 0; i < i2c_nxfers; i++) {
1003		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
1004
1005		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i2c_xfers[i].len) |
1006			COMMAND_PORT_TRANSFER_ARG;
1007
1008		cmd->cmd_lo = COMMAND_PORT_TID(i) |
1009			      COMMAND_PORT_DEV_INDEX(data->index) |
1010			      COMMAND_PORT_ROC;
1011
1012		if (i2c_xfers[i].flags & I2C_M_RD) {
1013			cmd->cmd_lo |= COMMAND_PORT_READ_TRANSFER;
1014			cmd->rx_buf = i2c_xfers[i].buf;
1015			cmd->rx_len = i2c_xfers[i].len;
1016		} else {
1017			cmd->tx_buf = i2c_xfers[i].buf;
1018			cmd->tx_len = i2c_xfers[i].len;
1019		}
1020
1021		if (i == (i2c_nxfers - 1))
1022			cmd->cmd_lo |= COMMAND_PORT_TOC;
1023	}
1024
1025	dw_i3c_master_enqueue_xfer(master, xfer);
1026	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
1027		dw_i3c_master_dequeue_xfer(master, xfer);
1028
1029	ret = xfer->ret;
1030	dw_i3c_master_free_xfer(xfer);
1031
 
 
1032	return ret;
1033}
1034
1035static int dw_i3c_master_attach_i2c_dev(struct i2c_dev_desc *dev)
1036{
1037	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1038	struct dw_i3c_master *master = to_dw_i3c_master(m);
1039	struct dw_i3c_i2c_dev_data *data;
1040	int pos;
1041
1042	pos = dw_i3c_master_get_free_pos(master);
1043	if (pos < 0)
1044		return pos;
1045
1046	data = kzalloc(sizeof(*data), GFP_KERNEL);
1047	if (!data)
1048		return -ENOMEM;
1049
1050	data->index = pos;
1051	master->addrs[pos] = dev->addr;
 
1052	master->free_pos &= ~BIT(pos);
1053	i2c_dev_set_master_data(dev, data);
1054
1055	writel(DEV_ADDR_TABLE_LEGACY_I2C_DEV |
1056	       DEV_ADDR_TABLE_STATIC_ADDR(dev->addr),
1057	       master->regs +
1058	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1059
1060	return 0;
1061}
1062
1063static void dw_i3c_master_detach_i2c_dev(struct i2c_dev_desc *dev)
1064{
1065	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1066	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1067	struct dw_i3c_master *master = to_dw_i3c_master(m);
1068
1069	writel(0,
1070	       master->regs +
1071	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1072
1073	i2c_dev_set_master_data(dev, NULL);
1074	master->addrs[data->index] = 0;
1075	master->free_pos |= BIT(data->index);
1076	kfree(data);
1077}
1078
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1079static irqreturn_t dw_i3c_master_irq_handler(int irq, void *dev_id)
1080{
1081	struct dw_i3c_master *master = dev_id;
1082	u32 status;
1083
1084	status = readl(master->regs + INTR_STATUS);
1085
1086	if (!(status & readl(master->regs + INTR_STATUS_EN))) {
1087		writel(INTR_ALL, master->regs + INTR_STATUS);
1088		return IRQ_NONE;
1089	}
1090
1091	spin_lock(&master->xferqueue.lock);
1092	dw_i3c_master_end_xfer_locked(master, status);
1093	if (status & INTR_TRANSFER_ERR_STAT)
1094		writel(INTR_TRANSFER_ERR_STAT, master->regs + INTR_STATUS);
1095	spin_unlock(&master->xferqueue.lock);
1096
 
 
 
1097	return IRQ_HANDLED;
1098}
1099
1100static const struct i3c_master_controller_ops dw_mipi_i3c_ops = {
1101	.bus_init = dw_i3c_master_bus_init,
1102	.bus_cleanup = dw_i3c_master_bus_cleanup,
1103	.attach_i3c_dev = dw_i3c_master_attach_i3c_dev,
1104	.reattach_i3c_dev = dw_i3c_master_reattach_i3c_dev,
1105	.detach_i3c_dev = dw_i3c_master_detach_i3c_dev,
1106	.do_daa = dw_i3c_master_daa,
1107	.supports_ccc_cmd = dw_i3c_master_supports_ccc_cmd,
1108	.send_ccc_cmd = dw_i3c_master_send_ccc_cmd,
1109	.priv_xfers = dw_i3c_master_priv_xfers,
1110	.attach_i2c_dev = dw_i3c_master_attach_i2c_dev,
1111	.detach_i2c_dev = dw_i3c_master_detach_i2c_dev,
1112	.i2c_xfers = dw_i3c_master_i2c_xfers,
 
 
 
 
 
 
 
1113};
1114
1115static int dw_i3c_probe(struct platform_device *pdev)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1116{
1117	struct dw_i3c_master *master;
1118	int ret, irq;
1119
1120	master = devm_kzalloc(&pdev->dev, sizeof(*master), GFP_KERNEL);
1121	if (!master)
1122		return -ENOMEM;
 
1123
1124	master->regs = devm_platform_ioremap_resource(pdev, 0);
1125	if (IS_ERR(master->regs))
1126		return PTR_ERR(master->regs);
1127
1128	master->core_clk = devm_clk_get(&pdev->dev, NULL);
1129	if (IS_ERR(master->core_clk))
1130		return PTR_ERR(master->core_clk);
1131
 
 
 
 
1132	master->core_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
1133								    "core_rst");
1134	if (IS_ERR(master->core_rst))
1135		return PTR_ERR(master->core_rst);
1136
1137	ret = clk_prepare_enable(master->core_clk);
1138	if (ret)
1139		goto err_disable_core_clk;
1140
1141	reset_control_deassert(master->core_rst);
1142
1143	spin_lock_init(&master->xferqueue.lock);
1144	INIT_LIST_HEAD(&master->xferqueue.list);
1145
1146	writel(INTR_ALL, master->regs + INTR_STATUS);
1147	irq = platform_get_irq(pdev, 0);
1148	ret = devm_request_irq(&pdev->dev, irq,
1149			       dw_i3c_master_irq_handler, 0,
1150			       dev_name(&pdev->dev), master);
1151	if (ret)
1152		goto err_assert_rst;
1153
1154	platform_set_drvdata(pdev, master);
1155
 
 
 
 
 
1156	/* Information regarding the FIFOs/QUEUEs depth */
1157	ret = readl(master->regs + QUEUE_STATUS_LEVEL);
1158	master->caps.cmdfifodepth = QUEUE_STATUS_LEVEL_CMD(ret);
1159
1160	ret = readl(master->regs + DATA_BUFFER_STATUS_LEVEL);
1161	master->caps.datafifodepth = DATA_BUFFER_STATUS_LEVEL_TX(ret);
1162
1163	ret = readl(master->regs + DEVICE_ADDR_TABLE_POINTER);
1164	master->datstartaddr = ret;
1165	master->maxdevs = ret >> 16;
1166	master->free_pos = GENMASK(master->maxdevs - 1, 0);
1167
 
 
 
1168	ret = i3c_master_register(&master->base, &pdev->dev,
1169				  &dw_mipi_i3c_ops, false);
1170	if (ret)
1171		goto err_assert_rst;
1172
1173	return 0;
1174
 
 
 
 
 
1175err_assert_rst:
1176	reset_control_assert(master->core_rst);
1177
1178err_disable_core_clk:
1179	clk_disable_unprepare(master->core_clk);
1180
1181	return ret;
1182}
 
 
 
 
 
 
 
 
 
 
 
 
1183
1184static int dw_i3c_remove(struct platform_device *pdev)
 
 
 
 
 
 
 
 
 
 
 
 
 
1185{
1186	struct dw_i3c_master *master = platform_get_drvdata(pdev);
1187	int ret;
1188
1189	ret = i3c_master_unregister(&master->base);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1190	if (ret)
1191		return ret;
1192
1193	reset_control_assert(master->core_rst);
 
 
 
 
 
 
 
1194
 
 
 
1195	clk_disable_unprepare(master->core_clk);
 
 
 
 
 
 
 
 
 
 
 
 
 
1196
 
 
 
 
 
 
 
 
 
 
 
 
 
1197	return 0;
1198}
1199
 
 
 
 
 
1200static const struct of_device_id dw_i3c_master_of_match[] = {
1201	{ .compatible = "snps,dw-i3c-master-1.00a", },
1202	{},
1203};
1204MODULE_DEVICE_TABLE(of, dw_i3c_master_of_match);
1205
 
 
 
 
 
 
1206static struct platform_driver dw_i3c_driver = {
1207	.probe = dw_i3c_probe,
1208	.remove = dw_i3c_remove,
1209	.driver = {
1210		.name = "dw-i3c-master",
1211		.of_match_table = of_match_ptr(dw_i3c_master_of_match),
 
 
1212	},
1213};
1214module_platform_driver(dw_i3c_driver);
1215
1216MODULE_AUTHOR("Vitor Soares <vitor.soares@synopsys.com>");
1217MODULE_DESCRIPTION("DesignWare MIPI I3C driver");
1218MODULE_LICENSE("GPL v2");
v6.13.7
   1// SPDX-License-Identifier: GPL-2.0
   2/*
   3 * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
   4 *
   5 * Author: Vitor Soares <vitor.soares@synopsys.com>
   6 */
   7
   8#include <linux/bitops.h>
   9#include <linux/clk.h>
  10#include <linux/completion.h>
  11#include <linux/err.h>
  12#include <linux/errno.h>
  13#include <linux/i3c/master.h>
  14#include <linux/interrupt.h>
  15#include <linux/ioport.h>
  16#include <linux/iopoll.h>
  17#include <linux/list.h>
  18#include <linux/module.h>
  19#include <linux/of.h>
  20#include <linux/pinctrl/consumer.h>
  21#include <linux/platform_device.h>
  22#include <linux/pm_runtime.h>
  23#include <linux/reset.h>
  24#include <linux/slab.h>
  25
  26#include "dw-i3c-master.h"
  27
  28#define DEVICE_CTRL			0x0
  29#define DEV_CTRL_ENABLE			BIT(31)
  30#define DEV_CTRL_RESUME			BIT(30)
  31#define DEV_CTRL_HOT_JOIN_NACK		BIT(8)
  32#define DEV_CTRL_I2C_SLAVE_PRESENT	BIT(7)
  33
  34#define DEVICE_ADDR			0x4
  35#define DEV_ADDR_DYNAMIC_ADDR_VALID	BIT(31)
  36#define DEV_ADDR_DYNAMIC(x)		(((x) << 16) & GENMASK(22, 16))
  37
  38#define HW_CAPABILITY			0x8
  39#define COMMAND_QUEUE_PORT		0xc
  40#define COMMAND_PORT_TOC		BIT(30)
  41#define COMMAND_PORT_READ_TRANSFER	BIT(28)
  42#define COMMAND_PORT_SDAP		BIT(27)
  43#define COMMAND_PORT_ROC		BIT(26)
  44#define COMMAND_PORT_SPEED(x)		(((x) << 21) & GENMASK(23, 21))
  45#define COMMAND_PORT_DEV_INDEX(x)	(((x) << 16) & GENMASK(20, 16))
  46#define COMMAND_PORT_CP			BIT(15)
  47#define COMMAND_PORT_CMD(x)		(((x) << 7) & GENMASK(14, 7))
  48#define COMMAND_PORT_TID(x)		(((x) << 3) & GENMASK(6, 3))
  49
  50#define COMMAND_PORT_ARG_DATA_LEN(x)	(((x) << 16) & GENMASK(31, 16))
  51#define COMMAND_PORT_ARG_DATA_LEN_MAX	65536
  52#define COMMAND_PORT_TRANSFER_ARG	0x01
  53
  54#define COMMAND_PORT_SDA_DATA_BYTE_3(x)	(((x) << 24) & GENMASK(31, 24))
  55#define COMMAND_PORT_SDA_DATA_BYTE_2(x)	(((x) << 16) & GENMASK(23, 16))
  56#define COMMAND_PORT_SDA_DATA_BYTE_1(x)	(((x) << 8) & GENMASK(15, 8))
  57#define COMMAND_PORT_SDA_BYTE_STRB_3	BIT(5)
  58#define COMMAND_PORT_SDA_BYTE_STRB_2	BIT(4)
  59#define COMMAND_PORT_SDA_BYTE_STRB_1	BIT(3)
  60#define COMMAND_PORT_SHORT_DATA_ARG	0x02
  61
  62#define COMMAND_PORT_DEV_COUNT(x)	(((x) << 21) & GENMASK(25, 21))
  63#define COMMAND_PORT_ADDR_ASSGN_CMD	0x03
  64
  65#define RESPONSE_QUEUE_PORT		0x10
  66#define RESPONSE_PORT_ERR_STATUS(x)	(((x) & GENMASK(31, 28)) >> 28)
  67#define RESPONSE_NO_ERROR		0
  68#define RESPONSE_ERROR_CRC		1
  69#define RESPONSE_ERROR_PARITY		2
  70#define RESPONSE_ERROR_FRAME		3
  71#define RESPONSE_ERROR_IBA_NACK		4
  72#define RESPONSE_ERROR_ADDRESS_NACK	5
  73#define RESPONSE_ERROR_OVER_UNDER_FLOW	6
  74#define RESPONSE_ERROR_TRANSF_ABORT	8
  75#define RESPONSE_ERROR_I2C_W_NACK_ERR	9
  76#define RESPONSE_PORT_TID(x)		(((x) & GENMASK(27, 24)) >> 24)
  77#define RESPONSE_PORT_DATA_LEN(x)	((x) & GENMASK(15, 0))
  78
  79#define RX_TX_DATA_PORT			0x14
  80#define IBI_QUEUE_STATUS		0x18
  81#define IBI_QUEUE_STATUS_IBI_ID(x)	(((x) & GENMASK(15, 8)) >> 8)
  82#define IBI_QUEUE_STATUS_DATA_LEN(x)	((x) & GENMASK(7, 0))
  83#define IBI_QUEUE_IBI_ADDR(x)		(IBI_QUEUE_STATUS_IBI_ID(x) >> 1)
  84#define IBI_QUEUE_IBI_RNW(x)		(IBI_QUEUE_STATUS_IBI_ID(x) & BIT(0))
  85#define IBI_TYPE_MR(x)                                                         \
  86	((IBI_QUEUE_IBI_ADDR(x) != I3C_HOT_JOIN_ADDR) && !IBI_QUEUE_IBI_RNW(x))
  87#define IBI_TYPE_HJ(x)                                                         \
  88	((IBI_QUEUE_IBI_ADDR(x) == I3C_HOT_JOIN_ADDR) && !IBI_QUEUE_IBI_RNW(x))
  89#define IBI_TYPE_SIRQ(x)                                                        \
  90	((IBI_QUEUE_IBI_ADDR(x) != I3C_HOT_JOIN_ADDR) && IBI_QUEUE_IBI_RNW(x))
  91
  92#define QUEUE_THLD_CTRL			0x1c
  93#define QUEUE_THLD_CTRL_IBI_STAT_MASK	GENMASK(31, 24)
  94#define QUEUE_THLD_CTRL_IBI_STAT(x)	(((x) - 1) << 24)
  95#define QUEUE_THLD_CTRL_IBI_DATA_MASK	GENMASK(20, 16)
  96#define QUEUE_THLD_CTRL_IBI_DATA(x)	((x) << 16)
  97#define QUEUE_THLD_CTRL_RESP_BUF_MASK	GENMASK(15, 8)
  98#define QUEUE_THLD_CTRL_RESP_BUF(x)	(((x) - 1) << 8)
  99
 100#define DATA_BUFFER_THLD_CTRL		0x20
 101#define DATA_BUFFER_THLD_CTRL_RX_BUF	GENMASK(11, 8)
 102
 103#define IBI_QUEUE_CTRL			0x24
 104#define IBI_MR_REQ_REJECT		0x2C
 105#define IBI_SIR_REQ_REJECT		0x30
 106#define IBI_REQ_REJECT_ALL		GENMASK(31, 0)
 107
 108#define RESET_CTRL			0x34
 109#define RESET_CTRL_IBI_QUEUE		BIT(5)
 110#define RESET_CTRL_RX_FIFO		BIT(4)
 111#define RESET_CTRL_TX_FIFO		BIT(3)
 112#define RESET_CTRL_RESP_QUEUE		BIT(2)
 113#define RESET_CTRL_CMD_QUEUE		BIT(1)
 114#define RESET_CTRL_SOFT			BIT(0)
 115
 116#define SLV_EVENT_CTRL			0x38
 117#define INTR_STATUS			0x3c
 118#define INTR_STATUS_EN			0x40
 119#define INTR_SIGNAL_EN			0x44
 120#define INTR_FORCE			0x48
 121#define INTR_BUSOWNER_UPDATE_STAT	BIT(13)
 122#define INTR_IBI_UPDATED_STAT		BIT(12)
 123#define INTR_READ_REQ_RECV_STAT		BIT(11)
 124#define INTR_DEFSLV_STAT		BIT(10)
 125#define INTR_TRANSFER_ERR_STAT		BIT(9)
 126#define INTR_DYN_ADDR_ASSGN_STAT	BIT(8)
 127#define INTR_CCC_UPDATED_STAT		BIT(6)
 128#define INTR_TRANSFER_ABORT_STAT	BIT(5)
 129#define INTR_RESP_READY_STAT		BIT(4)
 130#define INTR_CMD_QUEUE_READY_STAT	BIT(3)
 131#define INTR_IBI_THLD_STAT		BIT(2)
 132#define INTR_RX_THLD_STAT		BIT(1)
 133#define INTR_TX_THLD_STAT		BIT(0)
 134#define INTR_ALL			(INTR_BUSOWNER_UPDATE_STAT |	\
 135					INTR_IBI_UPDATED_STAT |		\
 136					INTR_READ_REQ_RECV_STAT |	\
 137					INTR_DEFSLV_STAT |		\
 138					INTR_TRANSFER_ERR_STAT |	\
 139					INTR_DYN_ADDR_ASSGN_STAT |	\
 140					INTR_CCC_UPDATED_STAT |		\
 141					INTR_TRANSFER_ABORT_STAT |	\
 142					INTR_RESP_READY_STAT |		\
 143					INTR_CMD_QUEUE_READY_STAT |	\
 144					INTR_IBI_THLD_STAT |		\
 145					INTR_TX_THLD_STAT |		\
 146					INTR_RX_THLD_STAT)
 147
 148#define INTR_MASTER_MASK		(INTR_TRANSFER_ERR_STAT |	\
 149					 INTR_RESP_READY_STAT)
 150
 151#define QUEUE_STATUS_LEVEL		0x4c
 152#define QUEUE_STATUS_IBI_STATUS_CNT(x)	(((x) & GENMASK(28, 24)) >> 24)
 153#define QUEUE_STATUS_IBI_BUF_BLR(x)	(((x) & GENMASK(23, 16)) >> 16)
 154#define QUEUE_STATUS_LEVEL_RESP(x)	(((x) & GENMASK(15, 8)) >> 8)
 155#define QUEUE_STATUS_LEVEL_CMD(x)	((x) & GENMASK(7, 0))
 156
 157#define DATA_BUFFER_STATUS_LEVEL	0x50
 158#define DATA_BUFFER_STATUS_LEVEL_TX(x)	((x) & GENMASK(7, 0))
 159
 160#define PRESENT_STATE			0x54
 161#define CCC_DEVICE_STATUS		0x58
 162#define DEVICE_ADDR_TABLE_POINTER	0x5c
 163#define DEVICE_ADDR_TABLE_DEPTH(x)	(((x) & GENMASK(31, 16)) >> 16)
 164#define DEVICE_ADDR_TABLE_ADDR(x)	((x) & GENMASK(7, 0))
 165
 166#define DEV_CHAR_TABLE_POINTER		0x60
 167#define VENDOR_SPECIFIC_REG_POINTER	0x6c
 168#define SLV_PID_VALUE			0x74
 169#define SLV_CHAR_CTRL			0x78
 170#define SLV_MAX_LEN			0x7c
 171#define MAX_READ_TURNAROUND		0x80
 172#define MAX_DATA_SPEED			0x84
 173#define SLV_DEBUG_STATUS		0x88
 174#define SLV_INTR_REQ			0x8c
 175#define DEVICE_CTRL_EXTENDED		0xb0
 176#define SCL_I3C_OD_TIMING		0xb4
 177#define SCL_I3C_PP_TIMING		0xb8
 178#define SCL_I3C_TIMING_HCNT(x)		(((x) << 16) & GENMASK(23, 16))
 179#define SCL_I3C_TIMING_LCNT(x)		((x) & GENMASK(7, 0))
 180#define SCL_I3C_TIMING_CNT_MIN		5
 181
 182#define SCL_I2C_FM_TIMING		0xbc
 183#define SCL_I2C_FM_TIMING_HCNT(x)	(((x) << 16) & GENMASK(31, 16))
 184#define SCL_I2C_FM_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 185
 186#define SCL_I2C_FMP_TIMING		0xc0
 187#define SCL_I2C_FMP_TIMING_HCNT(x)	(((x) << 16) & GENMASK(23, 16))
 188#define SCL_I2C_FMP_TIMING_LCNT(x)	((x) & GENMASK(15, 0))
 189
 190#define SCL_EXT_LCNT_TIMING		0xc8
 191#define SCL_EXT_LCNT_4(x)		(((x) << 24) & GENMASK(31, 24))
 192#define SCL_EXT_LCNT_3(x)		(((x) << 16) & GENMASK(23, 16))
 193#define SCL_EXT_LCNT_2(x)		(((x) << 8) & GENMASK(15, 8))
 194#define SCL_EXT_LCNT_1(x)		((x) & GENMASK(7, 0))
 195
 196#define SCL_EXT_TERMN_LCNT_TIMING	0xcc
 197#define BUS_FREE_TIMING			0xd4
 198#define BUS_I3C_MST_FREE(x)		((x) & GENMASK(15, 0))
 199
 200#define BUS_IDLE_TIMING			0xd8
 201#define I3C_VER_ID			0xe0
 202#define I3C_VER_TYPE			0xe4
 203#define EXTENDED_CAPABILITY		0xe8
 204#define SLAVE_CONFIG			0xec
 205
 206#define DEV_ADDR_TABLE_IBI_MDB		BIT(12)
 207#define DEV_ADDR_TABLE_SIR_REJECT	BIT(13)
 208#define DEV_ADDR_TABLE_LEGACY_I2C_DEV	BIT(31)
 209#define DEV_ADDR_TABLE_DYNAMIC_ADDR(x)	(((x) << 16) & GENMASK(23, 16))
 210#define DEV_ADDR_TABLE_STATIC_ADDR(x)	((x) & GENMASK(6, 0))
 211#define DEV_ADDR_TABLE_LOC(start, idx)	((start) + ((idx) << 2))
 212
 
 
 213#define I3C_BUS_SDR1_SCL_RATE		8000000
 214#define I3C_BUS_SDR2_SCL_RATE		6000000
 215#define I3C_BUS_SDR3_SCL_RATE		4000000
 216#define I3C_BUS_SDR4_SCL_RATE		2000000
 217#define I3C_BUS_I2C_FM_TLOW_MIN_NS	1300
 218#define I3C_BUS_I2C_FMP_TLOW_MIN_NS	500
 219#define I3C_BUS_THIGH_MAX_NS		41
 220
 221#define XFER_TIMEOUT (msecs_to_jiffies(1000))
 222#define RPM_AUTOSUSPEND_TIMEOUT 1000 /* ms */
 223
 224/* Timing values to configure 12.5MHz frequency */
 225#define AMD_I3C_OD_TIMING          0x4C007C
 226#define AMD_I3C_PP_TIMING          0x8001A
 227
 228/* List of quirks */
 229#define AMD_I3C_OD_PP_TIMING		BIT(1)
 230
 231struct dw_i3c_cmd {
 232	u32 cmd_lo;
 233	u32 cmd_hi;
 234	u16 tx_len;
 235	const void *tx_buf;
 236	u16 rx_len;
 237	void *rx_buf;
 238	u8 error;
 239};
 240
 241struct dw_i3c_xfer {
 242	struct list_head node;
 243	struct completion comp;
 244	int ret;
 245	unsigned int ncmds;
 246	struct dw_i3c_cmd cmds[] __counted_by(ncmds);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 247};
 248
 249struct dw_i3c_i2c_dev_data {
 250	u8 index;
 251	struct i3c_generic_ibi_pool *ibi_pool;
 252};
 253
 254static u8 even_parity(u8 p)
 255{
 256	p ^= p >> 4;
 257	p &= 0xf;
 258
 259	return (0x9669 >> p) & 1;
 260}
 261
 262static bool dw_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m,
 263					   const struct i3c_ccc_cmd *cmd)
 264{
 265	if (cmd->ndests > 1)
 266		return false;
 267
 268	switch (cmd->id) {
 269	case I3C_CCC_ENEC(true):
 270	case I3C_CCC_ENEC(false):
 271	case I3C_CCC_DISEC(true):
 272	case I3C_CCC_DISEC(false):
 273	case I3C_CCC_ENTAS(0, true):
 274	case I3C_CCC_ENTAS(0, false):
 275	case I3C_CCC_RSTDAA(true):
 276	case I3C_CCC_RSTDAA(false):
 277	case I3C_CCC_ENTDAA:
 278	case I3C_CCC_SETMWL(true):
 279	case I3C_CCC_SETMWL(false):
 280	case I3C_CCC_SETMRL(true):
 281	case I3C_CCC_SETMRL(false):
 282	case I3C_CCC_ENTHDR(0):
 283	case I3C_CCC_SETDASA:
 284	case I3C_CCC_SETNEWDA:
 285	case I3C_CCC_GETMWL:
 286	case I3C_CCC_GETMRL:
 287	case I3C_CCC_GETPID:
 288	case I3C_CCC_GETBCR:
 289	case I3C_CCC_GETDCR:
 290	case I3C_CCC_GETSTATUS:
 291	case I3C_CCC_GETMXDS:
 292	case I3C_CCC_GETHDRCAP:
 293		return true;
 294	default:
 295		return false;
 296	}
 297}
 298
 299static inline struct dw_i3c_master *
 300to_dw_i3c_master(struct i3c_master_controller *master)
 301{
 302	return container_of(master, struct dw_i3c_master, base);
 303}
 304
 305static void dw_i3c_master_disable(struct dw_i3c_master *master)
 306{
 307	writel(readl(master->regs + DEVICE_CTRL) & ~DEV_CTRL_ENABLE,
 308	       master->regs + DEVICE_CTRL);
 309}
 310
 311static void dw_i3c_master_enable(struct dw_i3c_master *master)
 312{
 313	u32 dev_ctrl;
 314
 315	dev_ctrl = readl(master->regs + DEVICE_CTRL);
 316	/* For now don't support Hot-Join */
 317	dev_ctrl |= DEV_CTRL_HOT_JOIN_NACK;
 318	if (master->i2c_slv_prsnt)
 319		dev_ctrl |= DEV_CTRL_I2C_SLAVE_PRESENT;
 320	writel(dev_ctrl | DEV_CTRL_ENABLE,
 321	       master->regs + DEVICE_CTRL);
 322}
 323
 324static int dw_i3c_master_get_addr_pos(struct dw_i3c_master *master, u8 addr)
 325{
 326	int pos;
 327
 328	for (pos = 0; pos < master->maxdevs; pos++) {
 329		if (addr == master->devs[pos].addr)
 330			return pos;
 331	}
 332
 333	return -EINVAL;
 334}
 335
 336static int dw_i3c_master_get_free_pos(struct dw_i3c_master *master)
 337{
 338	if (!(master->free_pos & GENMASK(master->maxdevs - 1, 0)))
 339		return -ENOSPC;
 340
 341	return ffs(master->free_pos) - 1;
 342}
 343
 344static void dw_i3c_master_wr_tx_fifo(struct dw_i3c_master *master,
 345				     const u8 *bytes, int nbytes)
 346{
 347	writesl(master->regs + RX_TX_DATA_PORT, bytes, nbytes / 4);
 348	if (nbytes & 3) {
 349		u32 tmp = 0;
 350
 351		memcpy(&tmp, bytes + (nbytes & ~3), nbytes & 3);
 352		writesl(master->regs + RX_TX_DATA_PORT, &tmp, 1);
 353	}
 354}
 355
 356static void dw_i3c_master_read_fifo(struct dw_i3c_master *master,
 357				    int reg,  u8 *bytes, int nbytes)
 358{
 359	readsl(master->regs + reg, bytes, nbytes / 4);
 360	if (nbytes & 3) {
 361		u32 tmp;
 362
 363		readsl(master->regs + reg, &tmp, 1);
 364		memcpy(bytes + (nbytes & ~3), &tmp, nbytes & 3);
 365	}
 366}
 367
 368static void dw_i3c_master_read_rx_fifo(struct dw_i3c_master *master,
 369				       u8 *bytes, int nbytes)
 370{
 371	return dw_i3c_master_read_fifo(master, RX_TX_DATA_PORT, bytes, nbytes);
 372}
 373
 374static void dw_i3c_master_read_ibi_fifo(struct dw_i3c_master *master,
 375					u8 *bytes, int nbytes)
 376{
 377	return dw_i3c_master_read_fifo(master, IBI_QUEUE_STATUS, bytes, nbytes);
 378}
 379
 380static struct dw_i3c_xfer *
 381dw_i3c_master_alloc_xfer(struct dw_i3c_master *master, unsigned int ncmds)
 382{
 383	struct dw_i3c_xfer *xfer;
 384
 385	xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL);
 386	if (!xfer)
 387		return NULL;
 388
 389	INIT_LIST_HEAD(&xfer->node);
 390	xfer->ncmds = ncmds;
 391	xfer->ret = -ETIMEDOUT;
 392
 393	return xfer;
 394}
 395
 396static void dw_i3c_master_free_xfer(struct dw_i3c_xfer *xfer)
 397{
 398	kfree(xfer);
 399}
 400
 401static void dw_i3c_master_start_xfer_locked(struct dw_i3c_master *master)
 402{
 403	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 404	unsigned int i;
 405	u32 thld_ctrl;
 406
 407	if (!xfer)
 408		return;
 409
 410	for (i = 0; i < xfer->ncmds; i++) {
 411		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 412
 413		dw_i3c_master_wr_tx_fifo(master, cmd->tx_buf, cmd->tx_len);
 414	}
 415
 416	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 417	thld_ctrl &= ~QUEUE_THLD_CTRL_RESP_BUF_MASK;
 418	thld_ctrl |= QUEUE_THLD_CTRL_RESP_BUF(xfer->ncmds);
 419	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 420
 421	for (i = 0; i < xfer->ncmds; i++) {
 422		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 423
 424		writel(cmd->cmd_hi, master->regs + COMMAND_QUEUE_PORT);
 425		writel(cmd->cmd_lo, master->regs + COMMAND_QUEUE_PORT);
 426	}
 427}
 428
 429static void dw_i3c_master_enqueue_xfer(struct dw_i3c_master *master,
 430				       struct dw_i3c_xfer *xfer)
 431{
 432	unsigned long flags;
 433
 434	init_completion(&xfer->comp);
 435	spin_lock_irqsave(&master->xferqueue.lock, flags);
 436	if (master->xferqueue.cur) {
 437		list_add_tail(&xfer->node, &master->xferqueue.list);
 438	} else {
 439		master->xferqueue.cur = xfer;
 440		dw_i3c_master_start_xfer_locked(master);
 441	}
 442	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 443}
 444
 445static void dw_i3c_master_dequeue_xfer_locked(struct dw_i3c_master *master,
 446					      struct dw_i3c_xfer *xfer)
 447{
 448	if (master->xferqueue.cur == xfer) {
 449		u32 status;
 450
 451		master->xferqueue.cur = NULL;
 452
 453		writel(RESET_CTRL_RX_FIFO | RESET_CTRL_TX_FIFO |
 454		       RESET_CTRL_RESP_QUEUE | RESET_CTRL_CMD_QUEUE,
 455		       master->regs + RESET_CTRL);
 456
 457		readl_poll_timeout_atomic(master->regs + RESET_CTRL, status,
 458					  !status, 10, 1000000);
 459	} else {
 460		list_del_init(&xfer->node);
 461	}
 462}
 463
 464static void dw_i3c_master_dequeue_xfer(struct dw_i3c_master *master,
 465				       struct dw_i3c_xfer *xfer)
 466{
 467	unsigned long flags;
 468
 469	spin_lock_irqsave(&master->xferqueue.lock, flags);
 470	dw_i3c_master_dequeue_xfer_locked(master, xfer);
 471	spin_unlock_irqrestore(&master->xferqueue.lock, flags);
 472}
 473
 474static void dw_i3c_master_end_xfer_locked(struct dw_i3c_master *master, u32 isr)
 475{
 476	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
 477	int i, ret = 0;
 478	u32 nresp;
 479
 480	if (!xfer)
 481		return;
 482
 483	nresp = readl(master->regs + QUEUE_STATUS_LEVEL);
 484	nresp = QUEUE_STATUS_LEVEL_RESP(nresp);
 485
 486	for (i = 0; i < nresp; i++) {
 487		struct dw_i3c_cmd *cmd;
 488		u32 resp;
 489
 490		resp = readl(master->regs + RESPONSE_QUEUE_PORT);
 491
 492		cmd = &xfer->cmds[RESPONSE_PORT_TID(resp)];
 493		cmd->rx_len = RESPONSE_PORT_DATA_LEN(resp);
 494		cmd->error = RESPONSE_PORT_ERR_STATUS(resp);
 495		if (cmd->rx_len && !cmd->error)
 496			dw_i3c_master_read_rx_fifo(master, cmd->rx_buf,
 497						   cmd->rx_len);
 498	}
 499
 500	for (i = 0; i < nresp; i++) {
 501		switch (xfer->cmds[i].error) {
 502		case RESPONSE_NO_ERROR:
 503			break;
 504		case RESPONSE_ERROR_PARITY:
 505		case RESPONSE_ERROR_IBA_NACK:
 506		case RESPONSE_ERROR_TRANSF_ABORT:
 507		case RESPONSE_ERROR_CRC:
 508		case RESPONSE_ERROR_FRAME:
 509			ret = -EIO;
 510			break;
 511		case RESPONSE_ERROR_OVER_UNDER_FLOW:
 512			ret = -ENOSPC;
 513			break;
 514		case RESPONSE_ERROR_I2C_W_NACK_ERR:
 515		case RESPONSE_ERROR_ADDRESS_NACK:
 516		default:
 517			ret = -EINVAL;
 518			break;
 519		}
 520	}
 521
 522	xfer->ret = ret;
 523	complete(&xfer->comp);
 524
 525	if (ret < 0) {
 526		dw_i3c_master_dequeue_xfer_locked(master, xfer);
 527		writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_RESUME,
 528		       master->regs + DEVICE_CTRL);
 529	}
 530
 531	xfer = list_first_entry_or_null(&master->xferqueue.list,
 532					struct dw_i3c_xfer,
 533					node);
 534	if (xfer)
 535		list_del_init(&xfer->node);
 536
 537	master->xferqueue.cur = xfer;
 538	dw_i3c_master_start_xfer_locked(master);
 539}
 540
 541static void dw_i3c_master_set_intr_regs(struct dw_i3c_master *master)
 542{
 543	u32 thld_ctrl;
 544
 545	thld_ctrl = readl(master->regs + QUEUE_THLD_CTRL);
 546	thld_ctrl &= ~(QUEUE_THLD_CTRL_RESP_BUF_MASK |
 547		       QUEUE_THLD_CTRL_IBI_STAT_MASK |
 548		       QUEUE_THLD_CTRL_IBI_DATA_MASK);
 549	thld_ctrl |= QUEUE_THLD_CTRL_IBI_STAT(1) |
 550		QUEUE_THLD_CTRL_IBI_DATA(31);
 551	writel(thld_ctrl, master->regs + QUEUE_THLD_CTRL);
 552
 553	thld_ctrl = readl(master->regs + DATA_BUFFER_THLD_CTRL);
 554	thld_ctrl &= ~DATA_BUFFER_THLD_CTRL_RX_BUF;
 555	writel(thld_ctrl, master->regs + DATA_BUFFER_THLD_CTRL);
 556
 557	writel(INTR_ALL, master->regs + INTR_STATUS);
 558	writel(INTR_MASTER_MASK, master->regs + INTR_STATUS_EN);
 559	writel(INTR_MASTER_MASK, master->regs + INTR_SIGNAL_EN);
 560
 561	master->sir_rej_mask = IBI_REQ_REJECT_ALL;
 562	writel(master->sir_rej_mask, master->regs + IBI_SIR_REQ_REJECT);
 563
 564	writel(IBI_REQ_REJECT_ALL, master->regs + IBI_MR_REQ_REJECT);
 565}
 566
 567static int dw_i3c_clk_cfg(struct dw_i3c_master *master)
 568{
 569	unsigned long core_rate, core_period;
 570	u32 scl_timing;
 571	u8 hcnt, lcnt;
 572
 573	core_rate = clk_get_rate(master->core_clk);
 574	if (!core_rate)
 575		return -EINVAL;
 576
 577	core_period = DIV_ROUND_UP(1000000000, core_rate);
 578
 579	hcnt = DIV_ROUND_UP(I3C_BUS_THIGH_MAX_NS, core_period) - 1;
 580	if (hcnt < SCL_I3C_TIMING_CNT_MIN)
 581		hcnt = SCL_I3C_TIMING_CNT_MIN;
 582
 583	lcnt = DIV_ROUND_UP(core_rate, master->base.bus.scl_rate.i3c) - hcnt;
 584	if (lcnt < SCL_I3C_TIMING_CNT_MIN)
 585		lcnt = SCL_I3C_TIMING_CNT_MIN;
 586
 587	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 588	writel(scl_timing, master->regs + SCL_I3C_PP_TIMING);
 589	master->i3c_pp_timing = scl_timing;
 590
 591	/*
 592	 * In pure i3c mode, MST_FREE represents tCAS. In shared mode, this
 593	 * will be set up by dw_i2c_clk_cfg as tLOW.
 594	 */
 595	if (master->base.bus.mode == I3C_BUS_MODE_PURE) {
 596		writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 597		master->bus_free_timing = BUS_I3C_MST_FREE(lcnt);
 598	}
 599
 600	lcnt = max_t(u8,
 601		     DIV_ROUND_UP(I3C_BUS_TLOW_OD_MIN_NS, core_period), lcnt);
 602	scl_timing = SCL_I3C_TIMING_HCNT(hcnt) | SCL_I3C_TIMING_LCNT(lcnt);
 603	writel(scl_timing, master->regs + SCL_I3C_OD_TIMING);
 604	master->i3c_od_timing = scl_timing;
 605
 606	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR1_SCL_RATE) - hcnt;
 607	scl_timing = SCL_EXT_LCNT_1(lcnt);
 608	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR2_SCL_RATE) - hcnt;
 609	scl_timing |= SCL_EXT_LCNT_2(lcnt);
 610	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR3_SCL_RATE) - hcnt;
 611	scl_timing |= SCL_EXT_LCNT_3(lcnt);
 612	lcnt = DIV_ROUND_UP(core_rate, I3C_BUS_SDR4_SCL_RATE) - hcnt;
 613	scl_timing |= SCL_EXT_LCNT_4(lcnt);
 614	writel(scl_timing, master->regs + SCL_EXT_LCNT_TIMING);
 615	master->ext_lcnt_timing = scl_timing;
 616
 617	return 0;
 618}
 619
 620static int dw_i2c_clk_cfg(struct dw_i3c_master *master)
 621{
 622	unsigned long core_rate, core_period;
 623	u16 hcnt, lcnt;
 624	u32 scl_timing;
 625
 626	core_rate = clk_get_rate(master->core_clk);
 627	if (!core_rate)
 628		return -EINVAL;
 629
 630	core_period = DIV_ROUND_UP(1000000000, core_rate);
 631
 632	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FMP_TLOW_MIN_NS, core_period);
 633	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_PLUS_SCL_RATE) - lcnt;
 634	scl_timing = SCL_I2C_FMP_TIMING_HCNT(hcnt) |
 635		     SCL_I2C_FMP_TIMING_LCNT(lcnt);
 636	writel(scl_timing, master->regs + SCL_I2C_FMP_TIMING);
 637	master->i2c_fmp_timing = scl_timing;
 638
 639	lcnt = DIV_ROUND_UP(I3C_BUS_I2C_FM_TLOW_MIN_NS, core_period);
 640	hcnt = DIV_ROUND_UP(core_rate, I3C_BUS_I2C_FM_SCL_RATE) - lcnt;
 641	scl_timing = SCL_I2C_FM_TIMING_HCNT(hcnt) |
 642		     SCL_I2C_FM_TIMING_LCNT(lcnt);
 643	writel(scl_timing, master->regs + SCL_I2C_FM_TIMING);
 644	master->i2c_fm_timing = scl_timing;
 645
 646	writel(BUS_I3C_MST_FREE(lcnt), master->regs + BUS_FREE_TIMING);
 647	master->bus_free_timing = BUS_I3C_MST_FREE(lcnt);
 648
 649	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_I2C_SLAVE_PRESENT,
 650	       master->regs + DEVICE_CTRL);
 651	master->i2c_slv_prsnt = true;
 652
 653	return 0;
 654}
 655
 656static int dw_i3c_master_bus_init(struct i3c_master_controller *m)
 657{
 658	struct dw_i3c_master *master = to_dw_i3c_master(m);
 659	struct i3c_bus *bus = i3c_master_get_bus(m);
 660	struct i3c_device_info info = { };
 
 661	int ret;
 662
 663	ret = pm_runtime_resume_and_get(master->dev);
 664	if (ret < 0) {
 665		dev_err(master->dev,
 666			"<%s> cannot resume i3c bus master, err: %d\n",
 667			__func__, ret);
 668		return ret;
 669	}
 670
 671	ret = master->platform_ops->init(master);
 672	if (ret)
 673		goto rpm_out;
 674
 675	switch (bus->mode) {
 676	case I3C_BUS_MODE_MIXED_FAST:
 677	case I3C_BUS_MODE_MIXED_LIMITED:
 678		ret = dw_i2c_clk_cfg(master);
 679		if (ret)
 680			goto rpm_out;
 681		fallthrough;
 682	case I3C_BUS_MODE_PURE:
 683		ret = dw_i3c_clk_cfg(master);
 684		if (ret)
 685			goto rpm_out;
 686		break;
 687	default:
 688		ret = -EINVAL;
 689		goto rpm_out;
 690	}
 691
 
 
 
 
 
 
 
 
 
 
 
 
 692	ret = i3c_master_get_free_addr(m, 0);
 693	if (ret < 0)
 694		goto rpm_out;
 695
 696	writel(DEV_ADDR_DYNAMIC_ADDR_VALID | DEV_ADDR_DYNAMIC(ret),
 697	       master->regs + DEVICE_ADDR);
 698	master->dev_addr = ret;
 699	memset(&info, 0, sizeof(info));
 700	info.dyn_addr = ret;
 701
 702	ret = i3c_master_set_info(&master->base, &info);
 703	if (ret)
 704		goto rpm_out;
 
 
 
 
 
 
 
 705
 706	dw_i3c_master_set_intr_regs(master);
 707	dw_i3c_master_enable(master);
 708
 709rpm_out:
 710	pm_runtime_mark_last_busy(master->dev);
 711	pm_runtime_put_autosuspend(master->dev);
 712	return ret;
 713}
 714
 715static void dw_i3c_master_bus_cleanup(struct i3c_master_controller *m)
 716{
 717	struct dw_i3c_master *master = to_dw_i3c_master(m);
 718
 719	dw_i3c_master_disable(master);
 720}
 721
 722static int dw_i3c_ccc_set(struct dw_i3c_master *master,
 723			  struct i3c_ccc_cmd *ccc)
 724{
 725	struct dw_i3c_xfer *xfer;
 726	struct dw_i3c_cmd *cmd;
 727	int ret, pos = 0;
 728
 729	if (ccc->id & I3C_CCC_DIRECT) {
 730		pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 731		if (pos < 0)
 732			return pos;
 733	}
 734
 735	xfer = dw_i3c_master_alloc_xfer(master, 1);
 736	if (!xfer)
 737		return -ENOMEM;
 738
 739	cmd = xfer->cmds;
 740	cmd->tx_buf = ccc->dests[0].payload.data;
 741	cmd->tx_len = ccc->dests[0].payload.len;
 742
 743	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 744		      COMMAND_PORT_TRANSFER_ARG;
 745
 746	cmd->cmd_lo = COMMAND_PORT_CP |
 747		      COMMAND_PORT_DEV_INDEX(pos) |
 748		      COMMAND_PORT_CMD(ccc->id) |
 749		      COMMAND_PORT_TOC |
 750		      COMMAND_PORT_ROC;
 751
 752	dw_i3c_master_enqueue_xfer(master, xfer);
 753	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 754		dw_i3c_master_dequeue_xfer(master, xfer);
 755
 756	ret = xfer->ret;
 757	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 758		ccc->err = I3C_ERROR_M2;
 759
 760	dw_i3c_master_free_xfer(xfer);
 761
 762	return ret;
 763}
 764
 765static int dw_i3c_ccc_get(struct dw_i3c_master *master, struct i3c_ccc_cmd *ccc)
 766{
 767	struct dw_i3c_xfer *xfer;
 768	struct dw_i3c_cmd *cmd;
 769	int ret, pos;
 770
 771	pos = dw_i3c_master_get_addr_pos(master, ccc->dests[0].addr);
 772	if (pos < 0)
 773		return pos;
 774
 775	xfer = dw_i3c_master_alloc_xfer(master, 1);
 776	if (!xfer)
 777		return -ENOMEM;
 778
 779	cmd = xfer->cmds;
 780	cmd->rx_buf = ccc->dests[0].payload.data;
 781	cmd->rx_len = ccc->dests[0].payload.len;
 782
 783	cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(ccc->dests[0].payload.len) |
 784		      COMMAND_PORT_TRANSFER_ARG;
 785
 786	cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 787		      COMMAND_PORT_CP |
 788		      COMMAND_PORT_DEV_INDEX(pos) |
 789		      COMMAND_PORT_CMD(ccc->id) |
 790		      COMMAND_PORT_TOC |
 791		      COMMAND_PORT_ROC;
 792
 793	dw_i3c_master_enqueue_xfer(master, xfer);
 794	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 795		dw_i3c_master_dequeue_xfer(master, xfer);
 796
 797	ret = xfer->ret;
 798	if (xfer->cmds[0].error == RESPONSE_ERROR_IBA_NACK)
 799		ccc->err = I3C_ERROR_M2;
 800	dw_i3c_master_free_xfer(xfer);
 801
 802	return ret;
 803}
 804
 805static void amd_configure_od_pp_quirk(struct dw_i3c_master *master)
 806{
 807	master->i3c_od_timing = AMD_I3C_OD_TIMING;
 808	master->i3c_pp_timing = AMD_I3C_PP_TIMING;
 809}
 810
 811static int dw_i3c_master_send_ccc_cmd(struct i3c_master_controller *m,
 812				      struct i3c_ccc_cmd *ccc)
 813{
 814	struct dw_i3c_master *master = to_dw_i3c_master(m);
 815	int ret = 0;
 816
 817	if (ccc->id == I3C_CCC_ENTDAA)
 818		return -EINVAL;
 819
 820	/* AMD platform specific OD and PP timings */
 821	if (master->quirks & AMD_I3C_OD_PP_TIMING) {
 822		amd_configure_od_pp_quirk(master);
 823		writel(master->i3c_pp_timing, master->regs + SCL_I3C_PP_TIMING);
 824		writel(master->i3c_od_timing, master->regs + SCL_I3C_OD_TIMING);
 825	}
 826
 827	ret = pm_runtime_resume_and_get(master->dev);
 828	if (ret < 0) {
 829		dev_err(master->dev,
 830			"<%s> cannot resume i3c bus master, err: %d\n",
 831			__func__, ret);
 832		return ret;
 833	}
 834
 835	if (ccc->rnw)
 836		ret = dw_i3c_ccc_get(master, ccc);
 837	else
 838		ret = dw_i3c_ccc_set(master, ccc);
 839
 840	pm_runtime_mark_last_busy(master->dev);
 841	pm_runtime_put_autosuspend(master->dev);
 842	return ret;
 843}
 844
 845static int dw_i3c_master_daa(struct i3c_master_controller *m)
 846{
 847	struct dw_i3c_master *master = to_dw_i3c_master(m);
 848	struct dw_i3c_xfer *xfer;
 849	struct dw_i3c_cmd *cmd;
 850	u32 olddevs, newdevs;
 851	u8 p, last_addr = 0;
 852	int ret, pos;
 853
 854	ret = pm_runtime_resume_and_get(master->dev);
 855	if (ret < 0) {
 856		dev_err(master->dev,
 857			"<%s> cannot resume i3c bus master, err: %d\n",
 858			__func__, ret);
 859		return ret;
 860	}
 861
 862	olddevs = ~(master->free_pos);
 863
 864	/* Prepare DAT before launching DAA. */
 865	for (pos = 0; pos < master->maxdevs; pos++) {
 866		if (olddevs & BIT(pos))
 867			continue;
 868
 869		ret = i3c_master_get_free_addr(m, last_addr + 1);
 870		if (ret < 0) {
 871			ret = -ENOSPC;
 872			goto rpm_out;
 873		}
 874
 875		master->devs[pos].addr = ret;
 876		p = even_parity(ret);
 877		last_addr = ret;
 878		ret |= (p << 7);
 879
 880		writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(ret),
 881		       master->regs +
 882		       DEV_ADDR_TABLE_LOC(master->datstartaddr, pos));
 883
 884		ret = 0;
 885	}
 886
 887	xfer = dw_i3c_master_alloc_xfer(master, 1);
 888	if (!xfer) {
 889		ret = -ENOMEM;
 890		goto rpm_out;
 891	}
 892
 893	pos = dw_i3c_master_get_free_pos(master);
 894	if (pos < 0) {
 895		dw_i3c_master_free_xfer(xfer);
 896		ret = pos;
 897		goto rpm_out;
 898	}
 899	cmd = &xfer->cmds[0];
 900	cmd->cmd_hi = 0x1;
 901	cmd->cmd_lo = COMMAND_PORT_DEV_COUNT(master->maxdevs - pos) |
 902		      COMMAND_PORT_DEV_INDEX(pos) |
 903		      COMMAND_PORT_CMD(I3C_CCC_ENTDAA) |
 904		      COMMAND_PORT_ADDR_ASSGN_CMD |
 905		      COMMAND_PORT_TOC |
 906		      COMMAND_PORT_ROC;
 907
 908	dw_i3c_master_enqueue_xfer(master, xfer);
 909	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 910		dw_i3c_master_dequeue_xfer(master, xfer);
 911
 912	newdevs = GENMASK(master->maxdevs - cmd->rx_len - 1, 0);
 913	newdevs &= ~olddevs;
 914
 915	for (pos = 0; pos < master->maxdevs; pos++) {
 916		if (newdevs & BIT(pos))
 917			i3c_master_add_i3c_dev_locked(m, master->devs[pos].addr);
 918	}
 919
 920	dw_i3c_master_free_xfer(xfer);
 921
 922rpm_out:
 923	pm_runtime_mark_last_busy(master->dev);
 924	pm_runtime_put_autosuspend(master->dev);
 925	return ret;
 926}
 927
 928static int dw_i3c_master_priv_xfers(struct i3c_dev_desc *dev,
 929				    struct i3c_priv_xfer *i3c_xfers,
 930				    int i3c_nxfers)
 931{
 932	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
 933	struct i3c_master_controller *m = i3c_dev_get_master(dev);
 934	struct dw_i3c_master *master = to_dw_i3c_master(m);
 935	unsigned int nrxwords = 0, ntxwords = 0;
 936	struct dw_i3c_xfer *xfer;
 937	int i, ret = 0;
 938
 939	if (!i3c_nxfers)
 940		return 0;
 941
 942	if (i3c_nxfers > master->caps.cmdfifodepth)
 943		return -ENOTSUPP;
 944
 945	for (i = 0; i < i3c_nxfers; i++) {
 946		if (i3c_xfers[i].rnw)
 947			nrxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 948		else
 949			ntxwords += DIV_ROUND_UP(i3c_xfers[i].len, 4);
 950	}
 951
 952	if (ntxwords > master->caps.datafifodepth ||
 953	    nrxwords > master->caps.datafifodepth)
 954		return -ENOTSUPP;
 955
 956	xfer = dw_i3c_master_alloc_xfer(master, i3c_nxfers);
 957	if (!xfer)
 958		return -ENOMEM;
 959
 960	ret = pm_runtime_resume_and_get(master->dev);
 961	if (ret < 0) {
 962		dev_err(master->dev,
 963			"<%s> cannot resume i3c bus master, err: %d\n",
 964			__func__, ret);
 965		return ret;
 966	}
 967
 968	for (i = 0; i < i3c_nxfers; i++) {
 969		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
 970
 971		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i3c_xfers[i].len) |
 972			COMMAND_PORT_TRANSFER_ARG;
 973
 974		if (i3c_xfers[i].rnw) {
 975			cmd->rx_buf = i3c_xfers[i].data.in;
 976			cmd->rx_len = i3c_xfers[i].len;
 977			cmd->cmd_lo = COMMAND_PORT_READ_TRANSFER |
 978				      COMMAND_PORT_SPEED(dev->info.max_read_ds);
 979
 980		} else {
 981			cmd->tx_buf = i3c_xfers[i].data.out;
 982			cmd->tx_len = i3c_xfers[i].len;
 983			cmd->cmd_lo =
 984				COMMAND_PORT_SPEED(dev->info.max_write_ds);
 985		}
 986
 987		cmd->cmd_lo |= COMMAND_PORT_TID(i) |
 988			       COMMAND_PORT_DEV_INDEX(data->index) |
 989			       COMMAND_PORT_ROC;
 990
 991		if (i == (i3c_nxfers - 1))
 992			cmd->cmd_lo |= COMMAND_PORT_TOC;
 993	}
 994
 995	dw_i3c_master_enqueue_xfer(master, xfer);
 996	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
 997		dw_i3c_master_dequeue_xfer(master, xfer);
 998
 999	for (i = 0; i < i3c_nxfers; i++) {
1000		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
1001
1002		if (i3c_xfers[i].rnw)
1003			i3c_xfers[i].len = cmd->rx_len;
1004	}
1005
1006	ret = xfer->ret;
1007	dw_i3c_master_free_xfer(xfer);
1008
1009	pm_runtime_mark_last_busy(master->dev);
1010	pm_runtime_put_autosuspend(master->dev);
1011	return ret;
1012}
1013
1014static int dw_i3c_master_reattach_i3c_dev(struct i3c_dev_desc *dev,
1015					  u8 old_dyn_addr)
1016{
1017	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1018	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1019	struct dw_i3c_master *master = to_dw_i3c_master(m);
1020	int pos;
1021
1022	pos = dw_i3c_master_get_free_pos(master);
1023
1024	if (data->index > pos && pos > 0) {
1025		writel(0,
1026		       master->regs +
1027		       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1028
1029		master->devs[data->index].addr = 0;
1030		master->free_pos |= BIT(data->index);
1031
1032		data->index = pos;
1033		master->devs[pos].addr = dev->info.dyn_addr;
1034		master->free_pos &= ~BIT(pos);
1035	}
1036
1037	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(dev->info.dyn_addr),
1038	       master->regs +
1039	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1040
1041	master->devs[data->index].addr = dev->info.dyn_addr;
1042
1043	return 0;
1044}
1045
1046static int dw_i3c_master_attach_i3c_dev(struct i3c_dev_desc *dev)
1047{
1048	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1049	struct dw_i3c_master *master = to_dw_i3c_master(m);
1050	struct dw_i3c_i2c_dev_data *data;
1051	int pos;
1052
1053	pos = dw_i3c_master_get_free_pos(master);
1054	if (pos < 0)
1055		return pos;
1056
1057	data = kzalloc(sizeof(*data), GFP_KERNEL);
1058	if (!data)
1059		return -ENOMEM;
1060
1061	data->index = pos;
1062	master->devs[pos].addr = dev->info.dyn_addr ? : dev->info.static_addr;
1063	master->free_pos &= ~BIT(pos);
1064	i3c_dev_set_master_data(dev, data);
1065
1066	writel(DEV_ADDR_TABLE_DYNAMIC_ADDR(master->devs[pos].addr),
1067	       master->regs +
1068	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1069
1070	return 0;
1071}
1072
1073static void dw_i3c_master_detach_i3c_dev(struct i3c_dev_desc *dev)
1074{
1075	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1076	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1077	struct dw_i3c_master *master = to_dw_i3c_master(m);
1078
1079	writel(0,
1080	       master->regs +
1081	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1082
1083	i3c_dev_set_master_data(dev, NULL);
1084	master->devs[data->index].addr = 0;
1085	master->free_pos |= BIT(data->index);
1086	kfree(data);
1087}
1088
1089static int dw_i3c_master_i2c_xfers(struct i2c_dev_desc *dev,
1090				   const struct i2c_msg *i2c_xfers,
1091				   int i2c_nxfers)
1092{
1093	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1094	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1095	struct dw_i3c_master *master = to_dw_i3c_master(m);
1096	unsigned int nrxwords = 0, ntxwords = 0;
1097	struct dw_i3c_xfer *xfer;
1098	int i, ret = 0;
1099
1100	if (!i2c_nxfers)
1101		return 0;
1102
1103	if (i2c_nxfers > master->caps.cmdfifodepth)
1104		return -ENOTSUPP;
1105
1106	for (i = 0; i < i2c_nxfers; i++) {
1107		if (i2c_xfers[i].flags & I2C_M_RD)
1108			nrxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
1109		else
1110			ntxwords += DIV_ROUND_UP(i2c_xfers[i].len, 4);
1111	}
1112
1113	if (ntxwords > master->caps.datafifodepth ||
1114	    nrxwords > master->caps.datafifodepth)
1115		return -ENOTSUPP;
1116
1117	xfer = dw_i3c_master_alloc_xfer(master, i2c_nxfers);
1118	if (!xfer)
1119		return -ENOMEM;
1120
1121	ret = pm_runtime_resume_and_get(master->dev);
1122	if (ret < 0) {
1123		dev_err(master->dev,
1124			"<%s> cannot resume i3c bus master, err: %d\n",
1125			__func__, ret);
1126		return ret;
1127	}
1128
1129	for (i = 0; i < i2c_nxfers; i++) {
1130		struct dw_i3c_cmd *cmd = &xfer->cmds[i];
1131
1132		cmd->cmd_hi = COMMAND_PORT_ARG_DATA_LEN(i2c_xfers[i].len) |
1133			COMMAND_PORT_TRANSFER_ARG;
1134
1135		cmd->cmd_lo = COMMAND_PORT_TID(i) |
1136			      COMMAND_PORT_DEV_INDEX(data->index) |
1137			      COMMAND_PORT_ROC;
1138
1139		if (i2c_xfers[i].flags & I2C_M_RD) {
1140			cmd->cmd_lo |= COMMAND_PORT_READ_TRANSFER;
1141			cmd->rx_buf = i2c_xfers[i].buf;
1142			cmd->rx_len = i2c_xfers[i].len;
1143		} else {
1144			cmd->tx_buf = i2c_xfers[i].buf;
1145			cmd->tx_len = i2c_xfers[i].len;
1146		}
1147
1148		if (i == (i2c_nxfers - 1))
1149			cmd->cmd_lo |= COMMAND_PORT_TOC;
1150	}
1151
1152	dw_i3c_master_enqueue_xfer(master, xfer);
1153	if (!wait_for_completion_timeout(&xfer->comp, XFER_TIMEOUT))
1154		dw_i3c_master_dequeue_xfer(master, xfer);
1155
1156	ret = xfer->ret;
1157	dw_i3c_master_free_xfer(xfer);
1158
1159	pm_runtime_mark_last_busy(master->dev);
1160	pm_runtime_put_autosuspend(master->dev);
1161	return ret;
1162}
1163
1164static int dw_i3c_master_attach_i2c_dev(struct i2c_dev_desc *dev)
1165{
1166	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1167	struct dw_i3c_master *master = to_dw_i3c_master(m);
1168	struct dw_i3c_i2c_dev_data *data;
1169	int pos;
1170
1171	pos = dw_i3c_master_get_free_pos(master);
1172	if (pos < 0)
1173		return pos;
1174
1175	data = kzalloc(sizeof(*data), GFP_KERNEL);
1176	if (!data)
1177		return -ENOMEM;
1178
1179	data->index = pos;
1180	master->devs[pos].addr = dev->addr;
1181	master->devs[pos].is_i2c_addr = true;
1182	master->free_pos &= ~BIT(pos);
1183	i2c_dev_set_master_data(dev, data);
1184
1185	writel(DEV_ADDR_TABLE_LEGACY_I2C_DEV |
1186	       DEV_ADDR_TABLE_STATIC_ADDR(dev->addr),
1187	       master->regs +
1188	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1189
1190	return 0;
1191}
1192
1193static void dw_i3c_master_detach_i2c_dev(struct i2c_dev_desc *dev)
1194{
1195	struct dw_i3c_i2c_dev_data *data = i2c_dev_get_master_data(dev);
1196	struct i3c_master_controller *m = i2c_dev_get_master(dev);
1197	struct dw_i3c_master *master = to_dw_i3c_master(m);
1198
1199	writel(0,
1200	       master->regs +
1201	       DEV_ADDR_TABLE_LOC(master->datstartaddr, data->index));
1202
1203	i2c_dev_set_master_data(dev, NULL);
1204	master->devs[data->index].addr = 0;
1205	master->free_pos |= BIT(data->index);
1206	kfree(data);
1207}
1208
1209static int dw_i3c_master_request_ibi(struct i3c_dev_desc *dev,
1210				     const struct i3c_ibi_setup *req)
1211{
1212	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1213	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1214	struct dw_i3c_master *master = to_dw_i3c_master(m);
1215	unsigned long flags;
1216
1217	data->ibi_pool = i3c_generic_ibi_alloc_pool(dev, req);
1218	if (IS_ERR(data->ibi_pool))
1219		return PTR_ERR(data->ibi_pool);
1220
1221	spin_lock_irqsave(&master->devs_lock, flags);
1222	master->devs[data->index].ibi_dev = dev;
1223	spin_unlock_irqrestore(&master->devs_lock, flags);
1224
1225	return 0;
1226}
1227
1228static void dw_i3c_master_free_ibi(struct i3c_dev_desc *dev)
1229{
1230	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1231	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1232	struct dw_i3c_master *master = to_dw_i3c_master(m);
1233	unsigned long flags;
1234
1235	spin_lock_irqsave(&master->devs_lock, flags);
1236	master->devs[data->index].ibi_dev = NULL;
1237	spin_unlock_irqrestore(&master->devs_lock, flags);
1238
1239	i3c_generic_ibi_free_pool(data->ibi_pool);
1240	data->ibi_pool = NULL;
1241}
1242
1243static void dw_i3c_master_enable_sir_signal(struct dw_i3c_master *master, bool enable)
1244{
1245	u32 reg;
1246
1247	reg = readl(master->regs + INTR_STATUS_EN);
1248	reg &= ~INTR_IBI_THLD_STAT;
1249	if (enable)
1250		reg |= INTR_IBI_THLD_STAT;
1251	writel(reg, master->regs + INTR_STATUS_EN);
1252
1253	reg = readl(master->regs + INTR_SIGNAL_EN);
1254	reg &= ~INTR_IBI_THLD_STAT;
1255	if (enable)
1256		reg |= INTR_IBI_THLD_STAT;
1257	writel(reg, master->regs + INTR_SIGNAL_EN);
1258}
1259
1260static void dw_i3c_master_set_sir_enabled(struct dw_i3c_master *master,
1261					  struct i3c_dev_desc *dev,
1262					  u8 idx, bool enable)
1263{
1264	unsigned long flags;
1265	u32 dat_entry, reg;
1266	bool global;
1267
1268	dat_entry = DEV_ADDR_TABLE_LOC(master->datstartaddr, idx);
1269
1270	spin_lock_irqsave(&master->devs_lock, flags);
1271	reg = readl(master->regs + dat_entry);
1272	if (enable) {
1273		reg &= ~DEV_ADDR_TABLE_SIR_REJECT;
1274		if (dev->info.bcr & I3C_BCR_IBI_PAYLOAD)
1275			reg |= DEV_ADDR_TABLE_IBI_MDB;
1276	} else {
1277		reg |= DEV_ADDR_TABLE_SIR_REJECT;
1278	}
1279	master->platform_ops->set_dat_ibi(master, dev, enable, &reg);
1280	writel(reg, master->regs + dat_entry);
1281
1282	if (enable) {
1283		global = (master->sir_rej_mask == IBI_REQ_REJECT_ALL);
1284		master->sir_rej_mask &= ~BIT(idx);
1285	} else {
1286		bool hj_rejected = !!(readl(master->regs + DEVICE_CTRL) & DEV_CTRL_HOT_JOIN_NACK);
1287
1288		master->sir_rej_mask |= BIT(idx);
1289		global = (master->sir_rej_mask == IBI_REQ_REJECT_ALL) && hj_rejected;
1290	}
1291	writel(master->sir_rej_mask, master->regs + IBI_SIR_REQ_REJECT);
1292
1293	if (global)
1294		dw_i3c_master_enable_sir_signal(master, enable);
1295
1296
1297	spin_unlock_irqrestore(&master->devs_lock, flags);
1298}
1299
1300static int dw_i3c_master_enable_hotjoin(struct i3c_master_controller *m)
1301{
1302	struct dw_i3c_master *master = to_dw_i3c_master(m);
1303	int ret;
1304
1305	ret = pm_runtime_resume_and_get(master->dev);
1306	if (ret < 0) {
1307		dev_err(master->dev,
1308			"<%s> cannot resume i3c bus master, err: %d\n",
1309			__func__, ret);
1310		return ret;
1311	}
1312
1313	dw_i3c_master_enable_sir_signal(master, true);
1314	writel(readl(master->regs + DEVICE_CTRL) & ~DEV_CTRL_HOT_JOIN_NACK,
1315	       master->regs + DEVICE_CTRL);
1316
1317	return 0;
1318}
1319
1320static int dw_i3c_master_disable_hotjoin(struct i3c_master_controller *m)
1321{
1322	struct dw_i3c_master *master = to_dw_i3c_master(m);
1323
1324	writel(readl(master->regs + DEVICE_CTRL) | DEV_CTRL_HOT_JOIN_NACK,
1325	       master->regs + DEVICE_CTRL);
1326
1327	pm_runtime_mark_last_busy(master->dev);
1328	pm_runtime_put_autosuspend(master->dev);
1329	return 0;
1330}
1331
1332static int dw_i3c_master_enable_ibi(struct i3c_dev_desc *dev)
1333{
1334	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1335	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1336	struct dw_i3c_master *master = to_dw_i3c_master(m);
1337	int rc;
1338
1339	rc = pm_runtime_resume_and_get(master->dev);
1340	if (rc < 0) {
1341		dev_err(master->dev,
1342			"<%s> cannot resume i3c bus master, err: %d\n",
1343			__func__, rc);
1344		return rc;
1345	}
1346
1347	dw_i3c_master_set_sir_enabled(master, dev, data->index, true);
1348
1349	rc = i3c_master_enec_locked(m, dev->info.dyn_addr, I3C_CCC_EVENT_SIR);
1350
1351	if (rc) {
1352		dw_i3c_master_set_sir_enabled(master, dev, data->index, false);
1353		pm_runtime_mark_last_busy(master->dev);
1354		pm_runtime_put_autosuspend(master->dev);
1355	}
1356
1357	return rc;
1358}
1359
1360static int dw_i3c_master_disable_ibi(struct i3c_dev_desc *dev)
1361{
1362	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1363	struct i3c_master_controller *m = i3c_dev_get_master(dev);
1364	struct dw_i3c_master *master = to_dw_i3c_master(m);
1365	int rc;
1366
1367	rc = i3c_master_disec_locked(m, dev->info.dyn_addr, I3C_CCC_EVENT_SIR);
1368	if (rc)
1369		return rc;
1370
1371	dw_i3c_master_set_sir_enabled(master, dev, data->index, false);
1372
1373	pm_runtime_mark_last_busy(master->dev);
1374	pm_runtime_put_autosuspend(master->dev);
1375	return 0;
1376}
1377
1378static void dw_i3c_master_recycle_ibi_slot(struct i3c_dev_desc *dev,
1379					   struct i3c_ibi_slot *slot)
1380{
1381	struct dw_i3c_i2c_dev_data *data = i3c_dev_get_master_data(dev);
1382
1383	i3c_generic_ibi_recycle_slot(data->ibi_pool, slot);
1384}
1385
1386static void dw_i3c_master_drain_ibi_queue(struct dw_i3c_master *master,
1387					  int len)
1388{
1389	int i;
1390
1391	for (i = 0; i < DIV_ROUND_UP(len, 4); i++)
1392		readl(master->regs + IBI_QUEUE_STATUS);
1393}
1394
1395static void dw_i3c_master_handle_ibi_sir(struct dw_i3c_master *master,
1396					 u32 status)
1397{
1398	struct dw_i3c_i2c_dev_data *data;
1399	struct i3c_ibi_slot *slot;
1400	struct i3c_dev_desc *dev;
1401	unsigned long flags;
1402	u8 addr, len;
1403	int idx;
1404
1405	addr = IBI_QUEUE_IBI_ADDR(status);
1406	len = IBI_QUEUE_STATUS_DATA_LEN(status);
1407
1408	/*
1409	 * We be tempted to check the error status in bit 30; however, due
1410	 * to the PEC errata workaround on some platform implementations (see
1411	 * ast2600_i3c_set_dat_ibi()), those will almost always have a PEC
1412	 * error on IBI payload data, as well as losing the last byte of
1413	 * payload.
1414	 *
1415	 * If we implement error status checking on that bit, we may need
1416	 * a new platform op to validate it.
1417	 */
1418
1419	spin_lock_irqsave(&master->devs_lock, flags);
1420	idx = dw_i3c_master_get_addr_pos(master, addr);
1421	if (idx < 0) {
1422		dev_dbg_ratelimited(&master->base.dev,
1423			 "IBI from unknown addr 0x%x\n", addr);
1424		goto err_drain;
1425	}
1426
1427	dev = master->devs[idx].ibi_dev;
1428	if (!dev || !dev->ibi) {
1429		dev_dbg_ratelimited(&master->base.dev,
1430			 "IBI from non-requested dev idx %d\n", idx);
1431		goto err_drain;
1432	}
1433
1434	data = i3c_dev_get_master_data(dev);
1435	slot = i3c_generic_ibi_get_free_slot(data->ibi_pool);
1436	if (!slot) {
1437		dev_dbg_ratelimited(&master->base.dev,
1438				    "No IBI slots available\n");
1439		goto err_drain;
1440	}
1441
1442	if (dev->ibi->max_payload_len < len) {
1443		dev_dbg_ratelimited(&master->base.dev,
1444				    "IBI payload len %d greater than max %d\n",
1445				    len, dev->ibi->max_payload_len);
1446		goto err_drain;
1447	}
1448
1449	if (len) {
1450		dw_i3c_master_read_ibi_fifo(master, slot->data, len);
1451		slot->len = len;
1452	}
1453	i3c_master_queue_ibi(dev, slot);
1454
1455	spin_unlock_irqrestore(&master->devs_lock, flags);
1456
1457	return;
1458
1459err_drain:
1460	dw_i3c_master_drain_ibi_queue(master, len);
1461
1462	spin_unlock_irqrestore(&master->devs_lock, flags);
1463}
1464
1465/* "ibis": referring to In-Band Interrupts, and not
1466 * https://en.wikipedia.org/wiki/Australian_white_ibis. The latter should
1467 * not be handled.
1468 */
1469static void dw_i3c_master_irq_handle_ibis(struct dw_i3c_master *master)
1470{
1471	unsigned int i, len, n_ibis;
1472	u32 reg;
1473
1474	reg = readl(master->regs + QUEUE_STATUS_LEVEL);
1475	n_ibis = QUEUE_STATUS_IBI_STATUS_CNT(reg);
1476	if (!n_ibis)
1477		return;
1478
1479	for (i = 0; i < n_ibis; i++) {
1480		reg = readl(master->regs + IBI_QUEUE_STATUS);
1481
1482		if (IBI_TYPE_SIRQ(reg)) {
1483			dw_i3c_master_handle_ibi_sir(master, reg);
1484		} else if (IBI_TYPE_HJ(reg)) {
1485			queue_work(master->base.wq, &master->hj_work);
1486		} else {
1487			len = IBI_QUEUE_STATUS_DATA_LEN(reg);
1488			dev_info(&master->base.dev,
1489				 "unsupported IBI type 0x%lx len %d\n",
1490				 IBI_QUEUE_STATUS_IBI_ID(reg), len);
1491			dw_i3c_master_drain_ibi_queue(master, len);
1492		}
1493	}
1494}
1495
1496static irqreturn_t dw_i3c_master_irq_handler(int irq, void *dev_id)
1497{
1498	struct dw_i3c_master *master = dev_id;
1499	u32 status;
1500
1501	status = readl(master->regs + INTR_STATUS);
1502
1503	if (!(status & readl(master->regs + INTR_STATUS_EN))) {
1504		writel(INTR_ALL, master->regs + INTR_STATUS);
1505		return IRQ_NONE;
1506	}
1507
1508	spin_lock(&master->xferqueue.lock);
1509	dw_i3c_master_end_xfer_locked(master, status);
1510	if (status & INTR_TRANSFER_ERR_STAT)
1511		writel(INTR_TRANSFER_ERR_STAT, master->regs + INTR_STATUS);
1512	spin_unlock(&master->xferqueue.lock);
1513
1514	if (status & INTR_IBI_THLD_STAT)
1515		dw_i3c_master_irq_handle_ibis(master);
1516
1517	return IRQ_HANDLED;
1518}
1519
1520static const struct i3c_master_controller_ops dw_mipi_i3c_ops = {
1521	.bus_init = dw_i3c_master_bus_init,
1522	.bus_cleanup = dw_i3c_master_bus_cleanup,
1523	.attach_i3c_dev = dw_i3c_master_attach_i3c_dev,
1524	.reattach_i3c_dev = dw_i3c_master_reattach_i3c_dev,
1525	.detach_i3c_dev = dw_i3c_master_detach_i3c_dev,
1526	.do_daa = dw_i3c_master_daa,
1527	.supports_ccc_cmd = dw_i3c_master_supports_ccc_cmd,
1528	.send_ccc_cmd = dw_i3c_master_send_ccc_cmd,
1529	.priv_xfers = dw_i3c_master_priv_xfers,
1530	.attach_i2c_dev = dw_i3c_master_attach_i2c_dev,
1531	.detach_i2c_dev = dw_i3c_master_detach_i2c_dev,
1532	.i2c_xfers = dw_i3c_master_i2c_xfers,
1533	.request_ibi = dw_i3c_master_request_ibi,
1534	.free_ibi = dw_i3c_master_free_ibi,
1535	.enable_ibi = dw_i3c_master_enable_ibi,
1536	.disable_ibi = dw_i3c_master_disable_ibi,
1537	.recycle_ibi_slot = dw_i3c_master_recycle_ibi_slot,
1538	.enable_hotjoin = dw_i3c_master_enable_hotjoin,
1539	.disable_hotjoin = dw_i3c_master_disable_hotjoin,
1540};
1541
1542/* default platform ops implementations */
1543static int dw_i3c_platform_init_nop(struct dw_i3c_master *i3c)
1544{
1545	return 0;
1546}
1547
1548static void dw_i3c_platform_set_dat_ibi_nop(struct dw_i3c_master *i3c,
1549					struct i3c_dev_desc *dev,
1550					bool enable, u32 *dat)
1551{
1552}
1553
1554static const struct dw_i3c_platform_ops dw_i3c_platform_ops_default = {
1555	.init = dw_i3c_platform_init_nop,
1556	.set_dat_ibi = dw_i3c_platform_set_dat_ibi_nop,
1557};
1558
1559static void dw_i3c_hj_work(struct work_struct *work)
1560{
1561	struct dw_i3c_master *master =
1562		container_of(work, typeof(*master), hj_work);
1563
1564	i3c_master_do_daa(&master->base);
1565}
1566
1567int dw_i3c_common_probe(struct dw_i3c_master *master,
1568			struct platform_device *pdev)
1569{
 
1570	int ret, irq;
1571
1572	if (!master->platform_ops)
1573		master->platform_ops = &dw_i3c_platform_ops_default;
1574
1575	master->dev = &pdev->dev;
1576
1577	master->regs = devm_platform_ioremap_resource(pdev, 0);
1578	if (IS_ERR(master->regs))
1579		return PTR_ERR(master->regs);
1580
1581	master->core_clk = devm_clk_get_enabled(&pdev->dev, NULL);
1582	if (IS_ERR(master->core_clk))
1583		return PTR_ERR(master->core_clk);
1584
1585	master->pclk = devm_clk_get_optional_enabled(&pdev->dev, "pclk");
1586	if (IS_ERR(master->pclk))
1587		return PTR_ERR(master->pclk);
1588
1589	master->core_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
1590								    "core_rst");
1591	if (IS_ERR(master->core_rst))
1592		return PTR_ERR(master->core_rst);
1593
 
 
 
 
1594	reset_control_deassert(master->core_rst);
1595
1596	spin_lock_init(&master->xferqueue.lock);
1597	INIT_LIST_HEAD(&master->xferqueue.list);
1598
1599	writel(INTR_ALL, master->regs + INTR_STATUS);
1600	irq = platform_get_irq(pdev, 0);
1601	ret = devm_request_irq(&pdev->dev, irq,
1602			       dw_i3c_master_irq_handler, 0,
1603			       dev_name(&pdev->dev), master);
1604	if (ret)
1605		goto err_assert_rst;
1606
1607	platform_set_drvdata(pdev, master);
1608
1609	pm_runtime_set_autosuspend_delay(&pdev->dev, RPM_AUTOSUSPEND_TIMEOUT);
1610	pm_runtime_use_autosuspend(&pdev->dev);
1611	pm_runtime_set_active(&pdev->dev);
1612	pm_runtime_enable(&pdev->dev);
1613
1614	/* Information regarding the FIFOs/QUEUEs depth */
1615	ret = readl(master->regs + QUEUE_STATUS_LEVEL);
1616	master->caps.cmdfifodepth = QUEUE_STATUS_LEVEL_CMD(ret);
1617
1618	ret = readl(master->regs + DATA_BUFFER_STATUS_LEVEL);
1619	master->caps.datafifodepth = DATA_BUFFER_STATUS_LEVEL_TX(ret);
1620
1621	ret = readl(master->regs + DEVICE_ADDR_TABLE_POINTER);
1622	master->datstartaddr = ret;
1623	master->maxdevs = ret >> 16;
1624	master->free_pos = GENMASK(master->maxdevs - 1, 0);
1625
1626	master->quirks = (unsigned long)device_get_match_data(&pdev->dev);
1627
1628	INIT_WORK(&master->hj_work, dw_i3c_hj_work);
1629	ret = i3c_master_register(&master->base, &pdev->dev,
1630				  &dw_mipi_i3c_ops, false);
1631	if (ret)
1632		goto err_disable_pm;
1633
1634	return 0;
1635
1636err_disable_pm:
1637	pm_runtime_disable(&pdev->dev);
1638	pm_runtime_set_suspended(&pdev->dev);
1639	pm_runtime_dont_use_autosuspend(&pdev->dev);
1640
1641err_assert_rst:
1642	reset_control_assert(master->core_rst);
1643
 
 
 
1644	return ret;
1645}
1646EXPORT_SYMBOL_GPL(dw_i3c_common_probe);
1647
1648void dw_i3c_common_remove(struct dw_i3c_master *master)
1649{
1650	cancel_work_sync(&master->hj_work);
1651	i3c_master_unregister(&master->base);
1652
1653	pm_runtime_disable(master->dev);
1654	pm_runtime_set_suspended(master->dev);
1655	pm_runtime_dont_use_autosuspend(master->dev);
1656}
1657EXPORT_SYMBOL_GPL(dw_i3c_common_remove);
1658
1659/* base platform implementation */
1660
1661static int dw_i3c_probe(struct platform_device *pdev)
1662{
1663	struct dw_i3c_master *master;
1664
1665	master = devm_kzalloc(&pdev->dev, sizeof(*master), GFP_KERNEL);
1666	if (!master)
1667		return -ENOMEM;
1668
1669	return dw_i3c_common_probe(master, pdev);
1670}
1671
1672static void dw_i3c_remove(struct platform_device *pdev)
1673{
1674	struct dw_i3c_master *master = platform_get_drvdata(pdev);
 
1675
1676	dw_i3c_common_remove(master);
1677}
1678
1679static void dw_i3c_master_restore_addrs(struct dw_i3c_master *master)
1680{
1681	u32 pos, reg_val;
1682
1683	writel(DEV_ADDR_DYNAMIC_ADDR_VALID | DEV_ADDR_DYNAMIC(master->dev_addr),
1684	       master->regs + DEVICE_ADDR);
1685
1686	for (pos = 0; pos < master->maxdevs; pos++) {
1687		if (master->free_pos & BIT(pos))
1688			continue;
1689
1690		if (master->devs[pos].is_i2c_addr)
1691			reg_val = DEV_ADDR_TABLE_LEGACY_I2C_DEV |
1692			       DEV_ADDR_TABLE_STATIC_ADDR(master->devs[pos].addr);
1693		else
1694			reg_val = DEV_ADDR_TABLE_DYNAMIC_ADDR(master->devs[pos].addr);
1695
1696		writel(reg_val, master->regs + DEV_ADDR_TABLE_LOC(master->datstartaddr, pos));
1697	}
1698}
1699
1700static void dw_i3c_master_restore_timing_regs(struct dw_i3c_master *master)
1701{
1702	/* AMD platform specific OD and PP timings */
1703	if (master->quirks & AMD_I3C_OD_PP_TIMING)
1704		amd_configure_od_pp_quirk(master);
1705
1706	writel(master->i3c_pp_timing, master->regs + SCL_I3C_PP_TIMING);
1707	writel(master->bus_free_timing, master->regs + BUS_FREE_TIMING);
1708	writel(master->i3c_od_timing, master->regs + SCL_I3C_OD_TIMING);
1709	writel(master->ext_lcnt_timing, master->regs + SCL_EXT_LCNT_TIMING);
1710
1711	if (master->i2c_slv_prsnt) {
1712		writel(master->i2c_fmp_timing, master->regs + SCL_I2C_FMP_TIMING);
1713		writel(master->i2c_fm_timing, master->regs + SCL_I2C_FM_TIMING);
1714	}
1715}
1716
1717static int dw_i3c_master_enable_clks(struct dw_i3c_master *master)
1718{
1719	int ret = 0;
1720
1721	ret = clk_prepare_enable(master->core_clk);
1722	if (ret)
1723		return ret;
1724
1725	ret = clk_prepare_enable(master->pclk);
1726	if (ret) {
1727		clk_disable_unprepare(master->core_clk);
1728		return ret;
1729	}
1730
1731	return 0;
1732}
1733
1734static inline void dw_i3c_master_disable_clks(struct dw_i3c_master *master)
1735{
1736	clk_disable_unprepare(master->pclk);
1737	clk_disable_unprepare(master->core_clk);
1738}
1739
1740static int __maybe_unused dw_i3c_master_runtime_suspend(struct device *dev)
1741{
1742	struct dw_i3c_master *master = dev_get_drvdata(dev);
1743
1744	dw_i3c_master_disable(master);
1745
1746	reset_control_assert(master->core_rst);
1747	dw_i3c_master_disable_clks(master);
1748	pinctrl_pm_select_sleep_state(dev);
1749	return 0;
1750}
1751
1752static int __maybe_unused dw_i3c_master_runtime_resume(struct device *dev)
1753{
1754	struct dw_i3c_master *master = dev_get_drvdata(dev);
1755
1756	pinctrl_pm_select_default_state(dev);
1757	dw_i3c_master_enable_clks(master);
1758	reset_control_deassert(master->core_rst);
1759
1760	dw_i3c_master_set_intr_regs(master);
1761	dw_i3c_master_restore_timing_regs(master);
1762	dw_i3c_master_restore_addrs(master);
1763
1764	dw_i3c_master_enable(master);
1765	return 0;
1766}
1767
1768static const struct dev_pm_ops dw_i3c_pm_ops = {
1769	SET_SYSTEM_SLEEP_PM_OPS(pm_runtime_force_suspend, pm_runtime_force_resume)
1770	SET_RUNTIME_PM_OPS(dw_i3c_master_runtime_suspend, dw_i3c_master_runtime_resume, NULL)
1771};
1772
1773static const struct of_device_id dw_i3c_master_of_match[] = {
1774	{ .compatible = "snps,dw-i3c-master-1.00a", },
1775	{},
1776};
1777MODULE_DEVICE_TABLE(of, dw_i3c_master_of_match);
1778
1779static const struct acpi_device_id amd_i3c_device_match[] = {
1780	{ "AMDI0015", AMD_I3C_OD_PP_TIMING },
1781	{ }
1782};
1783MODULE_DEVICE_TABLE(acpi, amd_i3c_device_match);
1784
1785static struct platform_driver dw_i3c_driver = {
1786	.probe = dw_i3c_probe,
1787	.remove = dw_i3c_remove,
1788	.driver = {
1789		.name = "dw-i3c-master",
1790		.of_match_table = dw_i3c_master_of_match,
1791		.acpi_match_table = amd_i3c_device_match,
1792		.pm = &dw_i3c_pm_ops,
1793	},
1794};
1795module_platform_driver(dw_i3c_driver);
1796
1797MODULE_AUTHOR("Vitor Soares <vitor.soares@synopsys.com>");
1798MODULE_DESCRIPTION("DesignWare MIPI I3C driver");
1799MODULE_LICENSE("GPL v2");