[PATCH v3 2/2] i3c: master: Add driver for MediaTek IP

From: Qii Wang
Date: Tue Jul 09 2019 - 09:10:04 EST


Add a driver for MediaTek I3C master IP.

Signed-off-by: Qii Wang <qii.wang@xxxxxxxxxxxx>
---
drivers/i3c/master/Kconfig | 10 +
drivers/i3c/master/Makefile | 1 +
drivers/i3c/master/i3c-master-mtk.c | 1239 +++++++++++++++++++++++++++++++++++
3 files changed, 1250 insertions(+)
create mode 100644 drivers/i3c/master/i3c-master-mtk.c

diff --git a/drivers/i3c/master/Kconfig b/drivers/i3c/master/Kconfig
index 26c6b58..acc00d9 100644
--- a/drivers/i3c/master/Kconfig
+++ b/drivers/i3c/master/Kconfig
@@ -20,3 +20,13 @@ config DW_I3C_MASTER

This driver can also be built as a module. If so, the module
will be called dw-i3c-master.
+
+config MTK_I3C_MASTER
+ tristate "MediaTek I3C master driver"
+ depends on I3C
+ depends on HAS_IOMEM
+ depends on !(ALPHA || PARISC)
+ help
+ This selects the MediaTek(R) I3C master controller driver.
+ If you want to use MediaTek(R) I3C interface, say Y here.
+ If unsure, say N or M.
diff --git a/drivers/i3c/master/Makefile b/drivers/i3c/master/Makefile
index fc53939..fe7ccf5 100644
--- a/drivers/i3c/master/Makefile
+++ b/drivers/i3c/master/Makefile
@@ -1,2 +1,3 @@
obj-$(CONFIG_CDNS_I3C_MASTER) += i3c-master-cdns.o
obj-$(CONFIG_DW_I3C_MASTER) += dw-i3c-master.o
+obj-$(CONFIG_MTK_I3C_MASTER) += i3c-master-mtk.o
diff --git a/drivers/i3c/master/i3c-master-mtk.c b/drivers/i3c/master/i3c-master-mtk.c
new file mode 100644
index 0000000..3981149
--- /dev/null
+++ b/drivers/i3c/master/i3c-master-mtk.c
@@ -0,0 +1,1239 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 MediaTek Design Systems Inc.
+ *
+ * Author: Qii Wang <qii.wang@xxxxxxxxxxxx>
+ */
+
+#include <linux/clk.h>
+#include <linux/dma-mapping.h>
+#include <linux/i3c/master.h>
+#include <linux/interrupt.h>
+#include <linux/iopoll.h>
+#include <linux/platform_device.h>
+#include <linux/slab.h>
+
+#define DRV_NAME "i3c-master-mtk"
+
+#define SLAVE_ADDR 0x04
+#define INTR_MASK 0x08
+#define INTR_STAT 0x0c
+#define INTR_TRANSAC_COMP BIT(0)
+#define INTR_ACKERR GENMASK(2, 1)
+#define INTR_ARB_LOST BIT(3)
+#define INTR_RS_MULTI BIT(4)
+#define INTR_MAS_ERR BIT(8)
+#define INTR_ALL (INTR_MAS_ERR | INTR_ARB_LOST |\
+ INTR_ACKERR | INTR_TRANSAC_COMP)
+
+#define CONTROL 0x10
+#define CONTROL_WRAPPER BIT(0)
+#define CONTROL_RS BIT(1)
+#define CONTROL_DMA_EN BIT(2)
+#define CONTROL_CLK_EXT_EN BIT(3)
+#define CONTROL_DIR_CHANGE BIT(4)
+#define CONTROL_ACKERR_DET_EN BIT(5)
+#define CONTROL_LEN_CHANGE BIT(6)
+#define CONTROL_DMAACK_EN BIT(8)
+#define CONTROL_ASYNC_MODE BIT(9)
+
+#define TRANSFER_LEN 0x14
+#define TRANSAC_LEN 0x18
+#define TRANSAC_LEN_WRRD 0x0002
+#define TRANS_ONE_LEN 0x0001
+
+#define DELAY_LEN 0x1c
+#define DELAY_LEN_DEFAULT 0x000a
+
+#define TIMING 0x20
+#define TIMING_VALUE(sample_cnt, step_cnt) ({ \
+ typeof(sample_cnt) sample_cnt_ = (sample_cnt); \
+ typeof(step_cnt) step_cnt_ = (step_cnt); \
+ (((sample_cnt_) << 8) | (step_cnt_)); \
+})
+
+#define START 0x24
+#define START_EN BIT(0)
+#define START_MUL_TRIG BIT(14)
+#define START_MUL_CNFG BIT(15)
+
+#define EXT_CONF 0x28
+#define EXT_CONF_DEFAULT 0x0a1f
+
+#define LTIMING 0x2c
+#define LTIMING_VALUE(sample_cnt, step_cnt) ({ \
+ typeof(sample_cnt) sample_cnt_ = (sample_cnt); \
+ typeof(step_cnt) step_cnt_ = (step_cnt); \
+ (((sample_cnt_) << 6) | (step_cnt_) | \
+ ((sample_cnt_) << 12) | ((step_cnt_) << 9)); \
+})
+
+#define HS 0x30
+#define HS_CLR_VALUE 0x0000
+#define HS_DEFAULT_VALUE 0x0083
+#define HS_VALUE(sample_cnt, step_cnt) ({ \
+ typeof(sample_cnt) sample_cnt_ = (sample_cnt); \
+ typeof(step_cnt) step_cnt_ = (step_cnt); \
+ (HS_DEFAULT_VALUE | \
+ ((sample_cnt_) << 12) | ((step_cnt_) << 8)); \
+})
+
+#define IO_CONFIG 0x34
+#define IO_CONFIG_PUSH_PULL 0x0000
+
+#define FIFO_ADDR_CLR 0x38
+#define FIFO_CLR 0x0003
+
+#define MCU_INTR 0x40
+#define MCU_INTR_EN BIT(0)
+
+#define TRANSFER_LEN_AUX 0x44
+#define CLOCK_DIV 0x48
+#define CLOCK_DIV_DEFAULT ((INTER_CLK_DIV - 1) << 8 |\
+ (INTER_CLK_DIV - 1))
+
+#define SOFTRESET 0x50
+#define SOFT_RST BIT(0)
+
+#define TRAFFIC 0x54
+#define TRAFFIC_DAA_EN BIT(4)
+#define TRAFFIC_TBIT BIT(7)
+#define TRAFFIC_HEAD_ONLY BIT(9)
+#define TRAFFIC_SKIP_SLV_ADDR BIT(10)
+#define TRAFFIC_HANDOFF BIT(14)
+
+#define DEF_DA 0x68
+#define DEF_DAA_SLV_PARITY BIT(8)
+
+#define SHAPE 0x6c
+#define SHAPE_T_STALL BIT(1)
+#define SHAPE_T_PARITY BIT(2)
+
+#define HFIFO_DATA 0x70
+#define NINTH_BIT_IGNORE 0
+#define NINTH_BIT_ACK 1
+#define NINTH_BIT_NACK 2
+#define NINTH_BIT_ODD_PAR 3
+#define INST_WITH_HS BIT(10)
+#define UNLOCK_HFIFO BIT(15)
+#define HFIFO_DATA_08 0x8208
+#define HFIFO_DATA_7E (UNLOCK_HFIFO |\
+ (I3C_BROADCAST_ADDR << 1) |\
+ (NINTH_BIT_ACK << 8))
+#define HFIFO_HEAD (UNLOCK_HFIFO | INST_WITH_HS |\
+ (NINTH_BIT_ODD_PAR << 8))
+
+#define FIFO_STAT 0xf4
+#define FIFO_DAA 0x0080
+
+#define DMA_INT_FLAG 0x0
+#define DMA_EN 0x08
+#define DMA_RST 0x0c
+#define DMA_CON 0x18
+#define DMA_TX_MEM_ADDR 0x1c
+#define DMA_RX_MEM_ADDR 0x20
+#define DMA_TX_LEN 0x24
+#define DMA_RX_LEN 0x28
+#define DMA_TX_4G_MODE 0x54
+#define DMA_RX_4G_MODE 0x58
+
+#define DMA_EN_START BIT(0)
+#define DMA_RST_HARD BIT(1)
+#define DMA_4G_MODE BIT(0)
+#define DMA_CLR_FLAG 0x0000
+#define DMA_CON_TX 0x0000
+#define DMA_CON_RX 0x0001
+
+#define INTER_CLK_DIV 5
+#define MAX_SAMPLE_CNT 8
+#define MAX_STEP_CNT 64
+#define MAX_HS_STEP_CNT 8
+#define MAX_I3C_DEVS 16
+
+#define MTK_I3C_BCR 0x01
+#define MTK_I3C_DCR 0x02
+#define MTK_I3C_PID 0x03040506
+
+enum mtk_trans_op {
+ MASTER_WR = 1,
+ MASTER_RD,
+ MASTER_WRRD,
+ /* I3C private op */
+ MASTER_CCC_BROADCAST,
+ MASTER_DAA,
+};
+
+enum mtk_trans_mode {
+ I2C_TRANSFER = 1,
+ I3C_TRANSFER,
+ I3C_CCC,
+};
+
+struct daa_addr_ary {
+ u8 addr;
+ bool used;
+};
+
+struct daa_anchor {
+ struct daa_addr_ary daa_addr[MAX_I3C_DEVS];
+ int idx;
+};
+
+struct mtk_i3c_cmd {
+ enum mtk_trans_op op;
+ u8 ccc_id;
+ u16 addr; /* device addr */
+
+ u16 tx_len;
+ const void *tx_buf;
+ u16 rx_len;
+ void *rx_buf;
+};
+
+struct mtk_i3c_xfer {
+ struct list_head node;
+ struct completion complete; /* i3c transfer stop */
+
+ int ret;
+ u16 irq_stat; /* interrupt status */
+ enum mtk_trans_mode mode;
+ bool ignore_restart_irq;
+ bool auto_restart;
+ unsigned int ncmds;
+ unsigned int trans_num;
+ u8 *msg_buf_w;
+ u8 *msg_buf_r;
+ dma_addr_t rpaddr;
+ dma_addr_t wpaddr;
+ struct mtk_i3c_cmd cmds[0];
+};
+
+struct mtk_i3c_master {
+ struct device *dev;
+ struct i3c_master_controller mas_ctrler;
+
+ /* set in i3c probe */
+ void __iomem *regs; /* i3c base addr */
+ void __iomem *dma_regs; /* apdma base address*/
+ struct clk *clk_main; /* main clock for i3c bus */
+ struct clk *clk_dma; /* DMA clock for i3c via DMA */
+ struct clk *clk_arb; /* Arbitrator clock for i3c */
+ struct daa_anchor daa_anchor;
+
+ u16 timing_reg[2];
+ u16 ltiming_reg[2];
+ u16 high_speed_reg[2];
+
+ struct {
+ struct list_head list;
+ struct mtk_i3c_xfer *cur;
+ spinlock_t lock; /* Lock for stats update */
+ } xferqueue;
+};
+
+static u16 mtk_i3c_readw(struct mtk_i3c_master *master, u16 offset)
+{
+ return readw(master->regs + offset);
+}
+
+static void mtk_i3c_writew(struct mtk_i3c_master *master, u16 val, u16 offset)
+{
+ writew(val, master->regs + offset);
+}
+
+/**
+ * Calculate i3c port speed
+ *
+ * Hardware design:
+ * i3c_bus_freq = parent_clk / (clock_div * 2 * sample_cnt * step_cnt)
+ * clock_div: fixed in hardware, but may be various in different SoCs
+ *
+ * The calculation want to pick the highest bus frequency that is still
+ * less than or equal to master->speed_hz. The calculation try to get
+ * sample_cnt and step_cn
+ */
+static int mtk_i3c_calculate_speed(struct mtk_i3c_master *master,
+ unsigned int clk_src,
+ unsigned int target_speed,
+ unsigned int *timing_step_cnt,
+ unsigned int *timing_sample_cnt)
+{
+ unsigned int sample_cnt, step_cnt, max_step_cnt, opt_div;
+ unsigned int best_mul, cnt_mul, base_step_cnt;
+ unsigned int base_sample_cnt = MAX_SAMPLE_CNT;
+
+ if (target_speed > I3C_BUS_I2C_FM_PLUS_SCL_RATE)
+ max_step_cnt = MAX_HS_STEP_CNT;
+ else
+ max_step_cnt = MAX_STEP_CNT;
+
+ base_step_cnt = max_step_cnt;
+ /* Find the best combination */
+ opt_div = DIV_ROUND_UP(clk_src >> 1, target_speed);
+ best_mul = MAX_SAMPLE_CNT * max_step_cnt;
+
+ /* Search for the best pair (sample_cnt, step_cnt) with
+ * 1 < sample_cnt < MAX_SAMPLE_CNT
+ * 1 < step_cnt < max_step_cnt
+ * sample_cnt * step_cnt >= opt_div
+ * optimizing for sample_cnt * step_cnt being minimal
+ */
+ for (sample_cnt = 2; sample_cnt <= MAX_SAMPLE_CNT; sample_cnt++) {
+ step_cnt = DIV_ROUND_UP(opt_div, sample_cnt);
+ cnt_mul = step_cnt * sample_cnt;
+ if (step_cnt > max_step_cnt || step_cnt < 2)
+ continue;
+
+ if (cnt_mul < best_mul) {
+ best_mul = cnt_mul;
+ base_sample_cnt = sample_cnt;
+ base_step_cnt = step_cnt;
+ if (best_mul == opt_div)
+ break;
+ }
+ }
+
+ sample_cnt = base_sample_cnt;
+ step_cnt = base_step_cnt;
+
+ if ((clk_src / (2 * sample_cnt * step_cnt)) > target_speed) {
+ dev_err(master->dev, "Unsupport speed (%uhz)\n", target_speed);
+ return -EINVAL;
+ }
+
+ *timing_step_cnt = step_cnt - 1;
+ *timing_sample_cnt = sample_cnt - 1;
+
+ return 0;
+}
+
+static int mtk_i3c_set_speed(struct mtk_i3c_master *master,
+ unsigned int clk_src)
+{
+ struct i3c_bus *bus = i3c_master_get_bus(&master->mas_ctrler);
+ unsigned int step_cnt, sample_cnt;
+ int ret;
+
+ ret = mtk_i3c_calculate_speed(master, clk_src, bus->scl_rate.i2c,
+ &step_cnt, &sample_cnt);
+ if (ret < 0)
+ return ret;
+
+ master->timing_reg[0] = TIMING_VALUE(sample_cnt, step_cnt);
+ /* Disable the high speed transaction */
+ master->high_speed_reg[0] = HS_CLR_VALUE;
+ master->ltiming_reg[0] = LTIMING_VALUE(sample_cnt, step_cnt);
+
+ ret = mtk_i3c_calculate_speed(master, clk_src, bus->scl_rate.i3c,
+ &step_cnt, &sample_cnt);
+ if (ret < 0)
+ return ret;
+
+ master->timing_reg[1] = TIMING_VALUE(sample_cnt, step_cnt);
+ master->high_speed_reg[1] = HS_VALUE(sample_cnt, step_cnt);
+ master->ltiming_reg[1] = LTIMING_VALUE(sample_cnt, step_cnt);
+
+ return 0;
+}
+
+static inline u32 mtk_i3c_set_4g_mode(dma_addr_t addr)
+{
+ return (addr & BIT_ULL(32)) ? DMA_4G_MODE : DMA_CLR_FLAG;
+}
+
+static inline struct mtk_i3c_master *
+to_mtk_i3c_master(struct i3c_master_controller *master)
+{
+ return container_of(master, struct mtk_i3c_master, mas_ctrler);
+}
+
+static u8 mtk_i3c_master_get_free_addr(struct mtk_i3c_master *master)
+{
+ int slot = 0;
+
+ for (slot = 0; slot < MAX_I3C_DEVS; slot++) {
+ if (master->daa_anchor.daa_addr[slot].used)
+ continue;
+
+ master->daa_anchor.idx = slot;
+ master->daa_anchor.daa_addr[slot].used = true;
+ return master->daa_anchor.daa_addr[slot].addr;
+ }
+
+ return 0;
+}
+
+static int mtk_i3c_master_clock_enable(struct mtk_i3c_master *master)
+{
+ int ret;
+
+ ret = clk_prepare_enable(master->clk_dma);
+ if (ret)
+ return ret;
+
+ ret = clk_prepare_enable(master->clk_main);
+ if (ret)
+ goto err_main;
+
+ ret = clk_prepare_enable(master->clk_arb);
+ if (ret)
+ goto err_arb;
+
+ return 0;
+
+err_arb:
+ clk_disable_unprepare(master->clk_main);
+err_main:
+ clk_disable_unprepare(master->clk_dma);
+
+ return ret;
+}
+
+static void mtk_i3c_master_clock_disable(struct mtk_i3c_master *master)
+{
+ clk_disable_unprepare(master->clk_arb);
+ clk_disable_unprepare(master->clk_main);
+ clk_disable_unprepare(master->clk_dma);
+}
+
+static inline void mtk_i3c_master_init_hw(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *mtk_xfer = master->xferqueue.cur;
+
+ if (!mtk_xfer->trans_num)
+ mtk_i3c_writew(master, SOFT_RST, SOFTRESET);
+
+ mtk_i3c_writew(master, IO_CONFIG_PUSH_PULL, IO_CONFIG);
+ mtk_i3c_writew(master, CLOCK_DIV_DEFAULT, CLOCK_DIV);
+
+ mtk_i3c_writew(master, DELAY_LEN_DEFAULT, DELAY_LEN);
+ mtk_i3c_writew(master, EXT_CONF_DEFAULT, EXT_CONF);
+ mtk_i3c_writew(master, FIFO_CLR, FIFO_ADDR_CLR);
+
+ /* DMA hard reset */
+ writel(DMA_RST_HARD, master->dma_regs + DMA_RST);
+ writel(DMA_CLR_FLAG, master->dma_regs + DMA_RST);
+}
+
+static int mtk_i3c_master_apdma_tx(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ u32 reg_4g_mode;
+
+ xfer->msg_buf_w = kzalloc(cmd->tx_len, GFP_KERNEL);
+ if (!xfer->msg_buf_w)
+ return -ENOMEM;
+
+ memcpy(xfer->msg_buf_w, cmd->tx_buf, cmd->tx_len);
+
+ writel(DMA_CLR_FLAG, master->dma_regs + DMA_INT_FLAG);
+ writel(DMA_CON_TX, master->dma_regs + DMA_CON);
+
+ xfer->wpaddr = dma_map_single(master->dev, xfer->msg_buf_w,
+ cmd->tx_len, DMA_TO_DEVICE);
+ if (dma_mapping_error(master->dev, xfer->wpaddr)) {
+ kfree(xfer->msg_buf_w);
+ return -ENOMEM;
+ }
+
+ reg_4g_mode = mtk_i3c_set_4g_mode(xfer->wpaddr);
+ writel(reg_4g_mode, master->dma_regs + DMA_TX_4G_MODE);
+ writel((u32)xfer->wpaddr, master->dma_regs + DMA_TX_MEM_ADDR);
+ writel(cmd->tx_len, master->dma_regs + DMA_TX_LEN);
+
+ return 0;
+}
+
+static int mtk_i3c_master_apdma_rx(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ u32 reg_4g_mode;
+
+ xfer->msg_buf_r = kzalloc(cmd->rx_len, GFP_KERNEL);
+ if (!xfer->msg_buf_r)
+ return -ENOMEM;
+
+ writel(DMA_CLR_FLAG, master->dma_regs + DMA_INT_FLAG);
+ writel(DMA_CON_RX, master->dma_regs + DMA_CON);
+
+ xfer->rpaddr = dma_map_single(master->dev, xfer->msg_buf_r,
+ cmd->rx_len, DMA_FROM_DEVICE);
+ if (dma_mapping_error(master->dev, xfer->rpaddr)) {
+ kfree(xfer->msg_buf_r);
+ return -ENOMEM;
+ }
+
+ reg_4g_mode = mtk_i3c_set_4g_mode(xfer->rpaddr);
+ writel(reg_4g_mode, master->dma_regs + DMA_RX_4G_MODE);
+ writel((u32)xfer->rpaddr, master->dma_regs + DMA_RX_MEM_ADDR);
+ writel(cmd->rx_len, master->dma_regs + DMA_RX_LEN);
+
+ return 0;
+}
+
+static void mtk_i3c_master_apdma_end(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ u16 int_reg = 0;
+
+ if (xfer->auto_restart)
+ int_reg = INTR_RS_MULTI;
+
+ int_reg |= INTR_ALL;
+
+ /* Clear interrupt mask */
+ mtk_i3c_writew(master, ~int_reg, INTR_MASK);
+
+ if (cmd->op == MASTER_WR) {
+ dma_unmap_single(master->dev, xfer->wpaddr,
+ cmd->tx_len, DMA_TO_DEVICE);
+ kfree(xfer->msg_buf_w);
+
+ } else if (cmd->op == MASTER_RD) {
+ dma_unmap_single(master->dev, xfer->rpaddr,
+ cmd->rx_len, DMA_FROM_DEVICE);
+ memcpy(cmd->rx_buf, xfer->msg_buf_r, cmd->rx_len);
+ kfree(xfer->msg_buf_r);
+ } else if (cmd->op == MASTER_WRRD) {
+ dma_unmap_single(master->dev, xfer->wpaddr, cmd->tx_len,
+ DMA_TO_DEVICE);
+ dma_unmap_single(master->dev, xfer->rpaddr, cmd->rx_len,
+ DMA_FROM_DEVICE);
+ memcpy(cmd->rx_buf, xfer->msg_buf_r, cmd->rx_len);
+ kfree(xfer->msg_buf_w);
+ kfree(xfer->msg_buf_r);
+ }
+}
+
+static int mtk_i3c_master_start_transfer(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ u16 addr_reg, control_reg, start_reg, tmp_rg_val;
+ u16 int_reg = 0, data_size;
+ int ret = 0;
+ u8 *ptr;
+
+ xfer->irq_stat = 0;
+
+ mtk_i3c_master_init_hw(master);
+
+ control_reg = CONTROL_ACKERR_DET_EN | CONTROL_CLK_EXT_EN | CONTROL_RS;
+ if (cmd->op != MASTER_DAA)
+ control_reg |= CONTROL_DMA_EN | CONTROL_DMAACK_EN |
+ CONTROL_ASYNC_MODE;
+
+ if (cmd->op == MASTER_WRRD)
+ control_reg |= CONTROL_DIR_CHANGE;
+
+ mtk_i3c_writew(master, control_reg, CONTROL);
+
+ addr_reg = cmd->addr << 1;
+ if (cmd->op == MASTER_RD || cmd->op == MASTER_DAA)
+ addr_reg |= 0x1;
+
+ mtk_i3c_writew(master, addr_reg, SLAVE_ADDR);
+
+ if (xfer->auto_restart)
+ int_reg = INTR_RS_MULTI;
+
+ int_reg |= INTR_ALL;
+ /* Clear interrupt status */
+ mtk_i3c_writew(master, int_reg, INTR_STAT);
+ /* Enable interrupt */
+ mtk_i3c_writew(master, int_reg, INTR_MASK);
+
+ /* Set transfer and transaction len */
+ switch (cmd->op) {
+ case MASTER_WRRD:
+ mtk_i3c_writew(master, cmd->tx_len, TRANSFER_LEN);
+ mtk_i3c_writew(master, cmd->rx_len, TRANSFER_LEN_AUX);
+ mtk_i3c_writew(master, TRANSAC_LEN_WRRD, TRANSAC_LEN);
+ break;
+ case MASTER_WR:
+ mtk_i3c_writew(master, cmd->tx_len, TRANSFER_LEN);
+ mtk_i3c_writew(master, xfer->ncmds, TRANSAC_LEN);
+ break;
+ case MASTER_RD:
+ mtk_i3c_writew(master, cmd->rx_len, TRANSFER_LEN);
+ mtk_i3c_writew(master, xfer->ncmds, TRANSAC_LEN);
+ break;
+ case MASTER_CCC_BROADCAST:
+ mtk_i3c_writew(master, TRANS_ONE_LEN, TRANSFER_LEN);
+ mtk_i3c_writew(master, TRANS_ONE_LEN, TRANSAC_LEN);
+ break;
+ case MASTER_DAA:
+ mtk_i3c_writew(master, cmd->rx_len + cmd->tx_len, TRANSFER_LEN);
+ mtk_i3c_writew(master, MAX_I3C_DEVS, TRANSAC_LEN);
+ break;
+ }
+
+ mtk_i3c_writew(master, master->timing_reg[1], TIMING);
+ mtk_i3c_writew(master, master->ltiming_reg[1], LTIMING);
+ mtk_i3c_writew(master, master->high_speed_reg[1], HS);
+
+ if (xfer->mode == I2C_TRANSFER) {
+ mtk_i3c_writew(master, master->timing_reg[0], TIMING);
+ mtk_i3c_writew(master, master->ltiming_reg[0], LTIMING);
+ mtk_i3c_writew(master, master->high_speed_reg[0], HS);
+ mtk_i3c_writew(master, 0, TRAFFIC);
+ mtk_i3c_writew(master, 0, SHAPE);
+ } else if ((xfer->mode == I3C_TRANSFER) && (!xfer->trans_num)) {
+ mtk_i3c_writew(master, HFIFO_DATA_7E, HFIFO_DATA);
+ mtk_i3c_writew(master, TRAFFIC_HANDOFF | TRAFFIC_TBIT, TRAFFIC);
+ mtk_i3c_writew(master, SHAPE_T_STALL | SHAPE_T_PARITY, SHAPE);
+ } else if (xfer->mode == I3C_CCC) {
+ tmp_rg_val = TRAFFIC_HANDOFF | TRAFFIC_TBIT;
+ mtk_i3c_writew(master, HFIFO_DATA_7E, HFIFO_DATA);
+ mtk_i3c_writew(master, HFIFO_HEAD | cmd->ccc_id, HFIFO_DATA);
+
+ if (cmd->op == MASTER_DAA) {
+ addr_reg = mtk_i3c_master_get_free_addr(master);
+ addr_reg |= DEF_DAA_SLV_PARITY;
+ tmp_rg_val |= TRAFFIC_DAA_EN;
+ mtk_i3c_writew(master, addr_reg, DEF_DA);
+ } else if (cmd->op == MASTER_CCC_BROADCAST) {
+ tmp_rg_val |= TRAFFIC_HEAD_ONLY | TRAFFIC_SKIP_SLV_ADDR;
+ data_size = cmd->tx_len;
+ ptr = (u8 *)cmd->tx_buf;
+ while (data_size--)
+ mtk_i3c_writew(master, HFIFO_HEAD | *ptr++,
+ HFIFO_DATA);
+ }
+
+ mtk_i3c_writew(master, tmp_rg_val, TRAFFIC);
+ mtk_i3c_writew(master, SHAPE_T_STALL, SHAPE);
+ }
+
+ /* Prepare buffer data to start transfer */
+ if (cmd->op == MASTER_RD) {
+ ret = mtk_i3c_master_apdma_rx(master);
+ } else if (cmd->op == MASTER_WR) {
+ ret = mtk_i3c_master_apdma_tx(master);
+ } else if (cmd->op == MASTER_WRRD) {
+ ret = mtk_i3c_master_apdma_rx(master);
+ ret |= mtk_i3c_master_apdma_tx(master);
+ }
+
+ if (ret)
+ return ret;
+
+ if (cmd->op != MASTER_DAA)
+ writel(DMA_EN_START, master->dma_regs + DMA_EN);
+
+ start_reg = START_EN;
+ if (xfer->auto_restart) {
+ start_reg |= START_MUL_TRIG;
+ if ((xfer->ncmds - xfer->trans_num) >= 2 ||
+ cmd->op == MASTER_DAA)
+ start_reg |= START_MUL_CNFG;
+ }
+
+ mtk_i3c_writew(master, MCU_INTR_EN, MCU_INTR);
+ mtk_i3c_writew(master, start_reg, START);
+
+ return 0;
+}
+
+static struct mtk_i3c_xfer *
+mtk_i3c_master_alloc_xfer(struct mtk_i3c_master *master, unsigned int ncmds)
+{
+ struct mtk_i3c_xfer *xfer;
+
+ xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL);
+ if (!xfer)
+ return NULL;
+
+ INIT_LIST_HEAD(&xfer->node);
+ xfer->ncmds = ncmds;
+ xfer->ret = -ETIMEDOUT;
+
+ return xfer;
+}
+
+static void mtk_i3c_master_free_xfer(struct mtk_i3c_xfer *xfer)
+{
+ kfree(xfer);
+}
+
+static void mtk_i3c_master_start_xfer_locked(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ unsigned int ret = 0;
+
+ if (!xfer)
+ return;
+
+ xfer->trans_num = 0;
+ ret = mtk_i3c_master_start_transfer(master);
+ xfer->ret = ret;
+}
+
+static void mtk_i3c_master_end_xfer_locked(struct mtk_i3c_master *master)
+{
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ int slot = 0;
+
+ if (!xfer)
+ return;
+
+ if (xfer->irq_stat & (INTR_ACKERR | INTR_MAS_ERR)) {
+ if (cmd->op == MASTER_DAA) {
+ slot = master->daa_anchor.idx;
+ master->daa_anchor.daa_addr[slot].used = false;
+ }
+
+ dev_err(master->dev, "Addr: %x, device ACK error\n", cmd->addr);
+ xfer->ret = -ENXIO;
+ }
+
+ mtk_i3c_master_apdma_end(master);
+
+ complete(&xfer->complete);
+ xfer = list_first_entry_or_null(&master->xferqueue.list,
+ struct mtk_i3c_xfer, node);
+ if (xfer)
+ list_del_init(&xfer->node);
+
+ master->xferqueue.cur = xfer;
+ mtk_i3c_master_start_xfer_locked(master);
+}
+
+static irqreturn_t mtk_i3c_master_irq(int irqno, void *dev_id)
+{
+ struct mtk_i3c_master *master = dev_id;
+ struct mtk_i3c_xfer *xfer = master->xferqueue.cur;
+ struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num];
+ u16 restart_flag = 0, intr_stat, fifo_stat;
+ unsigned long flags;
+ int ret = 0;
+ u8 addr;
+
+ if (cmd->op == MASTER_DAA) {
+ fifo_stat = mtk_i3c_readw(master, FIFO_STAT);
+ intr_stat = mtk_i3c_readw(master, INTR_STAT);
+
+ if (!(fifo_stat & FIFO_DAA) && !(intr_stat & INTR_TRANSAC_COMP))
+ return IRQ_NONE;
+
+ mtk_i3c_writew(master, FIFO_CLR, FIFO_ADDR_CLR);
+ }
+
+ if (xfer->auto_restart)
+ restart_flag = INTR_RS_MULTI;
+
+ intr_stat = mtk_i3c_readw(master, INTR_STAT);
+ mtk_i3c_writew(master, intr_stat, INTR_STAT);
+
+ spin_lock_irqsave(&master->xferqueue.lock, flags);
+ /*
+ * when occurs ack error, i3c controller generate two interrupts
+ * first is the ack error interrupt, then the complete interrupt
+ * i3c->irq_stat need keep the two interrupt value.
+ */
+ xfer->irq_stat |= intr_stat;
+
+ if (xfer->irq_stat & INTR_TRANSAC_COMP) {
+ mtk_i3c_master_end_xfer_locked(master);
+ goto exit_irq;
+ }
+
+ if (xfer->irq_stat & restart_flag) {
+ xfer->irq_stat = 0;
+
+ if (xfer->ignore_restart_irq) {
+ xfer->ignore_restart_irq = false;
+
+ mtk_i3c_writew(master, START_MUL_CNFG | START_MUL_TRIG |
+ START_EN, START);
+ } else if (cmd->op == MASTER_DAA) {
+ addr = mtk_i3c_master_get_free_addr(master);
+ mtk_i3c_writew(master, addr, DEF_DA);
+ mtk_i3c_writew(master, START_MUL_CNFG | START_MUL_TRIG |
+ START_EN, START);
+ } else {
+ mtk_i3c_master_apdma_end(master);
+
+ xfer->trans_num++;
+ ret = mtk_i3c_master_start_transfer(master);
+ xfer->ret = ret;
+ }
+ }
+
+exit_irq:
+ spin_unlock_irqrestore(&master->xferqueue.lock, flags);
+ return IRQ_HANDLED;
+}
+
+static void mtk_i3c_master_queue_xfer(struct mtk_i3c_master *master,
+ struct mtk_i3c_xfer *xfer)
+{
+ unsigned long flags;
+
+ init_completion(&xfer->complete);
+ spin_lock_irqsave(&master->xferqueue.lock, flags);
+ if (master->xferqueue.cur) {
+ list_add_tail(&xfer->node, &master->xferqueue.list);
+ } else {
+ master->xferqueue.cur = xfer;
+ mtk_i3c_master_start_xfer_locked(master);
+ }
+ spin_unlock_irqrestore(&master->xferqueue.lock, flags);
+}
+
+static void mtk_i3c_master_unqueue_xfer(struct mtk_i3c_master *master,
+ struct mtk_i3c_xfer *xfer)
+{
+ struct mtk_i3c_cmd *ccmd = &xfer->cmds[xfer->trans_num];
+ unsigned long flags;
+
+ dev_err(master->dev, "addr: %x, transfer timeout\n", ccmd->addr);
+ xfer->ret = -ETIMEDOUT;
+
+ spin_lock_irqsave(&master->xferqueue.lock, flags);
+ mtk_i3c_master_apdma_end(master);
+ if (master->xferqueue.cur == xfer)
+ master->xferqueue.cur = NULL;
+ else
+ list_del_init(&xfer->node);
+ spin_unlock_irqrestore(&master->xferqueue.lock, flags);
+}
+
+static bool mtk_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m,
+ const struct i3c_ccc_cmd *cmd)
+{
+ if (cmd->ndests > 1)
+ return false;
+
+ switch (cmd->id) {
+ case I3C_CCC_ENEC(true):
+ case I3C_CCC_ENEC(false):
+ case I3C_CCC_DISEC(true):
+ case I3C_CCC_DISEC(false):
+ case I3C_CCC_ENTAS(0, true):
+ case I3C_CCC_ENTAS(0, false):
+ case I3C_CCC_RSTDAA(true):
+ case I3C_CCC_RSTDAA(false):
+ /* FALL THROUGH */
+ case I3C_CCC_ENTDAA:
+ case I3C_CCC_SETMWL(true):
+ case I3C_CCC_SETMWL(false):
+ case I3C_CCC_SETMRL(true):
+ case I3C_CCC_SETMRL(false):
+ /* FALL THROUGH */
+ case I3C_CCC_DEFSLVS:
+ case I3C_CCC_ENTHDR(0):
+ case I3C_CCC_SETDASA:
+ case I3C_CCC_SETNEWDA:
+ case I3C_CCC_GETMWL:
+ case I3C_CCC_GETMRL:
+ case I3C_CCC_GETPID:
+ case I3C_CCC_GETBCR:
+ case I3C_CCC_GETDCR:
+ case I3C_CCC_GETSTATUS:
+ case I3C_CCC_GETACCMST:
+ case I3C_CCC_GETMXDS:
+ case I3C_CCC_GETHDRCAP:
+ return true;
+ default:
+ break;
+ }
+
+ return false;
+}
+
+static int mtk_i3c_master_send_ccc_cmd(struct i3c_master_controller *m,
+ struct i3c_ccc_cmd *cmd)
+{
+ struct mtk_i3c_master *master = to_mtk_i3c_master(m);
+ struct mtk_i3c_xfer *mtk_xfer;
+ struct mtk_i3c_cmd *ccmd;
+ int ret = 0;
+
+ mtk_xfer = mtk_i3c_master_alloc_xfer(master, 1);
+ if (!mtk_xfer)
+ return -ENOMEM;
+
+ mtk_xfer->mode = I3C_CCC;
+ mtk_xfer->ignore_restart_irq = false;
+ mtk_xfer->auto_restart = false;
+
+ ccmd = mtk_xfer->cmds;
+ ccmd->addr = cmd->dests[0].addr;
+ ccmd->ccc_id = cmd->id;
+
+ if (cmd->rnw) {
+ ccmd->op = MASTER_RD;
+ ccmd->rx_len = cmd->dests[0].payload.len;
+ ccmd->rx_buf = cmd->dests[0].payload.data;
+ } else {
+ ccmd->op = MASTER_WR;
+ ccmd->tx_len = cmd->dests[0].payload.len;
+ ccmd->tx_buf = cmd->dests[0].payload.data;
+ }
+
+ if (ccmd->ccc_id < I3C_CCC_DIRECT) {
+ ccmd->op = MASTER_CCC_BROADCAST;
+ if (ccmd->ccc_id == I3C_CCC_ENTDAA) {
+ ccmd->op = MASTER_DAA;
+ ccmd->rx_len = 8;
+ ccmd->tx_len = 1;
+ mtk_xfer->ignore_restart_irq = true;
+ mtk_xfer->auto_restart = true;
+ }
+ }
+
+ mtk_i3c_master_queue_xfer(master, mtk_xfer);
+ if (!wait_for_completion_timeout(&mtk_xfer->complete,
+ msecs_to_jiffies(2000)))
+ mtk_i3c_master_unqueue_xfer(master, mtk_xfer);
+
+ ret = mtk_xfer->ret;
+
+ mtk_i3c_master_free_xfer(mtk_xfer);
+ if (ret == -ENXIO)
+ return I3C_ERROR_M2;
+
+ return ret;
+}
+
+static int mtk_i3c_master_priv_xfers(struct i3c_dev_desc *dev,
+ struct i3c_priv_xfer *xfers,
+ int nxfers)
+{
+ struct i3c_master_controller *m = i3c_dev_get_master(dev);
+ struct mtk_i3c_master *master = to_mtk_i3c_master(m);
+ struct mtk_i3c_xfer *mtk_xfer;
+ int i, ret = 0;
+
+ if (!nxfers)
+ return 0;
+
+ mtk_xfer = mtk_i3c_master_alloc_xfer(master, nxfers);
+ if (!mtk_xfer)
+ return -ENOMEM;
+
+ mtk_xfer->mode = I3C_TRANSFER;
+
+ if (nxfers == 2 && !xfers[0].rnw && xfers[1].rnw)
+ mtk_xfer->auto_restart = false;
+ else
+ mtk_xfer->auto_restart = true;
+
+ if (mtk_xfer->auto_restart && nxfers >= 2)
+ mtk_xfer->ignore_restart_irq = true;
+ else
+ mtk_xfer->ignore_restart_irq = false;
+
+ for (i = 0; i < nxfers; i++) {
+ struct mtk_i3c_cmd *ccmd = &mtk_xfer->cmds[i];
+
+ ccmd->addr = dev->info.dyn_addr;
+ ccmd->ccc_id = 0;
+
+ if (!mtk_xfer->auto_restart) {
+ /* combined two messages into one transaction */
+ ccmd->op = MASTER_WRRD;
+ ccmd->tx_len = xfers[i].len;
+ ccmd->rx_len = xfers[i + 1].len;
+ ccmd->tx_buf = xfers->data.out;
+ ccmd->rx_buf = xfers[i + 1].data.in;
+ break;
+ }
+
+ if (xfers[i].rnw) {
+ ccmd->op = MASTER_RD;
+ ccmd->rx_len = xfers->len;
+ ccmd->rx_buf = xfers->data.in;
+ } else {
+ ccmd->op = MASTER_WR;
+ ccmd->tx_len = xfers->len;
+ ccmd->tx_buf = xfers->data.out;
+ }
+ }
+
+ mtk_i3c_master_queue_xfer(master, mtk_xfer);
+ if (!wait_for_completion_timeout(&mtk_xfer->complete,
+ msecs_to_jiffies(2000)))
+ mtk_i3c_master_unqueue_xfer(master, mtk_xfer);
+
+ ret = mtk_xfer->ret;
+
+ mtk_i3c_master_free_xfer(mtk_xfer);
+ return ret;
+}
+
+static int mtk_i3c_master_i2c_xfers(struct i2c_dev_desc *dev,
+ const struct i2c_msg *xfers,
+ int nxfers)
+{
+ struct i3c_master_controller *m = i2c_dev_get_master(dev);
+ struct mtk_i3c_master *master = to_mtk_i3c_master(m);
+ struct mtk_i3c_xfer *mtk_xfer;
+ int i, ret = 0;
+
+ mtk_xfer = mtk_i3c_master_alloc_xfer(master, nxfers);
+ if (!mtk_xfer)
+ return -ENOMEM;
+
+ mtk_xfer->mode = I2C_TRANSFER;
+
+ if (!(xfers[0].flags & I2C_M_RD) && (xfers[1].flags & I2C_M_RD) &&
+ xfers[0].addr == xfers[1].addr && nxfers == 2)
+ mtk_xfer->auto_restart = false;
+ else
+ mtk_xfer->auto_restart = true;
+
+ mtk_xfer->ignore_restart_irq = false;
+
+ for (i = 0; i < nxfers; i++) {
+ struct mtk_i3c_cmd *ccmd = &mtk_xfer->cmds[i];
+
+ ccmd->addr = xfers->addr;
+ ccmd->ccc_id = 0;
+
+ if (!mtk_xfer->auto_restart) {
+ /* combined two messages into one transaction */
+ ccmd->op = MASTER_WRRD;
+ ccmd->tx_len = xfers[i].len;
+ ccmd->rx_len = xfers[i + 1].len;
+ ccmd->tx_buf = xfers[i].buf;
+ ccmd->rx_buf = xfers[i + 1].buf;
+ break;
+ }
+
+ if (xfers[i].flags & I2C_M_RD) {
+ ccmd->op = MASTER_RD;
+ ccmd->rx_len = xfers[i].len;
+ ccmd->rx_buf = xfers[i].buf;
+ } else {
+ ccmd->op = MASTER_WR;
+ ccmd->tx_len = xfers[i].len;
+ ccmd->tx_buf = xfers[i].buf;
+ }
+ }
+
+ mtk_i3c_master_queue_xfer(master, mtk_xfer);
+ if (!wait_for_completion_timeout(&mtk_xfer->complete,
+ msecs_to_jiffies(2000)))
+ mtk_i3c_master_unqueue_xfer(master, mtk_xfer);
+
+ ret = mtk_xfer->ret;
+
+ mtk_i3c_master_free_xfer(mtk_xfer);
+ return ret;
+}
+
+static u32 mtk_i3c_master_i2c_funcs(struct i3c_master_controller *m)
+{
+ return I2C_FUNC_SMBUS_EMUL | I2C_FUNC_I2C;
+}
+
+static int mtk_i3c_master_do_daa(struct i3c_master_controller *m)
+{
+ struct mtk_i3c_master *master = to_mtk_i3c_master(m);
+ int ret = 0, slot;
+ u8 last_addr = 0;
+
+ for (slot = 0; slot < MAX_I3C_DEVS; slot++) {
+ if (master->daa_anchor.daa_addr[slot].used)
+ continue;
+
+ ret = i3c_master_get_free_addr(m, last_addr + 1);
+ if (ret < 0)
+ return -ENOSPC;
+
+ last_addr = ret;
+ master->daa_anchor.daa_addr[slot].addr = last_addr;
+ }
+
+ ret = i3c_master_entdaa_locked(&master->mas_ctrler);
+ if (ret && ret != I3C_ERROR_M2)
+ return ret;
+
+ /*
+ * Clear all retaining registers filled during DAA. We already
+ * have the addressed assigned to them in the addrs array.
+ */
+ for (slot = 1; slot < MAX_I3C_DEVS; slot++) {
+ if (master->daa_anchor.daa_addr[slot].used) {
+ last_addr = master->daa_anchor.daa_addr[slot].addr;
+ i3c_master_add_i3c_dev_locked(m, last_addr);
+ }
+ }
+
+ /*
+ * Clear slots that ended up not being used. Can be caused by I3C
+ * device creation failure or when the I3C device was already known
+ * by the system but with a different address (in this case the device
+ * already has a slot and does not need a new one).
+ */
+ i3c_master_defslvs_locked(&master->mas_ctrler);
+
+ /* Unmask Hot-Join and Mastership request interrupts. */
+ i3c_master_enec_locked(m, I3C_BROADCAST_ADDR, I3C_CCC_EVENT_HJ |
+ I3C_CCC_EVENT_MR);
+
+ return 0;
+}
+
+static int mtk_i3c_master_bus_init(struct i3c_master_controller *m)
+{
+ struct mtk_i3c_master *master = to_mtk_i3c_master(m);
+ struct i3c_device_info info = { };
+ unsigned long source_clk_rate;
+ int ret;
+
+ source_clk_rate = clk_get_rate(master->clk_main) / INTER_CLK_DIV;
+
+ ret = mtk_i3c_set_speed(master, source_clk_rate);
+ if (ret) {
+ dev_err(master->dev, "Failed to set the bus speed.\n");
+ return -EINVAL;
+ }
+
+ /* Get an address for the master. */
+ ret = i3c_master_get_free_addr(m, 0);
+ if (ret < 0)
+ return ret;
+
+ master->daa_anchor.daa_addr[0].addr = (u8)ret;
+ master->daa_anchor.daa_addr[0].used = true;
+ master->daa_anchor.idx = 0;
+
+ /*
+ * In I3C protocol host controller is also with device role,
+ * so the driver should provide dcr, bcr, and pid info
+ * of host controller itself
+ */
+ memset(&info, 0, sizeof(info));
+ info.dyn_addr = ret;
+ info.dcr = MTK_I3C_DCR;
+ info.bcr = MTK_I3C_BCR;
+ info.pid = MTK_I3C_PID;
+
+ if (info.bcr & I3C_BCR_HDR_CAP)
+ info.hdr_cap = I3C_CCC_HDR_MODE(I3C_HDR_DDR);
+
+ return i3c_master_set_info(&master->mas_ctrler, &info);
+}
+
+static const struct i3c_master_controller_ops mtk_i3c_master_ops = {
+ .bus_init = mtk_i3c_master_bus_init,
+ .do_daa = mtk_i3c_master_do_daa,
+ .supports_ccc_cmd = mtk_i3c_master_supports_ccc_cmd,
+ .send_ccc_cmd = mtk_i3c_master_send_ccc_cmd,
+ .priv_xfers = mtk_i3c_master_priv_xfers,
+ .i2c_xfers = mtk_i3c_master_i2c_xfers,
+ .i2c_funcs = mtk_i3c_master_i2c_funcs,
+};
+
+static int mtk_i3c_master_probe(struct platform_device *pdev)
+{
+ struct device *dev = &pdev->dev;
+ struct mtk_i3c_master *master;
+ struct resource *res;
+ int ret, irqnr;
+
+ master = devm_kzalloc(dev, sizeof(*master), GFP_KERNEL);
+ if (!master)
+ return -ENOMEM;
+
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "main");
+ master->regs = devm_ioremap_resource(dev, res);
+ if (IS_ERR(master->regs))
+ return PTR_ERR(master->regs);
+
+ res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dma");
+ master->dma_regs = devm_ioremap_resource(dev, res);
+ if (IS_ERR(master->dma_regs))
+ return PTR_ERR(master->dma_regs);
+
+ irqnr = platform_get_irq(pdev, 0);
+ if (irqnr < 0)
+ return irqnr;
+
+ ret = devm_request_irq(dev, irqnr, mtk_i3c_master_irq,
+ IRQF_TRIGGER_NONE, DRV_NAME, master);
+ if (ret < 0) {
+ dev_err(dev, "Request I3C IRQ %d fail\n", irqnr);
+ return ret;
+ }
+
+ spin_lock_init(&master->xferqueue.lock);
+ INIT_LIST_HEAD(&master->xferqueue.list);
+
+ if (dma_set_mask(dev, DMA_BIT_MASK(33))) {
+ dev_err(dev, "dma_set_mask return error.\n");
+ return -EINVAL;
+ }
+
+ master->clk_main = devm_clk_get(dev, "main");
+ if (IS_ERR(master->clk_main)) {
+ dev_err(dev, "cannot get main clock\n");
+ return PTR_ERR(master->clk_main);
+ }
+ master->clk_dma = devm_clk_get(dev, "dma");
+ if (IS_ERR(master->clk_dma)) {
+ dev_err(dev, "cannot get dma clock\n");
+ return PTR_ERR(master->clk_dma);
+ }
+
+ master->clk_arb = devm_clk_get_optional(dev, "arb");
+ if (IS_ERR(master->clk_arb))
+ return PTR_ERR(master->clk_arb);
+
+ ret = mtk_i3c_master_clock_enable(master);
+ if (ret) {
+ dev_err(dev, "clock enable failed!\n");
+ return ret;
+ }
+
+ master->dev = dev;
+ platform_set_drvdata(pdev, master);
+
+ ret = i3c_master_register(&master->mas_ctrler, dev,
+ &mtk_i3c_master_ops, false);
+ if (ret) {
+ dev_err(dev, "Failed to add i3c bus to i3c core\n");
+ mtk_i3c_master_clock_disable(master);
+ return ret;
+ }
+
+ return 0;
+}
+
+static int mtk_i3c_master_remove(struct platform_device *pdev)
+{
+ struct mtk_i3c_master *master = platform_get_drvdata(pdev);
+
+ i3c_master_unregister(&master->mas_ctrler);
+ mtk_i3c_master_clock_disable(master);
+
+ return 0;
+}
+
+static const struct of_device_id mtk_i3c_master_of_ids[] = {
+ { .compatible = "mediatek,i3c-master" },
+ { /* sentinel */ },
+};
+MODULE_DEVICE_TABLE(of, mtk_i3c_master_of_ids);
+
+static struct platform_driver mtk_i3c_master_driver = {
+ .probe = mtk_i3c_master_probe,
+ .remove = mtk_i3c_master_remove,
+ .driver = {
+ .name = DRV_NAME,
+ .of_match_table = mtk_i3c_master_of_ids,
+ },
+};
+module_platform_driver(mtk_i3c_master_driver);
+
+MODULE_AUTHOR("Qii Wang <qii.wang@xxxxxxxxxxxx>");
+MODULE_DESCRIPTION("MediaTek I3C master driver");
+MODULE_LICENSE("GPL v2");
+MODULE_ALIAS("platform:mtk-i3c-master");
--
1.7.9.5