Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Loic Poulain | 3973 | 96.88% | 1 | 6.25% |
Vladimir Zapolskiy | 64 | 1.56% | 4 | 25.00% |
Harshit Mogalapalli | 17 | 0.41% | 1 | 6.25% |
Matti Lehtimäki | 12 | 0.29% | 1 | 6.25% |
Konrad Dybcio | 7 | 0.17% | 1 | 6.25% |
Bryan O'Donoghue | 7 | 0.17% | 1 | 6.25% |
Md Sadre Alam | 5 | 0.12% | 1 | 6.25% |
Jonathan Marek | 5 | 0.12% | 1 | 6.25% |
Liao Chang | 4 | 0.10% | 1 | 6.25% |
Robert Foss | 2 | 0.05% | 1 | 6.25% |
Uwe Kleine-König | 2 | 0.05% | 1 | 6.25% |
Andy Shevchenko | 2 | 0.05% | 1 | 6.25% |
Rayyan Ansari | 1 | 0.02% | 1 | 6.25% |
Total | 4101 | 16 |
// SPDX-License-Identifier: GPL-2.0 // Copyright (c) 2012-2016, The Linux Foundation. All rights reserved. // Copyright (c) 2017-2022 Linaro Limited. #include <linux/clk.h> #include <linux/completion.h> #include <linux/i2c.h> #include <linux/io.h> #include <linux/interrupt.h> #include <linux/module.h> #include <linux/of.h> #include <linux/platform_device.h> #include <linux/pm_runtime.h> #define CCI_HW_VERSION 0x0 #define CCI_RESET_CMD 0x004 #define CCI_RESET_CMD_MASK 0x0f73f3f7 #define CCI_RESET_CMD_M0_MASK 0x000003f1 #define CCI_RESET_CMD_M1_MASK 0x0003f001 #define CCI_QUEUE_START 0x008 #define CCI_HALT_REQ 0x034 #define CCI_HALT_REQ_I2C_M0_Q0Q1 BIT(0) #define CCI_HALT_REQ_I2C_M1_Q0Q1 BIT(1) #define CCI_I2C_Mm_SCL_CTL(m) (0x100 + 0x100 * (m)) #define CCI_I2C_Mm_SDA_CTL_0(m) (0x104 + 0x100 * (m)) #define CCI_I2C_Mm_SDA_CTL_1(m) (0x108 + 0x100 * (m)) #define CCI_I2C_Mm_SDA_CTL_2(m) (0x10c + 0x100 * (m)) #define CCI_I2C_Mm_MISC_CTL(m) (0x110 + 0x100 * (m)) #define CCI_I2C_Mm_READ_DATA(m) (0x118 + 0x100 * (m)) #define CCI_I2C_Mm_READ_BUF_LEVEL(m) (0x11c + 0x100 * (m)) #define CCI_I2C_Mm_Qn_EXEC_WORD_CNT(m, n) (0x300 + 0x200 * (m) + 0x100 * (n)) #define CCI_I2C_Mm_Qn_CUR_WORD_CNT(m, n) (0x304 + 0x200 * (m) + 0x100 * (n)) #define CCI_I2C_Mm_Qn_CUR_CMD(m, n) (0x308 + 0x200 * (m) + 0x100 * (n)) #define CCI_I2C_Mm_Qn_REPORT_STATUS(m, n) (0x30c + 0x200 * (m) + 0x100 * (n)) #define CCI_I2C_Mm_Qn_LOAD_DATA(m, n) (0x310 + 0x200 * (m) + 0x100 * (n)) #define CCI_IRQ_GLOBAL_CLEAR_CMD 0xc00 #define CCI_IRQ_MASK_0 0xc04 #define CCI_IRQ_MASK_0_I2C_M0_RD_DONE BIT(0) #define CCI_IRQ_MASK_0_I2C_M0_Q0_REPORT BIT(4) #define CCI_IRQ_MASK_0_I2C_M0_Q1_REPORT BIT(8) #define CCI_IRQ_MASK_0_I2C_M1_RD_DONE BIT(12) #define CCI_IRQ_MASK_0_I2C_M1_Q0_REPORT BIT(16) #define CCI_IRQ_MASK_0_I2C_M1_Q1_REPORT BIT(20) #define CCI_IRQ_MASK_0_RST_DONE_ACK BIT(24) #define CCI_IRQ_MASK_0_I2C_M0_Q0Q1_HALT_ACK BIT(25) #define CCI_IRQ_MASK_0_I2C_M1_Q0Q1_HALT_ACK BIT(26) #define CCI_IRQ_MASK_0_I2C_M0_ERROR 0x18000ee6 #define CCI_IRQ_MASK_0_I2C_M1_ERROR 0x60ee6000 #define CCI_IRQ_CLEAR_0 0xc08 #define CCI_IRQ_STATUS_0 0xc0c #define CCI_IRQ_STATUS_0_I2C_M0_RD_DONE BIT(0) #define CCI_IRQ_STATUS_0_I2C_M0_Q0_REPORT BIT(4) #define CCI_IRQ_STATUS_0_I2C_M0_Q1_REPORT BIT(8) #define CCI_IRQ_STATUS_0_I2C_M1_RD_DONE BIT(12) #define CCI_IRQ_STATUS_0_I2C_M1_Q0_REPORT BIT(16) #define CCI_IRQ_STATUS_0_I2C_M1_Q1_REPORT BIT(20) #define CCI_IRQ_STATUS_0_RST_DONE_ACK BIT(24) #define CCI_IRQ_STATUS_0_I2C_M0_Q0Q1_HALT_ACK BIT(25) #define CCI_IRQ_STATUS_0_I2C_M1_Q0Q1_HALT_ACK BIT(26) #define CCI_IRQ_STATUS_0_I2C_M0_Q0_NACK_ERR BIT(27) #define CCI_IRQ_STATUS_0_I2C_M0_Q1_NACK_ERR BIT(28) #define CCI_IRQ_STATUS_0_I2C_M1_Q0_NACK_ERR BIT(29) #define CCI_IRQ_STATUS_0_I2C_M1_Q1_NACK_ERR BIT(30) #define CCI_IRQ_STATUS_0_I2C_M0_ERROR 0x18000ee6 #define CCI_IRQ_STATUS_0_I2C_M1_ERROR 0x60ee6000 #define CCI_TIMEOUT (msecs_to_jiffies(100)) #define NUM_MASTERS 2 #define NUM_QUEUES 2 /* Max number of resources + 1 for a NULL terminator */ #define CCI_RES_MAX 6 #define CCI_I2C_SET_PARAM 1 #define CCI_I2C_REPORT 8 #define CCI_I2C_WRITE 9 #define CCI_I2C_READ 10 #define CCI_I2C_REPORT_IRQ_EN BIT(8) enum { I2C_MODE_STANDARD, I2C_MODE_FAST, I2C_MODE_FAST_PLUS, }; enum cci_i2c_queue_t { QUEUE_0, QUEUE_1 }; struct hw_params { u16 thigh; /* HIGH period of the SCL clock in clock ticks */ u16 tlow; /* LOW period of the SCL clock */ u16 tsu_sto; /* set-up time for STOP condition */ u16 tsu_sta; /* set-up time for a repeated START condition */ u16 thd_dat; /* data hold time */ u16 thd_sta; /* hold time (repeated) START condition */ u16 tbuf; /* bus free time between a STOP and START condition */ u8 scl_stretch_en; u16 trdhld; u16 tsp; /* pulse width of spikes suppressed by the input filter */ }; struct cci; struct cci_master { struct i2c_adapter adap; u16 master; u8 mode; int status; struct completion irq_complete; struct cci *cci; }; struct cci_data { unsigned int num_masters; struct i2c_adapter_quirks quirks; u16 queue_size[NUM_QUEUES]; unsigned long cci_clk_rate; struct hw_params params[3]; }; struct cci { struct device *dev; void __iomem *base; unsigned int irq; const struct cci_data *data; struct clk_bulk_data *clocks; int nclocks; struct cci_master master[NUM_MASTERS]; }; static irqreturn_t cci_isr(int irq, void *dev) { struct cci *cci = dev; u32 val, reset = 0; int ret = IRQ_NONE; val = readl(cci->base + CCI_IRQ_STATUS_0); writel(val, cci->base + CCI_IRQ_CLEAR_0); writel(0x1, cci->base + CCI_IRQ_GLOBAL_CLEAR_CMD); if (val & CCI_IRQ_STATUS_0_RST_DONE_ACK) { complete(&cci->master[0].irq_complete); if (cci->master[1].master) complete(&cci->master[1].irq_complete); ret = IRQ_HANDLED; } if (val & CCI_IRQ_STATUS_0_I2C_M0_RD_DONE || val & CCI_IRQ_STATUS_0_I2C_M0_Q0_REPORT || val & CCI_IRQ_STATUS_0_I2C_M0_Q1_REPORT) { cci->master[0].status = 0; complete(&cci->master[0].irq_complete); ret = IRQ_HANDLED; } if (val & CCI_IRQ_STATUS_0_I2C_M1_RD_DONE || val & CCI_IRQ_STATUS_0_I2C_M1_Q0_REPORT || val & CCI_IRQ_STATUS_0_I2C_M1_Q1_REPORT) { cci->master[1].status = 0; complete(&cci->master[1].irq_complete); ret = IRQ_HANDLED; } if (unlikely(val & CCI_IRQ_STATUS_0_I2C_M0_Q0Q1_HALT_ACK)) { reset = CCI_RESET_CMD_M0_MASK; ret = IRQ_HANDLED; } if (unlikely(val & CCI_IRQ_STATUS_0_I2C_M1_Q0Q1_HALT_ACK)) { reset = CCI_RESET_CMD_M1_MASK; ret = IRQ_HANDLED; } if (unlikely(reset)) writel(reset, cci->base + CCI_RESET_CMD); if (unlikely(val & CCI_IRQ_STATUS_0_I2C_M0_ERROR)) { if (val & CCI_IRQ_STATUS_0_I2C_M0_Q0_NACK_ERR || val & CCI_IRQ_STATUS_0_I2C_M0_Q1_NACK_ERR) cci->master[0].status = -ENXIO; else cci->master[0].status = -EIO; writel(CCI_HALT_REQ_I2C_M0_Q0Q1, cci->base + CCI_HALT_REQ); ret = IRQ_HANDLED; } if (unlikely(val & CCI_IRQ_STATUS_0_I2C_M1_ERROR)) { if (val & CCI_IRQ_STATUS_0_I2C_M1_Q0_NACK_ERR || val & CCI_IRQ_STATUS_0_I2C_M1_Q1_NACK_ERR) cci->master[1].status = -ENXIO; else cci->master[1].status = -EIO; writel(CCI_HALT_REQ_I2C_M1_Q0Q1, cci->base + CCI_HALT_REQ); ret = IRQ_HANDLED; } return ret; } static int cci_halt(struct cci *cci, u8 master_num) { struct cci_master *master; u32 val; if (master_num >= cci->data->num_masters) { dev_err(cci->dev, "Unsupported master idx (%u)\n", master_num); return -EINVAL; } val = BIT(master_num); master = &cci->master[master_num]; reinit_completion(&master->irq_complete); writel(val, cci->base + CCI_HALT_REQ); if (!wait_for_completion_timeout(&master->irq_complete, CCI_TIMEOUT)) { dev_err(cci->dev, "CCI halt timeout\n"); return -ETIMEDOUT; } return 0; } static int cci_reset(struct cci *cci) { /* * we reset the whole controller, here and for implicity use * master[0].xxx for waiting on it. */ reinit_completion(&cci->master[0].irq_complete); writel(CCI_RESET_CMD_MASK, cci->base + CCI_RESET_CMD); if (!wait_for_completion_timeout(&cci->master[0].irq_complete, CCI_TIMEOUT)) { dev_err(cci->dev, "CCI reset timeout\n"); return -ETIMEDOUT; } return 0; } static int cci_init(struct cci *cci) { u32 val = CCI_IRQ_MASK_0_I2C_M0_RD_DONE | CCI_IRQ_MASK_0_I2C_M0_Q0_REPORT | CCI_IRQ_MASK_0_I2C_M0_Q1_REPORT | CCI_IRQ_MASK_0_I2C_M1_RD_DONE | CCI_IRQ_MASK_0_I2C_M1_Q0_REPORT | CCI_IRQ_MASK_0_I2C_M1_Q1_REPORT | CCI_IRQ_MASK_0_RST_DONE_ACK | CCI_IRQ_MASK_0_I2C_M0_Q0Q1_HALT_ACK | CCI_IRQ_MASK_0_I2C_M1_Q0Q1_HALT_ACK | CCI_IRQ_MASK_0_I2C_M0_ERROR | CCI_IRQ_MASK_0_I2C_M1_ERROR; int i; writel(val, cci->base + CCI_IRQ_MASK_0); for (i = 0; i < cci->data->num_masters; i++) { int mode = cci->master[i].mode; const struct hw_params *hw; if (!cci->master[i].cci) continue; hw = &cci->data->params[mode]; val = hw->thigh << 16 | hw->tlow; writel(val, cci->base + CCI_I2C_Mm_SCL_CTL(i)); val = hw->tsu_sto << 16 | hw->tsu_sta; writel(val, cci->base + CCI_I2C_Mm_SDA_CTL_0(i)); val = hw->thd_dat << 16 | hw->thd_sta; writel(val, cci->base + CCI_I2C_Mm_SDA_CTL_1(i)); val = hw->tbuf; writel(val, cci->base + CCI_I2C_Mm_SDA_CTL_2(i)); val = hw->scl_stretch_en << 8 | hw->trdhld << 4 | hw->tsp; writel(val, cci->base + CCI_I2C_Mm_MISC_CTL(i)); } return 0; } static int cci_run_queue(struct cci *cci, u8 master, u8 queue) { u32 val; val = readl(cci->base + CCI_I2C_Mm_Qn_CUR_WORD_CNT(master, queue)); writel(val, cci->base + CCI_I2C_Mm_Qn_EXEC_WORD_CNT(master, queue)); reinit_completion(&cci->master[master].irq_complete); val = BIT(master * 2 + queue); writel(val, cci->base + CCI_QUEUE_START); if (!wait_for_completion_timeout(&cci->master[master].irq_complete, CCI_TIMEOUT)) { dev_err(cci->dev, "master %d queue %d timeout\n", master, queue); cci_reset(cci); cci_init(cci); return -ETIMEDOUT; } return cci->master[master].status; } static int cci_validate_queue(struct cci *cci, u8 master, u8 queue) { u32 val; val = readl(cci->base + CCI_I2C_Mm_Qn_CUR_WORD_CNT(master, queue)); if (val == cci->data->queue_size[queue]) return -EINVAL; if (!val) return 0; val = CCI_I2C_REPORT | CCI_I2C_REPORT_IRQ_EN; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); return cci_run_queue(cci, master, queue); } static int cci_i2c_read(struct cci *cci, u16 master, u16 addr, u8 *buf, u16 len) { u32 val, words_read, words_exp; u8 queue = QUEUE_1; int i, index = 0, ret; bool first = true; /* * Call validate queue to make sure queue is empty before starting. * This is to avoid overflow / underflow of queue. */ ret = cci_validate_queue(cci, master, queue); if (ret < 0) return ret; val = CCI_I2C_SET_PARAM | (addr & 0x7f) << 4; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); val = CCI_I2C_READ | len << 4; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); ret = cci_run_queue(cci, master, queue); if (ret < 0) return ret; words_read = readl(cci->base + CCI_I2C_Mm_READ_BUF_LEVEL(master)); words_exp = len / 4 + 1; if (words_read != words_exp) { dev_err(cci->dev, "words read = %d, words expected = %d\n", words_read, words_exp); return -EIO; } do { val = readl(cci->base + CCI_I2C_Mm_READ_DATA(master)); for (i = 0; i < 4 && index < len; i++) { if (first) { /* The LS byte of this register represents the * first byte read from the slave during a read * access. */ first = false; continue; } buf[index++] = (val >> (i * 8)) & 0xff; } } while (--words_read); return 0; } static int cci_i2c_write(struct cci *cci, u16 master, u16 addr, u8 *buf, u16 len) { u8 queue = QUEUE_0; u8 load[12] = { 0 }; int i = 0, j, ret; u32 val; /* * Call validate queue to make sure queue is empty before starting. * This is to avoid overflow / underflow of queue. */ ret = cci_validate_queue(cci, master, queue); if (ret < 0) return ret; val = CCI_I2C_SET_PARAM | (addr & 0x7f) << 4; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); load[i++] = CCI_I2C_WRITE | len << 4; for (j = 0; j < len; j++) load[i++] = buf[j]; for (j = 0; j < i; j += 4) { val = load[j]; val |= load[j + 1] << 8; val |= load[j + 2] << 16; val |= load[j + 3] << 24; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); } val = CCI_I2C_REPORT | CCI_I2C_REPORT_IRQ_EN; writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue)); return cci_run_queue(cci, master, queue); } static int cci_xfer(struct i2c_adapter *adap, struct i2c_msg msgs[], int num) { struct cci_master *cci_master = i2c_get_adapdata(adap); struct cci *cci = cci_master->cci; int i, ret; ret = pm_runtime_get_sync(cci->dev); if (ret < 0) goto err; for (i = 0; i < num; i++) { if (msgs[i].flags & I2C_M_RD) ret = cci_i2c_read(cci, cci_master->master, msgs[i].addr, msgs[i].buf, msgs[i].len); else ret = cci_i2c_write(cci, cci_master->master, msgs[i].addr, msgs[i].buf, msgs[i].len); if (ret < 0) break; } if (!ret) ret = num; err: pm_runtime_mark_last_busy(cci->dev); pm_runtime_put_autosuspend(cci->dev); return ret; } static u32 cci_func(struct i2c_adapter *adap) { return I2C_FUNC_I2C | I2C_FUNC_SMBUS_EMUL; } static const struct i2c_algorithm cci_algo = { .master_xfer = cci_xfer, .functionality = cci_func, }; static int cci_enable_clocks(struct cci *cci) { return clk_bulk_prepare_enable(cci->nclocks, cci->clocks); } static void cci_disable_clocks(struct cci *cci) { clk_bulk_disable_unprepare(cci->nclocks, cci->clocks); } static int __maybe_unused cci_suspend_runtime(struct device *dev) { struct cci *cci = dev_get_drvdata(dev); cci_disable_clocks(cci); return 0; } static int __maybe_unused cci_resume_runtime(struct device *dev) { struct cci *cci = dev_get_drvdata(dev); int ret; ret = cci_enable_clocks(cci); if (ret) return ret; cci_init(cci); return 0; } static int __maybe_unused cci_suspend(struct device *dev) { if (!pm_runtime_suspended(dev)) return cci_suspend_runtime(dev); return 0; } static int __maybe_unused cci_resume(struct device *dev) { cci_resume_runtime(dev); pm_runtime_mark_last_busy(dev); pm_request_autosuspend(dev); return 0; } static const struct dev_pm_ops qcom_cci_pm = { SET_SYSTEM_SLEEP_PM_OPS(cci_suspend, cci_resume) SET_RUNTIME_PM_OPS(cci_suspend_runtime, cci_resume_runtime, NULL) }; static int cci_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; unsigned long cci_clk_rate = 0; struct device_node *child; struct resource *r; struct cci *cci; int ret, i; u32 val; cci = devm_kzalloc(dev, sizeof(*cci), GFP_KERNEL); if (!cci) return -ENOMEM; cci->dev = dev; platform_set_drvdata(pdev, cci); cci->data = device_get_match_data(dev); if (!cci->data) return -ENOENT; for_each_available_child_of_node(dev->of_node, child) { struct cci_master *master; u32 idx; ret = of_property_read_u32(child, "reg", &idx); if (ret) { dev_err(dev, "%pOF invalid 'reg' property", child); continue; } if (idx >= cci->data->num_masters) { dev_err(dev, "%pOF invalid 'reg' value: %u (max is %u)", child, idx, cci->data->num_masters - 1); continue; } master = &cci->master[idx]; master->adap.quirks = &cci->data->quirks; master->adap.algo = &cci_algo; master->adap.dev.parent = dev; master->adap.dev.of_node = of_node_get(child); master->master = idx; master->cci = cci; i2c_set_adapdata(&master->adap, master); snprintf(master->adap.name, sizeof(master->adap.name), "Qualcomm-CCI"); master->mode = I2C_MODE_STANDARD; ret = of_property_read_u32(child, "clock-frequency", &val); if (!ret) { if (val == I2C_MAX_FAST_MODE_FREQ) master->mode = I2C_MODE_FAST; else if (val == I2C_MAX_FAST_MODE_PLUS_FREQ) master->mode = I2C_MODE_FAST_PLUS; } init_completion(&master->irq_complete); } /* Memory */ cci->base = devm_platform_get_and_ioremap_resource(pdev, 0, &r); if (IS_ERR(cci->base)) return PTR_ERR(cci->base); /* Clocks */ ret = devm_clk_bulk_get_all(dev, &cci->clocks); if (ret < 0) return dev_err_probe(dev, ret, "failed to get clocks\n"); else if (!ret) return dev_err_probe(dev, -EINVAL, "not enough clocks in DT\n"); cci->nclocks = ret; /* Retrieve CCI clock rate */ for (i = 0; i < cci->nclocks; i++) { if (!strcmp(cci->clocks[i].id, "cci")) { cci_clk_rate = clk_get_rate(cci->clocks[i].clk); break; } } if (cci_clk_rate != cci->data->cci_clk_rate) { /* cci clock set by the bootloader or via assigned clock rate * in DT. */ dev_warn(dev, "Found %lu cci clk rate while %lu was expected\n", cci_clk_rate, cci->data->cci_clk_rate); } ret = cci_enable_clocks(cci); if (ret < 0) return ret; /* Interrupt */ ret = platform_get_irq(pdev, 0); if (ret < 0) goto disable_clocks; cci->irq = ret; ret = devm_request_irq(dev, cci->irq, cci_isr, 0, dev_name(dev), cci); if (ret < 0) { dev_err(dev, "request_irq failed, ret: %d\n", ret); goto disable_clocks; } val = readl(cci->base + CCI_HW_VERSION); dev_dbg(dev, "CCI HW version = 0x%08x", val); ret = cci_reset(cci); if (ret < 0) goto error; ret = cci_init(cci); if (ret < 0) goto error; pm_runtime_set_autosuspend_delay(dev, MSEC_PER_SEC); pm_runtime_use_autosuspend(dev); pm_runtime_set_active(dev); pm_runtime_enable(dev); for (i = 0; i < cci->data->num_masters; i++) { if (!cci->master[i].cci) continue; ret = i2c_add_adapter(&cci->master[i].adap); if (ret < 0) { of_node_put(cci->master[i].adap.dev.of_node); goto error_i2c; } } return 0; error_i2c: pm_runtime_disable(dev); pm_runtime_dont_use_autosuspend(dev); for (--i ; i >= 0; i--) { if (cci->master[i].cci) { i2c_del_adapter(&cci->master[i].adap); of_node_put(cci->master[i].adap.dev.of_node); } } error: disable_irq(cci->irq); disable_clocks: cci_disable_clocks(cci); return ret; } static void cci_remove(struct platform_device *pdev) { struct cci *cci = platform_get_drvdata(pdev); int i; for (i = 0; i < cci->data->num_masters; i++) { if (cci->master[i].cci) { i2c_del_adapter(&cci->master[i].adap); of_node_put(cci->master[i].adap.dev.of_node); } cci_halt(cci, i); } disable_irq(cci->irq); pm_runtime_disable(&pdev->dev); pm_runtime_set_suspended(&pdev->dev); } static const struct cci_data cci_v1_data = { .num_masters = 1, .queue_size = { 64, 16 }, .quirks = { .max_write_len = 10, .max_read_len = 12, }, .cci_clk_rate = 19200000, .params[I2C_MODE_STANDARD] = { .thigh = 78, .tlow = 114, .tsu_sto = 28, .tsu_sta = 28, .thd_dat = 10, .thd_sta = 77, .tbuf = 118, .scl_stretch_en = 0, .trdhld = 6, .tsp = 1 }, .params[I2C_MODE_FAST] = { .thigh = 20, .tlow = 28, .tsu_sto = 21, .tsu_sta = 21, .thd_dat = 13, .thd_sta = 18, .tbuf = 32, .scl_stretch_en = 0, .trdhld = 6, .tsp = 3 }, }; static const struct cci_data cci_v1_5_data = { .num_masters = 2, .queue_size = { 64, 16 }, .quirks = { .max_write_len = 10, .max_read_len = 12, }, .cci_clk_rate = 19200000, .params[I2C_MODE_STANDARD] = { .thigh = 78, .tlow = 114, .tsu_sto = 28, .tsu_sta = 28, .thd_dat = 10, .thd_sta = 77, .tbuf = 118, .scl_stretch_en = 0, .trdhld = 6, .tsp = 1 }, .params[I2C_MODE_FAST] = { .thigh = 20, .tlow = 28, .tsu_sto = 21, .tsu_sta = 21, .thd_dat = 13, .thd_sta = 18, .tbuf = 32, .scl_stretch_en = 0, .trdhld = 6, .tsp = 3 }, }; static const struct cci_data cci_v2_data = { .num_masters = 2, .queue_size = { 64, 16 }, .quirks = { .max_write_len = 11, .max_read_len = 12, }, .cci_clk_rate = 37500000, .params[I2C_MODE_STANDARD] = { .thigh = 201, .tlow = 174, .tsu_sto = 204, .tsu_sta = 231, .thd_dat = 22, .thd_sta = 162, .tbuf = 227, .scl_stretch_en = 0, .trdhld = 6, .tsp = 3 }, .params[I2C_MODE_FAST] = { .thigh = 38, .tlow = 56, .tsu_sto = 40, .tsu_sta = 40, .thd_dat = 22, .thd_sta = 35, .tbuf = 62, .scl_stretch_en = 0, .trdhld = 6, .tsp = 3 }, .params[I2C_MODE_FAST_PLUS] = { .thigh = 16, .tlow = 22, .tsu_sto = 17, .tsu_sta = 18, .thd_dat = 16, .thd_sta = 15, .tbuf = 24, .scl_stretch_en = 0, .trdhld = 3, .tsp = 3 }, }; static const struct of_device_id cci_dt_match[] = { { .compatible = "qcom,msm8226-cci", .data = &cci_v1_data}, { .compatible = "qcom,msm8974-cci", .data = &cci_v1_5_data}, { .compatible = "qcom,msm8996-cci", .data = &cci_v2_data}, /* * Legacy compatibles kept for backwards compatibility. * Do not add any new ones unless they introduce a new config */ { .compatible = "qcom,msm8916-cci", .data = &cci_v1_data}, { .compatible = "qcom,sdm845-cci", .data = &cci_v2_data}, { .compatible = "qcom,sm8250-cci", .data = &cci_v2_data}, { .compatible = "qcom,sm8450-cci", .data = &cci_v2_data}, {} }; MODULE_DEVICE_TABLE(of, cci_dt_match); static struct platform_driver qcom_cci_driver = { .probe = cci_probe, .remove_new = cci_remove, .driver = { .name = "i2c-qcom-cci", .of_match_table = cci_dt_match, .pm = &qcom_cci_pm, }, }; module_platform_driver(qcom_cci_driver); MODULE_DESCRIPTION("Qualcomm Camera Control Interface driver"); MODULE_AUTHOR("Todor Tomov <todor.tomov@linaro.org>"); MODULE_AUTHOR("Loic Poulain <loic.poulain@linaro.org>"); MODULE_LICENSE("GPL v2");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1