Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Wei Wang | 6245 | 69.01% | 16 | 22.54% |
rui_feng | 950 | 10.50% | 5 | 7.04% |
RickyWu | 711 | 7.86% | 7 | 9.86% |
Micky Ching | 588 | 6.50% | 11 | 15.49% |
Kai-Heng Feng | 139 | 1.54% | 3 | 4.23% |
Tom Rix | 134 | 1.48% | 1 | 1.41% |
Vaibhav Gupta | 60 | 0.66% | 2 | 2.82% |
Steven Feng | 55 | 0.61% | 1 | 1.41% |
Thomas Gleixner | 35 | 0.39% | 2 | 2.82% |
Christophe Jaillet | 29 | 0.32% | 2 | 2.82% |
Tejun Heo | 22 | 0.24% | 1 | 1.41% |
Roger Tseng | 20 | 0.22% | 2 | 2.82% |
Björn Helgaas | 16 | 0.18% | 3 | 4.23% |
Colin Ian King | 9 | 0.10% | 2 | 2.82% |
Greg Kroah-Hartman | 8 | 0.09% | 1 | 1.41% |
Jingoo Han | 6 | 0.07% | 1 | 1.41% |
Klaus Doth | 5 | 0.06% | 1 | 1.41% |
Samuel Ortiz | 3 | 0.03% | 1 | 1.41% |
Arnd Bergmann | 3 | 0.03% | 1 | 1.41% |
Keita Suzuki | 3 | 0.03% | 1 | 1.41% |
Jiri Slaby | 2 | 0.02% | 1 | 1.41% |
Stephen Rothwell | 2 | 0.02% | 1 | 1.41% |
Javier Martinez Canillas | 1 | 0.01% | 1 | 1.41% |
Zhang Jiaming | 1 | 0.01% | 1 | 1.41% |
Li Yang | 1 | 0.01% | 1 | 1.41% |
Borislav Petkov | 1 | 0.01% | 1 | 1.41% |
Christoph Hellwig | 1 | 0.01% | 1 | 1.41% |
Total | 9050 | 71 |
// SPDX-License-Identifier: GPL-2.0-or-later /* Driver for Realtek PCI-Express card reader * * Copyright(c) 2009-2013 Realtek Semiconductor Corp. All rights reserved. * * Author: * Wei WANG <wei_wang@realsil.com.cn> */ #include <linux/pci.h> #include <linux/module.h> #include <linux/slab.h> #include <linux/dma-mapping.h> #include <linux/highmem.h> #include <linux/interrupt.h> #include <linux/delay.h> #include <linux/idr.h> #include <linux/platform_device.h> #include <linux/mfd/core.h> #include <linux/rtsx_pci.h> #include <linux/mmc/card.h> #include <asm/unaligned.h> #include <linux/pm.h> #include <linux/pm_runtime.h> #include "rtsx_pcr.h" #include "rts5261.h" #include "rts5228.h" static bool msi_en = true; module_param(msi_en, bool, S_IRUGO | S_IWUSR); MODULE_PARM_DESC(msi_en, "Enable MSI"); static DEFINE_IDR(rtsx_pci_idr); static DEFINE_SPINLOCK(rtsx_pci_lock); static struct mfd_cell rtsx_pcr_cells[] = { [RTSX_SD_CARD] = { .name = DRV_NAME_RTSX_PCI_SDMMC, }, }; static const struct pci_device_id rtsx_pci_ids[] = { { PCI_DEVICE(0x10EC, 0x5209), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5229), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5289), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5227), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x522A), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5249), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5287), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5286), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x524A), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x525A), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5260), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5261), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { PCI_DEVICE(0x10EC, 0x5228), PCI_CLASS_OTHERS << 16, 0xFF0000 }, { 0, } }; MODULE_DEVICE_TABLE(pci, rtsx_pci_ids); static int rtsx_comm_set_ltr_latency(struct rtsx_pcr *pcr, u32 latency) { rtsx_pci_write_register(pcr, MSGTXDATA0, MASK_8_BIT_DEF, (u8) (latency & 0xFF)); rtsx_pci_write_register(pcr, MSGTXDATA1, MASK_8_BIT_DEF, (u8)((latency >> 8) & 0xFF)); rtsx_pci_write_register(pcr, MSGTXDATA2, MASK_8_BIT_DEF, (u8)((latency >> 16) & 0xFF)); rtsx_pci_write_register(pcr, MSGTXDATA3, MASK_8_BIT_DEF, (u8)((latency >> 24) & 0xFF)); rtsx_pci_write_register(pcr, LTR_CTL, LTR_TX_EN_MASK | LTR_LATENCY_MODE_MASK, LTR_TX_EN_1 | LTR_LATENCY_MODE_SW); return 0; } int rtsx_set_ltr_latency(struct rtsx_pcr *pcr, u32 latency) { return rtsx_comm_set_ltr_latency(pcr, latency); } static void rtsx_comm_set_aspm(struct rtsx_pcr *pcr, bool enable) { if (pcr->aspm_enabled == enable) return; if (pcr->aspm_mode == ASPM_MODE_CFG) { pcie_capability_clear_and_set_word(pcr->pci, PCI_EXP_LNKCTL, PCI_EXP_LNKCTL_ASPMC, enable ? pcr->aspm_en : 0); } else if (pcr->aspm_mode == ASPM_MODE_REG) { if (pcr->aspm_en & 0x02) rtsx_pci_write_register(pcr, ASPM_FORCE_CTL, FORCE_ASPM_CTL0 | FORCE_ASPM_CTL1, enable ? 0 : FORCE_ASPM_CTL0 | FORCE_ASPM_CTL1); else rtsx_pci_write_register(pcr, ASPM_FORCE_CTL, FORCE_ASPM_CTL0 | FORCE_ASPM_CTL1, FORCE_ASPM_CTL0 | FORCE_ASPM_CTL1); } if (!enable && (pcr->aspm_en & 0x02)) mdelay(10); pcr->aspm_enabled = enable; } static void rtsx_disable_aspm(struct rtsx_pcr *pcr) { if (pcr->ops->set_aspm) pcr->ops->set_aspm(pcr, false); else rtsx_comm_set_aspm(pcr, false); } int rtsx_set_l1off_sub(struct rtsx_pcr *pcr, u8 val) { rtsx_pci_write_register(pcr, L1SUB_CONFIG3, 0xFF, val); return 0; } static void rtsx_set_l1off_sub_cfg_d0(struct rtsx_pcr *pcr, int active) { if (pcr->ops->set_l1off_cfg_sub_d0) pcr->ops->set_l1off_cfg_sub_d0(pcr, active); } static void rtsx_comm_pm_full_on(struct rtsx_pcr *pcr) { struct rtsx_cr_option *option = &pcr->option; rtsx_disable_aspm(pcr); /* Fixes DMA transfer timeout issue after disabling ASPM on RTS5260 */ msleep(1); if (option->ltr_enabled) rtsx_set_ltr_latency(pcr, option->ltr_active_latency); if (rtsx_check_dev_flag(pcr, LTR_L1SS_PWR_GATE_EN)) rtsx_set_l1off_sub_cfg_d0(pcr, 1); } static void rtsx_pm_full_on(struct rtsx_pcr *pcr) { rtsx_comm_pm_full_on(pcr); } void rtsx_pci_start_run(struct rtsx_pcr *pcr) { /* If pci device removed, don't queue idle work any more */ if (pcr->remove_pci) return; if (pcr->state != PDEV_STAT_RUN) { pcr->state = PDEV_STAT_RUN; if (pcr->ops->enable_auto_blink) pcr->ops->enable_auto_blink(pcr); rtsx_pm_full_on(pcr); } } EXPORT_SYMBOL_GPL(rtsx_pci_start_run); int rtsx_pci_write_register(struct rtsx_pcr *pcr, u16 addr, u8 mask, u8 data) { int i; u32 val = HAIMR_WRITE_START; val |= (u32)(addr & 0x3FFF) << 16; val |= (u32)mask << 8; val |= (u32)data; rtsx_pci_writel(pcr, RTSX_HAIMR, val); for (i = 0; i < MAX_RW_REG_CNT; i++) { val = rtsx_pci_readl(pcr, RTSX_HAIMR); if ((val & HAIMR_TRANS_END) == 0) { if (data != (u8)val) return -EIO; return 0; } } return -ETIMEDOUT; } EXPORT_SYMBOL_GPL(rtsx_pci_write_register); int rtsx_pci_read_register(struct rtsx_pcr *pcr, u16 addr, u8 *data) { u32 val = HAIMR_READ_START; int i; val |= (u32)(addr & 0x3FFF) << 16; rtsx_pci_writel(pcr, RTSX_HAIMR, val); for (i = 0; i < MAX_RW_REG_CNT; i++) { val = rtsx_pci_readl(pcr, RTSX_HAIMR); if ((val & HAIMR_TRANS_END) == 0) break; } if (i >= MAX_RW_REG_CNT) return -ETIMEDOUT; if (data) *data = (u8)(val & 0xFF); return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_read_register); int __rtsx_pci_write_phy_register(struct rtsx_pcr *pcr, u8 addr, u16 val) { int err, i, finished = 0; u8 tmp; rtsx_pci_write_register(pcr, PHYDATA0, 0xFF, (u8)val); rtsx_pci_write_register(pcr, PHYDATA1, 0xFF, (u8)(val >> 8)); rtsx_pci_write_register(pcr, PHYADDR, 0xFF, addr); rtsx_pci_write_register(pcr, PHYRWCTL, 0xFF, 0x81); for (i = 0; i < 100000; i++) { err = rtsx_pci_read_register(pcr, PHYRWCTL, &tmp); if (err < 0) return err; if (!(tmp & 0x80)) { finished = 1; break; } } if (!finished) return -ETIMEDOUT; return 0; } int rtsx_pci_write_phy_register(struct rtsx_pcr *pcr, u8 addr, u16 val) { if (pcr->ops->write_phy) return pcr->ops->write_phy(pcr, addr, val); return __rtsx_pci_write_phy_register(pcr, addr, val); } EXPORT_SYMBOL_GPL(rtsx_pci_write_phy_register); int __rtsx_pci_read_phy_register(struct rtsx_pcr *pcr, u8 addr, u16 *val) { int err, i, finished = 0; u16 data; u8 tmp, val1, val2; rtsx_pci_write_register(pcr, PHYADDR, 0xFF, addr); rtsx_pci_write_register(pcr, PHYRWCTL, 0xFF, 0x80); for (i = 0; i < 100000; i++) { err = rtsx_pci_read_register(pcr, PHYRWCTL, &tmp); if (err < 0) return err; if (!(tmp & 0x80)) { finished = 1; break; } } if (!finished) return -ETIMEDOUT; rtsx_pci_read_register(pcr, PHYDATA0, &val1); rtsx_pci_read_register(pcr, PHYDATA1, &val2); data = val1 | (val2 << 8); if (val) *val = data; return 0; } int rtsx_pci_read_phy_register(struct rtsx_pcr *pcr, u8 addr, u16 *val) { if (pcr->ops->read_phy) return pcr->ops->read_phy(pcr, addr, val); return __rtsx_pci_read_phy_register(pcr, addr, val); } EXPORT_SYMBOL_GPL(rtsx_pci_read_phy_register); void rtsx_pci_stop_cmd(struct rtsx_pcr *pcr) { if (pcr->ops->stop_cmd) return pcr->ops->stop_cmd(pcr); rtsx_pci_writel(pcr, RTSX_HCBCTLR, STOP_CMD); rtsx_pci_writel(pcr, RTSX_HDBCTLR, STOP_DMA); rtsx_pci_write_register(pcr, DMACTL, 0x80, 0x80); rtsx_pci_write_register(pcr, RBCTL, 0x80, 0x80); } EXPORT_SYMBOL_GPL(rtsx_pci_stop_cmd); void rtsx_pci_add_cmd(struct rtsx_pcr *pcr, u8 cmd_type, u16 reg_addr, u8 mask, u8 data) { unsigned long flags; u32 val = 0; u32 *ptr = (u32 *)(pcr->host_cmds_ptr); val |= (u32)(cmd_type & 0x03) << 30; val |= (u32)(reg_addr & 0x3FFF) << 16; val |= (u32)mask << 8; val |= (u32)data; spin_lock_irqsave(&pcr->lock, flags); ptr += pcr->ci; if (pcr->ci < (HOST_CMDS_BUF_LEN / 4)) { put_unaligned_le32(val, ptr); ptr++; pcr->ci++; } spin_unlock_irqrestore(&pcr->lock, flags); } EXPORT_SYMBOL_GPL(rtsx_pci_add_cmd); void rtsx_pci_send_cmd_no_wait(struct rtsx_pcr *pcr) { u32 val = 1 << 31; rtsx_pci_writel(pcr, RTSX_HCBAR, pcr->host_cmds_addr); val |= (u32)(pcr->ci * 4) & 0x00FFFFFF; /* Hardware Auto Response */ val |= 0x40000000; rtsx_pci_writel(pcr, RTSX_HCBCTLR, val); } EXPORT_SYMBOL_GPL(rtsx_pci_send_cmd_no_wait); int rtsx_pci_send_cmd(struct rtsx_pcr *pcr, int timeout) { struct completion trans_done; u32 val = 1 << 31; long timeleft; unsigned long flags; int err = 0; spin_lock_irqsave(&pcr->lock, flags); /* set up data structures for the wakeup system */ pcr->done = &trans_done; pcr->trans_result = TRANS_NOT_READY; init_completion(&trans_done); rtsx_pci_writel(pcr, RTSX_HCBAR, pcr->host_cmds_addr); val |= (u32)(pcr->ci * 4) & 0x00FFFFFF; /* Hardware Auto Response */ val |= 0x40000000; rtsx_pci_writel(pcr, RTSX_HCBCTLR, val); spin_unlock_irqrestore(&pcr->lock, flags); /* Wait for TRANS_OK_INT */ timeleft = wait_for_completion_interruptible_timeout( &trans_done, msecs_to_jiffies(timeout)); if (timeleft <= 0) { pcr_dbg(pcr, "Timeout (%s %d)\n", __func__, __LINE__); err = -ETIMEDOUT; goto finish_send_cmd; } spin_lock_irqsave(&pcr->lock, flags); if (pcr->trans_result == TRANS_RESULT_FAIL) err = -EINVAL; else if (pcr->trans_result == TRANS_RESULT_OK) err = 0; else if (pcr->trans_result == TRANS_NO_DEVICE) err = -ENODEV; spin_unlock_irqrestore(&pcr->lock, flags); finish_send_cmd: spin_lock_irqsave(&pcr->lock, flags); pcr->done = NULL; spin_unlock_irqrestore(&pcr->lock, flags); if ((err < 0) && (err != -ENODEV)) rtsx_pci_stop_cmd(pcr); if (pcr->finish_me) complete(pcr->finish_me); return err; } EXPORT_SYMBOL_GPL(rtsx_pci_send_cmd); static void rtsx_pci_add_sg_tbl(struct rtsx_pcr *pcr, dma_addr_t addr, unsigned int len, int end) { u64 *ptr = (u64 *)(pcr->host_sg_tbl_ptr) + pcr->sgi; u64 val; u8 option = RTSX_SG_VALID | RTSX_SG_TRANS_DATA; pcr_dbg(pcr, "DMA addr: 0x%x, Len: 0x%x\n", (unsigned int)addr, len); if (end) option |= RTSX_SG_END; if ((PCI_PID(pcr) == PID_5261) || (PCI_PID(pcr) == PID_5228)) { if (len > 0xFFFF) val = ((u64)addr << 32) | (((u64)len & 0xFFFF) << 16) | (((u64)len >> 16) << 6) | option; else val = ((u64)addr << 32) | ((u64)len << 16) | option; } else { val = ((u64)addr << 32) | ((u64)len << 12) | option; } put_unaligned_le64(val, ptr); pcr->sgi++; } int rtsx_pci_transfer_data(struct rtsx_pcr *pcr, struct scatterlist *sglist, int num_sg, bool read, int timeout) { int err = 0, count; pcr_dbg(pcr, "--> %s: num_sg = %d\n", __func__, num_sg); count = rtsx_pci_dma_map_sg(pcr, sglist, num_sg, read); if (count < 1) return -EINVAL; pcr_dbg(pcr, "DMA mapping count: %d\n", count); err = rtsx_pci_dma_transfer(pcr, sglist, count, read, timeout); rtsx_pci_dma_unmap_sg(pcr, sglist, num_sg, read); return err; } EXPORT_SYMBOL_GPL(rtsx_pci_transfer_data); int rtsx_pci_dma_map_sg(struct rtsx_pcr *pcr, struct scatterlist *sglist, int num_sg, bool read) { enum dma_data_direction dir = read ? DMA_FROM_DEVICE : DMA_TO_DEVICE; if (pcr->remove_pci) return -EINVAL; if ((sglist == NULL) || (num_sg <= 0)) return -EINVAL; return dma_map_sg(&(pcr->pci->dev), sglist, num_sg, dir); } EXPORT_SYMBOL_GPL(rtsx_pci_dma_map_sg); void rtsx_pci_dma_unmap_sg(struct rtsx_pcr *pcr, struct scatterlist *sglist, int num_sg, bool read) { enum dma_data_direction dir = read ? DMA_FROM_DEVICE : DMA_TO_DEVICE; dma_unmap_sg(&(pcr->pci->dev), sglist, num_sg, dir); } EXPORT_SYMBOL_GPL(rtsx_pci_dma_unmap_sg); int rtsx_pci_dma_transfer(struct rtsx_pcr *pcr, struct scatterlist *sglist, int count, bool read, int timeout) { struct completion trans_done; struct scatterlist *sg; dma_addr_t addr; long timeleft; unsigned long flags; unsigned int len; int i, err = 0; u32 val; u8 dir = read ? DEVICE_TO_HOST : HOST_TO_DEVICE; if (pcr->remove_pci) return -ENODEV; if ((sglist == NULL) || (count < 1)) return -EINVAL; val = ((u32)(dir & 0x01) << 29) | TRIG_DMA | ADMA_MODE; pcr->sgi = 0; for_each_sg(sglist, sg, count, i) { addr = sg_dma_address(sg); len = sg_dma_len(sg); rtsx_pci_add_sg_tbl(pcr, addr, len, i == count - 1); } spin_lock_irqsave(&pcr->lock, flags); pcr->done = &trans_done; pcr->trans_result = TRANS_NOT_READY; init_completion(&trans_done); rtsx_pci_writel(pcr, RTSX_HDBAR, pcr->host_sg_tbl_addr); rtsx_pci_writel(pcr, RTSX_HDBCTLR, val); spin_unlock_irqrestore(&pcr->lock, flags); timeleft = wait_for_completion_interruptible_timeout( &trans_done, msecs_to_jiffies(timeout)); if (timeleft <= 0) { pcr_dbg(pcr, "Timeout (%s %d)\n", __func__, __LINE__); err = -ETIMEDOUT; goto out; } spin_lock_irqsave(&pcr->lock, flags); if (pcr->trans_result == TRANS_RESULT_FAIL) { err = -EILSEQ; if (pcr->dma_error_count < RTS_MAX_TIMES_FREQ_REDUCTION) pcr->dma_error_count++; } else if (pcr->trans_result == TRANS_NO_DEVICE) err = -ENODEV; spin_unlock_irqrestore(&pcr->lock, flags); out: spin_lock_irqsave(&pcr->lock, flags); pcr->done = NULL; spin_unlock_irqrestore(&pcr->lock, flags); if ((err < 0) && (err != -ENODEV)) rtsx_pci_stop_cmd(pcr); if (pcr->finish_me) complete(pcr->finish_me); return err; } EXPORT_SYMBOL_GPL(rtsx_pci_dma_transfer); int rtsx_pci_read_ppbuf(struct rtsx_pcr *pcr, u8 *buf, int buf_len) { int err; int i, j; u16 reg; u8 *ptr; if (buf_len > 512) buf_len = 512; ptr = buf; reg = PPBUF_BASE2; for (i = 0; i < buf_len / 256; i++) { rtsx_pci_init_cmd(pcr); for (j = 0; j < 256; j++) rtsx_pci_add_cmd(pcr, READ_REG_CMD, reg++, 0, 0); err = rtsx_pci_send_cmd(pcr, 250); if (err < 0) return err; memcpy(ptr, rtsx_pci_get_cmd_data(pcr), 256); ptr += 256; } if (buf_len % 256) { rtsx_pci_init_cmd(pcr); for (j = 0; j < buf_len % 256; j++) rtsx_pci_add_cmd(pcr, READ_REG_CMD, reg++, 0, 0); err = rtsx_pci_send_cmd(pcr, 250); if (err < 0) return err; } memcpy(ptr, rtsx_pci_get_cmd_data(pcr), buf_len % 256); return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_read_ppbuf); int rtsx_pci_write_ppbuf(struct rtsx_pcr *pcr, u8 *buf, int buf_len) { int err; int i, j; u16 reg; u8 *ptr; if (buf_len > 512) buf_len = 512; ptr = buf; reg = PPBUF_BASE2; for (i = 0; i < buf_len / 256; i++) { rtsx_pci_init_cmd(pcr); for (j = 0; j < 256; j++) { rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, reg++, 0xFF, *ptr); ptr++; } err = rtsx_pci_send_cmd(pcr, 250); if (err < 0) return err; } if (buf_len % 256) { rtsx_pci_init_cmd(pcr); for (j = 0; j < buf_len % 256; j++) { rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, reg++, 0xFF, *ptr); ptr++; } err = rtsx_pci_send_cmd(pcr, 250); if (err < 0) return err; } return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_write_ppbuf); static int rtsx_pci_set_pull_ctl(struct rtsx_pcr *pcr, const u32 *tbl) { rtsx_pci_init_cmd(pcr); while (*tbl & 0xFFFF0000) { rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, (u16)(*tbl >> 16), 0xFF, (u8)(*tbl)); tbl++; } return rtsx_pci_send_cmd(pcr, 100); } int rtsx_pci_card_pull_ctl_enable(struct rtsx_pcr *pcr, int card) { const u32 *tbl; if (card == RTSX_SD_CARD) tbl = pcr->sd_pull_ctl_enable_tbl; else if (card == RTSX_MS_CARD) tbl = pcr->ms_pull_ctl_enable_tbl; else return -EINVAL; return rtsx_pci_set_pull_ctl(pcr, tbl); } EXPORT_SYMBOL_GPL(rtsx_pci_card_pull_ctl_enable); int rtsx_pci_card_pull_ctl_disable(struct rtsx_pcr *pcr, int card) { const u32 *tbl; if (card == RTSX_SD_CARD) tbl = pcr->sd_pull_ctl_disable_tbl; else if (card == RTSX_MS_CARD) tbl = pcr->ms_pull_ctl_disable_tbl; else return -EINVAL; return rtsx_pci_set_pull_ctl(pcr, tbl); } EXPORT_SYMBOL_GPL(rtsx_pci_card_pull_ctl_disable); static void rtsx_pci_enable_bus_int(struct rtsx_pcr *pcr) { struct rtsx_hw_param *hw_param = &pcr->hw_param; pcr->bier = TRANS_OK_INT_EN | TRANS_FAIL_INT_EN | SD_INT_EN | hw_param->interrupt_en; if (pcr->num_slots > 1) pcr->bier |= MS_INT_EN; /* Enable Bus Interrupt */ rtsx_pci_writel(pcr, RTSX_BIER, pcr->bier); pcr_dbg(pcr, "RTSX_BIER: 0x%08x\n", pcr->bier); } static inline u8 double_ssc_depth(u8 depth) { return ((depth > 1) ? (depth - 1) : depth); } static u8 revise_ssc_depth(u8 ssc_depth, u8 div) { if (div > CLK_DIV_1) { if (ssc_depth > (div - 1)) ssc_depth -= (div - 1); else ssc_depth = SSC_DEPTH_4M; } return ssc_depth; } int rtsx_pci_switch_clock(struct rtsx_pcr *pcr, unsigned int card_clock, u8 ssc_depth, bool initial_mode, bool double_clk, bool vpclk) { int err, clk; u8 n, clk_divider, mcu_cnt, div; static const u8 depth[] = { [RTSX_SSC_DEPTH_4M] = SSC_DEPTH_4M, [RTSX_SSC_DEPTH_2M] = SSC_DEPTH_2M, [RTSX_SSC_DEPTH_1M] = SSC_DEPTH_1M, [RTSX_SSC_DEPTH_500K] = SSC_DEPTH_500K, [RTSX_SSC_DEPTH_250K] = SSC_DEPTH_250K, }; if (PCI_PID(pcr) == PID_5261) return rts5261_pci_switch_clock(pcr, card_clock, ssc_depth, initial_mode, double_clk, vpclk); if (PCI_PID(pcr) == PID_5228) return rts5228_pci_switch_clock(pcr, card_clock, ssc_depth, initial_mode, double_clk, vpclk); if (initial_mode) { /* We use 250k(around) here, in initial stage */ clk_divider = SD_CLK_DIVIDE_128; card_clock = 30000000; } else { clk_divider = SD_CLK_DIVIDE_0; } err = rtsx_pci_write_register(pcr, SD_CFG1, SD_CLK_DIVIDE_MASK, clk_divider); if (err < 0) return err; /* Reduce card clock by 20MHz each time a DMA transfer error occurs */ if (card_clock == UHS_SDR104_MAX_DTR && pcr->dma_error_count && PCI_PID(pcr) == RTS5227_DEVICE_ID) card_clock = UHS_SDR104_MAX_DTR - (pcr->dma_error_count * 20000000); card_clock /= 1000000; pcr_dbg(pcr, "Switch card clock to %dMHz\n", card_clock); clk = card_clock; if (!initial_mode && double_clk) clk = card_clock * 2; pcr_dbg(pcr, "Internal SSC clock: %dMHz (cur_clock = %d)\n", clk, pcr->cur_clock); if (clk == pcr->cur_clock) return 0; if (pcr->ops->conv_clk_and_div_n) n = (u8)pcr->ops->conv_clk_and_div_n(clk, CLK_TO_DIV_N); else n = (u8)(clk - 2); if ((clk <= 2) || (n > MAX_DIV_N_PCR)) return -EINVAL; mcu_cnt = (u8)(125/clk + 3); if (mcu_cnt > 15) mcu_cnt = 15; /* Make sure that the SSC clock div_n is not less than MIN_DIV_N_PCR */ div = CLK_DIV_1; while ((n < MIN_DIV_N_PCR) && (div < CLK_DIV_8)) { if (pcr->ops->conv_clk_and_div_n) { int dbl_clk = pcr->ops->conv_clk_and_div_n(n, DIV_N_TO_CLK) * 2; n = (u8)pcr->ops->conv_clk_and_div_n(dbl_clk, CLK_TO_DIV_N); } else { n = (n + 2) * 2 - 2; } div++; } pcr_dbg(pcr, "n = %d, div = %d\n", n, div); ssc_depth = depth[ssc_depth]; if (double_clk) ssc_depth = double_ssc_depth(ssc_depth); ssc_depth = revise_ssc_depth(ssc_depth, div); pcr_dbg(pcr, "ssc_depth = %d\n", ssc_depth); rtsx_pci_init_cmd(pcr); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CLK_CTL, CLK_LOW_FREQ, CLK_LOW_FREQ); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CLK_DIV, 0xFF, (div << 4) | mcu_cnt); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL1, SSC_RSTB, 0); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL2, SSC_DEPTH_MASK, ssc_depth); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_DIV_N_0, 0xFF, n); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL1, SSC_RSTB, SSC_RSTB); if (vpclk) { rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SD_VPCLK0_CTL, PHASE_NOT_RESET, 0); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SD_VPCLK0_CTL, PHASE_NOT_RESET, PHASE_NOT_RESET); } err = rtsx_pci_send_cmd(pcr, 2000); if (err < 0) return err; /* Wait SSC clock stable */ udelay(SSC_CLOCK_STABLE_WAIT); err = rtsx_pci_write_register(pcr, CLK_CTL, CLK_LOW_FREQ, 0); if (err < 0) return err; pcr->cur_clock = clk; return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_switch_clock); int rtsx_pci_card_power_on(struct rtsx_pcr *pcr, int card) { if (pcr->ops->card_power_on) return pcr->ops->card_power_on(pcr, card); return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_card_power_on); int rtsx_pci_card_power_off(struct rtsx_pcr *pcr, int card) { if (pcr->ops->card_power_off) return pcr->ops->card_power_off(pcr, card); return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_card_power_off); int rtsx_pci_card_exclusive_check(struct rtsx_pcr *pcr, int card) { static const unsigned int cd_mask[] = { [RTSX_SD_CARD] = SD_EXIST, [RTSX_MS_CARD] = MS_EXIST }; if (!(pcr->flags & PCR_MS_PMOS)) { /* When using single PMOS, accessing card is not permitted * if the existing card is not the designated one. */ if (pcr->card_exist & (~cd_mask[card])) return -EIO; } return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_card_exclusive_check); int rtsx_pci_switch_output_voltage(struct rtsx_pcr *pcr, u8 voltage) { if (pcr->ops->switch_output_voltage) return pcr->ops->switch_output_voltage(pcr, voltage); return 0; } EXPORT_SYMBOL_GPL(rtsx_pci_switch_output_voltage); unsigned int rtsx_pci_card_exist(struct rtsx_pcr *pcr) { unsigned int val; val = rtsx_pci_readl(pcr, RTSX_BIPR); if (pcr->ops->cd_deglitch) val = pcr->ops->cd_deglitch(pcr); return val; } EXPORT_SYMBOL_GPL(rtsx_pci_card_exist); void rtsx_pci_complete_unfinished_transfer(struct rtsx_pcr *pcr) { struct completion finish; pcr->finish_me = &finish; init_completion(&finish); if (pcr->done) complete(pcr->done); if (!pcr->remove_pci) rtsx_pci_stop_cmd(pcr); wait_for_completion_interruptible_timeout(&finish, msecs_to_jiffies(2)); pcr->finish_me = NULL; } EXPORT_SYMBOL_GPL(rtsx_pci_complete_unfinished_transfer); static void rtsx_pci_card_detect(struct work_struct *work) { struct delayed_work *dwork; struct rtsx_pcr *pcr; unsigned long flags; unsigned int card_detect = 0, card_inserted, card_removed; u32 irq_status; dwork = to_delayed_work(work); pcr = container_of(dwork, struct rtsx_pcr, carddet_work); pcr_dbg(pcr, "--> %s\n", __func__); mutex_lock(&pcr->pcr_mutex); spin_lock_irqsave(&pcr->lock, flags); irq_status = rtsx_pci_readl(pcr, RTSX_BIPR); pcr_dbg(pcr, "irq_status: 0x%08x\n", irq_status); irq_status &= CARD_EXIST; card_inserted = pcr->card_inserted & irq_status; card_removed = pcr->card_removed; pcr->card_inserted = 0; pcr->card_removed = 0; spin_unlock_irqrestore(&pcr->lock, flags); if (card_inserted || card_removed) { pcr_dbg(pcr, "card_inserted: 0x%x, card_removed: 0x%x\n", card_inserted, card_removed); if (pcr->ops->cd_deglitch) card_inserted = pcr->ops->cd_deglitch(pcr); card_detect = card_inserted | card_removed; pcr->card_exist |= card_inserted; pcr->card_exist &= ~card_removed; } mutex_unlock(&pcr->pcr_mutex); if ((card_detect & SD_EXIST) && pcr->slots[RTSX_SD_CARD].card_event) pcr->slots[RTSX_SD_CARD].card_event( pcr->slots[RTSX_SD_CARD].p_dev); if ((card_detect & MS_EXIST) && pcr->slots[RTSX_MS_CARD].card_event) pcr->slots[RTSX_MS_CARD].card_event( pcr->slots[RTSX_MS_CARD].p_dev); } static void rtsx_pci_process_ocp(struct rtsx_pcr *pcr) { if (pcr->ops->process_ocp) { pcr->ops->process_ocp(pcr); } else { if (!pcr->option.ocp_en) return; rtsx_pci_get_ocpstat(pcr, &pcr->ocp_stat); if (pcr->ocp_stat & (SD_OC_NOW | SD_OC_EVER)) { rtsx_pci_card_power_off(pcr, RTSX_SD_CARD); rtsx_pci_write_register(pcr, CARD_OE, SD_OUTPUT_EN, 0); rtsx_pci_clear_ocpstat(pcr); pcr->ocp_stat = 0; } } } static int rtsx_pci_process_ocp_interrupt(struct rtsx_pcr *pcr) { if (pcr->option.ocp_en) rtsx_pci_process_ocp(pcr); return 0; } static irqreturn_t rtsx_pci_isr(int irq, void *dev_id) { struct rtsx_pcr *pcr = dev_id; u32 int_reg; if (!pcr) return IRQ_NONE; spin_lock(&pcr->lock); int_reg = rtsx_pci_readl(pcr, RTSX_BIPR); /* Clear interrupt flag */ rtsx_pci_writel(pcr, RTSX_BIPR, int_reg); if ((int_reg & pcr->bier) == 0) { spin_unlock(&pcr->lock); return IRQ_NONE; } if (int_reg == 0xFFFFFFFF) { spin_unlock(&pcr->lock); return IRQ_HANDLED; } int_reg &= (pcr->bier | 0x7FFFFF); if (int_reg & SD_OC_INT) rtsx_pci_process_ocp_interrupt(pcr); if (int_reg & SD_INT) { if (int_reg & SD_EXIST) { pcr->card_inserted |= SD_EXIST; } else { pcr->card_removed |= SD_EXIST; pcr->card_inserted &= ~SD_EXIST; if (PCI_PID(pcr) == PID_5261) { rtsx_pci_write_register(pcr, RTS5261_FW_STATUS, RTS5261_EXPRESS_LINK_FAIL_MASK, 0); pcr->extra_caps |= EXTRA_CAPS_SD_EXPRESS; } } pcr->dma_error_count = 0; } if (int_reg & MS_INT) { if (int_reg & MS_EXIST) { pcr->card_inserted |= MS_EXIST; } else { pcr->card_removed |= MS_EXIST; pcr->card_inserted &= ~MS_EXIST; } } if (int_reg & (NEED_COMPLETE_INT | DELINK_INT)) { if (int_reg & (TRANS_FAIL_INT | DELINK_INT)) { pcr->trans_result = TRANS_RESULT_FAIL; if (pcr->done) complete(pcr->done); } else if (int_reg & TRANS_OK_INT) { pcr->trans_result = TRANS_RESULT_OK; if (pcr->done) complete(pcr->done); } } if ((pcr->card_inserted || pcr->card_removed) && !(int_reg & SD_OC_INT)) schedule_delayed_work(&pcr->carddet_work, msecs_to_jiffies(200)); spin_unlock(&pcr->lock); return IRQ_HANDLED; } static int rtsx_pci_acquire_irq(struct rtsx_pcr *pcr) { pcr_dbg(pcr, "%s: pcr->msi_en = %d, pci->irq = %d\n", __func__, pcr->msi_en, pcr->pci->irq); if (request_irq(pcr->pci->irq, rtsx_pci_isr, pcr->msi_en ? 0 : IRQF_SHARED, DRV_NAME_RTSX_PCI, pcr)) { dev_err(&(pcr->pci->dev), "rtsx_sdmmc: unable to grab IRQ %d, disabling device\n", pcr->pci->irq); return -1; } pcr->irq = pcr->pci->irq; pci_intx(pcr->pci, !pcr->msi_en); return 0; } static void rtsx_base_force_power_down(struct rtsx_pcr *pcr) { /* Set relink_time to 0 */ rtsx_pci_write_register(pcr, AUTOLOAD_CFG_BASE + 1, MASK_8_BIT_DEF, 0); rtsx_pci_write_register(pcr, AUTOLOAD_CFG_BASE + 2, MASK_8_BIT_DEF, 0); rtsx_pci_write_register(pcr, AUTOLOAD_CFG_BASE + 3, RELINK_TIME_MASK, 0); rtsx_pci_write_register(pcr, pcr->reg_pm_ctrl3, D3_DELINK_MODE_EN, D3_DELINK_MODE_EN); rtsx_pci_write_register(pcr, FPDCTL, ALL_POWER_DOWN, ALL_POWER_DOWN); } static void __maybe_unused rtsx_pci_power_off(struct rtsx_pcr *pcr, u8 pm_state, bool runtime) { if (pcr->ops->turn_off_led) pcr->ops->turn_off_led(pcr); rtsx_pci_writel(pcr, RTSX_BIER, 0); pcr->bier = 0; rtsx_pci_write_register(pcr, PETXCFG, 0x08, 0x08); rtsx_pci_write_register(pcr, HOST_SLEEP_STATE, 0x03, pm_state); if (pcr->ops->force_power_down) pcr->ops->force_power_down(pcr, pm_state, runtime); else rtsx_base_force_power_down(pcr); } void rtsx_pci_enable_ocp(struct rtsx_pcr *pcr) { u8 val = SD_OCP_INT_EN | SD_DETECT_EN; if (pcr->ops->enable_ocp) { pcr->ops->enable_ocp(pcr); } else { rtsx_pci_write_register(pcr, FPDCTL, OC_POWER_DOWN, 0); rtsx_pci_write_register(pcr, REG_OCPCTL, 0xFF, val); } } void rtsx_pci_disable_ocp(struct rtsx_pcr *pcr) { u8 mask = SD_OCP_INT_EN | SD_DETECT_EN; if (pcr->ops->disable_ocp) { pcr->ops->disable_ocp(pcr); } else { rtsx_pci_write_register(pcr, REG_OCPCTL, mask, 0); rtsx_pci_write_register(pcr, FPDCTL, OC_POWER_DOWN, OC_POWER_DOWN); } } void rtsx_pci_init_ocp(struct rtsx_pcr *pcr) { if (pcr->ops->init_ocp) { pcr->ops->init_ocp(pcr); } else { struct rtsx_cr_option *option = &(pcr->option); if (option->ocp_en) { u8 val = option->sd_800mA_ocp_thd; rtsx_pci_write_register(pcr, FPDCTL, OC_POWER_DOWN, 0); rtsx_pci_write_register(pcr, REG_OCPPARA1, SD_OCP_TIME_MASK, SD_OCP_TIME_800); rtsx_pci_write_register(pcr, REG_OCPPARA2, SD_OCP_THD_MASK, val); rtsx_pci_write_register(pcr, REG_OCPGLITCH, SD_OCP_GLITCH_MASK, pcr->hw_param.ocp_glitch); rtsx_pci_enable_ocp(pcr); } } } int rtsx_pci_get_ocpstat(struct rtsx_pcr *pcr, u8 *val) { if (pcr->ops->get_ocpstat) return pcr->ops->get_ocpstat(pcr, val); else return rtsx_pci_read_register(pcr, REG_OCPSTAT, val); } void rtsx_pci_clear_ocpstat(struct rtsx_pcr *pcr) { if (pcr->ops->clear_ocpstat) { pcr->ops->clear_ocpstat(pcr); } else { u8 mask = SD_OCP_INT_CLR | SD_OC_CLR; u8 val = SD_OCP_INT_CLR | SD_OC_CLR; rtsx_pci_write_register(pcr, REG_OCPCTL, mask, val); udelay(100); rtsx_pci_write_register(pcr, REG_OCPCTL, mask, 0); } } void rtsx_pci_enable_oobs_polling(struct rtsx_pcr *pcr) { u16 val; if ((PCI_PID(pcr) != PID_525A) && (PCI_PID(pcr) != PID_5260)) { rtsx_pci_read_phy_register(pcr, 0x01, &val); val |= 1<<9; rtsx_pci_write_phy_register(pcr, 0x01, val); } rtsx_pci_write_register(pcr, REG_CFG_OOBS_OFF_TIMER, 0xFF, 0x32); rtsx_pci_write_register(pcr, REG_CFG_OOBS_ON_TIMER, 0xFF, 0x05); rtsx_pci_write_register(pcr, REG_CFG_VCM_ON_TIMER, 0xFF, 0x83); rtsx_pci_write_register(pcr, REG_CFG_OOBS_POLLING, 0xFF, 0xDE); } void rtsx_pci_disable_oobs_polling(struct rtsx_pcr *pcr) { u16 val; if ((PCI_PID(pcr) != PID_525A) && (PCI_PID(pcr) != PID_5260)) { rtsx_pci_read_phy_register(pcr, 0x01, &val); val &= ~(1<<9); rtsx_pci_write_phy_register(pcr, 0x01, val); } rtsx_pci_write_register(pcr, REG_CFG_VCM_ON_TIMER, 0xFF, 0x03); rtsx_pci_write_register(pcr, REG_CFG_OOBS_POLLING, 0xFF, 0x00); } int rtsx_sd_power_off_card3v3(struct rtsx_pcr *pcr) { rtsx_pci_write_register(pcr, CARD_CLK_EN, SD_CLK_EN | MS_CLK_EN | SD40_CLK_EN, 0); rtsx_pci_write_register(pcr, CARD_OE, SD_OUTPUT_EN, 0); rtsx_pci_card_power_off(pcr, RTSX_SD_CARD); msleep(50); rtsx_pci_card_pull_ctl_disable(pcr, RTSX_SD_CARD); return 0; } int rtsx_ms_power_off_card3v3(struct rtsx_pcr *pcr) { rtsx_pci_write_register(pcr, CARD_CLK_EN, SD_CLK_EN | MS_CLK_EN | SD40_CLK_EN, 0); rtsx_pci_card_pull_ctl_disable(pcr, RTSX_MS_CARD); rtsx_pci_write_register(pcr, CARD_OE, MS_OUTPUT_EN, 0); rtsx_pci_card_power_off(pcr, RTSX_MS_CARD); return 0; } static int rtsx_pci_init_hw(struct rtsx_pcr *pcr) { struct pci_dev *pdev = pcr->pci; int err; if (PCI_PID(pcr) == PID_5228) rtsx_pci_write_register(pcr, RTS5228_LDO1_CFG1, RTS5228_LDO1_SR_TIME_MASK, RTS5228_LDO1_SR_0_5); rtsx_pci_writel(pcr, RTSX_HCBAR, pcr->host_cmds_addr); rtsx_pci_enable_bus_int(pcr); /* Power on SSC */ if (PCI_PID(pcr) == PID_5261) { /* Gating real mcu clock */ err = rtsx_pci_write_register(pcr, RTS5261_FW_CFG1, RTS5261_MCU_CLOCK_GATING, 0); err = rtsx_pci_write_register(pcr, RTS5261_REG_FPDCTL, SSC_POWER_DOWN, 0); } else { err = rtsx_pci_write_register(pcr, FPDCTL, SSC_POWER_DOWN, 0); } if (err < 0) return err; /* Wait SSC power stable */ udelay(200); rtsx_disable_aspm(pcr); if (pcr->ops->optimize_phy) { err = pcr->ops->optimize_phy(pcr); if (err < 0) return err; } rtsx_pci_init_cmd(pcr); /* Set mcu_cnt to 7 to ensure data can be sampled properly */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CLK_DIV, 0x07, 0x07); rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, HOST_SLEEP_STATE, 0x03, 0x00); /* Disable card clock */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CARD_CLK_EN, 0x1E, 0); /* Reset delink mode */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CHANGE_LINK_STATE, 0x0A, 0); /* Card driving select */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CARD_DRIVE_SEL, 0xFF, pcr->card_drive_sel); /* Enable SSC Clock */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL1, 0xFF, SSC_8X_EN | SSC_SEL_4M); if (PCI_PID(pcr) == PID_5261) rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL2, 0xFF, RTS5261_SSC_DEPTH_2M); else if (PCI_PID(pcr) == PID_5228) rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL2, 0xFF, RTS5228_SSC_DEPTH_2M); else rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SSC_CTL2, 0xFF, 0x12); /* Disable cd_pwr_save */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, CHANGE_LINK_STATE, 0x16, 0x10); /* Clear Link Ready Interrupt */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, IRQSTAT0, LINK_RDY_INT, LINK_RDY_INT); /* Enlarge the estimation window of PERST# glitch * to reduce the chance of invalid card interrupt */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, PERST_GLITCH_WIDTH, 0xFF, 0x80); /* Update RC oscillator to 400k * bit[0] F_HIGH: for RC oscillator, Rst_value is 1'b1 * 1: 2M 0: 400k */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, RCCTL, 0x01, 0x00); /* Set interrupt write clear * bit 1: U_elbi_if_rd_clr_en * 1: Enable ELBI interrupt[31:22] & [7:0] flag read clear * 0: ELBI interrupt flag[31:22] & [7:0] only can be write clear */ rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, NFTS_TX_CTRL, 0x02, 0); err = rtsx_pci_send_cmd(pcr, 100); if (err < 0) return err; switch (PCI_PID(pcr)) { case PID_5250: case PID_524A: case PID_525A: case PID_5260: case PID_5261: case PID_5228: rtsx_pci_write_register(pcr, PM_CLK_FORCE_CTL, 1, 1); break; default: break; } /*init ocp*/ rtsx_pci_init_ocp(pcr); /* Enable clk_request_n to enable clock power management */ pcie_capability_clear_and_set_word(pcr->pci, PCI_EXP_LNKCTL, 0, PCI_EXP_LNKCTL_CLKREQ_EN); /* Enter L1 when host tx idle */ pci_write_config_byte(pdev, 0x70F, 0x5B); if (pcr->ops->extra_init_hw) { err = pcr->ops->extra_init_hw(pcr); if (err < 0) return err; } if (pcr->aspm_mode == ASPM_MODE_REG) rtsx_pci_write_register(pcr, ASPM_FORCE_CTL, 0x30, 0x30); /* No CD interrupt if probing driver with card inserted. * So we need to initialize pcr->card_exist here. */ if (pcr->ops->cd_deglitch) pcr->card_exist = pcr->ops->cd_deglitch(pcr); else pcr->card_exist = rtsx_pci_readl(pcr, RTSX_BIPR) & CARD_EXIST; return 0; } static int rtsx_pci_init_chip(struct rtsx_pcr *pcr) { int err; u16 cfg_val; u8 val; spin_lock_init(&pcr->lock); mutex_init(&pcr->pcr_mutex); switch (PCI_PID(pcr)) { default: case 0x5209: rts5209_init_params(pcr); break; case 0x5229: rts5229_init_params(pcr); break; case 0x5289: rtl8411_init_params(pcr); break; case 0x5227: rts5227_init_params(pcr); break; case 0x522A: rts522a_init_params(pcr); break; case 0x5249: rts5249_init_params(pcr); break; case 0x524A: rts524a_init_params(pcr); break; case 0x525A: rts525a_init_params(pcr); break; case 0x5287: rtl8411b_init_params(pcr); break; case 0x5286: rtl8402_init_params(pcr); break; case 0x5260: rts5260_init_params(pcr); break; case 0x5261: rts5261_init_params(pcr); break; case 0x5228: rts5228_init_params(pcr); break; } pcr_dbg(pcr, "PID: 0x%04x, IC version: 0x%02x\n", PCI_PID(pcr), pcr->ic_version); pcr->slots = kcalloc(pcr->num_slots, sizeof(struct rtsx_slot), GFP_KERNEL); if (!pcr->slots) return -ENOMEM; if (pcr->aspm_mode == ASPM_MODE_CFG) { pcie_capability_read_word(pcr->pci, PCI_EXP_LNKCTL, &cfg_val); if (cfg_val & PCI_EXP_LNKCTL_ASPM_L1) pcr->aspm_enabled = true; else pcr->aspm_enabled = false; } else if (pcr->aspm_mode == ASPM_MODE_REG) { rtsx_pci_read_register(pcr, ASPM_FORCE_CTL, &val); if (val & FORCE_ASPM_CTL0 && val & FORCE_ASPM_CTL1) pcr->aspm_enabled = false; else pcr->aspm_enabled = true; } if (pcr->ops->fetch_vendor_settings) pcr->ops->fetch_vendor_settings(pcr); pcr_dbg(pcr, "pcr->aspm_en = 0x%x\n", pcr->aspm_en); pcr_dbg(pcr, "pcr->sd30_drive_sel_1v8 = 0x%x\n", pcr->sd30_drive_sel_1v8); pcr_dbg(pcr, "pcr->sd30_drive_sel_3v3 = 0x%x\n", pcr->sd30_drive_sel_3v3); pcr_dbg(pcr, "pcr->card_drive_sel = 0x%x\n", pcr->card_drive_sel); pcr_dbg(pcr, "pcr->flags = 0x%x\n", pcr->flags); pcr->state = PDEV_STAT_IDLE; err = rtsx_pci_init_hw(pcr); if (err < 0) { kfree(pcr->slots); return err; } return 0; } static int rtsx_pci_probe(struct pci_dev *pcidev, const struct pci_device_id *id) { struct rtsx_pcr *pcr; struct pcr_handle *handle; u32 base, len; int ret, i, bar = 0; dev_dbg(&(pcidev->dev), ": Realtek PCI-E Card Reader found at %s [%04x:%04x] (rev %x)\n", pci_name(pcidev), (int)pcidev->vendor, (int)pcidev->device, (int)pcidev->revision); ret = dma_set_mask(&pcidev->dev, DMA_BIT_MASK(32)); if (ret < 0) return ret; ret = pci_enable_device(pcidev); if (ret) return ret; ret = pci_request_regions(pcidev, DRV_NAME_RTSX_PCI); if (ret) goto disable; pcr = kzalloc(sizeof(*pcr), GFP_KERNEL); if (!pcr) { ret = -ENOMEM; goto release_pci; } handle = kzalloc(sizeof(*handle), GFP_KERNEL); if (!handle) { ret = -ENOMEM; goto free_pcr; } handle->pcr = pcr; idr_preload(GFP_KERNEL); spin_lock(&rtsx_pci_lock); ret = idr_alloc(&rtsx_pci_idr, pcr, 0, 0, GFP_NOWAIT); if (ret >= 0) pcr->id = ret; spin_unlock(&rtsx_pci_lock); idr_preload_end(); if (ret < 0) goto free_handle; pcr->pci = pcidev; dev_set_drvdata(&pcidev->dev, handle); if (CHK_PCI_PID(pcr, 0x525A)) bar = 1; len = pci_resource_len(pcidev, bar); base = pci_resource_start(pcidev, bar); pcr->remap_addr = ioremap(base, len); if (!pcr->remap_addr) { ret = -ENOMEM; goto free_idr; } pcr->rtsx_resv_buf = dma_alloc_coherent(&(pcidev->dev), RTSX_RESV_BUF_LEN, &(pcr->rtsx_resv_buf_addr), GFP_KERNEL); if (pcr->rtsx_resv_buf == NULL) { ret = -ENXIO; goto unmap; } pcr->host_cmds_ptr = pcr->rtsx_resv_buf; pcr->host_cmds_addr = pcr->rtsx_resv_buf_addr; pcr->host_sg_tbl_ptr = pcr->rtsx_resv_buf + HOST_CMDS_BUF_LEN; pcr->host_sg_tbl_addr = pcr->rtsx_resv_buf_addr + HOST_CMDS_BUF_LEN; pcr->card_inserted = 0; pcr->card_removed = 0; INIT_DELAYED_WORK(&pcr->carddet_work, rtsx_pci_card_detect); pcr->msi_en = msi_en; if (pcr->msi_en) { ret = pci_enable_msi(pcidev); if (ret) pcr->msi_en = false; } ret = rtsx_pci_acquire_irq(pcr); if (ret < 0) goto disable_msi; pci_set_master(pcidev); synchronize_irq(pcr->irq); ret = rtsx_pci_init_chip(pcr); if (ret < 0) goto disable_irq; for (i = 0; i < ARRAY_SIZE(rtsx_pcr_cells); i++) { rtsx_pcr_cells[i].platform_data = handle; rtsx_pcr_cells[i].pdata_size = sizeof(*handle); } ret = mfd_add_devices(&pcidev->dev, pcr->id, rtsx_pcr_cells, ARRAY_SIZE(rtsx_pcr_cells), NULL, 0, NULL); if (ret < 0) goto free_slots; pm_runtime_allow(&pcidev->dev); pm_runtime_put(&pcidev->dev); return 0; free_slots: kfree(pcr->slots); disable_irq: free_irq(pcr->irq, (void *)pcr); disable_msi: if (pcr->msi_en) pci_disable_msi(pcr->pci); dma_free_coherent(&(pcr->pci->dev), RTSX_RESV_BUF_LEN, pcr->rtsx_resv_buf, pcr->rtsx_resv_buf_addr); unmap: iounmap(pcr->remap_addr); free_idr: spin_lock(&rtsx_pci_lock); idr_remove(&rtsx_pci_idr, pcr->id); spin_unlock(&rtsx_pci_lock); free_handle: kfree(handle); free_pcr: kfree(pcr); release_pci: pci_release_regions(pcidev); disable: pci_disable_device(pcidev); return ret; } static void rtsx_pci_remove(struct pci_dev *pcidev) { struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; pcr->remove_pci = true; pm_runtime_get_sync(&pcidev->dev); pm_runtime_forbid(&pcidev->dev); /* Disable interrupts at the pcr level */ spin_lock_irq(&pcr->lock); rtsx_pci_writel(pcr, RTSX_BIER, 0); pcr->bier = 0; spin_unlock_irq(&pcr->lock); cancel_delayed_work_sync(&pcr->carddet_work); mfd_remove_devices(&pcidev->dev); dma_free_coherent(&(pcr->pci->dev), RTSX_RESV_BUF_LEN, pcr->rtsx_resv_buf, pcr->rtsx_resv_buf_addr); free_irq(pcr->irq, (void *)pcr); if (pcr->msi_en) pci_disable_msi(pcr->pci); iounmap(pcr->remap_addr); pci_release_regions(pcidev); pci_disable_device(pcidev); spin_lock(&rtsx_pci_lock); idr_remove(&rtsx_pci_idr, pcr->id); spin_unlock(&rtsx_pci_lock); kfree(pcr->slots); kfree(pcr); kfree(handle); dev_dbg(&(pcidev->dev), ": Realtek PCI-E Card Reader at %s [%04x:%04x] has been removed\n", pci_name(pcidev), (int)pcidev->vendor, (int)pcidev->device); } static int __maybe_unused rtsx_pci_suspend(struct device *dev_d) { struct pci_dev *pcidev = to_pci_dev(dev_d); struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; dev_dbg(&(pcidev->dev), "--> %s\n", __func__); cancel_delayed_work_sync(&pcr->carddet_work); mutex_lock(&pcr->pcr_mutex); rtsx_pci_power_off(pcr, HOST_ENTER_S3, false); mutex_unlock(&pcr->pcr_mutex); return 0; } static int __maybe_unused rtsx_pci_resume(struct device *dev_d) { struct pci_dev *pcidev = to_pci_dev(dev_d); struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; int ret = 0; dev_dbg(&(pcidev->dev), "--> %s\n", __func__); mutex_lock(&pcr->pcr_mutex); ret = rtsx_pci_write_register(pcr, HOST_SLEEP_STATE, 0x03, 0x00); if (ret) goto out; ret = rtsx_pci_init_hw(pcr); if (ret) goto out; out: mutex_unlock(&pcr->pcr_mutex); return ret; } #ifdef CONFIG_PM static void rtsx_enable_aspm(struct rtsx_pcr *pcr) { if (pcr->ops->set_aspm) pcr->ops->set_aspm(pcr, true); else rtsx_comm_set_aspm(pcr, true); } static void rtsx_comm_pm_power_saving(struct rtsx_pcr *pcr) { struct rtsx_cr_option *option = &pcr->option; if (option->ltr_enabled) { u32 latency = option->ltr_l1off_latency; if (rtsx_check_dev_flag(pcr, L1_SNOOZE_TEST_EN)) mdelay(option->l1_snooze_delay); rtsx_set_ltr_latency(pcr, latency); } if (rtsx_check_dev_flag(pcr, LTR_L1SS_PWR_GATE_EN)) rtsx_set_l1off_sub_cfg_d0(pcr, 0); rtsx_enable_aspm(pcr); } static void rtsx_pm_power_saving(struct rtsx_pcr *pcr) { rtsx_comm_pm_power_saving(pcr); } static void rtsx_pci_shutdown(struct pci_dev *pcidev) { struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; dev_dbg(&(pcidev->dev), "--> %s\n", __func__); rtsx_pci_power_off(pcr, HOST_ENTER_S1, false); pci_disable_device(pcidev); free_irq(pcr->irq, (void *)pcr); if (pcr->msi_en) pci_disable_msi(pcr->pci); } static int rtsx_pci_runtime_idle(struct device *device) { struct pci_dev *pcidev = to_pci_dev(device); struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; dev_dbg(device, "--> %s\n", __func__); mutex_lock(&pcr->pcr_mutex); pcr->state = PDEV_STAT_IDLE; if (pcr->ops->disable_auto_blink) pcr->ops->disable_auto_blink(pcr); if (pcr->ops->turn_off_led) pcr->ops->turn_off_led(pcr); rtsx_pm_power_saving(pcr); mutex_unlock(&pcr->pcr_mutex); if (pcr->rtd3_en) pm_schedule_suspend(device, 10000); return -EBUSY; } static int rtsx_pci_runtime_suspend(struct device *device) { struct pci_dev *pcidev = to_pci_dev(device); struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; dev_dbg(device, "--> %s\n", __func__); cancel_delayed_work_sync(&pcr->carddet_work); mutex_lock(&pcr->pcr_mutex); rtsx_pci_power_off(pcr, HOST_ENTER_S3, true); mutex_unlock(&pcr->pcr_mutex); return 0; } static int rtsx_pci_runtime_resume(struct device *device) { struct pci_dev *pcidev = to_pci_dev(device); struct pcr_handle *handle = pci_get_drvdata(pcidev); struct rtsx_pcr *pcr = handle->pcr; dev_dbg(device, "--> %s\n", __func__); mutex_lock(&pcr->pcr_mutex); rtsx_pci_write_register(pcr, HOST_SLEEP_STATE, 0x03, 0x00); rtsx_pci_init_hw(pcr); if (pcr->slots[RTSX_SD_CARD].p_dev != NULL) { pcr->slots[RTSX_SD_CARD].card_event( pcr->slots[RTSX_SD_CARD].p_dev); } mutex_unlock(&pcr->pcr_mutex); return 0; } #else /* CONFIG_PM */ #define rtsx_pci_shutdown NULL #define rtsx_pci_runtime_suspend NULL #define rtsx_pic_runtime_resume NULL #endif /* CONFIG_PM */ static const struct dev_pm_ops rtsx_pci_pm_ops = { SET_SYSTEM_SLEEP_PM_OPS(rtsx_pci_suspend, rtsx_pci_resume) SET_RUNTIME_PM_OPS(rtsx_pci_runtime_suspend, rtsx_pci_runtime_resume, rtsx_pci_runtime_idle) }; static struct pci_driver rtsx_pci_driver = { .name = DRV_NAME_RTSX_PCI, .id_table = rtsx_pci_ids, .probe = rtsx_pci_probe, .remove = rtsx_pci_remove, .driver.pm = &rtsx_pci_pm_ops, .shutdown = rtsx_pci_shutdown, }; module_pci_driver(rtsx_pci_driver); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Wei WANG <wei_wang@realsil.com.cn>"); MODULE_DESCRIPTION("Realtek PCI-E Card Reader Driver");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1