Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Keith Busch | 856 | 39.45% | 17 | 26.56% |
Dongdong Liu | 430 | 19.82% | 3 | 4.69% |
Lukas Wunner | 262 | 12.07% | 3 | 4.69% |
Björn Helgaas | 208 | 9.59% | 14 | 21.88% |
Smita Koralahalli | 110 | 5.07% | 1 | 1.56% |
Kai-Heng Feng | 72 | 3.32% | 1 | 1.56% |
Ilpo Järvinen | 51 | 2.35% | 4 | 6.25% |
Kuppuswamy Sathyanarayanan | 45 | 2.07% | 5 | 7.81% |
Frederick Lawler | 37 | 1.71% | 1 | 1.56% |
Oza Pawandeep | 21 | 0.97% | 3 | 4.69% |
Mika Westerberg | 18 | 0.83% | 3 | 4.69% |
Alex Williamson | 16 | 0.74% | 1 | 1.56% |
Yanmin Zhang | 14 | 0.65% | 1 | 1.56% |
Matthew Wilcox | 7 | 0.32% | 1 | 1.56% |
Naveen Naidu | 7 | 0.32% | 1 | 1.56% |
Yicong Yang | 5 | 0.23% | 1 | 1.56% |
Guenter Roeck | 4 | 0.18% | 1 | 1.56% |
Olof Johansson | 3 | 0.14% | 1 | 1.56% |
Dely Sy | 3 | 0.14% | 1 | 1.56% |
Paul Gortmaker | 1 | 0.05% | 1 | 1.56% |
Total | 2170 | 64 |
// SPDX-License-Identifier: GPL-2.0 /* * PCI Express Downstream Port Containment services driver * Author: Keith Busch <keith.busch@intel.com> * * Copyright (C) 2016 Intel Corp. */ #define dev_fmt(fmt) "DPC: " fmt #include <linux/aer.h> #include <linux/bitfield.h> #include <linux/delay.h> #include <linux/interrupt.h> #include <linux/init.h> #include <linux/pci.h> #include "portdrv.h" #include "../pci.h" #define PCI_EXP_DPC_CTL_EN_MASK (PCI_EXP_DPC_CTL_EN_FATAL | \ PCI_EXP_DPC_CTL_EN_NONFATAL) static const char * const rp_pio_error_string[] = { "Configuration Request received UR Completion", /* Bit Position 0 */ "Configuration Request received CA Completion", /* Bit Position 1 */ "Configuration Request Completion Timeout", /* Bit Position 2 */ NULL, NULL, NULL, NULL, NULL, "I/O Request received UR Completion", /* Bit Position 8 */ "I/O Request received CA Completion", /* Bit Position 9 */ "I/O Request Completion Timeout", /* Bit Position 10 */ NULL, NULL, NULL, NULL, NULL, "Memory Request received UR Completion", /* Bit Position 16 */ "Memory Request received CA Completion", /* Bit Position 17 */ "Memory Request Completion Timeout", /* Bit Position 18 */ }; void pci_save_dpc_state(struct pci_dev *dev) { struct pci_cap_saved_state *save_state; u16 *cap; if (!pci_is_pcie(dev)) return; save_state = pci_find_saved_ext_cap(dev, PCI_EXT_CAP_ID_DPC); if (!save_state) return; cap = (u16 *)&save_state->cap.data[0]; pci_read_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, cap); } void pci_restore_dpc_state(struct pci_dev *dev) { struct pci_cap_saved_state *save_state; u16 *cap; if (!pci_is_pcie(dev)) return; save_state = pci_find_saved_ext_cap(dev, PCI_EXT_CAP_ID_DPC); if (!save_state) return; cap = (u16 *)&save_state->cap.data[0]; pci_write_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, *cap); } static DECLARE_WAIT_QUEUE_HEAD(dpc_completed_waitqueue); #ifdef CONFIG_HOTPLUG_PCI_PCIE static bool dpc_completed(struct pci_dev *pdev) { u16 status; pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_STATUS, &status); if ((!PCI_POSSIBLE_ERROR(status)) && (status & PCI_EXP_DPC_STATUS_TRIGGER)) return false; if (test_bit(PCI_DPC_RECOVERING, &pdev->priv_flags)) return false; return true; } /** * pci_dpc_recovered - whether DPC triggered and has recovered successfully * @pdev: PCI device * * Return true if DPC was triggered for @pdev and has recovered successfully. * Wait for recovery if it hasn't completed yet. Called from the PCIe hotplug * driver to recognize and ignore Link Down/Up events caused by DPC. */ bool pci_dpc_recovered(struct pci_dev *pdev) { struct pci_host_bridge *host; if (!pdev->dpc_cap) return false; /* * Synchronization between hotplug and DPC is not supported * if DPC is owned by firmware and EDR is not enabled. */ host = pci_find_host_bridge(pdev->bus); if (!host->native_dpc && !IS_ENABLED(CONFIG_PCIE_EDR)) return false; /* * Need a timeout in case DPC never completes due to failure of * dpc_wait_rp_inactive(). The spec doesn't mandate a time limit, * but reports indicate that DPC completes within 4 seconds. */ wait_event_timeout(dpc_completed_waitqueue, dpc_completed(pdev), msecs_to_jiffies(4000)); return test_and_clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags); } #endif /* CONFIG_HOTPLUG_PCI_PCIE */ static int dpc_wait_rp_inactive(struct pci_dev *pdev) { unsigned long timeout = jiffies + HZ; u16 cap = pdev->dpc_cap, status; pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); while (status & PCI_EXP_DPC_RP_BUSY && !time_after(jiffies, timeout)) { msleep(10); pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); } if (status & PCI_EXP_DPC_RP_BUSY) { pci_warn(pdev, "root port still busy\n"); return -EBUSY; } return 0; } pci_ers_result_t dpc_reset_link(struct pci_dev *pdev) { pci_ers_result_t ret; u16 cap; set_bit(PCI_DPC_RECOVERING, &pdev->priv_flags); /* * DPC disables the Link automatically in hardware, so it has * already been reset by the time we get here. */ cap = pdev->dpc_cap; /* * Wait until the Link is inactive, then clear DPC Trigger Status * to allow the Port to leave DPC. */ if (!pcie_wait_for_link(pdev, false)) pci_info(pdev, "Data Link Layer Link Active not cleared in 1000 msec\n"); if (pdev->dpc_rp_extensions && dpc_wait_rp_inactive(pdev)) { clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags); ret = PCI_ERS_RESULT_DISCONNECT; goto out; } pci_write_config_word(pdev, cap + PCI_EXP_DPC_STATUS, PCI_EXP_DPC_STATUS_TRIGGER); if (pci_bridge_wait_for_secondary_bus(pdev, "DPC")) { clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags); ret = PCI_ERS_RESULT_DISCONNECT; } else { set_bit(PCI_DPC_RECOVERED, &pdev->priv_flags); ret = PCI_ERS_RESULT_RECOVERED; } out: clear_bit(PCI_DPC_RECOVERING, &pdev->priv_flags); wake_up_all(&dpc_completed_waitqueue); return ret; } static void dpc_process_rp_pio_error(struct pci_dev *pdev) { u16 cap = pdev->dpc_cap, dpc_status, first_error; u32 status, mask, sev, syserr, exc, log, prefix; struct pcie_tlp_log tlp_log; int i; pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_STATUS, &status); pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_MASK, &mask); pci_err(pdev, "rp_pio_status: %#010x, rp_pio_mask: %#010x\n", status, mask); pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_SEVERITY, &sev); pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_SYSERROR, &syserr); pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_EXCEPTION, &exc); pci_err(pdev, "RP PIO severity=%#010x, syserror=%#010x, exception=%#010x\n", sev, syserr, exc); /* Get First Error Pointer */ pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &dpc_status); first_error = FIELD_GET(PCI_EXP_DPC_RP_PIO_FEP, dpc_status); for (i = 0; i < ARRAY_SIZE(rp_pio_error_string); i++) { if ((status & ~mask) & (1 << i)) pci_err(pdev, "[%2d] %s%s\n", i, rp_pio_error_string[i], first_error == i ? " (First)" : ""); } if (pdev->dpc_rp_log_size < 4) goto clear_status; pcie_read_tlp_log(pdev, cap + PCI_EXP_DPC_RP_PIO_HEADER_LOG, &tlp_log); pci_err(pdev, "TLP Header: %#010x %#010x %#010x %#010x\n", tlp_log.dw[0], tlp_log.dw[1], tlp_log.dw[2], tlp_log.dw[3]); if (pdev->dpc_rp_log_size < 5) goto clear_status; pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_IMPSPEC_LOG, &log); pci_err(pdev, "RP PIO ImpSpec Log %#010x\n", log); for (i = 0; i < pdev->dpc_rp_log_size - 5; i++) { pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_TLPPREFIX_LOG + i * 4, &prefix); pci_err(pdev, "TLP Prefix Header: dw%d, %#010x\n", i, prefix); } clear_status: pci_write_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_STATUS, status); } static int dpc_get_aer_uncorrect_severity(struct pci_dev *dev, struct aer_err_info *info) { int pos = dev->aer_cap; u32 status, mask, sev; pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_STATUS, &status); pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_MASK, &mask); status &= ~mask; if (!status) return 0; pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_SEVER, &sev); status &= sev; if (status) info->severity = AER_FATAL; else info->severity = AER_NONFATAL; return 1; } void dpc_process_error(struct pci_dev *pdev) { u16 cap = pdev->dpc_cap, status, source, reason, ext_reason; struct aer_err_info info; pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); pci_read_config_word(pdev, cap + PCI_EXP_DPC_SOURCE_ID, &source); pci_info(pdev, "containment event, status:%#06x source:%#06x\n", status, source); reason = status & PCI_EXP_DPC_STATUS_TRIGGER_RSN; ext_reason = status & PCI_EXP_DPC_STATUS_TRIGGER_RSN_EXT; pci_warn(pdev, "%s detected\n", (reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_UNCOR) ? "unmasked uncorrectable error" : (reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_NFE) ? "ERR_NONFATAL" : (reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_FE) ? "ERR_FATAL" : (ext_reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_RP_PIO) ? "RP PIO error" : (ext_reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_SW_TRIGGER) ? "software trigger" : "reserved error"); /* show RP PIO error detail information */ if (pdev->dpc_rp_extensions && reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_IN_EXT && ext_reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_RP_PIO) dpc_process_rp_pio_error(pdev); else if (reason == PCI_EXP_DPC_STATUS_TRIGGER_RSN_UNCOR && dpc_get_aer_uncorrect_severity(pdev, &info) && aer_get_device_error_info(pdev, &info)) { aer_print_error(pdev, &info); pci_aer_clear_nonfatal_status(pdev); pci_aer_clear_fatal_status(pdev); } } static void pci_clear_surpdn_errors(struct pci_dev *pdev) { if (pdev->dpc_rp_extensions) pci_write_config_dword(pdev, pdev->dpc_cap + PCI_EXP_DPC_RP_PIO_STATUS, ~0); /* * In practice, Surprise Down errors have been observed to also set * error bits in the Status Register as well as the Fatal Error * Detected bit in the Device Status Register. */ pci_write_config_word(pdev, PCI_STATUS, 0xffff); pcie_capability_write_word(pdev, PCI_EXP_DEVSTA, PCI_EXP_DEVSTA_FED); } static void dpc_handle_surprise_removal(struct pci_dev *pdev) { if (!pcie_wait_for_link(pdev, false)) { pci_info(pdev, "Data Link Layer Link Active not cleared in 1000 msec\n"); goto out; } if (pdev->dpc_rp_extensions && dpc_wait_rp_inactive(pdev)) goto out; pci_aer_raw_clear_status(pdev); pci_clear_surpdn_errors(pdev); pci_write_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_STATUS, PCI_EXP_DPC_STATUS_TRIGGER); out: clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags); wake_up_all(&dpc_completed_waitqueue); } static bool dpc_is_surprise_removal(struct pci_dev *pdev) { u16 status; if (!pdev->is_hotplug_bridge) return false; if (pci_read_config_word(pdev, pdev->aer_cap + PCI_ERR_UNCOR_STATUS, &status)) return false; return status & PCI_ERR_UNC_SURPDN; } static irqreturn_t dpc_handler(int irq, void *context) { struct pci_dev *pdev = context; /* * According to PCIe r6.0 sec 6.7.6, errors are an expected side effect * of async removal and should be ignored by software. */ if (dpc_is_surprise_removal(pdev)) { dpc_handle_surprise_removal(pdev); return IRQ_HANDLED; } dpc_process_error(pdev); /* We configure DPC so it only triggers on ERR_FATAL */ pcie_do_recovery(pdev, pci_channel_io_frozen, dpc_reset_link); return IRQ_HANDLED; } static irqreturn_t dpc_irq(int irq, void *context) { struct pci_dev *pdev = context; u16 cap = pdev->dpc_cap, status; pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); if (!(status & PCI_EXP_DPC_STATUS_INTERRUPT) || PCI_POSSIBLE_ERROR(status)) return IRQ_NONE; pci_write_config_word(pdev, cap + PCI_EXP_DPC_STATUS, PCI_EXP_DPC_STATUS_INTERRUPT); if (status & PCI_EXP_DPC_STATUS_TRIGGER) return IRQ_WAKE_THREAD; return IRQ_HANDLED; } void pci_dpc_init(struct pci_dev *pdev) { u16 cap; pdev->dpc_cap = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_DPC); if (!pdev->dpc_cap) return; pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CAP, &cap); if (!(cap & PCI_EXP_DPC_CAP_RP_EXT)) return; pdev->dpc_rp_extensions = true; /* Quirks may set dpc_rp_log_size if device or firmware is buggy */ if (!pdev->dpc_rp_log_size) { pdev->dpc_rp_log_size = FIELD_GET(PCI_EXP_DPC_RP_PIO_LOG_SIZE, cap); if (pdev->dpc_rp_log_size < 4 || pdev->dpc_rp_log_size > 9) { pci_err(pdev, "RP PIO log size %u is invalid\n", pdev->dpc_rp_log_size); pdev->dpc_rp_log_size = 0; } } } static void dpc_enable(struct pcie_device *dev) { struct pci_dev *pdev = dev->port; int dpc = pdev->dpc_cap; u16 ctl; /* * Clear DPC Interrupt Status so we don't get an interrupt for an * old event when setting DPC Interrupt Enable. */ pci_write_config_word(pdev, dpc + PCI_EXP_DPC_STATUS, PCI_EXP_DPC_STATUS_INTERRUPT); pci_read_config_word(pdev, dpc + PCI_EXP_DPC_CTL, &ctl); ctl &= ~PCI_EXP_DPC_CTL_EN_MASK; ctl |= PCI_EXP_DPC_CTL_EN_FATAL | PCI_EXP_DPC_CTL_INT_EN; pci_write_config_word(pdev, dpc + PCI_EXP_DPC_CTL, ctl); } static void dpc_disable(struct pcie_device *dev) { struct pci_dev *pdev = dev->port; int dpc = pdev->dpc_cap; u16 ctl; /* Disable DPC triggering and DPC interrupts */ pci_read_config_word(pdev, dpc + PCI_EXP_DPC_CTL, &ctl); ctl &= ~(PCI_EXP_DPC_CTL_EN_FATAL | PCI_EXP_DPC_CTL_INT_EN); pci_write_config_word(pdev, dpc + PCI_EXP_DPC_CTL, ctl); } #define FLAG(x, y) (((x) & (y)) ? '+' : '-') static int dpc_probe(struct pcie_device *dev) { struct pci_dev *pdev = dev->port; struct device *device = &dev->device; int status; u16 cap; if (!pcie_aer_is_native(pdev) && !pcie_ports_dpc_native) return -ENOTSUPP; status = devm_request_threaded_irq(device, dev->irq, dpc_irq, dpc_handler, IRQF_SHARED, "pcie-dpc", pdev); if (status) { pci_warn(pdev, "request IRQ%d failed: %d\n", dev->irq, status); return status; } pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CAP, &cap); dpc_enable(dev); pci_info(pdev, "enabled with IRQ %d\n", dev->irq); pci_info(pdev, "error containment capabilities: Int Msg #%d, RPExt%c PoisonedTLP%c SwTrigger%c RP PIO Log %d, DL_ActiveErr%c\n", cap & PCI_EXP_DPC_IRQ, FLAG(cap, PCI_EXP_DPC_CAP_RP_EXT), FLAG(cap, PCI_EXP_DPC_CAP_POISONED_TLP), FLAG(cap, PCI_EXP_DPC_CAP_SW_TRIGGER), pdev->dpc_rp_log_size, FLAG(cap, PCI_EXP_DPC_CAP_DL_ACTIVE)); pci_add_ext_cap_save_buffer(pdev, PCI_EXT_CAP_ID_DPC, sizeof(u16)); return status; } static int dpc_suspend(struct pcie_device *dev) { dpc_disable(dev); return 0; } static int dpc_resume(struct pcie_device *dev) { dpc_enable(dev); return 0; } static void dpc_remove(struct pcie_device *dev) { dpc_disable(dev); } static struct pcie_port_service_driver dpcdriver = { .name = "dpc", .port_type = PCIE_ANY_PORT, .service = PCIE_PORT_SERVICE_DPC, .probe = dpc_probe, .suspend = dpc_suspend, .resume = dpc_resume, .remove = dpc_remove, }; int __init pcie_dpc_init(void) { return pcie_port_service_register(&dpcdriver); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1