Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Hao Wu | 629 | 33.28% | 2 | 11.11% |
Zhang Yi | 465 | 24.60% | 1 | 5.56% |
Matthew Gerlach | 463 | 24.50% | 4 | 22.22% |
Xu Yilun | 266 | 14.07% | 5 | 27.78% |
Martin Hundeböll | 20 | 1.06% | 1 | 5.56% |
Russ Weight | 14 | 0.74% | 1 | 5.56% |
Christophe Jaillet | 11 | 0.58% | 1 | 5.56% |
Björn Helgaas | 11 | 0.58% | 1 | 5.56% |
Peter Colberg | 8 | 0.42% | 1 | 5.56% |
Xiongfeng Wang | 3 | 0.16% | 1 | 5.56% |
Total | 1890 | 18 |
// SPDX-License-Identifier: GPL-2.0 /* * Driver for FPGA Device Feature List (DFL) PCIe device * * Copyright (C) 2017-2018 Intel Corporation, Inc. * * Authors: * Zhang Yi <Yi.Z.Zhang@intel.com> * Xiao Guangrong <guangrong.xiao@linux.intel.com> * Joseph Grecco <joe.grecco@intel.com> * Enno Luebbers <enno.luebbers@intel.com> * Tim Whisonant <tim.whisonant@intel.com> * Ananda Ravuri <ananda.ravuri@intel.com> * Henry Mitchel <henry.mitchel@intel.com> */ #include <linux/pci.h> #include <linux/dma-mapping.h> #include <linux/types.h> #include <linux/kernel.h> #include <linux/module.h> #include <linux/stddef.h> #include <linux/errno.h> #include "dfl.h" #define DRV_VERSION "0.8" #define DRV_NAME "dfl-pci" #define PCI_VSEC_ID_INTEL_DFLS 0x43 #define PCI_VNDR_DFLS_CNT 0x8 #define PCI_VNDR_DFLS_RES 0xc #define PCI_VNDR_DFLS_RES_BAR_MASK GENMASK(2, 0) #define PCI_VNDR_DFLS_RES_OFF_MASK GENMASK(31, 3) struct cci_drvdata { struct dfl_fpga_cdev *cdev; /* container device */ }; static void __iomem *cci_pci_ioremap_bar0(struct pci_dev *pcidev) { if (pcim_iomap_regions(pcidev, BIT(0), DRV_NAME)) return NULL; return pcim_iomap_table(pcidev)[0]; } static int cci_pci_alloc_irq(struct pci_dev *pcidev) { int ret, nvec = pci_msix_vec_count(pcidev); if (nvec <= 0) { dev_dbg(&pcidev->dev, "fpga interrupt not supported\n"); return 0; } ret = pci_alloc_irq_vectors(pcidev, nvec, nvec, PCI_IRQ_MSIX); if (ret < 0) return ret; return nvec; } static void cci_pci_free_irq(struct pci_dev *pcidev) { pci_free_irq_vectors(pcidev); } /* PCI Device ID */ #define PCIE_DEVICE_ID_PF_INT_5_X 0xBCBD #define PCIE_DEVICE_ID_PF_INT_6_X 0xBCC0 #define PCIE_DEVICE_ID_PF_DSC_1_X 0x09C4 #define PCIE_DEVICE_ID_INTEL_PAC_N3000 0x0B30 #define PCIE_DEVICE_ID_INTEL_PAC_D5005 0x0B2B #define PCIE_DEVICE_ID_SILICOM_PAC_N5010 0x1000 #define PCIE_DEVICE_ID_SILICOM_PAC_N5011 0x1001 #define PCIE_DEVICE_ID_INTEL_DFL 0xbcce /* PCI Subdevice ID for PCIE_DEVICE_ID_INTEL_DFL */ #define PCIE_SUBDEVICE_ID_INTEL_D5005 0x138d #define PCIE_SUBDEVICE_ID_INTEL_N6000 0x1770 #define PCIE_SUBDEVICE_ID_INTEL_N6001 0x1771 #define PCIE_SUBDEVICE_ID_INTEL_C6100 0x17d4 /* VF Device */ #define PCIE_DEVICE_ID_VF_INT_5_X 0xBCBF #define PCIE_DEVICE_ID_VF_INT_6_X 0xBCC1 #define PCIE_DEVICE_ID_VF_DSC_1_X 0x09C5 #define PCIE_DEVICE_ID_INTEL_PAC_D5005_VF 0x0B2C #define PCIE_DEVICE_ID_INTEL_DFL_VF 0xbccf static struct pci_device_id cci_pcie_id_tbl[] = { {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_INT_5_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_INT_5_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_INT_6_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_INT_6_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_PF_DSC_1_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_VF_DSC_1_X),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_PAC_N3000),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_PAC_D5005),}, {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_PAC_D5005_VF),}, {PCI_DEVICE(PCI_VENDOR_ID_SILICOM_DENMARK, PCIE_DEVICE_ID_SILICOM_PAC_N5010),}, {PCI_DEVICE(PCI_VENDOR_ID_SILICOM_DENMARK, PCIE_DEVICE_ID_SILICOM_PAC_N5011),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_D5005),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_N6000),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL_VF, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_N6000),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_N6001),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL_VF, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_N6001),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_C6100),}, {PCI_DEVICE_SUB(PCI_VENDOR_ID_INTEL, PCIE_DEVICE_ID_INTEL_DFL_VF, PCI_VENDOR_ID_INTEL, PCIE_SUBDEVICE_ID_INTEL_C6100),}, {0,} }; MODULE_DEVICE_TABLE(pci, cci_pcie_id_tbl); static int cci_init_drvdata(struct pci_dev *pcidev) { struct cci_drvdata *drvdata; drvdata = devm_kzalloc(&pcidev->dev, sizeof(*drvdata), GFP_KERNEL); if (!drvdata) return -ENOMEM; pci_set_drvdata(pcidev, drvdata); return 0; } static void cci_remove_feature_devs(struct pci_dev *pcidev) { struct cci_drvdata *drvdata = pci_get_drvdata(pcidev); /* remove all children feature devices */ dfl_fpga_feature_devs_remove(drvdata->cdev); cci_pci_free_irq(pcidev); } static int *cci_pci_create_irq_table(struct pci_dev *pcidev, unsigned int nvec) { unsigned int i; int *table; table = kcalloc(nvec, sizeof(int), GFP_KERNEL); if (!table) return table; for (i = 0; i < nvec; i++) table[i] = pci_irq_vector(pcidev, i); return table; } static int find_dfls_by_vsec(struct pci_dev *pcidev, struct dfl_fpga_enum_info *info) { u32 bir, offset, dfl_cnt, dfl_res; int dfl_res_off, i, bars, voff; resource_size_t start, len; voff = pci_find_vsec_capability(pcidev, PCI_VENDOR_ID_INTEL, PCI_VSEC_ID_INTEL_DFLS); if (!voff) { dev_dbg(&pcidev->dev, "%s no DFL VSEC found\n", __func__); return -ENODEV; } dfl_cnt = 0; pci_read_config_dword(pcidev, voff + PCI_VNDR_DFLS_CNT, &dfl_cnt); if (dfl_cnt > PCI_STD_NUM_BARS) { dev_err(&pcidev->dev, "%s too many DFLs %d > %d\n", __func__, dfl_cnt, PCI_STD_NUM_BARS); return -EINVAL; } dfl_res_off = voff + PCI_VNDR_DFLS_RES; if (dfl_res_off + (dfl_cnt * sizeof(u32)) > PCI_CFG_SPACE_EXP_SIZE) { dev_err(&pcidev->dev, "%s DFL VSEC too big for PCIe config space\n", __func__); return -EINVAL; } for (i = 0, bars = 0; i < dfl_cnt; i++, dfl_res_off += sizeof(u32)) { dfl_res = GENMASK(31, 0); pci_read_config_dword(pcidev, dfl_res_off, &dfl_res); bir = dfl_res & PCI_VNDR_DFLS_RES_BAR_MASK; if (bir >= PCI_STD_NUM_BARS) { dev_err(&pcidev->dev, "%s bad bir number %d\n", __func__, bir); return -EINVAL; } if (bars & BIT(bir)) { dev_err(&pcidev->dev, "%s DFL for BAR %d already specified\n", __func__, bir); return -EINVAL; } bars |= BIT(bir); len = pci_resource_len(pcidev, bir); offset = dfl_res & PCI_VNDR_DFLS_RES_OFF_MASK; if (offset >= len) { dev_err(&pcidev->dev, "%s bad offset %u >= %pa\n", __func__, offset, &len); return -EINVAL; } dev_dbg(&pcidev->dev, "%s BAR %d offset 0x%x\n", __func__, bir, offset); len -= offset; start = pci_resource_start(pcidev, bir) + offset; dfl_fpga_enum_info_add_dfl(info, start, len); } return 0; } /* default method of finding dfls starting at offset 0 of bar 0 */ static int find_dfls_by_default(struct pci_dev *pcidev, struct dfl_fpga_enum_info *info) { int port_num, bar, i, ret = 0; resource_size_t start, len; void __iomem *base; u32 offset; u64 v; /* start to find Device Feature List from Bar 0 */ base = cci_pci_ioremap_bar0(pcidev); if (!base) return -ENOMEM; /* * PF device has FME and Ports/AFUs, and VF device only has one * Port/AFU. Check them and add related "Device Feature List" info * for the next step enumeration. */ if (dfl_feature_is_fme(base)) { start = pci_resource_start(pcidev, 0); len = pci_resource_len(pcidev, 0); dfl_fpga_enum_info_add_dfl(info, start, len); /* * find more Device Feature Lists (e.g. Ports) per information * indicated by FME module. */ v = readq(base + FME_HDR_CAP); port_num = FIELD_GET(FME_CAP_NUM_PORTS, v); WARN_ON(port_num > MAX_DFL_FPGA_PORT_NUM); for (i = 0; i < port_num; i++) { v = readq(base + FME_HDR_PORT_OFST(i)); /* skip ports which are not implemented. */ if (!(v & FME_PORT_OFST_IMP)) continue; /* * add Port's Device Feature List information for next * step enumeration. */ bar = FIELD_GET(FME_PORT_OFST_BAR_ID, v); offset = FIELD_GET(FME_PORT_OFST_DFH_OFST, v); if (bar == FME_PORT_OFST_BAR_SKIP) { continue; } else if (bar >= PCI_STD_NUM_BARS) { dev_err(&pcidev->dev, "bad BAR %d for port %d\n", bar, i); ret = -EINVAL; break; } start = pci_resource_start(pcidev, bar) + offset; len = pci_resource_len(pcidev, bar) - offset; dfl_fpga_enum_info_add_dfl(info, start, len); } } else if (dfl_feature_is_port(base)) { start = pci_resource_start(pcidev, 0); len = pci_resource_len(pcidev, 0); dfl_fpga_enum_info_add_dfl(info, start, len); } else { ret = -ENODEV; } /* release I/O mappings for next step enumeration */ pcim_iounmap_regions(pcidev, BIT(0)); return ret; } /* enumerate feature devices under pci device */ static int cci_enumerate_feature_devs(struct pci_dev *pcidev) { struct cci_drvdata *drvdata = pci_get_drvdata(pcidev); struct dfl_fpga_enum_info *info; struct dfl_fpga_cdev *cdev; int nvec, ret = 0; int *irq_table; /* allocate enumeration info via pci_dev */ info = dfl_fpga_enum_info_alloc(&pcidev->dev); if (!info) return -ENOMEM; /* add irq info for enumeration if the device support irq */ nvec = cci_pci_alloc_irq(pcidev); if (nvec < 0) { dev_err(&pcidev->dev, "Fail to alloc irq %d.\n", nvec); ret = nvec; goto enum_info_free_exit; } else if (nvec) { irq_table = cci_pci_create_irq_table(pcidev, nvec); if (!irq_table) { ret = -ENOMEM; goto irq_free_exit; } ret = dfl_fpga_enum_info_add_irq(info, nvec, irq_table); kfree(irq_table); if (ret) goto irq_free_exit; } ret = find_dfls_by_vsec(pcidev, info); if (ret == -ENODEV) ret = find_dfls_by_default(pcidev, info); if (ret) goto irq_free_exit; /* start enumeration with prepared enumeration information */ cdev = dfl_fpga_feature_devs_enumerate(info); if (IS_ERR(cdev)) { dev_err(&pcidev->dev, "Enumeration failure\n"); ret = PTR_ERR(cdev); goto irq_free_exit; } drvdata->cdev = cdev; irq_free_exit: if (ret) cci_pci_free_irq(pcidev); enum_info_free_exit: dfl_fpga_enum_info_free(info); return ret; } static int cci_pci_probe(struct pci_dev *pcidev, const struct pci_device_id *pcidevid) { int ret; ret = pcim_enable_device(pcidev); if (ret < 0) { dev_err(&pcidev->dev, "Failed to enable device %d.\n", ret); return ret; } pci_set_master(pcidev); ret = dma_set_mask_and_coherent(&pcidev->dev, DMA_BIT_MASK(64)); if (ret) ret = dma_set_mask_and_coherent(&pcidev->dev, DMA_BIT_MASK(32)); if (ret) { dev_err(&pcidev->dev, "No suitable DMA support available.\n"); return ret; } ret = cci_init_drvdata(pcidev); if (ret) { dev_err(&pcidev->dev, "Fail to init drvdata %d.\n", ret); return ret; } ret = cci_enumerate_feature_devs(pcidev); if (ret) { dev_err(&pcidev->dev, "enumeration failure %d.\n", ret); return ret; } return 0; } static int cci_pci_sriov_configure(struct pci_dev *pcidev, int num_vfs) { struct cci_drvdata *drvdata = pci_get_drvdata(pcidev); struct dfl_fpga_cdev *cdev = drvdata->cdev; if (!num_vfs) { /* * disable SRIOV and then put released ports back to default * PF access mode. */ pci_disable_sriov(pcidev); dfl_fpga_cdev_config_ports_pf(cdev); } else { int ret; /* * before enable SRIOV, put released ports into VF access mode * first of all. */ ret = dfl_fpga_cdev_config_ports_vf(cdev, num_vfs); if (ret) return ret; ret = pci_enable_sriov(pcidev, num_vfs); if (ret) { dfl_fpga_cdev_config_ports_pf(cdev); return ret; } } return num_vfs; } static void cci_pci_remove(struct pci_dev *pcidev) { if (dev_is_pf(&pcidev->dev)) cci_pci_sriov_configure(pcidev, 0); cci_remove_feature_devs(pcidev); } static struct pci_driver cci_pci_driver = { .name = DRV_NAME, .id_table = cci_pcie_id_tbl, .probe = cci_pci_probe, .remove = cci_pci_remove, .sriov_configure = cci_pci_sriov_configure, }; module_pci_driver(cci_pci_driver); MODULE_DESCRIPTION("FPGA DFL PCIe Device Driver"); MODULE_AUTHOR("Intel Corporation"); MODULE_LICENSE("GPL v2");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1