Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Atish Patra | 511 | 76.50% | 2 | 22.22% |
Alexandre Ghiti | 105 | 15.72% | 2 | 22.22% |
Vadim Shakirov | 30 | 4.49% | 2 | 22.22% |
Jonathan Cameron | 11 | 1.65% | 1 | 11.11% |
Sergey Matyukevich | 10 | 1.50% | 1 | 11.11% |
Conor Dooley | 1 | 0.15% | 1 | 11.11% |
Total | 668 | 9 |
// SPDX-License-Identifier: GPL-2.0 /* * RISC-V performance counter support. * * Copyright (C) 2021 Western Digital Corporation or its affiliates. * * This implementation is based on old RISC-V perf and ARM perf event code * which are in turn based on sparc64 and x86 code. */ #include <linux/mod_devicetable.h> #include <linux/perf/riscv_pmu.h> #include <linux/platform_device.h> #define RISCV_PMU_LEGACY_CYCLE 0 #define RISCV_PMU_LEGACY_INSTRET 2 static bool pmu_init_done; static int pmu_legacy_ctr_get_idx(struct perf_event *event) { struct perf_event_attr *attr = &event->attr; if (event->attr.type != PERF_TYPE_HARDWARE) return -EOPNOTSUPP; if (attr->config == PERF_COUNT_HW_CPU_CYCLES) return RISCV_PMU_LEGACY_CYCLE; else if (attr->config == PERF_COUNT_HW_INSTRUCTIONS) return RISCV_PMU_LEGACY_INSTRET; else return -EOPNOTSUPP; } /* For legacy config & counter index are same */ static int pmu_legacy_event_map(struct perf_event *event, u64 *config) { return pmu_legacy_ctr_get_idx(event); } /* cycle & instret are always 64 bit, one bit less according to SBI spec */ static int pmu_legacy_ctr_get_width(int idx) { return 63; } static u64 pmu_legacy_read_ctr(struct perf_event *event) { struct hw_perf_event *hwc = &event->hw; int idx = hwc->idx; u64 val; if (idx == RISCV_PMU_LEGACY_CYCLE) { val = riscv_pmu_ctr_read_csr(CSR_CYCLE); if (IS_ENABLED(CONFIG_32BIT)) val = (u64)riscv_pmu_ctr_read_csr(CSR_CYCLEH) << 32 | val; } else if (idx == RISCV_PMU_LEGACY_INSTRET) { val = riscv_pmu_ctr_read_csr(CSR_INSTRET); if (IS_ENABLED(CONFIG_32BIT)) val = ((u64)riscv_pmu_ctr_read_csr(CSR_INSTRETH)) << 32 | val; } else return 0; return val; } static void pmu_legacy_ctr_start(struct perf_event *event, u64 ival) { struct hw_perf_event *hwc = &event->hw; u64 initial_val = pmu_legacy_read_ctr(event); /** * The legacy method doesn't really have a start/stop method. * It also can not update the counter with a initial value. * But we still need to set the prev_count so that read() can compute * the delta. Just use the current counter value to set the prev_count. */ local64_set(&hwc->prev_count, initial_val); } static uint8_t pmu_legacy_csr_index(struct perf_event *event) { return event->hw.idx; } static void pmu_legacy_event_mapped(struct perf_event *event, struct mm_struct *mm) { if (event->attr.config != PERF_COUNT_HW_CPU_CYCLES && event->attr.config != PERF_COUNT_HW_INSTRUCTIONS) return; event->hw.flags |= PERF_EVENT_FLAG_USER_READ_CNT; } static void pmu_legacy_event_unmapped(struct perf_event *event, struct mm_struct *mm) { if (event->attr.config != PERF_COUNT_HW_CPU_CYCLES && event->attr.config != PERF_COUNT_HW_INSTRUCTIONS) return; event->hw.flags &= ~PERF_EVENT_FLAG_USER_READ_CNT; } /* * This is just a simple implementation to allow legacy implementations * compatible with new RISC-V PMU driver framework. * This driver only allows reading two counters i.e CYCLE & INSTRET. * However, it can not start or stop the counter. Thus, it is not very useful * will be removed in future. */ static void pmu_legacy_init(struct riscv_pmu *pmu) { pr_info("Legacy PMU implementation is available\n"); pmu->cmask = BIT(RISCV_PMU_LEGACY_CYCLE) | BIT(RISCV_PMU_LEGACY_INSTRET); pmu->ctr_start = pmu_legacy_ctr_start; pmu->ctr_stop = NULL; pmu->event_map = pmu_legacy_event_map; pmu->ctr_get_idx = pmu_legacy_ctr_get_idx; pmu->ctr_get_width = pmu_legacy_ctr_get_width; pmu->ctr_clear_idx = NULL; pmu->ctr_read = pmu_legacy_read_ctr; pmu->event_mapped = pmu_legacy_event_mapped; pmu->event_unmapped = pmu_legacy_event_unmapped; pmu->csr_index = pmu_legacy_csr_index; pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; perf_pmu_register(&pmu->pmu, "cpu", PERF_TYPE_RAW); } static int pmu_legacy_device_probe(struct platform_device *pdev) { struct riscv_pmu *pmu = NULL; pmu = riscv_pmu_alloc(); if (!pmu) return -ENOMEM; pmu->pmu.parent = &pdev->dev; pmu_legacy_init(pmu); return 0; } static struct platform_driver pmu_legacy_driver = { .probe = pmu_legacy_device_probe, .driver = { .name = RISCV_PMU_LEGACY_PDEV_NAME, }, }; static int __init riscv_pmu_legacy_devinit(void) { int ret; struct platform_device *pdev; if (likely(pmu_init_done)) return 0; ret = platform_driver_register(&pmu_legacy_driver); if (ret) return ret; pdev = platform_device_register_simple(RISCV_PMU_LEGACY_PDEV_NAME, -1, NULL, 0); if (IS_ERR(pdev)) { platform_driver_unregister(&pmu_legacy_driver); return PTR_ERR(pdev); } return ret; } late_initcall(riscv_pmu_legacy_devinit); void riscv_pmu_legacy_skip_init(void) { pmu_init_done = true; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1