Contributors: 1
Author Tokens Token Proportion Commits Commit Proportion
Junhao He 1951 100.00% 1 100.00%
Total 1951 1


// SPDX-License-Identifier: GPL-2.0-only
/*
 * HiSilicon SoC MN uncore Hardware event counters support
 *
 * Copyright (c) 2025 HiSilicon Technologies Co., Ltd.
 */
#include <linux/cpuhotplug.h>
#include <linux/interrupt.h>
#include <linux/iopoll.h>
#include <linux/irq.h>
#include <linux/list.h>
#include <linux/mod_devicetable.h>
#include <linux/property.h>

#include "hisi_uncore_pmu.h"

/* Dynamic CPU hotplug state used by MN PMU */
static enum cpuhp_state hisi_mn_pmu_online;

/* MN register definition */
#define HISI_MN_DYNAMIC_CTRL_REG	0x400
#define   HISI_MN_DYNAMIC_CTRL_EN	BIT(0)
#define HISI_MN_PERF_CTRL_REG		0x408
#define   HISI_MN_PERF_CTRL_EN		BIT(6)
#define HISI_MN_INT_MASK_REG		0x800
#define HISI_MN_INT_STATUS_REG		0x808
#define HISI_MN_INT_CLEAR_REG		0x80C
#define HISI_MN_EVENT_CTRL_REG		0x1C00
#define HISI_MN_VERSION_REG		0x1C04
#define HISI_MN_EVTYPE0_REG		0x1d00
#define   HISI_MN_EVTYPE_MASK		GENMASK(7, 0)
#define HISI_MN_CNTR0_REG		0x1e00
#define HISI_MN_EVTYPE_REGn(evtype0, n)	((evtype0) + (n) * 4)
#define HISI_MN_CNTR_REGn(cntr0, n)	((cntr0) + (n) * 8)

#define HISI_MN_NR_COUNTERS		4
#define HISI_MN_TIMEOUT_US		500U

struct hisi_mn_pmu_regs {
	u32 version;
	u32 dyn_ctrl;
	u32 perf_ctrl;
	u32 int_mask;
	u32 int_clear;
	u32 int_status;
	u32 event_ctrl;
	u32 event_type0;
	u32 event_cntr0;
};

/*
 * Each event request takes a certain amount of time to complete. If
 * we counting the latency related event, we need to wait for the all
 * requests complete. Otherwise, the value of counter is slightly larger.
 */
static void hisi_mn_pmu_counter_flush(struct hisi_pmu *mn_pmu)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	int ret;
	u32 val;

	val = readl(mn_pmu->base + reg_info->dyn_ctrl);
	val |= HISI_MN_DYNAMIC_CTRL_EN;
	writel(val, mn_pmu->base + reg_info->dyn_ctrl);

	ret = readl_poll_timeout_atomic(mn_pmu->base + reg_info->dyn_ctrl,
					val, !(val & HISI_MN_DYNAMIC_CTRL_EN),
					1, HISI_MN_TIMEOUT_US);
	if (ret)
		dev_warn(mn_pmu->dev, "Counter flush timeout\n");
}

static u64 hisi_mn_pmu_read_counter(struct hisi_pmu *mn_pmu,
				    struct hw_perf_event *hwc)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;

	return readq(mn_pmu->base + HISI_MN_CNTR_REGn(reg_info->event_cntr0, hwc->idx));
}

static void hisi_mn_pmu_write_counter(struct hisi_pmu *mn_pmu,
				      struct hw_perf_event *hwc, u64 val)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;

	writeq(val, mn_pmu->base + HISI_MN_CNTR_REGn(reg_info->event_cntr0, hwc->idx));
}

static void hisi_mn_pmu_write_evtype(struct hisi_pmu *mn_pmu, int idx, u32 type)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	/*
	 * Select the appropriate event select register.
	 * There are 2 32-bit event select registers for the
	 * 8 hardware counters, each event code is 8-bit wide.
	 */
	val = readl(mn_pmu->base + HISI_MN_EVTYPE_REGn(reg_info->event_type0, idx / 4));
	val &= ~(HISI_MN_EVTYPE_MASK << HISI_PMU_EVTYPE_SHIFT(idx));
	val |= (type << HISI_PMU_EVTYPE_SHIFT(idx));
	writel(val, mn_pmu->base + HISI_MN_EVTYPE_REGn(reg_info->event_type0, idx / 4));
}

static void hisi_mn_pmu_start_counters(struct hisi_pmu *mn_pmu)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->perf_ctrl);
	val |= HISI_MN_PERF_CTRL_EN;
	writel(val, mn_pmu->base + reg_info->perf_ctrl);
}

static void hisi_mn_pmu_stop_counters(struct hisi_pmu *mn_pmu)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->perf_ctrl);
	val &= ~HISI_MN_PERF_CTRL_EN;
	writel(val, mn_pmu->base + reg_info->perf_ctrl);

	hisi_mn_pmu_counter_flush(mn_pmu);
}

static void hisi_mn_pmu_enable_counter(struct hisi_pmu *mn_pmu,
				       struct hw_perf_event *hwc)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->event_ctrl);
	val |= BIT(hwc->idx);
	writel(val, mn_pmu->base + reg_info->event_ctrl);
}

static void hisi_mn_pmu_disable_counter(struct hisi_pmu *mn_pmu,
					struct hw_perf_event *hwc)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->event_ctrl);
	val &= ~BIT(hwc->idx);
	writel(val, mn_pmu->base + reg_info->event_ctrl);
}

static void hisi_mn_pmu_enable_counter_int(struct hisi_pmu *mn_pmu,
					   struct hw_perf_event *hwc)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->int_mask);
	val &= ~BIT(hwc->idx);
	writel(val, mn_pmu->base + reg_info->int_mask);
}

static void hisi_mn_pmu_disable_counter_int(struct hisi_pmu *mn_pmu,
					    struct hw_perf_event *hwc)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;
	u32 val;

	val = readl(mn_pmu->base + reg_info->int_mask);
	val |= BIT(hwc->idx);
	writel(val, mn_pmu->base + reg_info->int_mask);
}

static u32 hisi_mn_pmu_get_int_status(struct hisi_pmu *mn_pmu)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;

	return readl(mn_pmu->base + reg_info->int_status);
}

static void hisi_mn_pmu_clear_int_status(struct hisi_pmu *mn_pmu, int idx)
{
	struct hisi_mn_pmu_regs *reg_info = mn_pmu->dev_info->private;

	writel(BIT(idx), mn_pmu->base + reg_info->int_clear);
}

static struct attribute *hisi_mn_pmu_format_attr[] = {
	HISI_PMU_FORMAT_ATTR(event, "config:0-7"),
	NULL
};

static const struct attribute_group hisi_mn_pmu_format_group = {
	.name = "format",
	.attrs = hisi_mn_pmu_format_attr,
};

static struct attribute *hisi_mn_pmu_events_attr[] = {
	HISI_PMU_EVENT_ATTR(req_eobarrier_num,		0x00),
	HISI_PMU_EVENT_ATTR(req_ecbarrier_num,		0x01),
	HISI_PMU_EVENT_ATTR(req_dvmop_num,		0x02),
	HISI_PMU_EVENT_ATTR(req_dvmsync_num,		0x03),
	HISI_PMU_EVENT_ATTR(req_retry_num,		0x04),
	HISI_PMU_EVENT_ATTR(req_writenosnp_num,		0x05),
	HISI_PMU_EVENT_ATTR(req_readnosnp_num,		0x06),
	HISI_PMU_EVENT_ATTR(snp_dvm_num,		0x07),
	HISI_PMU_EVENT_ATTR(snp_dvmsync_num,		0x08),
	HISI_PMU_EVENT_ATTR(l3t_req_dvm_num,		0x09),
	HISI_PMU_EVENT_ATTR(l3t_req_dvmsync_num,	0x0A),
	HISI_PMU_EVENT_ATTR(mn_req_dvm_num,		0x0B),
	HISI_PMU_EVENT_ATTR(mn_req_dvmsync_num,		0x0C),
	HISI_PMU_EVENT_ATTR(pa_req_dvm_num,		0x0D),
	HISI_PMU_EVENT_ATTR(pa_req_dvmsync_num,		0x0E),
	HISI_PMU_EVENT_ATTR(snp_dvm_latency,		0x80),
	HISI_PMU_EVENT_ATTR(snp_dvmsync_latency,	0x81),
	HISI_PMU_EVENT_ATTR(l3t_req_dvm_latency,	0x82),
	HISI_PMU_EVENT_ATTR(l3t_req_dvmsync_latency,	0x83),
	HISI_PMU_EVENT_ATTR(mn_req_dvm_latency,		0x84),
	HISI_PMU_EVENT_ATTR(mn_req_dvmsync_latency,	0x85),
	HISI_PMU_EVENT_ATTR(pa_req_dvm_latency,		0x86),
	HISI_PMU_EVENT_ATTR(pa_req_dvmsync_latency,	0x87),
	NULL
};

static const struct attribute_group hisi_mn_pmu_events_group = {
	.name = "events",
	.attrs = hisi_mn_pmu_events_attr,
};

static const struct attribute_group *hisi_mn_pmu_attr_groups[] = {
	&hisi_mn_pmu_format_group,
	&hisi_mn_pmu_events_group,
	&hisi_pmu_cpumask_attr_group,
	&hisi_pmu_identifier_group,
	NULL
};

static const struct hisi_uncore_ops hisi_uncore_mn_ops = {
	.write_evtype		= hisi_mn_pmu_write_evtype,
	.get_event_idx		= hisi_uncore_pmu_get_event_idx,
	.start_counters		= hisi_mn_pmu_start_counters,
	.stop_counters		= hisi_mn_pmu_stop_counters,
	.enable_counter		= hisi_mn_pmu_enable_counter,
	.disable_counter	= hisi_mn_pmu_disable_counter,
	.enable_counter_int	= hisi_mn_pmu_enable_counter_int,
	.disable_counter_int	= hisi_mn_pmu_disable_counter_int,
	.write_counter		= hisi_mn_pmu_write_counter,
	.read_counter		= hisi_mn_pmu_read_counter,
	.get_int_status		= hisi_mn_pmu_get_int_status,
	.clear_int_status	= hisi_mn_pmu_clear_int_status,
};

static int hisi_mn_pmu_dev_init(struct platform_device *pdev,
				struct hisi_pmu *mn_pmu)
{
	struct hisi_mn_pmu_regs *reg_info;
	int ret;

	hisi_uncore_pmu_init_topology(mn_pmu, &pdev->dev);

	if (mn_pmu->topo.scl_id < 0)
		return dev_err_probe(&pdev->dev, -EINVAL,
				     "Failed to read MN scl id\n");

	if (mn_pmu->topo.index_id < 0)
		return dev_err_probe(&pdev->dev, -EINVAL,
				     "Failed to read MN index id\n");

	mn_pmu->base = devm_platform_ioremap_resource(pdev, 0);
	if (IS_ERR(mn_pmu->base))
		return dev_err_probe(&pdev->dev, PTR_ERR(mn_pmu->base),
				     "Failed to ioremap resource\n");

	ret = hisi_uncore_pmu_init_irq(mn_pmu, pdev);
	if (ret)
		return ret;

	mn_pmu->dev_info = device_get_match_data(&pdev->dev);
	if (!mn_pmu->dev_info)
		return -ENODEV;

	mn_pmu->pmu_events.attr_groups = mn_pmu->dev_info->attr_groups;
	mn_pmu->counter_bits = mn_pmu->dev_info->counter_bits;
	mn_pmu->check_event = mn_pmu->dev_info->check_event;
	mn_pmu->num_counters = HISI_MN_NR_COUNTERS;
	mn_pmu->ops = &hisi_uncore_mn_ops;
	mn_pmu->dev = &pdev->dev;
	mn_pmu->on_cpu = -1;

	reg_info = mn_pmu->dev_info->private;
	mn_pmu->identifier = readl(mn_pmu->base + reg_info->version);

	return 0;
}

static void hisi_mn_pmu_remove_cpuhp(void *hotplug_node)
{
	cpuhp_state_remove_instance_nocalls(hisi_mn_pmu_online, hotplug_node);
}

static void hisi_mn_pmu_unregister(void *pmu)
{
	perf_pmu_unregister(pmu);
}

static int hisi_mn_pmu_probe(struct platform_device *pdev)
{
	struct hisi_pmu *mn_pmu;
	char *name;
	int ret;

	mn_pmu = devm_kzalloc(&pdev->dev, sizeof(*mn_pmu), GFP_KERNEL);
	if (!mn_pmu)
		return -ENOMEM;

	platform_set_drvdata(pdev, mn_pmu);

	ret = hisi_mn_pmu_dev_init(pdev, mn_pmu);
	if (ret)
		return ret;

	name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "hisi_scl%d_mn%d",
				mn_pmu->topo.scl_id, mn_pmu->topo.index_id);
	if (!name)
		return -ENOMEM;

	ret = cpuhp_state_add_instance(hisi_mn_pmu_online, &mn_pmu->node);
	if (ret)
		return dev_err_probe(&pdev->dev, ret, "Failed to register cpu hotplug\n");

	ret = devm_add_action_or_reset(&pdev->dev, hisi_mn_pmu_remove_cpuhp, &mn_pmu->node);
	if (ret)
		return ret;

	hisi_pmu_init(mn_pmu, THIS_MODULE);

	ret = perf_pmu_register(&mn_pmu->pmu, name, -1);
	if (ret)
		return dev_err_probe(mn_pmu->dev, ret, "Failed to register MN PMU\n");

	return devm_add_action_or_reset(&pdev->dev, hisi_mn_pmu_unregister, &mn_pmu->pmu);
}

static struct hisi_mn_pmu_regs hisi_mn_v1_pmu_regs = {
	.version = HISI_MN_VERSION_REG,
	.dyn_ctrl = HISI_MN_DYNAMIC_CTRL_REG,
	.perf_ctrl = HISI_MN_PERF_CTRL_REG,
	.int_mask = HISI_MN_INT_MASK_REG,
	.int_clear = HISI_MN_INT_CLEAR_REG,
	.int_status = HISI_MN_INT_STATUS_REG,
	.event_ctrl = HISI_MN_EVENT_CTRL_REG,
	.event_type0 = HISI_MN_EVTYPE0_REG,
	.event_cntr0 = HISI_MN_CNTR0_REG,
};

static const struct hisi_pmu_dev_info hisi_mn_v1 = {
	.attr_groups = hisi_mn_pmu_attr_groups,
	.counter_bits = 48,
	.check_event = HISI_MN_EVTYPE_MASK,
	.private = &hisi_mn_v1_pmu_regs,
};

static const struct acpi_device_id hisi_mn_pmu_acpi_match[] = {
	{ "HISI0222", (kernel_ulong_t) &hisi_mn_v1 },
	{ }
};
MODULE_DEVICE_TABLE(acpi, hisi_mn_pmu_acpi_match);

static struct platform_driver hisi_mn_pmu_driver = {
	.driver = {
		.name = "hisi_mn_pmu",
		.acpi_match_table = hisi_mn_pmu_acpi_match,
		/*
		 * We have not worked out a safe bind/unbind process,
		 * Forcefully unbinding during sampling will lead to a
		 * kernel panic, so this is not supported yet.
		 */
		.suppress_bind_attrs = true,
	},
	.probe = hisi_mn_pmu_probe,
};

static int __init hisi_mn_pmu_module_init(void)
{
	int ret;

	ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN, "perf/hisi/mn:online",
				      hisi_uncore_pmu_online_cpu,
				      hisi_uncore_pmu_offline_cpu);
	if (ret < 0) {
		pr_err("hisi_mn_pmu: Failed to setup MN PMU hotplug: %d\n", ret);
		return ret;
	}
	hisi_mn_pmu_online = ret;

	ret = platform_driver_register(&hisi_mn_pmu_driver);
	if (ret)
		cpuhp_remove_multi_state(hisi_mn_pmu_online);

	return ret;
}
module_init(hisi_mn_pmu_module_init);

static void __exit hisi_mn_pmu_module_exit(void)
{
	platform_driver_unregister(&hisi_mn_pmu_driver);
	cpuhp_remove_multi_state(hisi_mn_pmu_online);
}
module_exit(hisi_mn_pmu_module_exit);

MODULE_IMPORT_NS("HISI_PMU");
MODULE_DESCRIPTION("HiSilicon SoC MN uncore PMU driver");
MODULE_LICENSE("GPL");
MODULE_AUTHOR("Junhao He <hejunhao3@huawei.com>");