[v2,2/3] drivers/perf: hisi: Add support for HiSilicon UC PMU driver

Message ID 20230531104625.18296-3-hejunhao3@huawei.com
State New
Headers
Series Add support for HiSilicon SoC uncore PMU |

Commit Message

hejunhao May 31, 2023, 10:46 a.m. UTC
  On HiSilicon Hip09 platform, there is a UC (unified cache) module
on each chip SCCL (Super CPU Cluster). UC is a cache that provides
coherence between NUMA and UMA domains. It is located between L2
and Memory System. While PA uncore PMU model is the same as other
Hip09 PMU modules and many PMU events are supported. Let's support
the PMU driver using the HiSilicon uncore PMU framework.

* rd_req_en : rd_req_en is the abbreviation of read request tracetag enable
and allows user to count only read operations.
details are listed in the hisi-pmu document.

* srcid_en & srcid: allows user to filter statistics that come from
specific CPU/ICL by configuration source ID.

* uring_channel: Allows users to filter statistical information based on
the specified tx request uring channel.
uring_channel only supported events: [0x47 ~ 0x59].

Signed-off-by: Junhao He <hejunhao3@huawei.com>
---
 drivers/perf/hisilicon/Makefile             |   2 +-
 drivers/perf/hisilicon/hisi_uncore_pmu.c    |   4 +-
 drivers/perf/hisilicon/hisi_uncore_pmu.h    |   6 +
 drivers/perf/hisilicon/hisi_uncore_uc_pmu.c | 577 ++++++++++++++++++++
 4 files changed, 587 insertions(+), 2 deletions(-)
 create mode 100644 drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
  

Comments

Jonathan Cameron May 31, 2023, 3:59 p.m. UTC | #1
On Wed, 31 May 2023 18:46:24 +0800
Junhao He <hejunhao3@huawei.com> wrote:

Hi Junhao,

A few small comments inline.

> On HiSilicon Hip09 platform, there is a UC (unified cache) module
> on each chip SCCL (Super CPU Cluster). UC is a cache that provides
> coherence between NUMA and UMA domains. It is located between L2
> and Memory System. While PA uncore PMU model is the same as other
> Hip09 PMU modules and many PMU events are supported. 

I don't follow what this sentence means.  Normally you'd have
While A, B is different..


> Let's support
> the PMU driver using the HiSilicon uncore PMU framework.
> 
> * rd_req_en : rd_req_en is the abbreviation of read request tracetag enable
> and allows user to count only read operations.
> details are listed in the hisi-pmu document.
Details are .. Also no need for the ine break
  and allows user to count only read operations. Details are listed
  in the hisi-pmu document at ....

> 
> * srcid_en & srcid: allows user to filter statistics that come from

Allows
for consistency with the uring_channel description that follows.

> specific CPU/ICL by configuration source ID.
> 
> * uring_channel: Allows users to filter statistical information based on
> the specified tx request uring channel.
> uring_channel only supported events: [0x47 ~ 0x59].
> 
> Signed-off-by: Junhao He <hejunhao3@huawei.com>


> diff --git a/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
> new file mode 100644
> index 000000000000..d27f28584fd7
> --- /dev/null
> +++ b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
> @@ -0,0 +1,577 @@
...




> +static int hisi_uc_pmu_init_data(struct platform_device *pdev,
> +				 struct hisi_pmu *uc_pmu)
> +{
> +	/*
> +	 * Use SCCL (Super CPU Cluster) ID and CCL (CPU Cluster) ID to
> +	 * identify the topology information of UC PMU devices in the chip.
> +	 */

From patch description, I'd assume there is only one of these
per sccl so why do we care about the cluster level or the sub-id?
Perhaps that description is missleading?

> +	if (device_property_read_u32(&pdev->dev, "hisilicon,scl-id",
> +				     &uc_pmu->sccl_id)) {
> +		dev_err(&pdev->dev, "Can not read uc sccl-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	if (device_property_read_u32(&pdev->dev, "hisilicon,ccl-id",
> +				     &uc_pmu->ccl_id)) {
> +		dev_err(&pdev->dev, "Can not read uc ccl-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	if (device_property_read_u32(&pdev->dev, "hisilicon,sub-id",
> +				     &uc_pmu->sub_id)) {
> +		dev_err(&pdev->dev, "Can not read sub-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	uc_pmu->base = devm_platform_ioremap_resource(pdev, 0);
> +	if (IS_ERR(uc_pmu->base)) {
> +		dev_err(&pdev->dev, "ioremap failed for uc_pmu resource\n");
> +		return PTR_ERR(uc_pmu->base);
> +	}
> +
> +	uc_pmu->identifier = readl(uc_pmu->base + HISI_UC_VERSION_REG);
> +
> +	return 0;
> +}


> +static struct platform_driver hisi_uc_pmu_driver = {
> +	.driver = {
> +		.name = "hisi_uc_pmu",
> +		.acpi_match_table = hisi_uc_pmu_acpi_match,
> +		/*
> +		 * We have not worked out a safe bind/unbind process,
> +		 * so this is not supported yet.

If you can reference more info on this that would be great.
Perhaps a thread talking about why?

> +		 */
> +		.suppress_bind_attrs = true,
> +	},
> +	.probe = hisi_uc_pmu_probe,
> +};
  
Yicong Yang June 1, 2023, 6:41 a.m. UTC | #2
n 2023/5/31 18:46, Junhao He wrote:
> On HiSilicon Hip09 platform, there is a UC (unified cache) module
> on each chip SCCL (Super CPU Cluster). UC is a cache that provides
> coherence between NUMA and UMA domains. It is located between L2
> and Memory System. While PA uncore PMU model is the same as other
> Hip09 PMU modules and many PMU events are supported. Let's support
> the PMU driver using the HiSilicon uncore PMU framework.
> 
> * rd_req_en : rd_req_en is the abbreviation of read request tracetag enable
> and allows user to count only read operations.
> details are listed in the hisi-pmu document.
> 
> * srcid_en & srcid: allows user to filter statistics that come from
> specific CPU/ICL by configuration source ID.
> 
> * uring_channel: Allows users to filter statistical information based on
> the specified tx request uring channel.
> uring_channel only supported events: [0x47 ~ 0x59].
> 
> Signed-off-by: Junhao He <hejunhao3@huawei.com>

Since my comments get addressed, so

Reviewed-by: Yicong Yang <yangyicong@hisilicon.com>

> ---
>  drivers/perf/hisilicon/Makefile             |   2 +-
>  drivers/perf/hisilicon/hisi_uncore_pmu.c    |   4 +-
>  drivers/perf/hisilicon/hisi_uncore_pmu.h    |   6 +
>  drivers/perf/hisilicon/hisi_uncore_uc_pmu.c | 577 ++++++++++++++++++++
>  4 files changed, 587 insertions(+), 2 deletions(-)
>  create mode 100644 drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
> 
> diff --git a/drivers/perf/hisilicon/Makefile b/drivers/perf/hisilicon/Makefile
> index 4d2c9abe3372..48dcc8381ea7 100644
> --- a/drivers/perf/hisilicon/Makefile
> +++ b/drivers/perf/hisilicon/Makefile
> @@ -1,7 +1,7 @@
>  # SPDX-License-Identifier: GPL-2.0-only
>  obj-$(CONFIG_HISI_PMU) += hisi_uncore_pmu.o hisi_uncore_l3c_pmu.o \
>  			  hisi_uncore_hha_pmu.o hisi_uncore_ddrc_pmu.o hisi_uncore_sllc_pmu.o \
> -			  hisi_uncore_pa_pmu.o hisi_uncore_cpa_pmu.o
> +			  hisi_uncore_pa_pmu.o hisi_uncore_cpa_pmu.o hisi_uncore_uc_pmu.o
>  
>  obj-$(CONFIG_HISI_PCIE_PMU) += hisi_pcie_pmu.o
>  obj-$(CONFIG_HNS3_PMU) += hns3_pmu.o
> diff --git a/drivers/perf/hisilicon/hisi_uncore_pmu.c b/drivers/perf/hisilicon/hisi_uncore_pmu.c
> index 2823f381930d..04031450d5fe 100644
> --- a/drivers/perf/hisilicon/hisi_uncore_pmu.c
> +++ b/drivers/perf/hisilicon/hisi_uncore_pmu.c
> @@ -20,7 +20,6 @@
>  
>  #include "hisi_uncore_pmu.h"
>  
> -#define HISI_GET_EVENTID(ev) (ev->hw.config_base & 0xff)
>  #define HISI_MAX_PERIOD(nr) (GENMASK_ULL((nr) - 1, 0))
>  
>  /*
> @@ -226,6 +225,9 @@ int hisi_uncore_pmu_event_init(struct perf_event *event)
>  	hwc->idx		= -1;
>  	hwc->config_base	= event->attr.config;
>  
> +	if (hisi_pmu->ops->check_filter && hisi_pmu->ops->check_filter(event))
> +		return -EINVAL;
> +
>  	/* Enforce to use the same CPU for all events in this PMU */
>  	event->cpu = hisi_pmu->on_cpu;
>  
> diff --git a/drivers/perf/hisilicon/hisi_uncore_pmu.h b/drivers/perf/hisilicon/hisi_uncore_pmu.h
> index a8d6d6905f3f..27b6122aa486 100644
> --- a/drivers/perf/hisilicon/hisi_uncore_pmu.h
> +++ b/drivers/perf/hisilicon/hisi_uncore_pmu.h
> @@ -44,9 +44,15 @@
>  		return FIELD_GET(GENMASK_ULL(hi, lo), event->attr.config);  \
>  	}
>  
> +#define HISI_GET_EVENTID(ev) (ev->hw.config_base & 0xff)
> +
> +#define HISI_PMU_EVTYPE_BITS		8
> +#define HISI_PMU_EVTYPE_SHIFT(idx)	((idx) % 4 * HISI_PMU_EVTYPE_BITS)
> +
>  struct hisi_pmu;
>  
>  struct hisi_uncore_ops {
> +	int (*check_filter)(struct perf_event *event);
>  	void (*write_evtype)(struct hisi_pmu *, int, u32);
>  	int (*get_event_idx)(struct perf_event *);
>  	u64 (*read_counter)(struct hisi_pmu *, struct hw_perf_event *);
> diff --git a/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
> new file mode 100644
> index 000000000000..d27f28584fd7
> --- /dev/null
> +++ b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
> @@ -0,0 +1,577 @@
> +// SPDX-License-Identifier: GPL-2.0-only
> +/*
> + * HiSilicon SoC UC (unified cache) uncore Hardware event counters support
> + *
> + * Copyright (C) 2023 HiSilicon Limited
> + *
> + * This code is based on the uncore PMUs like hisi_uncore_l3c_pmu.
> + */
> +#include <linux/cpuhotplug.h>
> +#include <linux/interrupt.h>
> +#include <linux/irq.h>
> +#include <linux/list.h>
> +#include <linux/mod_devicetable.h>
> +#include <linux/property.h>
> +
> +#include "hisi_uncore_pmu.h"
> +
> +/* Dynamic CPU hotplug state used by UC PMU */
> +static enum cpuhp_state hisi_uc_pmu_online;
> +
> +/* UC register definition */
> +#define HISI_UC_INT_MASK_REG		0x0800
> +#define HISI_UC_INT_STS_REG		0x0808
> +#define HISI_UC_INT_CLEAR_REG		0x080c
> +#define HISI_UC_TRACETAG_CTRL_REG	0x1b2c
> +#define HISI_UC_TRACETAG_REQ_MSK	GENMASK(9, 7)
> +#define HISI_UC_TRACETAG_MARK_EN	BIT(0)
> +#define HISI_UC_TRACETAG_REQ_EN		(HISI_UC_TRACETAG_MARK_EN | BIT(2))
> +#define HISI_UC_TRACETAG_SRCID_EN	BIT(3)
> +#define HISI_UC_SRCID_CTRL_REG		0x1b40
> +#define HISI_UC_SRCID_MSK		GENMASK(14, 1)
> +#define HISI_UC_EVENT_CTRL_REG		0x1c00
> +#define HISI_UC_EVENT_TRACETAG_EN	BIT(29)
> +#define HISI_UC_EVENT_URING_MSK		GENMASK(28, 27)
> +#define HISI_UC_EVENT_GLB_EN		BIT(26)
> +#define HISI_UC_VERSION_REG		0x1cf0
> +#define HISI_UC_EVTYPE_REGn(n)		(0x1d00 + (n) * 4)
> +#define HISI_UC_EVTYPE_MASK		GENMASK(7, 0)
> +#define HISI_UC_CNTR_REGn(n)		(0x1e00 + (n) * 8)
> +
> +#define HISI_UC_NR_COUNTERS		0x8
> +#define HISI_UC_V2_NR_EVENTS		0xFF
> +#define HISI_UC_CNTR_REG_BITS		64
> +
> +#define HISI_UC_RD_REQ_TRACETAG		0x4
> +#define HISI_UC_URING_EVENT_MIN		0x47
> +#define HISI_UC_URING_EVENT_MAX		0x59
> +
> +HISI_PMU_EVENT_ATTR_EXTRACTOR(rd_req_en, config1, 0, 0);
> +HISI_PMU_EVENT_ATTR_EXTRACTOR(uring_channel, config1, 5, 4);
> +HISI_PMU_EVENT_ATTR_EXTRACTOR(srcid, config1, 19, 6);
> +HISI_PMU_EVENT_ATTR_EXTRACTOR(srcid_en, config1, 20, 20);
> +
> +static int hisi_uc_pmu_check_filter(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +
> +	if (hisi_get_srcid_en(event) && !hisi_get_rd_req_en(event)) {
> +		dev_err(uc_pmu->dev,
> +			"Failed to set srcid: depending on read request enabled!\n");
> +		return -EINVAL;
> +	}
> +
> +	if (!hisi_get_uring_channel(event))
> +		return 0;
> +
> +	if ((HISI_GET_EVENTID(event) < HISI_UC_URING_EVENT_MIN) ||
> +	    (HISI_GET_EVENTID(event) > HISI_UC_URING_EVENT_MAX))
> +		dev_warn(uc_pmu->dev,
> +			 "Only events: [%#x ~ %#x] support channel filtering!",
> +			 HISI_UC_URING_EVENT_MIN, HISI_UC_URING_EVENT_MAX);
> +
> +	return 0;
> +}
> +
> +static void hisi_uc_pmu_config_req_tracetag(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 val;
> +
> +	if (!hisi_get_rd_req_en(event))
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	/* The request-type has been configured */
> +	if (FIELD_GET(HISI_UC_TRACETAG_REQ_MSK, val) == HISI_UC_RD_REQ_TRACETAG)
> +		return;
> +
> +	/* Set request-type for tracetag, only read request is supported! */
> +	val &= ~HISI_UC_TRACETAG_REQ_MSK;
> +	val |= FIELD_PREP(HISI_UC_TRACETAG_REQ_MSK, HISI_UC_RD_REQ_TRACETAG);
> +	val |= HISI_UC_TRACETAG_REQ_EN;
> +	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_clear_req_tracetag(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 val;
> +
> +	if (!hisi_get_rd_req_en(event))
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	/* Do nothing, the request-type tracetag has been cleaned up */
> +	if (FIELD_GET(HISI_UC_TRACETAG_REQ_MSK, val) == 0)
> +		return;
> +
> +	/* Clear request-type */
> +	val &= ~HISI_UC_TRACETAG_REQ_MSK;
> +	val &= ~HISI_UC_TRACETAG_REQ_EN;
> +	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_config_srcid_tracetag(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 val;
> +
> +	if (!hisi_get_srcid_en(event))
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	/* Do nothing, the source id has been configured */
> +	if (FIELD_GET(HISI_UC_TRACETAG_SRCID_EN, val))
> +		return;
> +
> +	/* Enable source id tracetag */
> +	val |= HISI_UC_TRACETAG_SRCID_EN;
> +	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	val = readl(uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
> +	val &= ~HISI_UC_SRCID_MSK;
> +	val |= FIELD_PREP(HISI_UC_SRCID_MSK, hisi_get_srcid(event));
> +	writel(val, uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
> +
> +	/* Depend on request-type tracetag enabled */
> +	hisi_uc_pmu_config_req_tracetag(event);
> +}
> +
> +static void hisi_uc_pmu_clear_srcid_tracetag(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 val;
> +
> +	if (!hisi_get_srcid_en(event))
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	/* Do nothing, the source id has been cleaned up */
> +	if (FIELD_GET(HISI_UC_TRACETAG_SRCID_EN, val) == 0)
> +		return;
> +
> +	hisi_uc_pmu_clear_req_tracetag(event);
> +
> +	/* Disable source id tracetag */
> +	val &= ~HISI_UC_TRACETAG_SRCID_EN;
> +	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
> +
> +	val = readl(uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
> +	val &= ~HISI_UC_SRCID_MSK;
> +	writel(val, uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_config_uring_channel(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 uring_channel = hisi_get_uring_channel(event);
> +	u32 val;
> +
> +	/* Do nothing if not being set or is set explicitly to zero (default) */
> +	if (uring_channel == 0)
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +
> +	/* Do nothing, the uring_channel has been configured */
> +	if (uring_channel == FIELD_GET(HISI_UC_EVENT_URING_MSK, val))
> +		return;
> +
> +	val &= ~HISI_UC_EVENT_URING_MSK;
> +	val |= FIELD_PREP(HISI_UC_EVENT_URING_MSK, uring_channel);
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_clear_uring_channel(struct perf_event *event)
> +{
> +	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
> +	u32 val;
> +
> +	/* Do nothing if not being set or is set explicitly to zero (default) */
> +	if (hisi_get_uring_channel(event) == 0)
> +		return;
> +
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +
> +	/* Do nothing, the uring_channel has been cleaned up */
> +	if (FIELD_GET(HISI_UC_EVENT_URING_MSK, val) == 0)
> +		return;
> +
> +	val &= ~HISI_UC_EVENT_URING_MSK;
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_enable_filter(struct perf_event *event)
> +{
> +	if (event->attr.config1 == 0)
> +		return;
> +
> +	hisi_uc_pmu_config_uring_channel(event);
> +	hisi_uc_pmu_config_req_tracetag(event);
> +	hisi_uc_pmu_config_srcid_tracetag(event);
> +}
> +
> +static void hisi_uc_pmu_disable_filter(struct perf_event *event)
> +{
> +	if (event->attr.config1 == 0)
> +		return;
> +
> +	hisi_uc_pmu_clear_srcid_tracetag(event);
> +	hisi_uc_pmu_clear_req_tracetag(event);
> +	hisi_uc_pmu_clear_uring_channel(event);
> +}
> +
> +static void hisi_uc_pmu_write_evtype(struct hisi_pmu *uc_pmu, int idx, u32 type)
> +{
> +	u32 val;
> +
> +	/*
> +	 * Select the appropriate event select register.
> +	 * There are 2 32-bit event select registers for the
> +	 * 8 hardware counters, each event code is 8-bit wide.
> +	 */
> +	val = readl(uc_pmu->base + HISI_UC_EVTYPE_REGn(idx / 4));
> +	val &= ~(HISI_UC_EVTYPE_MASK << HISI_PMU_EVTYPE_SHIFT(idx));
> +	val |= (type << HISI_PMU_EVTYPE_SHIFT(idx));
> +	writel(val, uc_pmu->base + HISI_UC_EVTYPE_REGn(idx / 4));
> +}
> +
> +static void hisi_uc_pmu_start_counters(struct hisi_pmu *uc_pmu)
> +{
> +	u32 val;
> +
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +	val |= HISI_UC_EVENT_GLB_EN;
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_stop_counters(struct hisi_pmu *uc_pmu)
> +{
> +	u32 val;
> +
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +	val &= ~HISI_UC_EVENT_GLB_EN;
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_enable_counter(struct hisi_pmu *uc_pmu,
> +					struct hw_perf_event *hwc)
> +{
> +	u32 val;
> +
> +	/* Enable counter index */
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +	val |= (1 << hwc->idx);
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static void hisi_uc_pmu_disable_counter(struct hisi_pmu *uc_pmu,
> +					struct hw_perf_event *hwc)
> +{
> +	u32 val;
> +
> +	/* Clear counter index */
> +	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +	val &= ~(1 << hwc->idx);
> +	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
> +}
> +
> +static u64 hisi_uc_pmu_read_counter(struct hisi_pmu *uc_pmu,
> +				    struct hw_perf_event *hwc)
> +{
> +	return readq(uc_pmu->base + HISI_UC_CNTR_REGn(hwc->idx));
> +}
> +
> +static void hisi_uc_pmu_write_counter(struct hisi_pmu *uc_pmu,
> +				      struct hw_perf_event *hwc, u64 val)
> +{
> +	hisi_uc_pmu_start_counters(uc_pmu);
> +	writeq(val, uc_pmu->base + HISI_UC_CNTR_REGn(hwc->idx));
> +}
> +
> +static void hisi_uc_pmu_enable_counter_int(struct hisi_pmu *uc_pmu,
> +					   struct hw_perf_event *hwc)
> +{
> +	u32 val;
> +
> +	val = readl(uc_pmu->base + HISI_UC_INT_MASK_REG);
> +	val &= ~(1 << hwc->idx);
> +	writel(val, uc_pmu->base + HISI_UC_INT_MASK_REG);
> +}
> +
> +static void hisi_uc_pmu_disable_counter_int(struct hisi_pmu *uc_pmu,
> +					    struct hw_perf_event *hwc)
> +{
> +	u32 val;
> +
> +	val = readl(uc_pmu->base + HISI_UC_INT_MASK_REG);
> +	val |= (1 << hwc->idx);
> +	writel(val, uc_pmu->base + HISI_UC_INT_MASK_REG);
> +}
> +
> +static u32 hisi_uc_pmu_get_int_status(struct hisi_pmu *uc_pmu)
> +{
> +	return readl(uc_pmu->base + HISI_UC_INT_STS_REG);
> +}
> +
> +static void hisi_uc_pmu_clear_int_status(struct hisi_pmu *uc_pmu, int idx)
> +{
> +	writel(1 << idx, uc_pmu->base + HISI_UC_INT_CLEAR_REG);
> +}
> +
> +static int hisi_uc_pmu_init_data(struct platform_device *pdev,
> +				 struct hisi_pmu *uc_pmu)
> +{
> +	/*
> +	 * Use SCCL (Super CPU Cluster) ID and CCL (CPU Cluster) ID to
> +	 * identify the topology information of UC PMU devices in the chip.
> +	 */
> +	if (device_property_read_u32(&pdev->dev, "hisilicon,scl-id",
> +				     &uc_pmu->sccl_id)) {
> +		dev_err(&pdev->dev, "Can not read uc sccl-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	if (device_property_read_u32(&pdev->dev, "hisilicon,ccl-id",
> +				     &uc_pmu->ccl_id)) {
> +		dev_err(&pdev->dev, "Can not read uc ccl-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	if (device_property_read_u32(&pdev->dev, "hisilicon,sub-id",
> +				     &uc_pmu->sub_id)) {
> +		dev_err(&pdev->dev, "Can not read sub-id!\n");
> +		return -EINVAL;
> +	}
> +
> +	uc_pmu->base = devm_platform_ioremap_resource(pdev, 0);
> +	if (IS_ERR(uc_pmu->base)) {
> +		dev_err(&pdev->dev, "ioremap failed for uc_pmu resource\n");
> +		return PTR_ERR(uc_pmu->base);
> +	}
> +
> +	uc_pmu->identifier = readl(uc_pmu->base + HISI_UC_VERSION_REG);
> +
> +	return 0;
> +}
> +
> +static struct attribute *hisi_uc_pmu_format_attr[] = {
> +	HISI_PMU_FORMAT_ATTR(event, "config:0-7"),
> +	HISI_PMU_FORMAT_ATTR(rd_req_en, "config1:0-0"),
> +	HISI_PMU_FORMAT_ATTR(uring_channel, "config1:4-5"),
> +	HISI_PMU_FORMAT_ATTR(srcid, "config1:6-19"),
> +	HISI_PMU_FORMAT_ATTR(srcid_en, "config1:20-20"),
> +	NULL
> +};
> +
> +static const struct attribute_group hisi_uc_pmu_format_group = {
> +	.name = "format",
> +	.attrs = hisi_uc_pmu_format_attr,
> +};
> +
> +static struct attribute *hisi_uc_pmu_events_attr[] = {
> +	HISI_PMU_EVENT_ATTR(sq_time,		0x00),
> +	HISI_PMU_EVENT_ATTR(pq_time,		0x01),
> +	HISI_PMU_EVENT_ATTR(hbm_time,		0x02),
> +	HISI_PMU_EVENT_ATTR(iq_comp_time_cring,	0x03),
> +	HISI_PMU_EVENT_ATTR(iq_comp_time_uring,	0x05),
> +	HISI_PMU_EVENT_ATTR(cpu_rd,		0x10),
> +	HISI_PMU_EVENT_ATTR(cpu_rd64,		0x17),
> +	HISI_PMU_EVENT_ATTR(cpu_rs64,		0x19),
> +	HISI_PMU_EVENT_ATTR(cpu_mru,		0x1a),
> +	HISI_PMU_EVENT_ATTR(cycles,		0x9c),
> +	HISI_PMU_EVENT_ATTR(spipe_hit,		0xb3),
> +	HISI_PMU_EVENT_ATTR(hpipe_hit,		0xdb),
> +	HISI_PMU_EVENT_ATTR(cring_rxdat_cnt,	0xfa),
> +	HISI_PMU_EVENT_ATTR(cring_txdat_cnt,	0xfb),
> +	HISI_PMU_EVENT_ATTR(uring_rxdat_cnt,	0xfc),
> +	HISI_PMU_EVENT_ATTR(uring_txdat_cnt,	0xfd),
> +	NULL
> +};
> +
> +static const struct attribute_group hisi_uc_pmu_events_group = {
> +	.name = "events",
> +	.attrs = hisi_uc_pmu_events_attr,
> +};
> +
> +static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL);
> +
> +static struct attribute *hisi_uc_pmu_cpumask_attrs[] = {
> +	&dev_attr_cpumask.attr,
> +	NULL,
> +};
> +
> +static const struct attribute_group hisi_uc_pmu_cpumask_attr_group = {
> +	.attrs = hisi_uc_pmu_cpumask_attrs,
> +};
> +
> +static struct device_attribute hisi_uc_pmu_identifier_attr =
> +	__ATTR(identifier, 0444, hisi_uncore_pmu_identifier_attr_show, NULL);
> +
> +static struct attribute *hisi_uc_pmu_identifier_attrs[] = {
> +	&hisi_uc_pmu_identifier_attr.attr,
> +	NULL
> +};
> +
> +static const struct attribute_group hisi_uc_pmu_identifier_group = {
> +	.attrs = hisi_uc_pmu_identifier_attrs,
> +};
> +
> +static const struct attribute_group *hisi_uc_pmu_attr_groups[] = {
> +	&hisi_uc_pmu_format_group,
> +	&hisi_uc_pmu_events_group,
> +	&hisi_uc_pmu_cpumask_attr_group,
> +	&hisi_uc_pmu_identifier_group,
> +	NULL
> +};
> +
> +static const struct hisi_uncore_ops hisi_uncore_uc_pmu_ops = {
> +	.check_filter		= hisi_uc_pmu_check_filter,
> +	.write_evtype		= hisi_uc_pmu_write_evtype,
> +	.get_event_idx		= hisi_uncore_pmu_get_event_idx,
> +	.start_counters		= hisi_uc_pmu_start_counters,
> +	.stop_counters		= hisi_uc_pmu_stop_counters,
> +	.enable_counter		= hisi_uc_pmu_enable_counter,
> +	.disable_counter	= hisi_uc_pmu_disable_counter,
> +	.enable_counter_int	= hisi_uc_pmu_enable_counter_int,
> +	.disable_counter_int	= hisi_uc_pmu_disable_counter_int,
> +	.write_counter		= hisi_uc_pmu_write_counter,
> +	.read_counter		= hisi_uc_pmu_read_counter,
> +	.get_int_status		= hisi_uc_pmu_get_int_status,
> +	.clear_int_status	= hisi_uc_pmu_clear_int_status,
> +	.enable_filter		= hisi_uc_pmu_enable_filter,
> +	.disable_filter		= hisi_uc_pmu_disable_filter,
> +};
> +
> +static int hisi_uc_pmu_dev_probe(struct platform_device *pdev,
> +				 struct hisi_pmu *uc_pmu)
> +{
> +	int ret;
> +
> +	ret = hisi_uc_pmu_init_data(pdev, uc_pmu);
> +	if (ret)
> +		return ret;
> +
> +	ret = hisi_uncore_pmu_init_irq(uc_pmu, pdev);
> +	if (ret)
> +		return ret;
> +
> +	uc_pmu->pmu_events.attr_groups = hisi_uc_pmu_attr_groups;
> +	uc_pmu->check_event = HISI_UC_EVTYPE_MASK;
> +	uc_pmu->ops = &hisi_uncore_uc_pmu_ops;
> +	uc_pmu->counter_bits = HISI_UC_CNTR_REG_BITS;
> +	uc_pmu->num_counters = HISI_UC_NR_COUNTERS;
> +	uc_pmu->dev = &pdev->dev;
> +	uc_pmu->on_cpu = -1;
> +
> +	return 0;
> +}
> +
> +static void hisi_uc_pmu_remove_cpuhp_instance(void *hotplug_node)
> +{
> +	cpuhp_state_remove_instance_nocalls(hisi_uc_pmu_online, hotplug_node);
> +}
> +
> +static void hisi_uc_pmu_unregister_pmu(void *pmu)
> +{
> +	perf_pmu_unregister(pmu);
> +}
> +
> +static int hisi_uc_pmu_probe(struct platform_device *pdev)
> +{
> +	struct hisi_pmu *uc_pmu;
> +	char *name;
> +	int ret;
> +
> +	uc_pmu = devm_kzalloc(&pdev->dev, sizeof(*uc_pmu), GFP_KERNEL);
> +	if (!uc_pmu)
> +		return -ENOMEM;
> +
> +	platform_set_drvdata(pdev, uc_pmu);
> +
> +	ret = hisi_uc_pmu_dev_probe(pdev, uc_pmu);
> +	if (ret)
> +		return ret;
> +
> +	name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "hisi_sccl%d_uc%d_%u",
> +			      uc_pmu->sccl_id, uc_pmu->ccl_id, uc_pmu->sub_id);
> +	if (!name)
> +		return -ENOMEM;
> +
> +	ret = cpuhp_state_add_instance(hisi_uc_pmu_online, &uc_pmu->node);
> +	if (ret)
> +		return dev_err_probe(&pdev->dev, ret, "Error registering hotplug\n");
> +
> +	ret = devm_add_action_or_reset(&pdev->dev,
> +				       hisi_uc_pmu_remove_cpuhp_instance,
> +				       &uc_pmu->node);
> +	if (ret)
> +		return ret;
> +
> +	hisi_pmu_init(uc_pmu, THIS_MODULE);
> +
> +	ret = perf_pmu_register(&uc_pmu->pmu, name, -1);
> +	if (ret)
> +		return ret;
> +
> +	return devm_add_action_or_reset(&pdev->dev,
> +					hisi_uc_pmu_unregister_pmu,
> +					&uc_pmu->pmu);
> +}
> +
> +static const struct acpi_device_id hisi_uc_pmu_acpi_match[] = {
> +	{ "HISI0291", },
> +	{}
> +};
> +MODULE_DEVICE_TABLE(acpi, hisi_uc_pmu_acpi_match);
> +
> +static struct platform_driver hisi_uc_pmu_driver = {
> +	.driver = {
> +		.name = "hisi_uc_pmu",
> +		.acpi_match_table = hisi_uc_pmu_acpi_match,
> +		/*
> +		 * We have not worked out a safe bind/unbind process,
> +		 * so this is not supported yet.
> +		 */
> +		.suppress_bind_attrs = true,
> +	},
> +	.probe = hisi_uc_pmu_probe,
> +};
> +
> +static int __init hisi_uc_pmu_module_init(void)
> +{
> +	int ret;
> +
> +	ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN,
> +				      "perf/hisi/uc:online",
> +				      hisi_uncore_pmu_online_cpu,
> +				      hisi_uncore_pmu_offline_cpu);
> +	if (ret < 0) {
> +		pr_err("UC PMU: Error setup hotplug, ret = %d\n", ret);
> +		return ret;
> +	}
> +	hisi_uc_pmu_online = ret;
> +
> +	ret = platform_driver_register(&hisi_uc_pmu_driver);
> +	if (ret)
> +		cpuhp_remove_multi_state(hisi_uc_pmu_online);
> +
> +	return ret;
> +}
> +module_init(hisi_uc_pmu_module_init);
> +
> +static void __exit hisi_uc_pmu_module_exit(void)
> +{
> +	platform_driver_unregister(&hisi_uc_pmu_driver);
> +	cpuhp_remove_multi_state(hisi_uc_pmu_online);
> +}
> +module_exit(hisi_uc_pmu_module_exit);
> +
> +MODULE_DESCRIPTION("HiSilicon SoC UC uncore PMU driver");
> +MODULE_LICENSE("GPL");
> +MODULE_AUTHOR("Junhao He <hejunhao3@huawei.com>");
>
  
Yicong Yang June 1, 2023, 6:53 a.m. UTC | #3
On 2023/5/31 23:59, Jonathan Cameron wrote:
> On Wed, 31 May 2023 18:46:24 +0800
> Junhao He <hejunhao3@huawei.com> wrote:
> 
> Hi Junhao,
> 
> A few small comments inline.
> 
>> On HiSilicon Hip09 platform, there is a UC (unified cache) module
>> on each chip SCCL (Super CPU Cluster). UC is a cache that provides
>> coherence between NUMA and UMA domains. It is located between L2
>> and Memory System. While PA uncore PMU model is the same as other
>> Hip09 PMU modules and many PMU events are supported. 
> 
> I don't follow what this sentence means.  Normally you'd have
> While A, B is different..
> 
> 
>> Let's support
>> the PMU driver using the HiSilicon uncore PMU framework.
>>
>> * rd_req_en : rd_req_en is the abbreviation of read request tracetag enable
>> and allows user to count only read operations.
>> details are listed in the hisi-pmu document.
> Details are .. Also no need for the ine break
>   and allows user to count only read operations. Details are listed
>   in the hisi-pmu document at ....
> 
>>
>> * srcid_en & srcid: allows user to filter statistics that come from
> 
> Allows
> for consistency with the uring_channel description that follows.
> 
>> specific CPU/ICL by configuration source ID.
>>
>> * uring_channel: Allows users to filter statistical information based on
>> the specified tx request uring channel.
>> uring_channel only supported events: [0x47 ~ 0x59].
>>
>> Signed-off-by: Junhao He <hejunhao3@huawei.com>
> 
> 
>> diff --git a/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
>> new file mode 100644
>> index 000000000000..d27f28584fd7
>> --- /dev/null
>> +++ b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
>> @@ -0,0 +1,577 @@
> ...
> 
> 
> 
> 
>> +static int hisi_uc_pmu_init_data(struct platform_device *pdev,
>> +				 struct hisi_pmu *uc_pmu)
>> +{
>> +	/*
>> +	 * Use SCCL (Super CPU Cluster) ID and CCL (CPU Cluster) ID to
>> +	 * identify the topology information of UC PMU devices in the chip.
>> +	 */
> 
>>From patch description, I'd assume there is only one of these
> per sccl so why do we care about the cluster level or the sub-id?
> Perhaps that description is missleading?
> 
>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,scl-id",
>> +				     &uc_pmu->sccl_id)) {
>> +		dev_err(&pdev->dev, "Can not read uc sccl-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,ccl-id",
>> +				     &uc_pmu->ccl_id)) {
>> +		dev_err(&pdev->dev, "Can not read uc ccl-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,sub-id",
>> +				     &uc_pmu->sub_id)) {
>> +		dev_err(&pdev->dev, "Can not read sub-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	uc_pmu->base = devm_platform_ioremap_resource(pdev, 0);
>> +	if (IS_ERR(uc_pmu->base)) {
>> +		dev_err(&pdev->dev, "ioremap failed for uc_pmu resource\n");
>> +		return PTR_ERR(uc_pmu->base);
>> +	}
>> +
>> +	uc_pmu->identifier = readl(uc_pmu->base + HISI_UC_VERSION_REG);
>> +
>> +	return 0;
>> +}
> 
> 
>> +static struct platform_driver hisi_uc_pmu_driver = {
>> +	.driver = {
>> +		.name = "hisi_uc_pmu",
>> +		.acpi_match_table = hisi_uc_pmu_acpi_match,
>> +		/*
>> +		 * We have not worked out a safe bind/unbind process,
>> +		 * so this is not supported yet.
> 
> If you can reference more info on this that would be great.
> Perhaps a thread talking about why?
> 

We handle like this from this patch,
https://lore.kernel.org/linux-arm-kernel/1594975763-32966-1-git-send-email-liuqi115@huawei.com/

>> +		 */
>> +		.suppress_bind_attrs = true,
>> +	},
>> +	.probe = hisi_uc_pmu_probe,
>> +};
> .
>
  
hejunhao June 7, 2023, 11 a.m. UTC | #4
Hi Jonathan,


On 2023/5/31 23:59, Jonathan Cameron wrote:
> On Wed, 31 May 2023 18:46:24 +0800
> Junhao He <hejunhao3@huawei.com> wrote:
>
> Hi Junhao,
>
> A few small comments inline.
>
>> On HiSilicon Hip09 platform, there is a UC (unified cache) module
>> on each chip SCCL (Super CPU Cluster). UC is a cache that provides
>> coherence between NUMA and UMA domains. It is located between L2
>> and Memory System. While PA uncore PMU model is the same as other
>> Hip09 PMU modules and many PMU events are supported.
> I don't follow what this sentence means.  Normally you'd have
> While A, B is different..
>

Ok, will fix it.

>> Let's support
>> the PMU driver using the HiSilicon uncore PMU framework.
>>
>> * rd_req_en : rd_req_en is the abbreviation of read request tracetag enable
>> and allows user to count only read operations.
>> details are listed in the hisi-pmu document.
> Details are .. Also no need for the ine break
>    and allows user to count only read operations. Details are listed
>    in the hisi-pmu document at ....

Thanks, I'm going to replace this

>> * srcid_en & srcid: allows user to filter statistics that come from
> Allows
> for consistency with the uring_channel description that follows.

Yes, I will do that.

>> specific CPU/ICL by configuration source ID.
>>
>> * uring_channel: Allows users to filter statistical information based on
>> the specified tx request uring channel.
>> uring_channel only supported events: [0x47 ~ 0x59].
>>
>> Signed-off-by: Junhao He <hejunhao3@huawei.com>
>
>> diff --git a/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
>> new file mode 100644
>> index 000000000000..d27f28584fd7
>> --- /dev/null
>> +++ b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
>> @@ -0,0 +1,577 @@
> ...
>
>
>
>
>> +static int hisi_uc_pmu_init_data(struct platform_device *pdev,
>> +				 struct hisi_pmu *uc_pmu)
>> +{
>> +	/*
>> +	 * Use SCCL (Super CPU Cluster) ID and CCL (CPU Cluster) ID to
>> +	 * identify the topology information of UC PMU devices in the chip.
>> +	 */
> >From patch description, I'd assume there is only one of these
> per sccl so why do we care about the cluster level or the sub-id?
> Perhaps that description is missleading?

They have some CCLs per SCCL and then 4 uc pmu per CCL.
The patch description is misleading and I will fix this in the next release.
Thanks.

>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,scl-id",
>> +				     &uc_pmu->sccl_id)) {
>> +		dev_err(&pdev->dev, "Can not read uc sccl-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,ccl-id",
>> +				     &uc_pmu->ccl_id)) {
>> +		dev_err(&pdev->dev, "Can not read uc ccl-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	if (device_property_read_u32(&pdev->dev, "hisilicon,sub-id",
>> +				     &uc_pmu->sub_id)) {
>> +		dev_err(&pdev->dev, "Can not read sub-id!\n");
>> +		return -EINVAL;
>> +	}
>> +
>> +	uc_pmu->base = devm_platform_ioremap_resource(pdev, 0);
>> +	if (IS_ERR(uc_pmu->base)) {
>> +		dev_err(&pdev->dev, "ioremap failed for uc_pmu resource\n");
>> +		return PTR_ERR(uc_pmu->base);
>> +	}
>> +
>> +	uc_pmu->identifier = readl(uc_pmu->base + HISI_UC_VERSION_REG);
>> +
>> +	return 0;
>> +}
>
>> +static struct platform_driver hisi_uc_pmu_driver = {
>> +	.driver = {
>> +		.name = "hisi_uc_pmu",
>> +		.acpi_match_table = hisi_uc_pmu_acpi_match,
>> +		/*
>> +		 * We have not worked out a safe bind/unbind process,
>> +		 * so this is not supported yet.
> If you can reference more info on this that would be great.
> Perhaps a thread talking about why?
forcefully unbinding during sampling will lead to a
kernel panic, because the perf upper-layer framework
call a NULL pointer in this situation.

Best regards,
Junhao.
>> +		 */
>> +		.suppress_bind_attrs = true,
>> +	},
>> +	.probe = hisi_uc_pmu_probe,
>> +};
> .
>
  

Patch

diff --git a/drivers/perf/hisilicon/Makefile b/drivers/perf/hisilicon/Makefile
index 4d2c9abe3372..48dcc8381ea7 100644
--- a/drivers/perf/hisilicon/Makefile
+++ b/drivers/perf/hisilicon/Makefile
@@ -1,7 +1,7 @@ 
 # SPDX-License-Identifier: GPL-2.0-only
 obj-$(CONFIG_HISI_PMU) += hisi_uncore_pmu.o hisi_uncore_l3c_pmu.o \
 			  hisi_uncore_hha_pmu.o hisi_uncore_ddrc_pmu.o hisi_uncore_sllc_pmu.o \
-			  hisi_uncore_pa_pmu.o hisi_uncore_cpa_pmu.o
+			  hisi_uncore_pa_pmu.o hisi_uncore_cpa_pmu.o hisi_uncore_uc_pmu.o
 
 obj-$(CONFIG_HISI_PCIE_PMU) += hisi_pcie_pmu.o
 obj-$(CONFIG_HNS3_PMU) += hns3_pmu.o
diff --git a/drivers/perf/hisilicon/hisi_uncore_pmu.c b/drivers/perf/hisilicon/hisi_uncore_pmu.c
index 2823f381930d..04031450d5fe 100644
--- a/drivers/perf/hisilicon/hisi_uncore_pmu.c
+++ b/drivers/perf/hisilicon/hisi_uncore_pmu.c
@@ -20,7 +20,6 @@ 
 
 #include "hisi_uncore_pmu.h"
 
-#define HISI_GET_EVENTID(ev) (ev->hw.config_base & 0xff)
 #define HISI_MAX_PERIOD(nr) (GENMASK_ULL((nr) - 1, 0))
 
 /*
@@ -226,6 +225,9 @@  int hisi_uncore_pmu_event_init(struct perf_event *event)
 	hwc->idx		= -1;
 	hwc->config_base	= event->attr.config;
 
+	if (hisi_pmu->ops->check_filter && hisi_pmu->ops->check_filter(event))
+		return -EINVAL;
+
 	/* Enforce to use the same CPU for all events in this PMU */
 	event->cpu = hisi_pmu->on_cpu;
 
diff --git a/drivers/perf/hisilicon/hisi_uncore_pmu.h b/drivers/perf/hisilicon/hisi_uncore_pmu.h
index a8d6d6905f3f..27b6122aa486 100644
--- a/drivers/perf/hisilicon/hisi_uncore_pmu.h
+++ b/drivers/perf/hisilicon/hisi_uncore_pmu.h
@@ -44,9 +44,15 @@ 
 		return FIELD_GET(GENMASK_ULL(hi, lo), event->attr.config);  \
 	}
 
+#define HISI_GET_EVENTID(ev) (ev->hw.config_base & 0xff)
+
+#define HISI_PMU_EVTYPE_BITS		8
+#define HISI_PMU_EVTYPE_SHIFT(idx)	((idx) % 4 * HISI_PMU_EVTYPE_BITS)
+
 struct hisi_pmu;
 
 struct hisi_uncore_ops {
+	int (*check_filter)(struct perf_event *event);
 	void (*write_evtype)(struct hisi_pmu *, int, u32);
 	int (*get_event_idx)(struct perf_event *);
 	u64 (*read_counter)(struct hisi_pmu *, struct hw_perf_event *);
diff --git a/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
new file mode 100644
index 000000000000..d27f28584fd7
--- /dev/null
+++ b/drivers/perf/hisilicon/hisi_uncore_uc_pmu.c
@@ -0,0 +1,577 @@ 
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * HiSilicon SoC UC (unified cache) uncore Hardware event counters support
+ *
+ * Copyright (C) 2023 HiSilicon Limited
+ *
+ * This code is based on the uncore PMUs like hisi_uncore_l3c_pmu.
+ */
+#include <linux/cpuhotplug.h>
+#include <linux/interrupt.h>
+#include <linux/irq.h>
+#include <linux/list.h>
+#include <linux/mod_devicetable.h>
+#include <linux/property.h>
+
+#include "hisi_uncore_pmu.h"
+
+/* Dynamic CPU hotplug state used by UC PMU */
+static enum cpuhp_state hisi_uc_pmu_online;
+
+/* UC register definition */
+#define HISI_UC_INT_MASK_REG		0x0800
+#define HISI_UC_INT_STS_REG		0x0808
+#define HISI_UC_INT_CLEAR_REG		0x080c
+#define HISI_UC_TRACETAG_CTRL_REG	0x1b2c
+#define HISI_UC_TRACETAG_REQ_MSK	GENMASK(9, 7)
+#define HISI_UC_TRACETAG_MARK_EN	BIT(0)
+#define HISI_UC_TRACETAG_REQ_EN		(HISI_UC_TRACETAG_MARK_EN | BIT(2))
+#define HISI_UC_TRACETAG_SRCID_EN	BIT(3)
+#define HISI_UC_SRCID_CTRL_REG		0x1b40
+#define HISI_UC_SRCID_MSK		GENMASK(14, 1)
+#define HISI_UC_EVENT_CTRL_REG		0x1c00
+#define HISI_UC_EVENT_TRACETAG_EN	BIT(29)
+#define HISI_UC_EVENT_URING_MSK		GENMASK(28, 27)
+#define HISI_UC_EVENT_GLB_EN		BIT(26)
+#define HISI_UC_VERSION_REG		0x1cf0
+#define HISI_UC_EVTYPE_REGn(n)		(0x1d00 + (n) * 4)
+#define HISI_UC_EVTYPE_MASK		GENMASK(7, 0)
+#define HISI_UC_CNTR_REGn(n)		(0x1e00 + (n) * 8)
+
+#define HISI_UC_NR_COUNTERS		0x8
+#define HISI_UC_V2_NR_EVENTS		0xFF
+#define HISI_UC_CNTR_REG_BITS		64
+
+#define HISI_UC_RD_REQ_TRACETAG		0x4
+#define HISI_UC_URING_EVENT_MIN		0x47
+#define HISI_UC_URING_EVENT_MAX		0x59
+
+HISI_PMU_EVENT_ATTR_EXTRACTOR(rd_req_en, config1, 0, 0);
+HISI_PMU_EVENT_ATTR_EXTRACTOR(uring_channel, config1, 5, 4);
+HISI_PMU_EVENT_ATTR_EXTRACTOR(srcid, config1, 19, 6);
+HISI_PMU_EVENT_ATTR_EXTRACTOR(srcid_en, config1, 20, 20);
+
+static int hisi_uc_pmu_check_filter(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+
+	if (hisi_get_srcid_en(event) && !hisi_get_rd_req_en(event)) {
+		dev_err(uc_pmu->dev,
+			"Failed to set srcid: depending on read request enabled!\n");
+		return -EINVAL;
+	}
+
+	if (!hisi_get_uring_channel(event))
+		return 0;
+
+	if ((HISI_GET_EVENTID(event) < HISI_UC_URING_EVENT_MIN) ||
+	    (HISI_GET_EVENTID(event) > HISI_UC_URING_EVENT_MAX))
+		dev_warn(uc_pmu->dev,
+			 "Only events: [%#x ~ %#x] support channel filtering!",
+			 HISI_UC_URING_EVENT_MIN, HISI_UC_URING_EVENT_MAX);
+
+	return 0;
+}
+
+static void hisi_uc_pmu_config_req_tracetag(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 val;
+
+	if (!hisi_get_rd_req_en(event))
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	/* The request-type has been configured */
+	if (FIELD_GET(HISI_UC_TRACETAG_REQ_MSK, val) == HISI_UC_RD_REQ_TRACETAG)
+		return;
+
+	/* Set request-type for tracetag, only read request is supported! */
+	val &= ~HISI_UC_TRACETAG_REQ_MSK;
+	val |= FIELD_PREP(HISI_UC_TRACETAG_REQ_MSK, HISI_UC_RD_REQ_TRACETAG);
+	val |= HISI_UC_TRACETAG_REQ_EN;
+	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+}
+
+static void hisi_uc_pmu_clear_req_tracetag(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 val;
+
+	if (!hisi_get_rd_req_en(event))
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	/* Do nothing, the request-type tracetag has been cleaned up */
+	if (FIELD_GET(HISI_UC_TRACETAG_REQ_MSK, val) == 0)
+		return;
+
+	/* Clear request-type */
+	val &= ~HISI_UC_TRACETAG_REQ_MSK;
+	val &= ~HISI_UC_TRACETAG_REQ_EN;
+	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+}
+
+static void hisi_uc_pmu_config_srcid_tracetag(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 val;
+
+	if (!hisi_get_srcid_en(event))
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	/* Do nothing, the source id has been configured */
+	if (FIELD_GET(HISI_UC_TRACETAG_SRCID_EN, val))
+		return;
+
+	/* Enable source id tracetag */
+	val |= HISI_UC_TRACETAG_SRCID_EN;
+	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	val = readl(uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
+	val &= ~HISI_UC_SRCID_MSK;
+	val |= FIELD_PREP(HISI_UC_SRCID_MSK, hisi_get_srcid(event));
+	writel(val, uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
+
+	/* Depend on request-type tracetag enabled */
+	hisi_uc_pmu_config_req_tracetag(event);
+}
+
+static void hisi_uc_pmu_clear_srcid_tracetag(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 val;
+
+	if (!hisi_get_srcid_en(event))
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	/* Do nothing, the source id has been cleaned up */
+	if (FIELD_GET(HISI_UC_TRACETAG_SRCID_EN, val) == 0)
+		return;
+
+	hisi_uc_pmu_clear_req_tracetag(event);
+
+	/* Disable source id tracetag */
+	val &= ~HISI_UC_TRACETAG_SRCID_EN;
+	writel(val, uc_pmu->base + HISI_UC_TRACETAG_CTRL_REG);
+
+	val = readl(uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
+	val &= ~HISI_UC_SRCID_MSK;
+	writel(val, uc_pmu->base + HISI_UC_SRCID_CTRL_REG);
+}
+
+static void hisi_uc_pmu_config_uring_channel(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 uring_channel = hisi_get_uring_channel(event);
+	u32 val;
+
+	/* Do nothing if not being set or is set explicitly to zero (default) */
+	if (uring_channel == 0)
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+
+	/* Do nothing, the uring_channel has been configured */
+	if (uring_channel == FIELD_GET(HISI_UC_EVENT_URING_MSK, val))
+		return;
+
+	val &= ~HISI_UC_EVENT_URING_MSK;
+	val |= FIELD_PREP(HISI_UC_EVENT_URING_MSK, uring_channel);
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static void hisi_uc_pmu_clear_uring_channel(struct perf_event *event)
+{
+	struct hisi_pmu *uc_pmu = to_hisi_pmu(event->pmu);
+	u32 val;
+
+	/* Do nothing if not being set or is set explicitly to zero (default) */
+	if (hisi_get_uring_channel(event) == 0)
+		return;
+
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+
+	/* Do nothing, the uring_channel has been cleaned up */
+	if (FIELD_GET(HISI_UC_EVENT_URING_MSK, val) == 0)
+		return;
+
+	val &= ~HISI_UC_EVENT_URING_MSK;
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static void hisi_uc_pmu_enable_filter(struct perf_event *event)
+{
+	if (event->attr.config1 == 0)
+		return;
+
+	hisi_uc_pmu_config_uring_channel(event);
+	hisi_uc_pmu_config_req_tracetag(event);
+	hisi_uc_pmu_config_srcid_tracetag(event);
+}
+
+static void hisi_uc_pmu_disable_filter(struct perf_event *event)
+{
+	if (event->attr.config1 == 0)
+		return;
+
+	hisi_uc_pmu_clear_srcid_tracetag(event);
+	hisi_uc_pmu_clear_req_tracetag(event);
+	hisi_uc_pmu_clear_uring_channel(event);
+}
+
+static void hisi_uc_pmu_write_evtype(struct hisi_pmu *uc_pmu, int idx, u32 type)
+{
+	u32 val;
+
+	/*
+	 * Select the appropriate event select register.
+	 * There are 2 32-bit event select registers for the
+	 * 8 hardware counters, each event code is 8-bit wide.
+	 */
+	val = readl(uc_pmu->base + HISI_UC_EVTYPE_REGn(idx / 4));
+	val &= ~(HISI_UC_EVTYPE_MASK << HISI_PMU_EVTYPE_SHIFT(idx));
+	val |= (type << HISI_PMU_EVTYPE_SHIFT(idx));
+	writel(val, uc_pmu->base + HISI_UC_EVTYPE_REGn(idx / 4));
+}
+
+static void hisi_uc_pmu_start_counters(struct hisi_pmu *uc_pmu)
+{
+	u32 val;
+
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+	val |= HISI_UC_EVENT_GLB_EN;
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static void hisi_uc_pmu_stop_counters(struct hisi_pmu *uc_pmu)
+{
+	u32 val;
+
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+	val &= ~HISI_UC_EVENT_GLB_EN;
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static void hisi_uc_pmu_enable_counter(struct hisi_pmu *uc_pmu,
+					struct hw_perf_event *hwc)
+{
+	u32 val;
+
+	/* Enable counter index */
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+	val |= (1 << hwc->idx);
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static void hisi_uc_pmu_disable_counter(struct hisi_pmu *uc_pmu,
+					struct hw_perf_event *hwc)
+{
+	u32 val;
+
+	/* Clear counter index */
+	val = readl(uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+	val &= ~(1 << hwc->idx);
+	writel(val, uc_pmu->base + HISI_UC_EVENT_CTRL_REG);
+}
+
+static u64 hisi_uc_pmu_read_counter(struct hisi_pmu *uc_pmu,
+				    struct hw_perf_event *hwc)
+{
+	return readq(uc_pmu->base + HISI_UC_CNTR_REGn(hwc->idx));
+}
+
+static void hisi_uc_pmu_write_counter(struct hisi_pmu *uc_pmu,
+				      struct hw_perf_event *hwc, u64 val)
+{
+	hisi_uc_pmu_start_counters(uc_pmu);
+	writeq(val, uc_pmu->base + HISI_UC_CNTR_REGn(hwc->idx));
+}
+
+static void hisi_uc_pmu_enable_counter_int(struct hisi_pmu *uc_pmu,
+					   struct hw_perf_event *hwc)
+{
+	u32 val;
+
+	val = readl(uc_pmu->base + HISI_UC_INT_MASK_REG);
+	val &= ~(1 << hwc->idx);
+	writel(val, uc_pmu->base + HISI_UC_INT_MASK_REG);
+}
+
+static void hisi_uc_pmu_disable_counter_int(struct hisi_pmu *uc_pmu,
+					    struct hw_perf_event *hwc)
+{
+	u32 val;
+
+	val = readl(uc_pmu->base + HISI_UC_INT_MASK_REG);
+	val |= (1 << hwc->idx);
+	writel(val, uc_pmu->base + HISI_UC_INT_MASK_REG);
+}
+
+static u32 hisi_uc_pmu_get_int_status(struct hisi_pmu *uc_pmu)
+{
+	return readl(uc_pmu->base + HISI_UC_INT_STS_REG);
+}
+
+static void hisi_uc_pmu_clear_int_status(struct hisi_pmu *uc_pmu, int idx)
+{
+	writel(1 << idx, uc_pmu->base + HISI_UC_INT_CLEAR_REG);
+}
+
+static int hisi_uc_pmu_init_data(struct platform_device *pdev,
+				 struct hisi_pmu *uc_pmu)
+{
+	/*
+	 * Use SCCL (Super CPU Cluster) ID and CCL (CPU Cluster) ID to
+	 * identify the topology information of UC PMU devices in the chip.
+	 */
+	if (device_property_read_u32(&pdev->dev, "hisilicon,scl-id",
+				     &uc_pmu->sccl_id)) {
+		dev_err(&pdev->dev, "Can not read uc sccl-id!\n");
+		return -EINVAL;
+	}
+
+	if (device_property_read_u32(&pdev->dev, "hisilicon,ccl-id",
+				     &uc_pmu->ccl_id)) {
+		dev_err(&pdev->dev, "Can not read uc ccl-id!\n");
+		return -EINVAL;
+	}
+
+	if (device_property_read_u32(&pdev->dev, "hisilicon,sub-id",
+				     &uc_pmu->sub_id)) {
+		dev_err(&pdev->dev, "Can not read sub-id!\n");
+		return -EINVAL;
+	}
+
+	uc_pmu->base = devm_platform_ioremap_resource(pdev, 0);
+	if (IS_ERR(uc_pmu->base)) {
+		dev_err(&pdev->dev, "ioremap failed for uc_pmu resource\n");
+		return PTR_ERR(uc_pmu->base);
+	}
+
+	uc_pmu->identifier = readl(uc_pmu->base + HISI_UC_VERSION_REG);
+
+	return 0;
+}
+
+static struct attribute *hisi_uc_pmu_format_attr[] = {
+	HISI_PMU_FORMAT_ATTR(event, "config:0-7"),
+	HISI_PMU_FORMAT_ATTR(rd_req_en, "config1:0-0"),
+	HISI_PMU_FORMAT_ATTR(uring_channel, "config1:4-5"),
+	HISI_PMU_FORMAT_ATTR(srcid, "config1:6-19"),
+	HISI_PMU_FORMAT_ATTR(srcid_en, "config1:20-20"),
+	NULL
+};
+
+static const struct attribute_group hisi_uc_pmu_format_group = {
+	.name = "format",
+	.attrs = hisi_uc_pmu_format_attr,
+};
+
+static struct attribute *hisi_uc_pmu_events_attr[] = {
+	HISI_PMU_EVENT_ATTR(sq_time,		0x00),
+	HISI_PMU_EVENT_ATTR(pq_time,		0x01),
+	HISI_PMU_EVENT_ATTR(hbm_time,		0x02),
+	HISI_PMU_EVENT_ATTR(iq_comp_time_cring,	0x03),
+	HISI_PMU_EVENT_ATTR(iq_comp_time_uring,	0x05),
+	HISI_PMU_EVENT_ATTR(cpu_rd,		0x10),
+	HISI_PMU_EVENT_ATTR(cpu_rd64,		0x17),
+	HISI_PMU_EVENT_ATTR(cpu_rs64,		0x19),
+	HISI_PMU_EVENT_ATTR(cpu_mru,		0x1a),
+	HISI_PMU_EVENT_ATTR(cycles,		0x9c),
+	HISI_PMU_EVENT_ATTR(spipe_hit,		0xb3),
+	HISI_PMU_EVENT_ATTR(hpipe_hit,		0xdb),
+	HISI_PMU_EVENT_ATTR(cring_rxdat_cnt,	0xfa),
+	HISI_PMU_EVENT_ATTR(cring_txdat_cnt,	0xfb),
+	HISI_PMU_EVENT_ATTR(uring_rxdat_cnt,	0xfc),
+	HISI_PMU_EVENT_ATTR(uring_txdat_cnt,	0xfd),
+	NULL
+};
+
+static const struct attribute_group hisi_uc_pmu_events_group = {
+	.name = "events",
+	.attrs = hisi_uc_pmu_events_attr,
+};
+
+static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL);
+
+static struct attribute *hisi_uc_pmu_cpumask_attrs[] = {
+	&dev_attr_cpumask.attr,
+	NULL,
+};
+
+static const struct attribute_group hisi_uc_pmu_cpumask_attr_group = {
+	.attrs = hisi_uc_pmu_cpumask_attrs,
+};
+
+static struct device_attribute hisi_uc_pmu_identifier_attr =
+	__ATTR(identifier, 0444, hisi_uncore_pmu_identifier_attr_show, NULL);
+
+static struct attribute *hisi_uc_pmu_identifier_attrs[] = {
+	&hisi_uc_pmu_identifier_attr.attr,
+	NULL
+};
+
+static const struct attribute_group hisi_uc_pmu_identifier_group = {
+	.attrs = hisi_uc_pmu_identifier_attrs,
+};
+
+static const struct attribute_group *hisi_uc_pmu_attr_groups[] = {
+	&hisi_uc_pmu_format_group,
+	&hisi_uc_pmu_events_group,
+	&hisi_uc_pmu_cpumask_attr_group,
+	&hisi_uc_pmu_identifier_group,
+	NULL
+};
+
+static const struct hisi_uncore_ops hisi_uncore_uc_pmu_ops = {
+	.check_filter		= hisi_uc_pmu_check_filter,
+	.write_evtype		= hisi_uc_pmu_write_evtype,
+	.get_event_idx		= hisi_uncore_pmu_get_event_idx,
+	.start_counters		= hisi_uc_pmu_start_counters,
+	.stop_counters		= hisi_uc_pmu_stop_counters,
+	.enable_counter		= hisi_uc_pmu_enable_counter,
+	.disable_counter	= hisi_uc_pmu_disable_counter,
+	.enable_counter_int	= hisi_uc_pmu_enable_counter_int,
+	.disable_counter_int	= hisi_uc_pmu_disable_counter_int,
+	.write_counter		= hisi_uc_pmu_write_counter,
+	.read_counter		= hisi_uc_pmu_read_counter,
+	.get_int_status		= hisi_uc_pmu_get_int_status,
+	.clear_int_status	= hisi_uc_pmu_clear_int_status,
+	.enable_filter		= hisi_uc_pmu_enable_filter,
+	.disable_filter		= hisi_uc_pmu_disable_filter,
+};
+
+static int hisi_uc_pmu_dev_probe(struct platform_device *pdev,
+				 struct hisi_pmu *uc_pmu)
+{
+	int ret;
+
+	ret = hisi_uc_pmu_init_data(pdev, uc_pmu);
+	if (ret)
+		return ret;
+
+	ret = hisi_uncore_pmu_init_irq(uc_pmu, pdev);
+	if (ret)
+		return ret;
+
+	uc_pmu->pmu_events.attr_groups = hisi_uc_pmu_attr_groups;
+	uc_pmu->check_event = HISI_UC_EVTYPE_MASK;
+	uc_pmu->ops = &hisi_uncore_uc_pmu_ops;
+	uc_pmu->counter_bits = HISI_UC_CNTR_REG_BITS;
+	uc_pmu->num_counters = HISI_UC_NR_COUNTERS;
+	uc_pmu->dev = &pdev->dev;
+	uc_pmu->on_cpu = -1;
+
+	return 0;
+}
+
+static void hisi_uc_pmu_remove_cpuhp_instance(void *hotplug_node)
+{
+	cpuhp_state_remove_instance_nocalls(hisi_uc_pmu_online, hotplug_node);
+}
+
+static void hisi_uc_pmu_unregister_pmu(void *pmu)
+{
+	perf_pmu_unregister(pmu);
+}
+
+static int hisi_uc_pmu_probe(struct platform_device *pdev)
+{
+	struct hisi_pmu *uc_pmu;
+	char *name;
+	int ret;
+
+	uc_pmu = devm_kzalloc(&pdev->dev, sizeof(*uc_pmu), GFP_KERNEL);
+	if (!uc_pmu)
+		return -ENOMEM;
+
+	platform_set_drvdata(pdev, uc_pmu);
+
+	ret = hisi_uc_pmu_dev_probe(pdev, uc_pmu);
+	if (ret)
+		return ret;
+
+	name = devm_kasprintf(&pdev->dev, GFP_KERNEL, "hisi_sccl%d_uc%d_%u",
+			      uc_pmu->sccl_id, uc_pmu->ccl_id, uc_pmu->sub_id);
+	if (!name)
+		return -ENOMEM;
+
+	ret = cpuhp_state_add_instance(hisi_uc_pmu_online, &uc_pmu->node);
+	if (ret)
+		return dev_err_probe(&pdev->dev, ret, "Error registering hotplug\n");
+
+	ret = devm_add_action_or_reset(&pdev->dev,
+				       hisi_uc_pmu_remove_cpuhp_instance,
+				       &uc_pmu->node);
+	if (ret)
+		return ret;
+
+	hisi_pmu_init(uc_pmu, THIS_MODULE);
+
+	ret = perf_pmu_register(&uc_pmu->pmu, name, -1);
+	if (ret)
+		return ret;
+
+	return devm_add_action_or_reset(&pdev->dev,
+					hisi_uc_pmu_unregister_pmu,
+					&uc_pmu->pmu);
+}
+
+static const struct acpi_device_id hisi_uc_pmu_acpi_match[] = {
+	{ "HISI0291", },
+	{}
+};
+MODULE_DEVICE_TABLE(acpi, hisi_uc_pmu_acpi_match);
+
+static struct platform_driver hisi_uc_pmu_driver = {
+	.driver = {
+		.name = "hisi_uc_pmu",
+		.acpi_match_table = hisi_uc_pmu_acpi_match,
+		/*
+		 * We have not worked out a safe bind/unbind process,
+		 * so this is not supported yet.
+		 */
+		.suppress_bind_attrs = true,
+	},
+	.probe = hisi_uc_pmu_probe,
+};
+
+static int __init hisi_uc_pmu_module_init(void)
+{
+	int ret;
+
+	ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN,
+				      "perf/hisi/uc:online",
+				      hisi_uncore_pmu_online_cpu,
+				      hisi_uncore_pmu_offline_cpu);
+	if (ret < 0) {
+		pr_err("UC PMU: Error setup hotplug, ret = %d\n", ret);
+		return ret;
+	}
+	hisi_uc_pmu_online = ret;
+
+	ret = platform_driver_register(&hisi_uc_pmu_driver);
+	if (ret)
+		cpuhp_remove_multi_state(hisi_uc_pmu_online);
+
+	return ret;
+}
+module_init(hisi_uc_pmu_module_init);
+
+static void __exit hisi_uc_pmu_module_exit(void)
+{
+	platform_driver_unregister(&hisi_uc_pmu_driver);
+	cpuhp_remove_multi_state(hisi_uc_pmu_online);
+}
+module_exit(hisi_uc_pmu_module_exit);
+
+MODULE_DESCRIPTION("HiSilicon SoC UC uncore PMU driver");
+MODULE_LICENSE("GPL");
+MODULE_AUTHOR("Junhao He <hejunhao3@huawei.com>");