1 // SPDX-License-Identifier: GPL-2.0
3 * Alibaba DDR Sub-System Driveway PMU driver
5 * Copyright (C) 2022 Alibaba Inc
8 #define ALI_DRW_PMUNAME "ali_drw"
9 #define ALI_DRW_DRVNAME ALI_DRW_PMUNAME "_pmu"
10 #define pr_fmt(fmt) ALI_DRW_DRVNAME ": " fmt
12 #include <linux/acpi.h>
13 #include <linux/bitfield.h>
14 #include <linux/bitmap.h>
15 #include <linux/bitops.h>
16 #include <linux/cpuhotplug.h>
17 #include <linux/cpumask.h>
18 #include <linux/device.h>
19 #include <linux/errno.h>
20 #include <linux/interrupt.h>
21 #include <linux/irq.h>
22 #include <linux/kernel.h>
23 #include <linux/list.h>
24 #include <linux/module.h>
25 #include <linux/mutex.h>
26 #include <linux/perf_event.h>
27 #include <linux/platform_device.h>
28 #include <linux/printk.h>
29 #include <linux/rculist.h>
30 #include <linux/refcount.h>
33 #define ALI_DRW_PMU_COMMON_MAX_COUNTERS 16
34 #define ALI_DRW_PMU_TEST_SEL_COMMON_COUNTER_BASE 19
36 #define ALI_DRW_PMU_PA_SHIFT 12
37 #define ALI_DRW_PMU_CNT_INIT 0x00000000
38 #define ALI_DRW_CNT_MAX_PERIOD 0xffffffff
39 #define ALI_DRW_PMU_CYCLE_EVT_ID 0x80
41 #define ALI_DRW_PMU_CNT_CTRL 0xC00
42 #define ALI_DRW_PMU_CNT_RST BIT(2)
43 #define ALI_DRW_PMU_CNT_STOP BIT(1)
44 #define ALI_DRW_PMU_CNT_START BIT(0)
46 #define ALI_DRW_PMU_CNT_STATE 0xC04
47 #define ALI_DRW_PMU_TEST_CTRL 0xC08
48 #define ALI_DRW_PMU_CNT_PRELOAD 0xC0C
50 #define ALI_DRW_PMU_CYCLE_CNT_HIGH_MASK GENMASK(23, 0)
51 #define ALI_DRW_PMU_CYCLE_CNT_LOW_MASK GENMASK(31, 0)
52 #define ALI_DRW_PMU_CYCLE_CNT_HIGH 0xC10
53 #define ALI_DRW_PMU_CYCLE_CNT_LOW 0xC14
55 /* PMU EVENT SEL 0-3 are paired in 32-bit registers on a 4-byte stride */
56 #define ALI_DRW_PMU_EVENT_SEL0 0xC68
57 /* counter 0-3 use sel0, counter 4-7 use sel1...*/
58 #define ALI_DRW_PMU_EVENT_SELn(n) \
59 (ALI_DRW_PMU_EVENT_SEL0 + (n / 4) * 0x4)
60 #define ALI_DRW_PMCOM_CNT_EN BIT(7)
61 #define ALI_DRW_PMCOM_CNT_EVENT_MASK GENMASK(5, 0)
62 #define ALI_DRW_PMCOM_CNT_EVENT_OFFSET(n) \
65 /* PMU COMMON COUNTER 0-15, are paired in 32-bit registers on a 4-byte stride */
66 #define ALI_DRW_PMU_COMMON_COUNTER0 0xC78
67 #define ALI_DRW_PMU_COMMON_COUNTERn(n) \
68 (ALI_DRW_PMU_COMMON_COUNTER0 + 0x4 * (n))
70 #define ALI_DRW_PMU_OV_INTR_ENABLE_CTL 0xCB8
71 #define ALI_DRW_PMU_OV_INTR_DISABLE_CTL 0xCBC
72 #define ALI_DRW_PMU_OV_INTR_ENABLE_STATUS 0xCC0
73 #define ALI_DRW_PMU_OV_INTR_CLR 0xCC4
74 #define ALI_DRW_PMU_OV_INTR_STATUS 0xCC8
75 #define ALI_DRW_PMCOM_CNT_OV_INTR_MASK GENMASK(23, 8)
76 #define ALI_DRW_PMBW_CNT_OV_INTR_MASK GENMASK(7, 0)
77 #define ALI_DRW_PMU_OV_INTR_MASK GENMASK_ULL(63, 0)
79 static int ali_drw_cpuhp_state_num;
81 static LIST_HEAD(ali_drw_pmu_irqs);
82 static DEFINE_MUTEX(ali_drw_pmu_irqs_lock);
84 struct ali_drw_pmu_irq {
85 struct hlist_node node;
86 struct list_head irqs_node;
87 struct list_head pmus_node;
94 void __iomem *cfg_base;
97 struct list_head pmus_node;
98 struct ali_drw_pmu_irq *irq;
101 DECLARE_BITMAP(used_mask, ALI_DRW_PMU_COMMON_MAX_COUNTERS);
102 struct perf_event *events[ALI_DRW_PMU_COMMON_MAX_COUNTERS];
103 int evtids[ALI_DRW_PMU_COMMON_MAX_COUNTERS];
108 #define to_ali_drw_pmu(p) (container_of(p, struct ali_drw_pmu, pmu))
110 #define DRW_CONFIG_EVENTID GENMASK(7, 0)
111 #define GET_DRW_EVENTID(event) FIELD_GET(DRW_CONFIG_EVENTID, (event)->attr.config)
113 static ssize_t ali_drw_pmu_format_show(struct device *dev,
114 struct device_attribute *attr, char *buf)
116 struct dev_ext_attribute *eattr;
118 eattr = container_of(attr, struct dev_ext_attribute, attr);
120 return sprintf(buf, "%s\n", (char *)eattr->var);
124 * PMU event attributes
126 static ssize_t ali_drw_pmu_event_show(struct device *dev,
127 struct device_attribute *attr, char *page)
129 struct dev_ext_attribute *eattr;
131 eattr = container_of(attr, struct dev_ext_attribute, attr);
133 return sprintf(page, "config=0x%lx\n", (unsigned long)eattr->var);
136 #define ALI_DRW_PMU_ATTR(_name, _func, _config) \
137 (&((struct dev_ext_attribute[]) { \
138 { __ATTR(_name, 0444, _func, NULL), (void *)_config } \
141 #define ALI_DRW_PMU_FORMAT_ATTR(_name, _config) \
142 ALI_DRW_PMU_ATTR(_name, ali_drw_pmu_format_show, (void *)_config)
143 #define ALI_DRW_PMU_EVENT_ATTR(_name, _config) \
144 ALI_DRW_PMU_ATTR(_name, ali_drw_pmu_event_show, (unsigned long)_config)
146 static struct attribute *ali_drw_pmu_events_attrs[] = {
147 ALI_DRW_PMU_EVENT_ATTR(hif_rd_or_wr, 0x0),
148 ALI_DRW_PMU_EVENT_ATTR(hif_wr, 0x1),
149 ALI_DRW_PMU_EVENT_ATTR(hif_rd, 0x2),
150 ALI_DRW_PMU_EVENT_ATTR(hif_rmw, 0x3),
151 ALI_DRW_PMU_EVENT_ATTR(hif_hi_pri_rd, 0x4),
152 ALI_DRW_PMU_EVENT_ATTR(dfi_wr_data_cycles, 0x7),
153 ALI_DRW_PMU_EVENT_ATTR(dfi_rd_data_cycles, 0x8),
154 ALI_DRW_PMU_EVENT_ATTR(hpr_xact_when_critical, 0x9),
155 ALI_DRW_PMU_EVENT_ATTR(lpr_xact_when_critical, 0xA),
156 ALI_DRW_PMU_EVENT_ATTR(wr_xact_when_critical, 0xB),
157 ALI_DRW_PMU_EVENT_ATTR(op_is_activate, 0xC),
158 ALI_DRW_PMU_EVENT_ATTR(op_is_rd_or_wr, 0xD),
159 ALI_DRW_PMU_EVENT_ATTR(op_is_rd_activate, 0xE),
160 ALI_DRW_PMU_EVENT_ATTR(op_is_rd, 0xF),
161 ALI_DRW_PMU_EVENT_ATTR(op_is_wr, 0x10),
162 ALI_DRW_PMU_EVENT_ATTR(op_is_mwr, 0x11),
163 ALI_DRW_PMU_EVENT_ATTR(op_is_precharge, 0x12),
164 ALI_DRW_PMU_EVENT_ATTR(precharge_for_rdwr, 0x13),
165 ALI_DRW_PMU_EVENT_ATTR(precharge_for_other, 0x14),
166 ALI_DRW_PMU_EVENT_ATTR(rdwr_transitions, 0x15),
167 ALI_DRW_PMU_EVENT_ATTR(write_combine, 0x16),
168 ALI_DRW_PMU_EVENT_ATTR(war_hazard, 0x17),
169 ALI_DRW_PMU_EVENT_ATTR(raw_hazard, 0x18),
170 ALI_DRW_PMU_EVENT_ATTR(waw_hazard, 0x19),
171 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_selfref_rk0, 0x1A),
172 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_selfref_rk1, 0x1B),
173 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_selfref_rk2, 0x1C),
174 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_selfref_rk3, 0x1D),
175 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_powerdown_rk0, 0x1E),
176 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_powerdown_rk1, 0x1F),
177 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_powerdown_rk2, 0x20),
178 ALI_DRW_PMU_EVENT_ATTR(op_is_enter_powerdown_rk3, 0x21),
179 ALI_DRW_PMU_EVENT_ATTR(selfref_mode_rk0, 0x26),
180 ALI_DRW_PMU_EVENT_ATTR(selfref_mode_rk1, 0x27),
181 ALI_DRW_PMU_EVENT_ATTR(selfref_mode_rk2, 0x28),
182 ALI_DRW_PMU_EVENT_ATTR(selfref_mode_rk3, 0x29),
183 ALI_DRW_PMU_EVENT_ATTR(op_is_refresh, 0x2A),
184 ALI_DRW_PMU_EVENT_ATTR(op_is_crit_ref, 0x2B),
185 ALI_DRW_PMU_EVENT_ATTR(op_is_load_mode, 0x2D),
186 ALI_DRW_PMU_EVENT_ATTR(op_is_zqcl, 0x2E),
187 ALI_DRW_PMU_EVENT_ATTR(visible_window_limit_reached_rd, 0x30),
188 ALI_DRW_PMU_EVENT_ATTR(visible_window_limit_reached_wr, 0x31),
189 ALI_DRW_PMU_EVENT_ATTR(op_is_dqsosc_mpc, 0x34),
190 ALI_DRW_PMU_EVENT_ATTR(op_is_dqsosc_mrr, 0x35),
191 ALI_DRW_PMU_EVENT_ATTR(op_is_tcr_mrr, 0x36),
192 ALI_DRW_PMU_EVENT_ATTR(op_is_zqstart, 0x37),
193 ALI_DRW_PMU_EVENT_ATTR(op_is_zqlatch, 0x38),
194 ALI_DRW_PMU_EVENT_ATTR(chi_txreq, 0x39),
195 ALI_DRW_PMU_EVENT_ATTR(chi_txdat, 0x3A),
196 ALI_DRW_PMU_EVENT_ATTR(chi_rxdat, 0x3B),
197 ALI_DRW_PMU_EVENT_ATTR(chi_rxrsp, 0x3C),
198 ALI_DRW_PMU_EVENT_ATTR(tsz_vio, 0x3D),
199 ALI_DRW_PMU_EVENT_ATTR(cycle, 0x80),
203 static struct attribute_group ali_drw_pmu_events_attr_group = {
205 .attrs = ali_drw_pmu_events_attrs,
208 static struct attribute *ali_drw_pmu_format_attr[] = {
209 ALI_DRW_PMU_FORMAT_ATTR(event, "config:0-7"),
213 static const struct attribute_group ali_drw_pmu_format_group = {
215 .attrs = ali_drw_pmu_format_attr,
218 static ssize_t ali_drw_pmu_cpumask_show(struct device *dev,
219 struct device_attribute *attr,
222 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(dev_get_drvdata(dev));
224 return cpumap_print_to_pagebuf(true, buf, cpumask_of(drw_pmu->cpu));
227 static struct device_attribute ali_drw_pmu_cpumask_attr =
228 __ATTR(cpumask, 0444, ali_drw_pmu_cpumask_show, NULL);
230 static struct attribute *ali_drw_pmu_cpumask_attrs[] = {
231 &ali_drw_pmu_cpumask_attr.attr,
235 static const struct attribute_group ali_drw_pmu_cpumask_attr_group = {
236 .attrs = ali_drw_pmu_cpumask_attrs,
239 static const struct attribute_group *ali_drw_pmu_attr_groups[] = {
240 &ali_drw_pmu_events_attr_group,
241 &ali_drw_pmu_cpumask_attr_group,
242 &ali_drw_pmu_format_group,
246 /* find a counter for event, then in add func, hw.idx will equal to counter */
247 static int ali_drw_get_counter_idx(struct perf_event *event)
249 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
252 for (idx = 0; idx < ALI_DRW_PMU_COMMON_MAX_COUNTERS; ++idx) {
253 if (!test_and_set_bit(idx, drw_pmu->used_mask))
257 /* The counters are all in use. */
261 static u64 ali_drw_pmu_read_counter(struct perf_event *event)
263 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
264 u64 cycle_high, cycle_low;
266 if (GET_DRW_EVENTID(event) == ALI_DRW_PMU_CYCLE_EVT_ID) {
267 cycle_high = readl(drw_pmu->cfg_base + ALI_DRW_PMU_CYCLE_CNT_HIGH);
268 cycle_high &= ALI_DRW_PMU_CYCLE_CNT_HIGH_MASK;
269 cycle_low = readl(drw_pmu->cfg_base + ALI_DRW_PMU_CYCLE_CNT_LOW);
270 cycle_low &= ALI_DRW_PMU_CYCLE_CNT_LOW_MASK;
271 return (cycle_high << 32 | cycle_low);
274 return readl(drw_pmu->cfg_base +
275 ALI_DRW_PMU_COMMON_COUNTERn(event->hw.idx));
278 static void ali_drw_pmu_event_update(struct perf_event *event)
280 struct hw_perf_event *hwc = &event->hw;
281 u64 delta, prev, now;
284 prev = local64_read(&hwc->prev_count);
285 now = ali_drw_pmu_read_counter(event);
286 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev);
288 /* handle overflow. */
290 if (GET_DRW_EVENTID(event) == ALI_DRW_PMU_CYCLE_EVT_ID)
291 delta &= ALI_DRW_PMU_OV_INTR_MASK;
293 delta &= ALI_DRW_CNT_MAX_PERIOD;
294 local64_add(delta, &event->count);
297 static void ali_drw_pmu_event_set_period(struct perf_event *event)
300 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
302 /* set a preload counter for test purpose */
303 writel(ALI_DRW_PMU_TEST_SEL_COMMON_COUNTER_BASE + event->hw.idx,
304 drw_pmu->cfg_base + ALI_DRW_PMU_TEST_CTRL);
306 /* set conunter initial value */
307 pre_val = ALI_DRW_PMU_CNT_INIT;
308 writel(pre_val, drw_pmu->cfg_base + ALI_DRW_PMU_CNT_PRELOAD);
309 local64_set(&event->hw.prev_count, pre_val);
311 /* set sel mode to zero to start test */
312 writel(0x0, drw_pmu->cfg_base + ALI_DRW_PMU_TEST_CTRL);
315 static void ali_drw_pmu_enable_counter(struct perf_event *event)
317 u32 val, subval, reg, shift;
318 int counter = event->hw.idx;
319 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
321 reg = ALI_DRW_PMU_EVENT_SELn(counter);
322 val = readl(drw_pmu->cfg_base + reg);
323 subval = FIELD_PREP(ALI_DRW_PMCOM_CNT_EN, 1) |
324 FIELD_PREP(ALI_DRW_PMCOM_CNT_EVENT_MASK, drw_pmu->evtids[counter]);
326 shift = ALI_DRW_PMCOM_CNT_EVENT_OFFSET(counter);
327 val &= ~(GENMASK(7, 0) << shift);
328 val |= subval << shift;
330 writel(val, drw_pmu->cfg_base + reg);
333 static void ali_drw_pmu_disable_counter(struct perf_event *event)
335 u32 val, reg, subval, shift;
336 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
337 int counter = event->hw.idx;
339 reg = ALI_DRW_PMU_EVENT_SELn(counter);
340 val = readl(drw_pmu->cfg_base + reg);
341 subval = FIELD_PREP(ALI_DRW_PMCOM_CNT_EN, 0) |
342 FIELD_PREP(ALI_DRW_PMCOM_CNT_EVENT_MASK, 0);
344 shift = ALI_DRW_PMCOM_CNT_EVENT_OFFSET(counter);
345 val &= ~(GENMASK(7, 0) << shift);
346 val |= subval << shift;
348 writel(val, drw_pmu->cfg_base + reg);
351 static irqreturn_t ali_drw_pmu_isr(int irq_num, void *data)
353 struct ali_drw_pmu_irq *irq = data;
354 struct ali_drw_pmu *drw_pmu;
355 irqreturn_t ret = IRQ_NONE;
358 list_for_each_entry_rcu(drw_pmu, &irq->pmus_node, pmus_node) {
359 unsigned long status, clr_status;
360 struct perf_event *event;
363 for (idx = 0; idx < ALI_DRW_PMU_COMMON_MAX_COUNTERS; idx++) {
364 event = drw_pmu->events[idx];
367 ali_drw_pmu_disable_counter(event);
370 /* common counter intr status */
371 status = readl(drw_pmu->cfg_base + ALI_DRW_PMU_OV_INTR_STATUS);
372 status = FIELD_GET(ALI_DRW_PMCOM_CNT_OV_INTR_MASK, status);
374 for_each_set_bit(idx, &status,
375 ALI_DRW_PMU_COMMON_MAX_COUNTERS) {
376 event = drw_pmu->events[idx];
377 if (WARN_ON_ONCE(!event))
379 ali_drw_pmu_event_update(event);
380 ali_drw_pmu_event_set_period(event);
383 /* clear common counter intr status */
384 clr_status = FIELD_PREP(ALI_DRW_PMCOM_CNT_OV_INTR_MASK, 1);
386 drw_pmu->cfg_base + ALI_DRW_PMU_OV_INTR_CLR);
389 for (idx = 0; idx < ALI_DRW_PMU_COMMON_MAX_COUNTERS; idx++) {
390 event = drw_pmu->events[idx];
393 if (!(event->hw.state & PERF_HES_STOPPED))
394 ali_drw_pmu_enable_counter(event);
403 static struct ali_drw_pmu_irq *__ali_drw_pmu_init_irq(struct platform_device
407 struct ali_drw_pmu_irq *irq;
409 list_for_each_entry(irq, &ali_drw_pmu_irqs, irqs_node) {
410 if (irq->irq_num == irq_num
411 && refcount_inc_not_zero(&irq->refcount))
415 irq = kzalloc(sizeof(*irq), GFP_KERNEL);
417 return ERR_PTR(-ENOMEM);
419 INIT_LIST_HEAD(&irq->pmus_node);
421 /* Pick one CPU to be the preferred one to use */
422 irq->cpu = smp_processor_id();
423 refcount_set(&irq->refcount, 1);
426 * FIXME: one of DDRSS Driveway PMU overflow interrupt shares the same
427 * irq number with MPAM ERR_IRQ. To register DDRSS PMU and MPAM drivers
428 * successfully, add IRQF_SHARED flag. Howerer, PMU interrupt should not
429 * share with other component.
431 ret = devm_request_irq(&pdev->dev, irq_num, ali_drw_pmu_isr,
432 IRQF_SHARED, dev_name(&pdev->dev), irq);
435 "Fail to request IRQ:%d ret:%d\n", irq_num, ret);
439 ret = irq_set_affinity_hint(irq_num, cpumask_of(irq->cpu));
443 ret = cpuhp_state_add_instance_nocalls(ali_drw_cpuhp_state_num,
448 irq->irq_num = irq_num;
449 list_add(&irq->irqs_node, &ali_drw_pmu_irqs);
458 static int ali_drw_pmu_init_irq(struct ali_drw_pmu *drw_pmu,
459 struct platform_device *pdev)
462 struct ali_drw_pmu_irq *irq;
464 /* Read and init IRQ */
465 irq_num = platform_get_irq(pdev, 0);
469 mutex_lock(&ali_drw_pmu_irqs_lock);
470 irq = __ali_drw_pmu_init_irq(pdev, irq_num);
471 mutex_unlock(&ali_drw_pmu_irqs_lock);
478 mutex_lock(&ali_drw_pmu_irqs_lock);
479 list_add_rcu(&drw_pmu->pmus_node, &irq->pmus_node);
480 mutex_unlock(&ali_drw_pmu_irqs_lock);
485 static void ali_drw_pmu_uninit_irq(struct ali_drw_pmu *drw_pmu)
487 struct ali_drw_pmu_irq *irq = drw_pmu->irq;
489 mutex_lock(&ali_drw_pmu_irqs_lock);
490 list_del_rcu(&drw_pmu->pmus_node);
492 if (!refcount_dec_and_test(&irq->refcount)) {
493 mutex_unlock(&ali_drw_pmu_irqs_lock);
497 list_del(&irq->irqs_node);
498 mutex_unlock(&ali_drw_pmu_irqs_lock);
500 WARN_ON(irq_set_affinity_hint(irq->irq_num, NULL));
501 cpuhp_state_remove_instance_nocalls(ali_drw_cpuhp_state_num,
506 static int ali_drw_pmu_event_init(struct perf_event *event)
508 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
509 struct hw_perf_event *hwc = &event->hw;
510 struct perf_event *sibling;
511 struct device *dev = drw_pmu->pmu.dev;
513 if (event->attr.type != event->pmu->type)
516 if (is_sampling_event(event)) {
517 dev_err(dev, "Sampling not supported!\n");
521 if (event->attach_state & PERF_ATTACH_TASK) {
522 dev_err(dev, "Per-task counter cannot allocate!\n");
526 event->cpu = drw_pmu->cpu;
527 if (event->cpu < 0) {
528 dev_err(dev, "Per-task mode not supported!\n");
532 if (event->group_leader != event &&
533 !is_software_event(event->group_leader)) {
534 dev_err(dev, "driveway only allow one event!\n");
538 for_each_sibling_event(sibling, event->group_leader) {
539 if (sibling != event && !is_software_event(sibling)) {
540 dev_err(dev, "driveway event not allowed!\n");
545 /* reset all the pmu counters */
546 writel(ALI_DRW_PMU_CNT_RST, drw_pmu->cfg_base + ALI_DRW_PMU_CNT_CTRL);
553 static void ali_drw_pmu_start(struct perf_event *event, int flags)
555 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
559 if (GET_DRW_EVENTID(event) == ALI_DRW_PMU_CYCLE_EVT_ID) {
560 writel(ALI_DRW_PMU_CNT_START,
561 drw_pmu->cfg_base + ALI_DRW_PMU_CNT_CTRL);
565 ali_drw_pmu_event_set_period(event);
566 if (flags & PERF_EF_RELOAD) {
567 unsigned long prev_raw_count =
568 local64_read(&event->hw.prev_count);
569 writel(prev_raw_count,
570 drw_pmu->cfg_base + ALI_DRW_PMU_CNT_PRELOAD);
573 ali_drw_pmu_enable_counter(event);
575 writel(ALI_DRW_PMU_CNT_START, drw_pmu->cfg_base + ALI_DRW_PMU_CNT_CTRL);
578 static void ali_drw_pmu_stop(struct perf_event *event, int flags)
580 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
582 if (event->hw.state & PERF_HES_STOPPED)
585 if (GET_DRW_EVENTID(event) != ALI_DRW_PMU_CYCLE_EVT_ID)
586 ali_drw_pmu_disable_counter(event);
588 writel(ALI_DRW_PMU_CNT_STOP, drw_pmu->cfg_base + ALI_DRW_PMU_CNT_CTRL);
590 ali_drw_pmu_event_update(event);
591 event->hw.state |= PERF_HES_STOPPED | PERF_HES_UPTODATE;
594 static int ali_drw_pmu_add(struct perf_event *event, int flags)
596 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
597 struct hw_perf_event *hwc = &event->hw;
601 evtid = GET_DRW_EVENTID(event);
603 if (evtid != ALI_DRW_PMU_CYCLE_EVT_ID) {
604 idx = ali_drw_get_counter_idx(event);
607 drw_pmu->events[idx] = event;
608 drw_pmu->evtids[idx] = evtid;
612 hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE;
614 if (flags & PERF_EF_START)
615 ali_drw_pmu_start(event, PERF_EF_RELOAD);
617 /* Propagate our changes to the userspace mapping. */
618 perf_event_update_userpage(event);
623 static void ali_drw_pmu_del(struct perf_event *event, int flags)
625 struct ali_drw_pmu *drw_pmu = to_ali_drw_pmu(event->pmu);
626 struct hw_perf_event *hwc = &event->hw;
629 ali_drw_pmu_stop(event, PERF_EF_UPDATE);
631 if (idx >= 0 && idx < ALI_DRW_PMU_COMMON_MAX_COUNTERS) {
632 drw_pmu->events[idx] = NULL;
633 drw_pmu->evtids[idx] = 0;
634 clear_bit(idx, drw_pmu->used_mask);
637 perf_event_update_userpage(event);
640 static void ali_drw_pmu_read(struct perf_event *event)
642 ali_drw_pmu_event_update(event);
645 static int ali_drw_pmu_probe(struct platform_device *pdev)
647 struct ali_drw_pmu *drw_pmu;
648 struct resource *res;
652 drw_pmu = devm_kzalloc(&pdev->dev, sizeof(*drw_pmu), GFP_KERNEL);
656 drw_pmu->dev = &pdev->dev;
657 platform_set_drvdata(pdev, drw_pmu);
659 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
660 drw_pmu->cfg_base = devm_ioremap_resource(&pdev->dev, res);
661 if (!drw_pmu->cfg_base)
664 name = devm_kasprintf(drw_pmu->dev, GFP_KERNEL, "ali_drw_%llx",
665 (u64) (res->start >> ALI_DRW_PMU_PA_SHIFT));
669 writel(ALI_DRW_PMU_CNT_RST, drw_pmu->cfg_base + ALI_DRW_PMU_CNT_CTRL);
671 /* enable the generation of interrupt by all common counters */
672 writel(ALI_DRW_PMCOM_CNT_OV_INTR_MASK,
673 drw_pmu->cfg_base + ALI_DRW_PMU_OV_INTR_ENABLE_CTL);
675 /* clearing interrupt status */
676 writel(0xffffff, drw_pmu->cfg_base + ALI_DRW_PMU_OV_INTR_CLR);
678 drw_pmu->cpu = smp_processor_id();
680 ret = ali_drw_pmu_init_irq(drw_pmu, pdev);
684 drw_pmu->pmu = (struct pmu) {
685 .module = THIS_MODULE,
686 .task_ctx_nr = perf_invalid_context,
687 .event_init = ali_drw_pmu_event_init,
688 .add = ali_drw_pmu_add,
689 .del = ali_drw_pmu_del,
690 .start = ali_drw_pmu_start,
691 .stop = ali_drw_pmu_stop,
692 .read = ali_drw_pmu_read,
693 .attr_groups = ali_drw_pmu_attr_groups,
694 .capabilities = PERF_PMU_CAP_NO_EXCLUDE,
697 ret = perf_pmu_register(&drw_pmu->pmu, name, -1);
699 dev_err(drw_pmu->dev, "DRW Driveway PMU PMU register failed!\n");
700 ali_drw_pmu_uninit_irq(drw_pmu);
706 static int ali_drw_pmu_remove(struct platform_device *pdev)
708 struct ali_drw_pmu *drw_pmu = platform_get_drvdata(pdev);
710 /* disable the generation of interrupt by all common counters */
711 writel(ALI_DRW_PMCOM_CNT_OV_INTR_MASK,
712 drw_pmu->cfg_base + ALI_DRW_PMU_OV_INTR_DISABLE_CTL);
714 ali_drw_pmu_uninit_irq(drw_pmu);
715 perf_pmu_unregister(&drw_pmu->pmu);
720 static int ali_drw_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
722 struct ali_drw_pmu_irq *irq;
723 struct ali_drw_pmu *drw_pmu;
726 cpumask_t node_online_cpus;
728 irq = hlist_entry_safe(node, struct ali_drw_pmu_irq, node);
732 ret = cpumask_and(&node_online_cpus,
733 cpumask_of_node(cpu_to_node(cpu)), cpu_online_mask);
735 target = cpumask_any_but(&node_online_cpus, cpu);
737 target = cpumask_any_but(cpu_online_mask, cpu);
739 if (target >= nr_cpu_ids)
742 /* We're only reading, but this isn't the place to be involving RCU */
743 mutex_lock(&ali_drw_pmu_irqs_lock);
744 list_for_each_entry(drw_pmu, &irq->pmus_node, pmus_node)
745 perf_pmu_migrate_context(&drw_pmu->pmu, irq->cpu, target);
746 mutex_unlock(&ali_drw_pmu_irqs_lock);
748 WARN_ON(irq_set_affinity_hint(irq->irq_num, cpumask_of(target)));
755 * Due to historical reasons, the HID used in the production environment is
756 * ARMHD700, so we leave ARMHD700 as Compatible ID.
758 static const struct acpi_device_id ali_drw_acpi_match[] = {
764 MODULE_DEVICE_TABLE(acpi, ali_drw_acpi_match);
766 static struct platform_driver ali_drw_pmu_driver = {
768 .name = "ali_drw_pmu",
769 .acpi_match_table = ali_drw_acpi_match,
771 .probe = ali_drw_pmu_probe,
772 .remove = ali_drw_pmu_remove,
775 static int __init ali_drw_pmu_init(void)
779 ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN,
780 "ali_drw_pmu:online",
781 NULL, ali_drw_pmu_offline_cpu);
784 pr_err("DRW Driveway PMU: setup hotplug failed, ret = %d\n",
788 ali_drw_cpuhp_state_num = ret;
790 ret = platform_driver_register(&ali_drw_pmu_driver);
792 cpuhp_remove_multi_state(ali_drw_cpuhp_state_num);
797 static void __exit ali_drw_pmu_exit(void)
799 platform_driver_unregister(&ali_drw_pmu_driver);
800 cpuhp_remove_multi_state(ali_drw_cpuhp_state_num);
803 module_init(ali_drw_pmu_init);
804 module_exit(ali_drw_pmu_exit);
809 MODULE_DESCRIPTION("Alibaba DDR Sub-System Driveway PMU driver");
810 MODULE_LICENSE("GPL v2");