Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * CPU PMU driver for the Apple M1 and derivatives
4 *
5 * Copyright (C) 2021 Google LLC
6 *
7 * Author: Marc Zyngier <maz@kernel.org>
8 *
9 * Most of the information used in this driver was provided by the
10 * Asahi Linux project. The rest was experimentally discovered.
11 */
12
13#include <linux/of.h>
14#include <linux/perf/arm_pmu.h>
15#include <linux/platform_device.h>
16
17#include <asm/apple_m1_pmu.h>
18#include <asm/irq_regs.h>
19#include <asm/perf_event.h>
20
21#define M1_PMU_NR_COUNTERS 10
22
23#define M1_PMU_CFG_EVENT GENMASK(7, 0)
24
25#define ANY_BUT_0_1 GENMASK(9, 2)
26#define ONLY_2_TO_7 GENMASK(7, 2)
27#define ONLY_2_4_6 (BIT(2) | BIT(4) | BIT(6))
28#define ONLY_5_6_7 (BIT(5) | BIT(6) | BIT(7))
29
30/*
31 * Description of the events we actually know about, as well as those with
32 * a specific counter affinity. Yes, this is a grand total of two known
33 * counters, and the rest is anybody's guess.
34 *
35 * Not all counters can count all events. Counters #0 and #1 are wired to
36 * count cycles and instructions respectively, and some events have
37 * bizarre mappings (every other counter, or even *one* counter). These
38 * restrictions equally apply to both P and E cores.
39 *
40 * It is worth noting that the PMUs attached to P and E cores are likely
41 * to be different because the underlying uarches are different. At the
42 * moment, we don't really need to distinguish between the two because we
43 * know next to nothing about the events themselves, and we already have
44 * per cpu-type PMU abstractions.
45 *
46 * If we eventually find out that the events are different across
47 * implementations, we'll have to introduce per cpu-type tables.
48 */
49enum m1_pmu_events {
50 M1_PMU_PERFCTR_UNKNOWN_01 = 0x01,
51 M1_PMU_PERFCTR_CPU_CYCLES = 0x02,
52 M1_PMU_PERFCTR_INSTRUCTIONS = 0x8c,
53 M1_PMU_PERFCTR_UNKNOWN_8d = 0x8d,
54 M1_PMU_PERFCTR_UNKNOWN_8e = 0x8e,
55 M1_PMU_PERFCTR_UNKNOWN_8f = 0x8f,
56 M1_PMU_PERFCTR_UNKNOWN_90 = 0x90,
57 M1_PMU_PERFCTR_UNKNOWN_93 = 0x93,
58 M1_PMU_PERFCTR_UNKNOWN_94 = 0x94,
59 M1_PMU_PERFCTR_UNKNOWN_95 = 0x95,
60 M1_PMU_PERFCTR_UNKNOWN_96 = 0x96,
61 M1_PMU_PERFCTR_UNKNOWN_97 = 0x97,
62 M1_PMU_PERFCTR_UNKNOWN_98 = 0x98,
63 M1_PMU_PERFCTR_UNKNOWN_99 = 0x99,
64 M1_PMU_PERFCTR_UNKNOWN_9a = 0x9a,
65 M1_PMU_PERFCTR_UNKNOWN_9b = 0x9b,
66 M1_PMU_PERFCTR_UNKNOWN_9c = 0x9c,
67 M1_PMU_PERFCTR_UNKNOWN_9f = 0x9f,
68 M1_PMU_PERFCTR_UNKNOWN_bf = 0xbf,
69 M1_PMU_PERFCTR_UNKNOWN_c0 = 0xc0,
70 M1_PMU_PERFCTR_UNKNOWN_c1 = 0xc1,
71 M1_PMU_PERFCTR_UNKNOWN_c4 = 0xc4,
72 M1_PMU_PERFCTR_UNKNOWN_c5 = 0xc5,
73 M1_PMU_PERFCTR_UNKNOWN_c6 = 0xc6,
74 M1_PMU_PERFCTR_UNKNOWN_c8 = 0xc8,
75 M1_PMU_PERFCTR_UNKNOWN_ca = 0xca,
76 M1_PMU_PERFCTR_UNKNOWN_cb = 0xcb,
77 M1_PMU_PERFCTR_UNKNOWN_f5 = 0xf5,
78 M1_PMU_PERFCTR_UNKNOWN_f6 = 0xf6,
79 M1_PMU_PERFCTR_UNKNOWN_f7 = 0xf7,
80 M1_PMU_PERFCTR_UNKNOWN_f8 = 0xf8,
81 M1_PMU_PERFCTR_UNKNOWN_fd = 0xfd,
82 M1_PMU_PERFCTR_LAST = M1_PMU_CFG_EVENT,
83
84 /*
85 * From this point onwards, these are not actual HW events,
86 * but attributes that get stored in hw->config_base.
87 */
88 M1_PMU_CFG_COUNT_USER = BIT(8),
89 M1_PMU_CFG_COUNT_KERNEL = BIT(9),
90};
91
92/*
93 * Per-event affinity table. Most events can be installed on counter
94 * 2-9, but there are a number of exceptions. Note that this table
95 * has been created experimentally, and I wouldn't be surprised if more
96 * counters had strange affinities.
97 */
98static const u16 m1_pmu_event_affinity[M1_PMU_PERFCTR_LAST + 1] = {
99 [0 ... M1_PMU_PERFCTR_LAST] = ANY_BUT_0_1,
100 [M1_PMU_PERFCTR_UNKNOWN_01] = BIT(7),
101 [M1_PMU_PERFCTR_CPU_CYCLES] = ANY_BUT_0_1 | BIT(0),
102 [M1_PMU_PERFCTR_INSTRUCTIONS] = BIT(7) | BIT(1),
103 [M1_PMU_PERFCTR_UNKNOWN_8d] = ONLY_5_6_7,
104 [M1_PMU_PERFCTR_UNKNOWN_8e] = ONLY_5_6_7,
105 [M1_PMU_PERFCTR_UNKNOWN_8f] = ONLY_5_6_7,
106 [M1_PMU_PERFCTR_UNKNOWN_90] = ONLY_5_6_7,
107 [M1_PMU_PERFCTR_UNKNOWN_93] = ONLY_5_6_7,
108 [M1_PMU_PERFCTR_UNKNOWN_94] = ONLY_5_6_7,
109 [M1_PMU_PERFCTR_UNKNOWN_95] = ONLY_5_6_7,
110 [M1_PMU_PERFCTR_UNKNOWN_96] = ONLY_5_6_7,
111 [M1_PMU_PERFCTR_UNKNOWN_97] = BIT(7),
112 [M1_PMU_PERFCTR_UNKNOWN_98] = ONLY_5_6_7,
113 [M1_PMU_PERFCTR_UNKNOWN_99] = ONLY_5_6_7,
114 [M1_PMU_PERFCTR_UNKNOWN_9a] = BIT(7),
115 [M1_PMU_PERFCTR_UNKNOWN_9b] = ONLY_5_6_7,
116 [M1_PMU_PERFCTR_UNKNOWN_9c] = ONLY_5_6_7,
117 [M1_PMU_PERFCTR_UNKNOWN_9f] = BIT(7),
118 [M1_PMU_PERFCTR_UNKNOWN_bf] = ONLY_5_6_7,
119 [M1_PMU_PERFCTR_UNKNOWN_c0] = ONLY_5_6_7,
120 [M1_PMU_PERFCTR_UNKNOWN_c1] = ONLY_5_6_7,
121 [M1_PMU_PERFCTR_UNKNOWN_c4] = ONLY_5_6_7,
122 [M1_PMU_PERFCTR_UNKNOWN_c5] = ONLY_5_6_7,
123 [M1_PMU_PERFCTR_UNKNOWN_c6] = ONLY_5_6_7,
124 [M1_PMU_PERFCTR_UNKNOWN_c8] = ONLY_5_6_7,
125 [M1_PMU_PERFCTR_UNKNOWN_ca] = ONLY_5_6_7,
126 [M1_PMU_PERFCTR_UNKNOWN_cb] = ONLY_5_6_7,
127 [M1_PMU_PERFCTR_UNKNOWN_f5] = ONLY_2_4_6,
128 [M1_PMU_PERFCTR_UNKNOWN_f6] = ONLY_2_4_6,
129 [M1_PMU_PERFCTR_UNKNOWN_f7] = ONLY_2_4_6,
130 [M1_PMU_PERFCTR_UNKNOWN_f8] = ONLY_2_TO_7,
131 [M1_PMU_PERFCTR_UNKNOWN_fd] = ONLY_2_4_6,
132};
133
134static const unsigned m1_pmu_perf_map[PERF_COUNT_HW_MAX] = {
135 PERF_MAP_ALL_UNSUPPORTED,
136 [PERF_COUNT_HW_CPU_CYCLES] = M1_PMU_PERFCTR_CPU_CYCLES,
137 [PERF_COUNT_HW_INSTRUCTIONS] = M1_PMU_PERFCTR_INSTRUCTIONS,
138 /* No idea about the rest yet */
139};
140
141/* sysfs definitions */
142static ssize_t m1_pmu_events_sysfs_show(struct device *dev,
143 struct device_attribute *attr,
144 char *page)
145{
146 struct perf_pmu_events_attr *pmu_attr;
147
148 pmu_attr = container_of(attr, struct perf_pmu_events_attr, attr);
149
150 return sprintf(page, "event=0x%04llx\n", pmu_attr->id);
151}
152
153#define M1_PMU_EVENT_ATTR(name, config) \
154 PMU_EVENT_ATTR_ID(name, m1_pmu_events_sysfs_show, config)
155
156static struct attribute *m1_pmu_event_attrs[] = {
157 M1_PMU_EVENT_ATTR(cycles, M1_PMU_PERFCTR_CPU_CYCLES),
158 M1_PMU_EVENT_ATTR(instructions, M1_PMU_PERFCTR_INSTRUCTIONS),
159 NULL,
160};
161
162static const struct attribute_group m1_pmu_events_attr_group = {
163 .name = "events",
164 .attrs = m1_pmu_event_attrs,
165};
166
167PMU_FORMAT_ATTR(event, "config:0-7");
168
169static struct attribute *m1_pmu_format_attrs[] = {
170 &format_attr_event.attr,
171 NULL,
172};
173
174static const struct attribute_group m1_pmu_format_attr_group = {
175 .name = "format",
176 .attrs = m1_pmu_format_attrs,
177};
178
179/* Low level accessors. No synchronisation. */
180#define PMU_READ_COUNTER(_idx) \
181 case _idx: return read_sysreg_s(SYS_IMP_APL_PMC## _idx ##_EL1)
182
183#define PMU_WRITE_COUNTER(_val, _idx) \
184 case _idx: \
185 write_sysreg_s(_val, SYS_IMP_APL_PMC## _idx ##_EL1); \
186 return
187
188static u64 m1_pmu_read_hw_counter(unsigned int index)
189{
190 switch (index) {
191 PMU_READ_COUNTER(0);
192 PMU_READ_COUNTER(1);
193 PMU_READ_COUNTER(2);
194 PMU_READ_COUNTER(3);
195 PMU_READ_COUNTER(4);
196 PMU_READ_COUNTER(5);
197 PMU_READ_COUNTER(6);
198 PMU_READ_COUNTER(7);
199 PMU_READ_COUNTER(8);
200 PMU_READ_COUNTER(9);
201 }
202
203 BUG();
204}
205
206static void m1_pmu_write_hw_counter(u64 val, unsigned int index)
207{
208 switch (index) {
209 PMU_WRITE_COUNTER(val, 0);
210 PMU_WRITE_COUNTER(val, 1);
211 PMU_WRITE_COUNTER(val, 2);
212 PMU_WRITE_COUNTER(val, 3);
213 PMU_WRITE_COUNTER(val, 4);
214 PMU_WRITE_COUNTER(val, 5);
215 PMU_WRITE_COUNTER(val, 6);
216 PMU_WRITE_COUNTER(val, 7);
217 PMU_WRITE_COUNTER(val, 8);
218 PMU_WRITE_COUNTER(val, 9);
219 }
220
221 BUG();
222}
223
224#define get_bit_offset(index, mask) (__ffs(mask) + (index))
225
226static void __m1_pmu_enable_counter(unsigned int index, bool en)
227{
228 u64 val, bit;
229
230 switch (index) {
231 case 0 ... 7:
232 bit = BIT(get_bit_offset(index, PMCR0_CNT_ENABLE_0_7));
233 break;
234 case 8 ... 9:
235 bit = BIT(get_bit_offset(index - 8, PMCR0_CNT_ENABLE_8_9));
236 break;
237 default:
238 BUG();
239 }
240
241 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
242
243 if (en)
244 val |= bit;
245 else
246 val &= ~bit;
247
248 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
249}
250
251static void m1_pmu_enable_counter(unsigned int index)
252{
253 __m1_pmu_enable_counter(index, true);
254}
255
256static void m1_pmu_disable_counter(unsigned int index)
257{
258 __m1_pmu_enable_counter(index, false);
259}
260
261static void __m1_pmu_enable_counter_interrupt(unsigned int index, bool en)
262{
263 u64 val, bit;
264
265 switch (index) {
266 case 0 ... 7:
267 bit = BIT(get_bit_offset(index, PMCR0_PMI_ENABLE_0_7));
268 break;
269 case 8 ... 9:
270 bit = BIT(get_bit_offset(index - 8, PMCR0_PMI_ENABLE_8_9));
271 break;
272 default:
273 BUG();
274 }
275
276 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
277
278 if (en)
279 val |= bit;
280 else
281 val &= ~bit;
282
283 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
284}
285
286static void m1_pmu_enable_counter_interrupt(unsigned int index)
287{
288 __m1_pmu_enable_counter_interrupt(index, true);
289}
290
291static void m1_pmu_disable_counter_interrupt(unsigned int index)
292{
293 __m1_pmu_enable_counter_interrupt(index, false);
294}
295
296static void m1_pmu_configure_counter(unsigned int index, u8 event,
297 bool user, bool kernel)
298{
299 u64 val, user_bit, kernel_bit;
300 int shift;
301
302 switch (index) {
303 case 0 ... 7:
304 user_bit = BIT(get_bit_offset(index, PMCR1_COUNT_A64_EL0_0_7));
305 kernel_bit = BIT(get_bit_offset(index, PMCR1_COUNT_A64_EL1_0_7));
306 break;
307 case 8 ... 9:
308 user_bit = BIT(get_bit_offset(index - 8, PMCR1_COUNT_A64_EL0_8_9));
309 kernel_bit = BIT(get_bit_offset(index - 8, PMCR1_COUNT_A64_EL1_8_9));
310 break;
311 default:
312 BUG();
313 }
314
315 val = read_sysreg_s(SYS_IMP_APL_PMCR1_EL1);
316
317 if (user)
318 val |= user_bit;
319 else
320 val &= ~user_bit;
321
322 if (kernel)
323 val |= kernel_bit;
324 else
325 val &= ~kernel_bit;
326
327 write_sysreg_s(val, SYS_IMP_APL_PMCR1_EL1);
328
329 /*
330 * Counters 0 and 1 have fixed events. For anything else,
331 * place the event at the expected location in the relevant
332 * register (PMESR0 holds the event configuration for counters
333 * 2-5, resp. PMESR1 for counters 6-9).
334 */
335 switch (index) {
336 case 0 ... 1:
337 break;
338 case 2 ... 5:
339 shift = (index - 2) * 8;
340 val = read_sysreg_s(SYS_IMP_APL_PMESR0_EL1);
341 val &= ~((u64)0xff << shift);
342 val |= (u64)event << shift;
343 write_sysreg_s(val, SYS_IMP_APL_PMESR0_EL1);
344 break;
345 case 6 ... 9:
346 shift = (index - 6) * 8;
347 val = read_sysreg_s(SYS_IMP_APL_PMESR1_EL1);
348 val &= ~((u64)0xff << shift);
349 val |= (u64)event << shift;
350 write_sysreg_s(val, SYS_IMP_APL_PMESR1_EL1);
351 break;
352 }
353}
354
355/* arm_pmu backend */
356static void m1_pmu_enable_event(struct perf_event *event)
357{
358 bool user, kernel;
359 u8 evt;
360
361 evt = event->hw.config_base & M1_PMU_CFG_EVENT;
362 user = event->hw.config_base & M1_PMU_CFG_COUNT_USER;
363 kernel = event->hw.config_base & M1_PMU_CFG_COUNT_KERNEL;
364
365 m1_pmu_disable_counter_interrupt(event->hw.idx);
366 m1_pmu_disable_counter(event->hw.idx);
367 isb();
368
369 m1_pmu_configure_counter(event->hw.idx, evt, user, kernel);
370 m1_pmu_enable_counter(event->hw.idx);
371 m1_pmu_enable_counter_interrupt(event->hw.idx);
372 isb();
373}
374
375static void m1_pmu_disable_event(struct perf_event *event)
376{
377 m1_pmu_disable_counter_interrupt(event->hw.idx);
378 m1_pmu_disable_counter(event->hw.idx);
379 isb();
380}
381
382static irqreturn_t m1_pmu_handle_irq(struct arm_pmu *cpu_pmu)
383{
384 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events);
385 struct pt_regs *regs;
386 u64 overflow, state;
387 int idx;
388
389 overflow = read_sysreg_s(SYS_IMP_APL_PMSR_EL1);
390 if (!overflow) {
391 /* Spurious interrupt? */
392 state = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
393 state &= ~PMCR0_IACT;
394 write_sysreg_s(state, SYS_IMP_APL_PMCR0_EL1);
395 isb();
396 return IRQ_NONE;
397 }
398
399 cpu_pmu->stop(cpu_pmu);
400
401 regs = get_irq_regs();
402
403 for (idx = 0; idx < cpu_pmu->num_events; idx++) {
404 struct perf_event *event = cpuc->events[idx];
405 struct perf_sample_data data;
406
407 if (!event)
408 continue;
409
410 armpmu_event_update(event);
411 perf_sample_data_init(&data, 0, event->hw.last_period);
412 if (!armpmu_event_set_period(event))
413 continue;
414
415 if (perf_event_overflow(event, &data, regs))
416 m1_pmu_disable_event(event);
417 }
418
419 cpu_pmu->start(cpu_pmu);
420
421 return IRQ_HANDLED;
422}
423
424static u64 m1_pmu_read_counter(struct perf_event *event)
425{
426 return m1_pmu_read_hw_counter(event->hw.idx);
427}
428
429static void m1_pmu_write_counter(struct perf_event *event, u64 value)
430{
431 m1_pmu_write_hw_counter(value, event->hw.idx);
432 isb();
433}
434
435static int m1_pmu_get_event_idx(struct pmu_hw_events *cpuc,
436 struct perf_event *event)
437{
438 unsigned long evtype = event->hw.config_base & M1_PMU_CFG_EVENT;
439 unsigned long affinity = m1_pmu_event_affinity[evtype];
440 int idx;
441
442 /*
443 * Place the event on the first free counter that can count
444 * this event.
445 *
446 * We could do a better job if we had a view of all the events
447 * counting on the PMU at any given time, and by placing the
448 * most constraining events first.
449 */
450 for_each_set_bit(idx, &affinity, M1_PMU_NR_COUNTERS) {
451 if (!test_and_set_bit(idx, cpuc->used_mask))
452 return idx;
453 }
454
455 return -EAGAIN;
456}
457
458static void m1_pmu_clear_event_idx(struct pmu_hw_events *cpuc,
459 struct perf_event *event)
460{
461 clear_bit(event->hw.idx, cpuc->used_mask);
462}
463
464static void __m1_pmu_set_mode(u8 mode)
465{
466 u64 val;
467
468 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
469 val &= ~(PMCR0_IMODE | PMCR0_IACT);
470 val |= FIELD_PREP(PMCR0_IMODE, mode);
471 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
472 isb();
473}
474
475static void m1_pmu_start(struct arm_pmu *cpu_pmu)
476{
477 __m1_pmu_set_mode(PMCR0_IMODE_FIQ);
478}
479
480static void m1_pmu_stop(struct arm_pmu *cpu_pmu)
481{
482 __m1_pmu_set_mode(PMCR0_IMODE_OFF);
483}
484
485static int m1_pmu_map_event(struct perf_event *event)
486{
487 /*
488 * Although the counters are 48bit wide, bit 47 is what
489 * triggers the overflow interrupt. Advertise the counters
490 * being 47bit wide to mimick the behaviour of the ARM PMU.
491 */
492 event->hw.flags |= ARMPMU_EVT_47BIT;
493 return armpmu_map_event(event, &m1_pmu_perf_map, NULL, M1_PMU_CFG_EVENT);
494}
495
496static void m1_pmu_reset(void *info)
497{
498 int i;
499
500 __m1_pmu_set_mode(PMCR0_IMODE_OFF);
501
502 for (i = 0; i < M1_PMU_NR_COUNTERS; i++) {
503 m1_pmu_disable_counter(i);
504 m1_pmu_disable_counter_interrupt(i);
505 m1_pmu_write_hw_counter(0, i);
506 }
507
508 isb();
509}
510
511static int m1_pmu_set_event_filter(struct hw_perf_event *event,
512 struct perf_event_attr *attr)
513{
514 unsigned long config_base = 0;
515
516 if (!attr->exclude_guest)
517 return -EINVAL;
518 if (!attr->exclude_kernel)
519 config_base |= M1_PMU_CFG_COUNT_KERNEL;
520 if (!attr->exclude_user)
521 config_base |= M1_PMU_CFG_COUNT_USER;
522
523 event->config_base = config_base;
524
525 return 0;
526}
527
528static int m1_pmu_init(struct arm_pmu *cpu_pmu)
529{
530 cpu_pmu->handle_irq = m1_pmu_handle_irq;
531 cpu_pmu->enable = m1_pmu_enable_event;
532 cpu_pmu->disable = m1_pmu_disable_event;
533 cpu_pmu->read_counter = m1_pmu_read_counter;
534 cpu_pmu->write_counter = m1_pmu_write_counter;
535 cpu_pmu->get_event_idx = m1_pmu_get_event_idx;
536 cpu_pmu->clear_event_idx = m1_pmu_clear_event_idx;
537 cpu_pmu->start = m1_pmu_start;
538 cpu_pmu->stop = m1_pmu_stop;
539 cpu_pmu->map_event = m1_pmu_map_event;
540 cpu_pmu->reset = m1_pmu_reset;
541 cpu_pmu->set_event_filter = m1_pmu_set_event_filter;
542
543 cpu_pmu->num_events = M1_PMU_NR_COUNTERS;
544 cpu_pmu->attr_groups[ARMPMU_ATTR_GROUP_EVENTS] = &m1_pmu_events_attr_group;
545 cpu_pmu->attr_groups[ARMPMU_ATTR_GROUP_FORMATS] = &m1_pmu_format_attr_group;
546 return 0;
547}
548
549/* Device driver gunk */
550static int m1_pmu_ice_init(struct arm_pmu *cpu_pmu)
551{
552 cpu_pmu->name = "apple_icestorm_pmu";
553 return m1_pmu_init(cpu_pmu);
554}
555
556static int m1_pmu_fire_init(struct arm_pmu *cpu_pmu)
557{
558 cpu_pmu->name = "apple_firestorm_pmu";
559 return m1_pmu_init(cpu_pmu);
560}
561
562static const struct of_device_id m1_pmu_of_device_ids[] = {
563 { .compatible = "apple,icestorm-pmu", .data = m1_pmu_ice_init, },
564 { .compatible = "apple,firestorm-pmu", .data = m1_pmu_fire_init, },
565 { },
566};
567MODULE_DEVICE_TABLE(of, m1_pmu_of_device_ids);
568
569static int m1_pmu_device_probe(struct platform_device *pdev)
570{
571 return arm_pmu_device_probe(pdev, m1_pmu_of_device_ids, NULL);
572}
573
574static struct platform_driver m1_pmu_driver = {
575 .driver = {
576 .name = "apple-m1-cpu-pmu",
577 .of_match_table = m1_pmu_of_device_ids,
578 .suppress_bind_attrs = true,
579 },
580 .probe = m1_pmu_device_probe,
581};
582
583module_platform_driver(m1_pmu_driver);
584MODULE_LICENSE("GPL v2");
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * CPU PMU driver for the Apple M1 and derivatives
4 *
5 * Copyright (C) 2021 Google LLC
6 *
7 * Author: Marc Zyngier <maz@kernel.org>
8 *
9 * Most of the information used in this driver was provided by the
10 * Asahi Linux project. The rest was experimentally discovered.
11 */
12
13#include <linux/of.h>
14#include <linux/perf/arm_pmu.h>
15#include <linux/platform_device.h>
16
17#include <asm/apple_m1_pmu.h>
18#include <asm/irq_regs.h>
19#include <asm/perf_event.h>
20
21#define M1_PMU_NR_COUNTERS 10
22
23#define M1_PMU_CFG_EVENT GENMASK(7, 0)
24
25#define ANY_BUT_0_1 GENMASK(9, 2)
26#define ONLY_2_TO_7 GENMASK(7, 2)
27#define ONLY_2_4_6 (BIT(2) | BIT(4) | BIT(6))
28#define ONLY_5_6_7 (BIT(5) | BIT(6) | BIT(7))
29
30/*
31 * Description of the events we actually know about, as well as those with
32 * a specific counter affinity. Yes, this is a grand total of two known
33 * counters, and the rest is anybody's guess.
34 *
35 * Not all counters can count all events. Counters #0 and #1 are wired to
36 * count cycles and instructions respectively, and some events have
37 * bizarre mappings (every other counter, or even *one* counter). These
38 * restrictions equally apply to both P and E cores.
39 *
40 * It is worth noting that the PMUs attached to P and E cores are likely
41 * to be different because the underlying uarches are different. At the
42 * moment, we don't really need to distinguish between the two because we
43 * know next to nothing about the events themselves, and we already have
44 * per cpu-type PMU abstractions.
45 *
46 * If we eventually find out that the events are different across
47 * implementations, we'll have to introduce per cpu-type tables.
48 */
49enum m1_pmu_events {
50 M1_PMU_PERFCTR_RETIRE_UOP = 0x1,
51 M1_PMU_PERFCTR_CORE_ACTIVE_CYCLE = 0x2,
52 M1_PMU_PERFCTR_L1I_TLB_FILL = 0x4,
53 M1_PMU_PERFCTR_L1D_TLB_FILL = 0x5,
54 M1_PMU_PERFCTR_MMU_TABLE_WALK_INSTRUCTION = 0x7,
55 M1_PMU_PERFCTR_MMU_TABLE_WALK_DATA = 0x8,
56 M1_PMU_PERFCTR_L2_TLB_MISS_INSTRUCTION = 0xa,
57 M1_PMU_PERFCTR_L2_TLB_MISS_DATA = 0xb,
58 M1_PMU_PERFCTR_MMU_VIRTUAL_MEMORY_FAULT_NONSPEC = 0xd,
59 M1_PMU_PERFCTR_SCHEDULE_UOP = 0x52,
60 M1_PMU_PERFCTR_INTERRUPT_PENDING = 0x6c,
61 M1_PMU_PERFCTR_MAP_STALL_DISPATCH = 0x70,
62 M1_PMU_PERFCTR_MAP_REWIND = 0x75,
63 M1_PMU_PERFCTR_MAP_STALL = 0x76,
64 M1_PMU_PERFCTR_MAP_INT_UOP = 0x7c,
65 M1_PMU_PERFCTR_MAP_LDST_UOP = 0x7d,
66 M1_PMU_PERFCTR_MAP_SIMD_UOP = 0x7e,
67 M1_PMU_PERFCTR_FLUSH_RESTART_OTHER_NONSPEC = 0x84,
68 M1_PMU_PERFCTR_INST_ALL = 0x8c,
69 M1_PMU_PERFCTR_INST_BRANCH = 0x8d,
70 M1_PMU_PERFCTR_INST_BRANCH_CALL = 0x8e,
71 M1_PMU_PERFCTR_INST_BRANCH_RET = 0x8f,
72 M1_PMU_PERFCTR_INST_BRANCH_TAKEN = 0x90,
73 M1_PMU_PERFCTR_INST_BRANCH_INDIR = 0x93,
74 M1_PMU_PERFCTR_INST_BRANCH_COND = 0x94,
75 M1_PMU_PERFCTR_INST_INT_LD = 0x95,
76 M1_PMU_PERFCTR_INST_INT_ST = 0x96,
77 M1_PMU_PERFCTR_INST_INT_ALU = 0x97,
78 M1_PMU_PERFCTR_INST_SIMD_LD = 0x98,
79 M1_PMU_PERFCTR_INST_SIMD_ST = 0x99,
80 M1_PMU_PERFCTR_INST_SIMD_ALU = 0x9a,
81 M1_PMU_PERFCTR_INST_LDST = 0x9b,
82 M1_PMU_PERFCTR_INST_BARRIER = 0x9c,
83 M1_PMU_PERFCTR_UNKNOWN_9f = 0x9f,
84 M1_PMU_PERFCTR_L1D_TLB_ACCESS = 0xa0,
85 M1_PMU_PERFCTR_L1D_TLB_MISS = 0xa1,
86 M1_PMU_PERFCTR_L1D_CACHE_MISS_ST = 0xa2,
87 M1_PMU_PERFCTR_L1D_CACHE_MISS_LD = 0xa3,
88 M1_PMU_PERFCTR_LD_UNIT_UOP = 0xa6,
89 M1_PMU_PERFCTR_ST_UNIT_UOP = 0xa7,
90 M1_PMU_PERFCTR_L1D_CACHE_WRITEBACK = 0xa8,
91 M1_PMU_PERFCTR_LDST_X64_UOP = 0xb1,
92 M1_PMU_PERFCTR_LDST_XPG_UOP = 0xb2,
93 M1_PMU_PERFCTR_ATOMIC_OR_EXCLUSIVE_SUCC = 0xb3,
94 M1_PMU_PERFCTR_ATOMIC_OR_EXCLUSIVE_FAIL = 0xb4,
95 M1_PMU_PERFCTR_L1D_CACHE_MISS_LD_NONSPEC = 0xbf,
96 M1_PMU_PERFCTR_L1D_CACHE_MISS_ST_NONSPEC = 0xc0,
97 M1_PMU_PERFCTR_L1D_TLB_MISS_NONSPEC = 0xc1,
98 M1_PMU_PERFCTR_ST_MEMORY_ORDER_VIOLATION_NONSPEC = 0xc4,
99 M1_PMU_PERFCTR_BRANCH_COND_MISPRED_NONSPEC = 0xc5,
100 M1_PMU_PERFCTR_BRANCH_INDIR_MISPRED_NONSPEC = 0xc6,
101 M1_PMU_PERFCTR_BRANCH_RET_INDIR_MISPRED_NONSPEC = 0xc8,
102 M1_PMU_PERFCTR_BRANCH_CALL_INDIR_MISPRED_NONSPEC = 0xca,
103 M1_PMU_PERFCTR_BRANCH_MISPRED_NONSPEC = 0xcb,
104 M1_PMU_PERFCTR_L1I_TLB_MISS_DEMAND = 0xd4,
105 M1_PMU_PERFCTR_MAP_DISPATCH_BUBBLE = 0xd6,
106 M1_PMU_PERFCTR_L1I_CACHE_MISS_DEMAND = 0xdb,
107 M1_PMU_PERFCTR_FETCH_RESTART = 0xde,
108 M1_PMU_PERFCTR_ST_NT_UOP = 0xe5,
109 M1_PMU_PERFCTR_LD_NT_UOP = 0xe6,
110 M1_PMU_PERFCTR_UNKNOWN_f5 = 0xf5,
111 M1_PMU_PERFCTR_UNKNOWN_f6 = 0xf6,
112 M1_PMU_PERFCTR_UNKNOWN_f7 = 0xf7,
113 M1_PMU_PERFCTR_UNKNOWN_f8 = 0xf8,
114 M1_PMU_PERFCTR_UNKNOWN_fd = 0xfd,
115 M1_PMU_PERFCTR_LAST = M1_PMU_CFG_EVENT,
116
117 /*
118 * From this point onwards, these are not actual HW events,
119 * but attributes that get stored in hw->config_base.
120 */
121 M1_PMU_CFG_COUNT_USER = BIT(8),
122 M1_PMU_CFG_COUNT_KERNEL = BIT(9),
123};
124
125/*
126 * Per-event affinity table. Most events can be installed on counter
127 * 2-9, but there are a number of exceptions. Note that this table
128 * has been created experimentally, and I wouldn't be surprised if more
129 * counters had strange affinities.
130 */
131static const u16 m1_pmu_event_affinity[M1_PMU_PERFCTR_LAST + 1] = {
132 [0 ... M1_PMU_PERFCTR_LAST] = ANY_BUT_0_1,
133 [M1_PMU_PERFCTR_RETIRE_UOP] = BIT(7),
134 [M1_PMU_PERFCTR_CORE_ACTIVE_CYCLE] = ANY_BUT_0_1 | BIT(0),
135 [M1_PMU_PERFCTR_INST_ALL] = BIT(7) | BIT(1),
136 [M1_PMU_PERFCTR_INST_BRANCH] = ONLY_5_6_7,
137 [M1_PMU_PERFCTR_INST_BRANCH_CALL] = ONLY_5_6_7,
138 [M1_PMU_PERFCTR_INST_BRANCH_RET] = ONLY_5_6_7,
139 [M1_PMU_PERFCTR_INST_BRANCH_TAKEN] = ONLY_5_6_7,
140 [M1_PMU_PERFCTR_INST_BRANCH_INDIR] = ONLY_5_6_7,
141 [M1_PMU_PERFCTR_INST_BRANCH_COND] = ONLY_5_6_7,
142 [M1_PMU_PERFCTR_INST_INT_LD] = ONLY_5_6_7,
143 [M1_PMU_PERFCTR_INST_INT_ST] = BIT(7),
144 [M1_PMU_PERFCTR_INST_INT_ALU] = BIT(7),
145 [M1_PMU_PERFCTR_INST_SIMD_LD] = ONLY_5_6_7,
146 [M1_PMU_PERFCTR_INST_SIMD_ST] = ONLY_5_6_7,
147 [M1_PMU_PERFCTR_INST_SIMD_ALU] = BIT(7),
148 [M1_PMU_PERFCTR_INST_LDST] = BIT(7),
149 [M1_PMU_PERFCTR_INST_BARRIER] = ONLY_5_6_7,
150 [M1_PMU_PERFCTR_UNKNOWN_9f] = BIT(7),
151 [M1_PMU_PERFCTR_L1D_CACHE_MISS_LD_NONSPEC] = ONLY_5_6_7,
152 [M1_PMU_PERFCTR_L1D_CACHE_MISS_ST_NONSPEC] = ONLY_5_6_7,
153 [M1_PMU_PERFCTR_L1D_TLB_MISS_NONSPEC] = ONLY_5_6_7,
154 [M1_PMU_PERFCTR_ST_MEMORY_ORDER_VIOLATION_NONSPEC] = ONLY_5_6_7,
155 [M1_PMU_PERFCTR_BRANCH_COND_MISPRED_NONSPEC] = ONLY_5_6_7,
156 [M1_PMU_PERFCTR_BRANCH_INDIR_MISPRED_NONSPEC] = ONLY_5_6_7,
157 [M1_PMU_PERFCTR_BRANCH_RET_INDIR_MISPRED_NONSPEC] = ONLY_5_6_7,
158 [M1_PMU_PERFCTR_BRANCH_CALL_INDIR_MISPRED_NONSPEC] = ONLY_5_6_7,
159 [M1_PMU_PERFCTR_BRANCH_MISPRED_NONSPEC] = ONLY_5_6_7,
160 [M1_PMU_PERFCTR_UNKNOWN_f5] = ONLY_2_4_6,
161 [M1_PMU_PERFCTR_UNKNOWN_f6] = ONLY_2_4_6,
162 [M1_PMU_PERFCTR_UNKNOWN_f7] = ONLY_2_4_6,
163 [M1_PMU_PERFCTR_UNKNOWN_f8] = ONLY_2_TO_7,
164 [M1_PMU_PERFCTR_UNKNOWN_fd] = ONLY_2_4_6,
165};
166
167static const unsigned m1_pmu_perf_map[PERF_COUNT_HW_MAX] = {
168 PERF_MAP_ALL_UNSUPPORTED,
169 [PERF_COUNT_HW_CPU_CYCLES] = M1_PMU_PERFCTR_CORE_ACTIVE_CYCLE,
170 [PERF_COUNT_HW_INSTRUCTIONS] = M1_PMU_PERFCTR_INST_ALL,
171};
172
173/* sysfs definitions */
174static ssize_t m1_pmu_events_sysfs_show(struct device *dev,
175 struct device_attribute *attr,
176 char *page)
177{
178 struct perf_pmu_events_attr *pmu_attr;
179
180 pmu_attr = container_of(attr, struct perf_pmu_events_attr, attr);
181
182 return sprintf(page, "event=0x%04llx\n", pmu_attr->id);
183}
184
185#define M1_PMU_EVENT_ATTR(name, config) \
186 PMU_EVENT_ATTR_ID(name, m1_pmu_events_sysfs_show, config)
187
188static struct attribute *m1_pmu_event_attrs[] = {
189 M1_PMU_EVENT_ATTR(cycles, M1_PMU_PERFCTR_CORE_ACTIVE_CYCLE),
190 M1_PMU_EVENT_ATTR(instructions, M1_PMU_PERFCTR_INST_ALL),
191 NULL,
192};
193
194static const struct attribute_group m1_pmu_events_attr_group = {
195 .name = "events",
196 .attrs = m1_pmu_event_attrs,
197};
198
199PMU_FORMAT_ATTR(event, "config:0-7");
200
201static struct attribute *m1_pmu_format_attrs[] = {
202 &format_attr_event.attr,
203 NULL,
204};
205
206static const struct attribute_group m1_pmu_format_attr_group = {
207 .name = "format",
208 .attrs = m1_pmu_format_attrs,
209};
210
211/* Low level accessors. No synchronisation. */
212#define PMU_READ_COUNTER(_idx) \
213 case _idx: return read_sysreg_s(SYS_IMP_APL_PMC## _idx ##_EL1)
214
215#define PMU_WRITE_COUNTER(_val, _idx) \
216 case _idx: \
217 write_sysreg_s(_val, SYS_IMP_APL_PMC## _idx ##_EL1); \
218 return
219
220static u64 m1_pmu_read_hw_counter(unsigned int index)
221{
222 switch (index) {
223 PMU_READ_COUNTER(0);
224 PMU_READ_COUNTER(1);
225 PMU_READ_COUNTER(2);
226 PMU_READ_COUNTER(3);
227 PMU_READ_COUNTER(4);
228 PMU_READ_COUNTER(5);
229 PMU_READ_COUNTER(6);
230 PMU_READ_COUNTER(7);
231 PMU_READ_COUNTER(8);
232 PMU_READ_COUNTER(9);
233 }
234
235 BUG();
236}
237
238static void m1_pmu_write_hw_counter(u64 val, unsigned int index)
239{
240 switch (index) {
241 PMU_WRITE_COUNTER(val, 0);
242 PMU_WRITE_COUNTER(val, 1);
243 PMU_WRITE_COUNTER(val, 2);
244 PMU_WRITE_COUNTER(val, 3);
245 PMU_WRITE_COUNTER(val, 4);
246 PMU_WRITE_COUNTER(val, 5);
247 PMU_WRITE_COUNTER(val, 6);
248 PMU_WRITE_COUNTER(val, 7);
249 PMU_WRITE_COUNTER(val, 8);
250 PMU_WRITE_COUNTER(val, 9);
251 }
252
253 BUG();
254}
255
256#define get_bit_offset(index, mask) (__ffs(mask) + (index))
257
258static void __m1_pmu_enable_counter(unsigned int index, bool en)
259{
260 u64 val, bit;
261
262 switch (index) {
263 case 0 ... 7:
264 bit = BIT(get_bit_offset(index, PMCR0_CNT_ENABLE_0_7));
265 break;
266 case 8 ... 9:
267 bit = BIT(get_bit_offset(index - 8, PMCR0_CNT_ENABLE_8_9));
268 break;
269 default:
270 BUG();
271 }
272
273 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
274
275 if (en)
276 val |= bit;
277 else
278 val &= ~bit;
279
280 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
281}
282
283static void m1_pmu_enable_counter(unsigned int index)
284{
285 __m1_pmu_enable_counter(index, true);
286}
287
288static void m1_pmu_disable_counter(unsigned int index)
289{
290 __m1_pmu_enable_counter(index, false);
291}
292
293static void __m1_pmu_enable_counter_interrupt(unsigned int index, bool en)
294{
295 u64 val, bit;
296
297 switch (index) {
298 case 0 ... 7:
299 bit = BIT(get_bit_offset(index, PMCR0_PMI_ENABLE_0_7));
300 break;
301 case 8 ... 9:
302 bit = BIT(get_bit_offset(index - 8, PMCR0_PMI_ENABLE_8_9));
303 break;
304 default:
305 BUG();
306 }
307
308 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
309
310 if (en)
311 val |= bit;
312 else
313 val &= ~bit;
314
315 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
316}
317
318static void m1_pmu_enable_counter_interrupt(unsigned int index)
319{
320 __m1_pmu_enable_counter_interrupt(index, true);
321}
322
323static void m1_pmu_disable_counter_interrupt(unsigned int index)
324{
325 __m1_pmu_enable_counter_interrupt(index, false);
326}
327
328static void m1_pmu_configure_counter(unsigned int index, u8 event,
329 bool user, bool kernel)
330{
331 u64 val, user_bit, kernel_bit;
332 int shift;
333
334 switch (index) {
335 case 0 ... 7:
336 user_bit = BIT(get_bit_offset(index, PMCR1_COUNT_A64_EL0_0_7));
337 kernel_bit = BIT(get_bit_offset(index, PMCR1_COUNT_A64_EL1_0_7));
338 break;
339 case 8 ... 9:
340 user_bit = BIT(get_bit_offset(index - 8, PMCR1_COUNT_A64_EL0_8_9));
341 kernel_bit = BIT(get_bit_offset(index - 8, PMCR1_COUNT_A64_EL1_8_9));
342 break;
343 default:
344 BUG();
345 }
346
347 val = read_sysreg_s(SYS_IMP_APL_PMCR1_EL1);
348
349 if (user)
350 val |= user_bit;
351 else
352 val &= ~user_bit;
353
354 if (kernel)
355 val |= kernel_bit;
356 else
357 val &= ~kernel_bit;
358
359 write_sysreg_s(val, SYS_IMP_APL_PMCR1_EL1);
360
361 /*
362 * Counters 0 and 1 have fixed events. For anything else,
363 * place the event at the expected location in the relevant
364 * register (PMESR0 holds the event configuration for counters
365 * 2-5, resp. PMESR1 for counters 6-9).
366 */
367 switch (index) {
368 case 0 ... 1:
369 break;
370 case 2 ... 5:
371 shift = (index - 2) * 8;
372 val = read_sysreg_s(SYS_IMP_APL_PMESR0_EL1);
373 val &= ~((u64)0xff << shift);
374 val |= (u64)event << shift;
375 write_sysreg_s(val, SYS_IMP_APL_PMESR0_EL1);
376 break;
377 case 6 ... 9:
378 shift = (index - 6) * 8;
379 val = read_sysreg_s(SYS_IMP_APL_PMESR1_EL1);
380 val &= ~((u64)0xff << shift);
381 val |= (u64)event << shift;
382 write_sysreg_s(val, SYS_IMP_APL_PMESR1_EL1);
383 break;
384 }
385}
386
387/* arm_pmu backend */
388static void m1_pmu_enable_event(struct perf_event *event)
389{
390 bool user, kernel;
391 u8 evt;
392
393 evt = event->hw.config_base & M1_PMU_CFG_EVENT;
394 user = event->hw.config_base & M1_PMU_CFG_COUNT_USER;
395 kernel = event->hw.config_base & M1_PMU_CFG_COUNT_KERNEL;
396
397 m1_pmu_disable_counter_interrupt(event->hw.idx);
398 m1_pmu_disable_counter(event->hw.idx);
399 isb();
400
401 m1_pmu_configure_counter(event->hw.idx, evt, user, kernel);
402 m1_pmu_enable_counter(event->hw.idx);
403 m1_pmu_enable_counter_interrupt(event->hw.idx);
404 isb();
405}
406
407static void m1_pmu_disable_event(struct perf_event *event)
408{
409 m1_pmu_disable_counter_interrupt(event->hw.idx);
410 m1_pmu_disable_counter(event->hw.idx);
411 isb();
412}
413
414static irqreturn_t m1_pmu_handle_irq(struct arm_pmu *cpu_pmu)
415{
416 struct pmu_hw_events *cpuc = this_cpu_ptr(cpu_pmu->hw_events);
417 struct pt_regs *regs;
418 u64 overflow, state;
419 int idx;
420
421 overflow = read_sysreg_s(SYS_IMP_APL_PMSR_EL1);
422 if (!overflow) {
423 /* Spurious interrupt? */
424 state = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
425 state &= ~PMCR0_IACT;
426 write_sysreg_s(state, SYS_IMP_APL_PMCR0_EL1);
427 isb();
428 return IRQ_NONE;
429 }
430
431 cpu_pmu->stop(cpu_pmu);
432
433 regs = get_irq_regs();
434
435 for_each_set_bit(idx, cpu_pmu->cntr_mask, M1_PMU_NR_COUNTERS) {
436 struct perf_event *event = cpuc->events[idx];
437 struct perf_sample_data data;
438
439 if (!event)
440 continue;
441
442 armpmu_event_update(event);
443 perf_sample_data_init(&data, 0, event->hw.last_period);
444 if (!armpmu_event_set_period(event))
445 continue;
446
447 if (perf_event_overflow(event, &data, regs))
448 m1_pmu_disable_event(event);
449 }
450
451 cpu_pmu->start(cpu_pmu);
452
453 return IRQ_HANDLED;
454}
455
456static u64 m1_pmu_read_counter(struct perf_event *event)
457{
458 return m1_pmu_read_hw_counter(event->hw.idx);
459}
460
461static void m1_pmu_write_counter(struct perf_event *event, u64 value)
462{
463 m1_pmu_write_hw_counter(value, event->hw.idx);
464 isb();
465}
466
467static int m1_pmu_get_event_idx(struct pmu_hw_events *cpuc,
468 struct perf_event *event)
469{
470 unsigned long evtype = event->hw.config_base & M1_PMU_CFG_EVENT;
471 unsigned long affinity = m1_pmu_event_affinity[evtype];
472 int idx;
473
474 /*
475 * Place the event on the first free counter that can count
476 * this event.
477 *
478 * We could do a better job if we had a view of all the events
479 * counting on the PMU at any given time, and by placing the
480 * most constraining events first.
481 */
482 for_each_set_bit(idx, &affinity, M1_PMU_NR_COUNTERS) {
483 if (!test_and_set_bit(idx, cpuc->used_mask))
484 return idx;
485 }
486
487 return -EAGAIN;
488}
489
490static void m1_pmu_clear_event_idx(struct pmu_hw_events *cpuc,
491 struct perf_event *event)
492{
493 clear_bit(event->hw.idx, cpuc->used_mask);
494}
495
496static void __m1_pmu_set_mode(u8 mode)
497{
498 u64 val;
499
500 val = read_sysreg_s(SYS_IMP_APL_PMCR0_EL1);
501 val &= ~(PMCR0_IMODE | PMCR0_IACT);
502 val |= FIELD_PREP(PMCR0_IMODE, mode);
503 write_sysreg_s(val, SYS_IMP_APL_PMCR0_EL1);
504 isb();
505}
506
507static void m1_pmu_start(struct arm_pmu *cpu_pmu)
508{
509 __m1_pmu_set_mode(PMCR0_IMODE_FIQ);
510}
511
512static void m1_pmu_stop(struct arm_pmu *cpu_pmu)
513{
514 __m1_pmu_set_mode(PMCR0_IMODE_OFF);
515}
516
517static int m1_pmu_map_event(struct perf_event *event)
518{
519 /*
520 * Although the counters are 48bit wide, bit 47 is what
521 * triggers the overflow interrupt. Advertise the counters
522 * being 47bit wide to mimick the behaviour of the ARM PMU.
523 */
524 event->hw.flags |= ARMPMU_EVT_47BIT;
525 return armpmu_map_event(event, &m1_pmu_perf_map, NULL, M1_PMU_CFG_EVENT);
526}
527
528static int m2_pmu_map_event(struct perf_event *event)
529{
530 /*
531 * Same deal as the above, except that M2 has 64bit counters.
532 * Which, as far as we're concerned, actually means 63 bits.
533 * Yes, this is getting awkward.
534 */
535 event->hw.flags |= ARMPMU_EVT_63BIT;
536 return armpmu_map_event(event, &m1_pmu_perf_map, NULL, M1_PMU_CFG_EVENT);
537}
538
539static void m1_pmu_reset(void *info)
540{
541 int i;
542
543 __m1_pmu_set_mode(PMCR0_IMODE_OFF);
544
545 for (i = 0; i < M1_PMU_NR_COUNTERS; i++) {
546 m1_pmu_disable_counter(i);
547 m1_pmu_disable_counter_interrupt(i);
548 m1_pmu_write_hw_counter(0, i);
549 }
550
551 isb();
552}
553
554static int m1_pmu_set_event_filter(struct hw_perf_event *event,
555 struct perf_event_attr *attr)
556{
557 unsigned long config_base = 0;
558
559 if (!attr->exclude_guest) {
560 pr_debug("ARM performance counters do not support mode exclusion\n");
561 return -EOPNOTSUPP;
562 }
563 if (!attr->exclude_kernel)
564 config_base |= M1_PMU_CFG_COUNT_KERNEL;
565 if (!attr->exclude_user)
566 config_base |= M1_PMU_CFG_COUNT_USER;
567
568 event->config_base = config_base;
569
570 return 0;
571}
572
573static int m1_pmu_init(struct arm_pmu *cpu_pmu, u32 flags)
574{
575 cpu_pmu->handle_irq = m1_pmu_handle_irq;
576 cpu_pmu->enable = m1_pmu_enable_event;
577 cpu_pmu->disable = m1_pmu_disable_event;
578 cpu_pmu->read_counter = m1_pmu_read_counter;
579 cpu_pmu->write_counter = m1_pmu_write_counter;
580 cpu_pmu->get_event_idx = m1_pmu_get_event_idx;
581 cpu_pmu->clear_event_idx = m1_pmu_clear_event_idx;
582 cpu_pmu->start = m1_pmu_start;
583 cpu_pmu->stop = m1_pmu_stop;
584
585 if (flags & ARMPMU_EVT_47BIT)
586 cpu_pmu->map_event = m1_pmu_map_event;
587 else if (flags & ARMPMU_EVT_63BIT)
588 cpu_pmu->map_event = m2_pmu_map_event;
589 else
590 return WARN_ON(-EINVAL);
591
592 cpu_pmu->reset = m1_pmu_reset;
593 cpu_pmu->set_event_filter = m1_pmu_set_event_filter;
594
595 bitmap_set(cpu_pmu->cntr_mask, 0, M1_PMU_NR_COUNTERS);
596 cpu_pmu->attr_groups[ARMPMU_ATTR_GROUP_EVENTS] = &m1_pmu_events_attr_group;
597 cpu_pmu->attr_groups[ARMPMU_ATTR_GROUP_FORMATS] = &m1_pmu_format_attr_group;
598 return 0;
599}
600
601/* Device driver gunk */
602static int m1_pmu_ice_init(struct arm_pmu *cpu_pmu)
603{
604 cpu_pmu->name = "apple_icestorm_pmu";
605 return m1_pmu_init(cpu_pmu, ARMPMU_EVT_47BIT);
606}
607
608static int m1_pmu_fire_init(struct arm_pmu *cpu_pmu)
609{
610 cpu_pmu->name = "apple_firestorm_pmu";
611 return m1_pmu_init(cpu_pmu, ARMPMU_EVT_47BIT);
612}
613
614static int m2_pmu_avalanche_init(struct arm_pmu *cpu_pmu)
615{
616 cpu_pmu->name = "apple_avalanche_pmu";
617 return m1_pmu_init(cpu_pmu, ARMPMU_EVT_63BIT);
618}
619
620static int m2_pmu_blizzard_init(struct arm_pmu *cpu_pmu)
621{
622 cpu_pmu->name = "apple_blizzard_pmu";
623 return m1_pmu_init(cpu_pmu, ARMPMU_EVT_63BIT);
624}
625
626static const struct of_device_id m1_pmu_of_device_ids[] = {
627 { .compatible = "apple,avalanche-pmu", .data = m2_pmu_avalanche_init, },
628 { .compatible = "apple,blizzard-pmu", .data = m2_pmu_blizzard_init, },
629 { .compatible = "apple,icestorm-pmu", .data = m1_pmu_ice_init, },
630 { .compatible = "apple,firestorm-pmu", .data = m1_pmu_fire_init, },
631 { },
632};
633MODULE_DEVICE_TABLE(of, m1_pmu_of_device_ids);
634
635static int m1_pmu_device_probe(struct platform_device *pdev)
636{
637 return arm_pmu_device_probe(pdev, m1_pmu_of_device_ids, NULL);
638}
639
640static struct platform_driver m1_pmu_driver = {
641 .driver = {
642 .name = "apple-m1-cpu-pmu",
643 .of_match_table = m1_pmu_of_device_ids,
644 .suppress_bind_attrs = true,
645 },
646 .probe = m1_pmu_device_probe,
647};
648
649module_platform_driver(m1_pmu_driver);