Linux Audio

Check our new training course

Loading...
Note: File does not exist in v5.4.
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/*
  3 * Performance counter support for POWER10 processors.
  4 *
  5 * Copyright 2020 Madhavan Srinivasan, IBM Corporation.
  6 * Copyright 2020 Athira Rajeev, IBM Corporation.
  7 */
  8
  9#define pr_fmt(fmt)	"power10-pmu: " fmt
 10
 11#include "isa207-common.h"
 12
 13/*
 14 * Raw event encoding for Power10:
 15 *
 16 *        60        56        52        48        44        40        36        32
 17 * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
 18 *   | | [ ]   [ src_match ] [  src_mask ]   | [ ] [ l2l3_sel ]  [  thresh_ctl   ]
 19 *   | |  |                                  |  |                         |
 20 *   | |  *- IFM (Linux)                     |  |        thresh start/stop -*
 21 *   | *- BHRB (Linux)                       |  src_sel
 22 *   *- EBB (Linux)                          *invert_bit
 23 *
 24 *        28        24        20        16        12         8         4         0
 25 * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
 26 *   [   ] [  sample ]   [ ] [ ]   [ pmc ]   [unit ]   [ ] |  m   [    pmcxsel    ]
 27 *     |        |        |    |                        |   |  |
 28 *     |        |        |    |                        |   |  *- mark
 29 *     |        |        |    *- L1/L2/L3 cache_sel    |   |*-radix_scope_qual
 30 *     |        |        sdar_mode                     |
 31 *     |        *- sampling mode for marked events     *- combine
 32 *     |
 33 *     *- thresh_sel
 34 *
 35 * Below uses IBM bit numbering.
 36 *
 37 * MMCR1[x:y] = unit    (PMCxUNIT)
 38 * MMCR1[24]   = pmc1combine[0]
 39 * MMCR1[25]   = pmc1combine[1]
 40 * MMCR1[26]   = pmc2combine[0]
 41 * MMCR1[27]   = pmc2combine[1]
 42 * MMCR1[28]   = pmc3combine[0]
 43 * MMCR1[29]   = pmc3combine[1]
 44 * MMCR1[30]   = pmc4combine[0]
 45 * MMCR1[31]   = pmc4combine[1]
 46 *
 47 * if pmc == 3 and unit == 0 and pmcxsel[0:6] == 0b0101011
 48 *	MMCR1[20:27] = thresh_ctl
 49 * else if pmc == 4 and unit == 0xf and pmcxsel[0:6] == 0b0101001
 50 *	MMCR1[20:27] = thresh_ctl
 51 * else
 52 *	MMCRA[48:55] = thresh_ctl   (THRESH START/END)
 53 *
 54 * if thresh_sel:
 55 *	MMCRA[45:47] = thresh_sel
 56 *
 57 * if l2l3_sel:
 58 * MMCR2[56:60] = l2l3_sel[0:4]
 59 *
 60 * MMCR1[16] = cache_sel[0]
 61 * MMCR1[17] = cache_sel[1]
 62 * MMCR1[18] = radix_scope_qual
 63 *
 64 * if mark:
 65 *	MMCRA[63]    = 1		(SAMPLE_ENABLE)
 66 *	MMCRA[57:59] = sample[0:2]	(RAND_SAMP_ELIG)
 67 *	MMCRA[61:62] = sample[3:4]	(RAND_SAMP_MODE)
 68 *
 69 * if EBB and BHRB:
 70 *	MMCRA[32:33] = IFM
 71 *
 72 * MMCRA[SDAR_MODE]  = sdar_mode[0:1]
 73 */
 74
 75/*
 76 * Some power10 event codes.
 77 */
 78#define EVENT(_name, _code)     enum{_name = _code}
 79
 80#include "power10-events-list.h"
 81
 82#undef EVENT
 83
 84/* MMCRA IFM bits - POWER10 */
 85#define POWER10_MMCRA_IFM1		0x0000000040000000UL
 86#define POWER10_MMCRA_IFM2		0x0000000080000000UL
 87#define POWER10_MMCRA_IFM3		0x00000000C0000000UL
 88#define POWER10_MMCRA_BHRB_MASK		0x00000000C0000000UL
 89
 90extern u64 PERF_REG_EXTENDED_MASK;
 91
 92/* Table of alternatives, sorted by column 0 */
 93static const unsigned int power10_event_alternatives[][MAX_ALT] = {
 94	{ PM_INST_CMPL_ALT,		PM_INST_CMPL },
 95	{ PM_CYC_ALT,			PM_CYC },
 96};
 97
 98static int power10_get_alternatives(u64 event, unsigned int flags, u64 alt[])
 99{
100	int num_alt = 0;
101
102	num_alt = isa207_get_alternatives(event, alt,
103					  ARRAY_SIZE(power10_event_alternatives), flags,
104					  power10_event_alternatives);
105
106	return num_alt;
107}
108
109static int power10_check_attr_config(struct perf_event *ev)
110{
111	u64 val;
112	u64 event = ev->attr.config;
113
114	val = (event >> EVENT_SAMPLE_SHIFT) & EVENT_SAMPLE_MASK;
115	if (val == 0x10 || isa3XX_check_attr_config(ev))
116		return -EINVAL;
117
118	return 0;
119}
120
121GENERIC_EVENT_ATTR(cpu-cycles,			PM_CYC);
122GENERIC_EVENT_ATTR(instructions,		PM_INST_CMPL);
123GENERIC_EVENT_ATTR(branch-instructions,		PM_BR_CMPL);
124GENERIC_EVENT_ATTR(branch-misses,		PM_BR_MPRED_CMPL);
125GENERIC_EVENT_ATTR(cache-references,		PM_LD_REF_L1);
126GENERIC_EVENT_ATTR(cache-misses,		PM_LD_MISS_L1);
127GENERIC_EVENT_ATTR(mem-loads,			MEM_LOADS);
128GENERIC_EVENT_ATTR(mem-stores,			MEM_STORES);
129GENERIC_EVENT_ATTR(branch-instructions,		PM_BR_FIN);
130GENERIC_EVENT_ATTR(branch-misses,		PM_MPRED_BR_FIN);
131GENERIC_EVENT_ATTR(cache-misses,		PM_LD_DEMAND_MISS_L1_FIN);
132
133CACHE_EVENT_ATTR(L1-dcache-load-misses,		PM_LD_MISS_L1);
134CACHE_EVENT_ATTR(L1-dcache-loads,		PM_LD_REF_L1);
135CACHE_EVENT_ATTR(L1-dcache-prefetches,		PM_LD_PREFETCH_CACHE_LINE_MISS);
136CACHE_EVENT_ATTR(L1-dcache-store-misses,	PM_ST_MISS_L1);
137CACHE_EVENT_ATTR(L1-icache-load-misses,		PM_L1_ICACHE_MISS);
138CACHE_EVENT_ATTR(L1-icache-loads,		PM_INST_FROM_L1);
139CACHE_EVENT_ATTR(L1-icache-prefetches,		PM_IC_PREF_REQ);
140CACHE_EVENT_ATTR(LLC-load-misses,		PM_DATA_FROM_L3MISS);
141CACHE_EVENT_ATTR(LLC-loads,			PM_DATA_FROM_L3);
142CACHE_EVENT_ATTR(LLC-prefetches,		PM_L3_PF_MISS_L3);
143CACHE_EVENT_ATTR(LLC-store-misses,		PM_L2_ST_MISS);
144CACHE_EVENT_ATTR(LLC-stores,			PM_L2_ST);
145CACHE_EVENT_ATTR(branch-load-misses,		PM_BR_MPRED_CMPL);
146CACHE_EVENT_ATTR(branch-loads,			PM_BR_CMPL);
147CACHE_EVENT_ATTR(dTLB-load-misses,		PM_DTLB_MISS);
148CACHE_EVENT_ATTR(iTLB-load-misses,		PM_ITLB_MISS);
149
150static struct attribute *power10_events_attr_dd1[] = {
151	GENERIC_EVENT_PTR(PM_CYC),
152	GENERIC_EVENT_PTR(PM_INST_CMPL),
153	GENERIC_EVENT_PTR(PM_BR_CMPL),
154	GENERIC_EVENT_PTR(PM_BR_MPRED_CMPL),
155	GENERIC_EVENT_PTR(PM_LD_REF_L1),
156	GENERIC_EVENT_PTR(PM_LD_MISS_L1),
157	GENERIC_EVENT_PTR(MEM_LOADS),
158	GENERIC_EVENT_PTR(MEM_STORES),
159	CACHE_EVENT_PTR(PM_LD_MISS_L1),
160	CACHE_EVENT_PTR(PM_LD_REF_L1),
161	CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS),
162	CACHE_EVENT_PTR(PM_ST_MISS_L1),
163	CACHE_EVENT_PTR(PM_L1_ICACHE_MISS),
164	CACHE_EVENT_PTR(PM_INST_FROM_L1),
165	CACHE_EVENT_PTR(PM_IC_PREF_REQ),
166	CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS),
167	CACHE_EVENT_PTR(PM_DATA_FROM_L3),
168	CACHE_EVENT_PTR(PM_BR_MPRED_CMPL),
169	CACHE_EVENT_PTR(PM_BR_CMPL),
170	CACHE_EVENT_PTR(PM_DTLB_MISS),
171	CACHE_EVENT_PTR(PM_ITLB_MISS),
172	NULL
173};
174
175static struct attribute *power10_events_attr[] = {
176	GENERIC_EVENT_PTR(PM_CYC),
177	GENERIC_EVENT_PTR(PM_INST_CMPL),
178	GENERIC_EVENT_PTR(PM_BR_FIN),
179	GENERIC_EVENT_PTR(PM_MPRED_BR_FIN),
180	GENERIC_EVENT_PTR(PM_LD_REF_L1),
181	GENERIC_EVENT_PTR(PM_LD_DEMAND_MISS_L1_FIN),
182	GENERIC_EVENT_PTR(MEM_LOADS),
183	GENERIC_EVENT_PTR(MEM_STORES),
184	CACHE_EVENT_PTR(PM_LD_MISS_L1),
185	CACHE_EVENT_PTR(PM_LD_REF_L1),
186	CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS),
187	CACHE_EVENT_PTR(PM_ST_MISS_L1),
188	CACHE_EVENT_PTR(PM_L1_ICACHE_MISS),
189	CACHE_EVENT_PTR(PM_INST_FROM_L1),
190	CACHE_EVENT_PTR(PM_IC_PREF_REQ),
191	CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS),
192	CACHE_EVENT_PTR(PM_DATA_FROM_L3),
193	CACHE_EVENT_PTR(PM_L3_PF_MISS_L3),
194	CACHE_EVENT_PTR(PM_L2_ST_MISS),
195	CACHE_EVENT_PTR(PM_L2_ST),
196	CACHE_EVENT_PTR(PM_BR_MPRED_CMPL),
197	CACHE_EVENT_PTR(PM_BR_CMPL),
198	CACHE_EVENT_PTR(PM_DTLB_MISS),
199	CACHE_EVENT_PTR(PM_ITLB_MISS),
200	NULL
201};
202
203static const struct attribute_group power10_pmu_events_group_dd1 = {
204	.name = "events",
205	.attrs = power10_events_attr_dd1,
206};
207
208static const struct attribute_group power10_pmu_events_group = {
209	.name = "events",
210	.attrs = power10_events_attr,
211};
212
213PMU_FORMAT_ATTR(event,          "config:0-59");
214PMU_FORMAT_ATTR(pmcxsel,        "config:0-7");
215PMU_FORMAT_ATTR(mark,           "config:8");
216PMU_FORMAT_ATTR(combine,        "config:10-11");
217PMU_FORMAT_ATTR(unit,           "config:12-15");
218PMU_FORMAT_ATTR(pmc,            "config:16-19");
219PMU_FORMAT_ATTR(cache_sel,      "config:20-21");
220PMU_FORMAT_ATTR(sdar_mode,      "config:22-23");
221PMU_FORMAT_ATTR(sample_mode,    "config:24-28");
222PMU_FORMAT_ATTR(thresh_sel,     "config:29-31");
223PMU_FORMAT_ATTR(thresh_stop,    "config:32-35");
224PMU_FORMAT_ATTR(thresh_start,   "config:36-39");
225PMU_FORMAT_ATTR(l2l3_sel,       "config:40-44");
226PMU_FORMAT_ATTR(src_sel,        "config:45-46");
227PMU_FORMAT_ATTR(invert_bit,     "config:47");
228PMU_FORMAT_ATTR(src_mask,       "config:48-53");
229PMU_FORMAT_ATTR(src_match,      "config:54-59");
230PMU_FORMAT_ATTR(radix_scope,	"config:9");
231PMU_FORMAT_ATTR(thresh_cmp,     "config1:0-17");
232
233static struct attribute *power10_pmu_format_attr[] = {
234	&format_attr_event.attr,
235	&format_attr_pmcxsel.attr,
236	&format_attr_mark.attr,
237	&format_attr_combine.attr,
238	&format_attr_unit.attr,
239	&format_attr_pmc.attr,
240	&format_attr_cache_sel.attr,
241	&format_attr_sdar_mode.attr,
242	&format_attr_sample_mode.attr,
243	&format_attr_thresh_sel.attr,
244	&format_attr_thresh_stop.attr,
245	&format_attr_thresh_start.attr,
246	&format_attr_l2l3_sel.attr,
247	&format_attr_src_sel.attr,
248	&format_attr_invert_bit.attr,
249	&format_attr_src_mask.attr,
250	&format_attr_src_match.attr,
251	&format_attr_radix_scope.attr,
252	&format_attr_thresh_cmp.attr,
253	NULL,
254};
255
256static const struct attribute_group power10_pmu_format_group = {
257	.name = "format",
258	.attrs = power10_pmu_format_attr,
259};
260
261static struct attribute *power10_pmu_caps_attrs[] = {
262	NULL
263};
264
265static struct attribute_group power10_pmu_caps_group = {
266	.name  = "caps",
267	.attrs = power10_pmu_caps_attrs,
268};
269
270static const struct attribute_group *power10_pmu_attr_groups_dd1[] = {
271	&power10_pmu_format_group,
272	&power10_pmu_events_group_dd1,
273	&power10_pmu_caps_group,
274	NULL,
275};
276
277static const struct attribute_group *power10_pmu_attr_groups[] = {
278	&power10_pmu_format_group,
279	&power10_pmu_events_group,
280	&power10_pmu_caps_group,
281	NULL,
282};
283
284static int power10_generic_events_dd1[] = {
285	[PERF_COUNT_HW_CPU_CYCLES] =			PM_CYC,
286	[PERF_COUNT_HW_INSTRUCTIONS] =			PM_INST_CMPL,
287	[PERF_COUNT_HW_BRANCH_INSTRUCTIONS] =		PM_BR_CMPL,
288	[PERF_COUNT_HW_BRANCH_MISSES] =			PM_BR_MPRED_CMPL,
289	[PERF_COUNT_HW_CACHE_REFERENCES] =		PM_LD_REF_L1,
290	[PERF_COUNT_HW_CACHE_MISSES] =			PM_LD_MISS_L1,
291};
292
293static int power10_generic_events[] = {
294	[PERF_COUNT_HW_CPU_CYCLES] =			PM_CYC,
295	[PERF_COUNT_HW_INSTRUCTIONS] =			PM_INST_CMPL,
296	[PERF_COUNT_HW_BRANCH_INSTRUCTIONS] =		PM_BR_FIN,
297	[PERF_COUNT_HW_BRANCH_MISSES] =			PM_MPRED_BR_FIN,
298	[PERF_COUNT_HW_CACHE_REFERENCES] =		PM_LD_REF_L1,
299	[PERF_COUNT_HW_CACHE_MISSES] =			PM_LD_DEMAND_MISS_L1_FIN,
300};
301
302static u64 power10_bhrb_filter_map(u64 branch_sample_type)
303{
304	u64 pmu_bhrb_filter = 0;
305
306	/* BHRB and regular PMU events share the same privilege state
307	 * filter configuration. BHRB is always recorded along with a
308	 * regular PMU event. As the privilege state filter is handled
309	 * in the basic PMC configuration of the accompanying regular
310	 * PMU event, we ignore any separate BHRB specific request.
311	 */
312
313	/* No branch filter requested */
314	if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY)
315		return pmu_bhrb_filter;
316
317	/* Invalid branch filter options - HW does not support */
318	if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY_RETURN)
319		return -1;
320
321	if (branch_sample_type & PERF_SAMPLE_BRANCH_IND_CALL) {
322		pmu_bhrb_filter |= POWER10_MMCRA_IFM2;
323		return pmu_bhrb_filter;
324	}
325
326	if (branch_sample_type & PERF_SAMPLE_BRANCH_COND) {
327		pmu_bhrb_filter |= POWER10_MMCRA_IFM3;
328		return pmu_bhrb_filter;
329	}
330
331	if (branch_sample_type & PERF_SAMPLE_BRANCH_CALL)
332		return -1;
333
334	if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY_CALL) {
335		pmu_bhrb_filter |= POWER10_MMCRA_IFM1;
336		return pmu_bhrb_filter;
337	}
338
339	/* Every thing else is unsupported */
340	return -1;
341}
342
343static void power10_config_bhrb(u64 pmu_bhrb_filter)
344{
345	pmu_bhrb_filter &= POWER10_MMCRA_BHRB_MASK;
346
347	/* Enable BHRB filter in PMU */
348	mtspr(SPRN_MMCRA, (mfspr(SPRN_MMCRA) | pmu_bhrb_filter));
349}
350
351#define C(x)	PERF_COUNT_HW_CACHE_##x
352
353/*
354 * Table of generalized cache-related events.
355 * 0 means not supported, -1 means nonsensical, other values
356 * are event codes.
357 */
358static u64 power10_cache_events_dd1[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
359	[C(L1D)] = {
360		[C(OP_READ)] = {
361			[C(RESULT_ACCESS)] = PM_LD_REF_L1,
362			[C(RESULT_MISS)] = PM_LD_MISS_L1,
363		},
364		[C(OP_WRITE)] = {
365			[C(RESULT_ACCESS)] = 0,
366			[C(RESULT_MISS)] = PM_ST_MISS_L1,
367		},
368		[C(OP_PREFETCH)] = {
369			[C(RESULT_ACCESS)] = PM_LD_PREFETCH_CACHE_LINE_MISS,
370			[C(RESULT_MISS)] = 0,
371		},
372	},
373	[C(L1I)] = {
374		[C(OP_READ)] = {
375			[C(RESULT_ACCESS)] = PM_INST_FROM_L1,
376			[C(RESULT_MISS)] = PM_L1_ICACHE_MISS,
377		},
378		[C(OP_WRITE)] = {
379			[C(RESULT_ACCESS)] = PM_INST_FROM_L1MISS,
380			[C(RESULT_MISS)] = -1,
381		},
382		[C(OP_PREFETCH)] = {
383			[C(RESULT_ACCESS)] = PM_IC_PREF_REQ,
384			[C(RESULT_MISS)] = 0,
385		},
386	},
387	[C(LL)] = {
388		[C(OP_READ)] = {
389			[C(RESULT_ACCESS)] = PM_DATA_FROM_L3,
390			[C(RESULT_MISS)] = PM_DATA_FROM_L3MISS,
391		},
392		[C(OP_WRITE)] = {
393			[C(RESULT_ACCESS)] = -1,
394			[C(RESULT_MISS)] = -1,
395		},
396		[C(OP_PREFETCH)] = {
397			[C(RESULT_ACCESS)] = -1,
398			[C(RESULT_MISS)] = 0,
399		},
400	},
401	 [C(DTLB)] = {
402		[C(OP_READ)] = {
403			[C(RESULT_ACCESS)] = 0,
404			[C(RESULT_MISS)] = PM_DTLB_MISS,
405		},
406		[C(OP_WRITE)] = {
407			[C(RESULT_ACCESS)] = -1,
408			[C(RESULT_MISS)] = -1,
409		},
410		[C(OP_PREFETCH)] = {
411			[C(RESULT_ACCESS)] = -1,
412			[C(RESULT_MISS)] = -1,
413		},
414	},
415	[C(ITLB)] = {
416		[C(OP_READ)] = {
417			[C(RESULT_ACCESS)] = 0,
418			[C(RESULT_MISS)] = PM_ITLB_MISS,
419		},
420		[C(OP_WRITE)] = {
421			[C(RESULT_ACCESS)] = -1,
422			[C(RESULT_MISS)] = -1,
423		},
424		[C(OP_PREFETCH)] = {
425			[C(RESULT_ACCESS)] = -1,
426			[C(RESULT_MISS)] = -1,
427		},
428	},
429	[C(BPU)] = {
430		[C(OP_READ)] = {
431			[C(RESULT_ACCESS)] = PM_BR_CMPL,
432			[C(RESULT_MISS)] = PM_BR_MPRED_CMPL,
433		},
434		[C(OP_WRITE)] = {
435			[C(RESULT_ACCESS)] = -1,
436			[C(RESULT_MISS)] = -1,
437		},
438		[C(OP_PREFETCH)] = {
439			[C(RESULT_ACCESS)] = -1,
440			[C(RESULT_MISS)] = -1,
441		},
442	},
443	[C(NODE)] = {
444		[C(OP_READ)] = {
445			[C(RESULT_ACCESS)] = -1,
446			[C(RESULT_MISS)] = -1,
447		},
448		[C(OP_WRITE)] = {
449			[C(RESULT_ACCESS)] = -1,
450			[C(RESULT_MISS)] = -1,
451		},
452		[C(OP_PREFETCH)] = {
453			[C(RESULT_ACCESS)] = -1,
454			[C(RESULT_MISS)] = -1,
455		},
456	},
457};
458
459static u64 power10_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
460	[C(L1D)] = {
461		[C(OP_READ)] = {
462			[C(RESULT_ACCESS)] = PM_LD_REF_L1,
463			[C(RESULT_MISS)] = PM_LD_MISS_L1,
464		},
465		[C(OP_WRITE)] = {
466			[C(RESULT_ACCESS)] = 0,
467			[C(RESULT_MISS)] = PM_ST_MISS_L1,
468		},
469		[C(OP_PREFETCH)] = {
470			[C(RESULT_ACCESS)] = PM_LD_PREFETCH_CACHE_LINE_MISS,
471			[C(RESULT_MISS)] = 0,
472		},
473	},
474	[C(L1I)] = {
475		[C(OP_READ)] = {
476			[C(RESULT_ACCESS)] = PM_INST_FROM_L1,
477			[C(RESULT_MISS)] = PM_L1_ICACHE_MISS,
478		},
479		[C(OP_WRITE)] = {
480			[C(RESULT_ACCESS)] = PM_INST_FROM_L1MISS,
481			[C(RESULT_MISS)] = -1,
482		},
483		[C(OP_PREFETCH)] = {
484			[C(RESULT_ACCESS)] = PM_IC_PREF_REQ,
485			[C(RESULT_MISS)] = 0,
486		},
487	},
488	[C(LL)] = {
489		[C(OP_READ)] = {
490			[C(RESULT_ACCESS)] = PM_DATA_FROM_L3,
491			[C(RESULT_MISS)] = PM_DATA_FROM_L3MISS,
492		},
493		[C(OP_WRITE)] = {
494			[C(RESULT_ACCESS)] = PM_L2_ST,
495			[C(RESULT_MISS)] = PM_L2_ST_MISS,
496		},
497		[C(OP_PREFETCH)] = {
498			[C(RESULT_ACCESS)] = PM_L3_PF_MISS_L3,
499			[C(RESULT_MISS)] = 0,
500		},
501	},
502	 [C(DTLB)] = {
503		[C(OP_READ)] = {
504			[C(RESULT_ACCESS)] = 0,
505			[C(RESULT_MISS)] = PM_DTLB_MISS,
506		},
507		[C(OP_WRITE)] = {
508			[C(RESULT_ACCESS)] = -1,
509			[C(RESULT_MISS)] = -1,
510		},
511		[C(OP_PREFETCH)] = {
512			[C(RESULT_ACCESS)] = -1,
513			[C(RESULT_MISS)] = -1,
514		},
515	},
516	[C(ITLB)] = {
517		[C(OP_READ)] = {
518			[C(RESULT_ACCESS)] = 0,
519			[C(RESULT_MISS)] = PM_ITLB_MISS,
520		},
521		[C(OP_WRITE)] = {
522			[C(RESULT_ACCESS)] = -1,
523			[C(RESULT_MISS)] = -1,
524		},
525		[C(OP_PREFETCH)] = {
526			[C(RESULT_ACCESS)] = -1,
527			[C(RESULT_MISS)] = -1,
528		},
529	},
530	[C(BPU)] = {
531		[C(OP_READ)] = {
532			[C(RESULT_ACCESS)] = PM_BR_CMPL,
533			[C(RESULT_MISS)] = PM_BR_MPRED_CMPL,
534		},
535		[C(OP_WRITE)] = {
536			[C(RESULT_ACCESS)] = -1,
537			[C(RESULT_MISS)] = -1,
538		},
539		[C(OP_PREFETCH)] = {
540			[C(RESULT_ACCESS)] = -1,
541			[C(RESULT_MISS)] = -1,
542		},
543	},
544	[C(NODE)] = {
545		[C(OP_READ)] = {
546			[C(RESULT_ACCESS)] = -1,
547			[C(RESULT_MISS)] = -1,
548		},
549		[C(OP_WRITE)] = {
550			[C(RESULT_ACCESS)] = -1,
551			[C(RESULT_MISS)] = -1,
552		},
553		[C(OP_PREFETCH)] = {
554			[C(RESULT_ACCESS)] = -1,
555			[C(RESULT_MISS)] = -1,
556		},
557	},
558};
559
560#undef C
561
562/*
563 * Set the MMCR0[CC56RUN] bit to enable counting for
564 * PMC5 and PMC6 regardless of the state of CTRL[RUN],
565 * so that we can use counters 5 and 6 as PM_INST_CMPL and
566 * PM_CYC.
567 */
568static int power10_compute_mmcr(u64 event[], int n_ev,
569				unsigned int hwc[], struct mmcr_regs *mmcr,
570				struct perf_event *pevents[], u32 flags)
571{
572	int ret;
573
574	ret = isa207_compute_mmcr(event, n_ev, hwc, mmcr, pevents, flags);
575	if (!ret)
576		mmcr->mmcr0 |= MMCR0_C56RUN;
577	return ret;
578}
579
580static struct power_pmu power10_pmu = {
581	.name			= "POWER10",
582	.n_counter		= MAX_PMU_COUNTERS,
583	.add_fields		= ISA207_ADD_FIELDS,
584	.test_adder		= ISA207_TEST_ADDER,
585	.group_constraint_mask	= CNST_CACHE_PMC4_MASK,
586	.group_constraint_val	= CNST_CACHE_PMC4_VAL,
587	.compute_mmcr		= power10_compute_mmcr,
588	.config_bhrb		= power10_config_bhrb,
589	.bhrb_filter_map	= power10_bhrb_filter_map,
590	.get_constraint		= isa207_get_constraint,
591	.get_alternatives	= power10_get_alternatives,
592	.get_mem_data_src	= isa207_get_mem_data_src,
593	.get_mem_weight		= isa207_get_mem_weight,
594	.disable_pmc		= isa207_disable_pmc,
595	.flags			= PPMU_HAS_SIER | PPMU_ARCH_207S |
596				  PPMU_ARCH_31 | PPMU_HAS_ATTR_CONFIG1,
597	.n_generic		= ARRAY_SIZE(power10_generic_events),
598	.generic_events		= power10_generic_events,
599	.cache_events		= &power10_cache_events,
600	.attr_groups		= power10_pmu_attr_groups,
601	.bhrb_nr		= 32,
602	.capabilities           = PERF_PMU_CAP_EXTENDED_REGS,
603	.check_attr_config	= power10_check_attr_config,
604};
605
606int __init init_power10_pmu(void)
607{
608	unsigned int pvr;
609	int rc;
610
611	pvr = mfspr(SPRN_PVR);
612	if (PVR_VER(pvr) != PVR_POWER10)
613		return -ENODEV;
614
615	/* Add the ppmu flag for power10 DD1 */
616	if ((PVR_CFG(pvr) == 1))
617		power10_pmu.flags |= PPMU_P10_DD1;
618
619	/* Set the PERF_REG_EXTENDED_MASK here */
620	PERF_REG_EXTENDED_MASK = PERF_REG_PMU_MASK_31;
621
622	if ((PVR_CFG(pvr) == 1)) {
623		power10_pmu.generic_events = power10_generic_events_dd1;
624		power10_pmu.attr_groups = power10_pmu_attr_groups_dd1;
625		power10_pmu.cache_events = &power10_cache_events_dd1;
626	}
627
628	rc = register_power_pmu(&power10_pmu);
629	if (rc)
630		return rc;
631
632	/* Tell userspace that EBB is supported */
633	cur_cpu_spec->cpu_user_features2 |= PPC_FEATURE2_EBB;
634
635	return 0;
636}