Loading...
1/*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
4 * for more details.
5 *
6 * Copyright (C) 2004, 05, 06 by Ralf Baechle
7 * Copyright (C) 2005 by MIPS Technologies, Inc.
8 */
9#include <linux/cpumask.h>
10#include <linux/oprofile.h>
11#include <linux/interrupt.h>
12#include <linux/smp.h>
13#include <asm/irq_regs.h>
14#include <asm/time.h>
15
16#include "op_impl.h"
17
18#define M_PERFCTL_EXL (1UL << 0)
19#define M_PERFCTL_KERNEL (1UL << 1)
20#define M_PERFCTL_SUPERVISOR (1UL << 2)
21#define M_PERFCTL_USER (1UL << 3)
22#define M_PERFCTL_INTERRUPT_ENABLE (1UL << 4)
23#define M_PERFCTL_EVENT(event) (((event) & 0x3ff) << 5)
24#define M_PERFCTL_VPEID(vpe) ((vpe) << 16)
25#define M_PERFCTL_MT_EN(filter) ((filter) << 20)
26#define M_TC_EN_ALL M_PERFCTL_MT_EN(0)
27#define M_TC_EN_VPE M_PERFCTL_MT_EN(1)
28#define M_TC_EN_TC M_PERFCTL_MT_EN(2)
29#define M_PERFCTL_TCID(tcid) ((tcid) << 22)
30#define M_PERFCTL_WIDE (1UL << 30)
31#define M_PERFCTL_MORE (1UL << 31)
32
33#define M_COUNTER_OVERFLOW (1UL << 31)
34
35/* Netlogic XLR specific, count events in all threads in a core */
36#define M_PERFCTL_COUNT_ALL_THREADS (1UL << 13)
37
38static int (*save_perf_irq)(void);
39static int perfcount_irq;
40
41/*
42 * XLR has only one set of counters per core. Designate the
43 * first hardware thread in the core for setup and init.
44 * Skip CPUs with non-zero hardware thread id (4 hwt per core)
45 */
46#if defined(CONFIG_CPU_XLR) && defined(CONFIG_SMP)
47#define oprofile_skip_cpu(c) ((cpu_logical_map(c) & 0x3) != 0)
48#else
49#define oprofile_skip_cpu(c) 0
50#endif
51
52#ifdef CONFIG_MIPS_MT_SMP
53static int cpu_has_mipsmt_pertccounters;
54#define WHAT (M_TC_EN_VPE | \
55 M_PERFCTL_VPEID(cpu_data[smp_processor_id()].vpe_id))
56#define vpe_id() (cpu_has_mipsmt_pertccounters ? \
57 0 : cpu_data[smp_processor_id()].vpe_id)
58
59/*
60 * The number of bits to shift to convert between counters per core and
61 * counters per VPE. There is no reasonable interface atm to obtain the
62 * number of VPEs used by Linux and in the 34K this number is fixed to two
63 * anyways so we hardcore a few things here for the moment. The way it's
64 * done here will ensure that oprofile VSMP kernel will run right on a lesser
65 * core like a 24K also or with maxcpus=1.
66 */
67static inline unsigned int vpe_shift(void)
68{
69 if (num_possible_cpus() > 1)
70 return 1;
71
72 return 0;
73}
74
75#else
76
77#define WHAT 0
78#define vpe_id() 0
79
80static inline unsigned int vpe_shift(void)
81{
82 return 0;
83}
84
85#endif
86
87static inline unsigned int counters_total_to_per_cpu(unsigned int counters)
88{
89 return counters >> vpe_shift();
90}
91
92static inline unsigned int counters_per_cpu_to_total(unsigned int counters)
93{
94 return counters << vpe_shift();
95}
96
97#define __define_perf_accessors(r, n, np) \
98 \
99static inline unsigned int r_c0_ ## r ## n(void) \
100{ \
101 unsigned int cpu = vpe_id(); \
102 \
103 switch (cpu) { \
104 case 0: \
105 return read_c0_ ## r ## n(); \
106 case 1: \
107 return read_c0_ ## r ## np(); \
108 default: \
109 BUG(); \
110 } \
111 return 0; \
112} \
113 \
114static inline void w_c0_ ## r ## n(unsigned int value) \
115{ \
116 unsigned int cpu = vpe_id(); \
117 \
118 switch (cpu) { \
119 case 0: \
120 write_c0_ ## r ## n(value); \
121 return; \
122 case 1: \
123 write_c0_ ## r ## np(value); \
124 return; \
125 default: \
126 BUG(); \
127 } \
128 return; \
129} \
130
131__define_perf_accessors(perfcntr, 0, 2)
132__define_perf_accessors(perfcntr, 1, 3)
133__define_perf_accessors(perfcntr, 2, 0)
134__define_perf_accessors(perfcntr, 3, 1)
135
136__define_perf_accessors(perfctrl, 0, 2)
137__define_perf_accessors(perfctrl, 1, 3)
138__define_perf_accessors(perfctrl, 2, 0)
139__define_perf_accessors(perfctrl, 3, 1)
140
141struct op_mips_model op_model_mipsxx_ops;
142
143static struct mipsxx_register_config {
144 unsigned int control[4];
145 unsigned int counter[4];
146} reg;
147
148/* Compute all of the registers in preparation for enabling profiling. */
149
150static void mipsxx_reg_setup(struct op_counter_config *ctr)
151{
152 unsigned int counters = op_model_mipsxx_ops.num_counters;
153 int i;
154
155 /* Compute the performance counter control word. */
156 for (i = 0; i < counters; i++) {
157 reg.control[i] = 0;
158 reg.counter[i] = 0;
159
160 if (!ctr[i].enabled)
161 continue;
162
163 reg.control[i] = M_PERFCTL_EVENT(ctr[i].event) |
164 M_PERFCTL_INTERRUPT_ENABLE;
165 if (ctr[i].kernel)
166 reg.control[i] |= M_PERFCTL_KERNEL;
167 if (ctr[i].user)
168 reg.control[i] |= M_PERFCTL_USER;
169 if (ctr[i].exl)
170 reg.control[i] |= M_PERFCTL_EXL;
171 if (boot_cpu_type() == CPU_XLR)
172 reg.control[i] |= M_PERFCTL_COUNT_ALL_THREADS;
173 reg.counter[i] = 0x80000000 - ctr[i].count;
174 }
175}
176
177/* Program all of the registers in preparation for enabling profiling. */
178
179static void mipsxx_cpu_setup(void *args)
180{
181 unsigned int counters = op_model_mipsxx_ops.num_counters;
182
183 if (oprofile_skip_cpu(smp_processor_id()))
184 return;
185
186 switch (counters) {
187 case 4:
188 w_c0_perfctrl3(0);
189 w_c0_perfcntr3(reg.counter[3]);
190 case 3:
191 w_c0_perfctrl2(0);
192 w_c0_perfcntr2(reg.counter[2]);
193 case 2:
194 w_c0_perfctrl1(0);
195 w_c0_perfcntr1(reg.counter[1]);
196 case 1:
197 w_c0_perfctrl0(0);
198 w_c0_perfcntr0(reg.counter[0]);
199 }
200}
201
202/* Start all counters on current CPU */
203static void mipsxx_cpu_start(void *args)
204{
205 unsigned int counters = op_model_mipsxx_ops.num_counters;
206
207 if (oprofile_skip_cpu(smp_processor_id()))
208 return;
209
210 switch (counters) {
211 case 4:
212 w_c0_perfctrl3(WHAT | reg.control[3]);
213 case 3:
214 w_c0_perfctrl2(WHAT | reg.control[2]);
215 case 2:
216 w_c0_perfctrl1(WHAT | reg.control[1]);
217 case 1:
218 w_c0_perfctrl0(WHAT | reg.control[0]);
219 }
220}
221
222/* Stop all counters on current CPU */
223static void mipsxx_cpu_stop(void *args)
224{
225 unsigned int counters = op_model_mipsxx_ops.num_counters;
226
227 if (oprofile_skip_cpu(smp_processor_id()))
228 return;
229
230 switch (counters) {
231 case 4:
232 w_c0_perfctrl3(0);
233 case 3:
234 w_c0_perfctrl2(0);
235 case 2:
236 w_c0_perfctrl1(0);
237 case 1:
238 w_c0_perfctrl0(0);
239 }
240}
241
242static int mipsxx_perfcount_handler(void)
243{
244 unsigned int counters = op_model_mipsxx_ops.num_counters;
245 unsigned int control;
246 unsigned int counter;
247 int handled = IRQ_NONE;
248
249 if (cpu_has_mips_r2 && !(read_c0_cause() & CAUSEF_PCI))
250 return handled;
251
252 switch (counters) {
253#define HANDLE_COUNTER(n) \
254 case n + 1: \
255 control = r_c0_perfctrl ## n(); \
256 counter = r_c0_perfcntr ## n(); \
257 if ((control & M_PERFCTL_INTERRUPT_ENABLE) && \
258 (counter & M_COUNTER_OVERFLOW)) { \
259 oprofile_add_sample(get_irq_regs(), n); \
260 w_c0_perfcntr ## n(reg.counter[n]); \
261 handled = IRQ_HANDLED; \
262 }
263 HANDLE_COUNTER(3)
264 HANDLE_COUNTER(2)
265 HANDLE_COUNTER(1)
266 HANDLE_COUNTER(0)
267 }
268
269 return handled;
270}
271
272#define M_CONFIG1_PC (1 << 4)
273
274static inline int __n_counters(void)
275{
276 if (!(read_c0_config1() & M_CONFIG1_PC))
277 return 0;
278 if (!(read_c0_perfctrl0() & M_PERFCTL_MORE))
279 return 1;
280 if (!(read_c0_perfctrl1() & M_PERFCTL_MORE))
281 return 2;
282 if (!(read_c0_perfctrl2() & M_PERFCTL_MORE))
283 return 3;
284
285 return 4;
286}
287
288static inline int n_counters(void)
289{
290 int counters;
291
292 switch (current_cpu_type()) {
293 case CPU_R10000:
294 counters = 2;
295 break;
296
297 case CPU_R12000:
298 case CPU_R14000:
299 case CPU_R16000:
300 counters = 4;
301 break;
302
303 default:
304 counters = __n_counters();
305 }
306
307 return counters;
308}
309
310static void reset_counters(void *arg)
311{
312 int counters = (int)(long)arg;
313 switch (counters) {
314 case 4:
315 w_c0_perfctrl3(0);
316 w_c0_perfcntr3(0);
317 case 3:
318 w_c0_perfctrl2(0);
319 w_c0_perfcntr2(0);
320 case 2:
321 w_c0_perfctrl1(0);
322 w_c0_perfcntr1(0);
323 case 1:
324 w_c0_perfctrl0(0);
325 w_c0_perfcntr0(0);
326 }
327}
328
329static irqreturn_t mipsxx_perfcount_int(int irq, void *dev_id)
330{
331 return mipsxx_perfcount_handler();
332}
333
334static int __init mipsxx_init(void)
335{
336 int counters;
337
338 counters = n_counters();
339 if (counters == 0) {
340 printk(KERN_ERR "Oprofile: CPU has no performance counters\n");
341 return -ENODEV;
342 }
343
344#ifdef CONFIG_MIPS_MT_SMP
345 cpu_has_mipsmt_pertccounters = read_c0_config7() & (1<<19);
346 if (!cpu_has_mipsmt_pertccounters)
347 counters = counters_total_to_per_cpu(counters);
348#endif
349 on_each_cpu(reset_counters, (void *)(long)counters, 1);
350
351 op_model_mipsxx_ops.num_counters = counters;
352 switch (current_cpu_type()) {
353 case CPU_M14KC:
354 op_model_mipsxx_ops.cpu_type = "mips/M14Kc";
355 break;
356
357 case CPU_M14KEC:
358 op_model_mipsxx_ops.cpu_type = "mips/M14KEc";
359 break;
360
361 case CPU_20KC:
362 op_model_mipsxx_ops.cpu_type = "mips/20K";
363 break;
364
365 case CPU_24K:
366 op_model_mipsxx_ops.cpu_type = "mips/24K";
367 break;
368
369 case CPU_25KF:
370 op_model_mipsxx_ops.cpu_type = "mips/25K";
371 break;
372
373 case CPU_1004K:
374 case CPU_34K:
375 op_model_mipsxx_ops.cpu_type = "mips/34K";
376 break;
377
378 case CPU_1074K:
379 case CPU_74K:
380 op_model_mipsxx_ops.cpu_type = "mips/74K";
381 break;
382
383 case CPU_INTERAPTIV:
384 op_model_mipsxx_ops.cpu_type = "mips/interAptiv";
385 break;
386
387 case CPU_PROAPTIV:
388 op_model_mipsxx_ops.cpu_type = "mips/proAptiv";
389 break;
390
391 case CPU_P5600:
392 op_model_mipsxx_ops.cpu_type = "mips/P5600";
393 break;
394
395 case CPU_I6400:
396 op_model_mipsxx_ops.cpu_type = "mips/I6400";
397 break;
398
399 case CPU_M5150:
400 op_model_mipsxx_ops.cpu_type = "mips/M5150";
401 break;
402
403 case CPU_5KC:
404 op_model_mipsxx_ops.cpu_type = "mips/5K";
405 break;
406
407 case CPU_R10000:
408 if ((current_cpu_data.processor_id & 0xff) == 0x20)
409 op_model_mipsxx_ops.cpu_type = "mips/r10000-v2.x";
410 else
411 op_model_mipsxx_ops.cpu_type = "mips/r10000";
412 break;
413
414 case CPU_R12000:
415 case CPU_R14000:
416 op_model_mipsxx_ops.cpu_type = "mips/r12000";
417 break;
418
419 case CPU_R16000:
420 op_model_mipsxx_ops.cpu_type = "mips/r16000";
421 break;
422
423 case CPU_SB1:
424 case CPU_SB1A:
425 op_model_mipsxx_ops.cpu_type = "mips/sb1";
426 break;
427
428 case CPU_LOONGSON1:
429 op_model_mipsxx_ops.cpu_type = "mips/loongson1";
430 break;
431
432 case CPU_XLR:
433 op_model_mipsxx_ops.cpu_type = "mips/xlr";
434 break;
435
436 default:
437 printk(KERN_ERR "Profiling unsupported for this CPU\n");
438
439 return -ENODEV;
440 }
441
442 save_perf_irq = perf_irq;
443 perf_irq = mipsxx_perfcount_handler;
444
445 if (get_c0_perfcount_int)
446 perfcount_irq = get_c0_perfcount_int();
447 else if (cp0_perfcount_irq >= 0)
448 perfcount_irq = MIPS_CPU_IRQ_BASE + cp0_perfcount_irq;
449 else
450 perfcount_irq = -1;
451
452 if (perfcount_irq >= 0)
453 return request_irq(perfcount_irq, mipsxx_perfcount_int,
454 IRQF_PERCPU | IRQF_NOBALANCING |
455 IRQF_NO_THREAD | IRQF_NO_SUSPEND |
456 IRQF_SHARED,
457 "Perfcounter", save_perf_irq);
458
459 return 0;
460}
461
462static void mipsxx_exit(void)
463{
464 int counters = op_model_mipsxx_ops.num_counters;
465
466 if (perfcount_irq >= 0)
467 free_irq(perfcount_irq, save_perf_irq);
468
469 counters = counters_per_cpu_to_total(counters);
470 on_each_cpu(reset_counters, (void *)(long)counters, 1);
471
472 perf_irq = save_perf_irq;
473}
474
475struct op_mips_model op_model_mipsxx_ops = {
476 .reg_setup = mipsxx_reg_setup,
477 .cpu_setup = mipsxx_cpu_setup,
478 .init = mipsxx_init,
479 .exit = mipsxx_exit,
480 .cpu_start = mipsxx_cpu_start,
481 .cpu_stop = mipsxx_cpu_stop,
482};
1/*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
4 * for more details.
5 *
6 * Copyright (C) 2004, 05, 06 by Ralf Baechle
7 * Copyright (C) 2005 by MIPS Technologies, Inc.
8 */
9#include <linux/cpumask.h>
10#include <linux/oprofile.h>
11#include <linux/interrupt.h>
12#include <linux/smp.h>
13#include <asm/irq_regs.h>
14#include <asm/time.h>
15
16#include "op_impl.h"
17
18#define M_PERFCTL_EVENT(event) (((event) << MIPS_PERFCTRL_EVENT_S) & \
19 MIPS_PERFCTRL_EVENT)
20#define M_PERFCTL_VPEID(vpe) ((vpe) << MIPS_PERFCTRL_VPEID_S)
21
22#define M_COUNTER_OVERFLOW (1UL << 31)
23
24static int (*save_perf_irq)(void);
25static int perfcount_irq;
26
27/*
28 * XLR has only one set of counters per core. Designate the
29 * first hardware thread in the core for setup and init.
30 * Skip CPUs with non-zero hardware thread id (4 hwt per core)
31 */
32#if defined(CONFIG_CPU_XLR) && defined(CONFIG_SMP)
33#define oprofile_skip_cpu(c) ((cpu_logical_map(c) & 0x3) != 0)
34#else
35#define oprofile_skip_cpu(c) 0
36#endif
37
38#ifdef CONFIG_MIPS_MT_SMP
39#define WHAT (MIPS_PERFCTRL_MT_EN_VPE | \
40 M_PERFCTL_VPEID(cpu_vpe_id(¤t_cpu_data)))
41#define vpe_id() (cpu_has_mipsmt_pertccounters ? \
42 0 : cpu_vpe_id(¤t_cpu_data))
43
44/*
45 * The number of bits to shift to convert between counters per core and
46 * counters per VPE. There is no reasonable interface atm to obtain the
47 * number of VPEs used by Linux and in the 34K this number is fixed to two
48 * anyways so we hardcore a few things here for the moment. The way it's
49 * done here will ensure that oprofile VSMP kernel will run right on a lesser
50 * core like a 24K also or with maxcpus=1.
51 */
52static inline unsigned int vpe_shift(void)
53{
54 if (num_possible_cpus() > 1)
55 return 1;
56
57 return 0;
58}
59
60#else
61
62#define WHAT 0
63#define vpe_id() 0
64
65static inline unsigned int vpe_shift(void)
66{
67 return 0;
68}
69
70#endif
71
72static inline unsigned int counters_total_to_per_cpu(unsigned int counters)
73{
74 return counters >> vpe_shift();
75}
76
77static inline unsigned int counters_per_cpu_to_total(unsigned int counters)
78{
79 return counters << vpe_shift();
80}
81
82#define __define_perf_accessors(r, n, np) \
83 \
84static inline unsigned int r_c0_ ## r ## n(void) \
85{ \
86 unsigned int cpu = vpe_id(); \
87 \
88 switch (cpu) { \
89 case 0: \
90 return read_c0_ ## r ## n(); \
91 case 1: \
92 return read_c0_ ## r ## np(); \
93 default: \
94 BUG(); \
95 } \
96 return 0; \
97} \
98 \
99static inline void w_c0_ ## r ## n(unsigned int value) \
100{ \
101 unsigned int cpu = vpe_id(); \
102 \
103 switch (cpu) { \
104 case 0: \
105 write_c0_ ## r ## n(value); \
106 return; \
107 case 1: \
108 write_c0_ ## r ## np(value); \
109 return; \
110 default: \
111 BUG(); \
112 } \
113 return; \
114} \
115
116__define_perf_accessors(perfcntr, 0, 2)
117__define_perf_accessors(perfcntr, 1, 3)
118__define_perf_accessors(perfcntr, 2, 0)
119__define_perf_accessors(perfcntr, 3, 1)
120
121__define_perf_accessors(perfctrl, 0, 2)
122__define_perf_accessors(perfctrl, 1, 3)
123__define_perf_accessors(perfctrl, 2, 0)
124__define_perf_accessors(perfctrl, 3, 1)
125
126struct op_mips_model op_model_mipsxx_ops;
127
128static struct mipsxx_register_config {
129 unsigned int control[4];
130 unsigned int counter[4];
131} reg;
132
133/* Compute all of the registers in preparation for enabling profiling. */
134
135static void mipsxx_reg_setup(struct op_counter_config *ctr)
136{
137 unsigned int counters = op_model_mipsxx_ops.num_counters;
138 int i;
139
140 /* Compute the performance counter control word. */
141 for (i = 0; i < counters; i++) {
142 reg.control[i] = 0;
143 reg.counter[i] = 0;
144
145 if (!ctr[i].enabled)
146 continue;
147
148 reg.control[i] = M_PERFCTL_EVENT(ctr[i].event) |
149 MIPS_PERFCTRL_IE;
150 if (ctr[i].kernel)
151 reg.control[i] |= MIPS_PERFCTRL_K;
152 if (ctr[i].user)
153 reg.control[i] |= MIPS_PERFCTRL_U;
154 if (ctr[i].exl)
155 reg.control[i] |= MIPS_PERFCTRL_EXL;
156 if (boot_cpu_type() == CPU_XLR)
157 reg.control[i] |= XLR_PERFCTRL_ALLTHREADS;
158 reg.counter[i] = 0x80000000 - ctr[i].count;
159 }
160}
161
162/* Program all of the registers in preparation for enabling profiling. */
163
164static void mipsxx_cpu_setup(void *args)
165{
166 unsigned int counters = op_model_mipsxx_ops.num_counters;
167
168 if (oprofile_skip_cpu(smp_processor_id()))
169 return;
170
171 switch (counters) {
172 case 4:
173 w_c0_perfctrl3(0);
174 w_c0_perfcntr3(reg.counter[3]);
175 fallthrough;
176 case 3:
177 w_c0_perfctrl2(0);
178 w_c0_perfcntr2(reg.counter[2]);
179 fallthrough;
180 case 2:
181 w_c0_perfctrl1(0);
182 w_c0_perfcntr1(reg.counter[1]);
183 fallthrough;
184 case 1:
185 w_c0_perfctrl0(0);
186 w_c0_perfcntr0(reg.counter[0]);
187 }
188}
189
190/* Start all counters on current CPU */
191static void mipsxx_cpu_start(void *args)
192{
193 unsigned int counters = op_model_mipsxx_ops.num_counters;
194
195 if (oprofile_skip_cpu(smp_processor_id()))
196 return;
197
198 switch (counters) {
199 case 4:
200 w_c0_perfctrl3(WHAT | reg.control[3]);
201 fallthrough;
202 case 3:
203 w_c0_perfctrl2(WHAT | reg.control[2]);
204 fallthrough;
205 case 2:
206 w_c0_perfctrl1(WHAT | reg.control[1]);
207 fallthrough;
208 case 1:
209 w_c0_perfctrl0(WHAT | reg.control[0]);
210 }
211}
212
213/* Stop all counters on current CPU */
214static void mipsxx_cpu_stop(void *args)
215{
216 unsigned int counters = op_model_mipsxx_ops.num_counters;
217
218 if (oprofile_skip_cpu(smp_processor_id()))
219 return;
220
221 switch (counters) {
222 case 4:
223 w_c0_perfctrl3(0);
224 fallthrough;
225 case 3:
226 w_c0_perfctrl2(0);
227 fallthrough;
228 case 2:
229 w_c0_perfctrl1(0);
230 fallthrough;
231 case 1:
232 w_c0_perfctrl0(0);
233 }
234}
235
236static int mipsxx_perfcount_handler(void)
237{
238 unsigned int counters = op_model_mipsxx_ops.num_counters;
239 unsigned int control;
240 unsigned int counter;
241 int handled = IRQ_NONE;
242
243 if (cpu_has_mips_r2 && !(read_c0_cause() & CAUSEF_PCI))
244 return handled;
245
246 switch (counters) {
247#define HANDLE_COUNTER(n) \
248 case n + 1: \
249 control = r_c0_perfctrl ## n(); \
250 counter = r_c0_perfcntr ## n(); \
251 if ((control & MIPS_PERFCTRL_IE) && \
252 (counter & M_COUNTER_OVERFLOW)) { \
253 oprofile_add_sample(get_irq_regs(), n); \
254 w_c0_perfcntr ## n(reg.counter[n]); \
255 handled = IRQ_HANDLED; \
256 }
257 HANDLE_COUNTER(3)
258 fallthrough;
259 HANDLE_COUNTER(2)
260 fallthrough;
261 HANDLE_COUNTER(1)
262 fallthrough;
263 HANDLE_COUNTER(0)
264 }
265
266 return handled;
267}
268
269static inline int __n_counters(void)
270{
271 if (!cpu_has_perf)
272 return 0;
273 if (!(read_c0_perfctrl0() & MIPS_PERFCTRL_M))
274 return 1;
275 if (!(read_c0_perfctrl1() & MIPS_PERFCTRL_M))
276 return 2;
277 if (!(read_c0_perfctrl2() & MIPS_PERFCTRL_M))
278 return 3;
279
280 return 4;
281}
282
283static inline int n_counters(void)
284{
285 int counters;
286
287 switch (current_cpu_type()) {
288 case CPU_R10000:
289 counters = 2;
290 break;
291
292 case CPU_R12000:
293 case CPU_R14000:
294 case CPU_R16000:
295 counters = 4;
296 break;
297
298 default:
299 counters = __n_counters();
300 }
301
302 return counters;
303}
304
305static void reset_counters(void *arg)
306{
307 int counters = (int)(long)arg;
308 switch (counters) {
309 case 4:
310 w_c0_perfctrl3(0);
311 w_c0_perfcntr3(0);
312 fallthrough;
313 case 3:
314 w_c0_perfctrl2(0);
315 w_c0_perfcntr2(0);
316 fallthrough;
317 case 2:
318 w_c0_perfctrl1(0);
319 w_c0_perfcntr1(0);
320 fallthrough;
321 case 1:
322 w_c0_perfctrl0(0);
323 w_c0_perfcntr0(0);
324 }
325}
326
327static irqreturn_t mipsxx_perfcount_int(int irq, void *dev_id)
328{
329 return mipsxx_perfcount_handler();
330}
331
332static int __init mipsxx_init(void)
333{
334 int counters;
335
336 counters = n_counters();
337 if (counters == 0) {
338 printk(KERN_ERR "Oprofile: CPU has no performance counters\n");
339 return -ENODEV;
340 }
341
342#ifdef CONFIG_MIPS_MT_SMP
343 if (!cpu_has_mipsmt_pertccounters)
344 counters = counters_total_to_per_cpu(counters);
345#endif
346 on_each_cpu(reset_counters, (void *)(long)counters, 1);
347
348 op_model_mipsxx_ops.num_counters = counters;
349 switch (current_cpu_type()) {
350 case CPU_M14KC:
351 op_model_mipsxx_ops.cpu_type = "mips/M14Kc";
352 break;
353
354 case CPU_M14KEC:
355 op_model_mipsxx_ops.cpu_type = "mips/M14KEc";
356 break;
357
358 case CPU_20KC:
359 op_model_mipsxx_ops.cpu_type = "mips/20K";
360 break;
361
362 case CPU_24K:
363 op_model_mipsxx_ops.cpu_type = "mips/24K";
364 break;
365
366 case CPU_25KF:
367 op_model_mipsxx_ops.cpu_type = "mips/25K";
368 break;
369
370 case CPU_1004K:
371 case CPU_34K:
372 op_model_mipsxx_ops.cpu_type = "mips/34K";
373 break;
374
375 case CPU_1074K:
376 case CPU_74K:
377 op_model_mipsxx_ops.cpu_type = "mips/74K";
378 break;
379
380 case CPU_INTERAPTIV:
381 op_model_mipsxx_ops.cpu_type = "mips/interAptiv";
382 break;
383
384 case CPU_PROAPTIV:
385 op_model_mipsxx_ops.cpu_type = "mips/proAptiv";
386 break;
387
388 case CPU_P5600:
389 op_model_mipsxx_ops.cpu_type = "mips/P5600";
390 break;
391
392 case CPU_I6400:
393 op_model_mipsxx_ops.cpu_type = "mips/I6400";
394 break;
395
396 case CPU_M5150:
397 op_model_mipsxx_ops.cpu_type = "mips/M5150";
398 break;
399
400 case CPU_5KC:
401 op_model_mipsxx_ops.cpu_type = "mips/5K";
402 break;
403
404 case CPU_R10000:
405 if ((current_cpu_data.processor_id & 0xff) == 0x20)
406 op_model_mipsxx_ops.cpu_type = "mips/r10000-v2.x";
407 else
408 op_model_mipsxx_ops.cpu_type = "mips/r10000";
409 break;
410
411 case CPU_R12000:
412 case CPU_R14000:
413 op_model_mipsxx_ops.cpu_type = "mips/r12000";
414 break;
415
416 case CPU_R16000:
417 op_model_mipsxx_ops.cpu_type = "mips/r16000";
418 break;
419
420 case CPU_SB1:
421 case CPU_SB1A:
422 op_model_mipsxx_ops.cpu_type = "mips/sb1";
423 break;
424
425 case CPU_LOONGSON32:
426 op_model_mipsxx_ops.cpu_type = "mips/loongson1";
427 break;
428
429 case CPU_XLR:
430 op_model_mipsxx_ops.cpu_type = "mips/xlr";
431 break;
432
433 default:
434 printk(KERN_ERR "Profiling unsupported for this CPU\n");
435
436 return -ENODEV;
437 }
438
439 save_perf_irq = perf_irq;
440 perf_irq = mipsxx_perfcount_handler;
441
442 if (get_c0_perfcount_int)
443 perfcount_irq = get_c0_perfcount_int();
444 else if (cp0_perfcount_irq >= 0)
445 perfcount_irq = MIPS_CPU_IRQ_BASE + cp0_perfcount_irq;
446 else
447 perfcount_irq = -1;
448
449 if (perfcount_irq >= 0)
450 return request_irq(perfcount_irq, mipsxx_perfcount_int,
451 IRQF_PERCPU | IRQF_NOBALANCING |
452 IRQF_NO_THREAD | IRQF_NO_SUSPEND |
453 IRQF_SHARED,
454 "Perfcounter", save_perf_irq);
455
456 return 0;
457}
458
459static void mipsxx_exit(void)
460{
461 int counters = op_model_mipsxx_ops.num_counters;
462
463 if (perfcount_irq >= 0)
464 free_irq(perfcount_irq, save_perf_irq);
465
466 counters = counters_per_cpu_to_total(counters);
467 on_each_cpu(reset_counters, (void *)(long)counters, 1);
468
469 perf_irq = save_perf_irq;
470}
471
472struct op_mips_model op_model_mipsxx_ops = {
473 .reg_setup = mipsxx_reg_setup,
474 .cpu_setup = mipsxx_cpu_setup,
475 .init = mipsxx_init,
476 .exit = mipsxx_exit,
477 .cpu_start = mipsxx_cpu_start,
478 .cpu_stop = mipsxx_cpu_stop,
479};