Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (c) 2017 Linaro Limited. All rights reserved.
4 *
5 * Author: Leo Yan <leo.yan@linaro.org>
6 */
7#include <linux/amba/bus.h>
8#include <linux/coresight.h>
9#include <linux/cpu.h>
10#include <linux/debugfs.h>
11#include <linux/delay.h>
12#include <linux/device.h>
13#include <linux/err.h>
14#include <linux/init.h>
15#include <linux/io.h>
16#include <linux/iopoll.h>
17#include <linux/kernel.h>
18#include <linux/module.h>
19#include <linux/moduleparam.h>
20#include <linux/panic_notifier.h>
21#include <linux/pm_qos.h>
22#include <linux/slab.h>
23#include <linux/smp.h>
24#include <linux/types.h>
25#include <linux/uaccess.h>
26
27#include "coresight-priv.h"
28
29#define EDPCSR 0x0A0
30#define EDCIDSR 0x0A4
31#define EDVIDSR 0x0A8
32#define EDPCSR_HI 0x0AC
33#define EDOSLAR 0x300
34#define EDPRCR 0x310
35#define EDPRSR 0x314
36#define EDDEVID1 0xFC4
37#define EDDEVID 0xFC8
38
39#define EDPCSR_PROHIBITED 0xFFFFFFFF
40
41/* bits definition for EDPCSR */
42#define EDPCSR_THUMB BIT(0)
43#define EDPCSR_ARM_INST_MASK GENMASK(31, 2)
44#define EDPCSR_THUMB_INST_MASK GENMASK(31, 1)
45
46/* bits definition for EDPRCR */
47#define EDPRCR_COREPURQ BIT(3)
48#define EDPRCR_CORENPDRQ BIT(0)
49
50/* bits definition for EDPRSR */
51#define EDPRSR_DLK BIT(6)
52#define EDPRSR_PU BIT(0)
53
54/* bits definition for EDVIDSR */
55#define EDVIDSR_NS BIT(31)
56#define EDVIDSR_E2 BIT(30)
57#define EDVIDSR_E3 BIT(29)
58#define EDVIDSR_HV BIT(28)
59#define EDVIDSR_VMID GENMASK(7, 0)
60
61/*
62 * bits definition for EDDEVID1:PSCROffset
63 *
64 * NOTE: armv8 and armv7 have different definition for the register,
65 * so consolidate the bits definition as below:
66 *
67 * 0b0000 - Sample offset applies based on the instruction state, we
68 * rely on EDDEVID to check if EDPCSR is implemented or not
69 * 0b0001 - No offset applies.
70 * 0b0010 - No offset applies, but do not use in AArch32 mode
71 *
72 */
73#define EDDEVID1_PCSR_OFFSET_MASK GENMASK(3, 0)
74#define EDDEVID1_PCSR_OFFSET_INS_SET (0x0)
75#define EDDEVID1_PCSR_NO_OFFSET_DIS_AARCH32 (0x2)
76
77/* bits definition for EDDEVID */
78#define EDDEVID_PCSAMPLE_MODE GENMASK(3, 0)
79#define EDDEVID_IMPL_EDPCSR (0x1)
80#define EDDEVID_IMPL_EDPCSR_EDCIDSR (0x2)
81#define EDDEVID_IMPL_FULL (0x3)
82
83#define DEBUG_WAIT_SLEEP 1000
84#define DEBUG_WAIT_TIMEOUT 32000
85
86struct debug_drvdata {
87 void __iomem *base;
88 struct device *dev;
89 int cpu;
90
91 bool edpcsr_present;
92 bool edcidsr_present;
93 bool edvidsr_present;
94 bool pc_has_offset;
95
96 u32 edpcsr;
97 u32 edpcsr_hi;
98 u32 edprsr;
99 u32 edvidsr;
100 u32 edcidsr;
101};
102
103static DEFINE_MUTEX(debug_lock);
104static DEFINE_PER_CPU(struct debug_drvdata *, debug_drvdata);
105static int debug_count;
106static struct dentry *debug_debugfs_dir;
107
108static bool debug_enable = IS_ENABLED(CONFIG_CORESIGHT_CPU_DEBUG_DEFAULT_ON);
109module_param_named(enable, debug_enable, bool, 0600);
110MODULE_PARM_DESC(enable, "Control to enable coresight CPU debug functionality");
111
112static void debug_os_unlock(struct debug_drvdata *drvdata)
113{
114 /* Unlocks the debug registers */
115 writel_relaxed(0x0, drvdata->base + EDOSLAR);
116
117 /* Make sure the registers are unlocked before accessing */
118 wmb();
119}
120
121/*
122 * According to ARM DDI 0487A.k, before access external debug
123 * registers should firstly check the access permission; if any
124 * below condition has been met then cannot access debug
125 * registers to avoid lockup issue:
126 *
127 * - CPU power domain is powered off;
128 * - The OS Double Lock is locked;
129 *
130 * By checking EDPRSR can get to know if meet these conditions.
131 */
132static bool debug_access_permitted(struct debug_drvdata *drvdata)
133{
134 /* CPU is powered off */
135 if (!(drvdata->edprsr & EDPRSR_PU))
136 return false;
137
138 /* The OS Double Lock is locked */
139 if (drvdata->edprsr & EDPRSR_DLK)
140 return false;
141
142 return true;
143}
144
145static void debug_force_cpu_powered_up(struct debug_drvdata *drvdata)
146{
147 u32 edprcr;
148
149try_again:
150
151 /*
152 * Send request to power management controller and assert
153 * DBGPWRUPREQ signal; if power management controller has
154 * sane implementation, it should enable CPU power domain
155 * in case CPU is in low power state.
156 */
157 edprcr = readl_relaxed(drvdata->base + EDPRCR);
158 edprcr |= EDPRCR_COREPURQ;
159 writel_relaxed(edprcr, drvdata->base + EDPRCR);
160
161 /* Wait for CPU to be powered up (timeout~=32ms) */
162 if (readx_poll_timeout_atomic(readl_relaxed, drvdata->base + EDPRSR,
163 drvdata->edprsr, (drvdata->edprsr & EDPRSR_PU),
164 DEBUG_WAIT_SLEEP, DEBUG_WAIT_TIMEOUT)) {
165 /*
166 * Unfortunately the CPU cannot be powered up, so return
167 * back and later has no permission to access other
168 * registers. For this case, should disable CPU low power
169 * states to ensure CPU power domain is enabled!
170 */
171 dev_err(drvdata->dev, "%s: power up request for CPU%d failed\n",
172 __func__, drvdata->cpu);
173 return;
174 }
175
176 /*
177 * At this point the CPU is powered up, so set the no powerdown
178 * request bit so we don't lose power and emulate power down.
179 */
180 edprcr = readl_relaxed(drvdata->base + EDPRCR);
181 edprcr |= EDPRCR_COREPURQ | EDPRCR_CORENPDRQ;
182 writel_relaxed(edprcr, drvdata->base + EDPRCR);
183
184 drvdata->edprsr = readl_relaxed(drvdata->base + EDPRSR);
185
186 /* The core power domain got switched off on use, try again */
187 if (unlikely(!(drvdata->edprsr & EDPRSR_PU)))
188 goto try_again;
189}
190
191static void debug_read_regs(struct debug_drvdata *drvdata)
192{
193 u32 save_edprcr;
194
195 CS_UNLOCK(drvdata->base);
196
197 /* Unlock os lock */
198 debug_os_unlock(drvdata);
199
200 /* Save EDPRCR register */
201 save_edprcr = readl_relaxed(drvdata->base + EDPRCR);
202
203 /*
204 * Ensure CPU power domain is enabled to let registers
205 * are accessiable.
206 */
207 debug_force_cpu_powered_up(drvdata);
208
209 if (!debug_access_permitted(drvdata))
210 goto out;
211
212 drvdata->edpcsr = readl_relaxed(drvdata->base + EDPCSR);
213
214 /*
215 * As described in ARM DDI 0487A.k, if the processing
216 * element (PE) is in debug state, or sample-based
217 * profiling is prohibited, EDPCSR reads as 0xFFFFFFFF;
218 * EDCIDSR, EDVIDSR and EDPCSR_HI registers also become
219 * UNKNOWN state. So directly bail out for this case.
220 */
221 if (drvdata->edpcsr == EDPCSR_PROHIBITED)
222 goto out;
223
224 /*
225 * A read of the EDPCSR normally has the side-effect of
226 * indirectly writing to EDCIDSR, EDVIDSR and EDPCSR_HI;
227 * at this point it's safe to read value from them.
228 */
229 if (IS_ENABLED(CONFIG_64BIT))
230 drvdata->edpcsr_hi = readl_relaxed(drvdata->base + EDPCSR_HI);
231
232 if (drvdata->edcidsr_present)
233 drvdata->edcidsr = readl_relaxed(drvdata->base + EDCIDSR);
234
235 if (drvdata->edvidsr_present)
236 drvdata->edvidsr = readl_relaxed(drvdata->base + EDVIDSR);
237
238out:
239 /* Restore EDPRCR register */
240 writel_relaxed(save_edprcr, drvdata->base + EDPRCR);
241
242 CS_LOCK(drvdata->base);
243}
244
245#ifdef CONFIG_64BIT
246static unsigned long debug_adjust_pc(struct debug_drvdata *drvdata)
247{
248 return (unsigned long)drvdata->edpcsr_hi << 32 |
249 (unsigned long)drvdata->edpcsr;
250}
251#else
252static unsigned long debug_adjust_pc(struct debug_drvdata *drvdata)
253{
254 unsigned long arm_inst_offset = 0, thumb_inst_offset = 0;
255 unsigned long pc;
256
257 pc = (unsigned long)drvdata->edpcsr;
258
259 if (drvdata->pc_has_offset) {
260 arm_inst_offset = 8;
261 thumb_inst_offset = 4;
262 }
263
264 /* Handle thumb instruction */
265 if (pc & EDPCSR_THUMB) {
266 pc = (pc & EDPCSR_THUMB_INST_MASK) - thumb_inst_offset;
267 return pc;
268 }
269
270 /*
271 * Handle arm instruction offset, if the arm instruction
272 * is not 4 byte alignment then it's possible the case
273 * for implementation defined; keep original value for this
274 * case and print info for notice.
275 */
276 if (pc & BIT(1))
277 dev_emerg(drvdata->dev,
278 "Instruction offset is implementation defined\n");
279 else
280 pc = (pc & EDPCSR_ARM_INST_MASK) - arm_inst_offset;
281
282 return pc;
283}
284#endif
285
286static void debug_dump_regs(struct debug_drvdata *drvdata)
287{
288 struct device *dev = drvdata->dev;
289 unsigned long pc;
290
291 dev_emerg(dev, " EDPRSR: %08x (Power:%s DLK:%s)\n",
292 drvdata->edprsr,
293 drvdata->edprsr & EDPRSR_PU ? "On" : "Off",
294 drvdata->edprsr & EDPRSR_DLK ? "Lock" : "Unlock");
295
296 if (!debug_access_permitted(drvdata)) {
297 dev_emerg(dev, "No permission to access debug registers!\n");
298 return;
299 }
300
301 if (drvdata->edpcsr == EDPCSR_PROHIBITED) {
302 dev_emerg(dev, "CPU is in Debug state or profiling is prohibited!\n");
303 return;
304 }
305
306 pc = debug_adjust_pc(drvdata);
307 dev_emerg(dev, " EDPCSR: %pS\n", (void *)pc);
308
309 if (drvdata->edcidsr_present)
310 dev_emerg(dev, " EDCIDSR: %08x\n", drvdata->edcidsr);
311
312 if (drvdata->edvidsr_present)
313 dev_emerg(dev, " EDVIDSR: %08x (State:%s Mode:%s Width:%dbits VMID:%x)\n",
314 drvdata->edvidsr,
315 drvdata->edvidsr & EDVIDSR_NS ?
316 "Non-secure" : "Secure",
317 drvdata->edvidsr & EDVIDSR_E3 ? "EL3" :
318 (drvdata->edvidsr & EDVIDSR_E2 ?
319 "EL2" : "EL1/0"),
320 drvdata->edvidsr & EDVIDSR_HV ? 64 : 32,
321 drvdata->edvidsr & (u32)EDVIDSR_VMID);
322}
323
324static void debug_init_arch_data(void *info)
325{
326 struct debug_drvdata *drvdata = info;
327 u32 mode, pcsr_offset;
328 u32 eddevid, eddevid1;
329
330 CS_UNLOCK(drvdata->base);
331
332 /* Read device info */
333 eddevid = readl_relaxed(drvdata->base + EDDEVID);
334 eddevid1 = readl_relaxed(drvdata->base + EDDEVID1);
335
336 CS_LOCK(drvdata->base);
337
338 /* Parse implementation feature */
339 mode = eddevid & EDDEVID_PCSAMPLE_MODE;
340 pcsr_offset = eddevid1 & EDDEVID1_PCSR_OFFSET_MASK;
341
342 drvdata->edpcsr_present = false;
343 drvdata->edcidsr_present = false;
344 drvdata->edvidsr_present = false;
345 drvdata->pc_has_offset = false;
346
347 switch (mode) {
348 case EDDEVID_IMPL_FULL:
349 drvdata->edvidsr_present = true;
350 fallthrough;
351 case EDDEVID_IMPL_EDPCSR_EDCIDSR:
352 drvdata->edcidsr_present = true;
353 fallthrough;
354 case EDDEVID_IMPL_EDPCSR:
355 /*
356 * In ARM DDI 0487A.k, the EDDEVID1.PCSROffset is used to
357 * define if has the offset for PC sampling value; if read
358 * back EDDEVID1.PCSROffset == 0x2, then this means the debug
359 * module does not sample the instruction set state when
360 * armv8 CPU in AArch32 state.
361 */
362 drvdata->edpcsr_present =
363 ((IS_ENABLED(CONFIG_64BIT) && pcsr_offset != 0) ||
364 (pcsr_offset != EDDEVID1_PCSR_NO_OFFSET_DIS_AARCH32));
365
366 drvdata->pc_has_offset =
367 (pcsr_offset == EDDEVID1_PCSR_OFFSET_INS_SET);
368 break;
369 default:
370 break;
371 }
372}
373
374/*
375 * Dump out information on panic.
376 */
377static int debug_notifier_call(struct notifier_block *self,
378 unsigned long v, void *p)
379{
380 int cpu;
381 struct debug_drvdata *drvdata;
382
383 /* Bail out if we can't acquire the mutex or the functionality is off */
384 if (!mutex_trylock(&debug_lock))
385 return NOTIFY_DONE;
386
387 if (!debug_enable)
388 goto skip_dump;
389
390 pr_emerg("ARM external debug module:\n");
391
392 for_each_possible_cpu(cpu) {
393 drvdata = per_cpu(debug_drvdata, cpu);
394 if (!drvdata)
395 continue;
396
397 dev_emerg(drvdata->dev, "CPU[%d]:\n", drvdata->cpu);
398
399 debug_read_regs(drvdata);
400 debug_dump_regs(drvdata);
401 }
402
403skip_dump:
404 mutex_unlock(&debug_lock);
405 return NOTIFY_DONE;
406}
407
408static struct notifier_block debug_notifier = {
409 .notifier_call = debug_notifier_call,
410};
411
412static int debug_enable_func(void)
413{
414 struct debug_drvdata *drvdata;
415 int cpu, ret = 0;
416 cpumask_t mask;
417
418 /*
419 * Use cpumask to track which debug power domains have
420 * been powered on and use it to handle failure case.
421 */
422 cpumask_clear(&mask);
423
424 for_each_possible_cpu(cpu) {
425 drvdata = per_cpu(debug_drvdata, cpu);
426 if (!drvdata)
427 continue;
428
429 ret = pm_runtime_get_sync(drvdata->dev);
430 if (ret < 0)
431 goto err;
432 else
433 cpumask_set_cpu(cpu, &mask);
434 }
435
436 return 0;
437
438err:
439 /*
440 * If pm_runtime_get_sync() has failed, need rollback on
441 * all the other CPUs that have been enabled before that.
442 */
443 for_each_cpu(cpu, &mask) {
444 drvdata = per_cpu(debug_drvdata, cpu);
445 pm_runtime_put_noidle(drvdata->dev);
446 }
447
448 return ret;
449}
450
451static int debug_disable_func(void)
452{
453 struct debug_drvdata *drvdata;
454 int cpu, ret, err = 0;
455
456 /*
457 * Disable debug power domains, records the error and keep
458 * circling through all other CPUs when an error has been
459 * encountered.
460 */
461 for_each_possible_cpu(cpu) {
462 drvdata = per_cpu(debug_drvdata, cpu);
463 if (!drvdata)
464 continue;
465
466 ret = pm_runtime_put(drvdata->dev);
467 if (ret < 0)
468 err = ret;
469 }
470
471 return err;
472}
473
474static ssize_t debug_func_knob_write(struct file *f,
475 const char __user *buf, size_t count, loff_t *ppos)
476{
477 u8 val;
478 int ret;
479
480 ret = kstrtou8_from_user(buf, count, 2, &val);
481 if (ret)
482 return ret;
483
484 mutex_lock(&debug_lock);
485
486 if (val == debug_enable)
487 goto out;
488
489 if (val)
490 ret = debug_enable_func();
491 else
492 ret = debug_disable_func();
493
494 if (ret) {
495 pr_err("%s: unable to %s debug function: %d\n",
496 __func__, val ? "enable" : "disable", ret);
497 goto err;
498 }
499
500 debug_enable = val;
501out:
502 ret = count;
503err:
504 mutex_unlock(&debug_lock);
505 return ret;
506}
507
508static ssize_t debug_func_knob_read(struct file *f,
509 char __user *ubuf, size_t count, loff_t *ppos)
510{
511 ssize_t ret;
512 char buf[3];
513
514 mutex_lock(&debug_lock);
515 snprintf(buf, sizeof(buf), "%d\n", debug_enable);
516 mutex_unlock(&debug_lock);
517
518 ret = simple_read_from_buffer(ubuf, count, ppos, buf, sizeof(buf));
519 return ret;
520}
521
522static const struct file_operations debug_func_knob_fops = {
523 .open = simple_open,
524 .read = debug_func_knob_read,
525 .write = debug_func_knob_write,
526};
527
528static int debug_func_init(void)
529{
530 int ret;
531
532 /* Create debugfs node */
533 debug_debugfs_dir = debugfs_create_dir("coresight_cpu_debug", NULL);
534 debugfs_create_file("enable", 0644, debug_debugfs_dir, NULL,
535 &debug_func_knob_fops);
536
537 /* Register function to be called for panic */
538 ret = atomic_notifier_chain_register(&panic_notifier_list,
539 &debug_notifier);
540 if (ret) {
541 pr_err("%s: unable to register notifier: %d\n",
542 __func__, ret);
543 goto err;
544 }
545
546 return 0;
547
548err:
549 debugfs_remove_recursive(debug_debugfs_dir);
550 return ret;
551}
552
553static void debug_func_exit(void)
554{
555 atomic_notifier_chain_unregister(&panic_notifier_list,
556 &debug_notifier);
557 debugfs_remove_recursive(debug_debugfs_dir);
558}
559
560static int debug_probe(struct amba_device *adev, const struct amba_id *id)
561{
562 void __iomem *base;
563 struct device *dev = &adev->dev;
564 struct debug_drvdata *drvdata;
565 struct resource *res = &adev->res;
566 int ret;
567
568 drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL);
569 if (!drvdata)
570 return -ENOMEM;
571
572 drvdata->cpu = coresight_get_cpu(dev);
573 if (drvdata->cpu < 0)
574 return drvdata->cpu;
575
576 if (per_cpu(debug_drvdata, drvdata->cpu)) {
577 dev_err(dev, "CPU%d drvdata has already been initialized\n",
578 drvdata->cpu);
579 return -EBUSY;
580 }
581
582 drvdata->dev = &adev->dev;
583 amba_set_drvdata(adev, drvdata);
584
585 /* Validity for the resource is already checked by the AMBA core */
586 base = devm_ioremap_resource(dev, res);
587 if (IS_ERR(base))
588 return PTR_ERR(base);
589
590 drvdata->base = base;
591
592 cpus_read_lock();
593 per_cpu(debug_drvdata, drvdata->cpu) = drvdata;
594 ret = smp_call_function_single(drvdata->cpu, debug_init_arch_data,
595 drvdata, 1);
596 cpus_read_unlock();
597
598 if (ret) {
599 dev_err(dev, "CPU%d debug arch init failed\n", drvdata->cpu);
600 goto err;
601 }
602
603 if (!drvdata->edpcsr_present) {
604 dev_err(dev, "CPU%d sample-based profiling isn't implemented\n",
605 drvdata->cpu);
606 ret = -ENXIO;
607 goto err;
608 }
609
610 if (!debug_count++) {
611 ret = debug_func_init();
612 if (ret)
613 goto err_func_init;
614 }
615
616 mutex_lock(&debug_lock);
617 /* Turn off debug power domain if debugging is disabled */
618 if (!debug_enable)
619 pm_runtime_put(dev);
620 mutex_unlock(&debug_lock);
621
622 dev_info(dev, "Coresight debug-CPU%d initialized\n", drvdata->cpu);
623 return 0;
624
625err_func_init:
626 debug_count--;
627err:
628 per_cpu(debug_drvdata, drvdata->cpu) = NULL;
629 return ret;
630}
631
632static void debug_remove(struct amba_device *adev)
633{
634 struct device *dev = &adev->dev;
635 struct debug_drvdata *drvdata = amba_get_drvdata(adev);
636
637 per_cpu(debug_drvdata, drvdata->cpu) = NULL;
638
639 mutex_lock(&debug_lock);
640 /* Turn off debug power domain before rmmod the module */
641 if (debug_enable)
642 pm_runtime_put(dev);
643 mutex_unlock(&debug_lock);
644
645 if (!--debug_count)
646 debug_func_exit();
647}
648
649static const struct amba_cs_uci_id uci_id_debug[] = {
650 {
651 /* CPU Debug UCI data */
652 .devarch = 0x47706a15,
653 .devarch_mask = 0xfff0ffff,
654 .devtype = 0x00000015,
655 }
656};
657
658static const struct amba_id debug_ids[] = {
659 CS_AMBA_ID(0x000bbd03), /* Cortex-A53 */
660 CS_AMBA_ID(0x000bbd07), /* Cortex-A57 */
661 CS_AMBA_ID(0x000bbd08), /* Cortex-A72 */
662 CS_AMBA_ID(0x000bbd09), /* Cortex-A73 */
663 CS_AMBA_UCI_ID(0x000f0205, uci_id_debug), /* Qualcomm Kryo */
664 CS_AMBA_UCI_ID(0x000f0211, uci_id_debug), /* Qualcomm Kryo */
665 {},
666};
667
668MODULE_DEVICE_TABLE(amba, debug_ids);
669
670static struct amba_driver debug_driver = {
671 .drv = {
672 .name = "coresight-cpu-debug",
673 .suppress_bind_attrs = true,
674 },
675 .probe = debug_probe,
676 .remove = debug_remove,
677 .id_table = debug_ids,
678};
679
680module_amba_driver(debug_driver);
681
682MODULE_AUTHOR("Leo Yan <leo.yan@linaro.org>");
683MODULE_DESCRIPTION("ARM Coresight CPU Debug Driver");
684MODULE_LICENSE("GPL");
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (c) 2017 Linaro Limited. All rights reserved.
4 *
5 * Author: Leo Yan <leo.yan@linaro.org>
6 */
7#include <linux/acpi.h>
8#include <linux/amba/bus.h>
9#include <linux/coresight.h>
10#include <linux/cpu.h>
11#include <linux/debugfs.h>
12#include <linux/delay.h>
13#include <linux/device.h>
14#include <linux/err.h>
15#include <linux/init.h>
16#include <linux/io.h>
17#include <linux/iopoll.h>
18#include <linux/kernel.h>
19#include <linux/module.h>
20#include <linux/moduleparam.h>
21#include <linux/panic_notifier.h>
22#include <linux/platform_device.h>
23#include <linux/pm_qos.h>
24#include <linux/slab.h>
25#include <linux/smp.h>
26#include <linux/types.h>
27#include <linux/uaccess.h>
28
29#include "coresight-priv.h"
30
31#define EDPCSR 0x0A0
32#define EDCIDSR 0x0A4
33#define EDVIDSR 0x0A8
34#define EDPCSR_HI 0x0AC
35#define EDOSLAR 0x300
36#define EDPRCR 0x310
37#define EDPRSR 0x314
38#define EDDEVID1 0xFC4
39#define EDDEVID 0xFC8
40
41#define EDPCSR_PROHIBITED 0xFFFFFFFF
42
43/* bits definition for EDPCSR */
44#define EDPCSR_THUMB BIT(0)
45#define EDPCSR_ARM_INST_MASK GENMASK(31, 2)
46#define EDPCSR_THUMB_INST_MASK GENMASK(31, 1)
47
48/* bits definition for EDPRCR */
49#define EDPRCR_COREPURQ BIT(3)
50#define EDPRCR_CORENPDRQ BIT(0)
51
52/* bits definition for EDPRSR */
53#define EDPRSR_DLK BIT(6)
54#define EDPRSR_PU BIT(0)
55
56/* bits definition for EDVIDSR */
57#define EDVIDSR_NS BIT(31)
58#define EDVIDSR_E2 BIT(30)
59#define EDVIDSR_E3 BIT(29)
60#define EDVIDSR_HV BIT(28)
61#define EDVIDSR_VMID GENMASK(7, 0)
62
63/*
64 * bits definition for EDDEVID1:PSCROffset
65 *
66 * NOTE: armv8 and armv7 have different definition for the register,
67 * so consolidate the bits definition as below:
68 *
69 * 0b0000 - Sample offset applies based on the instruction state, we
70 * rely on EDDEVID to check if EDPCSR is implemented or not
71 * 0b0001 - No offset applies.
72 * 0b0010 - No offset applies, but do not use in AArch32 mode
73 *
74 */
75#define EDDEVID1_PCSR_OFFSET_MASK GENMASK(3, 0)
76#define EDDEVID1_PCSR_OFFSET_INS_SET (0x0)
77#define EDDEVID1_PCSR_NO_OFFSET_DIS_AARCH32 (0x2)
78
79/* bits definition for EDDEVID */
80#define EDDEVID_PCSAMPLE_MODE GENMASK(3, 0)
81#define EDDEVID_IMPL_EDPCSR (0x1)
82#define EDDEVID_IMPL_EDPCSR_EDCIDSR (0x2)
83#define EDDEVID_IMPL_FULL (0x3)
84
85#define DEBUG_WAIT_SLEEP 1000
86#define DEBUG_WAIT_TIMEOUT 32000
87
88struct debug_drvdata {
89 struct clk *pclk;
90 void __iomem *base;
91 struct device *dev;
92 int cpu;
93
94 bool edpcsr_present;
95 bool edcidsr_present;
96 bool edvidsr_present;
97 bool pc_has_offset;
98
99 u32 edpcsr;
100 u32 edpcsr_hi;
101 u32 edprsr;
102 u32 edvidsr;
103 u32 edcidsr;
104};
105
106static DEFINE_MUTEX(debug_lock);
107static DEFINE_PER_CPU(struct debug_drvdata *, debug_drvdata);
108static int debug_count;
109static struct dentry *debug_debugfs_dir;
110
111static bool debug_enable = IS_ENABLED(CONFIG_CORESIGHT_CPU_DEBUG_DEFAULT_ON);
112module_param_named(enable, debug_enable, bool, 0600);
113MODULE_PARM_DESC(enable, "Control to enable coresight CPU debug functionality");
114
115static void debug_os_unlock(struct debug_drvdata *drvdata)
116{
117 /* Unlocks the debug registers */
118 writel_relaxed(0x0, drvdata->base + EDOSLAR);
119
120 /* Make sure the registers are unlocked before accessing */
121 wmb();
122}
123
124/*
125 * According to ARM DDI 0487A.k, before access external debug
126 * registers should firstly check the access permission; if any
127 * below condition has been met then cannot access debug
128 * registers to avoid lockup issue:
129 *
130 * - CPU power domain is powered off;
131 * - The OS Double Lock is locked;
132 *
133 * By checking EDPRSR can get to know if meet these conditions.
134 */
135static bool debug_access_permitted(struct debug_drvdata *drvdata)
136{
137 /* CPU is powered off */
138 if (!(drvdata->edprsr & EDPRSR_PU))
139 return false;
140
141 /* The OS Double Lock is locked */
142 if (drvdata->edprsr & EDPRSR_DLK)
143 return false;
144
145 return true;
146}
147
148static void debug_force_cpu_powered_up(struct debug_drvdata *drvdata)
149{
150 u32 edprcr;
151
152try_again:
153
154 /*
155 * Send request to power management controller and assert
156 * DBGPWRUPREQ signal; if power management controller has
157 * sane implementation, it should enable CPU power domain
158 * in case CPU is in low power state.
159 */
160 edprcr = readl_relaxed(drvdata->base + EDPRCR);
161 edprcr |= EDPRCR_COREPURQ;
162 writel_relaxed(edprcr, drvdata->base + EDPRCR);
163
164 /* Wait for CPU to be powered up (timeout~=32ms) */
165 if (readx_poll_timeout_atomic(readl_relaxed, drvdata->base + EDPRSR,
166 drvdata->edprsr, (drvdata->edprsr & EDPRSR_PU),
167 DEBUG_WAIT_SLEEP, DEBUG_WAIT_TIMEOUT)) {
168 /*
169 * Unfortunately the CPU cannot be powered up, so return
170 * back and later has no permission to access other
171 * registers. For this case, should disable CPU low power
172 * states to ensure CPU power domain is enabled!
173 */
174 dev_err(drvdata->dev, "%s: power up request for CPU%d failed\n",
175 __func__, drvdata->cpu);
176 return;
177 }
178
179 /*
180 * At this point the CPU is powered up, so set the no powerdown
181 * request bit so we don't lose power and emulate power down.
182 */
183 edprcr = readl_relaxed(drvdata->base + EDPRCR);
184 edprcr |= EDPRCR_COREPURQ | EDPRCR_CORENPDRQ;
185 writel_relaxed(edprcr, drvdata->base + EDPRCR);
186
187 drvdata->edprsr = readl_relaxed(drvdata->base + EDPRSR);
188
189 /* The core power domain got switched off on use, try again */
190 if (unlikely(!(drvdata->edprsr & EDPRSR_PU)))
191 goto try_again;
192}
193
194static void debug_read_regs(struct debug_drvdata *drvdata)
195{
196 u32 save_edprcr;
197
198 CS_UNLOCK(drvdata->base);
199
200 /* Unlock os lock */
201 debug_os_unlock(drvdata);
202
203 /* Save EDPRCR register */
204 save_edprcr = readl_relaxed(drvdata->base + EDPRCR);
205
206 /*
207 * Ensure CPU power domain is enabled to let registers
208 * are accessiable.
209 */
210 debug_force_cpu_powered_up(drvdata);
211
212 if (!debug_access_permitted(drvdata))
213 goto out;
214
215 drvdata->edpcsr = readl_relaxed(drvdata->base + EDPCSR);
216
217 /*
218 * As described in ARM DDI 0487A.k, if the processing
219 * element (PE) is in debug state, or sample-based
220 * profiling is prohibited, EDPCSR reads as 0xFFFFFFFF;
221 * EDCIDSR, EDVIDSR and EDPCSR_HI registers also become
222 * UNKNOWN state. So directly bail out for this case.
223 */
224 if (drvdata->edpcsr == EDPCSR_PROHIBITED)
225 goto out;
226
227 /*
228 * A read of the EDPCSR normally has the side-effect of
229 * indirectly writing to EDCIDSR, EDVIDSR and EDPCSR_HI;
230 * at this point it's safe to read value from them.
231 */
232 if (IS_ENABLED(CONFIG_64BIT))
233 drvdata->edpcsr_hi = readl_relaxed(drvdata->base + EDPCSR_HI);
234
235 if (drvdata->edcidsr_present)
236 drvdata->edcidsr = readl_relaxed(drvdata->base + EDCIDSR);
237
238 if (drvdata->edvidsr_present)
239 drvdata->edvidsr = readl_relaxed(drvdata->base + EDVIDSR);
240
241out:
242 /* Restore EDPRCR register */
243 writel_relaxed(save_edprcr, drvdata->base + EDPRCR);
244
245 CS_LOCK(drvdata->base);
246}
247
248#ifdef CONFIG_64BIT
249static unsigned long debug_adjust_pc(struct debug_drvdata *drvdata)
250{
251 return (unsigned long)drvdata->edpcsr_hi << 32 |
252 (unsigned long)drvdata->edpcsr;
253}
254#else
255static unsigned long debug_adjust_pc(struct debug_drvdata *drvdata)
256{
257 unsigned long arm_inst_offset = 0, thumb_inst_offset = 0;
258 unsigned long pc;
259
260 pc = (unsigned long)drvdata->edpcsr;
261
262 if (drvdata->pc_has_offset) {
263 arm_inst_offset = 8;
264 thumb_inst_offset = 4;
265 }
266
267 /* Handle thumb instruction */
268 if (pc & EDPCSR_THUMB) {
269 pc = (pc & EDPCSR_THUMB_INST_MASK) - thumb_inst_offset;
270 return pc;
271 }
272
273 /*
274 * Handle arm instruction offset, if the arm instruction
275 * is not 4 byte alignment then it's possible the case
276 * for implementation defined; keep original value for this
277 * case and print info for notice.
278 */
279 if (pc & BIT(1))
280 dev_emerg(drvdata->dev,
281 "Instruction offset is implementation defined\n");
282 else
283 pc = (pc & EDPCSR_ARM_INST_MASK) - arm_inst_offset;
284
285 return pc;
286}
287#endif
288
289static void debug_dump_regs(struct debug_drvdata *drvdata)
290{
291 struct device *dev = drvdata->dev;
292 unsigned long pc;
293
294 dev_emerg(dev, " EDPRSR: %08x (Power:%s DLK:%s)\n",
295 drvdata->edprsr,
296 drvdata->edprsr & EDPRSR_PU ? "On" : "Off",
297 drvdata->edprsr & EDPRSR_DLK ? "Lock" : "Unlock");
298
299 if (!debug_access_permitted(drvdata)) {
300 dev_emerg(dev, "No permission to access debug registers!\n");
301 return;
302 }
303
304 if (drvdata->edpcsr == EDPCSR_PROHIBITED) {
305 dev_emerg(dev, "CPU is in Debug state or profiling is prohibited!\n");
306 return;
307 }
308
309 pc = debug_adjust_pc(drvdata);
310 dev_emerg(dev, " EDPCSR: %pS\n", (void *)pc);
311
312 if (drvdata->edcidsr_present)
313 dev_emerg(dev, " EDCIDSR: %08x\n", drvdata->edcidsr);
314
315 if (drvdata->edvidsr_present)
316 dev_emerg(dev, " EDVIDSR: %08x (State:%s Mode:%s Width:%dbits VMID:%x)\n",
317 drvdata->edvidsr,
318 drvdata->edvidsr & EDVIDSR_NS ?
319 "Non-secure" : "Secure",
320 drvdata->edvidsr & EDVIDSR_E3 ? "EL3" :
321 (drvdata->edvidsr & EDVIDSR_E2 ?
322 "EL2" : "EL1/0"),
323 drvdata->edvidsr & EDVIDSR_HV ? 64 : 32,
324 drvdata->edvidsr & (u32)EDVIDSR_VMID);
325}
326
327static void debug_init_arch_data(void *info)
328{
329 struct debug_drvdata *drvdata = info;
330 u32 mode, pcsr_offset;
331 u32 eddevid, eddevid1;
332
333 CS_UNLOCK(drvdata->base);
334
335 /* Read device info */
336 eddevid = readl_relaxed(drvdata->base + EDDEVID);
337 eddevid1 = readl_relaxed(drvdata->base + EDDEVID1);
338
339 CS_LOCK(drvdata->base);
340
341 /* Parse implementation feature */
342 mode = eddevid & EDDEVID_PCSAMPLE_MODE;
343 pcsr_offset = eddevid1 & EDDEVID1_PCSR_OFFSET_MASK;
344
345 drvdata->edpcsr_present = false;
346 drvdata->edcidsr_present = false;
347 drvdata->edvidsr_present = false;
348 drvdata->pc_has_offset = false;
349
350 switch (mode) {
351 case EDDEVID_IMPL_FULL:
352 drvdata->edvidsr_present = true;
353 fallthrough;
354 case EDDEVID_IMPL_EDPCSR_EDCIDSR:
355 drvdata->edcidsr_present = true;
356 fallthrough;
357 case EDDEVID_IMPL_EDPCSR:
358 /*
359 * In ARM DDI 0487A.k, the EDDEVID1.PCSROffset is used to
360 * define if has the offset for PC sampling value; if read
361 * back EDDEVID1.PCSROffset == 0x2, then this means the debug
362 * module does not sample the instruction set state when
363 * armv8 CPU in AArch32 state.
364 */
365 drvdata->edpcsr_present =
366 ((IS_ENABLED(CONFIG_64BIT) && pcsr_offset != 0) ||
367 (pcsr_offset != EDDEVID1_PCSR_NO_OFFSET_DIS_AARCH32));
368
369 drvdata->pc_has_offset =
370 (pcsr_offset == EDDEVID1_PCSR_OFFSET_INS_SET);
371 break;
372 default:
373 break;
374 }
375}
376
377/*
378 * Dump out information on panic.
379 */
380static int debug_notifier_call(struct notifier_block *self,
381 unsigned long v, void *p)
382{
383 int cpu;
384 struct debug_drvdata *drvdata;
385
386 /* Bail out if we can't acquire the mutex or the functionality is off */
387 if (!mutex_trylock(&debug_lock))
388 return NOTIFY_DONE;
389
390 if (!debug_enable)
391 goto skip_dump;
392
393 pr_emerg("ARM external debug module:\n");
394
395 for_each_possible_cpu(cpu) {
396 drvdata = per_cpu(debug_drvdata, cpu);
397 if (!drvdata)
398 continue;
399
400 dev_emerg(drvdata->dev, "CPU[%d]:\n", drvdata->cpu);
401
402 debug_read_regs(drvdata);
403 debug_dump_regs(drvdata);
404 }
405
406skip_dump:
407 mutex_unlock(&debug_lock);
408 return NOTIFY_DONE;
409}
410
411static struct notifier_block debug_notifier = {
412 .notifier_call = debug_notifier_call,
413};
414
415static int debug_enable_func(void)
416{
417 struct debug_drvdata *drvdata;
418 int cpu, ret = 0;
419 cpumask_t mask;
420
421 /*
422 * Use cpumask to track which debug power domains have
423 * been powered on and use it to handle failure case.
424 */
425 cpumask_clear(&mask);
426
427 for_each_possible_cpu(cpu) {
428 drvdata = per_cpu(debug_drvdata, cpu);
429 if (!drvdata)
430 continue;
431
432 ret = pm_runtime_get_sync(drvdata->dev);
433 if (ret < 0)
434 goto err;
435 else
436 cpumask_set_cpu(cpu, &mask);
437 }
438
439 return 0;
440
441err:
442 /*
443 * If pm_runtime_get_sync() has failed, need rollback on
444 * all the other CPUs that have been enabled before that.
445 */
446 for_each_cpu(cpu, &mask) {
447 drvdata = per_cpu(debug_drvdata, cpu);
448 pm_runtime_put_noidle(drvdata->dev);
449 }
450
451 return ret;
452}
453
454static int debug_disable_func(void)
455{
456 struct debug_drvdata *drvdata;
457 int cpu, ret, err = 0;
458
459 /*
460 * Disable debug power domains, records the error and keep
461 * circling through all other CPUs when an error has been
462 * encountered.
463 */
464 for_each_possible_cpu(cpu) {
465 drvdata = per_cpu(debug_drvdata, cpu);
466 if (!drvdata)
467 continue;
468
469 ret = pm_runtime_put(drvdata->dev);
470 if (ret < 0)
471 err = ret;
472 }
473
474 return err;
475}
476
477static ssize_t debug_func_knob_write(struct file *f,
478 const char __user *buf, size_t count, loff_t *ppos)
479{
480 u8 val;
481 int ret;
482
483 ret = kstrtou8_from_user(buf, count, 2, &val);
484 if (ret)
485 return ret;
486
487 mutex_lock(&debug_lock);
488
489 if (val == debug_enable)
490 goto out;
491
492 if (val)
493 ret = debug_enable_func();
494 else
495 ret = debug_disable_func();
496
497 if (ret) {
498 pr_err("%s: unable to %s debug function: %d\n",
499 __func__, val ? "enable" : "disable", ret);
500 goto err;
501 }
502
503 debug_enable = val;
504out:
505 ret = count;
506err:
507 mutex_unlock(&debug_lock);
508 return ret;
509}
510
511static ssize_t debug_func_knob_read(struct file *f,
512 char __user *ubuf, size_t count, loff_t *ppos)
513{
514 ssize_t ret;
515 char buf[3];
516
517 mutex_lock(&debug_lock);
518 snprintf(buf, sizeof(buf), "%d\n", debug_enable);
519 mutex_unlock(&debug_lock);
520
521 ret = simple_read_from_buffer(ubuf, count, ppos, buf, sizeof(buf));
522 return ret;
523}
524
525static const struct file_operations debug_func_knob_fops = {
526 .open = simple_open,
527 .read = debug_func_knob_read,
528 .write = debug_func_knob_write,
529};
530
531static int debug_func_init(void)
532{
533 int ret;
534
535 /* Create debugfs node */
536 debug_debugfs_dir = debugfs_create_dir("coresight_cpu_debug", NULL);
537 debugfs_create_file("enable", 0644, debug_debugfs_dir, NULL,
538 &debug_func_knob_fops);
539
540 /* Register function to be called for panic */
541 ret = atomic_notifier_chain_register(&panic_notifier_list,
542 &debug_notifier);
543 if (ret) {
544 pr_err("%s: unable to register notifier: %d\n",
545 __func__, ret);
546 goto err;
547 }
548
549 return 0;
550
551err:
552 debugfs_remove_recursive(debug_debugfs_dir);
553 return ret;
554}
555
556static void debug_func_exit(void)
557{
558 atomic_notifier_chain_unregister(&panic_notifier_list,
559 &debug_notifier);
560 debugfs_remove_recursive(debug_debugfs_dir);
561}
562
563static int __debug_probe(struct device *dev, struct resource *res)
564{
565 struct debug_drvdata *drvdata = dev_get_drvdata(dev);
566 void __iomem *base;
567 int ret;
568
569 drvdata->cpu = coresight_get_cpu(dev);
570 if (drvdata->cpu < 0)
571 return drvdata->cpu;
572
573 if (per_cpu(debug_drvdata, drvdata->cpu)) {
574 dev_err(dev, "CPU%d drvdata has already been initialized\n",
575 drvdata->cpu);
576 return -EBUSY;
577 }
578
579 drvdata->dev = dev;
580 base = devm_ioremap_resource(dev, res);
581 if (IS_ERR(base))
582 return PTR_ERR(base);
583
584 drvdata->base = base;
585
586 cpus_read_lock();
587 per_cpu(debug_drvdata, drvdata->cpu) = drvdata;
588 ret = smp_call_function_single(drvdata->cpu, debug_init_arch_data,
589 drvdata, 1);
590 cpus_read_unlock();
591
592 if (ret) {
593 dev_err(dev, "CPU%d debug arch init failed\n", drvdata->cpu);
594 goto err;
595 }
596
597 if (!drvdata->edpcsr_present) {
598 dev_err(dev, "CPU%d sample-based profiling isn't implemented\n",
599 drvdata->cpu);
600 ret = -ENXIO;
601 goto err;
602 }
603
604 if (!debug_count++) {
605 ret = debug_func_init();
606 if (ret)
607 goto err_func_init;
608 }
609
610 mutex_lock(&debug_lock);
611 /* Turn off debug power domain if debugging is disabled */
612 if (!debug_enable)
613 pm_runtime_put(dev);
614 mutex_unlock(&debug_lock);
615
616 dev_info(dev, "Coresight debug-CPU%d initialized\n", drvdata->cpu);
617 return 0;
618
619err_func_init:
620 debug_count--;
621err:
622 per_cpu(debug_drvdata, drvdata->cpu) = NULL;
623 return ret;
624}
625
626static int debug_probe(struct amba_device *adev, const struct amba_id *id)
627{
628 struct debug_drvdata *drvdata;
629
630 drvdata = devm_kzalloc(&adev->dev, sizeof(*drvdata), GFP_KERNEL);
631 if (!drvdata)
632 return -ENOMEM;
633
634 amba_set_drvdata(adev, drvdata);
635 return __debug_probe(&adev->dev, &adev->res);
636}
637
638static void __debug_remove(struct device *dev)
639{
640 struct debug_drvdata *drvdata = dev_get_drvdata(dev);
641
642 per_cpu(debug_drvdata, drvdata->cpu) = NULL;
643
644 mutex_lock(&debug_lock);
645 /* Turn off debug power domain before rmmod the module */
646 if (debug_enable)
647 pm_runtime_put(dev);
648 mutex_unlock(&debug_lock);
649
650 if (!--debug_count)
651 debug_func_exit();
652}
653
654static void debug_remove(struct amba_device *adev)
655{
656 __debug_remove(&adev->dev);
657}
658
659static const struct amba_cs_uci_id uci_id_debug[] = {
660 {
661 /* CPU Debug UCI data */
662 .devarch = 0x47706a15,
663 .devarch_mask = 0xfff0ffff,
664 .devtype = 0x00000015,
665 }
666};
667
668static const struct amba_id debug_ids[] = {
669 CS_AMBA_ID(0x000bbd03), /* Cortex-A53 */
670 CS_AMBA_ID(0x000bbd07), /* Cortex-A57 */
671 CS_AMBA_ID(0x000bbd08), /* Cortex-A72 */
672 CS_AMBA_ID(0x000bbd09), /* Cortex-A73 */
673 CS_AMBA_UCI_ID(0x000f0205, uci_id_debug), /* Qualcomm Kryo */
674 CS_AMBA_UCI_ID(0x000f0211, uci_id_debug), /* Qualcomm Kryo */
675 {},
676};
677
678MODULE_DEVICE_TABLE(amba, debug_ids);
679
680static struct amba_driver debug_driver = {
681 .drv = {
682 .name = "coresight-cpu-debug",
683 .suppress_bind_attrs = true,
684 },
685 .probe = debug_probe,
686 .remove = debug_remove,
687 .id_table = debug_ids,
688};
689
690static int debug_platform_probe(struct platform_device *pdev)
691{
692 struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
693 struct debug_drvdata *drvdata;
694 int ret = 0;
695
696 drvdata = devm_kzalloc(&pdev->dev, sizeof(*drvdata), GFP_KERNEL);
697 if (!drvdata)
698 return -ENOMEM;
699
700 drvdata->pclk = coresight_get_enable_apb_pclk(&pdev->dev);
701 if (IS_ERR(drvdata->pclk))
702 return -ENODEV;
703
704 dev_set_drvdata(&pdev->dev, drvdata);
705 pm_runtime_get_noresume(&pdev->dev);
706 pm_runtime_set_active(&pdev->dev);
707 pm_runtime_enable(&pdev->dev);
708
709 ret = __debug_probe(&pdev->dev, res);
710 if (ret) {
711 pm_runtime_put_noidle(&pdev->dev);
712 pm_runtime_disable(&pdev->dev);
713 if (!IS_ERR_OR_NULL(drvdata->pclk))
714 clk_put(drvdata->pclk);
715 }
716 return ret;
717}
718
719static void debug_platform_remove(struct platform_device *pdev)
720{
721 struct debug_drvdata *drvdata = dev_get_drvdata(&pdev->dev);
722
723 if (WARN_ON(!drvdata))
724 return;
725
726 __debug_remove(&pdev->dev);
727 pm_runtime_disable(&pdev->dev);
728 if (!IS_ERR_OR_NULL(drvdata->pclk))
729 clk_put(drvdata->pclk);
730}
731
732#ifdef CONFIG_ACPI
733static const struct acpi_device_id debug_platform_ids[] = {
734 {"ARMHC503", 0, 0, 0}, /* ARM CoreSight Debug */
735 {},
736};
737MODULE_DEVICE_TABLE(acpi, debug_platform_ids);
738#endif
739
740#ifdef CONFIG_PM
741static int debug_runtime_suspend(struct device *dev)
742{
743 struct debug_drvdata *drvdata = dev_get_drvdata(dev);
744
745 if (drvdata && !IS_ERR_OR_NULL(drvdata->pclk))
746 clk_disable_unprepare(drvdata->pclk);
747 return 0;
748}
749
750static int debug_runtime_resume(struct device *dev)
751{
752 struct debug_drvdata *drvdata = dev_get_drvdata(dev);
753
754 if (drvdata && !IS_ERR_OR_NULL(drvdata->pclk))
755 clk_prepare_enable(drvdata->pclk);
756 return 0;
757}
758#endif
759
760static const struct dev_pm_ops debug_dev_pm_ops = {
761 SET_RUNTIME_PM_OPS(debug_runtime_suspend, debug_runtime_resume, NULL)
762};
763
764static struct platform_driver debug_platform_driver = {
765 .probe = debug_platform_probe,
766 .remove = debug_platform_remove,
767 .driver = {
768 .name = "coresight-debug-platform",
769 .acpi_match_table = ACPI_PTR(debug_platform_ids),
770 .suppress_bind_attrs = true,
771 .pm = &debug_dev_pm_ops,
772 },
773};
774
775static int __init debug_init(void)
776{
777 return coresight_init_driver("debug", &debug_driver, &debug_platform_driver);
778}
779
780static void __exit debug_exit(void)
781{
782 coresight_remove_driver(&debug_driver, &debug_platform_driver);
783}
784module_init(debug_init);
785module_exit(debug_exit);
786
787MODULE_AUTHOR("Leo Yan <leo.yan@linaro.org>");
788MODULE_DESCRIPTION("ARM Coresight CPU Debug Driver");
789MODULE_LICENSE("GPL");