Linux Audio

Check our new training course

Loading...
v5.9
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright (C) 2012 Regents of the University of California
  4 * Copyright (C) 2017 SiFive
  5 *
  6 * All RISC-V systems have a timer attached to every hart.  These timers can
  7 * either be read from the "time" and "timeh" CSRs, and can use the SBI to
  8 * setup events, or directly accessed using MMIO registers.
  9 */
 
 
 
 
 10#include <linux/clocksource.h>
 11#include <linux/clockchips.h>
 12#include <linux/cpu.h>
 13#include <linux/delay.h>
 14#include <linux/irq.h>
 15#include <linux/irqdomain.h>
 
 16#include <linux/sched_clock.h>
 17#include <linux/io-64-nonatomic-lo-hi.h>
 18#include <linux/interrupt.h>
 19#include <linux/of_irq.h>
 
 
 20#include <asm/smp.h>
 
 21#include <asm/sbi.h>
 22#include <asm/timex.h>
 23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 24static int riscv_clock_next_event(unsigned long delta,
 25		struct clock_event_device *ce)
 26{
 27	csr_set(CSR_IE, IE_TIE);
 28	sbi_set_timer(get_cycles64() + delta);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 29	return 0;
 30}
 31
 32static unsigned int riscv_clock_event_irq;
 33static DEFINE_PER_CPU(struct clock_event_device, riscv_clock_event) = {
 34	.name			= "riscv_timer_clockevent",
 35	.features		= CLOCK_EVT_FEAT_ONESHOT,
 36	.rating			= 100,
 37	.set_next_event		= riscv_clock_next_event,
 
 38};
 39
 40/*
 41 * It is guaranteed that all the timers across all the harts are synchronized
 42 * within one tick of each other, so while this could technically go
 43 * backwards when hopping between CPUs, practically it won't happen.
 44 */
 45static unsigned long long riscv_clocksource_rdtime(struct clocksource *cs)
 46{
 47	return get_cycles64();
 48}
 49
 50static u64 notrace riscv_sched_clock(void)
 51{
 52	return get_cycles64();
 53}
 54
 55static struct clocksource riscv_clocksource = {
 56	.name		= "riscv_clocksource",
 57	.rating		= 300,
 58	.mask		= CLOCKSOURCE_MASK(64),
 59	.flags		= CLOCK_SOURCE_IS_CONTINUOUS,
 60	.read		= riscv_clocksource_rdtime,
 
 
 
 
 
 61};
 62
 63static int riscv_timer_starting_cpu(unsigned int cpu)
 64{
 65	struct clock_event_device *ce = per_cpu_ptr(&riscv_clock_event, cpu);
 66
 67	ce->cpumask = cpumask_of(cpu);
 68	ce->irq = riscv_clock_event_irq;
 
 
 
 
 69	clockevents_config_and_register(ce, riscv_timebase, 100, 0x7fffffff);
 70
 71	enable_percpu_irq(riscv_clock_event_irq,
 72			  irq_get_trigger_type(riscv_clock_event_irq));
 73	return 0;
 74}
 75
 76static int riscv_timer_dying_cpu(unsigned int cpu)
 77{
 78	disable_percpu_irq(riscv_clock_event_irq);
 79	return 0;
 80}
 81
 
 
 
 
 
 
 
 82/* called directly from the low-level interrupt handler */
 83static irqreturn_t riscv_timer_interrupt(int irq, void *dev_id)
 84{
 85	struct clock_event_device *evdev = this_cpu_ptr(&riscv_clock_event);
 86
 87	csr_clear(CSR_IE, IE_TIE);
 88	evdev->event_handler(evdev);
 89
 90	return IRQ_HANDLED;
 91}
 92
 93static int __init riscv_timer_init_dt(struct device_node *n)
 94{
 95	int cpuid, hartid, error;
 96	struct device_node *child;
 97	struct irq_domain *domain;
 
 98
 99	hartid = riscv_of_processor_hartid(n);
100	if (hartid < 0) {
101		pr_warn("Not valid hartid for node [%pOF] error = [%d]\n",
102			n, hartid);
103		return hartid;
104	}
105
106	cpuid = riscv_hartid_to_cpuid(hartid);
107	if (cpuid < 0) {
108		pr_warn("Invalid cpuid for hartid [%d]\n", hartid);
109		return cpuid;
110	}
111
112	if (cpuid != smp_processor_id())
113		return 0;
114
115	domain = NULL;
116	child = of_get_compatible_child(n, "riscv,cpu-intc");
117	if (!child) {
118		pr_err("Failed to find INTC node [%pOF]\n", n);
119		return -ENODEV;
120	}
121	domain = irq_find_host(child);
122	of_node_put(child);
123	if (!domain) {
124		pr_err("Failed to find IRQ domain for node [%pOF]\n", n);
 
125		return -ENODEV;
126	}
127
128	riscv_clock_event_irq = irq_create_mapping(domain, RV_IRQ_TIMER);
129	if (!riscv_clock_event_irq) {
130		pr_err("Failed to map timer interrupt for node [%pOF]\n", n);
131		return -ENODEV;
132	}
133
134	pr_info("%s: Registering clocksource cpuid [%d] hartid [%d]\n",
135	       __func__, cpuid, hartid);
136	error = clocksource_register_hz(&riscv_clocksource, riscv_timebase);
137	if (error) {
138		pr_err("RISCV timer register failed [%d] for cpu = [%d]\n",
139		       error, cpuid);
140		return error;
141	}
142
143	sched_clock_register(riscv_sched_clock, 64, riscv_timebase);
144
145	error = request_percpu_irq(riscv_clock_event_irq,
146				    riscv_timer_interrupt,
147				    "riscv-timer", &riscv_clock_event);
148	if (error) {
149		pr_err("registering percpu irq failed [%d]\n", error);
150		return error;
151	}
152
 
 
 
 
 
153	error = cpuhp_setup_state(CPUHP_AP_RISCV_TIMER_STARTING,
154			 "clockevents/riscv/timer:starting",
155			 riscv_timer_starting_cpu, riscv_timer_dying_cpu);
156	if (error)
157		pr_err("cpu hp setup state failed for RISCV timer [%d]\n",
158		       error);
 
159	return error;
160}
161
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162TIMER_OF_DECLARE(riscv_timer, "riscv", riscv_timer_init_dt);
v6.8
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright (C) 2012 Regents of the University of California
  4 * Copyright (C) 2017 SiFive
  5 *
  6 * All RISC-V systems have a timer attached to every hart.  These timers can
  7 * either be read from the "time" and "timeh" CSRs, and can use the SBI to
  8 * setup events, or directly accessed using MMIO registers.
  9 */
 10
 11#define pr_fmt(fmt) "riscv-timer: " fmt
 12
 13#include <linux/acpi.h>
 14#include <linux/clocksource.h>
 15#include <linux/clockchips.h>
 16#include <linux/cpu.h>
 17#include <linux/delay.h>
 18#include <linux/irq.h>
 19#include <linux/irqdomain.h>
 20#include <linux/module.h>
 21#include <linux/sched_clock.h>
 22#include <linux/io-64-nonatomic-lo-hi.h>
 23#include <linux/interrupt.h>
 24#include <linux/of_irq.h>
 25#include <linux/limits.h>
 26#include <clocksource/timer-riscv.h>
 27#include <asm/smp.h>
 28#include <asm/cpufeature.h>
 29#include <asm/sbi.h>
 30#include <asm/timex.h>
 31
 32static DEFINE_STATIC_KEY_FALSE(riscv_sstc_available);
 33static bool riscv_timer_cannot_wake_cpu;
 34
 35static void riscv_clock_event_stop(void)
 36{
 37	if (static_branch_likely(&riscv_sstc_available)) {
 38		csr_write(CSR_STIMECMP, ULONG_MAX);
 39		if (IS_ENABLED(CONFIG_32BIT))
 40			csr_write(CSR_STIMECMPH, ULONG_MAX);
 41	} else {
 42		sbi_set_timer(U64_MAX);
 43	}
 44}
 45
 46static int riscv_clock_next_event(unsigned long delta,
 47		struct clock_event_device *ce)
 48{
 49	u64 next_tval = get_cycles64() + delta;
 50
 51	if (static_branch_likely(&riscv_sstc_available)) {
 52#if defined(CONFIG_32BIT)
 53		csr_write(CSR_STIMECMP, next_tval & 0xFFFFFFFF);
 54		csr_write(CSR_STIMECMPH, next_tval >> 32);
 55#else
 56		csr_write(CSR_STIMECMP, next_tval);
 57#endif
 58	} else
 59		sbi_set_timer(next_tval);
 60
 61	return 0;
 62}
 63
 64static int riscv_clock_shutdown(struct clock_event_device *evt)
 65{
 66	riscv_clock_event_stop();
 67	return 0;
 68}
 69
 70static unsigned int riscv_clock_event_irq;
 71static DEFINE_PER_CPU(struct clock_event_device, riscv_clock_event) = {
 72	.name			= "riscv_timer_clockevent",
 73	.features		= CLOCK_EVT_FEAT_ONESHOT,
 74	.rating			= 100,
 75	.set_next_event		= riscv_clock_next_event,
 76	.set_state_shutdown	= riscv_clock_shutdown,
 77};
 78
 79/*
 80 * It is guaranteed that all the timers across all the harts are synchronized
 81 * within one tick of each other, so while this could technically go
 82 * backwards when hopping between CPUs, practically it won't happen.
 83 */
 84static unsigned long long riscv_clocksource_rdtime(struct clocksource *cs)
 85{
 86	return get_cycles64();
 87}
 88
 89static u64 notrace riscv_sched_clock(void)
 90{
 91	return get_cycles64();
 92}
 93
 94static struct clocksource riscv_clocksource = {
 95	.name		= "riscv_clocksource",
 96	.rating		= 400,
 97	.mask		= CLOCKSOURCE_MASK(64),
 98	.flags		= CLOCK_SOURCE_IS_CONTINUOUS,
 99	.read		= riscv_clocksource_rdtime,
100#if IS_ENABLED(CONFIG_GENERIC_GETTIMEOFDAY)
101	.vdso_clock_mode = VDSO_CLOCKMODE_ARCHTIMER,
102#else
103	.vdso_clock_mode = VDSO_CLOCKMODE_NONE,
104#endif
105};
106
107static int riscv_timer_starting_cpu(unsigned int cpu)
108{
109	struct clock_event_device *ce = per_cpu_ptr(&riscv_clock_event, cpu);
110
111	ce->cpumask = cpumask_of(cpu);
112	ce->irq = riscv_clock_event_irq;
113	if (riscv_timer_cannot_wake_cpu)
114		ce->features |= CLOCK_EVT_FEAT_C3STOP;
115	if (static_branch_likely(&riscv_sstc_available))
116		ce->rating = 450;
117	clockevents_config_and_register(ce, riscv_timebase, 100, 0x7fffffff);
118
119	enable_percpu_irq(riscv_clock_event_irq,
120			  irq_get_trigger_type(riscv_clock_event_irq));
121	return 0;
122}
123
124static int riscv_timer_dying_cpu(unsigned int cpu)
125{
126	disable_percpu_irq(riscv_clock_event_irq);
127	return 0;
128}
129
130void riscv_cs_get_mult_shift(u32 *mult, u32 *shift)
131{
132	*mult = riscv_clocksource.mult;
133	*shift = riscv_clocksource.shift;
134}
135EXPORT_SYMBOL_GPL(riscv_cs_get_mult_shift);
136
137/* called directly from the low-level interrupt handler */
138static irqreturn_t riscv_timer_interrupt(int irq, void *dev_id)
139{
140	struct clock_event_device *evdev = this_cpu_ptr(&riscv_clock_event);
141
142	riscv_clock_event_stop();
143	evdev->event_handler(evdev);
144
145	return IRQ_HANDLED;
146}
147
148static int __init riscv_timer_init_common(void)
149{
150	int error;
 
151	struct irq_domain *domain;
152	struct fwnode_handle *intc_fwnode = riscv_get_intc_hwnode();
153
154	domain = irq_find_matching_fwnode(intc_fwnode, DOMAIN_BUS_ANY);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
155	if (!domain) {
156		pr_err("Failed to find irq_domain for INTC node [%pfwP]\n",
157		       intc_fwnode);
158		return -ENODEV;
159	}
160
161	riscv_clock_event_irq = irq_create_mapping(domain, RV_IRQ_TIMER);
162	if (!riscv_clock_event_irq) {
163		pr_err("Failed to map timer interrupt for node [%pfwP]\n", intc_fwnode);
164		return -ENODEV;
165	}
166
 
 
167	error = clocksource_register_hz(&riscv_clocksource, riscv_timebase);
168	if (error) {
169		pr_err("RISCV timer registration failed [%d]\n", error);
 
170		return error;
171	}
172
173	sched_clock_register(riscv_sched_clock, 64, riscv_timebase);
174
175	error = request_percpu_irq(riscv_clock_event_irq,
176				    riscv_timer_interrupt,
177				    "riscv-timer", &riscv_clock_event);
178	if (error) {
179		pr_err("registering percpu irq failed [%d]\n", error);
180		return error;
181	}
182
183	if (riscv_isa_extension_available(NULL, SSTC)) {
184		pr_info("Timer interrupt in S-mode is available via sstc extension\n");
185		static_branch_enable(&riscv_sstc_available);
186	}
187
188	error = cpuhp_setup_state(CPUHP_AP_RISCV_TIMER_STARTING,
189			 "clockevents/riscv/timer:starting",
190			 riscv_timer_starting_cpu, riscv_timer_dying_cpu);
191	if (error)
192		pr_err("cpu hp setup state failed for RISCV timer [%d]\n",
193		       error);
194
195	return error;
196}
197
198static int __init riscv_timer_init_dt(struct device_node *n)
199{
200	int cpuid, error;
201	unsigned long hartid;
202	struct device_node *child;
203
204	error = riscv_of_processor_hartid(n, &hartid);
205	if (error < 0) {
206		pr_warn("Invalid hartid for node [%pOF] error = [%lu]\n",
207			n, hartid);
208		return error;
209	}
210
211	cpuid = riscv_hartid_to_cpuid(hartid);
212	if (cpuid < 0) {
213		pr_warn("Invalid cpuid for hartid [%lu]\n", hartid);
214		return cpuid;
215	}
216
217	if (cpuid != smp_processor_id())
218		return 0;
219
220	child = of_find_compatible_node(NULL, NULL, "riscv,timer");
221	if (child) {
222		riscv_timer_cannot_wake_cpu = of_property_read_bool(child,
223					"riscv,timer-cannot-wake-cpu");
224		of_node_put(child);
225	}
226
227	return riscv_timer_init_common();
228}
229
230TIMER_OF_DECLARE(riscv_timer, "riscv", riscv_timer_init_dt);
231
232#ifdef CONFIG_ACPI
233static int __init riscv_timer_acpi_init(struct acpi_table_header *table)
234{
235	struct acpi_table_rhct *rhct = (struct acpi_table_rhct *)table;
236
237	riscv_timer_cannot_wake_cpu = rhct->flags & ACPI_RHCT_TIMER_CANNOT_WAKEUP_CPU;
238
239	return riscv_timer_init_common();
240}
241
242TIMER_ACPI_DECLARE(aclint_mtimer, ACPI_SIG_RHCT, riscv_timer_acpi_init);
243
244#endif