Loading...
1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef __ASMARM_ARCH_TIMER_H
3#define __ASMARM_ARCH_TIMER_H
4
5#include <asm/barrier.h>
6#include <asm/errno.h>
7#include <asm/hwcap.h>
8#include <linux/clocksource.h>
9#include <linux/init.h>
10#include <linux/io-64-nonatomic-lo-hi.h>
11#include <linux/types.h>
12
13#include <clocksource/arm_arch_timer.h>
14
15#ifdef CONFIG_ARM_ARCH_TIMER
16/* 32bit ARM doesn't know anything about timer errata... */
17#define has_erratum_handler(h) (false)
18#define erratum_handler(h) (arch_timer_##h)
19
20int arch_timer_arch_init(void);
21
22/*
23 * These register accessors are marked inline so the compiler can
24 * nicely work out which register we want, and chuck away the rest of
25 * the code. At least it does so with a recent GCC (4.6.3).
26 */
27static __always_inline
28void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u64 val)
29{
30 if (access == ARCH_TIMER_PHYS_ACCESS) {
31 switch (reg) {
32 case ARCH_TIMER_REG_CTRL:
33 asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" ((u32)val));
34 isb();
35 break;
36 case ARCH_TIMER_REG_CVAL:
37 asm volatile("mcrr p15, 2, %Q0, %R0, c14" : : "r" (val));
38 break;
39 default:
40 BUILD_BUG();
41 }
42 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
43 switch (reg) {
44 case ARCH_TIMER_REG_CTRL:
45 asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" ((u32)val));
46 isb();
47 break;
48 case ARCH_TIMER_REG_CVAL:
49 asm volatile("mcrr p15, 3, %Q0, %R0, c14" : : "r" (val));
50 break;
51 default:
52 BUILD_BUG();
53 }
54 } else {
55 BUILD_BUG();
56 }
57}
58
59static __always_inline
60u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
61{
62 u32 val = 0;
63
64 if (access == ARCH_TIMER_PHYS_ACCESS) {
65 switch (reg) {
66 case ARCH_TIMER_REG_CTRL:
67 asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val));
68 break;
69 default:
70 BUILD_BUG();
71 }
72 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
73 switch (reg) {
74 case ARCH_TIMER_REG_CTRL:
75 asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val));
76 break;
77 default:
78 BUILD_BUG();
79 }
80 } else {
81 BUILD_BUG();
82 }
83
84 return val;
85}
86
87static inline u32 arch_timer_get_cntfrq(void)
88{
89 u32 val;
90 asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val));
91 return val;
92}
93
94static inline u64 __arch_counter_get_cntpct(void)
95{
96 u64 cval;
97
98 isb();
99 asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
100 return cval;
101}
102
103static inline u64 __arch_counter_get_cntpct_stable(void)
104{
105 return __arch_counter_get_cntpct();
106}
107
108static inline u64 __arch_counter_get_cntvct(void)
109{
110 u64 cval;
111
112 isb();
113 asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval));
114 return cval;
115}
116
117static inline u64 __arch_counter_get_cntvct_stable(void)
118{
119 return __arch_counter_get_cntvct();
120}
121
122static inline u32 arch_timer_get_cntkctl(void)
123{
124 u32 cntkctl;
125 asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl));
126 return cntkctl;
127}
128
129static inline void arch_timer_set_cntkctl(u32 cntkctl)
130{
131 asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl));
132 isb();
133}
134
135static inline void arch_timer_set_evtstrm_feature(void)
136{
137 elf_hwcap |= HWCAP_EVTSTRM;
138}
139
140static inline bool arch_timer_have_evtstrm_feature(void)
141{
142 return elf_hwcap & HWCAP_EVTSTRM;
143}
144#endif
145
146#endif
1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef __ASMARM_ARCH_TIMER_H
3#define __ASMARM_ARCH_TIMER_H
4
5#include <asm/barrier.h>
6#include <asm/errno.h>
7#include <asm/hwcap.h>
8#include <linux/clocksource.h>
9#include <linux/init.h>
10#include <linux/types.h>
11
12#include <clocksource/arm_arch_timer.h>
13
14#ifdef CONFIG_ARM_ARCH_TIMER
15/* 32bit ARM doesn't know anything about timer errata... */
16#define has_erratum_handler(h) (false)
17#define erratum_handler(h) (arch_timer_##h)
18
19int arch_timer_arch_init(void);
20
21/*
22 * These register accessors are marked inline so the compiler can
23 * nicely work out which register we want, and chuck away the rest of
24 * the code. At least it does so with a recent GCC (4.6.3).
25 */
26static __always_inline
27void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u32 val)
28{
29 if (access == ARCH_TIMER_PHYS_ACCESS) {
30 switch (reg) {
31 case ARCH_TIMER_REG_CTRL:
32 asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" (val));
33 break;
34 case ARCH_TIMER_REG_TVAL:
35 asm volatile("mcr p15, 0, %0, c14, c2, 0" : : "r" (val));
36 break;
37 }
38 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
39 switch (reg) {
40 case ARCH_TIMER_REG_CTRL:
41 asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" (val));
42 break;
43 case ARCH_TIMER_REG_TVAL:
44 asm volatile("mcr p15, 0, %0, c14, c3, 0" : : "r" (val));
45 break;
46 }
47 }
48
49 isb();
50}
51
52static __always_inline
53u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
54{
55 u32 val = 0;
56
57 if (access == ARCH_TIMER_PHYS_ACCESS) {
58 switch (reg) {
59 case ARCH_TIMER_REG_CTRL:
60 asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val));
61 break;
62 case ARCH_TIMER_REG_TVAL:
63 asm volatile("mrc p15, 0, %0, c14, c2, 0" : "=r" (val));
64 break;
65 }
66 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
67 switch (reg) {
68 case ARCH_TIMER_REG_CTRL:
69 asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val));
70 break;
71 case ARCH_TIMER_REG_TVAL:
72 asm volatile("mrc p15, 0, %0, c14, c3, 0" : "=r" (val));
73 break;
74 }
75 }
76
77 return val;
78}
79
80static inline u32 arch_timer_get_cntfrq(void)
81{
82 u32 val;
83 asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val));
84 return val;
85}
86
87static inline u64 __arch_counter_get_cntpct(void)
88{
89 u64 cval;
90
91 isb();
92 asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
93 return cval;
94}
95
96static inline u64 __arch_counter_get_cntpct_stable(void)
97{
98 return __arch_counter_get_cntpct();
99}
100
101static inline u64 __arch_counter_get_cntvct(void)
102{
103 u64 cval;
104
105 isb();
106 asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval));
107 return cval;
108}
109
110static inline u64 __arch_counter_get_cntvct_stable(void)
111{
112 return __arch_counter_get_cntvct();
113}
114
115static inline u32 arch_timer_get_cntkctl(void)
116{
117 u32 cntkctl;
118 asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl));
119 return cntkctl;
120}
121
122static inline void arch_timer_set_cntkctl(u32 cntkctl)
123{
124 asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl));
125 isb();
126}
127
128static inline void arch_timer_set_evtstrm_feature(void)
129{
130 elf_hwcap |= HWCAP_EVTSTRM;
131}
132
133static inline bool arch_timer_have_evtstrm_feature(void)
134{
135 return elf_hwcap & HWCAP_EVTSTRM;
136}
137#endif
138
139#endif