Linux Audio

Check our new training course

Loading...
v6.13.7
  1/* SPDX-License-Identifier: GPL-2.0 */
  2#ifndef __ASMARM_ARCH_TIMER_H
  3#define __ASMARM_ARCH_TIMER_H
  4
  5#include <asm/barrier.h>
  6#include <asm/errno.h>
  7#include <asm/hwcap.h>
  8#include <linux/clocksource.h>
  9#include <linux/init.h>
 10#include <linux/io-64-nonatomic-lo-hi.h>
 11#include <linux/types.h>
 12
 13#include <clocksource/arm_arch_timer.h>
 14
 15#ifdef CONFIG_ARM_ARCH_TIMER
 16/* 32bit ARM doesn't know anything about timer errata... */
 17#define has_erratum_handler(h)		(false)
 18#define erratum_handler(h)		(arch_timer_##h)
 19
 20int arch_timer_arch_init(void);
 21
 22/*
 23 * These register accessors are marked inline so the compiler can
 24 * nicely work out which register we want, and chuck away the rest of
 25 * the code. At least it does so with a recent GCC (4.6.3).
 26 */
 27static __always_inline
 28void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u64 val)
 29{
 30	if (access == ARCH_TIMER_PHYS_ACCESS) {
 31		switch (reg) {
 32		case ARCH_TIMER_REG_CTRL:
 33			asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" ((u32)val));
 34			isb();
 35			break;
 36		case ARCH_TIMER_REG_CVAL:
 37			asm volatile("mcrr p15, 2, %Q0, %R0, c14" : : "r" (val));
 38			break;
 39		default:
 40			BUILD_BUG();
 41		}
 42	} else if (access == ARCH_TIMER_VIRT_ACCESS) {
 43		switch (reg) {
 44		case ARCH_TIMER_REG_CTRL:
 45			asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" ((u32)val));
 46			isb();
 47			break;
 48		case ARCH_TIMER_REG_CVAL:
 49			asm volatile("mcrr p15, 3, %Q0, %R0, c14" : : "r" (val));
 50			break;
 51		default:
 52			BUILD_BUG();
 53		}
 54	} else {
 55		BUILD_BUG();
 56	}
 
 
 57}
 58
 59static __always_inline
 60u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
 61{
 62	u32 val = 0;
 63
 64	if (access == ARCH_TIMER_PHYS_ACCESS) {
 65		switch (reg) {
 66		case ARCH_TIMER_REG_CTRL:
 67			asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val));
 68			break;
 69		default:
 70			BUILD_BUG();
 
 71		}
 72	} else if (access == ARCH_TIMER_VIRT_ACCESS) {
 73		switch (reg) {
 74		case ARCH_TIMER_REG_CTRL:
 75			asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val));
 76			break;
 77		default:
 78			BUILD_BUG();
 
 79		}
 80	} else {
 81		BUILD_BUG();
 82	}
 83
 84	return val;
 85}
 86
 87static inline u32 arch_timer_get_cntfrq(void)
 88{
 89	u32 val;
 90	asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val));
 91	return val;
 92}
 93
 94static inline u64 __arch_counter_get_cntpct(void)
 95{
 96	u64 cval;
 97
 98	isb();
 99	asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
100	return cval;
101}
102
103static inline u64 __arch_counter_get_cntpct_stable(void)
104{
105	return __arch_counter_get_cntpct();
106}
107
108static inline u64 __arch_counter_get_cntvct(void)
109{
110	u64 cval;
111
112	isb();
113	asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval));
114	return cval;
115}
116
117static inline u64 __arch_counter_get_cntvct_stable(void)
118{
119	return __arch_counter_get_cntvct();
120}
121
122static inline u32 arch_timer_get_cntkctl(void)
123{
124	u32 cntkctl;
125	asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl));
126	return cntkctl;
127}
128
129static inline void arch_timer_set_cntkctl(u32 cntkctl)
130{
131	asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl));
132	isb();
133}
134
135static inline void arch_timer_set_evtstrm_feature(void)
136{
137	elf_hwcap |= HWCAP_EVTSTRM;
138}
139
140static inline bool arch_timer_have_evtstrm_feature(void)
141{
142	return elf_hwcap & HWCAP_EVTSTRM;
143}
144#endif
145
146#endif
v5.14.15
  1/* SPDX-License-Identifier: GPL-2.0 */
  2#ifndef __ASMARM_ARCH_TIMER_H
  3#define __ASMARM_ARCH_TIMER_H
  4
  5#include <asm/barrier.h>
  6#include <asm/errno.h>
  7#include <asm/hwcap.h>
  8#include <linux/clocksource.h>
  9#include <linux/init.h>
 
 10#include <linux/types.h>
 11
 12#include <clocksource/arm_arch_timer.h>
 13
 14#ifdef CONFIG_ARM_ARCH_TIMER
 15/* 32bit ARM doesn't know anything about timer errata... */
 16#define has_erratum_handler(h)		(false)
 17#define erratum_handler(h)		(arch_timer_##h)
 18
 19int arch_timer_arch_init(void);
 20
 21/*
 22 * These register accessors are marked inline so the compiler can
 23 * nicely work out which register we want, and chuck away the rest of
 24 * the code. At least it does so with a recent GCC (4.6.3).
 25 */
 26static __always_inline
 27void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u32 val)
 28{
 29	if (access == ARCH_TIMER_PHYS_ACCESS) {
 30		switch (reg) {
 31		case ARCH_TIMER_REG_CTRL:
 32			asm volatile("mcr p15, 0, %0, c14, c2, 1" : : "r" (val));
 
 33			break;
 34		case ARCH_TIMER_REG_TVAL:
 35			asm volatile("mcr p15, 0, %0, c14, c2, 0" : : "r" (val));
 36			break;
 
 
 37		}
 38	} else if (access == ARCH_TIMER_VIRT_ACCESS) {
 39		switch (reg) {
 40		case ARCH_TIMER_REG_CTRL:
 41			asm volatile("mcr p15, 0, %0, c14, c3, 1" : : "r" (val));
 
 42			break;
 43		case ARCH_TIMER_REG_TVAL:
 44			asm volatile("mcr p15, 0, %0, c14, c3, 0" : : "r" (val));
 45			break;
 
 
 46		}
 
 
 47	}
 48
 49	isb();
 50}
 51
 52static __always_inline
 53u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
 54{
 55	u32 val = 0;
 56
 57	if (access == ARCH_TIMER_PHYS_ACCESS) {
 58		switch (reg) {
 59		case ARCH_TIMER_REG_CTRL:
 60			asm volatile("mrc p15, 0, %0, c14, c2, 1" : "=r" (val));
 61			break;
 62		case ARCH_TIMER_REG_TVAL:
 63			asm volatile("mrc p15, 0, %0, c14, c2, 0" : "=r" (val));
 64			break;
 65		}
 66	} else if (access == ARCH_TIMER_VIRT_ACCESS) {
 67		switch (reg) {
 68		case ARCH_TIMER_REG_CTRL:
 69			asm volatile("mrc p15, 0, %0, c14, c3, 1" : "=r" (val));
 70			break;
 71		case ARCH_TIMER_REG_TVAL:
 72			asm volatile("mrc p15, 0, %0, c14, c3, 0" : "=r" (val));
 73			break;
 74		}
 
 
 75	}
 76
 77	return val;
 78}
 79
 80static inline u32 arch_timer_get_cntfrq(void)
 81{
 82	u32 val;
 83	asm volatile("mrc p15, 0, %0, c14, c0, 0" : "=r" (val));
 84	return val;
 85}
 86
 87static inline u64 __arch_counter_get_cntpct(void)
 88{
 89	u64 cval;
 90
 91	isb();
 92	asm volatile("mrrc p15, 0, %Q0, %R0, c14" : "=r" (cval));
 93	return cval;
 94}
 95
 96static inline u64 __arch_counter_get_cntpct_stable(void)
 97{
 98	return __arch_counter_get_cntpct();
 99}
100
101static inline u64 __arch_counter_get_cntvct(void)
102{
103	u64 cval;
104
105	isb();
106	asm volatile("mrrc p15, 1, %Q0, %R0, c14" : "=r" (cval));
107	return cval;
108}
109
110static inline u64 __arch_counter_get_cntvct_stable(void)
111{
112	return __arch_counter_get_cntvct();
113}
114
115static inline u32 arch_timer_get_cntkctl(void)
116{
117	u32 cntkctl;
118	asm volatile("mrc p15, 0, %0, c14, c1, 0" : "=r" (cntkctl));
119	return cntkctl;
120}
121
122static inline void arch_timer_set_cntkctl(u32 cntkctl)
123{
124	asm volatile("mcr p15, 0, %0, c14, c1, 0" : : "r" (cntkctl));
125	isb();
126}
127
128static inline void arch_timer_set_evtstrm_feature(void)
129{
130	elf_hwcap |= HWCAP_EVTSTRM;
131}
132
133static inline bool arch_timer_have_evtstrm_feature(void)
134{
135	return elf_hwcap & HWCAP_EVTSTRM;
136}
137#endif
138
139#endif