Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * arch/arm64/include/asm/arch_timer.h
4 *
5 * Copyright (C) 2012 ARM Ltd.
6 * Author: Marc Zyngier <marc.zyngier@arm.com>
7 */
8#ifndef __ASM_ARCH_TIMER_H
9#define __ASM_ARCH_TIMER_H
10
11#include <asm/barrier.h>
12#include <asm/hwcap.h>
13#include <asm/sysreg.h>
14
15#include <linux/bug.h>
16#include <linux/init.h>
17#include <linux/jump_label.h>
18#include <linux/smp.h>
19#include <linux/types.h>
20
21#include <clocksource/arm_arch_timer.h>
22
23#if IS_ENABLED(CONFIG_ARM_ARCH_TIMER_OOL_WORKAROUND)
24#define has_erratum_handler(h) \
25 ({ \
26 const struct arch_timer_erratum_workaround *__wa; \
27 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
28 (__wa && __wa->h); \
29 })
30
31#define erratum_handler(h) \
32 ({ \
33 const struct arch_timer_erratum_workaround *__wa; \
34 __wa = __this_cpu_read(timer_unstable_counter_workaround); \
35 (__wa && __wa->h) ? ({ isb(); __wa->h;}) : arch_timer_##h; \
36 })
37
38#else
39#define has_erratum_handler(h) false
40#define erratum_handler(h) (arch_timer_##h)
41#endif
42
43enum arch_timer_erratum_match_type {
44 ate_match_dt,
45 ate_match_local_cap_id,
46 ate_match_acpi_oem_info,
47};
48
49struct clock_event_device;
50
51struct arch_timer_erratum_workaround {
52 enum arch_timer_erratum_match_type match_type;
53 const void *id;
54 const char *desc;
55 u64 (*read_cntpct_el0)(void);
56 u64 (*read_cntvct_el0)(void);
57 int (*set_next_event_phys)(unsigned long, struct clock_event_device *);
58 int (*set_next_event_virt)(unsigned long, struct clock_event_device *);
59 bool disable_compat_vdso;
60};
61
62DECLARE_PER_CPU(const struct arch_timer_erratum_workaround *,
63 timer_unstable_counter_workaround);
64
65static inline notrace u64 arch_timer_read_cntpct_el0(void)
66{
67 u64 cnt;
68
69 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0",
70 "nop\n" __mrs_s("%0", SYS_CNTPCTSS_EL0),
71 ARM64_HAS_ECV)
72 : "=r" (cnt));
73
74 return cnt;
75}
76
77static inline notrace u64 arch_timer_read_cntvct_el0(void)
78{
79 u64 cnt;
80
81 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0",
82 "nop\n" __mrs_s("%0", SYS_CNTVCTSS_EL0),
83 ARM64_HAS_ECV)
84 : "=r" (cnt));
85
86 return cnt;
87}
88
89#define arch_timer_reg_read_stable(reg) \
90 ({ \
91 u64 _val; \
92 \
93 preempt_disable_notrace(); \
94 _val = erratum_handler(read_ ## reg)(); \
95 preempt_enable_notrace(); \
96 \
97 _val; \
98 })
99
100/*
101 * These register accessors are marked inline so the compiler can
102 * nicely work out which register we want, and chuck away the rest of
103 * the code.
104 */
105static __always_inline
106void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u64 val)
107{
108 if (access == ARCH_TIMER_PHYS_ACCESS) {
109 switch (reg) {
110 case ARCH_TIMER_REG_CTRL:
111 write_sysreg(val, cntp_ctl_el0);
112 isb();
113 break;
114 case ARCH_TIMER_REG_CVAL:
115 write_sysreg(val, cntp_cval_el0);
116 break;
117 default:
118 BUILD_BUG();
119 }
120 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
121 switch (reg) {
122 case ARCH_TIMER_REG_CTRL:
123 write_sysreg(val, cntv_ctl_el0);
124 isb();
125 break;
126 case ARCH_TIMER_REG_CVAL:
127 write_sysreg(val, cntv_cval_el0);
128 break;
129 default:
130 BUILD_BUG();
131 }
132 } else {
133 BUILD_BUG();
134 }
135}
136
137static __always_inline
138u64 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg)
139{
140 if (access == ARCH_TIMER_PHYS_ACCESS) {
141 switch (reg) {
142 case ARCH_TIMER_REG_CTRL:
143 return read_sysreg(cntp_ctl_el0);
144 default:
145 BUILD_BUG();
146 }
147 } else if (access == ARCH_TIMER_VIRT_ACCESS) {
148 switch (reg) {
149 case ARCH_TIMER_REG_CTRL:
150 return read_sysreg(cntv_ctl_el0);
151 default:
152 BUILD_BUG();
153 }
154 }
155
156 BUILD_BUG();
157 unreachable();
158}
159
160static inline u32 arch_timer_get_cntfrq(void)
161{
162 return read_sysreg(cntfrq_el0);
163}
164
165static inline u32 arch_timer_get_cntkctl(void)
166{
167 return read_sysreg(cntkctl_el1);
168}
169
170static inline void arch_timer_set_cntkctl(u32 cntkctl)
171{
172 write_sysreg(cntkctl, cntkctl_el1);
173 isb();
174}
175
176static __always_inline u64 __arch_counter_get_cntpct_stable(void)
177{
178 u64 cnt;
179
180 cnt = arch_timer_reg_read_stable(cntpct_el0);
181 arch_counter_enforce_ordering(cnt);
182 return cnt;
183}
184
185static __always_inline u64 __arch_counter_get_cntpct(void)
186{
187 u64 cnt;
188
189 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0",
190 "nop\n" __mrs_s("%0", SYS_CNTPCTSS_EL0),
191 ARM64_HAS_ECV)
192 : "=r" (cnt));
193 arch_counter_enforce_ordering(cnt);
194 return cnt;
195}
196
197static __always_inline u64 __arch_counter_get_cntvct_stable(void)
198{
199 u64 cnt;
200
201 cnt = arch_timer_reg_read_stable(cntvct_el0);
202 arch_counter_enforce_ordering(cnt);
203 return cnt;
204}
205
206static __always_inline u64 __arch_counter_get_cntvct(void)
207{
208 u64 cnt;
209
210 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0",
211 "nop\n" __mrs_s("%0", SYS_CNTVCTSS_EL0),
212 ARM64_HAS_ECV)
213 : "=r" (cnt));
214 arch_counter_enforce_ordering(cnt);
215 return cnt;
216}
217
218static inline int arch_timer_arch_init(void)
219{
220 return 0;
221}
222
223static inline void arch_timer_set_evtstrm_feature(void)
224{
225 cpu_set_named_feature(EVTSTRM);
226#ifdef CONFIG_COMPAT
227 compat_elf_hwcap |= COMPAT_HWCAP_EVTSTRM;
228#endif
229}
230
231static inline bool arch_timer_have_evtstrm_feature(void)
232{
233 return cpu_have_named_feature(EVTSTRM);
234}
235#endif