Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
fork
Configure Feed
Select the types of activity you want to include in your feed.
1/* SPDX-License-Identifier: GPL-2.0 */
2
3#ifndef _ASM_X86_CPU_ENTRY_AREA_H
4#define _ASM_X86_CPU_ENTRY_AREA_H
5
6#include <linux/percpu-defs.h>
7#include <asm/processor.h>
8#include <asm/intel_ds.h>
9#include <asm/pgtable_areas.h>
10
11#ifdef CONFIG_X86_64
12
13/* Macro to enforce the same ordering and stack sizes */
14#define ESTACKS_MEMBERS(guardsize, optional_stack_size) \
15 char DF_stack_guard[guardsize]; \
16 char DF_stack[EXCEPTION_STKSZ]; \
17 char NMI_stack_guard[guardsize]; \
18 char NMI_stack[EXCEPTION_STKSZ]; \
19 char DB_stack_guard[guardsize]; \
20 char DB_stack[EXCEPTION_STKSZ]; \
21 char MCE_stack_guard[guardsize]; \
22 char MCE_stack[EXCEPTION_STKSZ]; \
23 char VC_stack_guard[guardsize]; \
24 char VC_stack[optional_stack_size]; \
25 char VC2_stack_guard[guardsize]; \
26 char VC2_stack[optional_stack_size]; \
27 char IST_top_guard[guardsize]; \
28
29/* The exception stacks' physical storage. No guard pages required */
30struct exception_stacks {
31 ESTACKS_MEMBERS(0, 0)
32};
33
34/* The effective cpu entry area mapping with guard pages. */
35struct cea_exception_stacks {
36 ESTACKS_MEMBERS(PAGE_SIZE, EXCEPTION_STKSZ)
37};
38
39/*
40 * The exception stack ordering in [cea_]exception_stacks
41 */
42enum exception_stack_ordering {
43 ESTACK_DF,
44 ESTACK_NMI,
45 ESTACK_DB,
46 ESTACK_MCE,
47 ESTACK_VC,
48 ESTACK_VC2,
49 N_EXCEPTION_STACKS
50};
51
52#define CEA_ESTACK_SIZE(st) \
53 sizeof(((struct cea_exception_stacks *)0)->st## _stack)
54
55#define CEA_ESTACK_BOT(ceastp, st) \
56 ((unsigned long)&(ceastp)->st## _stack)
57
58#define CEA_ESTACK_TOP(ceastp, st) \
59 (CEA_ESTACK_BOT(ceastp, st) + CEA_ESTACK_SIZE(st))
60
61#define CEA_ESTACK_OFFS(st) \
62 offsetof(struct cea_exception_stacks, st## _stack)
63
64#define CEA_ESTACK_PAGES \
65 (sizeof(struct cea_exception_stacks) / PAGE_SIZE)
66
67#endif
68
69#ifdef CONFIG_X86_32
70struct doublefault_stack {
71 unsigned long stack[(PAGE_SIZE - sizeof(struct x86_hw_tss)) / sizeof(unsigned long)];
72 struct x86_hw_tss tss;
73} __aligned(PAGE_SIZE);
74#endif
75
76/*
77 * cpu_entry_area is a percpu region that contains things needed by the CPU
78 * and early entry/exit code. Real types aren't used for all fields here
79 * to avoid circular header dependencies.
80 *
81 * Every field is a virtual alias of some other allocated backing store.
82 * There is no direct allocation of a struct cpu_entry_area.
83 */
84struct cpu_entry_area {
85 char gdt[PAGE_SIZE];
86
87 /*
88 * The GDT is just below entry_stack and thus serves (on x86_64) as
89 * a read-only guard page. On 32-bit the GDT must be writeable, so
90 * it needs an extra guard page.
91 */
92#ifdef CONFIG_X86_32
93 char guard_entry_stack[PAGE_SIZE];
94#endif
95 struct entry_stack_page entry_stack_page;
96
97#ifdef CONFIG_X86_32
98 char guard_doublefault_stack[PAGE_SIZE];
99 struct doublefault_stack doublefault_stack;
100#endif
101
102 /*
103 * On x86_64, the TSS is mapped RO. On x86_32, it's mapped RW because
104 * we need task switches to work, and task switches write to the TSS.
105 */
106 struct tss_struct tss;
107
108#ifdef CONFIG_X86_64
109 /*
110 * Exception stacks used for IST entries with guard pages.
111 */
112 struct cea_exception_stacks estacks;
113#endif
114 /*
115 * Per CPU debug store for Intel performance monitoring. Wastes a
116 * full page at the moment.
117 */
118 struct debug_store cpu_debug_store;
119 /*
120 * The actual PEBS/BTS buffers must be mapped to user space
121 * Reserve enough fixmap PTEs.
122 */
123 struct debug_store_buffers cpu_debug_buffers;
124};
125
126#define CPU_ENTRY_AREA_SIZE (sizeof(struct cpu_entry_area))
127#define CPU_ENTRY_AREA_ARRAY_SIZE (CPU_ENTRY_AREA_SIZE * NR_CPUS)
128
129/* Total size includes the readonly IDT mapping page as well: */
130#define CPU_ENTRY_AREA_TOTAL_SIZE (CPU_ENTRY_AREA_ARRAY_SIZE + PAGE_SIZE)
131
132DECLARE_PER_CPU(struct cpu_entry_area *, cpu_entry_area);
133DECLARE_PER_CPU(struct cea_exception_stacks *, cea_exception_stacks);
134
135extern void setup_cpu_entry_areas(void);
136extern void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags);
137
138extern struct cpu_entry_area *get_cpu_entry_area(int cpu);
139
140static inline struct entry_stack *cpu_entry_stack(int cpu)
141{
142 return &get_cpu_entry_area(cpu)->entry_stack_page.stack;
143}
144
145#define __this_cpu_ist_top_va(name) \
146 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name)
147
148#define __this_cpu_ist_bottom_va(name) \
149 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
150
151#endif