123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081 |
- // SPDX-License-Identifier: GPL-2.0
- #ifndef _ASM_X86_CPU_ENTRY_AREA_H
- #define _ASM_X86_CPU_ENTRY_AREA_H
- #include <linux/percpu-defs.h>
- #include <asm/processor.h>
- #include <asm/intel_ds.h>
- /*
- * cpu_entry_area is a percpu region that contains things needed by the CPU
- * and early entry/exit code. Real types aren't used for all fields here
- * to avoid circular header dependencies.
- *
- * Every field is a virtual alias of some other allocated backing store.
- * There is no direct allocation of a struct cpu_entry_area.
- */
- struct cpu_entry_area {
- char gdt[PAGE_SIZE];
- /*
- * The GDT is just below entry_stack and thus serves (on x86_64) as
- * a a read-only guard page.
- */
- struct entry_stack_page entry_stack_page;
- /*
- * On x86_64, the TSS is mapped RO. On x86_32, it's mapped RW because
- * we need task switches to work, and task switches write to the TSS.
- */
- struct tss_struct tss;
- char entry_trampoline[PAGE_SIZE];
- #ifdef CONFIG_X86_64
- /*
- * Exception stacks used for IST entries.
- *
- * In the future, this should have a separate slot for each stack
- * with guard pages between them.
- */
- char exception_stacks[(N_EXCEPTION_STACKS - 1) * EXCEPTION_STKSZ + DEBUG_STKSZ];
- #endif
- #ifdef CONFIG_CPU_SUP_INTEL
- /*
- * Per CPU debug store for Intel performance monitoring. Wastes a
- * full page at the moment.
- */
- struct debug_store cpu_debug_store;
- /*
- * The actual PEBS/BTS buffers must be mapped to user space
- * Reserve enough fixmap PTEs.
- */
- struct debug_store_buffers cpu_debug_buffers;
- #endif
- };
- #define CPU_ENTRY_AREA_SIZE (sizeof(struct cpu_entry_area))
- #define CPU_ENTRY_AREA_TOT_SIZE (CPU_ENTRY_AREA_SIZE * NR_CPUS)
- DECLARE_PER_CPU(struct cpu_entry_area *, cpu_entry_area);
- extern void setup_cpu_entry_areas(void);
- extern void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags);
- #define CPU_ENTRY_AREA_RO_IDT CPU_ENTRY_AREA_BASE
- #define CPU_ENTRY_AREA_PER_CPU (CPU_ENTRY_AREA_RO_IDT + PAGE_SIZE)
- #define CPU_ENTRY_AREA_RO_IDT_VADDR ((void *)CPU_ENTRY_AREA_RO_IDT)
- #define CPU_ENTRY_AREA_MAP_SIZE \
- (CPU_ENTRY_AREA_PER_CPU + CPU_ENTRY_AREA_TOT_SIZE - CPU_ENTRY_AREA_BASE)
- extern struct cpu_entry_area *get_cpu_entry_area(int cpu);
- static inline struct entry_stack *cpu_entry_stack(int cpu)
- {
- return &get_cpu_entry_area(cpu)->entry_stack_page.stack;
- }
- #endif
|