|
@@ -47,7 +47,7 @@
|
|
* section. Since TSS's are completely CPU-local, we want them
|
|
* section. Since TSS's are completely CPU-local, we want them
|
|
* on exact cacheline boundaries, to eliminate cacheline ping-pong.
|
|
* on exact cacheline boundaries, to eliminate cacheline ping-pong.
|
|
*/
|
|
*/
|
|
-__visible DEFINE_PER_CPU_SHARED_ALIGNED(struct tss_struct, cpu_tss_rw) = {
|
|
|
|
|
|
+__visible DEFINE_PER_CPU_PAGE_ALIGNED(struct tss_struct, cpu_tss_rw) = {
|
|
.x86_tss = {
|
|
.x86_tss = {
|
|
/*
|
|
/*
|
|
* .sp0 is only used when entering ring 0 from a lower
|
|
* .sp0 is only used when entering ring 0 from a lower
|