|
@@ -0,0 +1,122 @@
|
|
|
+/*
|
|
|
+ * This file is subject to the terms and conditions of the GNU General Public
|
|
|
+ * License. See the file "COPYING" in the main directory of this archive
|
|
|
+ * for more details.
|
|
|
+ *
|
|
|
+ * FPU context handling code for KVM.
|
|
|
+ *
|
|
|
+ * Copyright (C) 2015 Imagination Technologies Ltd.
|
|
|
+ */
|
|
|
+
|
|
|
+#include <asm/asm.h>
|
|
|
+#include <asm/asm-offsets.h>
|
|
|
+#include <asm/fpregdef.h>
|
|
|
+#include <asm/mipsregs.h>
|
|
|
+#include <asm/regdef.h>
|
|
|
+
|
|
|
+ .set noreorder
|
|
|
+ .set noat
|
|
|
+
|
|
|
+LEAF(__kvm_save_fpu)
|
|
|
+ .set push
|
|
|
+ .set mips64r2
|
|
|
+ SET_HARDFLOAT
|
|
|
+ mfc0 t0, CP0_STATUS
|
|
|
+ sll t0, t0, 5 # is Status.FR set?
|
|
|
+ bgez t0, 1f # no: skip odd doubles
|
|
|
+ nop
|
|
|
+ sdc1 $f1, VCPU_FPR1(a0)
|
|
|
+ sdc1 $f3, VCPU_FPR3(a0)
|
|
|
+ sdc1 $f5, VCPU_FPR5(a0)
|
|
|
+ sdc1 $f7, VCPU_FPR7(a0)
|
|
|
+ sdc1 $f9, VCPU_FPR9(a0)
|
|
|
+ sdc1 $f11, VCPU_FPR11(a0)
|
|
|
+ sdc1 $f13, VCPU_FPR13(a0)
|
|
|
+ sdc1 $f15, VCPU_FPR15(a0)
|
|
|
+ sdc1 $f17, VCPU_FPR17(a0)
|
|
|
+ sdc1 $f19, VCPU_FPR19(a0)
|
|
|
+ sdc1 $f21, VCPU_FPR21(a0)
|
|
|
+ sdc1 $f23, VCPU_FPR23(a0)
|
|
|
+ sdc1 $f25, VCPU_FPR25(a0)
|
|
|
+ sdc1 $f27, VCPU_FPR27(a0)
|
|
|
+ sdc1 $f29, VCPU_FPR29(a0)
|
|
|
+ sdc1 $f31, VCPU_FPR31(a0)
|
|
|
+1: sdc1 $f0, VCPU_FPR0(a0)
|
|
|
+ sdc1 $f2, VCPU_FPR2(a0)
|
|
|
+ sdc1 $f4, VCPU_FPR4(a0)
|
|
|
+ sdc1 $f6, VCPU_FPR6(a0)
|
|
|
+ sdc1 $f8, VCPU_FPR8(a0)
|
|
|
+ sdc1 $f10, VCPU_FPR10(a0)
|
|
|
+ sdc1 $f12, VCPU_FPR12(a0)
|
|
|
+ sdc1 $f14, VCPU_FPR14(a0)
|
|
|
+ sdc1 $f16, VCPU_FPR16(a0)
|
|
|
+ sdc1 $f18, VCPU_FPR18(a0)
|
|
|
+ sdc1 $f20, VCPU_FPR20(a0)
|
|
|
+ sdc1 $f22, VCPU_FPR22(a0)
|
|
|
+ sdc1 $f24, VCPU_FPR24(a0)
|
|
|
+ sdc1 $f26, VCPU_FPR26(a0)
|
|
|
+ sdc1 $f28, VCPU_FPR28(a0)
|
|
|
+ jr ra
|
|
|
+ sdc1 $f30, VCPU_FPR30(a0)
|
|
|
+ .set pop
|
|
|
+ END(__kvm_save_fpu)
|
|
|
+
|
|
|
+LEAF(__kvm_restore_fpu)
|
|
|
+ .set push
|
|
|
+ .set mips64r2
|
|
|
+ SET_HARDFLOAT
|
|
|
+ mfc0 t0, CP0_STATUS
|
|
|
+ sll t0, t0, 5 # is Status.FR set?
|
|
|
+ bgez t0, 1f # no: skip odd doubles
|
|
|
+ nop
|
|
|
+ ldc1 $f1, VCPU_FPR1(a0)
|
|
|
+ ldc1 $f3, VCPU_FPR3(a0)
|
|
|
+ ldc1 $f5, VCPU_FPR5(a0)
|
|
|
+ ldc1 $f7, VCPU_FPR7(a0)
|
|
|
+ ldc1 $f9, VCPU_FPR9(a0)
|
|
|
+ ldc1 $f11, VCPU_FPR11(a0)
|
|
|
+ ldc1 $f13, VCPU_FPR13(a0)
|
|
|
+ ldc1 $f15, VCPU_FPR15(a0)
|
|
|
+ ldc1 $f17, VCPU_FPR17(a0)
|
|
|
+ ldc1 $f19, VCPU_FPR19(a0)
|
|
|
+ ldc1 $f21, VCPU_FPR21(a0)
|
|
|
+ ldc1 $f23, VCPU_FPR23(a0)
|
|
|
+ ldc1 $f25, VCPU_FPR25(a0)
|
|
|
+ ldc1 $f27, VCPU_FPR27(a0)
|
|
|
+ ldc1 $f29, VCPU_FPR29(a0)
|
|
|
+ ldc1 $f31, VCPU_FPR31(a0)
|
|
|
+1: ldc1 $f0, VCPU_FPR0(a0)
|
|
|
+ ldc1 $f2, VCPU_FPR2(a0)
|
|
|
+ ldc1 $f4, VCPU_FPR4(a0)
|
|
|
+ ldc1 $f6, VCPU_FPR6(a0)
|
|
|
+ ldc1 $f8, VCPU_FPR8(a0)
|
|
|
+ ldc1 $f10, VCPU_FPR10(a0)
|
|
|
+ ldc1 $f12, VCPU_FPR12(a0)
|
|
|
+ ldc1 $f14, VCPU_FPR14(a0)
|
|
|
+ ldc1 $f16, VCPU_FPR16(a0)
|
|
|
+ ldc1 $f18, VCPU_FPR18(a0)
|
|
|
+ ldc1 $f20, VCPU_FPR20(a0)
|
|
|
+ ldc1 $f22, VCPU_FPR22(a0)
|
|
|
+ ldc1 $f24, VCPU_FPR24(a0)
|
|
|
+ ldc1 $f26, VCPU_FPR26(a0)
|
|
|
+ ldc1 $f28, VCPU_FPR28(a0)
|
|
|
+ jr ra
|
|
|
+ ldc1 $f30, VCPU_FPR30(a0)
|
|
|
+ .set pop
|
|
|
+ END(__kvm_restore_fpu)
|
|
|
+
|
|
|
+LEAF(__kvm_restore_fcsr)
|
|
|
+ .set push
|
|
|
+ SET_HARDFLOAT
|
|
|
+ lw t0, VCPU_FCR31(a0)
|
|
|
+ /*
|
|
|
+ * The ctc1 must stay at this offset in __kvm_restore_fcsr.
|
|
|
+ * See kvm_mips_csr_die_notify() which handles t0 containing a value
|
|
|
+ * which triggers an FP Exception, which must be stepped over and
|
|
|
+ * ignored since the set cause bits must remain there for the guest.
|
|
|
+ */
|
|
|
+ ctc1 t0, fcr31
|
|
|
+ jr ra
|
|
|
+ nop
|
|
|
+ .set pop
|
|
|
+ END(__kvm_restore_fcsr)
|