cb8a2ef084
The kernel CONFIG_UNWINDER_ORC option enables the ORC unwinder, which is similar in concept to a DWARF unwinder. The difference is that the format of the ORC data is much simpler than DWARF, which in turn allows the ORC unwinder to be much simpler and faster. The ORC data consists of unwind tables which are generated by objtool. After analyzing all the code paths of a .o file, it determines information about the stack state at each instruction address in the file and outputs that information to the .orc_unwind and .orc_unwind_ip sections. The per-object ORC sections are combined at link time and are sorted and post-processed at boot time. The unwinder uses the resulting data to correlate instruction addresses with their stack states at run time. Most of the logic are similar with x86, in order to get ra info before ra is saved into stack, add ra_reg and ra_offset into orc_entry. At the same time, modify some arch-specific code to silence the objtool warnings. Co-developed-by: Jinyang He <hejinyang@loongson.cn> Signed-off-by: Jinyang He <hejinyang@loongson.cn> Co-developed-by: Youling Tang <tangyouling@loongson.cn> Signed-off-by: Youling Tang <tangyouling@loongson.cn> Signed-off-by: Tiezhu Yang <yangtiezhu@loongson.cn> Signed-off-by: Huacai Chen <chenhuacai@loongson.cn>
536 lines
15 KiB
ArmAsm
536 lines
15 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Author: Lu Zeng <zenglu@loongson.cn>
|
|
* Pei Huang <huangpei@loongson.cn>
|
|
* Huacai Chen <chenhuacai@loongson.cn>
|
|
*
|
|
* Copyright (C) 2020-2022 Loongson Technology Corporation Limited
|
|
*/
|
|
#include <linux/export.h>
|
|
#include <asm/asm.h>
|
|
#include <asm/asmmacro.h>
|
|
#include <asm/asm-extable.h>
|
|
#include <asm/asm-offsets.h>
|
|
#include <asm/errno.h>
|
|
#include <asm/fpregdef.h>
|
|
#include <asm/loongarch.h>
|
|
#include <asm/regdef.h>
|
|
#include <asm/unwind_hints.h>
|
|
|
|
#define FPU_REG_WIDTH 8
|
|
#define LSX_REG_WIDTH 16
|
|
#define LASX_REG_WIDTH 32
|
|
|
|
.macro EX insn, reg, src, offs
|
|
.ex\@: \insn \reg, \src, \offs
|
|
_asm_extable .ex\@, .L_fpu_fault
|
|
.endm
|
|
|
|
.macro sc_save_fp base
|
|
EX fst.d $f0, \base, (0 * FPU_REG_WIDTH)
|
|
EX fst.d $f1, \base, (1 * FPU_REG_WIDTH)
|
|
EX fst.d $f2, \base, (2 * FPU_REG_WIDTH)
|
|
EX fst.d $f3, \base, (3 * FPU_REG_WIDTH)
|
|
EX fst.d $f4, \base, (4 * FPU_REG_WIDTH)
|
|
EX fst.d $f5, \base, (5 * FPU_REG_WIDTH)
|
|
EX fst.d $f6, \base, (6 * FPU_REG_WIDTH)
|
|
EX fst.d $f7, \base, (7 * FPU_REG_WIDTH)
|
|
EX fst.d $f8, \base, (8 * FPU_REG_WIDTH)
|
|
EX fst.d $f9, \base, (9 * FPU_REG_WIDTH)
|
|
EX fst.d $f10, \base, (10 * FPU_REG_WIDTH)
|
|
EX fst.d $f11, \base, (11 * FPU_REG_WIDTH)
|
|
EX fst.d $f12, \base, (12 * FPU_REG_WIDTH)
|
|
EX fst.d $f13, \base, (13 * FPU_REG_WIDTH)
|
|
EX fst.d $f14, \base, (14 * FPU_REG_WIDTH)
|
|
EX fst.d $f15, \base, (15 * FPU_REG_WIDTH)
|
|
EX fst.d $f16, \base, (16 * FPU_REG_WIDTH)
|
|
EX fst.d $f17, \base, (17 * FPU_REG_WIDTH)
|
|
EX fst.d $f18, \base, (18 * FPU_REG_WIDTH)
|
|
EX fst.d $f19, \base, (19 * FPU_REG_WIDTH)
|
|
EX fst.d $f20, \base, (20 * FPU_REG_WIDTH)
|
|
EX fst.d $f21, \base, (21 * FPU_REG_WIDTH)
|
|
EX fst.d $f22, \base, (22 * FPU_REG_WIDTH)
|
|
EX fst.d $f23, \base, (23 * FPU_REG_WIDTH)
|
|
EX fst.d $f24, \base, (24 * FPU_REG_WIDTH)
|
|
EX fst.d $f25, \base, (25 * FPU_REG_WIDTH)
|
|
EX fst.d $f26, \base, (26 * FPU_REG_WIDTH)
|
|
EX fst.d $f27, \base, (27 * FPU_REG_WIDTH)
|
|
EX fst.d $f28, \base, (28 * FPU_REG_WIDTH)
|
|
EX fst.d $f29, \base, (29 * FPU_REG_WIDTH)
|
|
EX fst.d $f30, \base, (30 * FPU_REG_WIDTH)
|
|
EX fst.d $f31, \base, (31 * FPU_REG_WIDTH)
|
|
.endm
|
|
|
|
.macro sc_restore_fp base
|
|
EX fld.d $f0, \base, (0 * FPU_REG_WIDTH)
|
|
EX fld.d $f1, \base, (1 * FPU_REG_WIDTH)
|
|
EX fld.d $f2, \base, (2 * FPU_REG_WIDTH)
|
|
EX fld.d $f3, \base, (3 * FPU_REG_WIDTH)
|
|
EX fld.d $f4, \base, (4 * FPU_REG_WIDTH)
|
|
EX fld.d $f5, \base, (5 * FPU_REG_WIDTH)
|
|
EX fld.d $f6, \base, (6 * FPU_REG_WIDTH)
|
|
EX fld.d $f7, \base, (7 * FPU_REG_WIDTH)
|
|
EX fld.d $f8, \base, (8 * FPU_REG_WIDTH)
|
|
EX fld.d $f9, \base, (9 * FPU_REG_WIDTH)
|
|
EX fld.d $f10, \base, (10 * FPU_REG_WIDTH)
|
|
EX fld.d $f11, \base, (11 * FPU_REG_WIDTH)
|
|
EX fld.d $f12, \base, (12 * FPU_REG_WIDTH)
|
|
EX fld.d $f13, \base, (13 * FPU_REG_WIDTH)
|
|
EX fld.d $f14, \base, (14 * FPU_REG_WIDTH)
|
|
EX fld.d $f15, \base, (15 * FPU_REG_WIDTH)
|
|
EX fld.d $f16, \base, (16 * FPU_REG_WIDTH)
|
|
EX fld.d $f17, \base, (17 * FPU_REG_WIDTH)
|
|
EX fld.d $f18, \base, (18 * FPU_REG_WIDTH)
|
|
EX fld.d $f19, \base, (19 * FPU_REG_WIDTH)
|
|
EX fld.d $f20, \base, (20 * FPU_REG_WIDTH)
|
|
EX fld.d $f21, \base, (21 * FPU_REG_WIDTH)
|
|
EX fld.d $f22, \base, (22 * FPU_REG_WIDTH)
|
|
EX fld.d $f23, \base, (23 * FPU_REG_WIDTH)
|
|
EX fld.d $f24, \base, (24 * FPU_REG_WIDTH)
|
|
EX fld.d $f25, \base, (25 * FPU_REG_WIDTH)
|
|
EX fld.d $f26, \base, (26 * FPU_REG_WIDTH)
|
|
EX fld.d $f27, \base, (27 * FPU_REG_WIDTH)
|
|
EX fld.d $f28, \base, (28 * FPU_REG_WIDTH)
|
|
EX fld.d $f29, \base, (29 * FPU_REG_WIDTH)
|
|
EX fld.d $f30, \base, (30 * FPU_REG_WIDTH)
|
|
EX fld.d $f31, \base, (31 * FPU_REG_WIDTH)
|
|
.endm
|
|
|
|
.macro sc_save_fcc base, tmp0, tmp1
|
|
movcf2gr \tmp0, $fcc0
|
|
move \tmp1, \tmp0
|
|
movcf2gr \tmp0, $fcc1
|
|
bstrins.d \tmp1, \tmp0, 15, 8
|
|
movcf2gr \tmp0, $fcc2
|
|
bstrins.d \tmp1, \tmp0, 23, 16
|
|
movcf2gr \tmp0, $fcc3
|
|
bstrins.d \tmp1, \tmp0, 31, 24
|
|
movcf2gr \tmp0, $fcc4
|
|
bstrins.d \tmp1, \tmp0, 39, 32
|
|
movcf2gr \tmp0, $fcc5
|
|
bstrins.d \tmp1, \tmp0, 47, 40
|
|
movcf2gr \tmp0, $fcc6
|
|
bstrins.d \tmp1, \tmp0, 55, 48
|
|
movcf2gr \tmp0, $fcc7
|
|
bstrins.d \tmp1, \tmp0, 63, 56
|
|
EX st.d \tmp1, \base, 0
|
|
.endm
|
|
|
|
.macro sc_restore_fcc base, tmp0, tmp1
|
|
EX ld.d \tmp0, \base, 0
|
|
bstrpick.d \tmp1, \tmp0, 7, 0
|
|
movgr2cf $fcc0, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 15, 8
|
|
movgr2cf $fcc1, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 23, 16
|
|
movgr2cf $fcc2, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 31, 24
|
|
movgr2cf $fcc3, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 39, 32
|
|
movgr2cf $fcc4, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 47, 40
|
|
movgr2cf $fcc5, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 55, 48
|
|
movgr2cf $fcc6, \tmp1
|
|
bstrpick.d \tmp1, \tmp0, 63, 56
|
|
movgr2cf $fcc7, \tmp1
|
|
.endm
|
|
|
|
.macro sc_save_fcsr base, tmp0
|
|
movfcsr2gr \tmp0, fcsr0
|
|
EX st.w \tmp0, \base, 0
|
|
#if defined(CONFIG_CPU_HAS_LBT)
|
|
/* TM bit is always 0 if LBT not supported */
|
|
andi \tmp0, \tmp0, FPU_CSR_TM
|
|
beqz \tmp0, 1f
|
|
x86clrtm
|
|
1:
|
|
#endif
|
|
.endm
|
|
|
|
.macro sc_restore_fcsr base, tmp0
|
|
EX ld.w \tmp0, \base, 0
|
|
movgr2fcsr fcsr0, \tmp0
|
|
.endm
|
|
|
|
.macro sc_save_lsx base
|
|
#ifdef CONFIG_CPU_HAS_LSX
|
|
EX vst $vr0, \base, (0 * LSX_REG_WIDTH)
|
|
EX vst $vr1, \base, (1 * LSX_REG_WIDTH)
|
|
EX vst $vr2, \base, (2 * LSX_REG_WIDTH)
|
|
EX vst $vr3, \base, (3 * LSX_REG_WIDTH)
|
|
EX vst $vr4, \base, (4 * LSX_REG_WIDTH)
|
|
EX vst $vr5, \base, (5 * LSX_REG_WIDTH)
|
|
EX vst $vr6, \base, (6 * LSX_REG_WIDTH)
|
|
EX vst $vr7, \base, (7 * LSX_REG_WIDTH)
|
|
EX vst $vr8, \base, (8 * LSX_REG_WIDTH)
|
|
EX vst $vr9, \base, (9 * LSX_REG_WIDTH)
|
|
EX vst $vr10, \base, (10 * LSX_REG_WIDTH)
|
|
EX vst $vr11, \base, (11 * LSX_REG_WIDTH)
|
|
EX vst $vr12, \base, (12 * LSX_REG_WIDTH)
|
|
EX vst $vr13, \base, (13 * LSX_REG_WIDTH)
|
|
EX vst $vr14, \base, (14 * LSX_REG_WIDTH)
|
|
EX vst $vr15, \base, (15 * LSX_REG_WIDTH)
|
|
EX vst $vr16, \base, (16 * LSX_REG_WIDTH)
|
|
EX vst $vr17, \base, (17 * LSX_REG_WIDTH)
|
|
EX vst $vr18, \base, (18 * LSX_REG_WIDTH)
|
|
EX vst $vr19, \base, (19 * LSX_REG_WIDTH)
|
|
EX vst $vr20, \base, (20 * LSX_REG_WIDTH)
|
|
EX vst $vr21, \base, (21 * LSX_REG_WIDTH)
|
|
EX vst $vr22, \base, (22 * LSX_REG_WIDTH)
|
|
EX vst $vr23, \base, (23 * LSX_REG_WIDTH)
|
|
EX vst $vr24, \base, (24 * LSX_REG_WIDTH)
|
|
EX vst $vr25, \base, (25 * LSX_REG_WIDTH)
|
|
EX vst $vr26, \base, (26 * LSX_REG_WIDTH)
|
|
EX vst $vr27, \base, (27 * LSX_REG_WIDTH)
|
|
EX vst $vr28, \base, (28 * LSX_REG_WIDTH)
|
|
EX vst $vr29, \base, (29 * LSX_REG_WIDTH)
|
|
EX vst $vr30, \base, (30 * LSX_REG_WIDTH)
|
|
EX vst $vr31, \base, (31 * LSX_REG_WIDTH)
|
|
#endif
|
|
.endm
|
|
|
|
.macro sc_restore_lsx base
|
|
#ifdef CONFIG_CPU_HAS_LSX
|
|
EX vld $vr0, \base, (0 * LSX_REG_WIDTH)
|
|
EX vld $vr1, \base, (1 * LSX_REG_WIDTH)
|
|
EX vld $vr2, \base, (2 * LSX_REG_WIDTH)
|
|
EX vld $vr3, \base, (3 * LSX_REG_WIDTH)
|
|
EX vld $vr4, \base, (4 * LSX_REG_WIDTH)
|
|
EX vld $vr5, \base, (5 * LSX_REG_WIDTH)
|
|
EX vld $vr6, \base, (6 * LSX_REG_WIDTH)
|
|
EX vld $vr7, \base, (7 * LSX_REG_WIDTH)
|
|
EX vld $vr8, \base, (8 * LSX_REG_WIDTH)
|
|
EX vld $vr9, \base, (9 * LSX_REG_WIDTH)
|
|
EX vld $vr10, \base, (10 * LSX_REG_WIDTH)
|
|
EX vld $vr11, \base, (11 * LSX_REG_WIDTH)
|
|
EX vld $vr12, \base, (12 * LSX_REG_WIDTH)
|
|
EX vld $vr13, \base, (13 * LSX_REG_WIDTH)
|
|
EX vld $vr14, \base, (14 * LSX_REG_WIDTH)
|
|
EX vld $vr15, \base, (15 * LSX_REG_WIDTH)
|
|
EX vld $vr16, \base, (16 * LSX_REG_WIDTH)
|
|
EX vld $vr17, \base, (17 * LSX_REG_WIDTH)
|
|
EX vld $vr18, \base, (18 * LSX_REG_WIDTH)
|
|
EX vld $vr19, \base, (19 * LSX_REG_WIDTH)
|
|
EX vld $vr20, \base, (20 * LSX_REG_WIDTH)
|
|
EX vld $vr21, \base, (21 * LSX_REG_WIDTH)
|
|
EX vld $vr22, \base, (22 * LSX_REG_WIDTH)
|
|
EX vld $vr23, \base, (23 * LSX_REG_WIDTH)
|
|
EX vld $vr24, \base, (24 * LSX_REG_WIDTH)
|
|
EX vld $vr25, \base, (25 * LSX_REG_WIDTH)
|
|
EX vld $vr26, \base, (26 * LSX_REG_WIDTH)
|
|
EX vld $vr27, \base, (27 * LSX_REG_WIDTH)
|
|
EX vld $vr28, \base, (28 * LSX_REG_WIDTH)
|
|
EX vld $vr29, \base, (29 * LSX_REG_WIDTH)
|
|
EX vld $vr30, \base, (30 * LSX_REG_WIDTH)
|
|
EX vld $vr31, \base, (31 * LSX_REG_WIDTH)
|
|
#endif
|
|
.endm
|
|
|
|
.macro sc_save_lasx base
|
|
#ifdef CONFIG_CPU_HAS_LASX
|
|
EX xvst $xr0, \base, (0 * LASX_REG_WIDTH)
|
|
EX xvst $xr1, \base, (1 * LASX_REG_WIDTH)
|
|
EX xvst $xr2, \base, (2 * LASX_REG_WIDTH)
|
|
EX xvst $xr3, \base, (3 * LASX_REG_WIDTH)
|
|
EX xvst $xr4, \base, (4 * LASX_REG_WIDTH)
|
|
EX xvst $xr5, \base, (5 * LASX_REG_WIDTH)
|
|
EX xvst $xr6, \base, (6 * LASX_REG_WIDTH)
|
|
EX xvst $xr7, \base, (7 * LASX_REG_WIDTH)
|
|
EX xvst $xr8, \base, (8 * LASX_REG_WIDTH)
|
|
EX xvst $xr9, \base, (9 * LASX_REG_WIDTH)
|
|
EX xvst $xr10, \base, (10 * LASX_REG_WIDTH)
|
|
EX xvst $xr11, \base, (11 * LASX_REG_WIDTH)
|
|
EX xvst $xr12, \base, (12 * LASX_REG_WIDTH)
|
|
EX xvst $xr13, \base, (13 * LASX_REG_WIDTH)
|
|
EX xvst $xr14, \base, (14 * LASX_REG_WIDTH)
|
|
EX xvst $xr15, \base, (15 * LASX_REG_WIDTH)
|
|
EX xvst $xr16, \base, (16 * LASX_REG_WIDTH)
|
|
EX xvst $xr17, \base, (17 * LASX_REG_WIDTH)
|
|
EX xvst $xr18, \base, (18 * LASX_REG_WIDTH)
|
|
EX xvst $xr19, \base, (19 * LASX_REG_WIDTH)
|
|
EX xvst $xr20, \base, (20 * LASX_REG_WIDTH)
|
|
EX xvst $xr21, \base, (21 * LASX_REG_WIDTH)
|
|
EX xvst $xr22, \base, (22 * LASX_REG_WIDTH)
|
|
EX xvst $xr23, \base, (23 * LASX_REG_WIDTH)
|
|
EX xvst $xr24, \base, (24 * LASX_REG_WIDTH)
|
|
EX xvst $xr25, \base, (25 * LASX_REG_WIDTH)
|
|
EX xvst $xr26, \base, (26 * LASX_REG_WIDTH)
|
|
EX xvst $xr27, \base, (27 * LASX_REG_WIDTH)
|
|
EX xvst $xr28, \base, (28 * LASX_REG_WIDTH)
|
|
EX xvst $xr29, \base, (29 * LASX_REG_WIDTH)
|
|
EX xvst $xr30, \base, (30 * LASX_REG_WIDTH)
|
|
EX xvst $xr31, \base, (31 * LASX_REG_WIDTH)
|
|
#endif
|
|
.endm
|
|
|
|
.macro sc_restore_lasx base
|
|
#ifdef CONFIG_CPU_HAS_LASX
|
|
EX xvld $xr0, \base, (0 * LASX_REG_WIDTH)
|
|
EX xvld $xr1, \base, (1 * LASX_REG_WIDTH)
|
|
EX xvld $xr2, \base, (2 * LASX_REG_WIDTH)
|
|
EX xvld $xr3, \base, (3 * LASX_REG_WIDTH)
|
|
EX xvld $xr4, \base, (4 * LASX_REG_WIDTH)
|
|
EX xvld $xr5, \base, (5 * LASX_REG_WIDTH)
|
|
EX xvld $xr6, \base, (6 * LASX_REG_WIDTH)
|
|
EX xvld $xr7, \base, (7 * LASX_REG_WIDTH)
|
|
EX xvld $xr8, \base, (8 * LASX_REG_WIDTH)
|
|
EX xvld $xr9, \base, (9 * LASX_REG_WIDTH)
|
|
EX xvld $xr10, \base, (10 * LASX_REG_WIDTH)
|
|
EX xvld $xr11, \base, (11 * LASX_REG_WIDTH)
|
|
EX xvld $xr12, \base, (12 * LASX_REG_WIDTH)
|
|
EX xvld $xr13, \base, (13 * LASX_REG_WIDTH)
|
|
EX xvld $xr14, \base, (14 * LASX_REG_WIDTH)
|
|
EX xvld $xr15, \base, (15 * LASX_REG_WIDTH)
|
|
EX xvld $xr16, \base, (16 * LASX_REG_WIDTH)
|
|
EX xvld $xr17, \base, (17 * LASX_REG_WIDTH)
|
|
EX xvld $xr18, \base, (18 * LASX_REG_WIDTH)
|
|
EX xvld $xr19, \base, (19 * LASX_REG_WIDTH)
|
|
EX xvld $xr20, \base, (20 * LASX_REG_WIDTH)
|
|
EX xvld $xr21, \base, (21 * LASX_REG_WIDTH)
|
|
EX xvld $xr22, \base, (22 * LASX_REG_WIDTH)
|
|
EX xvld $xr23, \base, (23 * LASX_REG_WIDTH)
|
|
EX xvld $xr24, \base, (24 * LASX_REG_WIDTH)
|
|
EX xvld $xr25, \base, (25 * LASX_REG_WIDTH)
|
|
EX xvld $xr26, \base, (26 * LASX_REG_WIDTH)
|
|
EX xvld $xr27, \base, (27 * LASX_REG_WIDTH)
|
|
EX xvld $xr28, \base, (28 * LASX_REG_WIDTH)
|
|
EX xvld $xr29, \base, (29 * LASX_REG_WIDTH)
|
|
EX xvld $xr30, \base, (30 * LASX_REG_WIDTH)
|
|
EX xvld $xr31, \base, (31 * LASX_REG_WIDTH)
|
|
#endif
|
|
.endm
|
|
|
|
/*
|
|
* Save a thread's fp context.
|
|
*/
|
|
SYM_FUNC_START(_save_fp)
|
|
fpu_save_csr a0 t1
|
|
fpu_save_double a0 t1 # clobbers t1
|
|
fpu_save_cc a0 t1 t2 # clobbers t1, t2
|
|
jr ra
|
|
SYM_FUNC_END(_save_fp)
|
|
EXPORT_SYMBOL(_save_fp)
|
|
|
|
/*
|
|
* Restore a thread's fp context.
|
|
*/
|
|
SYM_FUNC_START(_restore_fp)
|
|
fpu_restore_double a0 t1 # clobbers t1
|
|
fpu_restore_csr a0 t1 t2
|
|
fpu_restore_cc a0 t1 t2 # clobbers t1, t2
|
|
jr ra
|
|
SYM_FUNC_END(_restore_fp)
|
|
|
|
#ifdef CONFIG_CPU_HAS_LSX
|
|
|
|
/*
|
|
* Save a thread's LSX vector context.
|
|
*/
|
|
SYM_FUNC_START(_save_lsx)
|
|
lsx_save_all a0 t1 t2
|
|
jr ra
|
|
SYM_FUNC_END(_save_lsx)
|
|
EXPORT_SYMBOL(_save_lsx)
|
|
|
|
/*
|
|
* Restore a thread's LSX vector context.
|
|
*/
|
|
SYM_FUNC_START(_restore_lsx)
|
|
lsx_restore_all a0 t1 t2
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lsx)
|
|
|
|
SYM_FUNC_START(_save_lsx_upper)
|
|
lsx_save_all_upper a0 t0 t1
|
|
jr ra
|
|
SYM_FUNC_END(_save_lsx_upper)
|
|
|
|
SYM_FUNC_START(_restore_lsx_upper)
|
|
lsx_restore_all_upper a0 t0 t1
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lsx_upper)
|
|
EXPORT_SYMBOL(_restore_lsx_upper)
|
|
|
|
SYM_FUNC_START(_init_lsx_upper)
|
|
lsx_init_all_upper t1
|
|
jr ra
|
|
SYM_FUNC_END(_init_lsx_upper)
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_HAS_LASX
|
|
|
|
/*
|
|
* Save a thread's LASX vector context.
|
|
*/
|
|
SYM_FUNC_START(_save_lasx)
|
|
lasx_save_all a0 t1 t2
|
|
jr ra
|
|
SYM_FUNC_END(_save_lasx)
|
|
EXPORT_SYMBOL(_save_lasx)
|
|
|
|
/*
|
|
* Restore a thread's LASX vector context.
|
|
*/
|
|
SYM_FUNC_START(_restore_lasx)
|
|
lasx_restore_all a0 t1 t2
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lasx)
|
|
|
|
SYM_FUNC_START(_save_lasx_upper)
|
|
lasx_save_all_upper a0 t0 t1
|
|
jr ra
|
|
SYM_FUNC_END(_save_lasx_upper)
|
|
|
|
SYM_FUNC_START(_restore_lasx_upper)
|
|
lasx_restore_all_upper a0 t0 t1
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lasx_upper)
|
|
EXPORT_SYMBOL(_restore_lasx_upper)
|
|
|
|
SYM_FUNC_START(_init_lasx_upper)
|
|
lasx_init_all_upper t1
|
|
jr ra
|
|
SYM_FUNC_END(_init_lasx_upper)
|
|
#endif
|
|
|
|
/*
|
|
* Load the FPU with signalling NANS. This bit pattern we're using has
|
|
* the property that no matter whether considered as single or as double
|
|
* precision represents signaling NANS.
|
|
*
|
|
* The value to initialize fcsr0 to comes in $a0.
|
|
*/
|
|
|
|
SYM_FUNC_START(_init_fpu)
|
|
li.w t1, CSR_EUEN_FPEN
|
|
csrxchg t1, t1, LOONGARCH_CSR_EUEN
|
|
|
|
movgr2fcsr fcsr0, a0
|
|
|
|
li.w t1, -1 # SNaN
|
|
|
|
movgr2fr.d $f0, t1
|
|
movgr2fr.d $f1, t1
|
|
movgr2fr.d $f2, t1
|
|
movgr2fr.d $f3, t1
|
|
movgr2fr.d $f4, t1
|
|
movgr2fr.d $f5, t1
|
|
movgr2fr.d $f6, t1
|
|
movgr2fr.d $f7, t1
|
|
movgr2fr.d $f8, t1
|
|
movgr2fr.d $f9, t1
|
|
movgr2fr.d $f10, t1
|
|
movgr2fr.d $f11, t1
|
|
movgr2fr.d $f12, t1
|
|
movgr2fr.d $f13, t1
|
|
movgr2fr.d $f14, t1
|
|
movgr2fr.d $f15, t1
|
|
movgr2fr.d $f16, t1
|
|
movgr2fr.d $f17, t1
|
|
movgr2fr.d $f18, t1
|
|
movgr2fr.d $f19, t1
|
|
movgr2fr.d $f20, t1
|
|
movgr2fr.d $f21, t1
|
|
movgr2fr.d $f22, t1
|
|
movgr2fr.d $f23, t1
|
|
movgr2fr.d $f24, t1
|
|
movgr2fr.d $f25, t1
|
|
movgr2fr.d $f26, t1
|
|
movgr2fr.d $f27, t1
|
|
movgr2fr.d $f28, t1
|
|
movgr2fr.d $f29, t1
|
|
movgr2fr.d $f30, t1
|
|
movgr2fr.d $f31, t1
|
|
|
|
jr ra
|
|
SYM_FUNC_END(_init_fpu)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_save_fp_context)
|
|
sc_save_fcc a1 t1 t2
|
|
sc_save_fcsr a2 t1
|
|
sc_save_fp a0
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_save_fp_context)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_restore_fp_context)
|
|
sc_restore_fp a0
|
|
sc_restore_fcc a1 t1 t2
|
|
sc_restore_fcsr a2 t1
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_restore_fp_context)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_save_lsx_context)
|
|
sc_save_fcc a1, t0, t1
|
|
sc_save_fcsr a2, t0
|
|
sc_save_lsx a0
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_save_lsx_context)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_restore_lsx_context)
|
|
sc_restore_lsx a0
|
|
sc_restore_fcc a1, t1, t2
|
|
sc_restore_fcsr a2, t1
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lsx_context)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_save_lasx_context)
|
|
sc_save_fcc a1, t0, t1
|
|
sc_save_fcsr a2, t0
|
|
sc_save_lasx a0
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_save_lasx_context)
|
|
|
|
/*
|
|
* a0: fpregs
|
|
* a1: fcc
|
|
* a2: fcsr
|
|
*/
|
|
SYM_FUNC_START(_restore_lasx_context)
|
|
sc_restore_lasx a0
|
|
sc_restore_fcc a1, t1, t2
|
|
sc_restore_fcsr a2, t1
|
|
li.w a0, 0 # success
|
|
jr ra
|
|
SYM_FUNC_END(_restore_lasx_context)
|
|
|
|
.L_fpu_fault:
|
|
li.w a0, -EFAULT # failure
|
|
jr ra
|
|
|
|
#ifdef CONFIG_CPU_HAS_LBT
|
|
STACK_FRAME_NON_STANDARD _restore_fp
|
|
STACK_FRAME_NON_STANDARD _restore_lsx
|
|
STACK_FRAME_NON_STANDARD _restore_lasx
|
|
#endif
|