7f65afb97f
This patch introduces definitions for the MSA control registers and functions which allow access to both the control & vector registers. If the toolchain being used to build the kernel includes support for MSA then this patch will make use of that support & use MSA instructions directly. However toolchain support for MSA is very new & far from a point where it can be reasonably expected that everyone building the kernel uses a toolchain with support. Thus fallbacks using .word assembler directives are also provided for now as a temporary measure. Signed-off-by: Paul Burton <paul.burton@imgtec.com> Cc: linux-mips@linux-mips.org Patchwork: https://patchwork.linux-mips.org/patch/6429/ Patchwork: https://patchwork.linux-mips.org/patch/6607/ Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
332 lines
7.0 KiB
C
332 lines
7.0 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 2003 Ralf Baechle
|
|
*/
|
|
#ifndef _ASM_ASMMACRO_H
|
|
#define _ASM_ASMMACRO_H
|
|
|
|
#include <asm/hazards.h>
|
|
|
|
#ifdef CONFIG_32BIT
|
|
#include <asm/asmmacro-32.h>
|
|
#endif
|
|
#ifdef CONFIG_64BIT
|
|
#include <asm/asmmacro-64.h>
|
|
#endif
|
|
#ifdef CONFIG_MIPS_MT_SMTC
|
|
#include <asm/mipsmtregs.h>
|
|
#endif
|
|
|
|
#ifdef CONFIG_MIPS_MT_SMTC
|
|
.macro local_irq_enable reg=t0
|
|
mfc0 \reg, CP0_TCSTATUS
|
|
ori \reg, \reg, TCSTATUS_IXMT
|
|
xori \reg, \reg, TCSTATUS_IXMT
|
|
mtc0 \reg, CP0_TCSTATUS
|
|
_ehb
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
mfc0 \reg, CP0_TCSTATUS
|
|
ori \reg, \reg, TCSTATUS_IXMT
|
|
mtc0 \reg, CP0_TCSTATUS
|
|
_ehb
|
|
.endm
|
|
#elif defined(CONFIG_CPU_MIPSR2)
|
|
.macro local_irq_enable reg=t0
|
|
ei
|
|
irq_enable_hazard
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
di
|
|
irq_disable_hazard
|
|
.endm
|
|
#else
|
|
.macro local_irq_enable reg=t0
|
|
mfc0 \reg, CP0_STATUS
|
|
ori \reg, \reg, 1
|
|
mtc0 \reg, CP0_STATUS
|
|
irq_enable_hazard
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
mfc0 \reg, CP0_STATUS
|
|
ori \reg, \reg, 1
|
|
xori \reg, \reg, 1
|
|
mtc0 \reg, CP0_STATUS
|
|
irq_disable_hazard
|
|
.endm
|
|
#endif /* CONFIG_MIPS_MT_SMTC */
|
|
|
|
.macro fpu_save_16even thread tmp=t0
|
|
cfc1 \tmp, fcr31
|
|
sdc1 $f0, THREAD_FPR0_LS64(\thread)
|
|
sdc1 $f2, THREAD_FPR2_LS64(\thread)
|
|
sdc1 $f4, THREAD_FPR4_LS64(\thread)
|
|
sdc1 $f6, THREAD_FPR6_LS64(\thread)
|
|
sdc1 $f8, THREAD_FPR8_LS64(\thread)
|
|
sdc1 $f10, THREAD_FPR10_LS64(\thread)
|
|
sdc1 $f12, THREAD_FPR12_LS64(\thread)
|
|
sdc1 $f14, THREAD_FPR14_LS64(\thread)
|
|
sdc1 $f16, THREAD_FPR16_LS64(\thread)
|
|
sdc1 $f18, THREAD_FPR18_LS64(\thread)
|
|
sdc1 $f20, THREAD_FPR20_LS64(\thread)
|
|
sdc1 $f22, THREAD_FPR22_LS64(\thread)
|
|
sdc1 $f24, THREAD_FPR24_LS64(\thread)
|
|
sdc1 $f26, THREAD_FPR26_LS64(\thread)
|
|
sdc1 $f28, THREAD_FPR28_LS64(\thread)
|
|
sdc1 $f30, THREAD_FPR30_LS64(\thread)
|
|
sw \tmp, THREAD_FCR31(\thread)
|
|
.endm
|
|
|
|
.macro fpu_save_16odd thread
|
|
.set push
|
|
.set mips64r2
|
|
sdc1 $f1, THREAD_FPR1_LS64(\thread)
|
|
sdc1 $f3, THREAD_FPR3_LS64(\thread)
|
|
sdc1 $f5, THREAD_FPR5_LS64(\thread)
|
|
sdc1 $f7, THREAD_FPR7_LS64(\thread)
|
|
sdc1 $f9, THREAD_FPR9_LS64(\thread)
|
|
sdc1 $f11, THREAD_FPR11_LS64(\thread)
|
|
sdc1 $f13, THREAD_FPR13_LS64(\thread)
|
|
sdc1 $f15, THREAD_FPR15_LS64(\thread)
|
|
sdc1 $f17, THREAD_FPR17_LS64(\thread)
|
|
sdc1 $f19, THREAD_FPR19_LS64(\thread)
|
|
sdc1 $f21, THREAD_FPR21_LS64(\thread)
|
|
sdc1 $f23, THREAD_FPR23_LS64(\thread)
|
|
sdc1 $f25, THREAD_FPR25_LS64(\thread)
|
|
sdc1 $f27, THREAD_FPR27_LS64(\thread)
|
|
sdc1 $f29, THREAD_FPR29_LS64(\thread)
|
|
sdc1 $f31, THREAD_FPR31_LS64(\thread)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro fpu_save_double thread status tmp
|
|
#if defined(CONFIG_MIPS64) || defined(CONFIG_CPU_MIPS32_R2)
|
|
sll \tmp, \status, 5
|
|
bgez \tmp, 10f
|
|
fpu_save_16odd \thread
|
|
10:
|
|
#endif
|
|
fpu_save_16even \thread \tmp
|
|
.endm
|
|
|
|
.macro fpu_restore_16even thread tmp=t0
|
|
lw \tmp, THREAD_FCR31(\thread)
|
|
ldc1 $f0, THREAD_FPR0_LS64(\thread)
|
|
ldc1 $f2, THREAD_FPR2_LS64(\thread)
|
|
ldc1 $f4, THREAD_FPR4_LS64(\thread)
|
|
ldc1 $f6, THREAD_FPR6_LS64(\thread)
|
|
ldc1 $f8, THREAD_FPR8_LS64(\thread)
|
|
ldc1 $f10, THREAD_FPR10_LS64(\thread)
|
|
ldc1 $f12, THREAD_FPR12_LS64(\thread)
|
|
ldc1 $f14, THREAD_FPR14_LS64(\thread)
|
|
ldc1 $f16, THREAD_FPR16_LS64(\thread)
|
|
ldc1 $f18, THREAD_FPR18_LS64(\thread)
|
|
ldc1 $f20, THREAD_FPR20_LS64(\thread)
|
|
ldc1 $f22, THREAD_FPR22_LS64(\thread)
|
|
ldc1 $f24, THREAD_FPR24_LS64(\thread)
|
|
ldc1 $f26, THREAD_FPR26_LS64(\thread)
|
|
ldc1 $f28, THREAD_FPR28_LS64(\thread)
|
|
ldc1 $f30, THREAD_FPR30_LS64(\thread)
|
|
ctc1 \tmp, fcr31
|
|
.endm
|
|
|
|
.macro fpu_restore_16odd thread
|
|
.set push
|
|
.set mips64r2
|
|
ldc1 $f1, THREAD_FPR1_LS64(\thread)
|
|
ldc1 $f3, THREAD_FPR3_LS64(\thread)
|
|
ldc1 $f5, THREAD_FPR5_LS64(\thread)
|
|
ldc1 $f7, THREAD_FPR7_LS64(\thread)
|
|
ldc1 $f9, THREAD_FPR9_LS64(\thread)
|
|
ldc1 $f11, THREAD_FPR11_LS64(\thread)
|
|
ldc1 $f13, THREAD_FPR13_LS64(\thread)
|
|
ldc1 $f15, THREAD_FPR15_LS64(\thread)
|
|
ldc1 $f17, THREAD_FPR17_LS64(\thread)
|
|
ldc1 $f19, THREAD_FPR19_LS64(\thread)
|
|
ldc1 $f21, THREAD_FPR21_LS64(\thread)
|
|
ldc1 $f23, THREAD_FPR23_LS64(\thread)
|
|
ldc1 $f25, THREAD_FPR25_LS64(\thread)
|
|
ldc1 $f27, THREAD_FPR27_LS64(\thread)
|
|
ldc1 $f29, THREAD_FPR29_LS64(\thread)
|
|
ldc1 $f31, THREAD_FPR31_LS64(\thread)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro fpu_restore_double thread status tmp
|
|
#if defined(CONFIG_MIPS64) || defined(CONFIG_CPU_MIPS32_R2)
|
|
sll \tmp, \status, 5
|
|
bgez \tmp, 10f # 16 register mode?
|
|
|
|
fpu_restore_16odd \thread
|
|
10:
|
|
#endif
|
|
fpu_restore_16even \thread \tmp
|
|
.endm
|
|
|
|
#ifdef CONFIG_CPU_MIPSR2
|
|
.macro _EXT rd, rs, p, s
|
|
ext \rd, \rs, \p, \s
|
|
.endm
|
|
#else /* !CONFIG_CPU_MIPSR2 */
|
|
.macro _EXT rd, rs, p, s
|
|
srl \rd, \rs, \p
|
|
andi \rd, \rd, (1 << \s) - 1
|
|
.endm
|
|
#endif /* !CONFIG_CPU_MIPSR2 */
|
|
|
|
/*
|
|
* Temporary until all gas have MT ASE support
|
|
*/
|
|
.macro DMT reg=0
|
|
.word 0x41600bc1 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro EMT reg=0
|
|
.word 0x41600be1 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro DVPE reg=0
|
|
.word 0x41600001 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro EVPE reg=0
|
|
.word 0x41600021 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro MFTR rt=0, rd=0, u=0, sel=0
|
|
.word 0x41000000 | (\rt << 16) | (\rd << 11) | (\u << 5) | (\sel)
|
|
.endm
|
|
|
|
.macro MTTR rt=0, rd=0, u=0, sel=0
|
|
.word 0x41800000 | (\rt << 16) | (\rd << 11) | (\u << 5) | (\sel)
|
|
.endm
|
|
|
|
#ifdef TOOLCHAIN_SUPPORTS_MSA
|
|
.macro ld_d wd, off, base
|
|
.set push
|
|
.set mips32r2
|
|
.set msa
|
|
ld.d $w\wd, \off(\base)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro st_d wd, off, base
|
|
.set push
|
|
.set mips32r2
|
|
.set msa
|
|
st.d $w\wd, \off(\base)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro copy_u_w rd, ws, n
|
|
.set push
|
|
.set mips32r2
|
|
.set msa
|
|
copy_u.w \rd, $w\ws[\n]
|
|
.set pop
|
|
.endm
|
|
|
|
.macro copy_u_d rd, ws, n
|
|
.set push
|
|
.set mips64r2
|
|
.set msa
|
|
copy_u.d \rd, $w\ws[\n]
|
|
.set pop
|
|
.endm
|
|
|
|
.macro insert_w wd, n, rs
|
|
.set push
|
|
.set mips32r2
|
|
.set msa
|
|
insert.w $w\wd[\n], \rs
|
|
.set pop
|
|
.endm
|
|
|
|
.macro insert_d wd, n, rs
|
|
.set push
|
|
.set mips64r2
|
|
.set msa
|
|
insert.d $w\wd[\n], \rs
|
|
.set pop
|
|
.endm
|
|
#else
|
|
/*
|
|
* Temporary until all toolchains in use include MSA support.
|
|
*/
|
|
.macro cfcmsa rd, cs
|
|
.set push
|
|
.set noat
|
|
.word 0x787e0059 | (\cs << 11)
|
|
move \rd, $1
|
|
.set pop
|
|
.endm
|
|
|
|
.macro ctcmsa cd, rs
|
|
.set push
|
|
.set noat
|
|
move $1, \rs
|
|
.word 0x783e0819 | (\cd << 6)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro ld_d wd, off, base
|
|
.set push
|
|
.set noat
|
|
add $1, \base, \off
|
|
.word 0x78000823 | (\wd << 6)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro st_d wd, off, base
|
|
.set push
|
|
.set noat
|
|
add $1, \base, \off
|
|
.word 0x78000827 | (\wd << 6)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro copy_u_w rd, ws, n
|
|
.set push
|
|
.set noat
|
|
.word 0x78f00059 | (\n << 16) | (\ws << 11)
|
|
/* move triggers an assembler bug... */
|
|
or \rd, $1, zero
|
|
.set pop
|
|
.endm
|
|
|
|
.macro copy_u_d rd, ws, n
|
|
.set push
|
|
.set noat
|
|
.word 0x78f80059 | (\n << 16) | (\ws << 11)
|
|
/* move triggers an assembler bug... */
|
|
or \rd, $1, zero
|
|
.set pop
|
|
.endm
|
|
|
|
.macro insert_w wd, n, rs
|
|
.set push
|
|
.set noat
|
|
/* move triggers an assembler bug... */
|
|
or $1, \rs, zero
|
|
.word 0x79300819 | (\n << 16) | (\wd << 6)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro insert_d wd, n, rs
|
|
.set push
|
|
.set noat
|
|
/* move triggers an assembler bug... */
|
|
or $1, \rs, zero
|
|
.word 0x79380819 | (\n << 16) | (\wd << 6)
|
|
.set pop
|
|
.endm
|
|
#endif
|
|
|
|
#endif /* _ASM_ASMMACRO_H */
|