User will use its Vector registers only after the kernel really returns to the userspace. So we can delay restoring Vector registers as long as we are still running in kernel mode. So, add a thread flag to indicates the need of restoring Vector and do the restore at the last arch-specific exit-to-user hook. This save the context restoring cost when we switch over multiple processes that run V in kernel mode. For example, if the kernel performs a context swicth from A->B->C, and returns to C's userspace, then there is no need to restore B's V-register. Besides, this also prevents us from repeatedly restoring V context when executing kernel-mode Vector multiple times. The cost of this is that we must disable preemption and mark vector as busy during vstate_{save,restore}. Because then the V context will not get restored back immediately when a trap-causing context switch happens in the middle of vstate_{save,restore}. Signed-off-by: Andy Chiu <andy.chiu@sifive.com> Acked-by: Conor Dooley <conor.dooley@microchip.com> Tested-by: Björn Töpel <bjorn@rivosinc.com> Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com> Link: https://lore.kernel.org/r/20240115055929.4736-5-andy.chiu@sifive.com Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
238 lines
6.0 KiB
C
238 lines
6.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* Copyright (C) 2020 SiFive
|
|
*/
|
|
|
|
#ifndef __ASM_RISCV_VECTOR_H
|
|
#define __ASM_RISCV_VECTOR_H
|
|
|
|
#include <linux/types.h>
|
|
#include <uapi/asm-generic/errno.h>
|
|
|
|
#ifdef CONFIG_RISCV_ISA_V
|
|
|
|
#include <linux/stringify.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/sched/task_stack.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/cpufeature.h>
|
|
#include <asm/csr.h>
|
|
#include <asm/asm.h>
|
|
|
|
extern unsigned long riscv_v_vsize;
|
|
int riscv_v_setup_vsize(void);
|
|
bool riscv_v_first_use_handler(struct pt_regs *regs);
|
|
void kernel_vector_begin(void);
|
|
void kernel_vector_end(void);
|
|
void get_cpu_vector_context(void);
|
|
void put_cpu_vector_context(void);
|
|
|
|
static inline u32 riscv_v_flags(void)
|
|
{
|
|
return current->thread.riscv_v_flags;
|
|
}
|
|
|
|
static __always_inline bool has_vector(void)
|
|
{
|
|
return riscv_has_extension_unlikely(RISCV_ISA_EXT_v);
|
|
}
|
|
|
|
static inline void __riscv_v_vstate_clean(struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_VS) | SR_VS_CLEAN;
|
|
}
|
|
|
|
static inline void __riscv_v_vstate_dirty(struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_VS) | SR_VS_DIRTY;
|
|
}
|
|
|
|
static inline void riscv_v_vstate_off(struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_VS) | SR_VS_OFF;
|
|
}
|
|
|
|
static inline void riscv_v_vstate_on(struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_VS) | SR_VS_INITIAL;
|
|
}
|
|
|
|
static inline bool riscv_v_vstate_query(struct pt_regs *regs)
|
|
{
|
|
return (regs->status & SR_VS) != 0;
|
|
}
|
|
|
|
static __always_inline void riscv_v_enable(void)
|
|
{
|
|
csr_set(CSR_SSTATUS, SR_VS);
|
|
}
|
|
|
|
static __always_inline void riscv_v_disable(void)
|
|
{
|
|
csr_clear(CSR_SSTATUS, SR_VS);
|
|
}
|
|
|
|
static __always_inline void __vstate_csr_save(struct __riscv_v_ext_state *dest)
|
|
{
|
|
asm volatile (
|
|
"csrr %0, " __stringify(CSR_VSTART) "\n\t"
|
|
"csrr %1, " __stringify(CSR_VTYPE) "\n\t"
|
|
"csrr %2, " __stringify(CSR_VL) "\n\t"
|
|
"csrr %3, " __stringify(CSR_VCSR) "\n\t"
|
|
"csrr %4, " __stringify(CSR_VLENB) "\n\t"
|
|
: "=r" (dest->vstart), "=r" (dest->vtype), "=r" (dest->vl),
|
|
"=r" (dest->vcsr), "=r" (dest->vlenb) : :);
|
|
}
|
|
|
|
static __always_inline void __vstate_csr_restore(struct __riscv_v_ext_state *src)
|
|
{
|
|
asm volatile (
|
|
".option push\n\t"
|
|
".option arch, +v\n\t"
|
|
"vsetvl x0, %2, %1\n\t"
|
|
".option pop\n\t"
|
|
"csrw " __stringify(CSR_VSTART) ", %0\n\t"
|
|
"csrw " __stringify(CSR_VCSR) ", %3\n\t"
|
|
: : "r" (src->vstart), "r" (src->vtype), "r" (src->vl),
|
|
"r" (src->vcsr) :);
|
|
}
|
|
|
|
static inline void __riscv_v_vstate_save(struct __riscv_v_ext_state *save_to,
|
|
void *datap)
|
|
{
|
|
unsigned long vl;
|
|
|
|
riscv_v_enable();
|
|
__vstate_csr_save(save_to);
|
|
asm volatile (
|
|
".option push\n\t"
|
|
".option arch, +v\n\t"
|
|
"vsetvli %0, x0, e8, m8, ta, ma\n\t"
|
|
"vse8.v v0, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vse8.v v8, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vse8.v v16, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vse8.v v24, (%1)\n\t"
|
|
".option pop\n\t"
|
|
: "=&r" (vl) : "r" (datap) : "memory");
|
|
riscv_v_disable();
|
|
}
|
|
|
|
static inline void __riscv_v_vstate_restore(struct __riscv_v_ext_state *restore_from,
|
|
void *datap)
|
|
{
|
|
unsigned long vl;
|
|
|
|
riscv_v_enable();
|
|
asm volatile (
|
|
".option push\n\t"
|
|
".option arch, +v\n\t"
|
|
"vsetvli %0, x0, e8, m8, ta, ma\n\t"
|
|
"vle8.v v0, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vle8.v v8, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vle8.v v16, (%1)\n\t"
|
|
"add %1, %1, %0\n\t"
|
|
"vle8.v v24, (%1)\n\t"
|
|
".option pop\n\t"
|
|
: "=&r" (vl) : "r" (datap) : "memory");
|
|
__vstate_csr_restore(restore_from);
|
|
riscv_v_disable();
|
|
}
|
|
|
|
static inline void __riscv_v_vstate_discard(void)
|
|
{
|
|
unsigned long vl, vtype_inval = 1UL << (BITS_PER_LONG - 1);
|
|
|
|
riscv_v_enable();
|
|
asm volatile (
|
|
".option push\n\t"
|
|
".option arch, +v\n\t"
|
|
"vsetvli %0, x0, e8, m8, ta, ma\n\t"
|
|
"vmv.v.i v0, -1\n\t"
|
|
"vmv.v.i v8, -1\n\t"
|
|
"vmv.v.i v16, -1\n\t"
|
|
"vmv.v.i v24, -1\n\t"
|
|
"vsetvl %0, x0, %1\n\t"
|
|
".option pop\n\t"
|
|
: "=&r" (vl) : "r" (vtype_inval) : "memory");
|
|
riscv_v_disable();
|
|
}
|
|
|
|
static inline void riscv_v_vstate_discard(struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_VS) == SR_VS_OFF)
|
|
return;
|
|
|
|
__riscv_v_vstate_discard();
|
|
__riscv_v_vstate_dirty(regs);
|
|
}
|
|
|
|
static inline void riscv_v_vstate_save(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_VS) == SR_VS_DIRTY) {
|
|
struct __riscv_v_ext_state *vstate = &task->thread.vstate;
|
|
|
|
__riscv_v_vstate_save(vstate, vstate->datap);
|
|
__riscv_v_vstate_clean(regs);
|
|
}
|
|
}
|
|
|
|
static inline void riscv_v_vstate_restore(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_VS) != SR_VS_OFF) {
|
|
struct __riscv_v_ext_state *vstate = &task->thread.vstate;
|
|
|
|
__riscv_v_vstate_restore(vstate, vstate->datap);
|
|
__riscv_v_vstate_clean(regs);
|
|
}
|
|
}
|
|
|
|
static inline void riscv_v_vstate_set_restore(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_VS) != SR_VS_OFF) {
|
|
set_tsk_thread_flag(task, TIF_RISCV_V_DEFER_RESTORE);
|
|
riscv_v_vstate_on(regs);
|
|
}
|
|
}
|
|
|
|
static inline void __switch_to_vector(struct task_struct *prev,
|
|
struct task_struct *next)
|
|
{
|
|
struct pt_regs *regs;
|
|
|
|
regs = task_pt_regs(prev);
|
|
riscv_v_vstate_save(prev, regs);
|
|
riscv_v_vstate_set_restore(next, task_pt_regs(next));
|
|
}
|
|
|
|
void riscv_v_vstate_ctrl_init(struct task_struct *tsk);
|
|
bool riscv_v_vstate_ctrl_user_allowed(void);
|
|
|
|
#else /* ! CONFIG_RISCV_ISA_V */
|
|
|
|
struct pt_regs;
|
|
|
|
static inline int riscv_v_setup_vsize(void) { return -EOPNOTSUPP; }
|
|
static __always_inline bool has_vector(void) { return false; }
|
|
static inline bool riscv_v_first_use_handler(struct pt_regs *regs) { return false; }
|
|
static inline bool riscv_v_vstate_query(struct pt_regs *regs) { return false; }
|
|
static inline bool riscv_v_vstate_ctrl_user_allowed(void) { return false; }
|
|
#define riscv_v_vsize (0)
|
|
#define riscv_v_vstate_discard(regs) do {} while (0)
|
|
#define riscv_v_vstate_save(task, regs) do {} while (0)
|
|
#define riscv_v_vstate_restore(task, regs) do {} while (0)
|
|
#define __switch_to_vector(__prev, __next) do {} while (0)
|
|
#define riscv_v_vstate_off(regs) do {} while (0)
|
|
#define riscv_v_vstate_on(regs) do {} while (0)
|
|
|
|
#endif /* CONFIG_RISCV_ISA_V */
|
|
|
|
#endif /* ! __ASM_RISCV_VECTOR_H */
|