The stubs for kvm_own/lsx()/kvm_own_lasx() when CONFIG_CPU_HAS_LSX or CONFIG_CPU_HAS_LASX is not defined should have a return value since they return an int, so add "return -EINVAL;" to the stubs. Fixes the build error: In file included from ../arch/loongarch/include/asm/kvm_csr.h:12, from ../arch/loongarch/kvm/interrupt.c:8: ../arch/loongarch/include/asm/kvm_vcpu.h: In function 'kvm_own_lasx': ../arch/loongarch/include/asm/kvm_vcpu.h:73:39: error: no return statement in function returning non-void [-Werror=return-type] 73 | static inline int kvm_own_lasx(struct kvm_vcpu *vcpu) { } Fixes: db1ecca22edf ("LoongArch: KVM: Add LSX (128bit SIMD) support") Fixes: 118e10cd893d ("LoongArch: KVM: Add LASX (256bit SIMD) support") Signed-off-by: Randy Dunlap <rdunlap@infradead.org> Signed-off-by: Huacai Chen <chenhuacai@loongson.cn>
113 lines
3.7 KiB
C
113 lines
3.7 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (C) 2020-2023 Loongson Technology Corporation Limited
|
|
*/
|
|
|
|
#ifndef __ASM_LOONGARCH_KVM_VCPU_H__
|
|
#define __ASM_LOONGARCH_KVM_VCPU_H__
|
|
|
|
#include <linux/kvm_host.h>
|
|
#include <asm/loongarch.h>
|
|
|
|
/* Controlled by 0x5 guest estat */
|
|
#define CPU_SIP0 (_ULCAST_(1))
|
|
#define CPU_SIP1 (_ULCAST_(1) << 1)
|
|
#define CPU_PMU (_ULCAST_(1) << 10)
|
|
#define CPU_TIMER (_ULCAST_(1) << 11)
|
|
#define CPU_IPI (_ULCAST_(1) << 12)
|
|
|
|
/* Controlled by 0x52 guest exception VIP aligned to estat bit 5~12 */
|
|
#define CPU_IP0 (_ULCAST_(1))
|
|
#define CPU_IP1 (_ULCAST_(1) << 1)
|
|
#define CPU_IP2 (_ULCAST_(1) << 2)
|
|
#define CPU_IP3 (_ULCAST_(1) << 3)
|
|
#define CPU_IP4 (_ULCAST_(1) << 4)
|
|
#define CPU_IP5 (_ULCAST_(1) << 5)
|
|
#define CPU_IP6 (_ULCAST_(1) << 6)
|
|
#define CPU_IP7 (_ULCAST_(1) << 7)
|
|
|
|
#define MNSEC_PER_SEC (NSEC_PER_SEC >> 20)
|
|
|
|
/* KVM_IRQ_LINE irq field index values */
|
|
#define KVM_LOONGSON_IRQ_TYPE_SHIFT 24
|
|
#define KVM_LOONGSON_IRQ_TYPE_MASK 0xff
|
|
#define KVM_LOONGSON_IRQ_VCPU_SHIFT 16
|
|
#define KVM_LOONGSON_IRQ_VCPU_MASK 0xff
|
|
#define KVM_LOONGSON_IRQ_NUM_SHIFT 0
|
|
#define KVM_LOONGSON_IRQ_NUM_MASK 0xffff
|
|
|
|
typedef union loongarch_instruction larch_inst;
|
|
typedef int (*exit_handle_fn)(struct kvm_vcpu *);
|
|
|
|
int kvm_emu_mmio_read(struct kvm_vcpu *vcpu, larch_inst inst);
|
|
int kvm_emu_mmio_write(struct kvm_vcpu *vcpu, larch_inst inst);
|
|
int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
|
int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
|
int kvm_emu_idle(struct kvm_vcpu *vcpu);
|
|
int kvm_pending_timer(struct kvm_vcpu *vcpu);
|
|
int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault);
|
|
void kvm_deliver_intr(struct kvm_vcpu *vcpu);
|
|
void kvm_deliver_exception(struct kvm_vcpu *vcpu);
|
|
|
|
void kvm_own_fpu(struct kvm_vcpu *vcpu);
|
|
void kvm_lose_fpu(struct kvm_vcpu *vcpu);
|
|
void kvm_save_fpu(struct loongarch_fpu *fpu);
|
|
void kvm_restore_fpu(struct loongarch_fpu *fpu);
|
|
void kvm_restore_fcsr(struct loongarch_fpu *fpu);
|
|
|
|
#ifdef CONFIG_CPU_HAS_LSX
|
|
int kvm_own_lsx(struct kvm_vcpu *vcpu);
|
|
void kvm_save_lsx(struct loongarch_fpu *fpu);
|
|
void kvm_restore_lsx(struct loongarch_fpu *fpu);
|
|
#else
|
|
static inline int kvm_own_lsx(struct kvm_vcpu *vcpu) { return -EINVAL; }
|
|
static inline void kvm_save_lsx(struct loongarch_fpu *fpu) { }
|
|
static inline void kvm_restore_lsx(struct loongarch_fpu *fpu) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_HAS_LASX
|
|
int kvm_own_lasx(struct kvm_vcpu *vcpu);
|
|
void kvm_save_lasx(struct loongarch_fpu *fpu);
|
|
void kvm_restore_lasx(struct loongarch_fpu *fpu);
|
|
#else
|
|
static inline int kvm_own_lasx(struct kvm_vcpu *vcpu) { return -EINVAL; }
|
|
static inline void kvm_save_lasx(struct loongarch_fpu *fpu) { }
|
|
static inline void kvm_restore_lasx(struct loongarch_fpu *fpu) { }
|
|
#endif
|
|
|
|
void kvm_init_timer(struct kvm_vcpu *vcpu, unsigned long hz);
|
|
void kvm_reset_timer(struct kvm_vcpu *vcpu);
|
|
void kvm_save_timer(struct kvm_vcpu *vcpu);
|
|
void kvm_restore_timer(struct kvm_vcpu *vcpu);
|
|
|
|
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq);
|
|
|
|
/*
|
|
* Loongarch KVM guest interrupt handling
|
|
*/
|
|
static inline void kvm_queue_irq(struct kvm_vcpu *vcpu, unsigned int irq)
|
|
{
|
|
set_bit(irq, &vcpu->arch.irq_pending);
|
|
clear_bit(irq, &vcpu->arch.irq_clear);
|
|
}
|
|
|
|
static inline void kvm_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int irq)
|
|
{
|
|
clear_bit(irq, &vcpu->arch.irq_pending);
|
|
set_bit(irq, &vcpu->arch.irq_clear);
|
|
}
|
|
|
|
static inline int kvm_queue_exception(struct kvm_vcpu *vcpu,
|
|
unsigned int code, unsigned int subcode)
|
|
{
|
|
/* only one exception can be injected */
|
|
if (!vcpu->arch.exception_pending) {
|
|
set_bit(code, &vcpu->arch.exception_pending);
|
|
vcpu->arch.esubcode = subcode;
|
|
return 0;
|
|
} else
|
|
return -1;
|
|
}
|
|
|
|
#endif /* __ASM_LOONGARCH_KVM_VCPU_H__ */
|