kvm: x86: move ioapic.c and irq_comm.c back to arch/x86/
ia64 does not need them anymore. Ack notifiers become x86-specific too. Suggested-by: Gleb Natapov <gleb@kernel.org> Reviewed-by: Radim Krcmar <rkrcmar@redhat.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
This commit is contained in:
parent
c32a42721c
commit
6ef768fac9
@ -603,6 +603,9 @@ struct kvm_arch {
|
|||||||
|
|
||||||
struct kvm_xen_hvm_config xen_hvm_config;
|
struct kvm_xen_hvm_config xen_hvm_config;
|
||||||
|
|
||||||
|
/* reads protected by irq_srcu, writes by irq_lock */
|
||||||
|
struct hlist_head mask_notifier_list;
|
||||||
|
|
||||||
/* fields used by HYPER-V emulation */
|
/* fields used by HYPER-V emulation */
|
||||||
u64 hv_guest_os_id;
|
u64 hv_guest_os_id;
|
||||||
u64 hv_hypercall;
|
u64 hv_hypercall;
|
||||||
@ -819,6 +822,19 @@ int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
|
|||||||
const void *val, int bytes);
|
const void *val, int bytes);
|
||||||
u8 kvm_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn);
|
u8 kvm_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn);
|
||||||
|
|
||||||
|
struct kvm_irq_mask_notifier {
|
||||||
|
void (*func)(struct kvm_irq_mask_notifier *kimn, bool masked);
|
||||||
|
int irq;
|
||||||
|
struct hlist_node link;
|
||||||
|
};
|
||||||
|
|
||||||
|
void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq,
|
||||||
|
struct kvm_irq_mask_notifier *kimn);
|
||||||
|
void kvm_unregister_irq_mask_notifier(struct kvm *kvm, int irq,
|
||||||
|
struct kvm_irq_mask_notifier *kimn);
|
||||||
|
void kvm_fire_mask_notifiers(struct kvm *kvm, unsigned irqchip, unsigned pin,
|
||||||
|
bool mask);
|
||||||
|
|
||||||
extern bool tdp_enabled;
|
extern bool tdp_enabled;
|
||||||
|
|
||||||
u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu);
|
u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu);
|
||||||
|
@ -7,14 +7,13 @@ CFLAGS_vmx.o := -I.
|
|||||||
|
|
||||||
KVM := ../../../virt/kvm
|
KVM := ../../../virt/kvm
|
||||||
|
|
||||||
kvm-y += $(KVM)/kvm_main.o $(KVM)/ioapic.o \
|
kvm-y += $(KVM)/kvm_main.o $(KVM)/coalesced_mmio.o \
|
||||||
$(KVM)/coalesced_mmio.o $(KVM)/irq_comm.o \
|
|
||||||
$(KVM)/eventfd.o $(KVM)/irqchip.o $(KVM)/vfio.o
|
$(KVM)/eventfd.o $(KVM)/irqchip.o $(KVM)/vfio.o
|
||||||
kvm-$(CONFIG_KVM_DEVICE_ASSIGNMENT) += $(KVM)/assigned-dev.o $(KVM)/iommu.o
|
kvm-$(CONFIG_KVM_DEVICE_ASSIGNMENT) += $(KVM)/assigned-dev.o $(KVM)/iommu.o
|
||||||
kvm-$(CONFIG_KVM_ASYNC_PF) += $(KVM)/async_pf.o
|
kvm-$(CONFIG_KVM_ASYNC_PF) += $(KVM)/async_pf.o
|
||||||
|
|
||||||
kvm-y += x86.o mmu.o emulate.o i8259.o irq.o lapic.o \
|
kvm-y += x86.o mmu.o emulate.o i8259.o irq.o lapic.o \
|
||||||
i8254.o cpuid.o pmu.o
|
i8254.o ioapic.o irq_comm.o cpuid.o pmu.o
|
||||||
kvm-intel-y += vmx.o
|
kvm-intel-y += vmx.o
|
||||||
kvm-amd-y += svm.o
|
kvm-amd-y += svm.o
|
||||||
|
|
||||||
|
@ -96,7 +96,6 @@ int kvm_irq_delivery_to_apic(struct kvm *kvm, struct kvm_lapic *src,
|
|||||||
struct kvm_lapic_irq *irq, unsigned long *dest_map);
|
struct kvm_lapic_irq *irq, unsigned long *dest_map);
|
||||||
int kvm_get_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state);
|
int kvm_get_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state);
|
||||||
int kvm_set_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state);
|
int kvm_set_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state);
|
||||||
void kvm_vcpu_request_scan_ioapic(struct kvm *kvm);
|
|
||||||
void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap,
|
void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap,
|
||||||
u32 *tmr);
|
u32 *tmr);
|
||||||
|
|
@ -234,7 +234,7 @@ void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq,
|
|||||||
{
|
{
|
||||||
mutex_lock(&kvm->irq_lock);
|
mutex_lock(&kvm->irq_lock);
|
||||||
kimn->irq = irq;
|
kimn->irq = irq;
|
||||||
hlist_add_head_rcu(&kimn->link, &kvm->mask_notifier_list);
|
hlist_add_head_rcu(&kimn->link, &kvm->arch.mask_notifier_list);
|
||||||
mutex_unlock(&kvm->irq_lock);
|
mutex_unlock(&kvm->irq_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -256,7 +256,7 @@ void kvm_fire_mask_notifiers(struct kvm *kvm, unsigned irqchip, unsigned pin,
|
|||||||
idx = srcu_read_lock(&kvm->irq_srcu);
|
idx = srcu_read_lock(&kvm->irq_srcu);
|
||||||
gsi = kvm_irq_map_chip_pin(kvm, irqchip, pin);
|
gsi = kvm_irq_map_chip_pin(kvm, irqchip, pin);
|
||||||
if (gsi != -1)
|
if (gsi != -1)
|
||||||
hlist_for_each_entry_rcu(kimn, &kvm->mask_notifier_list, link)
|
hlist_for_each_entry_rcu(kimn, &kvm->arch.mask_notifier_list, link)
|
||||||
if (kimn->irq == gsi)
|
if (kimn->irq == gsi)
|
||||||
kimn->func(kimn, mask);
|
kimn->func(kimn, mask);
|
||||||
srcu_read_unlock(&kvm->irq_srcu, idx);
|
srcu_read_unlock(&kvm->irq_srcu, idx);
|
@ -7264,6 +7264,7 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type)
|
|||||||
if (type)
|
if (type)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
|
INIT_HLIST_HEAD(&kvm->arch.mask_notifier_list);
|
||||||
INIT_LIST_HEAD(&kvm->arch.active_mmu_pages);
|
INIT_LIST_HEAD(&kvm->arch.active_mmu_pages);
|
||||||
INIT_LIST_HEAD(&kvm->arch.zapped_obsolete_pages);
|
INIT_LIST_HEAD(&kvm->arch.zapped_obsolete_pages);
|
||||||
INIT_LIST_HEAD(&kvm->arch.assigned_dev_head);
|
INIT_LIST_HEAD(&kvm->arch.assigned_dev_head);
|
||||||
|
@ -395,7 +395,6 @@ struct kvm {
|
|||||||
* Update side is protected by irq_lock.
|
* Update side is protected by irq_lock.
|
||||||
*/
|
*/
|
||||||
struct kvm_irq_routing_table __rcu *irq_routing;
|
struct kvm_irq_routing_table __rcu *irq_routing;
|
||||||
struct hlist_head mask_notifier_list;
|
|
||||||
#endif
|
#endif
|
||||||
#ifdef CONFIG_HAVE_KVM_IRQFD
|
#ifdef CONFIG_HAVE_KVM_IRQFD
|
||||||
struct hlist_head irq_ack_notifier_list;
|
struct hlist_head irq_ack_notifier_list;
|
||||||
@ -447,6 +446,14 @@ void kvm_vcpu_uninit(struct kvm_vcpu *vcpu);
|
|||||||
int __must_check vcpu_load(struct kvm_vcpu *vcpu);
|
int __must_check vcpu_load(struct kvm_vcpu *vcpu);
|
||||||
void vcpu_put(struct kvm_vcpu *vcpu);
|
void vcpu_put(struct kvm_vcpu *vcpu);
|
||||||
|
|
||||||
|
#ifdef __KVM_HAVE_IOAPIC
|
||||||
|
void kvm_vcpu_request_scan_ioapic(struct kvm *kvm);
|
||||||
|
#else
|
||||||
|
static inline void kvm_vcpu_request_scan_ioapic(struct kvm *kvm)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_HAVE_KVM_IRQFD
|
#ifdef CONFIG_HAVE_KVM_IRQFD
|
||||||
int kvm_irqfd_init(void);
|
int kvm_irqfd_init(void);
|
||||||
void kvm_irqfd_exit(void);
|
void kvm_irqfd_exit(void);
|
||||||
@ -736,19 +743,6 @@ struct kvm_assigned_dev_kernel {
|
|||||||
struct pci_saved_state *pci_saved_state;
|
struct pci_saved_state *pci_saved_state;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct kvm_irq_mask_notifier {
|
|
||||||
void (*func)(struct kvm_irq_mask_notifier *kimn, bool masked);
|
|
||||||
int irq;
|
|
||||||
struct hlist_node link;
|
|
||||||
};
|
|
||||||
|
|
||||||
void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq,
|
|
||||||
struct kvm_irq_mask_notifier *kimn);
|
|
||||||
void kvm_unregister_irq_mask_notifier(struct kvm *kvm, int irq,
|
|
||||||
struct kvm_irq_mask_notifier *kimn);
|
|
||||||
void kvm_fire_mask_notifiers(struct kvm *kvm, unsigned irqchip, unsigned pin,
|
|
||||||
bool mask);
|
|
||||||
|
|
||||||
int kvm_irq_map_gsi(struct kvm *kvm,
|
int kvm_irq_map_gsi(struct kvm *kvm,
|
||||||
struct kvm_kernel_irq_routing_entry *entries, int gsi);
|
struct kvm_kernel_irq_routing_entry *entries, int gsi);
|
||||||
int kvm_irq_map_chip_pin(struct kvm *kvm, unsigned irqchip, unsigned pin);
|
int kvm_irq_map_chip_pin(struct kvm *kvm, unsigned irqchip, unsigned pin);
|
||||||
|
@ -36,9 +36,6 @@
|
|||||||
#include <linux/seqlock.h>
|
#include <linux/seqlock.h>
|
||||||
#include <trace/events/kvm.h>
|
#include <trace/events/kvm.h>
|
||||||
|
|
||||||
#ifdef __KVM_HAVE_IOAPIC
|
|
||||||
#include "ioapic.h"
|
|
||||||
#endif
|
|
||||||
#include "iodev.h"
|
#include "iodev.h"
|
||||||
|
|
||||||
#ifdef CONFIG_HAVE_KVM_IRQFD
|
#ifdef CONFIG_HAVE_KVM_IRQFD
|
||||||
@ -492,9 +489,7 @@ void kvm_register_irq_ack_notifier(struct kvm *kvm,
|
|||||||
mutex_lock(&kvm->irq_lock);
|
mutex_lock(&kvm->irq_lock);
|
||||||
hlist_add_head_rcu(&kian->link, &kvm->irq_ack_notifier_list);
|
hlist_add_head_rcu(&kian->link, &kvm->irq_ack_notifier_list);
|
||||||
mutex_unlock(&kvm->irq_lock);
|
mutex_unlock(&kvm->irq_lock);
|
||||||
#ifdef __KVM_HAVE_IOAPIC
|
|
||||||
kvm_vcpu_request_scan_ioapic(kvm);
|
kvm_vcpu_request_scan_ioapic(kvm);
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
|
void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
|
||||||
@ -504,9 +499,7 @@ void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
|
|||||||
hlist_del_init_rcu(&kian->link);
|
hlist_del_init_rcu(&kian->link);
|
||||||
mutex_unlock(&kvm->irq_lock);
|
mutex_unlock(&kvm->irq_lock);
|
||||||
synchronize_srcu(&kvm->irq_srcu);
|
synchronize_srcu(&kvm->irq_srcu);
|
||||||
#ifdef __KVM_HAVE_IOAPIC
|
|
||||||
kvm_vcpu_request_scan_ioapic(kvm);
|
kvm_vcpu_request_scan_ioapic(kvm);
|
||||||
#endif
|
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -468,9 +468,6 @@ static struct kvm *kvm_create_vm(unsigned long type)
|
|||||||
if (r)
|
if (r)
|
||||||
goto out_err_no_disable;
|
goto out_err_no_disable;
|
||||||
|
|
||||||
#ifdef CONFIG_HAVE_KVM_IRQCHIP
|
|
||||||
INIT_HLIST_HEAD(&kvm->mask_notifier_list);
|
|
||||||
#endif
|
|
||||||
#ifdef CONFIG_HAVE_KVM_IRQFD
|
#ifdef CONFIG_HAVE_KVM_IRQFD
|
||||||
INIT_HLIST_HEAD(&kvm->irq_ack_notifier_list);
|
INIT_HLIST_HEAD(&kvm->irq_ack_notifier_list);
|
||||||
#endif
|
#endif
|
||||||
|
Loading…
x
Reference in New Issue
Block a user