2711b8e2b7
The new AP state tracking and synchronization mechanism in the CPU hotplug core code allows to remove quite some x86 specific code: 1) The AP alive synchronization based on cpumasks 2) The decision whether an AP can be brought up again Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Tested-by: Michael Kelley <mikelley@microsoft.com> Tested-by: Oleksandr Natalenko <oleksandr@natalenko.name> Tested-by: Helge Deller <deller@gmx.de> # parisc Tested-by: Guilherme G. Piccoli <gpiccoli@igalia.com> # Steam Deck Link: https://lore.kernel.org/r/20230512205256.529657366@linutronix.de
91 lines
2.0 KiB
C
91 lines
2.0 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/thread_info.h>
|
|
#include <asm/smp.h>
|
|
|
|
#include <xen/events.h>
|
|
|
|
#include "xen-ops.h"
|
|
#include "smp.h"
|
|
|
|
|
|
static void __init xen_hvm_smp_prepare_boot_cpu(void)
|
|
{
|
|
BUG_ON(smp_processor_id() != 0);
|
|
native_smp_prepare_boot_cpu();
|
|
|
|
/*
|
|
* Setup vcpu_info for boot CPU. Secondary CPUs get their vcpu_info
|
|
* in xen_cpu_up_prepare_hvm().
|
|
*/
|
|
xen_vcpu_setup(0);
|
|
|
|
/*
|
|
* Called again in case the kernel boots on vcpu >= MAX_VIRT_CPUS.
|
|
* Refer to comments in xen_hvm_init_time_ops().
|
|
*/
|
|
xen_hvm_init_time_ops();
|
|
|
|
/*
|
|
* The alternative logic (which patches the unlock/lock) runs before
|
|
* the smp bootup up code is activated. Hence we need to set this up
|
|
* the core kernel is being patched. Otherwise we will have only
|
|
* modules patched but not core code.
|
|
*/
|
|
xen_init_spinlocks();
|
|
}
|
|
|
|
static void __init xen_hvm_smp_prepare_cpus(unsigned int max_cpus)
|
|
{
|
|
int cpu;
|
|
|
|
native_smp_prepare_cpus(max_cpus);
|
|
|
|
if (xen_have_vector_callback) {
|
|
WARN_ON(xen_smp_intr_init(0));
|
|
xen_init_lock_cpu(0);
|
|
}
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
if (cpu == 0)
|
|
continue;
|
|
|
|
/* Set default vcpu_id to make sure that we don't use cpu-0's */
|
|
per_cpu(xen_vcpu_id, cpu) = XEN_VCPU_ID_INVALID;
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
static void xen_hvm_cleanup_dead_cpu(unsigned int cpu)
|
|
{
|
|
if (xen_have_vector_callback) {
|
|
xen_smp_intr_free(cpu);
|
|
xen_uninit_lock_cpu(cpu);
|
|
xen_teardown_timer(cpu);
|
|
}
|
|
}
|
|
#else
|
|
static void xen_hvm_cleanup_dead_cpu(unsigned int cpu)
|
|
{
|
|
BUG();
|
|
}
|
|
#endif
|
|
|
|
void __init xen_hvm_smp_init(void)
|
|
{
|
|
smp_ops.smp_prepare_boot_cpu = xen_hvm_smp_prepare_boot_cpu;
|
|
smp_ops.smp_prepare_cpus = xen_hvm_smp_prepare_cpus;
|
|
smp_ops.smp_cpus_done = xen_smp_cpus_done;
|
|
smp_ops.cleanup_dead_cpu = xen_hvm_cleanup_dead_cpu;
|
|
|
|
if (!xen_have_vector_callback) {
|
|
#ifdef CONFIG_PARAVIRT_SPINLOCKS
|
|
nopvspin = true;
|
|
#endif
|
|
return;
|
|
}
|
|
|
|
smp_ops.smp_send_reschedule = xen_smp_send_reschedule;
|
|
smp_ops.send_call_func_ipi = xen_smp_send_call_function_ipi;
|
|
smp_ops.send_call_func_single_ipi = xen_smp_send_call_function_single_ipi;
|
|
}
|