2019-06-03 08:44:50 +03:00
// SPDX-License-Identifier: GPL-2.0-only
2012-03-05 15:49:30 +04:00
/*
* SMP initialisation and IPI support
* Based on arch / arm / kernel / smp . c
*
* Copyright ( C ) 2012 ARM Ltd .
*/
2015-05-13 16:12:47 +03:00
# include <linux/acpi.h>
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
# include <linux/arm_sdei.h>
2012-03-05 15:49:30 +04:00
# include <linux/delay.h>
# include <linux/init.h>
# include <linux/spinlock.h>
2017-02-01 21:08:20 +03:00
# include <linux/sched/mm.h>
2017-02-08 20:51:36 +03:00
# include <linux/sched/hotplug.h>
2017-02-08 20:51:37 +03:00
# include <linux/sched/task_stack.h>
2012-03-05 15:49:30 +04:00
# include <linux/interrupt.h>
# include <linux/cache.h>
# include <linux/profile.h>
# include <linux/errno.h>
# include <linux/mm.h>
# include <linux/err.h>
# include <linux/cpu.h>
# include <linux/smp.h>
# include <linux/seq_file.h>
# include <linux/irq.h>
2019-01-31 17:58:55 +03:00
# include <linux/irqchip/arm-gic-v3.h>
2012-03-05 15:49:30 +04:00
# include <linux/percpu.h>
# include <linux/clockchips.h>
# include <linux/completion.h>
# include <linux/of.h>
2014-05-12 19:48:51 +04:00
# include <linux/irq_work.h>
2020-06-20 19:19:00 +03:00
# include <linux/kernel_stat.h>
2017-04-03 05:24:36 +03:00
# include <linux/kexec.h>
2019-11-28 22:58:05 +03:00
# include <linux/kvm_host.h>
2012-03-05 15:49:30 +04:00
2014-11-14 18:54:08 +03:00
# include <asm/alternative.h>
2012-03-05 15:49:30 +04:00
# include <asm/atomic.h>
# include <asm/cacheflush.h>
2014-07-16 19:32:44 +04:00
# include <asm/cpu.h>
2012-03-05 15:49:30 +04:00
# include <asm/cputype.h>
2013-10-24 23:30:15 +04:00
# include <asm/cpu_ops.h>
2017-11-02 15:12:34 +03:00
# include <asm/daifflags.h>
2019-11-28 22:58:05 +03:00
# include <asm/kvm_mmu.h>
2012-03-05 15:49:30 +04:00
# include <asm/mmu_context.h>
2016-04-09 01:50:27 +03:00
# include <asm/numa.h>
2012-03-05 15:49:30 +04:00
# include <asm/processor.h>
2012-08-29 12:47:19 +04:00
# include <asm/smp_plat.h>
2012-03-05 15:49:30 +04:00
# include <asm/sections.h>
# include <asm/tlbflush.h>
# include <asm/ptrace.h>
2015-07-29 14:07:57 +03:00
# include <asm/virt.h>
2012-03-05 15:49:30 +04:00
2014-07-26 00:05:32 +04:00
# include <trace/events/ipi.h>
arm64: make cpu number a percpu variable
In the absence of CONFIG_THREAD_INFO_IN_TASK, core code maintains
thread_info::cpu, and low-level architecture code can access this to
build raw_smp_processor_id(). With CONFIG_THREAD_INFO_IN_TASK, core code
maintains task_struct::cpu, which for reasons of hte header soup is not
accessible to low-level arch code.
Instead, we can maintain a percpu variable containing the cpu number.
For both the old and new implementation of raw_smp_processor_id(), we
read a syreg into a GPR, add an offset, and load the result. As the
offset is now larger, it may not be folded into the load, but otherwise
the assembly shouldn't change much.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Laura Abbott <labbott@redhat.com>
Cc: James Morse <james.morse@arm.com>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-11-03 23:23:11 +03:00
DEFINE_PER_CPU_READ_MOSTLY ( int , cpu_number ) ;
EXPORT_PER_CPU_SYMBOL ( cpu_number ) ;
2012-03-05 15:49:30 +04:00
/*
* as from 2.5 , kernels no longer have an init_tasks structure
* so we need some other way of telling a new secondary core
* where to place its SVC stack
*/
struct secondary_data secondary_data ;
2016-02-23 13:31:42 +03:00
/* Number of CPUs which aren't online, but looping in kernel text. */
2020-04-23 09:33:26 +03:00
static int cpus_stuck_in_kernel ;
2012-03-05 15:49:30 +04:00
enum ipi_msg_type {
IPI_RESCHEDULE ,
IPI_CALL_FUNC ,
IPI_CPU_STOP ,
2017-04-03 05:24:36 +03:00
IPI_CPU_CRASH_STOP ,
2013-09-04 13:55:17 +04:00
IPI_TIMER ,
2014-05-12 19:48:51 +04:00
IPI_IRQ_WORK ,
2020-06-20 19:19:00 +03:00
IPI_WAKEUP ,
NR_IPI
2012-03-05 15:49:30 +04:00
} ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
static int ipi_irq_base __read_mostly ;
static int nr_ipi __read_mostly = NR_IPI ;
static struct irq_desc * ipi_desc [ NR_IPI ] __read_mostly ;
static void ipi_setup ( int cpu ) ;
2016-02-23 13:31:42 +03:00
# ifdef CONFIG_HOTPLUG_CPU
2020-09-18 15:33:18 +03:00
static void ipi_teardown ( int cpu ) ;
2016-02-23 13:31:42 +03:00
static int op_cpu_kill ( unsigned int cpu ) ;
# else
static inline int op_cpu_kill ( unsigned int cpu )
{
return - ENOSYS ;
}
# endif
2012-03-05 15:49:30 +04:00
/*
* Boot a secondary CPU , and assign it the specified idle task .
* This also gives us the initial stack to use for this CPU .
*/
2013-06-18 18:18:31 +04:00
static int boot_secondary ( unsigned int cpu , struct task_struct * idle )
2012-03-05 15:49:30 +04:00
{
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops = get_cpu_ops ( cpu ) ;
if ( ops - > cpu_boot )
return ops - > cpu_boot ( cpu ) ;
2012-03-05 15:49:30 +04:00
arm64: factor out spin-table boot method
The arm64 kernel has an internal holding pen, which is necessary for
some systems where we can't bring CPUs online individually and must hold
multiple CPUs in a safe area until the kernel is able to handle them.
The current SMP infrastructure for arm64 is closely coupled to this
holding pen, and alternative boot methods must launch CPUs into the pen,
where they sit before they are launched into the kernel proper.
With PSCI (and possibly other future boot methods), we can bring CPUs
online individually, and need not perform the secondary_holding_pen
dance. Instead, this patch factors the holding pen management code out
to the spin-table boot method code, as it is the only boot method
requiring the pen.
A new entry point for secondaries, secondary_entry is added for other
boot methods to use, which bypasses the holding pen and its associated
overhead when bringing CPUs online. The smp.pen.text section is also
removed, as the pen can live in head.text without problem.
The cpu_operations structure is extended with two new functions,
cpu_boot and cpu_postboot, for bringing a cpu into the kernel and
performing any post-boot cleanup required by a bootmethod (e.g.
resetting the secondary_holding_pen_release to INVALID_HWID).
Documentation is added for cpu_operations.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2013-10-24 23:30:16 +04:00
return - EOPNOTSUPP ;
2012-03-05 15:49:30 +04:00
}
static DECLARE_COMPLETION ( cpu_running ) ;
2013-06-18 18:18:31 +04:00
int __cpu_up ( unsigned int cpu , struct task_struct * idle )
2012-03-05 15:49:30 +04:00
{
int ret ;
2016-02-23 13:31:42 +03:00
long status ;
2012-03-05 15:49:30 +04:00
/*
* We need to tell the secondary core where to find its stack and the
* page tables .
*/
arm64: split thread_info from task stack
This patch moves arm64's struct thread_info from the task stack into
task_struct. This protects thread_info from corruption in the case of
stack overflows, and makes its address harder to determine if stack
addresses are leaked, making a number of attacks more difficult. Precise
detection and handling of overflow is left for subsequent patches.
Largely, this involves changing code to store the task_struct in sp_el0,
and acquire the thread_info from the task struct. Core code now
implements current_thread_info(), and as noted in <linux/sched.h> this
relies on offsetof(task_struct, thread_info) == 0, enforced by core
code.
This change means that the 'tsk' register used in entry.S now points to
a task_struct, rather than a thread_info as it used to. To make this
clear, the TI_* field offsets are renamed to TSK_TI_*, with asm-offsets
appropriately updated to account for the structural change.
Userspace clobbers sp_el0, and we can no longer restore this from the
stack. Instead, the current task is cached in a per-cpu variable that we
can safely access from early assembly as interrupts are disabled (and we
are thus not preemptible).
Both secondary entry and idle are updated to stash the sp and task
pointer separately.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Laura Abbott <labbott@redhat.com>
Cc: AKASHI Takahiro <takahiro.akashi@linaro.org>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: James Morse <james.morse@arm.com>
Cc: Kees Cook <keescook@chromium.org>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-11-03 23:23:13 +03:00
secondary_data . task = idle ;
2016-02-23 13:31:42 +03:00
update_cpu_boot_status ( CPU_MMU_OFF ) ;
2012-03-05 15:49:30 +04:00
2020-03-02 05:03:40 +03:00
/* Now bring the CPU into our world */
2012-03-05 15:49:30 +04:00
ret = boot_secondary ( cpu , idle ) ;
2020-03-02 05:03:40 +03:00
if ( ret ) {
2012-03-05 15:49:30 +04:00
pr_err ( " CPU%u: failed to boot: %d \n " , cpu , ret ) ;
2018-12-10 21:07:33 +03:00
return ret ;
2012-03-05 15:49:30 +04:00
}
2020-03-02 05:03:40 +03:00
/*
* CPU was successfully started , wait for it to come online or
* time out .
*/
wait_for_completion_timeout ( & cpu_running ,
msecs_to_jiffies ( 5000 ) ) ;
if ( cpu_online ( cpu ) )
return 0 ;
pr_crit ( " CPU%u: failed to come online \n " , cpu ) ;
arm64: split thread_info from task stack
This patch moves arm64's struct thread_info from the task stack into
task_struct. This protects thread_info from corruption in the case of
stack overflows, and makes its address harder to determine if stack
addresses are leaked, making a number of attacks more difficult. Precise
detection and handling of overflow is left for subsequent patches.
Largely, this involves changing code to store the task_struct in sp_el0,
and acquire the thread_info from the task struct. Core code now
implements current_thread_info(), and as noted in <linux/sched.h> this
relies on offsetof(task_struct, thread_info) == 0, enforced by core
code.
This change means that the 'tsk' register used in entry.S now points to
a task_struct, rather than a thread_info as it used to. To make this
clear, the TI_* field offsets are renamed to TSK_TI_*, with asm-offsets
appropriately updated to account for the structural change.
Userspace clobbers sp_el0, and we can no longer restore this from the
stack. Instead, the current task is cached in a per-cpu variable that we
can safely access from early assembly as interrupts are disabled (and we
are thus not preemptible).
Both secondary entry and idle are updated to stash the sp and task
pointer separately.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Laura Abbott <labbott@redhat.com>
Cc: AKASHI Takahiro <takahiro.akashi@linaro.org>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: James Morse <james.morse@arm.com>
Cc: Kees Cook <keescook@chromium.org>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-11-03 23:23:13 +03:00
secondary_data . task = NULL ;
2016-02-23 13:31:42 +03:00
status = READ_ONCE ( secondary_data . status ) ;
2020-03-02 05:03:40 +03:00
if ( status = = CPU_MMU_OFF )
status = READ_ONCE ( __early_cpu_boot_status ) ;
2016-02-23 13:31:42 +03:00
2020-03-02 05:03:40 +03:00
switch ( status & CPU_BOOT_STATUS_MASK ) {
default :
pr_err ( " CPU%u: failed in unknown state : 0x%lx \n " ,
cpu , status ) ;
cpus_stuck_in_kernel + + ;
break ;
case CPU_KILL_ME :
if ( ! op_cpu_kill ( cpu ) ) {
pr_crit ( " CPU%u: died during early boot \n " , cpu ) ;
2016-02-23 13:31:42 +03:00
break ;
}
2020-03-02 05:03:40 +03:00
pr_crit ( " CPU%u: may not have shut down cleanly \n " , cpu ) ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2020-03-02 05:03:40 +03:00
case CPU_STUCK_IN_KERNEL :
pr_crit ( " CPU%u: is stuck in kernel \n " , cpu ) ;
if ( status & CPU_STUCK_REASON_52_BIT_VA )
pr_crit ( " CPU%u: does not support 52-bit VAs \n " , cpu ) ;
if ( status & CPU_STUCK_REASON_NO_GRAN ) {
pr_crit ( " CPU%u: does not support %luK granule \n " ,
cpu , PAGE_SIZE / SZ_1K ) ;
}
cpus_stuck_in_kernel + + ;
break ;
case CPU_PANIC_KERNEL :
panic ( " CPU%u detected unsupported configuration \n " , cpu ) ;
2016-02-23 13:31:42 +03:00
}
2012-03-05 15:49:30 +04:00
2020-05-28 02:34:57 +03:00
return - EIO ;
2012-03-05 15:49:30 +04:00
}
2019-01-31 17:58:55 +03:00
static void init_gic_priority_masking ( void )
{
u32 cpuflags ;
if ( WARN_ON ( ! gic_enable_sre ( ) ) )
return ;
cpuflags = read_sysreg ( daif ) ;
WARN_ON ( ! ( cpuflags & PSR_I_BIT ) ) ;
2021-03-15 14:56:28 +03:00
WARN_ON ( ! ( cpuflags & PSR_F_BIT ) ) ;
2019-01-31 17:58:55 +03:00
2019-06-11 12:38:12 +03:00
gic_write_pmr ( GIC_PRIO_IRQON | GIC_PRIO_PSR_I_SET ) ;
2019-01-31 17:58:55 +03:00
}
2012-03-05 15:49:30 +04:00
/*
* This is the secondary CPU boot entry . We ' re using this CPUs
* idle thread stack , but a set of temporary page tables .
*/
2018-06-12 12:07:37 +03:00
asmlinkage notrace void secondary_start_kernel ( void )
2012-03-05 15:49:30 +04:00
{
2017-09-27 16:50:38 +03:00
u64 mpidr = read_cpuid_mpidr ( ) & MPIDR_HWID_BITMASK ;
2012-03-05 15:49:30 +04:00
struct mm_struct * mm = & init_mm ;
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops ;
2021-05-20 14:50:31 +03:00
unsigned int cpu = smp_processor_id ( ) ;
2012-03-05 15:49:30 +04:00
/*
* All kernel threads share the same mm context ; grab a
* reference and switch to it .
*/
2017-02-28 01:30:07 +03:00
mmgrab ( mm ) ;
2012-03-05 15:49:30 +04:00
current - > active_mm = mm ;
/*
* TTBR0 is only used for the identity mapping at this stage . Make it
* point to zero page to avoid speculatively fetching new entries .
*/
2016-01-25 14:44:58 +03:00
cpu_uninstall_idmap ( ) ;
2012-03-05 15:49:30 +04:00
2019-01-31 17:58:55 +03:00
if ( system_uses_irq_prio_masking ( ) )
init_gic_priority_masking ( ) ;
2020-10-28 21:26:14 +03:00
rcu_cpu_starting ( cpu ) ;
2012-03-05 15:49:30 +04:00
trace_hardirqs_off ( ) ;
arm64: Delay cpu feature capability checks
At the moment we run through the arm64_features capability list for
each CPU and set the capability if one of the CPU supports it. This
could be problematic in a heterogeneous system with differing capabilities.
Delay the CPU feature checks until all the enabled CPUs are up(i.e,
smp_cpus_done(), so that we can make better decisions based on the
overall system capability. Once we decide and advertise the capabilities
the alternatives can be applied. From this state, we cannot roll back
a feature to disabled based on the values from a new hotplugged CPU,
due to the runtime patching and other reasons. So, for all new CPUs,
we need to make sure that they have the established system capabilities.
Failing which, we bring the CPU down, preventing it from turning online.
Once the capabilities are decided, any new CPU booting up goes through
verification to ensure that it has all the enabled capabilities and also
invokes the respective enable() method on the CPU.
The CPU errata checks are not delayed and is still executed per-CPU
to detect the respective capabilities. If we ever come across a non-errata
capability that needs to be checked on each-CPU, we could introduce them via
a new capability table(or introduce a flag), which can be processed per CPU.
The next patch will make the feature checks use the system wide
safe value of a feature register.
NOTE: The enable() methods associated with the capability is scheduled
on all the CPUs (which is the only use case at the moment). If we need
a different type of 'enable()' which only needs to be run once on any CPU,
we should be able to handle that when needed.
Signed-off-by: Suzuki K. Poulose <suzuki.poulose@arm.com>
Tested-by: Dave Martin <Dave.Martin@arm.com>
[catalin.marinas@arm.com: static variable and coding style fixes]
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2015-10-19 16:24:50 +03:00
/*
* If the system has established the capabilities , make sure
* this CPU ticks all of those . If it doesn ' t , the CPU will
* fail to come online .
*/
2016-09-09 16:07:10 +03:00
check_local_cpu_capabilities ( ) ;
arm64: Delay cpu feature capability checks
At the moment we run through the arm64_features capability list for
each CPU and set the capability if one of the CPU supports it. This
could be problematic in a heterogeneous system with differing capabilities.
Delay the CPU feature checks until all the enabled CPUs are up(i.e,
smp_cpus_done(), so that we can make better decisions based on the
overall system capability. Once we decide and advertise the capabilities
the alternatives can be applied. From this state, we cannot roll back
a feature to disabled based on the values from a new hotplugged CPU,
due to the runtime patching and other reasons. So, for all new CPUs,
we need to make sure that they have the established system capabilities.
Failing which, we bring the CPU down, preventing it from turning online.
Once the capabilities are decided, any new CPU booting up goes through
verification to ensure that it has all the enabled capabilities and also
invokes the respective enable() method on the CPU.
The CPU errata checks are not delayed and is still executed per-CPU
to detect the respective capabilities. If we ever come across a non-errata
capability that needs to be checked on each-CPU, we could introduce them via
a new capability table(or introduce a flag), which can be processed per CPU.
The next patch will make the feature checks use the system wide
safe value of a feature register.
NOTE: The enable() methods associated with the capability is scheduled
on all the CPUs (which is the only use case at the moment). If we need
a different type of 'enable()' which only needs to be run once on any CPU,
we should be able to handle that when needed.
Signed-off-by: Suzuki K. Poulose <suzuki.poulose@arm.com>
Tested-by: Dave Martin <Dave.Martin@arm.com>
[catalin.marinas@arm.com: static variable and coding style fixes]
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2015-10-19 16:24:50 +03:00
2020-03-19 02:01:44 +03:00
ops = get_cpu_ops ( cpu ) ;
if ( ops - > cpu_postboot )
ops - > cpu_postboot ( ) ;
2012-03-05 15:49:30 +04:00
2014-07-16 19:32:44 +04:00
/*
* Log the CPU info before it is marked online and might get read .
*/
cpuinfo_store_cpu ( ) ;
2022-03-31 18:39:26 +03:00
store_cpu_topology ( cpu ) ;
2014-07-16 19:32:44 +04:00
2013-11-04 20:55:22 +04:00
/*
* Enable GIC and timers .
*/
notify_cpu_starting ( cpu ) ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
ipi_setup ( cpu ) ;
2018-07-06 14:02:43 +03:00
numa_add_cpu ( cpu ) ;
2014-03-04 11:51:17 +04:00
2012-03-05 15:49:30 +04:00
/*
* OK , now it ' s safe to let the boot CPU continue . Wait for
* the CPU migration code to notice that the CPU is online
* before we continue .
*/
2017-09-27 16:50:38 +03:00
pr_info ( " CPU%u: Booted secondary processor 0x%010lx [0x%08x] \n " ,
cpu , ( unsigned long ) mpidr ,
read_cpuid_id ( ) ) ;
2016-02-23 13:31:42 +03:00
update_cpu_boot_status ( CPU_BOOT_SUCCESS ) ;
2012-03-05 15:49:30 +04:00
set_cpu_online ( cpu , true ) ;
2012-11-07 21:00:05 +04:00
complete ( & cpu_running ) ;
2012-03-05 15:49:30 +04:00
2017-11-02 15:12:36 +03:00
local_daif_restore ( DAIF_PROCCTX ) ;
2013-07-19 18:08:15 +04:00
2012-03-05 15:49:30 +04:00
/*
* OK , it ' s off to the idle thread for us
*/
2016-02-26 21:43:40 +03:00
cpu_startup_entry ( CPUHP_AP_ONLINE_IDLE ) ;
2012-03-05 15:49:30 +04:00
}
2013-10-24 23:30:18 +04:00
# ifdef CONFIG_HOTPLUG_CPU
static int op_cpu_disable ( unsigned int cpu )
{
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops = get_cpu_ops ( cpu ) ;
2013-10-24 23:30:18 +04:00
/*
* If we don ' t have a cpu_die method , abort before we reach the point
* of no return . CPU0 may not have an cpu_ops , so test for it .
*/
2020-03-19 02:01:44 +03:00
if ( ! ops | | ! ops - > cpu_die )
2013-10-24 23:30:18 +04:00
return - EOPNOTSUPP ;
/*
* We may need to abort a hot unplug for some other mechanism - specific
* reason .
*/
2020-03-19 02:01:44 +03:00
if ( ops - > cpu_disable )
return ops - > cpu_disable ( cpu ) ;
2013-10-24 23:30:18 +04:00
return 0 ;
}
/*
* __cpu_disable runs on the processor to be shutdown .
*/
int __cpu_disable ( void )
{
unsigned int cpu = smp_processor_id ( ) ;
int ret ;
ret = op_cpu_disable ( cpu ) ;
if ( ret )
return ret ;
2018-07-06 14:02:46 +03:00
remove_cpu_topology ( cpu ) ;
numa_remove_cpu ( cpu ) ;
2013-10-24 23:30:18 +04:00
/*
* Take this CPU offline . Once we clear this , we can ' t return ,
* and we must not schedule until we ' re ready to give up the cpu .
*/
set_cpu_online ( cpu , false ) ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
ipi_teardown ( cpu ) ;
2013-10-24 23:30:18 +04:00
/*
* OK - migrate IRQs away from this CPU
*/
2015-09-24 12:32:14 +03:00
irq_migrate_all_off_this_cpu ( ) ;
2013-10-24 23:30:18 +04:00
return 0 ;
}
2014-05-07 18:18:36 +04:00
static int op_cpu_kill ( unsigned int cpu )
{
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops = get_cpu_ops ( cpu ) ;
2014-05-07 18:18:36 +04:00
/*
* If we have no means of synchronising with the dying CPU , then assume
* that it is really dead . We can only wait for an arbitrary length of
* time and hope that it ' s dead , so let ' s skip the wait and just hope .
*/
2020-03-19 02:01:44 +03:00
if ( ! ops - > cpu_kill )
2015-04-20 19:55:30 +03:00
return 0 ;
2014-05-07 18:18:36 +04:00
2020-03-19 02:01:44 +03:00
return ops - > cpu_kill ( cpu ) ;
2014-05-07 18:18:36 +04:00
}
2013-10-24 23:30:18 +04:00
/*
2023-05-13 00:07:33 +03:00
* Called on the thread which is asking for a CPU to be shutdown after the
* shutdown completed .
2013-10-24 23:30:18 +04:00
*/
2023-05-13 00:07:33 +03:00
void arch_cpuhp_cleanup_dead_cpu ( unsigned int cpu )
2013-10-24 23:30:18 +04:00
{
2015-04-20 19:55:30 +03:00
int err ;
2021-06-17 10:30:59 +03:00
pr_debug ( " CPU%u: shutdown \n " , cpu ) ;
2014-05-07 18:18:36 +04:00
/*
* Now that the dying CPU is beyond the point of no return w . r . t .
* in - kernel synchronisation , try to get the firwmare to help us to
* verify that it has really left the kernel before we consider
* clobbering anything it might still be using .
*/
2015-04-20 19:55:30 +03:00
err = op_cpu_kill ( cpu ) ;
if ( err )
2019-10-18 06:18:19 +03:00
pr_warn ( " CPU%d may not have shut down cleanly: %d \n " , cpu , err ) ;
2013-10-24 23:30:18 +04:00
}
/*
* Called from the idle thread for the CPU which has been shutdown .
*
*/
2023-02-16 21:42:01 +03:00
void __noreturn cpu_die ( void )
2013-10-24 23:30:18 +04:00
{
unsigned int cpu = smp_processor_id ( ) ;
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops = get_cpu_ops ( cpu ) ;
2013-10-24 23:30:18 +04:00
idle_task_exit ( ) ;
2017-11-02 15:12:34 +03:00
local_daif_mask ( ) ;
2013-10-24 23:30:18 +04:00
2023-05-13 00:07:33 +03:00
/* Tell cpuhp_bp_sync_dead() that this CPU is now safe to dispose of */
cpuhp_ap_report_dead ( ) ;
2013-10-24 23:30:18 +04:00
/*
* Actually shutdown the CPU . This must never fail . The specific hotplug
* mechanism must perform all required cache maintenance to ensure that
* no dirty lines are lost in the process of shutting down the CPU .
*/
2020-03-19 02:01:44 +03:00
ops - > cpu_die ( cpu ) ;
2013-10-24 23:30:18 +04:00
BUG ( ) ;
}
# endif
2020-03-19 02:01:44 +03:00
static void __cpu_try_die ( int cpu )
{
# ifdef CONFIG_HOTPLUG_CPU
const struct cpu_operations * ops = get_cpu_ops ( cpu ) ;
if ( ops & & ops - > cpu_die )
ops - > cpu_die ( cpu ) ;
# endif
}
2016-02-23 13:31:41 +03:00
/*
* Kill the calling secondary CPU , early in bringup before it is turned
* online .
*/
2023-04-13 02:49:34 +03:00
void __noreturn cpu_die_early ( void )
2016-02-23 13:31:41 +03:00
{
int cpu = smp_processor_id ( ) ;
pr_crit ( " CPU%d: will not boot \n " , cpu ) ;
/* Mark this CPU absent */
set_cpu_present ( cpu , 0 ) ;
2020-11-06 13:25:49 +03:00
rcu_report_dead ( cpu ) ;
2016-02-23 13:31:41 +03:00
2020-03-19 02:01:44 +03:00
if ( IS_ENABLED ( CONFIG_HOTPLUG_CPU ) ) {
update_cpu_boot_status ( CPU_KILL_ME ) ;
__cpu_try_die ( cpu ) ;
}
2016-02-23 13:31:42 +03:00
update_cpu_boot_status ( CPU_STUCK_IN_KERNEL ) ;
2016-02-23 13:31:41 +03:00
cpu_park_loop ( ) ;
}
2015-07-29 14:07:57 +03:00
static void __init hyp_mode_check ( void )
{
if ( is_hyp_mode_available ( ) )
pr_info ( " CPU: All CPU(s) started at EL2 \n " ) ;
else if ( is_hyp_mode_mismatched ( ) )
WARN_TAINT ( 1 , TAINT_CPU_OUT_OF_SPEC ,
" CPU: CPUs started in inconsistent modes " ) ;
else
pr_info ( " CPU: All CPU(s) started at EL1 \n " ) ;
2021-01-05 21:05:38 +03:00
if ( IS_ENABLED ( CONFIG_KVM ) & & ! is_kernel_in_hyp_mode ( ) ) {
2019-11-28 22:58:05 +03:00
kvm_compute_layout ( ) ;
2021-01-05 21:05:38 +03:00
kvm_apply_hyp_relocations ( ) ;
}
2015-07-29 14:07:57 +03:00
}
2012-03-05 15:49:30 +04:00
void __init smp_cpus_done ( unsigned int max_cpus )
{
2013-08-30 21:06:48 +04:00
pr_info ( " SMP: Total of %d processors activated. \n " , num_online_cpus ( ) ) ;
2015-10-19 16:24:39 +03:00
setup_cpu_features ( ) ;
2015-07-29 14:07:57 +03:00
hyp_mode_check ( ) ;
apply_alternatives_all ( ) ;
2017-03-09 23:52:01 +03:00
mark_linear_text_alias_ro ( ) ;
2012-03-05 15:49:30 +04:00
}
void __init smp_prepare_boot_cpu ( void )
{
2021-05-20 14:50:31 +03:00
/*
* The runtime per - cpu areas have been allocated by
* setup_per_cpu_areas ( ) , and CPU0 ' s boot time per - cpu area will be
* freed shortly , so we must move over to the runtime per - cpu area .
*/
2016-07-21 13:12:55 +03:00
set_my_cpu_offset ( per_cpu_offset ( smp_processor_id ( ) ) ) ;
2015-10-19 16:24:40 +03:00
cpuinfo_store_boot_cpu ( ) ;
2019-01-31 17:58:53 +03:00
/*
* We now know enough about the boot CPU to apply the
* alternatives that cannot wait until interrupt handling
* and / or scheduling is enabled .
*/
apply_boot_alternatives ( ) ;
2019-01-31 17:58:55 +03:00
/* Conditionally switch to GIC PMR for interrupt masking */
if ( system_uses_irq_prio_masking ( ) )
init_gic_priority_masking ( ) ;
2020-12-22 23:02:10 +03:00
kasan_init_hw_tags ( ) ;
2012-03-05 15:49:30 +04:00
}
2015-05-13 16:12:47 +03:00
/*
* Duplicate MPIDRs are a recipe for disaster . Scan all initialized
* entries and check for duplicates . If any is found just ignore the
* cpu . cpu_logical_map was initialized to INVALID_HWID to avoid
* matching valid MPIDR values .
*/
static bool __init is_mpidr_duplicate ( unsigned int cpu , u64 hwid )
{
unsigned int i ;
for ( i = 1 ; ( i < cpu ) & & ( i < NR_CPUS ) ; i + + )
if ( cpu_logical_map ( i ) = = hwid )
return true ;
return false ;
}
2015-05-13 16:12:46 +03:00
/*
* Initialize cpu operations for a logical cpu and
* set it in the possible mask on success
*/
static int __init smp_cpu_setup ( int cpu )
{
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops ;
2020-03-19 02:01:43 +03:00
if ( init_cpu_ops ( cpu ) )
2015-05-13 16:12:46 +03:00
return - ENODEV ;
2020-03-19 02:01:44 +03:00
ops = get_cpu_ops ( cpu ) ;
if ( ops - > cpu_init ( cpu ) )
2015-05-13 16:12:46 +03:00
return - ENODEV ;
set_cpu_possible ( cpu , true ) ;
return 0 ;
}
2015-05-13 16:12:47 +03:00
static bool bootcpu_valid __initdata ;
static unsigned int cpu_count = 1 ;
# ifdef CONFIG_ACPI
2017-04-11 11:39:54 +03:00
static struct acpi_madt_generic_interrupt cpu_madt_gicc [ NR_CPUS ] ;
struct acpi_madt_generic_interrupt * acpi_cpu_get_madt_gicc ( int cpu )
{
return & cpu_madt_gicc [ cpu ] ;
}
cpufreq: CPPC: Add per_cpu efficiency_class
In ACPI, describing power efficiency of CPUs can be done through the
following arm specific field:
ACPI 6.4, s5.2.12.14 'GIC CPU Interface (GICC) Structure',
'Processor Power Efficiency Class field':
Describes the relative power efficiency of the associated pro-
cessor. Lower efficiency class numbers are more efficient than
higher ones (e.g. efficiency class 0 should be treated as more
efficient than efficiency class 1). However, absolute values
of this number have no meaning: 2 isn’t necessarily half as
efficient as 1.
The efficiency_class field is stored in the GicC structure of the
ACPI MADT table and it's currently supported in Linux for arm64 only.
Thus, this new functionality is introduced for arm64 only.
To allow the cppc_cpufreq driver to know and preprocess the
efficiency_class values of all the CPUs, add a per_cpu efficiency_class
variable to store them.
At least 2 different efficiency classes must be present,
otherwise there is no use in creating an Energy Model.
The efficiency_class values are squeezed in [0:#efficiency_class-1]
while conserving the order. For instance, efficiency classes of:
[111, 212, 250]
will be mapped to:
[0 (was 111), 1 (was 212), 2 (was 250)].
Each policy being independently registered in the driver, populating
the per_cpu efficiency_class is done only once at the driver
initialization. This prevents from having each policy re-searching the
efficiency_class values of other CPUs. The EM will be registered in a
following patch.
The patch also exports acpi_cpu_get_madt_gicc() to fetch the GicC
structure of the ACPI MADT table for each CPU.
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Pierre Gondois <Pierre.Gondois@arm.com>
Acked-by: Viresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
2022-04-25 15:38:07 +03:00
EXPORT_SYMBOL_GPL ( acpi_cpu_get_madt_gicc ) ;
2017-04-11 11:39:54 +03:00
2015-05-13 16:12:47 +03:00
/*
* acpi_map_gic_cpu_interface - parse processor MADT entry
*
* Carry out sanity checks on MADT processor entry and initialize
* cpu_logical_map on success
*/
static void __init
acpi_map_gic_cpu_interface ( struct acpi_madt_generic_interrupt * processor )
{
u64 hwid = processor - > arm_mpidr ;
2015-07-03 10:29:06 +03:00
if ( ! ( processor - > flags & ACPI_MADT_ENABLED ) ) {
pr_debug ( " skipping disabled CPU entry with 0x%llx MPIDR \n " , hwid ) ;
2015-05-13 16:12:47 +03:00
return ;
}
2015-07-03 10:29:06 +03:00
if ( hwid & ~ MPIDR_HWID_BITMASK | | hwid = = INVALID_HWID ) {
pr_err ( " skipping CPU entry with invalid MPIDR 0x%llx \n " , hwid ) ;
2015-05-13 16:12:47 +03:00
return ;
}
if ( is_mpidr_duplicate ( cpu_count , hwid ) ) {
pr_err ( " duplicate CPU MPIDR 0x%llx in MADT \n " , hwid ) ;
return ;
}
/* Check if GICC structure of boot CPU is available in the MADT */
if ( cpu_logical_map ( 0 ) = = hwid ) {
if ( bootcpu_valid ) {
pr_err ( " duplicate boot CPU MPIDR: 0x%llx in MADT \n " ,
hwid ) ;
return ;
}
bootcpu_valid = true ;
2017-04-11 11:39:54 +03:00
cpu_madt_gicc [ 0 ] = * processor ;
2015-05-13 16:12:47 +03:00
return ;
}
if ( cpu_count > = NR_CPUS )
return ;
/* map the logical cpu id to cpu MPIDR */
2020-07-27 18:29:38 +03:00
set_cpu_logical_map ( cpu_count , hwid ) ;
2015-05-13 16:12:47 +03:00
2017-04-11 11:39:54 +03:00
cpu_madt_gicc [ cpu_count ] = * processor ;
2016-01-26 14:10:38 +03:00
/*
* Set - up the ACPI parking protocol cpu entries
* while initializing the cpu_logical_map to
* avoid parsing MADT entries multiple times for
* nothing ( ie a valid cpu_logical_map entry should
* contain a valid parking protocol data set to
* initialize the cpu if the parking protocol is
* the only available enable method ) .
*/
acpi_set_mailbox_entry ( cpu_count , processor ) ;
2015-05-13 16:12:47 +03:00
cpu_count + + ;
}
static int __init
2019-03-11 23:55:57 +03:00
acpi_parse_gic_cpu_interface ( union acpi_subtable_headers * header ,
2015-05-13 16:12:47 +03:00
const unsigned long end )
{
struct acpi_madt_generic_interrupt * processor ;
processor = ( struct acpi_madt_generic_interrupt * ) header ;
2015-07-07 02:16:48 +03:00
if ( BAD_MADT_GICC_ENTRY ( processor , end ) )
2015-05-13 16:12:47 +03:00
return - EINVAL ;
2019-03-11 23:55:57 +03:00
acpi_table_print_madt_entry ( & header - > common ) ;
2015-05-13 16:12:47 +03:00
acpi_map_gic_cpu_interface ( processor ) ;
return 0 ;
}
2018-06-25 16:05:52 +03:00
static void __init acpi_parse_and_init_cpus ( void )
{
int i ;
/*
* do a walk of MADT to determine how many CPUs
* we have including disabled CPUs , and get information
* we need for SMP init .
*/
acpi_table_parse_madt ( ACPI_MADT_TYPE_GENERIC_INTERRUPT ,
acpi_parse_gic_cpu_interface , 0 ) ;
/*
* In ACPI , SMP and CPU NUMA information is provided in separate
* static tables , namely the MADT and the SRAT .
*
* Thus , it is simpler to first create the cpu logical map through
* an MADT walk and then map the logical cpus to their node ids
* as separate steps .
*/
acpi_map_cpus_to_nodes ( ) ;
for ( i = 0 ; i < nr_cpu_ids ; i + + )
early_map_cpu_to_node ( i , acpi_numa_get_nid ( i ) ) ;
}
2015-05-13 16:12:47 +03:00
# else
2018-06-25 16:05:52 +03:00
# define acpi_parse_and_init_cpus(...) do { } while (0)
2015-05-13 16:12:47 +03:00
# endif
2012-03-05 15:49:30 +04:00
/*
2012-08-29 12:47:19 +04:00
* Enumerate the possible CPU set from the device tree and build the
* cpu logical map array containing MPIDR values related to logical
* cpus . Assumes that cpu_logical_map ( 0 ) has already been initialized .
2012-03-05 15:49:30 +04:00
*/
2015-11-12 15:04:42 +03:00
static void __init of_parse_and_init_cpus ( void )
2012-03-05 15:49:30 +04:00
{
2017-02-02 02:01:05 +03:00
struct device_node * dn ;
2012-03-05 15:49:30 +04:00
2018-08-27 17:43:01 +03:00
for_each_of_cpu_node ( dn ) {
2021-10-06 19:43:24 +03:00
u64 hwid = of_get_cpu_hwid ( dn , 0 ) ;
2012-08-29 12:47:19 +04:00
2021-10-06 19:43:24 +03:00
if ( hwid & ~ MPIDR_HWID_BITMASK )
2012-08-29 12:47:19 +04:00
goto next ;
2015-05-13 16:12:47 +03:00
if ( is_mpidr_duplicate ( cpu_count , hwid ) ) {
2017-07-19 00:42:42 +03:00
pr_err ( " %pOF: duplicate cpu reg properties in the DT \n " ,
dn ) ;
2012-08-29 12:47:19 +04:00
goto next ;
}
/*
* The numbering scheme requires that the boot CPU
* must be assigned logical id 0. Record it so that
* the logical map built from DT is validated and can
* be used .
*/
if ( hwid = = cpu_logical_map ( 0 ) ) {
if ( bootcpu_valid ) {
2017-07-19 00:42:42 +03:00
pr_err ( " %pOF: duplicate boot cpu reg property in DT \n " ,
dn ) ;
2012-08-29 12:47:19 +04:00
goto next ;
}
bootcpu_valid = true ;
2016-09-01 09:55:04 +03:00
early_map_cpu_to_node ( 0 , of_node_to_nid ( dn ) ) ;
2012-08-29 12:47:19 +04:00
/*
* cpu_logical_map has already been
* initialized and the boot cpu doesn ' t need
* the enable - method so continue without
* incrementing cpu .
*/
continue ;
}
2015-05-13 16:12:47 +03:00
if ( cpu_count > = NR_CPUS )
2012-03-05 15:49:30 +04:00
goto next ;
2012-08-29 12:47:19 +04:00
pr_debug ( " cpu logical map 0x%llx \n " , hwid ) ;
2020-07-27 18:29:38 +03:00
set_cpu_logical_map ( cpu_count , hwid ) ;
2016-04-09 01:50:27 +03:00
early_map_cpu_to_node ( cpu_count , of_node_to_nid ( dn ) ) ;
2012-03-05 15:49:30 +04:00
next :
2015-05-13 16:12:47 +03:00
cpu_count + + ;
2012-03-05 15:49:30 +04:00
}
2015-05-13 16:12:47 +03:00
}
/*
* Enumerate the possible CPU set from the device tree or ACPI and build the
* cpu logical map array containing MPIDR values related to logical
* cpus . Assumes that cpu_logical_map ( 0 ) has already been initialized .
*/
void __init smp_init_cpus ( void )
{
int i ;
if ( acpi_disabled )
of_parse_and_init_cpus ( ) ;
else
2018-06-25 16:05:52 +03:00
acpi_parse_and_init_cpus ( ) ;
2012-03-05 15:49:30 +04:00
2016-08-09 05:30:49 +03:00
if ( cpu_count > nr_cpu_ids )
2017-09-09 02:14:18 +03:00
pr_warn ( " Number of cores (%d) exceeds configured maximum of %u - clipping \n " ,
2016-08-09 05:30:49 +03:00
cpu_count , nr_cpu_ids ) ;
2012-08-29 12:47:19 +04:00
if ( ! bootcpu_valid ) {
2015-05-13 16:12:47 +03:00
pr_err ( " missing boot CPU MPIDR, not enabling secondaries \n " ) ;
2012-08-29 12:47:19 +04:00
return ;
}
/*
2015-05-13 16:12:46 +03:00
* We need to set the cpu_logical_map entries before enabling
* the cpus so that cpu processor description entries ( DT cpu nodes
* and ACPI MADT entries ) can be retrieved by matching the cpu hwid
* with entries in cpu_logical_map while initializing the cpus .
* If the cpu set - up fails , invalidate the cpu_logical_map entry .
2012-08-29 12:47:19 +04:00
*/
2016-08-09 05:30:49 +03:00
for ( i = 1 ; i < nr_cpu_ids ; i + + ) {
2015-05-13 16:12:46 +03:00
if ( cpu_logical_map ( i ) ! = INVALID_HWID ) {
if ( smp_cpu_setup ( i ) )
2020-07-27 18:29:38 +03:00
set_cpu_logical_map ( i , INVALID_HWID ) ;
2015-05-13 16:12:46 +03:00
}
}
2012-03-05 15:49:30 +04:00
}
void __init smp_prepare_cpus ( unsigned int max_cpus )
{
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops ;
2013-10-24 23:30:15 +04:00
int err ;
2016-04-22 14:25:35 +03:00
unsigned int cpu ;
arm64: Call numa_store_cpu_info() earlier.
The wq_numa_init() function makes a private CPU to node map by calling
cpu_to_node() early in the boot process, before the non-boot CPUs are
brought online. Since the default implementation of cpu_to_node()
returns zero for CPUs that have never been brought online, the
workqueue system's view is that *all* CPUs are on node zero.
When the unbound workqueue for a non-zero node is created, the
tsk_cpus_allowed() for the worker threads is the empty set because
there are, in the view of the workqueue system, no CPUs on non-zero
nodes. The code in try_to_wake_up() using this empty cpumask ends up
using the cpumask empty set value of NR_CPUS as an index into the
per-CPU area pointer array, and gets garbage as it is one past the end
of the array. This results in:
[ 0.881970] Unable to handle kernel paging request at virtual address fffffb1008b926a4
[ 1.970095] pgd = fffffc00094b0000
[ 1.973530] [fffffb1008b926a4] *pgd=0000000000000000, *pud=0000000000000000, *pmd=0000000000000000
[ 1.982610] Internal error: Oops: 96000004 [#1] SMP
[ 1.987541] Modules linked in:
[ 1.990631] CPU: 48 PID: 295 Comm: cpuhp/48 Tainted: G W 4.8.0-rc6-preempt-vol+ #9
[ 1.999435] Hardware name: Cavium ThunderX CN88XX board (DT)
[ 2.005159] task: fffffe0fe89cc300 task.stack: fffffe0fe8b8c000
[ 2.011158] PC is at try_to_wake_up+0x194/0x34c
[ 2.015737] LR is at try_to_wake_up+0x150/0x34c
[ 2.020318] pc : [<fffffc00080e7468>] lr : [<fffffc00080e7424>] pstate: 600000c5
[ 2.027803] sp : fffffe0fe8b8fb10
[ 2.031149] x29: fffffe0fe8b8fb10 x28: 0000000000000000
[ 2.036522] x27: fffffc0008c63bc8 x26: 0000000000001000
[ 2.041896] x25: fffffc0008c63c80 x24: fffffc0008bfb200
[ 2.047270] x23: 00000000000000c0 x22: 0000000000000004
[ 2.052642] x21: fffffe0fe89d25bc x20: 0000000000001000
[ 2.058014] x19: fffffe0fe89d1d00 x18: 0000000000000000
[ 2.063386] x17: 0000000000000000 x16: 0000000000000000
[ 2.068760] x15: 0000000000000018 x14: 0000000000000000
[ 2.074133] x13: 0000000000000000 x12: 0000000000000000
[ 2.079505] x11: 0000000000000000 x10: 0000000000000000
[ 2.084879] x9 : 0000000000000000 x8 : 0000000000000000
[ 2.090251] x7 : 0000000000000040 x6 : 0000000000000000
[ 2.095621] x5 : ffffffffffffffff x4 : 0000000000000000
[ 2.100991] x3 : 0000000000000000 x2 : 0000000000000000
[ 2.106364] x1 : fffffc0008be4c24 x0 : ffffff0ffffada80
[ 2.111737]
[ 2.113236] Process cpuhp/48 (pid: 295, stack limit = 0xfffffe0fe8b8c020)
[ 2.120102] Stack: (0xfffffe0fe8b8fb10 to 0xfffffe0fe8b90000)
[ 2.125914] fb00: fffffe0fe8b8fb80 fffffc00080e7648
.
.
.
[ 2.442859] Call trace:
[ 2.445327] Exception stack(0xfffffe0fe8b8f940 to 0xfffffe0fe8b8fa70)
[ 2.451843] f940: fffffe0fe89d1d00 0000040000000000 fffffe0fe8b8fb10 fffffc00080e7468
[ 2.459767] f960: fffffe0fe8b8f980 fffffc00080e4958 ffffff0ff91ab200 fffffc00080e4b64
[ 2.467690] f980: fffffe0fe8b8f9d0 fffffc00080e515c fffffe0fe8b8fa80 0000000000000000
[ 2.475614] f9a0: fffffe0fe8b8f9d0 fffffc00080e58e4 fffffe0fe8b8fa80 0000000000000000
[ 2.483540] f9c0: fffffe0fe8d10000 0000000000000040 fffffe0fe8b8fa50 fffffc00080e5ac4
[ 2.491465] f9e0: ffffff0ffffada80 fffffc0008be4c24 0000000000000000 0000000000000000
[ 2.499387] fa00: 0000000000000000 ffffffffffffffff 0000000000000000 0000000000000040
[ 2.507309] fa20: 0000000000000000 0000000000000000 0000000000000000 0000000000000000
[ 2.515233] fa40: 0000000000000000 0000000000000000 0000000000000000 0000000000000018
[ 2.523156] fa60: 0000000000000000 0000000000000000
[ 2.528089] [<fffffc00080e7468>] try_to_wake_up+0x194/0x34c
[ 2.533723] [<fffffc00080e7648>] wake_up_process+0x28/0x34
[ 2.539275] [<fffffc00080d3764>] create_worker+0x110/0x19c
[ 2.544824] [<fffffc00080d69dc>] alloc_unbound_pwq+0x3cc/0x4b0
[ 2.550724] [<fffffc00080d6bcc>] wq_update_unbound_numa+0x10c/0x1e4
[ 2.557066] [<fffffc00080d7d78>] workqueue_online_cpu+0x220/0x28c
[ 2.563234] [<fffffc00080bd288>] cpuhp_invoke_callback+0x6c/0x168
[ 2.569398] [<fffffc00080bdf74>] cpuhp_up_callbacks+0x44/0xe4
[ 2.575210] [<fffffc00080be194>] cpuhp_thread_fun+0x13c/0x148
[ 2.581027] [<fffffc00080dfbac>] smpboot_thread_fn+0x19c/0x1a8
[ 2.586929] [<fffffc00080dbd64>] kthread+0xdc/0xf0
[ 2.591776] [<fffffc0008083380>] ret_from_fork+0x10/0x50
[ 2.597147] Code: b00057e1 91304021 91005021 b8626822 (b8606821)
[ 2.603464] ---[ end trace 58c0cd36b88802bc ]---
[ 2.608138] Kernel panic - not syncing: Fatal exception
Fix by moving call to numa_store_cpu_info() for all CPUs into
smp_prepare_cpus(), which happens before wq_numa_init(). Since
smp_store_cpu_info() now contains only a single function call,
simplify by removing the function and out-lining its contents.
Suggested-by: Robert Richter <rric@kernel.org>
Fixes: 1a2db300348b ("arm64, numa: Add NUMA support for arm64 platforms.")
Cc: <stable@vger.kernel.org> # 4.7.x-
Signed-off-by: David Daney <david.daney@cavium.com>
Reviewed-by: Robert Richter <rrichter@cavium.com>
Tested-by: Yisheng Xie <xieyisheng1@huawei.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-09-20 21:46:35 +03:00
unsigned int this_cpu ;
2012-03-05 15:49:30 +04:00
2014-03-04 11:51:17 +04:00
init_cpu_topology ( ) ;
arm64: Call numa_store_cpu_info() earlier.
The wq_numa_init() function makes a private CPU to node map by calling
cpu_to_node() early in the boot process, before the non-boot CPUs are
brought online. Since the default implementation of cpu_to_node()
returns zero for CPUs that have never been brought online, the
workqueue system's view is that *all* CPUs are on node zero.
When the unbound workqueue for a non-zero node is created, the
tsk_cpus_allowed() for the worker threads is the empty set because
there are, in the view of the workqueue system, no CPUs on non-zero
nodes. The code in try_to_wake_up() using this empty cpumask ends up
using the cpumask empty set value of NR_CPUS as an index into the
per-CPU area pointer array, and gets garbage as it is one past the end
of the array. This results in:
[ 0.881970] Unable to handle kernel paging request at virtual address fffffb1008b926a4
[ 1.970095] pgd = fffffc00094b0000
[ 1.973530] [fffffb1008b926a4] *pgd=0000000000000000, *pud=0000000000000000, *pmd=0000000000000000
[ 1.982610] Internal error: Oops: 96000004 [#1] SMP
[ 1.987541] Modules linked in:
[ 1.990631] CPU: 48 PID: 295 Comm: cpuhp/48 Tainted: G W 4.8.0-rc6-preempt-vol+ #9
[ 1.999435] Hardware name: Cavium ThunderX CN88XX board (DT)
[ 2.005159] task: fffffe0fe89cc300 task.stack: fffffe0fe8b8c000
[ 2.011158] PC is at try_to_wake_up+0x194/0x34c
[ 2.015737] LR is at try_to_wake_up+0x150/0x34c
[ 2.020318] pc : [<fffffc00080e7468>] lr : [<fffffc00080e7424>] pstate: 600000c5
[ 2.027803] sp : fffffe0fe8b8fb10
[ 2.031149] x29: fffffe0fe8b8fb10 x28: 0000000000000000
[ 2.036522] x27: fffffc0008c63bc8 x26: 0000000000001000
[ 2.041896] x25: fffffc0008c63c80 x24: fffffc0008bfb200
[ 2.047270] x23: 00000000000000c0 x22: 0000000000000004
[ 2.052642] x21: fffffe0fe89d25bc x20: 0000000000001000
[ 2.058014] x19: fffffe0fe89d1d00 x18: 0000000000000000
[ 2.063386] x17: 0000000000000000 x16: 0000000000000000
[ 2.068760] x15: 0000000000000018 x14: 0000000000000000
[ 2.074133] x13: 0000000000000000 x12: 0000000000000000
[ 2.079505] x11: 0000000000000000 x10: 0000000000000000
[ 2.084879] x9 : 0000000000000000 x8 : 0000000000000000
[ 2.090251] x7 : 0000000000000040 x6 : 0000000000000000
[ 2.095621] x5 : ffffffffffffffff x4 : 0000000000000000
[ 2.100991] x3 : 0000000000000000 x2 : 0000000000000000
[ 2.106364] x1 : fffffc0008be4c24 x0 : ffffff0ffffada80
[ 2.111737]
[ 2.113236] Process cpuhp/48 (pid: 295, stack limit = 0xfffffe0fe8b8c020)
[ 2.120102] Stack: (0xfffffe0fe8b8fb10 to 0xfffffe0fe8b90000)
[ 2.125914] fb00: fffffe0fe8b8fb80 fffffc00080e7648
.
.
.
[ 2.442859] Call trace:
[ 2.445327] Exception stack(0xfffffe0fe8b8f940 to 0xfffffe0fe8b8fa70)
[ 2.451843] f940: fffffe0fe89d1d00 0000040000000000 fffffe0fe8b8fb10 fffffc00080e7468
[ 2.459767] f960: fffffe0fe8b8f980 fffffc00080e4958 ffffff0ff91ab200 fffffc00080e4b64
[ 2.467690] f980: fffffe0fe8b8f9d0 fffffc00080e515c fffffe0fe8b8fa80 0000000000000000
[ 2.475614] f9a0: fffffe0fe8b8f9d0 fffffc00080e58e4 fffffe0fe8b8fa80 0000000000000000
[ 2.483540] f9c0: fffffe0fe8d10000 0000000000000040 fffffe0fe8b8fa50 fffffc00080e5ac4
[ 2.491465] f9e0: ffffff0ffffada80 fffffc0008be4c24 0000000000000000 0000000000000000
[ 2.499387] fa00: 0000000000000000 ffffffffffffffff 0000000000000000 0000000000000040
[ 2.507309] fa20: 0000000000000000 0000000000000000 0000000000000000 0000000000000000
[ 2.515233] fa40: 0000000000000000 0000000000000000 0000000000000000 0000000000000018
[ 2.523156] fa60: 0000000000000000 0000000000000000
[ 2.528089] [<fffffc00080e7468>] try_to_wake_up+0x194/0x34c
[ 2.533723] [<fffffc00080e7648>] wake_up_process+0x28/0x34
[ 2.539275] [<fffffc00080d3764>] create_worker+0x110/0x19c
[ 2.544824] [<fffffc00080d69dc>] alloc_unbound_pwq+0x3cc/0x4b0
[ 2.550724] [<fffffc00080d6bcc>] wq_update_unbound_numa+0x10c/0x1e4
[ 2.557066] [<fffffc00080d7d78>] workqueue_online_cpu+0x220/0x28c
[ 2.563234] [<fffffc00080bd288>] cpuhp_invoke_callback+0x6c/0x168
[ 2.569398] [<fffffc00080bdf74>] cpuhp_up_callbacks+0x44/0xe4
[ 2.575210] [<fffffc00080be194>] cpuhp_thread_fun+0x13c/0x148
[ 2.581027] [<fffffc00080dfbac>] smpboot_thread_fn+0x19c/0x1a8
[ 2.586929] [<fffffc00080dbd64>] kthread+0xdc/0xf0
[ 2.591776] [<fffffc0008083380>] ret_from_fork+0x10/0x50
[ 2.597147] Code: b00057e1 91304021 91005021 b8626822 (b8606821)
[ 2.603464] ---[ end trace 58c0cd36b88802bc ]---
[ 2.608138] Kernel panic - not syncing: Fatal exception
Fix by moving call to numa_store_cpu_info() for all CPUs into
smp_prepare_cpus(), which happens before wq_numa_init(). Since
smp_store_cpu_info() now contains only a single function call,
simplify by removing the function and out-lining its contents.
Suggested-by: Robert Richter <rric@kernel.org>
Fixes: 1a2db300348b ("arm64, numa: Add NUMA support for arm64 platforms.")
Cc: <stable@vger.kernel.org> # 4.7.x-
Signed-off-by: David Daney <david.daney@cavium.com>
Reviewed-by: Robert Richter <rrichter@cavium.com>
Tested-by: Yisheng Xie <xieyisheng1@huawei.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-09-20 21:46:35 +03:00
this_cpu = smp_processor_id ( ) ;
store_cpu_topology ( this_cpu ) ;
numa_store_cpu_info ( this_cpu ) ;
2018-07-06 14:02:43 +03:00
numa_add_cpu ( this_cpu ) ;
2014-03-04 11:51:17 +04:00
2016-07-21 13:15:27 +03:00
/*
* If UP is mandated by " nosmp " ( which implies " maxcpus=0 " ) , don ' t set
* secondary CPUs present .
*/
if ( max_cpus = = 0 )
return ;
2012-03-05 15:49:30 +04:00
/*
* Initialise the present map ( which describes the set of CPUs
* actually populated at the present time ) and release the
* secondaries from the bootloader .
*/
for_each_possible_cpu ( cpu ) {
arm64: make cpu number a percpu variable
In the absence of CONFIG_THREAD_INFO_IN_TASK, core code maintains
thread_info::cpu, and low-level architecture code can access this to
build raw_smp_processor_id(). With CONFIG_THREAD_INFO_IN_TASK, core code
maintains task_struct::cpu, which for reasons of hte header soup is not
accessible to low-level arch code.
Instead, we can maintain a percpu variable containing the cpu number.
For both the old and new implementation of raw_smp_processor_id(), we
read a syreg into a GPR, add an offset, and load the result. As the
offset is now larger, it may not be folded into the load, but otherwise
the assembly shouldn't change much.
Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Tested-by: Laura Abbott <labbott@redhat.com>
Cc: James Morse <james.morse@arm.com>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-11-03 23:23:11 +03:00
per_cpu ( cpu_number , cpu ) = cpu ;
2013-01-02 19:24:22 +04:00
if ( cpu = = smp_processor_id ( ) )
continue ;
2020-03-19 02:01:44 +03:00
ops = get_cpu_ops ( cpu ) ;
if ( ! ops )
2012-03-05 15:49:30 +04:00
continue ;
2020-03-19 02:01:44 +03:00
err = ops - > cpu_prepare ( cpu ) ;
2013-01-02 19:24:22 +04:00
if ( err )
continue ;
2012-03-05 15:49:30 +04:00
set_cpu_present ( cpu , true ) ;
arm64: Call numa_store_cpu_info() earlier.
The wq_numa_init() function makes a private CPU to node map by calling
cpu_to_node() early in the boot process, before the non-boot CPUs are
brought online. Since the default implementation of cpu_to_node()
returns zero for CPUs that have never been brought online, the
workqueue system's view is that *all* CPUs are on node zero.
When the unbound workqueue for a non-zero node is created, the
tsk_cpus_allowed() for the worker threads is the empty set because
there are, in the view of the workqueue system, no CPUs on non-zero
nodes. The code in try_to_wake_up() using this empty cpumask ends up
using the cpumask empty set value of NR_CPUS as an index into the
per-CPU area pointer array, and gets garbage as it is one past the end
of the array. This results in:
[ 0.881970] Unable to handle kernel paging request at virtual address fffffb1008b926a4
[ 1.970095] pgd = fffffc00094b0000
[ 1.973530] [fffffb1008b926a4] *pgd=0000000000000000, *pud=0000000000000000, *pmd=0000000000000000
[ 1.982610] Internal error: Oops: 96000004 [#1] SMP
[ 1.987541] Modules linked in:
[ 1.990631] CPU: 48 PID: 295 Comm: cpuhp/48 Tainted: G W 4.8.0-rc6-preempt-vol+ #9
[ 1.999435] Hardware name: Cavium ThunderX CN88XX board (DT)
[ 2.005159] task: fffffe0fe89cc300 task.stack: fffffe0fe8b8c000
[ 2.011158] PC is at try_to_wake_up+0x194/0x34c
[ 2.015737] LR is at try_to_wake_up+0x150/0x34c
[ 2.020318] pc : [<fffffc00080e7468>] lr : [<fffffc00080e7424>] pstate: 600000c5
[ 2.027803] sp : fffffe0fe8b8fb10
[ 2.031149] x29: fffffe0fe8b8fb10 x28: 0000000000000000
[ 2.036522] x27: fffffc0008c63bc8 x26: 0000000000001000
[ 2.041896] x25: fffffc0008c63c80 x24: fffffc0008bfb200
[ 2.047270] x23: 00000000000000c0 x22: 0000000000000004
[ 2.052642] x21: fffffe0fe89d25bc x20: 0000000000001000
[ 2.058014] x19: fffffe0fe89d1d00 x18: 0000000000000000
[ 2.063386] x17: 0000000000000000 x16: 0000000000000000
[ 2.068760] x15: 0000000000000018 x14: 0000000000000000
[ 2.074133] x13: 0000000000000000 x12: 0000000000000000
[ 2.079505] x11: 0000000000000000 x10: 0000000000000000
[ 2.084879] x9 : 0000000000000000 x8 : 0000000000000000
[ 2.090251] x7 : 0000000000000040 x6 : 0000000000000000
[ 2.095621] x5 : ffffffffffffffff x4 : 0000000000000000
[ 2.100991] x3 : 0000000000000000 x2 : 0000000000000000
[ 2.106364] x1 : fffffc0008be4c24 x0 : ffffff0ffffada80
[ 2.111737]
[ 2.113236] Process cpuhp/48 (pid: 295, stack limit = 0xfffffe0fe8b8c020)
[ 2.120102] Stack: (0xfffffe0fe8b8fb10 to 0xfffffe0fe8b90000)
[ 2.125914] fb00: fffffe0fe8b8fb80 fffffc00080e7648
.
.
.
[ 2.442859] Call trace:
[ 2.445327] Exception stack(0xfffffe0fe8b8f940 to 0xfffffe0fe8b8fa70)
[ 2.451843] f940: fffffe0fe89d1d00 0000040000000000 fffffe0fe8b8fb10 fffffc00080e7468
[ 2.459767] f960: fffffe0fe8b8f980 fffffc00080e4958 ffffff0ff91ab200 fffffc00080e4b64
[ 2.467690] f980: fffffe0fe8b8f9d0 fffffc00080e515c fffffe0fe8b8fa80 0000000000000000
[ 2.475614] f9a0: fffffe0fe8b8f9d0 fffffc00080e58e4 fffffe0fe8b8fa80 0000000000000000
[ 2.483540] f9c0: fffffe0fe8d10000 0000000000000040 fffffe0fe8b8fa50 fffffc00080e5ac4
[ 2.491465] f9e0: ffffff0ffffada80 fffffc0008be4c24 0000000000000000 0000000000000000
[ 2.499387] fa00: 0000000000000000 ffffffffffffffff 0000000000000000 0000000000000040
[ 2.507309] fa20: 0000000000000000 0000000000000000 0000000000000000 0000000000000000
[ 2.515233] fa40: 0000000000000000 0000000000000000 0000000000000000 0000000000000018
[ 2.523156] fa60: 0000000000000000 0000000000000000
[ 2.528089] [<fffffc00080e7468>] try_to_wake_up+0x194/0x34c
[ 2.533723] [<fffffc00080e7648>] wake_up_process+0x28/0x34
[ 2.539275] [<fffffc00080d3764>] create_worker+0x110/0x19c
[ 2.544824] [<fffffc00080d69dc>] alloc_unbound_pwq+0x3cc/0x4b0
[ 2.550724] [<fffffc00080d6bcc>] wq_update_unbound_numa+0x10c/0x1e4
[ 2.557066] [<fffffc00080d7d78>] workqueue_online_cpu+0x220/0x28c
[ 2.563234] [<fffffc00080bd288>] cpuhp_invoke_callback+0x6c/0x168
[ 2.569398] [<fffffc00080bdf74>] cpuhp_up_callbacks+0x44/0xe4
[ 2.575210] [<fffffc00080be194>] cpuhp_thread_fun+0x13c/0x148
[ 2.581027] [<fffffc00080dfbac>] smpboot_thread_fn+0x19c/0x1a8
[ 2.586929] [<fffffc00080dbd64>] kthread+0xdc/0xf0
[ 2.591776] [<fffffc0008083380>] ret_from_fork+0x10/0x50
[ 2.597147] Code: b00057e1 91304021 91005021 b8626822 (b8606821)
[ 2.603464] ---[ end trace 58c0cd36b88802bc ]---
[ 2.608138] Kernel panic - not syncing: Fatal exception
Fix by moving call to numa_store_cpu_info() for all CPUs into
smp_prepare_cpus(), which happens before wq_numa_init(). Since
smp_store_cpu_info() now contains only a single function call,
simplify by removing the function and out-lining its contents.
Suggested-by: Robert Richter <rric@kernel.org>
Fixes: 1a2db300348b ("arm64, numa: Add NUMA support for arm64 platforms.")
Cc: <stable@vger.kernel.org> # 4.7.x-
Signed-off-by: David Daney <david.daney@cavium.com>
Reviewed-by: Robert Richter <rrichter@cavium.com>
Tested-by: Yisheng Xie <xieyisheng1@huawei.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2016-09-20 21:46:35 +03:00
numa_store_cpu_info ( cpu ) ;
2012-03-05 15:49:30 +04:00
}
}
2014-07-26 00:05:32 +04:00
static const char * ipi_types [ NR_IPI ] __tracepoint_string = {
2020-11-09 14:38:36 +03:00
[ IPI_RESCHEDULE ] = " Rescheduling interrupts " ,
[ IPI_CALL_FUNC ] = " Function call interrupts " ,
[ IPI_CPU_STOP ] = " CPU stop interrupts " ,
[ IPI_CPU_CRASH_STOP ] = " CPU stop (for crash dump) interrupts " ,
[ IPI_TIMER ] = " Timer broadcast interrupts " ,
[ IPI_IRQ_WORK ] = " IRQ work interrupts " ,
[ IPI_WAKEUP ] = " CPU wake-up interrupts " ,
2012-03-05 15:49:30 +04:00
} ;
2020-05-09 16:00:23 +03:00
static void smp_cross_call ( const struct cpumask * target , unsigned int ipinr ) ;
2014-07-26 00:05:32 +04:00
2020-06-20 19:19:00 +03:00
unsigned long irq_err_count ;
int arch_show_interrupts ( struct seq_file * p , int prec )
2012-03-05 15:49:30 +04:00
{
unsigned int cpu , i ;
for ( i = 0 ; i < NR_IPI ; i + + ) {
2014-07-26 00:05:32 +04:00
seq_printf ( p , " %*s%u:%s " , prec - 1 , " IPI " , i ,
2012-03-05 15:49:30 +04:00
prec > = 4 ? " " : " " ) ;
2013-11-07 19:25:44 +04:00
for_each_online_cpu ( cpu )
2020-12-10 22:25:46 +03:00
seq_printf ( p , " %10u " , irq_desc_kstat_cpu ( ipi_desc [ i ] , cpu ) ) ;
2012-03-05 15:49:30 +04:00
seq_printf ( p , " %s \n " , ipi_types [ i ] ) ;
}
2020-06-20 19:19:00 +03:00
seq_printf ( p , " %*s: %10lu \n " , prec , " Err " , irq_err_count ) ;
return 0 ;
2012-03-05 15:49:30 +04:00
}
2014-07-26 00:05:32 +04:00
void arch_send_call_function_ipi_mask ( const struct cpumask * mask )
{
smp_cross_call ( mask , IPI_CALL_FUNC ) ;
}
void arch_send_call_function_single_ipi ( int cpu )
{
2015-01-23 08:36:42 +03:00
smp_cross_call ( cpumask_of ( cpu ) , IPI_CALL_FUNC ) ;
2014-07-26 00:05:32 +04:00
}
2016-01-26 14:10:38 +03:00
# ifdef CONFIG_ARM64_ACPI_PARKING_PROTOCOL
void arch_send_wakeup_ipi_mask ( const struct cpumask * mask )
{
smp_cross_call ( mask , IPI_WAKEUP ) ;
}
# endif
2014-07-26 00:05:32 +04:00
# ifdef CONFIG_IRQ_WORK
void arch_irq_work_raise ( void )
{
2020-05-09 16:00:23 +03:00
smp_cross_call ( cpumask_of ( smp_processor_id ( ) ) , IPI_IRQ_WORK ) ;
2014-07-26 00:05:32 +04:00
}
# endif
2023-04-13 02:49:34 +03:00
static void __noreturn local_cpu_stop ( void )
2012-03-05 15:49:30 +04:00
{
2019-06-17 23:35:19 +03:00
set_cpu_online ( smp_processor_id ( ) , false ) ;
2012-03-05 15:49:30 +04:00
2017-11-02 15:12:34 +03:00
local_daif_mask ( ) ;
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
sdei_mask_local_cpu ( ) ;
2019-06-17 23:35:18 +03:00
cpu_park_loop ( ) ;
2012-03-05 15:49:30 +04:00
}
2019-06-17 23:35:19 +03:00
/*
* We need to implement panic_smp_self_stop ( ) for parallel panic ( ) calls , so
* that cpu_online_mask gets correctly updated and smp_send_stop ( ) can skip
* CPUs that have already stopped themselves .
*/
2023-04-13 02:49:35 +03:00
void __noreturn panic_smp_self_stop ( void )
2019-06-17 23:35:19 +03:00
{
local_cpu_stop ( ) ;
2012-03-05 15:49:30 +04:00
}
2017-04-03 05:24:36 +03:00
# ifdef CONFIG_KEXEC_CORE
static atomic_t waiting_for_crash_ipi = ATOMIC_INIT ( 0 ) ;
# endif
2023-04-13 02:49:34 +03:00
static void __noreturn ipi_cpu_crash_stop ( unsigned int cpu , struct pt_regs * regs )
2017-04-03 05:24:36 +03:00
{
# ifdef CONFIG_KEXEC_CORE
crash_save_cpu ( regs , cpu ) ;
atomic_dec ( & waiting_for_crash_ipi ) ;
local_irq_disable ( ) ;
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
sdei_mask_local_cpu ( ) ;
2017-04-03 05:24:36 +03:00
2020-03-19 02:01:44 +03:00
if ( IS_ENABLED ( CONFIG_HOTPLUG_CPU ) )
__cpu_try_die ( cpu ) ;
2017-04-03 05:24:36 +03:00
/* just in case */
cpu_park_loop ( ) ;
2023-04-13 02:49:34 +03:00
# else
BUG ( ) ;
2017-04-03 05:24:36 +03:00
# endif
}
2012-03-05 15:49:30 +04:00
/*
* Main handler for inter - processor interrupts
*/
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
static void do_handle_IPI ( int ipinr )
2012-03-05 15:49:30 +04:00
{
unsigned int cpu = smp_processor_id ( ) ;
2020-06-20 19:19:00 +03:00
if ( ( unsigned ) ipinr < NR_IPI )
2023-01-12 22:43:38 +03:00
trace_ipi_entry ( ipi_types [ ipinr ] ) ;
2012-03-05 15:49:30 +04:00
switch ( ipinr ) {
case IPI_RESCHEDULE :
scheduler_ipi ( ) ;
break ;
case IPI_CALL_FUNC :
generic_smp_call_function_interrupt ( ) ;
break ;
case IPI_CPU_STOP :
2019-06-17 23:35:19 +03:00
local_cpu_stop ( ) ;
2012-03-05 15:49:30 +04:00
break ;
2017-04-03 05:24:36 +03:00
case IPI_CPU_CRASH_STOP :
if ( IS_ENABLED ( CONFIG_KEXEC_CORE ) ) {
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
ipi_cpu_crash_stop ( cpu , get_irq_regs ( ) ) ;
2017-04-03 05:24:36 +03:00
unreachable ( ) ;
}
break ;
2013-09-04 13:55:17 +04:00
# ifdef CONFIG_GENERIC_CLOCKEVENTS_BROADCAST
case IPI_TIMER :
tick_receive_broadcast ( ) ;
break ;
# endif
2014-05-12 19:48:51 +04:00
# ifdef CONFIG_IRQ_WORK
case IPI_IRQ_WORK :
irq_work_run ( ) ;
break ;
# endif
2016-01-26 14:10:38 +03:00
# ifdef CONFIG_ARM64_ACPI_PARKING_PROTOCOL
case IPI_WAKEUP :
WARN_ONCE ( ! acpi_parking_protocol_valid ( cpu ) ,
" CPU%u: Wake-up IPI outside the ACPI parking protocol \n " ,
cpu ) ;
break ;
# endif
2012-03-05 15:49:30 +04:00
default :
pr_crit ( " CPU%u: Unknown IPI message 0x%x \n " , cpu , ipinr ) ;
break ;
}
2014-07-26 00:05:32 +04:00
if ( ( unsigned ) ipinr < NR_IPI )
2023-01-12 22:43:38 +03:00
trace_ipi_exit ( ipi_types [ ipinr ] ) ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
}
static irqreturn_t ipi_handler ( int irq , void * data )
{
do_handle_IPI ( irq - ipi_irq_base ) ;
return IRQ_HANDLED ;
}
2020-05-09 16:00:23 +03:00
static void smp_cross_call ( const struct cpumask * target , unsigned int ipinr )
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
{
2020-05-09 16:00:23 +03:00
trace_ipi_raise ( target , ipi_types [ ipinr ] ) ;
__ipi_send_mask ( ipi_desc [ ipinr ] , target ) ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
}
static void ipi_setup ( int cpu )
{
int i ;
2020-05-09 16:00:23 +03:00
if ( WARN_ON_ONCE ( ! ipi_irq_base ) )
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
return ;
for ( i = 0 ; i < nr_ipi ; i + + )
enable_percpu_irq ( ipi_irq_base + i , 0 ) ;
}
2020-09-18 15:33:18 +03:00
# ifdef CONFIG_HOTPLUG_CPU
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
static void ipi_teardown ( int cpu )
{
int i ;
2020-05-09 16:00:23 +03:00
if ( WARN_ON_ONCE ( ! ipi_irq_base ) )
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
return ;
for ( i = 0 ; i < nr_ipi ; i + + )
disable_percpu_irq ( ipi_irq_base + i ) ;
}
2020-09-18 15:33:18 +03:00
# endif
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
void __init set_smp_ipi_range ( int ipi_base , int n )
{
int i ;
WARN_ON ( n < NR_IPI ) ;
nr_ipi = min ( n , NR_IPI ) ;
for ( i = 0 ; i < nr_ipi ; i + + ) {
int err ;
err = request_percpu_irq ( ipi_base + i , ipi_handler ,
2020-06-20 19:19:00 +03:00
" IPI " , & cpu_number ) ;
arm64: Allow IPIs to be handled as normal interrupts
In order to deal with IPIs as normal interrupts, let's add
a new way to register them with the architecture code.
set_smp_ipi_range() takes a range of interrupts, and allows
the arch code to request them as if the were normal interrupts.
A standard handler is then called by the core IRQ code to deal
with the IPI.
This means that we don't need to call irq_enter/irq_exit, and
that we don't need to deal with set_irq_regs either. So let's
move the dispatcher into its own function, and leave handle_IPI()
as a compatibility function.
On the sending side, let's make use of ipi_send_mask, which
already exists for this purpose.
One of the major difference is that we end up, in some cases
(such as when performing IRQ time accounting on the scheduler
IPI), end up with nested irq_enter()/irq_exit() pairs.
Other than the (relatively small) overhead, there should be
no consequences to it (these pairs are designed to nest
correctly, and the accounting shouldn't be off).
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Marc Zyngier <maz@kernel.org>
2020-04-25 17:03:47 +03:00
WARN_ON ( err ) ;
ipi_desc [ i ] = irq_to_desc ( ipi_base + i ) ;
irq_set_status_flags ( ipi_base + i , IRQ_HIDDEN ) ;
}
ipi_irq_base = ipi_base ;
/* Setup the boot CPU immediately */
ipi_setup ( smp_processor_id ( ) ) ;
}
2023-03-07 17:35:56 +03:00
void arch_smp_send_reschedule ( int cpu )
2012-03-05 15:49:30 +04:00
{
smp_cross_call ( cpumask_of ( cpu ) , IPI_RESCHEDULE ) ;
}
2013-09-04 13:55:17 +04:00
# ifdef CONFIG_GENERIC_CLOCKEVENTS_BROADCAST
void tick_broadcast ( const struct cpumask * mask )
{
smp_cross_call ( mask , IPI_TIMER ) ;
}
# endif
2020-03-11 20:12:44 +03:00
/*
* The number of CPUs online , not counting this CPU ( which may not be
* fully online and so not counted in num_online_cpus ( ) ) .
*/
static inline unsigned int num_other_online_cpus ( void )
{
unsigned int this_cpu_online = cpu_online ( smp_processor_id ( ) ) ;
return num_online_cpus ( ) - this_cpu_online ;
}
2012-03-05 15:49:30 +04:00
void smp_send_stop ( void )
{
unsigned long timeout ;
2020-03-11 20:12:44 +03:00
if ( num_other_online_cpus ( ) ) {
2012-03-05 15:49:30 +04:00
cpumask_t mask ;
cpumask_copy ( & mask , cpu_online_mask ) ;
2015-03-05 03:19:18 +03:00
cpumask_clear_cpu ( smp_processor_id ( ) , & mask ) ;
2012-03-05 15:49:30 +04:00
2017-05-16 21:42:34 +03:00
if ( system_state < = SYSTEM_RUNNING )
2016-04-18 10:43:33 +03:00
pr_crit ( " SMP: stopping secondary CPUs \n " ) ;
2012-03-05 15:49:30 +04:00
smp_cross_call ( & mask , IPI_CPU_STOP ) ;
}
/* Wait up to one second for other CPUs to stop */
timeout = USEC_PER_SEC ;
2020-03-11 20:12:44 +03:00
while ( num_other_online_cpus ( ) & & timeout - - )
2012-03-05 15:49:30 +04:00
udelay ( 1 ) ;
2020-03-11 20:12:44 +03:00
if ( num_other_online_cpus ( ) )
2019-10-18 06:18:19 +03:00
pr_warn ( " SMP: failed to stop secondary CPUs %*pbl \n " ,
cpumask_pr_args ( cpu_online_mask ) ) ;
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
sdei_mask_local_cpu ( ) ;
2012-03-05 15:49:30 +04:00
}
2017-04-03 05:24:36 +03:00
# ifdef CONFIG_KEXEC_CORE
2017-08-17 05:24:27 +03:00
void crash_smp_send_stop ( void )
2017-04-03 05:24:36 +03:00
{
2017-08-17 05:24:27 +03:00
static int cpus_stopped ;
2017-04-03 05:24:36 +03:00
cpumask_t mask ;
unsigned long timeout ;
2017-08-17 05:24:27 +03:00
/*
* This function can be called twice in panic path , but obviously
* we execute this only once .
*/
if ( cpus_stopped )
return ;
cpus_stopped = 1 ;
2020-03-11 20:12:45 +03:00
/*
* If this cpu is the only one alive at this point in time , online or
* not , there are no stop messages to be sent around , so just back out .
*/
if ( num_other_online_cpus ( ) = = 0 ) {
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
sdei_mask_local_cpu ( ) ;
2017-04-03 05:24:36 +03:00
return ;
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
}
2017-04-03 05:24:36 +03:00
cpumask_copy ( & mask , cpu_online_mask ) ;
cpumask_clear_cpu ( smp_processor_id ( ) , & mask ) ;
2020-03-11 20:12:45 +03:00
atomic_set ( & waiting_for_crash_ipi , num_other_online_cpus ( ) ) ;
2017-04-03 05:24:36 +03:00
pr_crit ( " SMP: stopping secondary CPUs \n " ) ;
smp_cross_call ( & mask , IPI_CPU_CRASH_STOP ) ;
/* Wait up to one second for other CPUs to stop */
timeout = USEC_PER_SEC ;
while ( ( atomic_read ( & waiting_for_crash_ipi ) > 0 ) & & timeout - - )
udelay ( 1 ) ;
if ( atomic_read ( & waiting_for_crash_ipi ) > 0 )
2019-10-18 06:18:19 +03:00
pr_warn ( " SMP: failed to stop secondary CPUs %*pbl \n " ,
cpumask_pr_args ( & mask ) ) ;
arm64: kernel: Add arch-specific SDEI entry code and CPU masking
The Software Delegated Exception Interface (SDEI) is an ARM standard
for registering callbacks from the platform firmware into the OS.
This is typically used to implement RAS notifications.
Such notifications enter the kernel at the registered entry-point
with the register values of the interrupted CPU context. Because this
is not a CPU exception, it cannot reuse the existing entry code.
(crucially we don't implicitly know which exception level we interrupted),
Add the entry point to entry.S to set us up for calling into C code. If
the event interrupted code that had interrupts masked, we always return
to that location. Otherwise we pretend this was an IRQ, and use SDEI's
complete_and_resume call to return to vbar_el1 + offset.
This allows the kernel to deliver signals to user space processes. For
KVM this triggers the world switch, a quick spin round vcpu_run, then
back into the guest, unless there are pending signals.
Add sdei_mask_local_cpu() calls to the smp_send_stop() code, this covers
the panic() code-path, which doesn't invoke cpuhotplug notifiers.
Because we can interrupt entry-from/exit-to another EL, we can't trust the
value in sp_el0 or x29, even if we interrupted the kernel, in this case
the code in entry.S will save/restore sp_el0 and use the value in
__entry_task.
When we have VMAP stacks we can interrupt the stack-overflow test, which
stirs x0 into sp, meaning we have to have our own VMAP stacks. For now
these are allocated when we probe the interface. Future patches will add
refcounting hooks to allow the arch code to allocate them lazily.
Signed-off-by: James Morse <james.morse@arm.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2018-01-08 18:38:12 +03:00
sdei_mask_local_cpu ( ) ;
2017-04-03 05:24:36 +03:00
}
bool smp_crash_stop_failed ( void )
{
return ( atomic_read ( & waiting_for_crash_ipi ) > 0 ) ;
}
# endif
2016-06-22 12:06:12 +03:00
static bool have_cpu_die ( void )
{
# ifdef CONFIG_HOTPLUG_CPU
int any_cpu = raw_smp_processor_id ( ) ;
2020-03-19 02:01:44 +03:00
const struct cpu_operations * ops = get_cpu_ops ( any_cpu ) ;
2016-06-22 12:06:12 +03:00
2020-03-19 02:01:44 +03:00
if ( ops & & ops - > cpu_die )
2016-06-22 12:06:12 +03:00
return true ;
# endif
return false ;
}
bool cpus_are_stuck_in_kernel ( void )
{
bool smp_spin_tables = ( num_possible_cpus ( ) > 1 & & ! have_cpu_die ( ) ) ;
2021-10-08 16:58:35 +03:00
return ! ! cpus_stuck_in_kernel | | smp_spin_tables | |
is_protected_kvm_enabled ( ) ;
2016-06-22 12:06:12 +03:00
}