951c2a51ae
The per CPU hardirq_stack_ptr contains the pointer to the irq stack in the form that it is ready to be assigned to [ER]SP so that the first push ends up on the top entry of the stack. But the stack switching on 64 bit has the following rules: 1) Store the current stack pointer (RSP) in the top most stack entry to allow the unwinder to link back to the previous stack 2) Set RSP to the top most stack entry 3) Invoke functions on the irq stack 4) Pop RSP from the top most stack entry (stored in #1) so it's back to the original stack. That requires all stack switching code to decrement the stored pointer by 8 in order to be able to store the current RSP and then set RSP to that location. That's a pointless exercise. Do the -8 adjustment right when storing the pointer and make the data type a void pointer to avoid confusion vs. the struct irq_stack data type which is on 64bit only used to declare the backing store. Move the definition next to the inuse flag so they likely end up in the same cache line. Sticking them into a struct to enforce it is a seperate change. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Kees Cook <keescook@chromium.org> Link: https://lore.kernel.org/r/20210210002512.354260928@linutronix.de
216 lines
5.3 KiB
C
216 lines
5.3 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 1991, 1992 Linus Torvalds
|
|
* Copyright (C) 2000, 2001, 2002 Andi Kleen, SuSE Labs
|
|
*/
|
|
#include <linux/sched/debug.h>
|
|
#include <linux/kallsyms.h>
|
|
#include <linux/kprobes.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/hardirq.h>
|
|
#include <linux/kdebug.h>
|
|
#include <linux/export.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/kexec.h>
|
|
#include <linux/sysfs.h>
|
|
#include <linux/bug.h>
|
|
#include <linux/nmi.h>
|
|
|
|
#include <asm/cpu_entry_area.h>
|
|
#include <asm/stacktrace.h>
|
|
|
|
static const char * const exception_stack_names[] = {
|
|
[ ESTACK_DF ] = "#DF",
|
|
[ ESTACK_NMI ] = "NMI",
|
|
[ ESTACK_DB ] = "#DB",
|
|
[ ESTACK_MCE ] = "#MC",
|
|
[ ESTACK_VC ] = "#VC",
|
|
[ ESTACK_VC2 ] = "#VC2",
|
|
};
|
|
|
|
const char *stack_type_name(enum stack_type type)
|
|
{
|
|
BUILD_BUG_ON(N_EXCEPTION_STACKS != 6);
|
|
|
|
if (type == STACK_TYPE_IRQ)
|
|
return "IRQ";
|
|
|
|
if (type == STACK_TYPE_ENTRY) {
|
|
/*
|
|
* On 64-bit, we have a generic entry stack that we
|
|
* use for all the kernel entry points, including
|
|
* SYSENTER.
|
|
*/
|
|
return "ENTRY_TRAMPOLINE";
|
|
}
|
|
|
|
if (type >= STACK_TYPE_EXCEPTION && type <= STACK_TYPE_EXCEPTION_LAST)
|
|
return exception_stack_names[type - STACK_TYPE_EXCEPTION];
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* struct estack_pages - Page descriptor for exception stacks
|
|
* @offs: Offset from the start of the exception stack area
|
|
* @size: Size of the exception stack
|
|
* @type: Type to store in the stack_info struct
|
|
*/
|
|
struct estack_pages {
|
|
u32 offs;
|
|
u16 size;
|
|
u16 type;
|
|
};
|
|
|
|
#define EPAGERANGE(st) \
|
|
[PFN_DOWN(CEA_ESTACK_OFFS(st)) ... \
|
|
PFN_DOWN(CEA_ESTACK_OFFS(st) + CEA_ESTACK_SIZE(st) - 1)] = { \
|
|
.offs = CEA_ESTACK_OFFS(st), \
|
|
.size = CEA_ESTACK_SIZE(st), \
|
|
.type = STACK_TYPE_EXCEPTION + ESTACK_ ##st, }
|
|
|
|
/*
|
|
* Array of exception stack page descriptors. If the stack is larger than
|
|
* PAGE_SIZE, all pages covering a particular stack will have the same
|
|
* info. The guard pages including the not mapped DB2 stack are zeroed
|
|
* out.
|
|
*/
|
|
static const
|
|
struct estack_pages estack_pages[CEA_ESTACK_PAGES] ____cacheline_aligned = {
|
|
EPAGERANGE(DF),
|
|
EPAGERANGE(NMI),
|
|
EPAGERANGE(DB),
|
|
EPAGERANGE(MCE),
|
|
EPAGERANGE(VC),
|
|
EPAGERANGE(VC2),
|
|
};
|
|
|
|
static __always_inline bool in_exception_stack(unsigned long *stack, struct stack_info *info)
|
|
{
|
|
unsigned long begin, end, stk = (unsigned long)stack;
|
|
const struct estack_pages *ep;
|
|
struct pt_regs *regs;
|
|
unsigned int k;
|
|
|
|
BUILD_BUG_ON(N_EXCEPTION_STACKS != 6);
|
|
|
|
begin = (unsigned long)__this_cpu_read(cea_exception_stacks);
|
|
/*
|
|
* Handle the case where stack trace is collected _before_
|
|
* cea_exception_stacks had been initialized.
|
|
*/
|
|
if (!begin)
|
|
return false;
|
|
|
|
end = begin + sizeof(struct cea_exception_stacks);
|
|
/* Bail if @stack is outside the exception stack area. */
|
|
if (stk < begin || stk >= end)
|
|
return false;
|
|
|
|
/* Calc page offset from start of exception stacks */
|
|
k = (stk - begin) >> PAGE_SHIFT;
|
|
/* Lookup the page descriptor */
|
|
ep = &estack_pages[k];
|
|
/* Guard page? */
|
|
if (!ep->size)
|
|
return false;
|
|
|
|
begin += (unsigned long)ep->offs;
|
|
end = begin + (unsigned long)ep->size;
|
|
regs = (struct pt_regs *)end - 1;
|
|
|
|
info->type = ep->type;
|
|
info->begin = (unsigned long *)begin;
|
|
info->end = (unsigned long *)end;
|
|
info->next_sp = (unsigned long *)regs->sp;
|
|
return true;
|
|
}
|
|
|
|
static __always_inline bool in_irq_stack(unsigned long *stack, struct stack_info *info)
|
|
{
|
|
unsigned long *end = (unsigned long *)this_cpu_read(hardirq_stack_ptr);
|
|
unsigned long *begin;
|
|
|
|
/*
|
|
* @end points directly to the top most stack entry to avoid a -8
|
|
* adjustment in the stack switch hotpath. Adjust it back before
|
|
* calculating @begin.
|
|
*/
|
|
end++;
|
|
begin = end - (IRQ_STACK_SIZE / sizeof(long));
|
|
|
|
/*
|
|
* Due to the switching logic RSP can never be == @end because the
|
|
* final operation is 'popq %rsp' which means after that RSP points
|
|
* to the original stack and not to @end.
|
|
*/
|
|
if (stack < begin || stack >= end)
|
|
return false;
|
|
|
|
info->type = STACK_TYPE_IRQ;
|
|
info->begin = begin;
|
|
info->end = end;
|
|
|
|
/*
|
|
* The next stack pointer is stored at the top of the irq stack
|
|
* before switching to the irq stack. Actual stack entries are all
|
|
* below that.
|
|
*/
|
|
info->next_sp = (unsigned long *)*(end - 1);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool noinstr get_stack_info_noinstr(unsigned long *stack, struct task_struct *task,
|
|
struct stack_info *info)
|
|
{
|
|
if (in_task_stack(stack, task, info))
|
|
return true;
|
|
|
|
if (task != current)
|
|
return false;
|
|
|
|
if (in_exception_stack(stack, info))
|
|
return true;
|
|
|
|
if (in_irq_stack(stack, info))
|
|
return true;
|
|
|
|
if (in_entry_stack(stack, info))
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
int get_stack_info(unsigned long *stack, struct task_struct *task,
|
|
struct stack_info *info, unsigned long *visit_mask)
|
|
{
|
|
task = task ? : current;
|
|
|
|
if (!stack)
|
|
goto unknown;
|
|
|
|
if (!get_stack_info_noinstr(stack, task, info))
|
|
goto unknown;
|
|
|
|
/*
|
|
* Make sure we don't iterate through any given stack more than once.
|
|
* If it comes up a second time then there's something wrong going on:
|
|
* just break out and report an unknown stack type.
|
|
*/
|
|
if (visit_mask) {
|
|
if (*visit_mask & (1UL << info->type)) {
|
|
if (task == current)
|
|
printk_deferred_once(KERN_WARNING "WARNING: stack recursion on stack type %d\n", info->type);
|
|
goto unknown;
|
|
}
|
|
*visit_mask |= 1UL << info->type;
|
|
}
|
|
|
|
return 0;
|
|
|
|
unknown:
|
|
info->type = STACK_TYPE_UNKNOWN;
|
|
return -EINVAL;
|
|
}
|