x86, 64bit: Use a #PF handler to materialize early mappings on demand
Linear mode (CR0.PG = 0) is mutually exclusive with 64-bit mode; all 64-bit code has to use page tables. This makes it awkward before we have first set up properly all-covering page tables to access objects that are outside the static kernel range. So far we have dealt with that simply by mapping a fixed amount of low memory, but that fails in at least two upcoming use cases: 1. We will support load and run kernel, struct boot_params, ramdisk, command line, etc. above the 4 GiB mark. 2. need to access ramdisk early to get microcode to update that as early possible. We could use early_iomap to access them too, but it will make code to messy and hard to be unified with 32 bit. Hence, set up a #PF table and use a fixed number of buffers to set up page tables on demand. If the buffers fill up then we simply flush them and start over. These buffers are all in __initdata, so it does not increase RAM usage at runtime. Thus, with the help of the #PF handler, we can set the final kernel mapping from blank, and switch to init_level4_pgt later. During the switchover in head_64.S, before #PF handler is available, we use three pages to handle kernel crossing 1G, 512G boundaries with sharing page by playing games with page aliasing: the same page is mapped twice in the higher-level tables with appropriate wraparound. The kernel region itself will be properly mapped; other mappings may be spurious. early_make_pgtable is using kernel high mapping address to access pages to set page table. -v4: Add phys_base offset to make kexec happy, and add init_mapping_kernel() - Yinghai -v5: fix compiling with xen, and add back ident level3 and level2 for xen also move back init_level4_pgt from BSS to DATA again. because we have to clear it anyway. - Yinghai -v6: switch to init_level4_pgt in init_mem_mapping. - Yinghai -v7: remove not needed clear_page for init_level4_page it is with fill 512,8,0 already in head_64.S - Yinghai -v8: we need to keep that handler alive until init_mem_mapping and don't let early_trap_init to trash that early #PF handler. So split early_trap_pf_init out and move it down. - Yinghai -v9: switchover only cover kernel space instead of 1G so could avoid touch possible mem holes. - Yinghai -v11: change far jmp back to far return to initial_code, that is needed to fix failure that is reported by Konrad on AMD systems. - Yinghai Signed-off-by: Yinghai Lu <yinghai@kernel.org> Link: http://lkml.kernel.org/r/1359058816-7615-12-git-send-email-yinghai@kernel.org Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
This commit is contained in:
parent
4f7b92263a
commit
8170e6bed4
@ -1,6 +1,8 @@
|
|||||||
#ifndef _ASM_X86_PGTABLE_64_DEFS_H
|
#ifndef _ASM_X86_PGTABLE_64_DEFS_H
|
||||||
#define _ASM_X86_PGTABLE_64_DEFS_H
|
#define _ASM_X86_PGTABLE_64_DEFS_H
|
||||||
|
|
||||||
|
#include <asm/sparsemem.h>
|
||||||
|
|
||||||
#ifndef __ASSEMBLY__
|
#ifndef __ASSEMBLY__
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
|
|
||||||
@ -60,4 +62,6 @@ typedef struct { pteval_t pte; } pte_t;
|
|||||||
#define MODULES_END _AC(0xffffffffff000000, UL)
|
#define MODULES_END _AC(0xffffffffff000000, UL)
|
||||||
#define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
#define MODULES_LEN (MODULES_END - MODULES_VADDR)
|
||||||
|
|
||||||
|
#define EARLY_DYNAMIC_PAGE_TABLES 64
|
||||||
|
|
||||||
#endif /* _ASM_X86_PGTABLE_64_DEFS_H */
|
#endif /* _ASM_X86_PGTABLE_64_DEFS_H */
|
||||||
|
@ -731,6 +731,7 @@ extern void enable_sep_cpu(void);
|
|||||||
extern int sysenter_setup(void);
|
extern int sysenter_setup(void);
|
||||||
|
|
||||||
extern void early_trap_init(void);
|
extern void early_trap_init(void);
|
||||||
|
void early_trap_pf_init(void);
|
||||||
|
|
||||||
/* Defined in head.S */
|
/* Defined in head.S */
|
||||||
extern struct desc_ptr early_gdt_descr;
|
extern struct desc_ptr early_gdt_descr;
|
||||||
|
@ -27,11 +27,73 @@
|
|||||||
#include <asm/bios_ebda.h>
|
#include <asm/bios_ebda.h>
|
||||||
#include <asm/bootparam_utils.h>
|
#include <asm/bootparam_utils.h>
|
||||||
|
|
||||||
static void __init zap_identity_mappings(void)
|
/*
|
||||||
|
* Manage page tables very early on.
|
||||||
|
*/
|
||||||
|
extern pgd_t early_level4_pgt[PTRS_PER_PGD];
|
||||||
|
extern pmd_t early_dynamic_pgts[EARLY_DYNAMIC_PAGE_TABLES][PTRS_PER_PMD];
|
||||||
|
static unsigned int __initdata next_early_pgt = 2;
|
||||||
|
|
||||||
|
/* Wipe all early page tables except for the kernel symbol map */
|
||||||
|
static void __init reset_early_page_tables(void)
|
||||||
{
|
{
|
||||||
pgd_t *pgd = pgd_offset_k(0UL);
|
unsigned long i;
|
||||||
pgd_clear(pgd);
|
|
||||||
__flush_tlb_all();
|
for (i = 0; i < PTRS_PER_PGD-1; i++)
|
||||||
|
early_level4_pgt[i].pgd = 0;
|
||||||
|
|
||||||
|
next_early_pgt = 0;
|
||||||
|
|
||||||
|
write_cr3(__pa(early_level4_pgt));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* Create a new PMD entry */
|
||||||
|
int __init early_make_pgtable(unsigned long address)
|
||||||
|
{
|
||||||
|
unsigned long physaddr = address - __PAGE_OFFSET;
|
||||||
|
unsigned long i;
|
||||||
|
pgdval_t pgd, *pgd_p;
|
||||||
|
pudval_t *pud_p;
|
||||||
|
pmdval_t pmd, *pmd_p;
|
||||||
|
|
||||||
|
/* Invalid address or early pgt is done ? */
|
||||||
|
if (physaddr >= MAXMEM || read_cr3() != __pa(early_level4_pgt))
|
||||||
|
return -1;
|
||||||
|
|
||||||
|
i = (address >> PGDIR_SHIFT) & (PTRS_PER_PGD - 1);
|
||||||
|
pgd_p = &early_level4_pgt[i].pgd;
|
||||||
|
pgd = *pgd_p;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* The use of __START_KERNEL_map rather than __PAGE_OFFSET here is
|
||||||
|
* critical -- __PAGE_OFFSET would point us back into the dynamic
|
||||||
|
* range and we might end up looping forever...
|
||||||
|
*/
|
||||||
|
if (pgd && next_early_pgt < EARLY_DYNAMIC_PAGE_TABLES) {
|
||||||
|
pud_p = (pudval_t *)((pgd & PTE_PFN_MASK) + __START_KERNEL_map - phys_base);
|
||||||
|
} else {
|
||||||
|
if (next_early_pgt >= EARLY_DYNAMIC_PAGE_TABLES-1)
|
||||||
|
reset_early_page_tables();
|
||||||
|
|
||||||
|
pud_p = (pudval_t *)early_dynamic_pgts[next_early_pgt++];
|
||||||
|
for (i = 0; i < PTRS_PER_PUD; i++)
|
||||||
|
pud_p[i] = 0;
|
||||||
|
|
||||||
|
*pgd_p = (pgdval_t)pud_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE;
|
||||||
|
}
|
||||||
|
i = (address >> PUD_SHIFT) & (PTRS_PER_PUD - 1);
|
||||||
|
pud_p += i;
|
||||||
|
|
||||||
|
pmd_p = (pmdval_t *)early_dynamic_pgts[next_early_pgt++];
|
||||||
|
pmd = (physaddr & PUD_MASK) + (__PAGE_KERNEL_LARGE & ~_PAGE_GLOBAL);
|
||||||
|
for (i = 0; i < PTRS_PER_PMD; i++) {
|
||||||
|
pmd_p[i] = pmd;
|
||||||
|
pmd += PMD_SIZE;
|
||||||
|
}
|
||||||
|
|
||||||
|
*pud_p = (pudval_t)pmd_p - __START_KERNEL_map + phys_base + _KERNPG_TABLE;
|
||||||
|
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Don't add a printk in there. printk relies on the PDA which is not initialized
|
/* Don't add a printk in there. printk relies on the PDA which is not initialized
|
||||||
@ -72,12 +134,13 @@ void __init x86_64_start_kernel(char * real_mode_data)
|
|||||||
(__START_KERNEL & PGDIR_MASK)));
|
(__START_KERNEL & PGDIR_MASK)));
|
||||||
BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END);
|
BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END);
|
||||||
|
|
||||||
|
/* Kill off the identity-map trampoline */
|
||||||
|
reset_early_page_tables();
|
||||||
|
|
||||||
/* clear bss before set_intr_gate with early_idt_handler */
|
/* clear bss before set_intr_gate with early_idt_handler */
|
||||||
clear_bss();
|
clear_bss();
|
||||||
|
|
||||||
/* Make NULL pointers segfault */
|
/* XXX - this is wrong... we need to build page tables from scratch */
|
||||||
zap_identity_mappings();
|
|
||||||
|
|
||||||
max_pfn_mapped = KERNEL_IMAGE_SIZE >> PAGE_SHIFT;
|
max_pfn_mapped = KERNEL_IMAGE_SIZE >> PAGE_SHIFT;
|
||||||
|
|
||||||
for (i = 0; i < NUM_EXCEPTION_VECTORS; i++) {
|
for (i = 0; i < NUM_EXCEPTION_VECTORS; i++) {
|
||||||
@ -94,6 +157,10 @@ void __init x86_64_start_kernel(char * real_mode_data)
|
|||||||
if (console_loglevel == 10)
|
if (console_loglevel == 10)
|
||||||
early_printk("Kernel alive\n");
|
early_printk("Kernel alive\n");
|
||||||
|
|
||||||
|
clear_page(init_level4_pgt);
|
||||||
|
/* set init_level4_pgt kernel high mapping*/
|
||||||
|
init_level4_pgt[511] = early_level4_pgt[511];
|
||||||
|
|
||||||
x86_64_start_reservations(real_mode_data);
|
x86_64_start_reservations(real_mode_data);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -47,14 +47,13 @@ L3_START_KERNEL = pud_index(__START_KERNEL_map)
|
|||||||
.code64
|
.code64
|
||||||
.globl startup_64
|
.globl startup_64
|
||||||
startup_64:
|
startup_64:
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* At this point the CPU runs in 64bit mode CS.L = 1 CS.D = 1,
|
* At this point the CPU runs in 64bit mode CS.L = 1 CS.D = 1,
|
||||||
* and someone has loaded an identity mapped page table
|
* and someone has loaded an identity mapped page table
|
||||||
* for us. These identity mapped page tables map all of the
|
* for us. These identity mapped page tables map all of the
|
||||||
* kernel pages and possibly all of memory.
|
* kernel pages and possibly all of memory.
|
||||||
*
|
*
|
||||||
* %esi holds a physical pointer to real_mode_data.
|
* %rsi holds a physical pointer to real_mode_data.
|
||||||
*
|
*
|
||||||
* We come here either directly from a 64bit bootloader, or from
|
* We come here either directly from a 64bit bootloader, or from
|
||||||
* arch/x86_64/boot/compressed/head.S.
|
* arch/x86_64/boot/compressed/head.S.
|
||||||
@ -66,7 +65,8 @@ startup_64:
|
|||||||
* tables and then reload them.
|
* tables and then reload them.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/* Compute the delta between the address I am compiled to run at and the
|
/*
|
||||||
|
* Compute the delta between the address I am compiled to run at and the
|
||||||
* address I am actually running at.
|
* address I am actually running at.
|
||||||
*/
|
*/
|
||||||
leaq _text(%rip), %rbp
|
leaq _text(%rip), %rbp
|
||||||
@ -78,45 +78,62 @@ startup_64:
|
|||||||
testl %eax, %eax
|
testl %eax, %eax
|
||||||
jnz bad_address
|
jnz bad_address
|
||||||
|
|
||||||
/* Is the address too large? */
|
/*
|
||||||
leaq _text(%rip), %rdx
|
* Is the address too large?
|
||||||
movq $PGDIR_SIZE, %rax
|
|
||||||
cmpq %rax, %rdx
|
|
||||||
jae bad_address
|
|
||||||
|
|
||||||
/* Fixup the physical addresses in the page table
|
|
||||||
*/
|
*/
|
||||||
addq %rbp, init_level4_pgt + 0(%rip)
|
leaq _text(%rip), %rax
|
||||||
addq %rbp, init_level4_pgt + (L4_PAGE_OFFSET*8)(%rip)
|
shrq $MAX_PHYSMEM_BITS, %rax
|
||||||
addq %rbp, init_level4_pgt + (L4_START_KERNEL*8)(%rip)
|
jnz bad_address
|
||||||
|
|
||||||
addq %rbp, level3_ident_pgt + 0(%rip)
|
/*
|
||||||
|
* Fixup the physical addresses in the page table
|
||||||
|
*/
|
||||||
|
addq %rbp, early_level4_pgt + (L4_START_KERNEL*8)(%rip)
|
||||||
|
|
||||||
addq %rbp, level3_kernel_pgt + (510*8)(%rip)
|
addq %rbp, level3_kernel_pgt + (510*8)(%rip)
|
||||||
addq %rbp, level3_kernel_pgt + (511*8)(%rip)
|
addq %rbp, level3_kernel_pgt + (511*8)(%rip)
|
||||||
|
|
||||||
addq %rbp, level2_fixmap_pgt + (506*8)(%rip)
|
addq %rbp, level2_fixmap_pgt + (506*8)(%rip)
|
||||||
|
|
||||||
/* Add an Identity mapping if I am above 1G */
|
/*
|
||||||
|
* Set up the identity mapping for the switchover. These
|
||||||
|
* entries should *NOT* have the global bit set! This also
|
||||||
|
* creates a bunch of nonsense entries but that is fine --
|
||||||
|
* it avoids problems around wraparound.
|
||||||
|
*/
|
||||||
leaq _text(%rip), %rdi
|
leaq _text(%rip), %rdi
|
||||||
andq $PMD_PAGE_MASK, %rdi
|
leaq early_level4_pgt(%rip), %rbx
|
||||||
|
|
||||||
|
movq %rdi, %rax
|
||||||
|
shrq $PGDIR_SHIFT, %rax
|
||||||
|
|
||||||
|
leaq (4096 + _KERNPG_TABLE)(%rbx), %rdx
|
||||||
|
movq %rdx, 0(%rbx,%rax,8)
|
||||||
|
movq %rdx, 8(%rbx,%rax,8)
|
||||||
|
|
||||||
|
addq $4096, %rdx
|
||||||
movq %rdi, %rax
|
movq %rdi, %rax
|
||||||
shrq $PUD_SHIFT, %rax
|
shrq $PUD_SHIFT, %rax
|
||||||
andq $(PTRS_PER_PUD - 1), %rax
|
andl $(PTRS_PER_PUD-1), %eax
|
||||||
jz ident_complete
|
movq %rdx, (4096+0)(%rbx,%rax,8)
|
||||||
|
movq %rdx, (4096+8)(%rbx,%rax,8)
|
||||||
leaq (level2_spare_pgt - __START_KERNEL_map + _KERNPG_TABLE)(%rbp), %rdx
|
|
||||||
leaq level3_ident_pgt(%rip), %rbx
|
|
||||||
movq %rdx, 0(%rbx, %rax, 8)
|
|
||||||
|
|
||||||
|
addq $8192, %rbx
|
||||||
movq %rdi, %rax
|
movq %rdi, %rax
|
||||||
shrq $PMD_SHIFT, %rax
|
shrq $PMD_SHIFT, %rdi
|
||||||
andq $(PTRS_PER_PMD - 1), %rax
|
addq $(__PAGE_KERNEL_LARGE_EXEC & ~_PAGE_GLOBAL), %rax
|
||||||
leaq __PAGE_KERNEL_IDENT_LARGE_EXEC(%rdi), %rdx
|
leaq (_end - 1)(%rip), %rcx
|
||||||
leaq level2_spare_pgt(%rip), %rbx
|
shrq $PMD_SHIFT, %rcx
|
||||||
movq %rdx, 0(%rbx, %rax, 8)
|
subq %rdi, %rcx
|
||||||
ident_complete:
|
incl %ecx
|
||||||
|
|
||||||
|
1:
|
||||||
|
andq $(PTRS_PER_PMD - 1), %rdi
|
||||||
|
movq %rax, (%rbx,%rdi,8)
|
||||||
|
incq %rdi
|
||||||
|
addq $PMD_SIZE, %rax
|
||||||
|
decl %ecx
|
||||||
|
jnz 1b
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Fixup the kernel text+data virtual addresses. Note that
|
* Fixup the kernel text+data virtual addresses. Note that
|
||||||
@ -124,7 +141,6 @@ ident_complete:
|
|||||||
* cleanup_highmap() fixes this up along with the mappings
|
* cleanup_highmap() fixes this up along with the mappings
|
||||||
* beyond _end.
|
* beyond _end.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
leaq level2_kernel_pgt(%rip), %rdi
|
leaq level2_kernel_pgt(%rip), %rdi
|
||||||
leaq 4096(%rdi), %r8
|
leaq 4096(%rdi), %r8
|
||||||
/* See if it is a valid page table entry */
|
/* See if it is a valid page table entry */
|
||||||
@ -139,17 +155,14 @@ ident_complete:
|
|||||||
/* Fixup phys_base */
|
/* Fixup phys_base */
|
||||||
addq %rbp, phys_base(%rip)
|
addq %rbp, phys_base(%rip)
|
||||||
|
|
||||||
/* Due to ENTRY(), sometimes the empty space gets filled with
|
movq $(early_level4_pgt - __START_KERNEL_map), %rax
|
||||||
* zeros. Better take a jmp than relying on empty space being
|
jmp 1f
|
||||||
* filled with 0x90 (nop)
|
|
||||||
*/
|
|
||||||
jmp secondary_startup_64
|
|
||||||
ENTRY(secondary_startup_64)
|
ENTRY(secondary_startup_64)
|
||||||
/*
|
/*
|
||||||
* At this point the CPU runs in 64bit mode CS.L = 1 CS.D = 1,
|
* At this point the CPU runs in 64bit mode CS.L = 1 CS.D = 1,
|
||||||
* and someone has loaded a mapped page table.
|
* and someone has loaded a mapped page table.
|
||||||
*
|
*
|
||||||
* %esi holds a physical pointer to real_mode_data.
|
* %rsi holds a physical pointer to real_mode_data.
|
||||||
*
|
*
|
||||||
* We come here either from startup_64 (using physical addresses)
|
* We come here either from startup_64 (using physical addresses)
|
||||||
* or from trampoline.S (using virtual addresses).
|
* or from trampoline.S (using virtual addresses).
|
||||||
@ -159,12 +172,14 @@ ENTRY(secondary_startup_64)
|
|||||||
* after the boot processor executes this code.
|
* after the boot processor executes this code.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
movq $(init_level4_pgt - __START_KERNEL_map), %rax
|
||||||
|
1:
|
||||||
|
|
||||||
/* Enable PAE mode and PGE */
|
/* Enable PAE mode and PGE */
|
||||||
movl $(X86_CR4_PAE | X86_CR4_PGE), %eax
|
movl $(X86_CR4_PAE | X86_CR4_PGE), %ecx
|
||||||
movq %rax, %cr4
|
movq %rcx, %cr4
|
||||||
|
|
||||||
/* Setup early boot stage 4 level pagetables. */
|
/* Setup early boot stage 4 level pagetables. */
|
||||||
movq $(init_level4_pgt - __START_KERNEL_map), %rax
|
|
||||||
addq phys_base(%rip), %rax
|
addq phys_base(%rip), %rax
|
||||||
movq %rax, %cr3
|
movq %rax, %cr3
|
||||||
|
|
||||||
@ -196,7 +211,7 @@ ENTRY(secondary_startup_64)
|
|||||||
movq %rax, %cr0
|
movq %rax, %cr0
|
||||||
|
|
||||||
/* Setup a boot time stack */
|
/* Setup a boot time stack */
|
||||||
movq stack_start(%rip),%rsp
|
movq stack_start(%rip), %rsp
|
||||||
|
|
||||||
/* zero EFLAGS after setting rsp */
|
/* zero EFLAGS after setting rsp */
|
||||||
pushq $0
|
pushq $0
|
||||||
@ -236,15 +251,33 @@ ENTRY(secondary_startup_64)
|
|||||||
movl initial_gs+4(%rip),%edx
|
movl initial_gs+4(%rip),%edx
|
||||||
wrmsr
|
wrmsr
|
||||||
|
|
||||||
/* esi is pointer to real mode structure with interesting info.
|
/* rsi is pointer to real mode structure with interesting info.
|
||||||
pass it to C */
|
pass it to C */
|
||||||
movl %esi, %edi
|
movq %rsi, %rdi
|
||||||
|
|
||||||
/* Finally jump to run C code and to be on real kernel address
|
/* Finally jump to run C code and to be on real kernel address
|
||||||
* Since we are running on identity-mapped space we have to jump
|
* Since we are running on identity-mapped space we have to jump
|
||||||
* to the full 64bit address, this is only possible as indirect
|
* to the full 64bit address, this is only possible as indirect
|
||||||
* jump. In addition we need to ensure %cs is set so we make this
|
* jump. In addition we need to ensure %cs is set so we make this
|
||||||
* a far return.
|
* a far return.
|
||||||
|
*
|
||||||
|
* Note: do not change to far jump indirect with 64bit offset.
|
||||||
|
*
|
||||||
|
* AMD does not support far jump indirect with 64bit offset.
|
||||||
|
* AMD64 Architecture Programmer's Manual, Volume 3: states only
|
||||||
|
* JMP FAR mem16:16 FF /5 Far jump indirect,
|
||||||
|
* with the target specified by a far pointer in memory.
|
||||||
|
* JMP FAR mem16:32 FF /5 Far jump indirect,
|
||||||
|
* with the target specified by a far pointer in memory.
|
||||||
|
*
|
||||||
|
* Intel64 does support 64bit offset.
|
||||||
|
* Software Developer Manual Vol 2: states:
|
||||||
|
* FF /5 JMP m16:16 Jump far, absolute indirect,
|
||||||
|
* address given in m16:16
|
||||||
|
* FF /5 JMP m16:32 Jump far, absolute indirect,
|
||||||
|
* address given in m16:32.
|
||||||
|
* REX.W + FF /5 JMP m16:64 Jump far, absolute indirect,
|
||||||
|
* address given in m16:64.
|
||||||
*/
|
*/
|
||||||
movq initial_code(%rip),%rax
|
movq initial_code(%rip),%rax
|
||||||
pushq $0 # fake return address to stop unwinder
|
pushq $0 # fake return address to stop unwinder
|
||||||
@ -270,13 +303,13 @@ ENDPROC(start_cpu0)
|
|||||||
|
|
||||||
/* SMP bootup changes these two */
|
/* SMP bootup changes these two */
|
||||||
__REFDATA
|
__REFDATA
|
||||||
.align 8
|
.balign 8
|
||||||
ENTRY(initial_code)
|
GLOBAL(initial_code)
|
||||||
.quad x86_64_start_kernel
|
.quad x86_64_start_kernel
|
||||||
ENTRY(initial_gs)
|
GLOBAL(initial_gs)
|
||||||
.quad INIT_PER_CPU_VAR(irq_stack_union)
|
.quad INIT_PER_CPU_VAR(irq_stack_union)
|
||||||
|
|
||||||
ENTRY(stack_start)
|
GLOBAL(stack_start)
|
||||||
.quad init_thread_union+THREAD_SIZE-8
|
.quad init_thread_union+THREAD_SIZE-8
|
||||||
.word 0
|
.word 0
|
||||||
__FINITDATA
|
__FINITDATA
|
||||||
@ -284,7 +317,7 @@ ENDPROC(start_cpu0)
|
|||||||
bad_address:
|
bad_address:
|
||||||
jmp bad_address
|
jmp bad_address
|
||||||
|
|
||||||
.section ".init.text","ax"
|
__INIT
|
||||||
.globl early_idt_handlers
|
.globl early_idt_handlers
|
||||||
early_idt_handlers:
|
early_idt_handlers:
|
||||||
# 104(%rsp) %rflags
|
# 104(%rsp) %rflags
|
||||||
@ -321,14 +354,22 @@ ENTRY(early_idt_handler)
|
|||||||
pushq %r11 # 0(%rsp)
|
pushq %r11 # 0(%rsp)
|
||||||
|
|
||||||
cmpl $__KERNEL_CS,96(%rsp)
|
cmpl $__KERNEL_CS,96(%rsp)
|
||||||
jne 10f
|
jne 11f
|
||||||
|
|
||||||
|
cmpl $14,72(%rsp) # Page fault?
|
||||||
|
jnz 10f
|
||||||
|
GET_CR2_INTO(%rdi) # can clobber any volatile register if pv
|
||||||
|
call early_make_pgtable
|
||||||
|
andl %eax,%eax
|
||||||
|
jz 20f # All good
|
||||||
|
|
||||||
|
10:
|
||||||
leaq 88(%rsp),%rdi # Pointer to %rip
|
leaq 88(%rsp),%rdi # Pointer to %rip
|
||||||
call early_fixup_exception
|
call early_fixup_exception
|
||||||
andl %eax,%eax
|
andl %eax,%eax
|
||||||
jnz 20f # Found an exception entry
|
jnz 20f # Found an exception entry
|
||||||
|
|
||||||
10:
|
11:
|
||||||
#ifdef CONFIG_EARLY_PRINTK
|
#ifdef CONFIG_EARLY_PRINTK
|
||||||
GET_CR2_INTO(%r9) # can clobber any volatile register if pv
|
GET_CR2_INTO(%r9) # can clobber any volatile register if pv
|
||||||
movl 80(%rsp),%r8d # error code
|
movl 80(%rsp),%r8d # error code
|
||||||
@ -350,7 +391,7 @@ ENTRY(early_idt_handler)
|
|||||||
1: hlt
|
1: hlt
|
||||||
jmp 1b
|
jmp 1b
|
||||||
|
|
||||||
20: # Exception table entry found
|
20: # Exception table entry found or page table generated
|
||||||
popq %r11
|
popq %r11
|
||||||
popq %r10
|
popq %r10
|
||||||
popq %r9
|
popq %r9
|
||||||
@ -364,6 +405,8 @@ ENTRY(early_idt_handler)
|
|||||||
decl early_recursion_flag(%rip)
|
decl early_recursion_flag(%rip)
|
||||||
INTERRUPT_RETURN
|
INTERRUPT_RETURN
|
||||||
|
|
||||||
|
__INITDATA
|
||||||
|
|
||||||
.balign 4
|
.balign 4
|
||||||
early_recursion_flag:
|
early_recursion_flag:
|
||||||
.long 0
|
.long 0
|
||||||
@ -374,11 +417,10 @@ early_idt_msg:
|
|||||||
early_idt_ripmsg:
|
early_idt_ripmsg:
|
||||||
.asciz "RIP %s\n"
|
.asciz "RIP %s\n"
|
||||||
#endif /* CONFIG_EARLY_PRINTK */
|
#endif /* CONFIG_EARLY_PRINTK */
|
||||||
.previous
|
|
||||||
|
|
||||||
#define NEXT_PAGE(name) \
|
#define NEXT_PAGE(name) \
|
||||||
.balign PAGE_SIZE; \
|
.balign PAGE_SIZE; \
|
||||||
ENTRY(name)
|
GLOBAL(name)
|
||||||
|
|
||||||
/* Automate the creation of 1 to 1 mapping pmd entries */
|
/* Automate the creation of 1 to 1 mapping pmd entries */
|
||||||
#define PMDS(START, PERM, COUNT) \
|
#define PMDS(START, PERM, COUNT) \
|
||||||
@ -388,24 +430,37 @@ ENTRY(name)
|
|||||||
i = i + 1 ; \
|
i = i + 1 ; \
|
||||||
.endr
|
.endr
|
||||||
|
|
||||||
.data
|
__INITDATA
|
||||||
/*
|
NEXT_PAGE(early_level4_pgt)
|
||||||
* This default setting generates an ident mapping at address 0x100000
|
.fill 511,8,0
|
||||||
* and a mapping for the kernel that precisely maps virtual address
|
|
||||||
* 0xffffffff80000000 to physical address 0x000000. (always using
|
|
||||||
* 2Mbyte large pages provided by PAE mode)
|
|
||||||
*/
|
|
||||||
NEXT_PAGE(init_level4_pgt)
|
|
||||||
.quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
|
||||||
.org init_level4_pgt + L4_PAGE_OFFSET*8, 0
|
|
||||||
.quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
|
||||||
.org init_level4_pgt + L4_START_KERNEL*8, 0
|
|
||||||
/* (2^48-(2*1024*1024*1024))/(2^39) = 511 */
|
|
||||||
.quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE
|
.quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE
|
||||||
|
|
||||||
|
NEXT_PAGE(early_dynamic_pgts)
|
||||||
|
.fill 512*EARLY_DYNAMIC_PAGE_TABLES,8,0
|
||||||
|
|
||||||
|
.data
|
||||||
|
|
||||||
|
#ifndef CONFIG_XEN
|
||||||
|
NEXT_PAGE(init_level4_pgt)
|
||||||
|
.fill 512,8,0
|
||||||
|
#else
|
||||||
|
NEXT_PAGE(init_level4_pgt)
|
||||||
|
.quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
||||||
|
.org init_level4_pgt + L4_PAGE_OFFSET*8, 0
|
||||||
|
.quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
||||||
|
.org init_level4_pgt + L4_START_KERNEL*8, 0
|
||||||
|
/* (2^48-(2*1024*1024*1024))/(2^39) = 511 */
|
||||||
|
.quad level3_kernel_pgt - __START_KERNEL_map + _PAGE_TABLE
|
||||||
|
|
||||||
NEXT_PAGE(level3_ident_pgt)
|
NEXT_PAGE(level3_ident_pgt)
|
||||||
.quad level2_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
.quad level2_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
||||||
.fill 511,8,0
|
.fill 511, 8, 0
|
||||||
|
NEXT_PAGE(level2_ident_pgt)
|
||||||
|
/* Since I easily can, map the first 1G.
|
||||||
|
* Don't set NX because code runs from these pages.
|
||||||
|
*/
|
||||||
|
PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, PTRS_PER_PMD)
|
||||||
|
#endif
|
||||||
|
|
||||||
NEXT_PAGE(level3_kernel_pgt)
|
NEXT_PAGE(level3_kernel_pgt)
|
||||||
.fill L3_START_KERNEL,8,0
|
.fill L3_START_KERNEL,8,0
|
||||||
@ -413,21 +468,6 @@ NEXT_PAGE(level3_kernel_pgt)
|
|||||||
.quad level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
.quad level2_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE
|
||||||
.quad level2_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
|
.quad level2_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
|
||||||
|
|
||||||
NEXT_PAGE(level2_fixmap_pgt)
|
|
||||||
.fill 506,8,0
|
|
||||||
.quad level1_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
|
|
||||||
/* 8MB reserved for vsyscalls + a 2MB hole = 4 + 1 entries */
|
|
||||||
.fill 5,8,0
|
|
||||||
|
|
||||||
NEXT_PAGE(level1_fixmap_pgt)
|
|
||||||
.fill 512,8,0
|
|
||||||
|
|
||||||
NEXT_PAGE(level2_ident_pgt)
|
|
||||||
/* Since I easily can, map the first 1G.
|
|
||||||
* Don't set NX because code runs from these pages.
|
|
||||||
*/
|
|
||||||
PMDS(0, __PAGE_KERNEL_IDENT_LARGE_EXEC, PTRS_PER_PMD)
|
|
||||||
|
|
||||||
NEXT_PAGE(level2_kernel_pgt)
|
NEXT_PAGE(level2_kernel_pgt)
|
||||||
/*
|
/*
|
||||||
* 512 MB kernel mapping. We spend a full page on this pagetable
|
* 512 MB kernel mapping. We spend a full page on this pagetable
|
||||||
@ -442,11 +482,16 @@ NEXT_PAGE(level2_kernel_pgt)
|
|||||||
PMDS(0, __PAGE_KERNEL_LARGE_EXEC,
|
PMDS(0, __PAGE_KERNEL_LARGE_EXEC,
|
||||||
KERNEL_IMAGE_SIZE/PMD_SIZE)
|
KERNEL_IMAGE_SIZE/PMD_SIZE)
|
||||||
|
|
||||||
NEXT_PAGE(level2_spare_pgt)
|
NEXT_PAGE(level2_fixmap_pgt)
|
||||||
.fill 512, 8, 0
|
.fill 506,8,0
|
||||||
|
.quad level1_fixmap_pgt - __START_KERNEL_map + _PAGE_TABLE
|
||||||
|
/* 8MB reserved for vsyscalls + a 2MB hole = 4 + 1 entries */
|
||||||
|
.fill 5,8,0
|
||||||
|
|
||||||
|
NEXT_PAGE(level1_fixmap_pgt)
|
||||||
|
.fill 512,8,0
|
||||||
|
|
||||||
#undef PMDS
|
#undef PMDS
|
||||||
#undef NEXT_PAGE
|
|
||||||
|
|
||||||
.data
|
.data
|
||||||
.align 16
|
.align 16
|
||||||
@ -472,6 +517,5 @@ ENTRY(nmi_idt_table)
|
|||||||
.skip IDT_ENTRIES * 16
|
.skip IDT_ENTRIES * 16
|
||||||
|
|
||||||
__PAGE_ALIGNED_BSS
|
__PAGE_ALIGNED_BSS
|
||||||
.align PAGE_SIZE
|
NEXT_PAGE(empty_zero_page)
|
||||||
ENTRY(empty_zero_page)
|
|
||||||
.skip PAGE_SIZE
|
.skip PAGE_SIZE
|
||||||
|
@ -1005,6 +1005,8 @@ void __init setup_arch(char **cmdline_p)
|
|||||||
|
|
||||||
init_mem_mapping();
|
init_mem_mapping();
|
||||||
|
|
||||||
|
early_trap_pf_init();
|
||||||
|
|
||||||
setup_real_mode();
|
setup_real_mode();
|
||||||
|
|
||||||
memblock.current_limit = get_max_mapped();
|
memblock.current_limit = get_max_mapped();
|
||||||
|
@ -688,10 +688,19 @@ void __init early_trap_init(void)
|
|||||||
set_intr_gate_ist(X86_TRAP_DB, &debug, DEBUG_STACK);
|
set_intr_gate_ist(X86_TRAP_DB, &debug, DEBUG_STACK);
|
||||||
/* int3 can be called from all */
|
/* int3 can be called from all */
|
||||||
set_system_intr_gate_ist(X86_TRAP_BP, &int3, DEBUG_STACK);
|
set_system_intr_gate_ist(X86_TRAP_BP, &int3, DEBUG_STACK);
|
||||||
|
#ifdef CONFIG_X86_32
|
||||||
set_intr_gate(X86_TRAP_PF, &page_fault);
|
set_intr_gate(X86_TRAP_PF, &page_fault);
|
||||||
|
#endif
|
||||||
load_idt(&idt_descr);
|
load_idt(&idt_descr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void __init early_trap_pf_init(void)
|
||||||
|
{
|
||||||
|
#ifdef CONFIG_X86_64
|
||||||
|
set_intr_gate(X86_TRAP_PF, &page_fault);
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
void __init trap_init(void)
|
void __init trap_init(void)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
|
@ -446,9 +446,10 @@ void __init init_mem_mapping(void)
|
|||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
early_ioremap_page_table_range_init();
|
early_ioremap_page_table_range_init();
|
||||||
|
#endif
|
||||||
|
|
||||||
load_cr3(swapper_pg_dir);
|
load_cr3(swapper_pg_dir);
|
||||||
__flush_tlb_all();
|
__flush_tlb_all();
|
||||||
#endif
|
|
||||||
|
|
||||||
early_memtest(0, max_pfn_mapped << PAGE_SHIFT);
|
early_memtest(0, max_pfn_mapped << PAGE_SHIFT);
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user