Make sure that inline assembler that expects 'r' operand receives 32 bit value. Before this fix in case of CONFIG_ARCH_PHYS_ADDR_T_64BIT and CONFIG_ARM_PATCH_PHYS_VIRT __phys_to_virt function passed 64 bit value to __pv_stub inline assembler where 'r' operand is expected. Compiler behavior in such case is not well specified. It worked in little endian case, but in big endian case incorrect code was generated, where compiler confused which part of 64 bit value it needed to modify. For example BE snippet looked like this: N:0x80904E08 : MOV r2,#0 N:0x80904E0C : SUB r2,r2,#0x81000000 when LE similar code looked like this N:0x808FCE2C : MOV r2,r0 N:0x808FCE30 : SUB r2,r2,#0xc0, 8 ; #0xc0000000 Note 'r0' register is va that have to be translated into phys To avoid this situation use explicit cast to 'unsigned long', which explicitly discard upper part of phys address and convert value to 32 bit. Also add comment so such cast will not be removed in the future. Signed-off-by: Victor Kamensky <victor.kamensky@linaro.org> Acked-by: Santosh Shilimkar <santosh.shilimkar@ti.com> Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
360 lines
9.2 KiB
C
360 lines
9.2 KiB
C
/*
|
|
* arch/arm/include/asm/memory.h
|
|
*
|
|
* Copyright (C) 2000-2002 Russell King
|
|
* modification for nommu, Hyok S. Choi, 2004
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* Note: this file should not be included by non-asm/.h files
|
|
*/
|
|
#ifndef __ASM_ARM_MEMORY_H
|
|
#define __ASM_ARM_MEMORY_H
|
|
|
|
#include <linux/compiler.h>
|
|
#include <linux/const.h>
|
|
#include <linux/types.h>
|
|
#include <linux/sizes.h>
|
|
|
|
#include <asm/cache.h>
|
|
|
|
#ifdef CONFIG_NEED_MACH_MEMORY_H
|
|
#include <mach/memory.h>
|
|
#endif
|
|
|
|
/*
|
|
* Allow for constants defined here to be used from assembly code
|
|
* by prepending the UL suffix only with actual C code compilation.
|
|
*/
|
|
#define UL(x) _AC(x, UL)
|
|
|
|
#ifdef CONFIG_MMU
|
|
|
|
/*
|
|
* PAGE_OFFSET - the virtual address of the start of the kernel image
|
|
* TASK_SIZE - the maximum size of a user space task.
|
|
* TASK_UNMAPPED_BASE - the lower boundary of the mmap VM area
|
|
*/
|
|
#define PAGE_OFFSET UL(CONFIG_PAGE_OFFSET)
|
|
#define TASK_SIZE (UL(CONFIG_PAGE_OFFSET) - UL(SZ_16M))
|
|
#define TASK_UNMAPPED_BASE ALIGN(TASK_SIZE / 3, SZ_16M)
|
|
|
|
/*
|
|
* The maximum size of a 26-bit user space task.
|
|
*/
|
|
#define TASK_SIZE_26 (UL(1) << 26)
|
|
|
|
/*
|
|
* The module space lives between the addresses given by TASK_SIZE
|
|
* and PAGE_OFFSET - it must be within 32MB of the kernel text.
|
|
*/
|
|
#ifndef CONFIG_THUMB2_KERNEL
|
|
#define MODULES_VADDR (PAGE_OFFSET - SZ_16M)
|
|
#else
|
|
/* smaller range for Thumb-2 symbols relocation (2^24)*/
|
|
#define MODULES_VADDR (PAGE_OFFSET - SZ_8M)
|
|
#endif
|
|
|
|
#if TASK_SIZE > MODULES_VADDR
|
|
#error Top of user space clashes with start of module space
|
|
#endif
|
|
|
|
/*
|
|
* The highmem pkmap virtual space shares the end of the module area.
|
|
*/
|
|
#ifdef CONFIG_HIGHMEM
|
|
#define MODULES_END (PAGE_OFFSET - PMD_SIZE)
|
|
#else
|
|
#define MODULES_END (PAGE_OFFSET)
|
|
#endif
|
|
|
|
/*
|
|
* The XIP kernel gets mapped at the bottom of the module vm area.
|
|
* Since we use sections to map it, this macro replaces the physical address
|
|
* with its virtual address while keeping offset from the base section.
|
|
*/
|
|
#define XIP_VIRT_ADDR(physaddr) (MODULES_VADDR + ((physaddr) & 0x000fffff))
|
|
|
|
/*
|
|
* Allow 16MB-aligned ioremap pages
|
|
*/
|
|
#define IOREMAP_MAX_ORDER 24
|
|
|
|
#define CONSISTENT_END (0xffe00000UL)
|
|
|
|
#else /* CONFIG_MMU */
|
|
|
|
/*
|
|
* The limitation of user task size can grow up to the end of free ram region.
|
|
* It is difficult to define and perhaps will never meet the original meaning
|
|
* of this define that was meant to.
|
|
* Fortunately, there is no reference for this in noMMU mode, for now.
|
|
*/
|
|
#ifndef TASK_SIZE
|
|
#define TASK_SIZE (CONFIG_DRAM_SIZE)
|
|
#endif
|
|
|
|
#ifndef TASK_UNMAPPED_BASE
|
|
#define TASK_UNMAPPED_BASE UL(0x00000000)
|
|
#endif
|
|
|
|
#ifndef PHYS_OFFSET
|
|
#define PHYS_OFFSET UL(CONFIG_DRAM_BASE)
|
|
#endif
|
|
|
|
#ifndef END_MEM
|
|
#define END_MEM (UL(CONFIG_DRAM_BASE) + CONFIG_DRAM_SIZE)
|
|
#endif
|
|
|
|
#ifndef PAGE_OFFSET
|
|
#define PAGE_OFFSET (PHYS_OFFSET)
|
|
#endif
|
|
|
|
/*
|
|
* The module can be at any place in ram in nommu mode.
|
|
*/
|
|
#define MODULES_END (END_MEM)
|
|
#define MODULES_VADDR (PHYS_OFFSET)
|
|
|
|
#define XIP_VIRT_ADDR(physaddr) (physaddr)
|
|
|
|
#endif /* !CONFIG_MMU */
|
|
|
|
/*
|
|
* We fix the TCM memories max 32 KiB ITCM resp DTCM at these
|
|
* locations
|
|
*/
|
|
#ifdef CONFIG_HAVE_TCM
|
|
#define ITCM_OFFSET UL(0xfffe0000)
|
|
#define DTCM_OFFSET UL(0xfffe8000)
|
|
#endif
|
|
|
|
/*
|
|
* Convert a physical address to a Page Frame Number and back
|
|
*/
|
|
#define __phys_to_pfn(paddr) ((unsigned long)((paddr) >> PAGE_SHIFT))
|
|
#define __pfn_to_phys(pfn) ((phys_addr_t)(pfn) << PAGE_SHIFT)
|
|
|
|
/*
|
|
* Convert a page to/from a physical address
|
|
*/
|
|
#define page_to_phys(page) (__pfn_to_phys(page_to_pfn(page)))
|
|
#define phys_to_page(phys) (pfn_to_page(__phys_to_pfn(phys)))
|
|
|
|
/*
|
|
* Minimum guaranted alignment in pgd_alloc(). The page table pointers passed
|
|
* around in head.S and proc-*.S are shifted by this amount, in order to
|
|
* leave spare high bits for systems with physical address extension. This
|
|
* does not fully accomodate the 40-bit addressing capability of ARM LPAE, but
|
|
* gives us about 38-bits or so.
|
|
*/
|
|
#ifdef CONFIG_ARM_LPAE
|
|
#define ARCH_PGD_SHIFT L1_CACHE_SHIFT
|
|
#else
|
|
#define ARCH_PGD_SHIFT 0
|
|
#endif
|
|
#define ARCH_PGD_MASK ((1 << ARCH_PGD_SHIFT) - 1)
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
/*
|
|
* Physical vs virtual RAM address space conversion. These are
|
|
* private definitions which should NOT be used outside memory.h
|
|
* files. Use virt_to_phys/phys_to_virt/__pa/__va instead.
|
|
*/
|
|
#ifndef __virt_to_phys
|
|
#ifdef CONFIG_ARM_PATCH_PHYS_VIRT
|
|
|
|
/*
|
|
* Constants used to force the right instruction encodings and shifts
|
|
* so that all we need to do is modify the 8-bit constant field.
|
|
*/
|
|
#define __PV_BITS_31_24 0x81000000
|
|
#define __PV_BITS_7_0 0x81
|
|
|
|
extern u64 __pv_phys_offset;
|
|
extern u64 __pv_offset;
|
|
extern void fixup_pv_table(const void *, unsigned long);
|
|
extern const void *__pv_table_begin, *__pv_table_end;
|
|
|
|
#define PHYS_OFFSET __pv_phys_offset
|
|
|
|
#define __pv_stub(from,to,instr,type) \
|
|
__asm__("@ __pv_stub\n" \
|
|
"1: " instr " %0, %1, %2\n" \
|
|
" .pushsection .pv_table,\"a\"\n" \
|
|
" .long 1b\n" \
|
|
" .popsection\n" \
|
|
: "=r" (to) \
|
|
: "r" (from), "I" (type))
|
|
|
|
#define __pv_stub_mov_hi(t) \
|
|
__asm__ volatile("@ __pv_stub_mov\n" \
|
|
"1: mov %R0, %1\n" \
|
|
" .pushsection .pv_table,\"a\"\n" \
|
|
" .long 1b\n" \
|
|
" .popsection\n" \
|
|
: "=r" (t) \
|
|
: "I" (__PV_BITS_7_0))
|
|
|
|
#define __pv_add_carry_stub(x, y) \
|
|
__asm__ volatile("@ __pv_add_carry_stub\n" \
|
|
"1: adds %Q0, %1, %2\n" \
|
|
" adc %R0, %R0, #0\n" \
|
|
" .pushsection .pv_table,\"a\"\n" \
|
|
" .long 1b\n" \
|
|
" .popsection\n" \
|
|
: "+r" (y) \
|
|
: "r" (x), "I" (__PV_BITS_31_24) \
|
|
: "cc")
|
|
|
|
static inline phys_addr_t __virt_to_phys(unsigned long x)
|
|
{
|
|
phys_addr_t t;
|
|
|
|
if (sizeof(phys_addr_t) == 4) {
|
|
__pv_stub(x, t, "add", __PV_BITS_31_24);
|
|
} else {
|
|
__pv_stub_mov_hi(t);
|
|
__pv_add_carry_stub(x, t);
|
|
}
|
|
return t;
|
|
}
|
|
|
|
static inline unsigned long __phys_to_virt(phys_addr_t x)
|
|
{
|
|
unsigned long t;
|
|
|
|
/*
|
|
* 'unsigned long' cast discard upper word when
|
|
* phys_addr_t is 64 bit, and makes sure that inline
|
|
* assembler expression receives 32 bit argument
|
|
* in place where 'r' 32 bit operand is expected.
|
|
*/
|
|
__pv_stub((unsigned long) x, t, "sub", __PV_BITS_31_24);
|
|
return t;
|
|
}
|
|
|
|
#else
|
|
|
|
static inline phys_addr_t __virt_to_phys(unsigned long x)
|
|
{
|
|
return (phys_addr_t)x - PAGE_OFFSET + PHYS_OFFSET;
|
|
}
|
|
|
|
static inline unsigned long __phys_to_virt(phys_addr_t x)
|
|
{
|
|
return x - PHYS_OFFSET + PAGE_OFFSET;
|
|
}
|
|
|
|
#endif
|
|
#endif
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#ifndef PHYS_OFFSET
|
|
#ifdef PLAT_PHYS_OFFSET
|
|
#define PHYS_OFFSET PLAT_PHYS_OFFSET
|
|
#else
|
|
#define PHYS_OFFSET UL(CONFIG_PHYS_OFFSET)
|
|
#endif
|
|
#endif
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
/*
|
|
* PFNs are used to describe any physical page; this means
|
|
* PFN 0 == physical address 0.
|
|
*
|
|
* This is the PFN of the first RAM page in the kernel
|
|
* direct-mapped view. We assume this is the first page
|
|
* of RAM in the mem_map as well.
|
|
*/
|
|
#define PHYS_PFN_OFFSET ((unsigned long)(PHYS_OFFSET >> PAGE_SHIFT))
|
|
|
|
/*
|
|
* These are *only* valid on the kernel direct mapped RAM memory.
|
|
* Note: Drivers should NOT use these. They are the wrong
|
|
* translation for translating DMA addresses. Use the driver
|
|
* DMA support - see dma-mapping.h.
|
|
*/
|
|
static inline phys_addr_t virt_to_phys(const volatile void *x)
|
|
{
|
|
return __virt_to_phys((unsigned long)(x));
|
|
}
|
|
|
|
static inline void *phys_to_virt(phys_addr_t x)
|
|
{
|
|
return (void *)__phys_to_virt(x);
|
|
}
|
|
|
|
/*
|
|
* Drivers should NOT use these either.
|
|
*/
|
|
#define __pa(x) __virt_to_phys((unsigned long)(x))
|
|
#define __va(x) ((void *)__phys_to_virt((phys_addr_t)(x)))
|
|
#define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT)
|
|
|
|
extern phys_addr_t (*arch_virt_to_idmap)(unsigned long x);
|
|
|
|
/*
|
|
* These are for systems that have a hardware interconnect supported alias of
|
|
* physical memory for idmap purposes. Most cases should leave these
|
|
* untouched.
|
|
*/
|
|
static inline phys_addr_t __virt_to_idmap(unsigned long x)
|
|
{
|
|
if (arch_virt_to_idmap)
|
|
return arch_virt_to_idmap(x);
|
|
else
|
|
return __virt_to_phys(x);
|
|
}
|
|
|
|
#define virt_to_idmap(x) __virt_to_idmap((unsigned long)(x))
|
|
|
|
/*
|
|
* Virtual <-> DMA view memory address translations
|
|
* Again, these are *only* valid on the kernel direct mapped RAM
|
|
* memory. Use of these is *deprecated* (and that doesn't mean
|
|
* use the __ prefixed forms instead.) See dma-mapping.h.
|
|
*/
|
|
#ifndef __virt_to_bus
|
|
#define __virt_to_bus __virt_to_phys
|
|
#define __bus_to_virt __phys_to_virt
|
|
#define __pfn_to_bus(x) __pfn_to_phys(x)
|
|
#define __bus_to_pfn(x) __phys_to_pfn(x)
|
|
#endif
|
|
|
|
#ifdef CONFIG_VIRT_TO_BUS
|
|
static inline __deprecated unsigned long virt_to_bus(void *x)
|
|
{
|
|
return __virt_to_bus((unsigned long)x);
|
|
}
|
|
|
|
static inline __deprecated void *bus_to_virt(unsigned long x)
|
|
{
|
|
return (void *)__bus_to_virt(x);
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Conversion between a struct page and a physical address.
|
|
*
|
|
* page_to_pfn(page) convert a struct page * to a PFN number
|
|
* pfn_to_page(pfn) convert a _valid_ PFN number to struct page *
|
|
*
|
|
* virt_to_page(k) convert a _valid_ virtual address to struct page *
|
|
* virt_addr_valid(k) indicates whether a virtual address is valid
|
|
*/
|
|
#define ARCH_PFN_OFFSET PHYS_PFN_OFFSET
|
|
|
|
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
|
|
#define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (unsigned long)high_memory)
|
|
|
|
#endif
|
|
|
|
#include <asm-generic/memory_model.h>
|
|
|
|
#endif
|