2005-04-16 15:20:36 -07:00
/*
* IA - 64 Huge TLB Page Support for Kernel .
*
* Copyright ( C ) 2002 - 2004 Rohit Seth < rohit . seth @ intel . com >
* Copyright ( C ) 2003 - 2004 Ken Chen < kenneth . w . chen @ intel . com >
*
* Sep , 2003 : add numa support
* Feb , 2004 : dynamic hugetlb page size via boot parameter
*/
# include <linux/init.h>
# include <linux/fs.h>
# include <linux/mm.h>
# include <linux/hugetlb.h>
# include <linux/pagemap.h>
# include <linux/slab.h>
# include <linux/sysctl.h>
2007-06-07 15:27:46 +05:30
# include <linux/log2.h>
2005-04-16 15:20:36 -07:00
# include <asm/mman.h>
# include <asm/pgalloc.h>
# include <asm/tlb.h>
# include <asm/tlbflush.h>
unsigned int hpage_shift = HPAGE_SHIFT_DEFAULT ;
2005-06-21 17:14:44 -07:00
pte_t *
2008-07-23 21:27:41 -07:00
huge_pte_alloc ( struct mm_struct * mm , unsigned long addr , unsigned long sz )
2005-04-16 15:20:36 -07:00
{
unsigned long taddr = htlbpage_to_page ( addr ) ;
pgd_t * pgd ;
pud_t * pud ;
pmd_t * pmd ;
pte_t * pte = NULL ;
pgd = pgd_offset ( mm , taddr ) ;
pud = pud_alloc ( mm , pgd , taddr ) ;
if ( pud ) {
pmd = pmd_alloc ( mm , pud , taddr ) ;
if ( pmd )
pte = pte_alloc_map ( mm , pmd , taddr ) ;
}
return pte ;
}
2005-06-21 17:14:44 -07:00
pte_t *
2005-04-16 15:20:36 -07:00
huge_pte_offset ( struct mm_struct * mm , unsigned long addr )
{
unsigned long taddr = htlbpage_to_page ( addr ) ;
pgd_t * pgd ;
pud_t * pud ;
pmd_t * pmd ;
pte_t * pte = NULL ;
pgd = pgd_offset ( mm , taddr ) ;
if ( pgd_present ( * pgd ) ) {
pud = pud_offset ( pgd , taddr ) ;
if ( pud_present ( * pud ) ) {
pmd = pmd_offset ( pud , taddr ) ;
if ( pmd_present ( * pmd ) )
pte = pte_offset_map ( pmd , taddr ) ;
}
}
return pte ;
}
2006-12-06 20:32:03 -08:00
int huge_pmd_unshare ( struct mm_struct * mm , unsigned long * addr , pte_t * ptep )
{
return 0 ;
}
2005-04-16 15:20:36 -07:00
# define mk_pte_huge(entry) { pte_val(entry) |= _PAGE_P; }
/*
2006-03-22 00:09:01 -08:00
* Don ' t actually need to do any preparation , but need to make sure
* the address is in the right region .
2005-04-16 15:20:36 -07:00
*/
2008-07-23 21:27:41 -07:00
int prepare_hugepage_range ( struct file * file ,
unsigned long addr , unsigned long len )
2005-04-16 15:20:36 -07:00
{
if ( len & ~ HPAGE_MASK )
return - EINVAL ;
if ( addr & ~ HPAGE_MASK )
return - EINVAL ;
2005-08-16 19:54:00 -07:00
if ( REGION_NUMBER ( addr ) ! = RGN_HPAGE )
2005-04-16 15:20:36 -07:00
return - EINVAL ;
return 0 ;
}
struct page * follow_huge_addr ( struct mm_struct * mm , unsigned long addr , int write )
{
struct page * page ;
pte_t * ptep ;
2005-08-16 19:54:00 -07:00
if ( REGION_NUMBER ( addr ) ! = RGN_HPAGE )
2005-04-16 15:20:36 -07:00
return ERR_PTR ( - EINVAL ) ;
ptep = huge_pte_offset ( mm , addr ) ;
if ( ! ptep | | pte_none ( * ptep ) )
return NULL ;
page = pte_page ( * ptep ) ;
page + = ( ( addr & ~ HPAGE_MASK ) > > PAGE_SHIFT ) ;
return page ;
}
int pmd_huge ( pmd_t pmd )
{
return 0 ;
}
2008-07-23 21:27:50 -07:00
int pud_huge ( pud_t pud )
{
return 0 ;
}
2005-04-16 15:20:36 -07:00
struct page *
follow_huge_pmd ( struct mm_struct * mm , unsigned long address , pmd_t * pmd , int write )
{
return NULL ;
}
2008-07-23 21:27:10 -07:00
void hugetlb_free_pgd_range ( struct mmu_gather * tlb ,
2005-04-19 13:29:16 -07:00
unsigned long addr , unsigned long end ,
unsigned long floor , unsigned long ceiling )
2005-04-16 15:20:36 -07:00
{
2005-04-19 13:29:16 -07:00
/*
2006-03-22 10:49:00 -08:00
* This is called to free hugetlb page tables .
2005-04-19 13:29:16 -07:00
*
* The offset of these addresses from the base of the hugetlb
* region must be scaled down by HPAGE_SIZE / PAGE_SIZE so that
* the standard free_pgd_range will free the right page tables .
*
* If floor and ceiling are also in the hugetlb region , they
* must likewise be scaled down ; but if outside , left unchanged .
*/
addr = htlbpage_to_page ( addr ) ;
end = htlbpage_to_page ( end ) ;
2006-03-22 10:49:00 -08:00
if ( REGION_NUMBER ( floor ) = = RGN_HPAGE )
2005-04-19 13:29:16 -07:00
floor = htlbpage_to_page ( floor ) ;
2006-03-22 10:49:00 -08:00
if ( REGION_NUMBER ( ceiling ) = = RGN_HPAGE )
2005-04-19 13:29:16 -07:00
ceiling = htlbpage_to_page ( ceiling ) ;
free_pgd_range ( tlb , addr , end , floor , ceiling ) ;
2005-04-16 15:20:36 -07:00
}
unsigned long hugetlb_get_unmapped_area ( struct file * file , unsigned long addr , unsigned long len ,
unsigned long pgoff , unsigned long flags )
{
struct vm_area_struct * vmm ;
if ( len > RGN_MAP_LIMIT )
return - ENOMEM ;
if ( len & ~ HPAGE_MASK )
return - EINVAL ;
2007-05-06 14:50:09 -07:00
/* Handle MAP_FIXED */
if ( flags & MAP_FIXED ) {
2008-07-23 21:27:41 -07:00
if ( prepare_hugepage_range ( file , addr , len ) )
2007-05-06 14:50:09 -07:00
return - EINVAL ;
return addr ;
}
2005-08-16 19:54:00 -07:00
/* This code assumes that RGN_HPAGE != 0. */
if ( ( REGION_NUMBER ( addr ) ! = RGN_HPAGE ) | | ( addr & ( HPAGE_SIZE - 1 ) ) )
2005-04-16 15:20:36 -07:00
addr = HPAGE_REGION_BASE ;
else
addr = ALIGN ( addr , HPAGE_SIZE ) ;
for ( vmm = find_vma ( current - > mm , addr ) ; ; vmm = vmm - > vm_next ) {
/* At this point: (!vmm || addr < vmm->vm_end). */
if ( REGION_OFFSET ( addr ) + len > RGN_MAP_LIMIT )
return - ENOMEM ;
if ( ! vmm | | ( addr + len ) < = vmm - > vm_start )
return addr ;
addr = ALIGN ( vmm - > vm_end , HPAGE_SIZE ) ;
}
}
static int __init hugetlb_setup_sz ( char * str )
{
u64 tr_pages ;
unsigned long long size ;
if ( ia64_pal_vm_page_size ( & tr_pages , NULL ) ! = 0 )
/*
* shouldn ' t happen , but just in case .
*/
tr_pages = 0x15557000UL ;
size = memparse ( str , & str ) ;
2007-06-07 15:27:46 +05:30
if ( * str | | ! is_power_of_2 ( size ) | | ! ( tr_pages & size ) | |
2005-04-16 15:20:36 -07:00
size < = PAGE_SIZE | |
size > = ( 1UL < < PAGE_SHIFT < < MAX_ORDER ) ) {
printk ( KERN_WARNING " Invalid huge page size specified \n " ) ;
return 1 ;
}
hpage_shift = __ffs ( size ) ;
/*
* boot cpu already executed ia64_mmu_init , and has HPAGE_SHIFT_DEFAULT
* override here with new page shift .
*/
ia64_set_rr ( HPAGE_REGION_BASE , hpage_shift < < 2 ) ;
2007-10-16 01:26:01 -07:00
return 0 ;
2005-04-16 15:20:36 -07:00
}
2007-10-16 01:26:01 -07:00
early_param ( " hugepagesz " , hugetlb_setup_sz ) ;