2019-06-04 11:11:33 +03:00
// SPDX-License-Identifier: GPL-2.0-only
2008-10-31 16:08:02 +03:00
/*
* linux / arch / arm / mm / copypage - v4wb . c
*
* Copyright ( C ) 1995 - 1999 Russell King
*/
# include <linux/init.h>
2008-10-31 18:08:35 +03:00
# include <linux/highmem.h>
2008-10-31 16:08:02 +03:00
/*
2008-10-31 18:08:35 +03:00
* ARMv4 optimised copy_user_highpage
2008-10-31 16:08:02 +03:00
*
* We flush the destination cache lines just before we write the data into the
* corresponding address . Since the Dcache is read - allocate , this removes the
* Dcache aliasing issue . The writes will be forwarded to the write buffer ,
* and merged as appropriate .
*
* Note : We rely on all ARMv4 processors implementing the " invalidate D line "
* instruction . If your processor does not supply this , you have to write your
2008-10-31 18:08:35 +03:00
* own copy_user_highpage that does the right thing .
2008-10-31 16:08:02 +03:00
*/
2018-11-07 19:49:00 +03:00
static void v4wb_copy_user_page ( void * kto , const void * kfrom )
2008-10-31 16:08:02 +03:00
{
2018-11-07 19:49:00 +03:00
int tmp ;
asm volatile ( " \
2019-02-18 02:58:29 +03:00
. syntax unified \ n \
2018-11-07 19:49:00 +03:00
ldmia % 1 ! , { r3 , r4 , ip , lr } @ 4 \ n \
1 : mcr p15 , 0 , % 0 , c7 , c6 , 1 @ 1 invalidate D line \ n \
stmia % 0 ! , { r3 , r4 , ip , lr } @ 4 \ n \
ldmia % 1 ! , { r3 , r4 , ip , lr } @ 4 + 1 \ n \
stmia % 0 ! , { r3 , r4 , ip , lr } @ 4 \ n \
ldmia % 1 ! , { r3 , r4 , ip , lr } @ 4 \ n \
mcr p15 , 0 , % 0 , c7 , c6 , 1 @ 1 invalidate D line \ n \
stmia % 0 ! , { r3 , r4 , ip , lr } @ 4 \ n \
ldmia % 1 ! , { r3 , r4 , ip , lr } @ 4 \ n \
subs % 2 , % 2 , # 1 @ 1 \ n \
stmia % 0 ! , { r3 , r4 , ip , lr } @ 4 \ n \
2019-02-18 02:58:29 +03:00
ldmiane % 1 ! , { r3 , r4 , ip , lr } @ 4 \ n \
2008-10-31 16:08:02 +03:00
bne 1 b @ 1 \ n \
2018-11-07 19:49:00 +03:00
mcr p15 , 0 , % 1 , c7 , c10 , 4 @ 1 drain WB "
: " +&r " ( kto ) , " +&r " ( kfrom ) , " =&r " ( tmp )
: " 2 " ( PAGE_SIZE / 64 )
: " r3 " , " r4 " , " ip " , " lr " ) ;
2008-10-31 16:08:02 +03:00
}
2008-10-31 18:08:35 +03:00
void v4wb_copy_user_highpage ( struct page * to , struct page * from ,
2009-10-05 18:17:45 +04:00
unsigned long vaddr , struct vm_area_struct * vma )
2008-10-31 18:08:35 +03:00
{
void * kto , * kfrom ;
2011-11-25 19:14:15 +04:00
kto = kmap_atomic ( to ) ;
kfrom = kmap_atomic ( from ) ;
2009-10-05 18:34:22 +04:00
flush_cache_page ( vma , vaddr , page_to_pfn ( from ) ) ;
2008-10-31 18:08:35 +03:00
v4wb_copy_user_page ( kto , kfrom ) ;
2011-11-25 19:14:15 +04:00
kunmap_atomic ( kfrom ) ;
kunmap_atomic ( kto ) ;
2008-10-31 18:08:35 +03:00
}
2008-10-31 16:08:02 +03:00
/*
* ARMv4 optimised clear_user_page
*
* Same story as above .
*/
2008-10-31 19:32:19 +03:00
void v4wb_clear_user_highpage ( struct page * page , unsigned long vaddr )
2008-10-31 16:08:02 +03:00
{
2011-11-25 19:14:15 +04:00
void * ptr , * kaddr = kmap_atomic ( page ) ;
2008-11-04 10:42:27 +03:00
asm volatile ( " \
mov r1 , % 2 @ 1 \ n \
2008-10-31 16:08:02 +03:00
mov r2 , # 0 @ 1 \ n \
mov r3 , # 0 @ 1 \ n \
mov ip , # 0 @ 1 \ n \
mov lr , # 0 @ 1 \ n \
2008-10-31 19:32:19 +03:00
1 : mcr p15 , 0 , % 0 , c7 , c6 , 1 @ 1 invalidate D line \ n \
stmia % 0 ! , { r2 , r3 , ip , lr } @ 4 \ n \
stmia % 0 ! , { r2 , r3 , ip , lr } @ 4 \ n \
mcr p15 , 0 , % 0 , c7 , c6 , 1 @ 1 invalidate D line \ n \
stmia % 0 ! , { r2 , r3 , ip , lr } @ 4 \ n \
stmia % 0 ! , { r2 , r3 , ip , lr } @ 4 \ n \
2008-10-31 16:08:02 +03:00
subs r1 , r1 , # 1 @ 1 \ n \
bne 1 b @ 1 \ n \
2008-10-31 19:32:19 +03:00
mcr p15 , 0 , r1 , c7 , c10 , 4 @ 1 drain WB "
2008-11-04 10:42:27 +03:00
: " =r " ( ptr )
: " 0 " ( kaddr ) , " I " ( PAGE_SIZE / 64 )
2008-10-31 19:32:19 +03:00
: " r1 " , " r2 " , " r3 " , " ip " , " lr " ) ;
2011-11-25 19:14:15 +04:00
kunmap_atomic ( kaddr ) ;
2008-10-31 16:08:02 +03:00
}
struct cpu_user_fns v4wb_user_fns __initdata = {
2008-10-31 19:32:19 +03:00
. cpu_clear_user_highpage = v4wb_clear_user_highpage ,
2008-10-31 18:08:35 +03:00
. cpu_copy_user_highpage = v4wb_copy_user_highpage ,
2008-10-31 16:08:02 +03:00
} ;