2012-01-11 17:25:17 +00:00
# ifndef __ASMARM_ARCH_TIMER_H
# define __ASMARM_ARCH_TIMER_H
2012-11-12 16:18:00 +00:00
# include <asm/barrier.h>
2012-07-06 15:46:45 +01:00
# include <asm/errno.h>
2012-09-07 18:09:58 +01:00
# include <linux/clocksource.h>
2012-11-12 14:33:44 +00:00
# include <linux/init.h>
2012-11-12 16:18:00 +00:00
# include <linux/types.h>
2012-07-06 15:46:45 +01:00
2012-11-12 14:33:44 +00:00
# include <clocksource/arm_arch_timer.h>
2012-01-11 17:25:17 +00:00
# ifdef CONFIG_ARM_ARCH_TIMER
2013-04-10 18:27:51 -05:00
int arch_timer_arch_init ( void ) ;
2012-11-12 16:18:00 +00:00
/*
* These register accessors are marked inline so the compiler can
* nicely work out which register we want , and chuck away the rest of
* the code . At least it does so with a recent GCC ( 4.6 .3 ) .
*/
2013-07-18 16:59:28 -07:00
static __always_inline
void arch_timer_reg_write ( int access , enum arch_timer_reg reg , u32 val )
2012-11-12 16:18:00 +00:00
{
if ( access = = ARCH_TIMER_PHYS_ACCESS ) {
switch ( reg ) {
case ARCH_TIMER_REG_CTRL :
asm volatile ( " mcr p15, 0, %0, c14, c2, 1 " : : " r " ( val ) ) ;
break ;
case ARCH_TIMER_REG_TVAL :
asm volatile ( " mcr p15, 0, %0, c14, c2, 0 " : : " r " ( val ) ) ;
break ;
}
2013-07-18 16:59:28 -07:00
} else if ( access = = ARCH_TIMER_VIRT_ACCESS ) {
2012-11-12 16:18:00 +00:00
switch ( reg ) {
case ARCH_TIMER_REG_CTRL :
asm volatile ( " mcr p15, 0, %0, c14, c3, 1 " : : " r " ( val ) ) ;
break ;
case ARCH_TIMER_REG_TVAL :
asm volatile ( " mcr p15, 0, %0, c14, c3, 0 " : : " r " ( val ) ) ;
break ;
}
}
2013-01-11 14:32:33 +00:00
isb ( ) ;
2012-11-12 16:18:00 +00:00
}
2013-07-18 16:59:28 -07:00
static __always_inline
u32 arch_timer_reg_read ( int access , enum arch_timer_reg reg )
2012-11-12 16:18:00 +00:00
{
u32 val = 0 ;
if ( access = = ARCH_TIMER_PHYS_ACCESS ) {
switch ( reg ) {
case ARCH_TIMER_REG_CTRL :
asm volatile ( " mrc p15, 0, %0, c14, c2, 1 " : " =r " ( val ) ) ;
break ;
case ARCH_TIMER_REG_TVAL :
asm volatile ( " mrc p15, 0, %0, c14, c2, 0 " : " =r " ( val ) ) ;
break ;
}
2013-07-18 16:59:28 -07:00
} else if ( access = = ARCH_TIMER_VIRT_ACCESS ) {
2012-11-12 16:18:00 +00:00
switch ( reg ) {
case ARCH_TIMER_REG_CTRL :
asm volatile ( " mrc p15, 0, %0, c14, c3, 1 " : " =r " ( val ) ) ;
break ;
case ARCH_TIMER_REG_TVAL :
asm volatile ( " mrc p15, 0, %0, c14, c3, 0 " : " =r " ( val ) ) ;
break ;
}
}
return val ;
}
static inline u32 arch_timer_get_cntfrq ( void )
{
u32 val ;
asm volatile ( " mrc p15, 0, %0, c14, c0, 0 " : " =r " ( val ) ) ;
return val ;
}
static inline u64 arch_counter_get_cntvct ( void )
{
u64 cval ;
2013-01-11 14:32:33 +00:00
isb ( ) ;
2012-11-12 16:18:00 +00:00
asm volatile ( " mrrc p15, 1, %Q0, %R0, c14 " : " =r " ( cval ) ) ;
return cval ;
}
2012-11-14 10:32:24 +00:00
static inline void __cpuinit arch_counter_set_user_access ( void )
{
u32 cntkctl ;
asm volatile ( " mrc p15, 0, %0, c14, c1, 0 " : " =r " ( cntkctl ) ) ;
/* disable user access to everything */
cntkctl & = ~ ( ( 3 < < 8 ) | ( 7 < < 0 ) ) ;
asm volatile ( " mcr p15, 0, %0, c14, c1, 0 " : : " r " ( cntkctl ) ) ;
}
2012-01-11 17:25:17 +00:00
# endif
# endif