2008-05-03 20:29:28 +04:00
/*
* sched_clock for unstable cpu clocks
*
* Copyright ( C ) 2008 Red Hat , Inc . , Peter Zijlstra < pzijlstr @ redhat . com >
*
2008-07-09 08:15:33 +04:00
* Updates and enhancements :
* Copyright ( C ) 2008 Red Hat , Inc . Steven Rostedt < srostedt @ redhat . com >
*
2008-05-03 20:29:28 +04:00
* Based on code by :
* Ingo Molnar < mingo @ redhat . com >
* Guillaume Chazarain < guichaz @ gmail . com >
*
* Create a semi stable clock from a mixture of other events , including :
* - gtod
* - sched_clock ( )
* - explicit idle events
*
* We use gtod as base and the unstable clock deltas . The deltas are filtered ,
2008-08-25 19:15:34 +04:00
* making it monotonic and keeping it within an expected window .
2008-05-03 20:29:28 +04:00
*
* Furthermore , explicit sleep and wakeup hooks allow us to account for time
* that is otherwise invisible ( TSC gets stopped ) .
*
* The clock : sched_clock_cpu ( ) is monotonic per cpu , and should be somewhat
2008-08-25 19:15:34 +04:00
* consistent between cpus ( never more than 2 jiffies difference ) .
2008-05-03 20:29:28 +04:00
*/
# include <linux/sched.h>
# include <linux/percpu.h>
# include <linux/spinlock.h>
# include <linux/ktime.h>
# include <linux/module.h>
2008-07-25 22:45:00 +04:00
/*
* Scheduler clock - returns current time in nanosec units .
* This is default implementation .
* Architectures and sub - architectures can override this .
*/
unsigned long long __attribute__ ( ( weak ) ) sched_clock ( void )
{
return ( unsigned long long ) jiffies * ( NSEC_PER_SEC / HZ ) ;
}
2008-05-03 20:29:28 +04:00
2008-08-11 10:59:03 +04:00
static __read_mostly int sched_clock_running ;
2008-05-03 20:29:28 +04:00
# ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
struct sched_clock_data {
/*
* Raw spinlock - this is a special case : this might be called
* from within instrumentation code so we dont want to do any
* instrumentation ourselves .
*/
raw_spinlock_t lock ;
u64 tick_raw ;
u64 tick_gtod ;
u64 clock ;
} ;
static DEFINE_PER_CPU_SHARED_ALIGNED ( struct sched_clock_data , sched_clock_data ) ;
static inline struct sched_clock_data * this_scd ( void )
{
return & __get_cpu_var ( sched_clock_data ) ;
}
static inline struct sched_clock_data * cpu_sdc ( int cpu )
{
return & per_cpu ( sched_clock_data , cpu ) ;
}
void sched_clock_init ( void )
{
u64 ktime_now = ktime_to_ns ( ktime_get ( ) ) ;
int cpu ;
for_each_possible_cpu ( cpu ) {
struct sched_clock_data * scd = cpu_sdc ( cpu ) ;
scd - > lock = ( raw_spinlock_t ) __RAW_SPIN_LOCK_UNLOCKED ;
2008-05-29 12:07:15 +04:00
scd - > tick_raw = 0 ;
2008-05-03 20:29:28 +04:00
scd - > tick_gtod = ktime_now ;
scd - > clock = ktime_now ;
}
2008-05-29 12:07:15 +04:00
sched_clock_running = 1 ;
2008-05-03 20:29:28 +04:00
}
2008-08-25 19:15:34 +04:00
/*
* min , max except they take wrapping into account
*/
static inline u64 wrap_min ( u64 x , u64 y )
{
return ( s64 ) ( x - y ) < 0 ? x : y ;
}
static inline u64 wrap_max ( u64 x , u64 y )
{
return ( s64 ) ( x - y ) > 0 ? x : y ;
}
2008-05-03 20:29:28 +04:00
/*
* update the percpu scd from the raw @ now value
*
* - filter out backward motion
2008-08-25 19:15:34 +04:00
* - use the GTOD tick value to create a window to filter crazy TSC values
2008-05-03 20:29:28 +04:00
*/
2008-07-30 12:15:55 +04:00
static u64 __update_sched_clock ( struct sched_clock_data * scd , u64 now )
2008-05-03 20:29:28 +04:00
{
2008-07-30 12:13:35 +04:00
s64 delta = now - scd - > tick_raw ;
2008-08-25 19:15:34 +04:00
u64 clock , min_clock , max_clock ;
2008-05-03 20:29:28 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
2008-08-25 19:15:34 +04:00
if ( unlikely ( delta < 0 ) )
delta = 0 ;
2008-05-03 20:29:28 +04:00
2008-08-25 19:15:34 +04:00
/*
* scd - > clock = clamp ( scd - > tick_gtod + delta ,
2008-12-15 02:46:01 +03:00
* max ( scd - > tick_gtod , scd - > clock ) ,
* scd - > tick_gtod + TICK_NSEC ) ;
2008-08-25 19:15:34 +04:00
*/
2008-05-03 20:29:28 +04:00
2008-08-25 19:15:34 +04:00
clock = scd - > tick_gtod + delta ;
min_clock = wrap_max ( scd - > tick_gtod , scd - > clock ) ;
2008-12-23 01:05:28 +03:00
max_clock = wrap_max ( scd - > clock , scd - > tick_gtod + TICK_NSEC ) ;
2008-05-03 20:29:28 +04:00
2008-08-25 19:15:34 +04:00
clock = wrap_max ( clock , min_clock ) ;
clock = wrap_min ( clock , max_clock ) ;
2008-05-03 20:29:28 +04:00
2008-04-14 10:50:02 +04:00
scd - > clock = clock ;
2008-07-30 12:15:55 +04:00
2008-08-25 19:15:34 +04:00
return scd - > clock ;
2008-05-03 20:29:28 +04:00
}
static void lock_double_clock ( struct sched_clock_data * data1 ,
struct sched_clock_data * data2 )
{
if ( data1 < data2 ) {
__raw_spin_lock ( & data1 - > lock ) ;
__raw_spin_lock ( & data2 - > lock ) ;
} else {
__raw_spin_lock ( & data2 - > lock ) ;
__raw_spin_lock ( & data1 - > lock ) ;
}
}
u64 sched_clock_cpu ( int cpu )
{
struct sched_clock_data * scd = cpu_sdc ( cpu ) ;
2008-07-30 12:22:07 +04:00
u64 now , clock , this_clock , remote_clock ;
2008-05-03 20:29:28 +04:00
2008-05-29 12:07:15 +04:00
if ( unlikely ( ! sched_clock_running ) )
return 0ull ;
2008-05-03 20:29:28 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
now = sched_clock ( ) ;
if ( cpu ! = raw_smp_processor_id ( ) ) {
struct sched_clock_data * my_scd = this_scd ( ) ;
lock_double_clock ( scd , my_scd ) ;
2008-07-30 12:22:07 +04:00
this_clock = __update_sched_clock ( my_scd , now ) ;
remote_clock = scd - > clock ;
/*
* Use the opportunity that we have both locks
* taken to couple the two clocks : we take the
* larger time as the latest time for both
* runqueues . ( this creates monotonic movement )
*/
2008-08-25 19:15:34 +04:00
if ( likely ( ( s64 ) ( remote_clock - this_clock ) < 0 ) ) {
2008-07-30 12:22:07 +04:00
clock = this_clock ;
scd - > clock = clock ;
} else {
/*
* Should be rare , but possible :
*/
clock = remote_clock ;
my_scd - > clock = remote_clock ;
}
2008-05-03 20:29:28 +04:00
__raw_spin_unlock ( & my_scd - > lock ) ;
} else {
__raw_spin_lock ( & scd - > lock ) ;
2008-07-30 12:22:07 +04:00
clock = __update_sched_clock ( scd , now ) ;
2008-05-03 20:29:28 +04:00
}
2008-04-14 10:50:02 +04:00
__raw_spin_unlock ( & scd - > lock ) ;
2008-05-03 20:29:28 +04:00
return clock ;
}
void sched_clock_tick ( void )
{
struct sched_clock_data * scd = this_scd ( ) ;
u64 now , now_gtod ;
2008-05-29 12:07:15 +04:00
if ( unlikely ( ! sched_clock_running ) )
return ;
2008-05-03 20:29:28 +04:00
WARN_ON_ONCE ( ! irqs_disabled ( ) ) ;
now_gtod = ktime_to_ns ( ktime_get ( ) ) ;
2008-07-09 08:15:32 +04:00
now = sched_clock ( ) ;
2008-05-03 20:29:28 +04:00
__raw_spin_lock ( & scd - > lock ) ;
scd - > tick_raw = now ;
scd - > tick_gtod = now_gtod ;
2008-08-25 19:15:34 +04:00
__update_sched_clock ( scd , now ) ;
2008-05-03 20:29:28 +04:00
__raw_spin_unlock ( & scd - > lock ) ;
}
/*
* We are going deep - idle ( irqs are disabled ) :
*/
void sched_clock_idle_sleep_event ( void )
{
sched_clock_cpu ( smp_processor_id ( ) ) ;
}
EXPORT_SYMBOL_GPL ( sched_clock_idle_sleep_event ) ;
/*
* We just idled delta nanoseconds ( called with irqs disabled ) :
*/
void sched_clock_idle_wakeup_event ( u64 delta_ns )
{
2008-12-23 01:05:28 +03:00
if ( timekeeping_suspended )
return ;
2008-08-25 19:15:34 +04:00
sched_clock_tick ( ) ;
2008-05-03 20:29:28 +04:00
touch_softlockup_watchdog ( ) ;
}
EXPORT_SYMBOL_GPL ( sched_clock_idle_wakeup_event ) ;
2008-08-11 10:59:03 +04:00
# else /* CONFIG_HAVE_UNSTABLE_SCHED_CLOCK */
void sched_clock_init ( void )
{
sched_clock_running = 1 ;
}
u64 sched_clock_cpu ( int cpu )
{
if ( unlikely ( ! sched_clock_running ) )
return 0 ;
return sched_clock ( ) ;
}
2008-05-03 20:29:28 +04:00
# endif
2008-06-27 15:41:15 +04:00
unsigned long long cpu_clock ( int cpu )
{
unsigned long long clock ;
unsigned long flags ;
2008-06-29 17:01:59 +04:00
local_irq_save ( flags ) ;
2008-06-27 15:41:15 +04:00
clock = sched_clock_cpu ( cpu ) ;
2008-06-29 17:01:59 +04:00
local_irq_restore ( flags ) ;
2008-06-27 15:41:15 +04:00
return clock ;
}
2008-06-27 16:49:35 +04:00
EXPORT_SYMBOL_GPL ( cpu_clock ) ;