2019-05-27 09:55:01 +03:00
// SPDX-License-Identifier: GPL-2.0-or-later
2011-01-24 15:19:12 +03:00
/*
* Copyright 2010 - 2011 Picochip Ltd . , Jamie Iles
2020-07-13 23:58:21 +03:00
* https : //www.picochip.com
2011-01-24 15:19:12 +03:00
*
* This file implements a driver for the Synopsys DesignWare watchdog device
2013-12-30 16:25:54 +04:00
* in the many subsystems . The watchdog has 16 different timeout periods
2011-01-24 15:19:12 +03:00
* and these are a function of the input clock frequency .
*
* The DesignWare watchdog cannot be stopped once it has been started so we
2016-02-29 00:12:19 +03:00
* do not implement a stop function . The watchdog core will continue to send
* heartbeat requests after the watchdog device has been closed .
2011-01-24 15:19:12 +03:00
*/
2012-02-16 03:06:19 +04:00
2011-01-24 15:19:12 +03:00
# include <linux/bitops.h>
2020-05-30 10:35:54 +03:00
# include <linux/limits.h>
# include <linux/kernel.h>
2011-01-24 15:19:12 +03:00
# include <linux/clk.h>
2014-09-23 11:42:12 +04:00
# include <linux/delay.h>
2011-01-24 15:19:12 +03:00
# include <linux/err.h>
# include <linux/io.h>
# include <linux/kernel.h>
# include <linux/module.h>
# include <linux/moduleparam.h>
2020-05-30 10:35:56 +03:00
# include <linux/interrupt.h>
2013-10-22 20:59:12 +04:00
# include <linux/of.h>
2011-01-24 15:19:12 +03:00
# include <linux/pm.h>
# include <linux/platform_device.h>
2017-05-22 11:51:39 +03:00
# include <linux/reset.h>
2011-01-24 15:19:12 +03:00
# include <linux/watchdog.h>
2020-05-30 10:35:57 +03:00
# include <linux/debugfs.h>
2011-01-24 15:19:12 +03:00
# define WDOG_CONTROL_REG_OFFSET 0x00
# define WDOG_CONTROL_REG_WDT_EN_MASK 0x01
2018-03-10 06:46:06 +03:00
# define WDOG_CONTROL_REG_RESP_MODE_MASK 0x02
2011-01-24 15:19:12 +03:00
# define WDOG_TIMEOUT_RANGE_REG_OFFSET 0x04
2014-09-23 11:42:11 +04:00
# define WDOG_TIMEOUT_RANGE_TOPINIT_SHIFT 4
2011-01-24 15:19:12 +03:00
# define WDOG_CURRENT_COUNT_REG_OFFSET 0x08
# define WDOG_COUNTER_RESTART_REG_OFFSET 0x0c
# define WDOG_COUNTER_RESTART_KICK_VALUE 0x76
2020-05-30 10:35:56 +03:00
# define WDOG_INTERRUPT_STATUS_REG_OFFSET 0x10
# define WDOG_INTERRUPT_CLEAR_REG_OFFSET 0x14
2020-05-30 10:35:57 +03:00
# define WDOG_COMP_PARAMS_5_REG_OFFSET 0xe4
# define WDOG_COMP_PARAMS_4_REG_OFFSET 0xe8
# define WDOG_COMP_PARAMS_3_REG_OFFSET 0xec
# define WDOG_COMP_PARAMS_2_REG_OFFSET 0xf0
2020-05-30 10:35:54 +03:00
# define WDOG_COMP_PARAMS_1_REG_OFFSET 0xf4
# define WDOG_COMP_PARAMS_1_USE_FIX_TOP BIT(6)
2020-05-30 10:35:57 +03:00
# define WDOG_COMP_VERSION_REG_OFFSET 0xf8
# define WDOG_COMP_TYPE_REG_OFFSET 0xfc
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:54 +03:00
/* There are sixteen TOPs (timeout periods) that can be set in the watchdog. */
# define DW_WDT_NUM_TOPS 16
# define DW_WDT_FIX_TOP(_idx) (1U << (16 + _idx))
2011-01-24 15:19:12 +03:00
2015-01-28 01:25:17 +03:00
# define DW_WDT_DEFAULT_SECONDS 30
2020-05-30 10:35:54 +03:00
static const u32 dw_wdt_fix_tops [ DW_WDT_NUM_TOPS ] = {
DW_WDT_FIX_TOP ( 0 ) , DW_WDT_FIX_TOP ( 1 ) , DW_WDT_FIX_TOP ( 2 ) ,
DW_WDT_FIX_TOP ( 3 ) , DW_WDT_FIX_TOP ( 4 ) , DW_WDT_FIX_TOP ( 5 ) ,
DW_WDT_FIX_TOP ( 6 ) , DW_WDT_FIX_TOP ( 7 ) , DW_WDT_FIX_TOP ( 8 ) ,
DW_WDT_FIX_TOP ( 9 ) , DW_WDT_FIX_TOP ( 10 ) , DW_WDT_FIX_TOP ( 11 ) ,
DW_WDT_FIX_TOP ( 12 ) , DW_WDT_FIX_TOP ( 13 ) , DW_WDT_FIX_TOP ( 14 ) ,
DW_WDT_FIX_TOP ( 15 )
} ;
2012-03-05 19:51:11 +04:00
static bool nowayout = WATCHDOG_NOWAYOUT ;
module_param ( nowayout , bool , 0 ) ;
2011-01-24 15:19:12 +03:00
MODULE_PARM_DESC ( nowayout , " Watchdog cannot be stopped once started "
" (default= " __MODULE_STRING ( WATCHDOG_NOWAYOUT ) " ) " ) ;
2020-05-30 10:35:56 +03:00
enum dw_wdt_rmod {
DW_WDT_RMOD_RESET = 1 ,
DW_WDT_RMOD_IRQ = 2
} ;
2020-05-30 10:35:54 +03:00
struct dw_wdt_timeout {
u32 top_val ;
unsigned int sec ;
unsigned int msec ;
} ;
2016-02-29 00:12:19 +03:00
struct dw_wdt {
2011-01-24 15:19:12 +03:00
void __iomem * regs ;
struct clk * clk ;
2020-05-30 10:35:55 +03:00
struct clk * pclk ;
2016-08-10 08:35:58 +03:00
unsigned long rate ;
2020-05-30 10:35:56 +03:00
enum dw_wdt_rmod rmod ;
2020-05-30 10:35:54 +03:00
struct dw_wdt_timeout timeouts [ DW_WDT_NUM_TOPS ] ;
2016-02-29 00:12:19 +03:00
struct watchdog_device wdd ;
2017-05-22 11:51:39 +03:00
struct reset_control * rst ;
2018-03-10 06:46:07 +03:00
/* Save/restore */
u32 control ;
u32 timeout ;
2020-05-30 10:35:57 +03:00
# ifdef CONFIG_DEBUG_FS
struct dentry * dbgfs_dir ;
# endif
2016-02-29 00:12:19 +03:00
} ;
# define to_dw_wdt(wdd) container_of(wdd, struct dw_wdt, wdd)
2011-01-24 15:19:12 +03:00
2016-02-29 00:12:19 +03:00
static inline int dw_wdt_is_enabled ( struct dw_wdt * dw_wdt )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
return readl ( dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) &
2011-01-24 15:19:12 +03:00
WDOG_CONTROL_REG_WDT_EN_MASK ;
}
2020-05-30 10:35:56 +03:00
static void dw_wdt_update_mode ( struct dw_wdt * dw_wdt , enum dw_wdt_rmod rmod )
{
u32 val ;
val = readl ( dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
if ( rmod = = DW_WDT_RMOD_IRQ )
val | = WDOG_CONTROL_REG_RESP_MODE_MASK ;
else
val & = ~ WDOG_CONTROL_REG_RESP_MODE_MASK ;
writel ( val , dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
dw_wdt - > rmod = rmod ;
}
2020-05-30 10:35:54 +03:00
static unsigned int dw_wdt_find_best_top ( struct dw_wdt * dw_wdt ,
unsigned int timeout , u32 * top_val )
2011-01-24 15:19:12 +03:00
{
2020-05-30 10:35:54 +03:00
int idx ;
2011-01-24 15:19:12 +03:00
/*
2020-05-30 10:35:54 +03:00
* Find a TOP with timeout greater or equal to the requested number .
* Note we ' ll select a TOP with maximum timeout if the requested
* timeout couldn ' t be reached .
2011-01-24 15:19:12 +03:00
*/
2020-05-30 10:35:54 +03:00
for ( idx = 0 ; idx < DW_WDT_NUM_TOPS ; + + idx ) {
if ( dw_wdt - > timeouts [ idx ] . sec > = timeout )
break ;
}
if ( idx = = DW_WDT_NUM_TOPS )
- - idx ;
* top_val = dw_wdt - > timeouts [ idx ] . top_val ;
return dw_wdt - > timeouts [ idx ] . sec ;
2011-01-24 15:19:12 +03:00
}
2020-05-30 10:35:54 +03:00
static unsigned int dw_wdt_get_min_timeout ( struct dw_wdt * dw_wdt )
2011-01-24 15:19:12 +03:00
{
2020-05-30 10:35:54 +03:00
int idx ;
/*
* We ' ll find a timeout greater or equal to one second anyway because
* the driver probe would have failed if there was none .
*/
for ( idx = 0 ; idx < DW_WDT_NUM_TOPS ; + + idx ) {
if ( dw_wdt - > timeouts [ idx ] . sec )
break ;
}
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:54 +03:00
return dw_wdt - > timeouts [ idx ] . sec ;
}
static unsigned int dw_wdt_get_max_timeout_ms ( struct dw_wdt * dw_wdt )
{
struct dw_wdt_timeout * timeout = & dw_wdt - > timeouts [ DW_WDT_NUM_TOPS - 1 ] ;
u64 msec ;
msec = ( u64 ) timeout - > sec * MSEC_PER_SEC + timeout - > msec ;
return msec < UINT_MAX ? msec : UINT_MAX ;
}
static unsigned int dw_wdt_get_timeout ( struct dw_wdt * dw_wdt )
{
int top_val = readl ( dw_wdt - > regs + WDOG_TIMEOUT_RANGE_REG_OFFSET ) & 0xF ;
int idx ;
for ( idx = 0 ; idx < DW_WDT_NUM_TOPS ; + + idx ) {
if ( dw_wdt - > timeouts [ idx ] . top_val = = top_val )
break ;
}
2020-05-30 10:35:56 +03:00
/*
* In IRQ mode due to the two stages counter , the actual timeout is
* twice greater than the TOP setting .
*/
return dw_wdt - > timeouts [ idx ] . sec * dw_wdt - > rmod ;
2011-01-24 15:19:12 +03:00
}
2016-02-29 00:12:19 +03:00
static int dw_wdt_ping ( struct watchdog_device * wdd )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
2011-01-24 15:19:12 +03:00
2016-02-29 00:12:19 +03:00
writel ( WDOG_COUNTER_RESTART_KICK_VALUE , dw_wdt - > regs +
2015-01-28 01:25:16 +03:00
WDOG_COUNTER_RESTART_REG_OFFSET ) ;
2016-02-29 00:12:19 +03:00
return 0 ;
2015-01-28 01:25:16 +03:00
}
2016-02-29 00:12:19 +03:00
static int dw_wdt_set_timeout ( struct watchdog_device * wdd , unsigned int top_s )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
2020-05-30 10:35:54 +03:00
unsigned int timeout ;
u32 top_val ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:56 +03:00
/*
* Note IRQ mode being enabled means having a non - zero pre - timeout
* setup . In this case we try to find a TOP as close to the half of the
* requested timeout as possible since DW Watchdog IRQ mode is designed
* in two stages way - first timeout rises the pre - timeout interrupt ,
* second timeout performs the system reset . So basically the effective
* watchdog - caused reset happens after two watchdog TOPs elapsed .
*/
timeout = dw_wdt_find_best_top ( dw_wdt , DIV_ROUND_UP ( top_s , dw_wdt - > rmod ) ,
& top_val ) ;
if ( dw_wdt - > rmod = = DW_WDT_RMOD_IRQ )
wdd - > pretimeout = timeout ;
else
wdd - > pretimeout = 0 ;
2011-01-24 15:19:12 +03:00
2015-01-28 01:25:16 +03:00
/*
* Set the new value in the watchdog . Some versions of dw_wdt
* have have TOPINIT in the TIMEOUT_RANGE register ( as per
* CP_WDT_DUAL_TOP in WDT_COMP_PARAMS_1 ) . On those we
* effectively get a pat of the watchdog right here .
*/
2014-09-23 11:42:11 +04:00
writel ( top_val | top_val < < WDOG_TIMEOUT_RANGE_TOPINIT_SHIFT ,
2016-02-29 00:12:19 +03:00
dw_wdt - > regs + WDOG_TIMEOUT_RANGE_REG_OFFSET ) ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:56 +03:00
/* Kick new TOP value into the watchdog counter if activated. */
if ( watchdog_active ( wdd ) )
dw_wdt_ping ( wdd ) ;
2019-11-25 05:04:13 +03:00
/*
* In case users set bigger timeout value than HW can support ,
* kernel ( watchdog_dev . c ) helps to feed watchdog before
* wdd - > max_hw_heartbeat_ms
*/
if ( top_s * 1000 < = wdd - > max_hw_heartbeat_ms )
2020-05-30 10:35:56 +03:00
wdd - > timeout = timeout * dw_wdt - > rmod ;
2019-11-25 05:04:13 +03:00
else
wdd - > timeout = top_s ;
2015-01-28 01:25:16 +03:00
2016-02-29 00:12:19 +03:00
return 0 ;
}
2020-05-30 10:35:56 +03:00
static int dw_wdt_set_pretimeout ( struct watchdog_device * wdd , unsigned int req )
{
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
/*
* We ignore actual value of the timeout passed from user - space
* using it as a flag whether the pretimeout functionality is intended
* to be activated .
*/
dw_wdt_update_mode ( dw_wdt , req ? DW_WDT_RMOD_IRQ : DW_WDT_RMOD_RESET ) ;
dw_wdt_set_timeout ( wdd , wdd - > timeout ) ;
return 0 ;
}
2018-03-10 06:46:06 +03:00
static void dw_wdt_arm_system_reset ( struct dw_wdt * dw_wdt )
{
u32 val = readl ( dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
2020-05-30 10:35:56 +03:00
/* Disable/enable interrupt mode depending on the RMOD flag. */
if ( dw_wdt - > rmod = = DW_WDT_RMOD_IRQ )
val | = WDOG_CONTROL_REG_RESP_MODE_MASK ;
else
val & = ~ WDOG_CONTROL_REG_RESP_MODE_MASK ;
2018-03-10 06:46:06 +03:00
/* Enable watchdog. */
val | = WDOG_CONTROL_REG_WDT_EN_MASK ;
writel ( val , dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
}
2016-02-29 00:12:19 +03:00
static int dw_wdt_start ( struct watchdog_device * wdd )
{
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
dw_wdt_set_timeout ( wdd , wdd - > timeout ) ;
2020-01-07 18:51:55 +03:00
dw_wdt_ping ( & dw_wdt - > wdd ) ;
2018-03-10 06:46:06 +03:00
dw_wdt_arm_system_reset ( dw_wdt ) ;
2016-02-29 00:12:19 +03:00
return 0 ;
2011-01-24 15:19:12 +03:00
}
2017-09-26 09:11:22 +03:00
static int dw_wdt_stop ( struct watchdog_device * wdd )
{
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
if ( ! dw_wdt - > rst ) {
set_bit ( WDOG_HW_RUNNING , & wdd - > status ) ;
return 0 ;
}
reset_control_assert ( dw_wdt - > rst ) ;
reset_control_deassert ( dw_wdt - > rst ) ;
return 0 ;
}
2017-01-04 23:27:21 +03:00
static int dw_wdt_restart ( struct watchdog_device * wdd ,
unsigned long action , void * data )
2014-09-23 11:42:12 +04:00
{
2017-01-04 23:27:21 +03:00
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
2014-09-23 11:42:12 +04:00
2016-02-29 00:12:19 +03:00
writel ( 0 , dw_wdt - > regs + WDOG_TIMEOUT_RANGE_REG_OFFSET ) ;
2020-05-30 10:35:56 +03:00
dw_wdt_update_mode ( dw_wdt , DW_WDT_RMOD_RESET ) ;
2018-03-10 06:46:06 +03:00
if ( dw_wdt_is_enabled ( dw_wdt ) )
2016-02-29 00:12:19 +03:00
writel ( WDOG_COUNTER_RESTART_KICK_VALUE ,
dw_wdt - > regs + WDOG_COUNTER_RESTART_REG_OFFSET ) ;
2014-09-23 11:42:12 +04:00
else
2018-03-10 06:46:06 +03:00
dw_wdt_arm_system_reset ( dw_wdt ) ;
2014-09-23 11:42:12 +04:00
/* wait for reset to assert... */
mdelay ( 500 ) ;
2017-01-04 23:27:21 +03:00
return 0 ;
2014-09-23 11:42:12 +04:00
}
2016-02-29 00:12:19 +03:00
static unsigned int dw_wdt_get_timeleft ( struct watchdog_device * wdd )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = to_dw_wdt ( wdd ) ;
2020-05-30 10:35:56 +03:00
unsigned int sec ;
u32 val ;
val = readl ( dw_wdt - > regs + WDOG_CURRENT_COUNT_REG_OFFSET ) ;
sec = val / dw_wdt - > rate ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:56 +03:00
if ( dw_wdt - > rmod = = DW_WDT_RMOD_IRQ ) {
val = readl ( dw_wdt - > regs + WDOG_INTERRUPT_STATUS_REG_OFFSET ) ;
if ( ! val )
sec + = wdd - > pretimeout ;
}
return sec ;
2011-01-24 15:19:12 +03:00
}
static const struct watchdog_info dw_wdt_ident = {
. options = WDIOF_KEEPALIVEPING | WDIOF_SETTIMEOUT |
WDIOF_MAGICCLOSE ,
. identity = " Synopsys DesignWare Watchdog " ,
} ;
2020-05-30 10:35:56 +03:00
static const struct watchdog_info dw_wdt_pt_ident = {
. options = WDIOF_KEEPALIVEPING | WDIOF_SETTIMEOUT |
WDIOF_PRETIMEOUT | WDIOF_MAGICCLOSE ,
. identity = " Synopsys DesignWare Watchdog " ,
} ;
2016-02-29 00:12:19 +03:00
static const struct watchdog_ops dw_wdt_ops = {
. owner = THIS_MODULE ,
. start = dw_wdt_start ,
2017-09-26 09:11:22 +03:00
. stop = dw_wdt_stop ,
2016-02-29 00:12:19 +03:00
. ping = dw_wdt_ping ,
. set_timeout = dw_wdt_set_timeout ,
2020-05-30 10:35:56 +03:00
. set_pretimeout = dw_wdt_set_pretimeout ,
2016-02-29 00:12:19 +03:00
. get_timeleft = dw_wdt_get_timeleft ,
2017-01-04 23:27:21 +03:00
. restart = dw_wdt_restart ,
2016-02-29 00:12:19 +03:00
} ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:56 +03:00
static irqreturn_t dw_wdt_irq ( int irq , void * devid )
{
struct dw_wdt * dw_wdt = devid ;
u32 val ;
/*
* We don ' t clear the IRQ status . It ' s supposed to be done by the
* following ping operations .
*/
val = readl ( dw_wdt - > regs + WDOG_INTERRUPT_STATUS_REG_OFFSET ) ;
if ( ! val )
return IRQ_NONE ;
watchdog_notify_pretimeout ( & dw_wdt - > wdd ) ;
return IRQ_HANDLED ;
}
2013-06-26 22:03:52 +04:00
# ifdef CONFIG_PM_SLEEP
2011-01-24 15:19:12 +03:00
static int dw_wdt_suspend ( struct device * dev )
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = dev_get_drvdata ( dev ) ;
2018-03-10 06:46:07 +03:00
dw_wdt - > control = readl ( dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
dw_wdt - > timeout = readl ( dw_wdt - > regs + WDOG_TIMEOUT_RANGE_REG_OFFSET ) ;
2020-05-30 10:35:55 +03:00
clk_disable_unprepare ( dw_wdt - > pclk ) ;
2016-02-29 00:12:19 +03:00
clk_disable_unprepare ( dw_wdt - > clk ) ;
2011-01-24 15:19:12 +03:00
return 0 ;
}
static int dw_wdt_resume ( struct device * dev )
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = dev_get_drvdata ( dev ) ;
int err = clk_prepare_enable ( dw_wdt - > clk ) ;
2011-01-24 15:19:12 +03:00
if ( err )
return err ;
2020-05-30 10:35:55 +03:00
err = clk_prepare_enable ( dw_wdt - > pclk ) ;
if ( err ) {
clk_disable_unprepare ( dw_wdt - > clk ) ;
return err ;
}
2018-03-10 06:46:07 +03:00
writel ( dw_wdt - > timeout , dw_wdt - > regs + WDOG_TIMEOUT_RANGE_REG_OFFSET ) ;
writel ( dw_wdt - > control , dw_wdt - > regs + WDOG_CONTROL_REG_OFFSET ) ;
2016-02-29 00:12:19 +03:00
dw_wdt_ping ( & dw_wdt - > wdd ) ;
2011-01-24 15:19:12 +03:00
return 0 ;
}
2013-06-26 22:03:52 +04:00
# endif /* CONFIG_PM_SLEEP */
2011-01-24 15:19:12 +03:00
2013-06-26 22:03:52 +04:00
static SIMPLE_DEV_PM_OPS ( dw_wdt_pm_ops , dw_wdt_suspend , dw_wdt_resume ) ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:54 +03:00
/*
* In case if DW WDT IP core is synthesized with fixed TOP feature disabled the
* TOPs array can be arbitrary ordered with nearly any sixteen uint numbers
* depending on the system engineer imagination . The next method handles the
* passed TOPs array to pre - calculate the effective timeouts and to sort the
* TOP items out in the ascending order with respect to the timeouts .
*/
static void dw_wdt_handle_tops ( struct dw_wdt * dw_wdt , const u32 * tops )
{
struct dw_wdt_timeout tout , * dst ;
int val , tidx ;
u64 msec ;
/*
* We walk over the passed TOPs array and calculate corresponding
* timeouts in seconds and milliseconds . The milliseconds granularity
* is needed to distinguish the TOPs with very close timeouts and to
* set the watchdog max heartbeat setting further .
*/
for ( val = 0 ; val < DW_WDT_NUM_TOPS ; + + val ) {
tout . top_val = val ;
tout . sec = tops [ val ] / dw_wdt - > rate ;
msec = ( u64 ) tops [ val ] * MSEC_PER_SEC ;
do_div ( msec , dw_wdt - > rate ) ;
tout . msec = msec - ( ( u64 ) tout . sec * MSEC_PER_SEC ) ;
/*
* Find a suitable place for the current TOP in the timeouts
* array so that the list is remained in the ascending order .
*/
for ( tidx = 0 ; tidx < val ; + + tidx ) {
dst = & dw_wdt - > timeouts [ tidx ] ;
if ( tout . sec > dst - > sec | | ( tout . sec = = dst - > sec & &
tout . msec > = dst - > msec ) )
continue ;
else
swap ( * dst , tout ) ;
}
dw_wdt - > timeouts [ val ] = tout ;
}
}
static int dw_wdt_init_timeouts ( struct dw_wdt * dw_wdt , struct device * dev )
{
u32 data , of_tops [ DW_WDT_NUM_TOPS ] ;
const u32 * tops ;
int ret ;
/*
* Retrieve custom or fixed counter values depending on the
* WDT_USE_FIX_TOP flag found in the component specific parameters
* # 1 register .
*/
data = readl ( dw_wdt - > regs + WDOG_COMP_PARAMS_1_REG_OFFSET ) ;
if ( data & WDOG_COMP_PARAMS_1_USE_FIX_TOP ) {
tops = dw_wdt_fix_tops ;
} else {
ret = of_property_read_variable_u32_array ( dev_of_node ( dev ) ,
" snps,watchdog-tops " , of_tops , DW_WDT_NUM_TOPS ,
DW_WDT_NUM_TOPS ) ;
if ( ret < 0 ) {
dev_warn ( dev , " No valid TOPs array specified \n " ) ;
tops = dw_wdt_fix_tops ;
} else {
tops = of_tops ;
}
}
/* Convert the specified TOPs into an array of watchdog timeouts. */
dw_wdt_handle_tops ( dw_wdt , tops ) ;
if ( ! dw_wdt - > timeouts [ DW_WDT_NUM_TOPS - 1 ] . sec ) {
dev_err ( dev , " No any valid TOP detected \n " ) ;
return - EINVAL ;
}
return 0 ;
}
2020-05-30 10:35:57 +03:00
# ifdef CONFIG_DEBUG_FS
# define DW_WDT_DBGFS_REG(_name, _off) \
{ \
. name = _name , \
. offset = _off \
}
static const struct debugfs_reg32 dw_wdt_dbgfs_regs [ ] = {
DW_WDT_DBGFS_REG ( " cr " , WDOG_CONTROL_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " torr " , WDOG_TIMEOUT_RANGE_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " ccvr " , WDOG_CURRENT_COUNT_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " crr " , WDOG_COUNTER_RESTART_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " stat " , WDOG_INTERRUPT_STATUS_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " param5 " , WDOG_COMP_PARAMS_5_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " param4 " , WDOG_COMP_PARAMS_4_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " param3 " , WDOG_COMP_PARAMS_3_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " param2 " , WDOG_COMP_PARAMS_2_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " param1 " , WDOG_COMP_PARAMS_1_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " version " , WDOG_COMP_VERSION_REG_OFFSET ) ,
DW_WDT_DBGFS_REG ( " type " , WDOG_COMP_TYPE_REG_OFFSET )
} ;
static void dw_wdt_dbgfs_init ( struct dw_wdt * dw_wdt )
{
struct device * dev = dw_wdt - > wdd . parent ;
struct debugfs_regset32 * regset ;
regset = devm_kzalloc ( dev , sizeof ( * regset ) , GFP_KERNEL ) ;
if ( ! regset )
return ;
regset - > regs = dw_wdt_dbgfs_regs ;
regset - > nregs = ARRAY_SIZE ( dw_wdt_dbgfs_regs ) ;
regset - > base = dw_wdt - > regs ;
dw_wdt - > dbgfs_dir = debugfs_create_dir ( dev_name ( dev ) , NULL ) ;
debugfs_create_regset32 ( " registers " , 0444 , dw_wdt - > dbgfs_dir , regset ) ;
}
static void dw_wdt_dbgfs_clear ( struct dw_wdt * dw_wdt )
{
debugfs_remove_recursive ( dw_wdt - > dbgfs_dir ) ;
}
# else /* !CONFIG_DEBUG_FS */
static void dw_wdt_dbgfs_init ( struct dw_wdt * dw_wdt ) { }
static void dw_wdt_dbgfs_clear ( struct dw_wdt * dw_wdt ) { }
# endif /* !CONFIG_DEBUG_FS */
2012-11-19 22:21:41 +04:00
static int dw_wdt_drv_probe ( struct platform_device * pdev )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
struct device * dev = & pdev - > dev ;
struct watchdog_device * wdd ;
struct dw_wdt * dw_wdt ;
2011-01-24 15:19:12 +03:00
int ret ;
2016-02-29 00:12:19 +03:00
dw_wdt = devm_kzalloc ( dev , sizeof ( * dw_wdt ) , GFP_KERNEL ) ;
if ( ! dw_wdt )
return - ENOMEM ;
watchdog: Convert to use devm_platform_ioremap_resource
Use devm_platform_ioremap_resource to reduce source code size,
improve readability, and reduce the likelyhood of bugs.
The conversion was done automatically with coccinelle using the
following semantic patch.
@r@
identifier res, pdev;
expression a;
expression index;
expression e;
@@
<+...
- res = platform_get_resource(pdev, IORESOURCE_MEM, index);
- a = devm_ioremap_resource(e, res);
+ a = devm_platform_ioremap_resource(pdev, index);
...+>
@depends on r@
identifier r.res;
@@
- struct resource *res;
... when != res
@@
identifier res, pdev;
expression index;
expression a;
@@
- struct resource *res = platform_get_resource(pdev, IORESOURCE_MEM, index);
- a = devm_ioremap_resource(&pdev->dev, res);
+ a = devm_platform_ioremap_resource(pdev, index);
Cc: Joel Stanley <joel@jms.id.au>
Cc: Nicolas Ferre <nicolas.ferre@microchip.com>
Cc: Alexandre Belloni <alexandre.belloni@bootlin.com>
Cc: Florian Fainelli <f.fainelli@gmail.com>
Cc: Linus Walleij <linus.walleij@linaro.org>
Cc: Baruch Siach <baruch@tkos.co.il>
Cc: Keguang Zhang <keguang.zhang@gmail.com>
Cc: Vladimir Zapolskiy <vz@mleia.com>
Cc: Kevin Hilman <khilman@baylibre.com>
Cc: Matthias Brugger <matthias.bgg@gmail.com>
Cc: Avi Fishman <avifishman70@gmail.com>
Cc: Nancy Yuen <yuenn@google.com>
Cc: Brendan Higgins <brendanhiggins@google.com>
Cc: Wan ZongShun <mcuos.com@gmail.com>
Cc: Michal Simek <michal.simek@xilinx.com>
Cc: Sylvain Lemieux <slemieux.tyco@gmail.com>
Cc: Kukjin Kim <kgene@kernel.org>
Cc: Barry Song <baohua@kernel.org>
Cc: Orson Zhai <orsonzhai@gmail.com>
Cc: Patrice Chotard <patrice.chotard@st.com>
Cc: Maxime Coquelin <mcoquelin.stm32@gmail.com>
Cc: Maxime Ripard <maxime.ripard@bootlin.com>
Cc: Chen-Yu Tsai <wens@csie.org>
Cc: Marc Gonzalez <marc.w.gonzalez@free.fr>
Cc: Thierry Reding <thierry.reding@gmail.com>
Cc: Shawn Guo <shawnguo@kernel.org>
Signed-off-by: Guenter Roeck <linux@roeck-us.net>
Acked-by: Alexandre Belloni <alexandre.belloni@bootlin.com>
Tested-by: Alexandre Belloni <alexandre.belloni@bootlin.com>
Acked-by: Joel Stanley <joel@jms.id.au>
Reviewed-by: Linus Walleij <linus.walleij@linaro.org>
Acked-by: Maxime Ripard <maxime.ripard@bootlin.com>
Acked-by: Michal Simek <michal.simek@xilinx.com> (cadence/xilinx wdts)
Acked-by: Thierry Reding <treding@nvidia.com>
Reviewed-by: Florian Fainelli <f.fainelli@gmail.com>
Acked-by: Patrice Chotard <patrice.chotard@st.com>
Acked-by: Vladimir Zapolskiy <vz@mleia.com>
Signed-off-by: Guenter Roeck <linux@roeck-us.net>
Signed-off-by: Wim Van Sebroeck <wim@linux-watchdog.org>
2019-04-02 22:01:53 +03:00
dw_wdt - > regs = devm_platform_ioremap_resource ( pdev , 0 ) ;
2016-02-29 00:12:19 +03:00
if ( IS_ERR ( dw_wdt - > regs ) )
return PTR_ERR ( dw_wdt - > regs ) ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:55 +03:00
/*
* Try to request the watchdog dedicated timer clock source . It must
* be supplied if asynchronous mode is enabled . Otherwise fallback
* to the common timer / bus clocks configuration , in which the very
* first found clock supply both timer and APB signals .
*/
dw_wdt - > clk = devm_clk_get ( dev , " tclk " ) ;
if ( IS_ERR ( dw_wdt - > clk ) ) {
dw_wdt - > clk = devm_clk_get ( dev , NULL ) ;
if ( IS_ERR ( dw_wdt - > clk ) )
return PTR_ERR ( dw_wdt - > clk ) ;
}
2011-01-24 15:19:12 +03:00
2016-02-29 00:12:19 +03:00
ret = clk_prepare_enable ( dw_wdt - > clk ) ;
2011-01-24 15:19:12 +03:00
if ( ret )
2013-04-29 13:15:26 +04:00
return ret ;
2011-01-24 15:19:12 +03:00
2016-08-10 08:35:58 +03:00
dw_wdt - > rate = clk_get_rate ( dw_wdt - > clk ) ;
if ( dw_wdt - > rate = = 0 ) {
ret = - EINVAL ;
goto out_disable_clk ;
}
2020-05-30 10:35:55 +03:00
/*
* Request APB clock if device is configured with async clocks mode .
* In this case both tclk and pclk clocks are supposed to be specified .
* Alas we can ' t know for sure whether async mode was really activated ,
* so the pclk phandle reference is left optional . If it couldn ' t be
* found we consider the device configured in synchronous clocks mode .
*/
dw_wdt - > pclk = devm_clk_get_optional ( dev , " pclk " ) ;
if ( IS_ERR ( dw_wdt - > pclk ) ) {
ret = PTR_ERR ( dw_wdt - > pclk ) ;
goto out_disable_clk ;
}
ret = clk_prepare_enable ( dw_wdt - > pclk ) ;
if ( ret )
goto out_disable_clk ;
2017-05-22 11:51:39 +03:00
dw_wdt - > rst = devm_reset_control_get_optional_shared ( & pdev - > dev , NULL ) ;
if ( IS_ERR ( dw_wdt - > rst ) ) {
ret = PTR_ERR ( dw_wdt - > rst ) ;
2020-05-30 10:35:55 +03:00
goto out_disable_pclk ;
2017-05-22 11:51:39 +03:00
}
2020-05-30 10:35:56 +03:00
/* Enable normal reset without pre-timeout by default. */
dw_wdt_update_mode ( dw_wdt , DW_WDT_RMOD_RESET ) ;
/*
* Pre - timeout IRQ is optional , since some hardware may lack support
* of it . Note we must request rising - edge IRQ , since the lane is left
* pending either until the next watchdog kick event or up to the
* system reset .
*/
ret = platform_get_irq_optional ( pdev , 0 ) ;
if ( ret > 0 ) {
ret = devm_request_irq ( dev , ret , dw_wdt_irq ,
IRQF_SHARED | IRQF_TRIGGER_RISING ,
pdev - > name , dw_wdt ) ;
if ( ret )
goto out_disable_pclk ;
dw_wdt - > wdd . info = & dw_wdt_pt_ident ;
} else {
if ( ret = = - EPROBE_DEFER )
goto out_disable_pclk ;
dw_wdt - > wdd . info = & dw_wdt_ident ;
}
2017-05-22 11:51:39 +03:00
reset_control_deassert ( dw_wdt - > rst ) ;
2020-05-30 10:35:54 +03:00
ret = dw_wdt_init_timeouts ( dw_wdt , dev ) ;
if ( ret )
goto out_disable_clk ;
2016-02-29 00:12:19 +03:00
wdd = & dw_wdt - > wdd ;
wdd - > ops = & dw_wdt_ops ;
2020-05-30 10:35:54 +03:00
wdd - > min_timeout = dw_wdt_get_min_timeout ( dw_wdt ) ;
wdd - > max_hw_heartbeat_ms = dw_wdt_get_max_timeout_ms ( dw_wdt ) ;
2016-02-29 00:12:19 +03:00
wdd - > parent = dev ;
watchdog_set_drvdata ( wdd , dw_wdt ) ;
watchdog_set_nowayout ( wdd , nowayout ) ;
watchdog_init_timeout ( wdd , 0 , dev ) ;
/*
* If the watchdog is already running , use its already configured
* timeout . Otherwise use the default or the value provided through
* devicetree .
*/
if ( dw_wdt_is_enabled ( dw_wdt ) ) {
2020-05-30 10:35:54 +03:00
wdd - > timeout = dw_wdt_get_timeout ( dw_wdt ) ;
2016-02-29 00:12:19 +03:00
set_bit ( WDOG_HW_RUNNING , & wdd - > status ) ;
} else {
wdd - > timeout = DW_WDT_DEFAULT_SECONDS ;
watchdog_init_timeout ( wdd , 0 , dev ) ;
}
platform_set_drvdata ( pdev , dw_wdt ) ;
2017-01-04 23:27:21 +03:00
watchdog_set_restart_priority ( wdd , 128 ) ;
2016-02-29 00:12:19 +03:00
ret = watchdog_register_device ( wdd ) ;
2011-01-24 15:19:12 +03:00
if ( ret )
2020-05-30 10:35:55 +03:00
goto out_disable_pclk ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:57 +03:00
dw_wdt_dbgfs_init ( dw_wdt ) ;
2011-01-24 15:19:12 +03:00
return 0 ;
2020-05-30 10:35:55 +03:00
out_disable_pclk :
clk_disable_unprepare ( dw_wdt - > pclk ) ;
2011-01-24 15:19:12 +03:00
out_disable_clk :
2016-02-29 00:12:19 +03:00
clk_disable_unprepare ( dw_wdt - > clk ) ;
2011-01-24 15:19:12 +03:00
return ret ;
}
2012-11-19 22:26:24 +04:00
static int dw_wdt_drv_remove ( struct platform_device * pdev )
2011-01-24 15:19:12 +03:00
{
2016-02-29 00:12:19 +03:00
struct dw_wdt * dw_wdt = platform_get_drvdata ( pdev ) ;
2011-01-24 15:19:12 +03:00
2020-05-30 10:35:57 +03:00
dw_wdt_dbgfs_clear ( dw_wdt ) ;
2016-02-29 00:12:19 +03:00
watchdog_unregister_device ( & dw_wdt - > wdd ) ;
2017-05-22 11:51:39 +03:00
reset_control_assert ( dw_wdt - > rst ) ;
2020-05-30 10:35:55 +03:00
clk_disable_unprepare ( dw_wdt - > pclk ) ;
2016-02-29 00:12:19 +03:00
clk_disable_unprepare ( dw_wdt - > clk ) ;
2011-01-24 15:19:12 +03:00
return 0 ;
}
2013-10-22 20:59:12 +04:00
# ifdef CONFIG_OF
static const struct of_device_id dw_wdt_of_match [ ] = {
{ . compatible = " snps,dw-wdt " , } ,
{ /* sentinel */ }
} ;
MODULE_DEVICE_TABLE ( of , dw_wdt_of_match ) ;
# endif
2011-01-24 15:19:12 +03:00
static struct platform_driver dw_wdt_driver = {
. probe = dw_wdt_drv_probe ,
2012-11-19 22:21:12 +04:00
. remove = dw_wdt_drv_remove ,
2011-01-24 15:19:12 +03:00
. driver = {
. name = " dw_wdt " ,
2013-10-22 20:59:12 +04:00
. of_match_table = of_match_ptr ( dw_wdt_of_match ) ,
2011-01-24 15:19:12 +03:00
. pm = & dw_wdt_pm_ops ,
} ,
} ;
2011-11-29 09:56:27 +04:00
module_platform_driver ( dw_wdt_driver ) ;
2011-01-24 15:19:12 +03:00
MODULE_AUTHOR ( " Jamie Iles " ) ;
MODULE_DESCRIPTION ( " Synopsys DesignWare Watchdog Driver " ) ;
MODULE_LICENSE ( " GPL " ) ;