2019-01-07 14:07:41 +03:00
/* SPDX-License-Identifier: GPL-2.0 */
2008-07-08 22:59:42 +04:00
/*
* Driver for the Synopsys DesignWare AHB DMA Controller
*
* Copyright ( C ) 2005 - 2007 Atmel Corporation
2011-05-24 12:34:09 +04:00
* Copyright ( C ) 2010 - 2011 ST Microelectronics
2017-01-17 14:57:30 +03:00
* Copyright ( C ) 2016 Intel Corporation
2008-07-08 22:59:42 +04:00
*/
2017-01-17 14:57:29 +03:00
# include <linux/bitops.h>
2013-06-05 16:26:45 +04:00
# include <linux/interrupt.h>
2013-01-10 12:53:03 +04:00
# include <linux/dmaengine.h>
2008-07-08 22:59:42 +04:00
2017-01-17 14:57:30 +03:00
# include <linux/io-64-nonatomic-hi-lo.h>
2016-11-25 17:59:07 +03:00
# include "internal.h"
2013-02-28 01:36:03 +04:00
# define DW_DMA_MAX_NR_REQUESTS 16
2008-07-08 22:59:42 +04:00
2012-02-01 14:42:28 +04:00
/* flow controller */
enum dw_dma_fc {
DW_DMA_FC_D_M2M ,
DW_DMA_FC_D_M2P ,
DW_DMA_FC_D_P2M ,
DW_DMA_FC_D_P2P ,
DW_DMA_FC_P_P2M ,
DW_DMA_FC_SP_P2P ,
DW_DMA_FC_P_M2P ,
DW_DMA_FC_DP_P2P ,
} ;
2008-07-08 22:59:42 +04:00
/*
* Redefine this macro to handle differences between 32 - and 64 - bit
* addressing , big vs . little endian , etc .
*/
# define DW_REG(name) u32 name; u32 __pad_##name
/* Hardware register definitions. */
struct dw_dma_chan_regs {
DW_REG ( SAR ) ; /* Source Address Register */
DW_REG ( DAR ) ; /* Destination Address Register */
DW_REG ( LLP ) ; /* Linked List Pointer */
u32 CTL_LO ; /* Control Register Low */
u32 CTL_HI ; /* Control Register High */
DW_REG ( SSTAT ) ;
DW_REG ( DSTAT ) ;
DW_REG ( SSTATAR ) ;
DW_REG ( DSTATAR ) ;
u32 CFG_LO ; /* Configuration Register Low */
u32 CFG_HI ; /* Configuration Register High */
DW_REG ( SGR ) ;
DW_REG ( DSR ) ;
} ;
struct dw_dma_irq_regs {
DW_REG ( XFER ) ;
DW_REG ( BLOCK ) ;
DW_REG ( SRC_TRAN ) ;
DW_REG ( DST_TRAN ) ;
DW_REG ( ERROR ) ;
} ;
struct dw_dma_regs {
/* per-channel registers */
struct dw_dma_chan_regs CHAN [ DW_DMA_MAX_NR_CHANNELS ] ;
/* irq handling */
struct dw_dma_irq_regs RAW ; /* r */
struct dw_dma_irq_regs STATUS ; /* r (raw & mask) */
struct dw_dma_irq_regs MASK ; /* rw (set = irq enabled) */
struct dw_dma_irq_regs CLEAR ; /* w (ack, affects "raw") */
DW_REG ( STATUS_INT ) ; /* r */
/* software handshaking */
DW_REG ( REQ_SRC ) ;
DW_REG ( REQ_DST ) ;
DW_REG ( SGL_REQ_SRC ) ;
DW_REG ( SGL_REQ_DST ) ;
DW_REG ( LAST_SRC ) ;
DW_REG ( LAST_DST ) ;
/* miscellaneous */
DW_REG ( CFG ) ;
DW_REG ( CH_EN ) ;
DW_REG ( ID ) ;
DW_REG ( TEST ) ;
2017-01-17 14:57:30 +03:00
/* iDMA 32-bit support */
DW_REG ( CLASS_PRIORITY0 ) ;
DW_REG ( CLASS_PRIORITY1 ) ;
2012-09-21 16:05:45 +04:00
2012-06-19 14:34:01 +04:00
/* optional encoded params, 0x3c8..0x3f7 */
2012-09-21 16:05:45 +04:00
u32 __reserved ;
/* per-channel configuration registers */
u32 DWC_PARAMS [ DW_DMA_MAX_NR_CHANNELS ] ;
u32 MULTI_BLK_TYPE ;
u32 MAX_BLK_SIZE ;
/* top-level parameters */
u32 DW_PARAMS ;
2017-01-17 14:57:30 +03:00
/* component ID */
u32 COMP_TYPE ;
u32 COMP_VERSION ;
/* iDMA 32-bit support */
DW_REG ( FIFO_PARTITION0 ) ;
DW_REG ( FIFO_PARTITION1 ) ;
DW_REG ( SAI_ERR ) ;
DW_REG ( GLOBAL_CFG ) ;
2008-07-08 22:59:42 +04:00
} ;
2012-09-21 16:05:45 +04:00
/* Bitfields in DW_PARAMS */
# define DW_PARAMS_NR_CHAN 8 /* number of channels */
# define DW_PARAMS_NR_MASTER 11 /* number of AHB masters */
# define DW_PARAMS_DATA_WIDTH(n) (15 + 2 * (n))
# define DW_PARAMS_DATA_WIDTH1 15 /* master 1 data width */
# define DW_PARAMS_DATA_WIDTH2 17 /* master 2 data width */
# define DW_PARAMS_DATA_WIDTH3 19 /* master 3 data width */
# define DW_PARAMS_DATA_WIDTH4 21 /* master 4 data width */
# define DW_PARAMS_EN 28 /* encoded parameters */
/* Bitfields in DWC_PARAMS */
# define DWC_PARAMS_MBLK_EN 11 /* multi block transfer */
2020-07-23 03:58:44 +03:00
# define DWC_PARAMS_HC_LLP 13 /* set LLP register to zero */
2020-07-23 03:58:47 +03:00
# define DWC_PARAMS_MSIZE 16 /* max group transaction size */
2012-09-21 16:05:45 +04:00
2014-09-23 18:18:10 +04:00
/* bursts size */
enum dw_dma_msize {
DW_DMA_MSIZE_1 ,
DW_DMA_MSIZE_4 ,
DW_DMA_MSIZE_8 ,
DW_DMA_MSIZE_16 ,
DW_DMA_MSIZE_32 ,
DW_DMA_MSIZE_64 ,
DW_DMA_MSIZE_128 ,
DW_DMA_MSIZE_256 ,
} ;
2016-03-18 17:24:44 +03:00
/* Bitfields in LLP */
# define DWC_LLP_LMS(x) ((x) & 3) /* list master select */
# define DWC_LLP_LOC(x) ((x) & ~3) /* next lli */
2008-07-08 22:59:42 +04:00
/* Bitfields in CTL_LO */
# define DWC_CTLL_INT_EN (1 << 0) /* irqs enabled? */
# define DWC_CTLL_DST_WIDTH(n) ((n)<<1) /* bytes per element */
# define DWC_CTLL_SRC_WIDTH(n) ((n)<<4)
# define DWC_CTLL_DST_INC (0<<7) /* DAR update/not */
# define DWC_CTLL_DST_DEC (1<<7)
# define DWC_CTLL_DST_FIX (2<<7)
2016-01-07 03:39:33 +03:00
# define DWC_CTLL_SRC_INC (0<<9) /* SAR update/not */
2008-07-08 22:59:42 +04:00
# define DWC_CTLL_SRC_DEC (1<<9)
# define DWC_CTLL_SRC_FIX (2<<9)
# define DWC_CTLL_DST_MSIZE(n) ((n)<<11) /* burst, #elements */
# define DWC_CTLL_SRC_MSIZE(n) ((n)<<14)
# define DWC_CTLL_S_GATH_EN (1 << 17) /* src gather, !FIX */
# define DWC_CTLL_D_SCAT_EN (1 << 18) /* dst scatter, !FIX */
2011-03-04 13:12:51 +03:00
# define DWC_CTLL_FC(n) ((n) << 20)
2008-07-08 22:59:42 +04:00
# define DWC_CTLL_FC_M2M (0 << 20) /* mem-to-mem */
# define DWC_CTLL_FC_M2P (1 << 20) /* mem-to-periph */
# define DWC_CTLL_FC_P2M (2 << 20) /* periph-to-mem */
# define DWC_CTLL_FC_P2P (3 << 20) /* periph-to-periph */
/* plus 4 transfer types for peripheral-as-flow-controller */
# define DWC_CTLL_DMS(n) ((n)<<23) /* dst master select */
# define DWC_CTLL_SMS(n) ((n)<<25) /* src master select */
# define DWC_CTLL_LLP_D_EN (1 << 27) /* dest block chain */
# define DWC_CTLL_LLP_S_EN (1 << 28) /* src block chain */
/* Bitfields in CTL_HI */
2017-01-17 14:57:29 +03:00
# define DWC_CTLH_BLOCK_TS_MASK GENMASK(11, 0)
# define DWC_CTLH_BLOCK_TS(x) ((x) & DWC_CTLH_BLOCK_TS_MASK)
# define DWC_CTLH_DONE (1 << 12)
2008-07-08 22:59:42 +04:00
2014-09-23 18:18:10 +04:00
/* Bitfields in CFG_LO */
2011-03-03 13:17:22 +03:00
# define DWC_CFGL_CH_PRIOR_MASK (0x7 << 5) /* priority mask */
# define DWC_CFGL_CH_PRIOR(x) ((x) << 5) /* priority */
2008-07-08 22:59:42 +04:00
# define DWC_CFGL_CH_SUSP (1 << 8) /* pause xfer */
# define DWC_CFGL_FIFO_EMPTY (1 << 9) /* pause xfer */
# define DWC_CFGL_HS_DST (1 << 10) /* handshake w/dst */
# define DWC_CFGL_HS_SRC (1 << 11) /* handshake w/src */
2014-09-23 18:18:10 +04:00
# define DWC_CFGL_LOCK_CH_XFER (0 << 12) /* scope of LOCK_CH */
# define DWC_CFGL_LOCK_CH_BLOCK (1 << 12)
# define DWC_CFGL_LOCK_CH_XACT (2 << 12)
# define DWC_CFGL_LOCK_BUS_XFER (0 << 14) /* scope of LOCK_BUS */
# define DWC_CFGL_LOCK_BUS_BLOCK (1 << 14)
# define DWC_CFGL_LOCK_BUS_XACT (2 << 14)
# define DWC_CFGL_LOCK_CH (1 << 15) /* channel lockout */
# define DWC_CFGL_LOCK_BUS (1 << 16) /* busmaster lockout */
# define DWC_CFGL_HS_DST_POL (1 << 18) /* dst handshake active low */
# define DWC_CFGL_HS_SRC_POL (1 << 19) /* src handshake active low */
2008-07-08 22:59:42 +04:00
# define DWC_CFGL_MAX_BURST(x) ((x) << 20)
# define DWC_CFGL_RELOAD_SAR (1 << 30)
# define DWC_CFGL_RELOAD_DAR (1 << 31)
2014-09-23 18:18:10 +04:00
/* Bitfields in CFG_HI */
# define DWC_CFGH_FCMODE (1 << 0)
# define DWC_CFGH_FIFO_MODE (1 << 1)
# define DWC_CFGH_PROTCTL(x) ((x) << 2)
2018-11-17 19:17:21 +03:00
# define DWC_CFGH_PROTCTL_DATA (0 << 2) /* data access - always set */
# define DWC_CFGH_PROTCTL_PRIV (1 << 2) /* privileged -> AHB HPROT[1] */
# define DWC_CFGH_PROTCTL_BUFFER (2 << 2) /* bufferable -> AHB HPROT[2] */
# define DWC_CFGH_PROTCTL_CACHE (4 << 2) /* cacheable -> AHB HPROT[3] */
2008-07-08 22:59:42 +04:00
# define DWC_CFGH_DS_UPD_EN (1 << 5)
# define DWC_CFGH_SS_UPD_EN (1 << 6)
2014-09-23 18:18:10 +04:00
# define DWC_CFGH_SRC_PER(x) ((x) << 7)
# define DWC_CFGH_DST_PER(x) ((x) << 11)
2008-07-08 22:59:42 +04:00
/* Bitfields in SGR */
# define DWC_SGR_SGI(x) ((x) << 0)
# define DWC_SGR_SGC(x) ((x) << 20)
/* Bitfields in DSR */
# define DWC_DSR_DSI(x) ((x) << 0)
# define DWC_DSR_DSC(x) ((x) << 20)
/* Bitfields in CFG */
# define DW_CFG_DMA_EN (1 << 0)
2017-01-17 14:57:30 +03:00
/* iDMA 32-bit support */
2019-01-07 14:07:40 +03:00
/* bursts size */
enum idma32_msize {
IDMA32_MSIZE_1 ,
IDMA32_MSIZE_2 ,
IDMA32_MSIZE_4 ,
IDMA32_MSIZE_8 ,
IDMA32_MSIZE_16 ,
IDMA32_MSIZE_32 ,
} ;
2017-01-17 14:57:30 +03:00
/* Bitfields in CTL_HI */
# define IDMA32C_CTLH_BLOCK_TS_MASK GENMASK(16, 0)
# define IDMA32C_CTLH_BLOCK_TS(x) ((x) & IDMA32C_CTLH_BLOCK_TS_MASK)
# define IDMA32C_CTLH_DONE (1 << 17)
/* Bitfields in CFG_LO */
# define IDMA32C_CFGL_DST_BURST_ALIGN (1 << 0) /* dst burst align */
# define IDMA32C_CFGL_SRC_BURST_ALIGN (1 << 1) /* src burst align */
# define IDMA32C_CFGL_CH_DRAIN (1 << 10) /* drain FIFO */
# define IDMA32C_CFGL_DST_OPT_BL (1 << 20) /* optimize dst burst length */
# define IDMA32C_CFGL_SRC_OPT_BL (1 << 21) /* optimize src burst length */
/* Bitfields in CFG_HI */
# define IDMA32C_CFGH_SRC_PER(x) ((x) << 0)
# define IDMA32C_CFGH_DST_PER(x) ((x) << 4)
# define IDMA32C_CFGH_RD_ISSUE_THD(x) ((x) << 8)
# define IDMA32C_CFGH_RW_ISSUE_THD(x) ((x) << 18)
# define IDMA32C_CFGH_SRC_PER_EXT(x) ((x) << 28) /* src peripheral extension */
# define IDMA32C_CFGH_DST_PER_EXT(x) ((x) << 30) /* dst peripheral extension */
/* Bitfields in FIFO_PARTITION */
# define IDMA32C_FP_PSIZE_CH0(x) ((x) << 0)
# define IDMA32C_FP_PSIZE_CH1(x) ((x) << 13)
# define IDMA32C_FP_UPDATE (1 << 26)
2009-04-01 17:47:02 +04:00
enum dw_dmac_flags {
DW_DMA_IS_CYCLIC = 0 ,
2012-09-21 16:05:49 +04:00
DW_DMA_IS_SOFT_LLP = 1 ,
2016-03-18 17:24:51 +03:00
DW_DMA_IS_PAUSED = 2 ,
2016-03-18 17:24:52 +03:00
DW_DMA_IS_INITIALIZED = 3 ,
2009-04-01 17:47:02 +04:00
} ;
2008-07-08 22:59:42 +04:00
struct dw_dma_chan {
2013-01-10 12:53:03 +04:00
struct dma_chan chan ;
void __iomem * ch_regs ;
u8 mask ;
u8 priority ;
enum dma_transfer_direction direction ;
2008-07-08 22:59:42 +04:00
2012-09-21 16:05:49 +04:00
/* software emulation of the LLP transfers */
struct list_head * tx_node_active ;
2008-07-08 22:59:42 +04:00
spinlock_t lock ;
/* these other elements are all protected by lock */
2009-04-01 17:47:02 +04:00
unsigned long flags ;
2008-07-08 22:59:42 +04:00
struct list_head active_list ;
struct list_head queue ;
unsigned int descs_allocated ;
2012-02-01 14:42:26 +04:00
2012-09-21 16:05:47 +04:00
/* hardware configuration */
unsigned int block_size ;
2012-09-21 16:05:49 +04:00
bool nollp ;
2020-07-23 03:58:47 +03:00
u32 max_burst ;
2013-03-26 18:53:57 +04:00
/* custom slave configuration */
2016-08-17 19:20:20 +03:00
struct dw_dma_slave dws ;
2012-09-21 16:05:47 +04:00
2014-12-22 17:54:14 +03:00
/* configuration passed via .device_config */
2012-02-01 14:42:26 +04:00
struct dma_slave_config dma_sconfig ;
2008-07-08 22:59:42 +04:00
} ;
static inline struct dw_dma_chan_regs __iomem *
__dwc_regs ( struct dw_dma_chan * dwc )
{
return dwc - > ch_regs ;
}
# define channel_readl(dwc, name) \
2017-05-09 19:18:37 +03:00
readl ( & ( __dwc_regs ( dwc ) - > name ) )
2008-07-08 22:59:42 +04:00
# define channel_writel(dwc, name, val) \
2017-05-09 19:18:37 +03:00
writel ( ( val ) , & ( __dwc_regs ( dwc ) - > name ) )
2008-07-08 22:59:42 +04:00
static inline struct dw_dma_chan * to_dw_dma_chan ( struct dma_chan * chan )
{
return container_of ( chan , struct dw_dma_chan , chan ) ;
}
struct dw_dma {
struct dma_device dma ;
2017-01-17 14:57:26 +03:00
char name [ 20 ] ;
2008-07-08 22:59:42 +04:00
void __iomem * regs ;
2013-01-16 17:48:50 +04:00
struct dma_pool * desc_pool ;
2008-07-08 22:59:42 +04:00
struct tasklet_struct tasklet ;
2014-03-05 17:48:12 +04:00
/* channels */
struct dw_dma_chan * chan ;
2008-07-08 22:59:42 +04:00
u8 all_chan_mask ;
2014-09-23 18:18:14 +04:00
u8 in_use ;
2008-07-08 22:59:42 +04:00
2019-01-07 14:07:38 +03:00
/* Channel operations */
void ( * initialize_chan ) ( struct dw_dma_chan * dwc ) ;
void ( * suspend_chan ) ( struct dw_dma_chan * dwc , bool drain ) ;
2019-01-07 14:07:39 +03:00
void ( * resume_chan ) ( struct dw_dma_chan * dwc , bool drain ) ;
2019-01-07 14:07:40 +03:00
u32 ( * prepare_ctllo ) ( struct dw_dma_chan * dwc ) ;
2019-01-07 14:07:38 +03:00
void ( * encode_maxburst ) ( struct dw_dma_chan * dwc , u32 * maxburst ) ;
u32 ( * bytes2block ) ( struct dw_dma_chan * dwc , size_t bytes ,
unsigned int width , size_t * len ) ;
size_t ( * block2bytes ) ( struct dw_dma_chan * dwc , u32 block , u32 width ) ;
/* Device operations */
void ( * set_device_name ) ( struct dw_dma * dw , int id ) ;
void ( * disable ) ( struct dw_dma * dw ) ;
void ( * enable ) ( struct dw_dma * dw ) ;
2016-04-27 14:15:39 +03:00
/* platform data */
struct dw_dma_platform_data * pdata ;
2008-07-08 22:59:42 +04:00
} ;
static inline struct dw_dma_regs __iomem * __dw_regs ( struct dw_dma * dw )
{
return dw - > regs ;
}
# define dma_readl(dw, name) \
2017-05-09 19:18:37 +03:00
readl ( & ( __dw_regs ( dw ) - > name ) )
2008-07-08 22:59:42 +04:00
# define dma_writel(dw, name, val) \
2017-05-09 19:18:37 +03:00
writel ( ( val ) , & ( __dw_regs ( dw ) - > name ) )
2008-07-08 22:59:42 +04:00
2017-01-17 14:57:30 +03:00
# define idma32_readq(dw, name) \
hi_lo_readq ( & ( __dw_regs ( dw ) - > name ) )
# define idma32_writeq(dw, name, val) \
hi_lo_writeq ( ( val ) , & ( __dw_regs ( dw ) - > name ) )
2008-07-08 22:59:42 +04:00
# define channel_set_bit(dw, reg, mask) \
dma_writel ( dw , reg , ( ( mask ) < < 8 ) | ( mask ) )
# define channel_clear_bit(dw, reg, mask) \
dma_writel ( dw , reg , ( ( mask ) < < 8 ) | 0 )
static inline struct dw_dma * to_dw_dma ( struct dma_device * ddev )
{
return container_of ( ddev , struct dw_dma , dma ) ;
}
/* LLI == Linked List Item; a.k.a. DMA block descriptor */
struct dw_lli {
/* values that are not changed by hardware */
2017-05-09 19:18:37 +03:00
__le32 sar ;
__le32 dar ;
__le32 llp ; /* chain to next lli */
__le32 ctllo ;
2008-07-08 22:59:42 +04:00
/* values that may get written back: */
2017-05-09 19:18:37 +03:00
__le32 ctlhi ;
2008-07-08 22:59:42 +04:00
/* sstat and dstat can snapshot peripheral register state.
* silicon config may discard either or both . . .
*/
2017-05-09 19:18:37 +03:00
__le32 sstat ;
__le32 dstat ;
2008-07-08 22:59:42 +04:00
} ;
struct dw_desc {
/* FIRST values the hardware uses */
struct dw_lli lli ;
2016-03-18 17:24:43 +03:00
# define lli_set(d, reg, v) ((d)->lli.reg |= cpu_to_le32(v))
# define lli_clear(d, reg, v) ((d)->lli.reg &= ~cpu_to_le32(v))
# define lli_read(d, reg) le32_to_cpu((d)->lli.reg)
# define lli_write(d, reg, v) ((d)->lli.reg = cpu_to_le32(v))
2008-07-08 22:59:42 +04:00
/* THEN values for driver housekeeping */
struct list_head desc_node ;
2009-09-09 04:53:02 +04:00
struct list_head tx_list ;
2008-07-08 22:59:42 +04:00
struct dma_async_tx_descriptor txd ;
size_t len ;
2013-01-25 13:48:01 +04:00
size_t total_len ;
2016-03-18 17:24:53 +03:00
u32 residue ;
2008-07-08 22:59:42 +04:00
} ;
2012-10-18 18:34:12 +04:00
# define to_dw_desc(h) list_entry(h, struct dw_desc, desc_node)
2008-07-08 22:59:42 +04:00
static inline struct dw_desc *
txd_to_dw_desc ( struct dma_async_tx_descriptor * txd )
{
return container_of ( txd , struct dw_desc , txd ) ;
}