2008-07-08 22:58:36 +04:00
/*
* offload engine driver for the Marvell XOR engine
* Copyright ( C ) 2007 , 2008 , Marvell International Ltd .
*
* This program is free software ; you can redistribute it and / or modify it
* under the terms and conditions of the GNU General Public License ,
* version 2 , as published by the Free Software Foundation .
*
* This program is distributed in the hope it will be useful , but WITHOUT
* ANY WARRANTY ; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE . See the GNU General Public License for
* more details .
*
* You should have received a copy of the GNU General Public License along with
* this program ; if not , write to the Free Software Foundation , Inc . ,
* 51 Franklin St - Fifth Floor , Boston , MA 02110 - 1301 USA .
*/
# include <linux/init.h>
# include <linux/module.h>
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 11:04:11 +03:00
# include <linux/slab.h>
2008-07-08 22:58:36 +04:00
# include <linux/delay.h>
# include <linux/dma-mapping.h>
# include <linux/spinlock.h>
# include <linux/interrupt.h>
# include <linux/platform_device.h>
# include <linux/memory.h>
2012-02-19 16:30:26 +04:00
# include <linux/clk.h>
2012-11-15 19:47:58 +04:00
# include <linux/of.h>
# include <linux/of_irq.h>
# include <linux/irqdomain.h>
2012-08-24 17:21:54 +04:00
# include <linux/platform_data/dma-mv_xor.h>
2012-03-07 02:34:26 +04:00
# include "dmaengine.h"
2008-07-08 22:58:36 +04:00
# include "mv_xor.h"
static void mv_xor_issue_pending ( struct dma_chan * chan ) ;
# define to_mv_xor_chan(chan) \
2012-11-15 17:57:44 +04:00
container_of ( chan , struct mv_xor_chan , dmachan )
2008-07-08 22:58:36 +04:00
# define to_mv_xor_slot(tx) \
container_of ( tx , struct mv_xor_desc_slot , async_tx )
2012-11-15 17:17:18 +04:00
# define mv_chan_to_devp(chan) \
2012-11-15 18:17:05 +04:00
( ( chan ) - > dmadev . dev )
2012-11-15 17:17:18 +04:00
2014-08-27 17:52:51 +04:00
static void mv_desc_init ( struct mv_xor_desc_slot * desc ,
2014-08-27 17:52:53 +04:00
dma_addr_t addr , u32 byte_count ,
enum dma_ctrl_flags flags )
2008-07-08 22:58:36 +04:00
{
struct mv_xor_desc * hw_desc = desc - > hw_desc ;
2014-08-27 17:52:52 +04:00
hw_desc - > status = XOR_DESC_DMA_OWNED ;
2008-07-08 22:58:36 +04:00
hw_desc - > phy_next_desc = 0 ;
2014-08-27 17:52:53 +04:00
/* Enable end-of-descriptor interrupts only for DMA_PREP_INTERRUPT */
hw_desc - > desc_command = ( flags & DMA_PREP_INTERRUPT ) ?
XOR_DESC_EOD_INT_EN : 0 ;
2014-08-27 17:52:51 +04:00
hw_desc - > phy_dest_addr = addr ;
2008-07-08 22:58:36 +04:00
hw_desc - > byte_count = byte_count ;
}
static void mv_desc_set_next_desc ( struct mv_xor_desc_slot * desc ,
u32 next_desc_addr )
{
struct mv_xor_desc * hw_desc = desc - > hw_desc ;
BUG_ON ( hw_desc - > phy_next_desc ) ;
hw_desc - > phy_next_desc = next_desc_addr ;
}
static void mv_desc_clear_next_desc ( struct mv_xor_desc_slot * desc )
{
struct mv_xor_desc * hw_desc = desc - > hw_desc ;
hw_desc - > phy_next_desc = 0 ;
}
static void mv_desc_set_src_addr ( struct mv_xor_desc_slot * desc ,
int index , dma_addr_t addr )
{
struct mv_xor_desc * hw_desc = desc - > hw_desc ;
2013-07-29 19:42:14 +04:00
hw_desc - > phy_src_addr [ mv_phy_src_idx ( index ) ] = addr ;
2008-07-08 22:58:36 +04:00
if ( desc - > type = = DMA_XOR )
hw_desc - > desc_command | = ( 1 < < index ) ;
}
static u32 mv_chan_get_current_desc ( struct mv_xor_chan * chan )
{
2013-07-29 19:42:13 +04:00
return readl_relaxed ( XOR_CURR_DESC ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_chan_set_next_descriptor ( struct mv_xor_chan * chan ,
u32 next_desc_addr )
{
2013-07-29 19:42:13 +04:00
writel_relaxed ( next_desc_addr , XOR_NEXT_DESC ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_chan_unmask_interrupts ( struct mv_xor_chan * chan )
{
2013-07-29 19:42:13 +04:00
u32 val = readl_relaxed ( XOR_INTR_MASK ( chan ) ) ;
2008-07-08 22:58:36 +04:00
val | = XOR_INTR_MASK_VALUE < < ( chan - > idx * 16 ) ;
2013-07-29 19:42:13 +04:00
writel_relaxed ( val , XOR_INTR_MASK ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static u32 mv_chan_get_intr_cause ( struct mv_xor_chan * chan )
{
2013-07-29 19:42:13 +04:00
u32 intr_cause = readl_relaxed ( XOR_INTR_CAUSE ( chan ) ) ;
2008-07-08 22:58:36 +04:00
intr_cause = ( intr_cause > > ( chan - > idx * 16 ) ) & 0xFFFF ;
return intr_cause ;
}
static void mv_xor_device_clear_eoc_cause ( struct mv_xor_chan * chan )
{
2014-08-27 17:52:53 +04:00
u32 val ;
val = XOR_INT_END_OF_DESC | XOR_INT_END_OF_CHAIN | XOR_INT_STOPPED ;
val = ~ ( val < < ( chan - > idx * 16 ) ) ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( chan ) , " %s, val 0x%08x \n " , __func__ , val ) ;
2013-07-29 19:42:13 +04:00
writel_relaxed ( val , XOR_INTR_CAUSE ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_xor_device_clear_err_status ( struct mv_xor_chan * chan )
{
u32 val = 0xFFFF0000 > > ( chan - > idx * 16 ) ;
2013-07-29 19:42:13 +04:00
writel_relaxed ( val , XOR_INTR_CAUSE ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_set_mode ( struct mv_xor_chan * chan ,
enum dma_transaction_type type )
{
u32 op_mode ;
2013-07-29 19:42:13 +04:00
u32 config = readl_relaxed ( XOR_CONFIG ( chan ) ) ;
2008-07-08 22:58:36 +04:00
switch ( type ) {
case DMA_XOR :
op_mode = XOR_OPERATION_MODE_XOR ;
break ;
case DMA_MEMCPY :
op_mode = XOR_OPERATION_MODE_MEMCPY ;
break ;
default :
2012-11-15 17:17:18 +04:00
dev_err ( mv_chan_to_devp ( chan ) ,
2012-10-28 12:05:44 +04:00
" error: unsupported operation %d \n " ,
2012-11-15 15:50:27 +04:00
type ) ;
2008-07-08 22:58:36 +04:00
BUG ( ) ;
return ;
}
config & = ~ 0x7 ;
config | = op_mode ;
2013-07-29 19:42:14 +04:00
# if defined(__BIG_ENDIAN)
config | = XOR_DESCRIPTOR_SWAP ;
# else
config & = ~ XOR_DESCRIPTOR_SWAP ;
# endif
2013-07-29 19:42:13 +04:00
writel_relaxed ( config , XOR_CONFIG ( chan ) ) ;
2008-07-08 22:58:36 +04:00
chan - > current_type = type ;
}
static void mv_chan_activate ( struct mv_xor_chan * chan )
{
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( chan ) , " activate chan. \n " ) ;
2014-05-22 01:02:35 +04:00
/* writel ensures all descriptors are flushed before activation */
writel ( BIT ( 0 ) , XOR_ACTIVATION ( chan ) ) ;
2008-07-08 22:58:36 +04:00
}
static char mv_chan_is_busy ( struct mv_xor_chan * chan )
{
2013-07-29 19:42:13 +04:00
u32 state = readl_relaxed ( XOR_ACTIVATION ( chan ) ) ;
2008-07-08 22:58:36 +04:00
state = ( state > > 4 ) & 0x3 ;
return ( state = = 1 ) ? 1 : 0 ;
}
/**
* mv_xor_free_slots - flags descriptor slots for reuse
* @ slot : Slot to free
* Caller must hold & mv_chan - > lock while calling this function
*/
static void mv_xor_free_slots ( struct mv_xor_chan * mv_chan ,
struct mv_xor_desc_slot * slot )
{
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s %d slot %p \n " ,
2008-07-08 22:58:36 +04:00
__func__ , __LINE__ , slot ) ;
2014-08-27 17:52:51 +04:00
slot - > slot_used = 0 ;
2008-07-08 22:58:36 +04:00
}
/*
* mv_xor_start_new_chain - program the engine to operate on new chain headed by
* sw_desc
* Caller must hold & mv_chan - > lock while calling this function
*/
static void mv_xor_start_new_chain ( struct mv_xor_chan * mv_chan ,
struct mv_xor_desc_slot * sw_desc )
{
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s %d: sw_desc %p \n " ,
2008-07-08 22:58:36 +04:00
__func__ , __LINE__ , sw_desc ) ;
2013-07-04 02:05:06 +04:00
/* set the hardware chain */
mv_chan_set_next_descriptor ( mv_chan , sw_desc - > async_tx . phys ) ;
2014-08-27 17:52:51 +04:00
mv_chan - > pending + + ;
2012-11-15 17:57:44 +04:00
mv_xor_issue_pending ( & mv_chan - > dmachan ) ;
2008-07-08 22:58:36 +04:00
}
static dma_cookie_t
mv_xor_run_tx_complete_actions ( struct mv_xor_desc_slot * desc ,
struct mv_xor_chan * mv_chan , dma_cookie_t cookie )
{
BUG_ON ( desc - > async_tx . cookie < 0 ) ;
if ( desc - > async_tx . cookie > 0 ) {
cookie = desc - > async_tx . cookie ;
/* call the callback (must not sleep or submit new
* operations to this channel )
*/
if ( desc - > async_tx . callback )
desc - > async_tx . callback (
desc - > async_tx . callback_param ) ;
2013-10-18 21:35:23 +04:00
dma_descriptor_unmap ( & desc - > async_tx ) ;
2008-07-08 22:58:36 +04:00
}
/* run dependent operations */
2009-01-06 03:14:31 +03:00
dma_run_dependencies ( & desc - > async_tx ) ;
2008-07-08 22:58:36 +04:00
return cookie ;
}
static int
mv_xor_clean_completed_slots ( struct mv_xor_chan * mv_chan )
{
struct mv_xor_desc_slot * iter , * _iter ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s %d \n " , __func__ , __LINE__ ) ;
2008-07-08 22:58:36 +04:00
list_for_each_entry_safe ( iter , _iter , & mv_chan - > completed_slots ,
completed_node ) {
if ( async_tx_test_ack ( & iter - > async_tx ) ) {
list_del ( & iter - > completed_node ) ;
mv_xor_free_slots ( mv_chan , iter ) ;
}
}
return 0 ;
}
static int
mv_xor_clean_slot ( struct mv_xor_desc_slot * desc ,
struct mv_xor_chan * mv_chan )
{
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s %d: desc %p flags %d \n " ,
2008-07-08 22:58:36 +04:00
__func__ , __LINE__ , desc , desc - > async_tx . flags ) ;
list_del ( & desc - > chain_node ) ;
/* the client is allowed to attach dependent operations
* until ' ack ' is set
*/
if ( ! async_tx_test_ack ( & desc - > async_tx ) ) {
/* move this slot to the completed_slots */
list_add_tail ( & desc - > completed_node , & mv_chan - > completed_slots ) ;
return 0 ;
}
mv_xor_free_slots ( mv_chan , desc ) ;
return 0 ;
}
static void __mv_xor_slot_cleanup ( struct mv_xor_chan * mv_chan )
{
struct mv_xor_desc_slot * iter , * _iter ;
dma_cookie_t cookie = 0 ;
int busy = mv_chan_is_busy ( mv_chan ) ;
u32 current_desc = mv_chan_get_current_desc ( mv_chan ) ;
int seen_current = 0 ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s %d \n " , __func__ , __LINE__ ) ;
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " current_desc %x \n " , current_desc ) ;
2008-07-08 22:58:36 +04:00
mv_xor_clean_completed_slots ( mv_chan ) ;
/* free completed slots from the chain starting with
* the oldest descriptor
*/
list_for_each_entry_safe ( iter , _iter , & mv_chan - > chain ,
chain_node ) {
prefetch ( _iter ) ;
prefetch ( & _iter - > async_tx ) ;
/* do not advance past the current descriptor loaded into the
* hardware channel , subsequent descriptors are either in
* process or have not been submitted
*/
if ( seen_current )
break ;
/* stop the search if we reach the current descriptor and the
* channel is busy
*/
if ( iter - > async_tx . phys = = current_desc ) {
seen_current = 1 ;
if ( busy )
break ;
}
cookie = mv_xor_run_tx_complete_actions ( iter , mv_chan , cookie ) ;
if ( mv_xor_clean_slot ( iter , mv_chan ) )
break ;
}
if ( ( busy = = 0 ) & & ! list_empty ( & mv_chan - > chain ) ) {
struct mv_xor_desc_slot * chain_head ;
chain_head = list_entry ( mv_chan - > chain . next ,
struct mv_xor_desc_slot ,
chain_node ) ;
mv_xor_start_new_chain ( mv_chan , chain_head ) ;
}
if ( cookie > 0 )
2012-11-15 17:57:44 +04:00
mv_chan - > dmachan . completed_cookie = cookie ;
2008-07-08 22:58:36 +04:00
}
static void
mv_xor_slot_cleanup ( struct mv_xor_chan * mv_chan )
{
spin_lock_bh ( & mv_chan - > lock ) ;
__mv_xor_slot_cleanup ( mv_chan ) ;
spin_unlock_bh ( & mv_chan - > lock ) ;
}
static void mv_xor_tasklet ( unsigned long data )
{
struct mv_xor_chan * chan = ( struct mv_xor_chan * ) data ;
2010-12-21 17:53:39 +03:00
mv_xor_slot_cleanup ( chan ) ;
2008-07-08 22:58:36 +04:00
}
static struct mv_xor_desc_slot *
2014-08-27 17:52:51 +04:00
mv_xor_alloc_slot ( struct mv_xor_chan * mv_chan )
2008-07-08 22:58:36 +04:00
{
2014-08-27 17:52:51 +04:00
struct mv_xor_desc_slot * iter , * _iter ;
int retry = 0 ;
2008-07-08 22:58:36 +04:00
/* start search from the last allocated descrtiptor
* if a contiguous allocation can not be found start searching
* from the beginning of the list
*/
retry :
if ( retry = = 0 )
iter = mv_chan - > last_used ;
else
iter = list_entry ( & mv_chan - > all_slots ,
struct mv_xor_desc_slot ,
slot_node ) ;
list_for_each_entry_safe_continue (
iter , _iter , & mv_chan - > all_slots , slot_node ) {
2014-08-27 17:52:51 +04:00
2008-07-08 22:58:36 +04:00
prefetch ( _iter ) ;
prefetch ( & _iter - > async_tx ) ;
2014-08-27 17:52:51 +04:00
if ( iter - > slot_used ) {
2008-07-08 22:58:36 +04:00
/* give up after finding the first busy slot
* on the second pass through the list
*/
if ( retry )
break ;
continue ;
}
2014-08-27 17:52:51 +04:00
/* pre-ack descriptor */
async_tx_ack ( & iter - > async_tx ) ;
iter - > slot_used = 1 ;
INIT_LIST_HEAD ( & iter - > chain_node ) ;
iter - > async_tx . cookie = - EBUSY ;
mv_chan - > last_used = iter ;
mv_desc_clear_next_desc ( iter ) ;
return iter ;
2008-07-08 22:58:36 +04:00
}
if ( ! retry + + )
goto retry ;
/* try to free some slots if the allocation fails */
tasklet_schedule ( & mv_chan - > irq_tasklet ) ;
return NULL ;
}
/************************ DMA engine API functions ****************************/
static dma_cookie_t
mv_xor_tx_submit ( struct dma_async_tx_descriptor * tx )
{
struct mv_xor_desc_slot * sw_desc = to_mv_xor_slot ( tx ) ;
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( tx - > chan ) ;
2014-08-27 17:52:51 +04:00
struct mv_xor_desc_slot * old_chain_tail ;
2008-07-08 22:58:36 +04:00
dma_cookie_t cookie ;
int new_hw_chain = 1 ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) ,
2008-07-08 22:58:36 +04:00
" %s sw_desc %p: async_tx %p \n " ,
__func__ , sw_desc , & sw_desc - > async_tx ) ;
spin_lock_bh ( & mv_chan - > lock ) ;
2012-03-07 02:34:46 +04:00
cookie = dma_cookie_assign ( tx ) ;
2008-07-08 22:58:36 +04:00
if ( list_empty ( & mv_chan - > chain ) )
2014-08-27 17:52:51 +04:00
list_add_tail ( & sw_desc - > chain_node , & mv_chan - > chain ) ;
2008-07-08 22:58:36 +04:00
else {
new_hw_chain = 0 ;
old_chain_tail = list_entry ( mv_chan - > chain . prev ,
struct mv_xor_desc_slot ,
chain_node ) ;
2014-08-27 17:52:51 +04:00
list_add_tail ( & sw_desc - > chain_node , & mv_chan - > chain ) ;
2008-07-08 22:58:36 +04:00
2014-02-04 05:13:23 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " Append to last desc %pa \n " ,
& old_chain_tail - > async_tx . phys ) ;
2008-07-08 22:58:36 +04:00
/* fix up the hardware chain */
2014-08-27 17:52:51 +04:00
mv_desc_set_next_desc ( old_chain_tail , sw_desc - > async_tx . phys ) ;
2008-07-08 22:58:36 +04:00
/* if the channel is not busy */
if ( ! mv_chan_is_busy ( mv_chan ) ) {
u32 current_desc = mv_chan_get_current_desc ( mv_chan ) ;
/*
* and the curren desc is the end of the chain before
* the append , then we need to start the channel
*/
if ( current_desc = = old_chain_tail - > async_tx . phys )
new_hw_chain = 1 ;
}
}
if ( new_hw_chain )
2014-08-27 17:52:51 +04:00
mv_xor_start_new_chain ( mv_chan , sw_desc ) ;
2008-07-08 22:58:36 +04:00
spin_unlock_bh ( & mv_chan - > lock ) ;
return cookie ;
}
/* returns the number of allocated descriptors */
2009-01-06 21:38:17 +03:00
static int mv_xor_alloc_chan_resources ( struct dma_chan * chan )
2008-07-08 22:58:36 +04:00
{
2014-02-04 05:13:23 +04:00
void * virt_desc ;
dma_addr_t dma_desc ;
2008-07-08 22:58:36 +04:00
int idx ;
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( chan ) ;
struct mv_xor_desc_slot * slot = NULL ;
2012-11-15 18:55:30 +04:00
int num_descs_in_pool = MV_XOR_POOL_SIZE / MV_XOR_SLOT_SIZE ;
2008-07-08 22:58:36 +04:00
/* Allocate descriptor slots */
idx = mv_chan - > slots_allocated ;
while ( idx < num_descs_in_pool ) {
slot = kzalloc ( sizeof ( * slot ) , GFP_KERNEL ) ;
if ( ! slot ) {
2014-08-27 17:52:49 +04:00
dev_info ( mv_chan_to_devp ( mv_chan ) ,
" channel only initialized %d descriptor slots " ,
idx ) ;
2008-07-08 22:58:36 +04:00
break ;
}
2014-02-04 05:13:23 +04:00
virt_desc = mv_chan - > dma_desc_pool_virt ;
slot - > hw_desc = virt_desc + idx * MV_XOR_SLOT_SIZE ;
2008-07-08 22:58:36 +04:00
dma_async_tx_descriptor_init ( & slot - > async_tx , chan ) ;
slot - > async_tx . tx_submit = mv_xor_tx_submit ;
INIT_LIST_HEAD ( & slot - > chain_node ) ;
INIT_LIST_HEAD ( & slot - > slot_node ) ;
2014-02-04 05:13:23 +04:00
dma_desc = mv_chan - > dma_desc_pool ;
slot - > async_tx . phys = dma_desc + idx * MV_XOR_SLOT_SIZE ;
2008-07-08 22:58:36 +04:00
slot - > idx = idx + + ;
spin_lock_bh ( & mv_chan - > lock ) ;
mv_chan - > slots_allocated = idx ;
list_add_tail ( & slot - > slot_node , & mv_chan - > all_slots ) ;
spin_unlock_bh ( & mv_chan - > lock ) ;
}
if ( mv_chan - > slots_allocated & & ! mv_chan - > last_used )
mv_chan - > last_used = list_entry ( mv_chan - > all_slots . next ,
struct mv_xor_desc_slot ,
slot_node ) ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) ,
2008-07-08 22:58:36 +04:00
" allocated %d descriptor slots last_used: %p \n " ,
mv_chan - > slots_allocated , mv_chan - > last_used ) ;
return mv_chan - > slots_allocated ? : - ENOMEM ;
}
static struct dma_async_tx_descriptor *
mv_xor_prep_dma_xor ( struct dma_chan * chan , dma_addr_t dest , dma_addr_t * src ,
unsigned int src_cnt , size_t len , unsigned long flags )
{
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( chan ) ;
2014-08-27 17:52:51 +04:00
struct mv_xor_desc_slot * sw_desc ;
2008-07-08 22:58:36 +04:00
if ( unlikely ( len < MV_XOR_MIN_BYTE_COUNT ) )
return NULL ;
2011-03-26 20:26:53 +03:00
BUG_ON ( len > MV_XOR_MAX_BYTE_COUNT ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) ,
2014-02-04 05:13:23 +04:00
" %s src_cnt: %d len: %u dest %pad flags: %ld \n " ,
__func__ , src_cnt , len , & dest , flags ) ;
2008-07-08 22:58:36 +04:00
spin_lock_bh ( & mv_chan - > lock ) ;
2014-08-27 17:52:51 +04:00
sw_desc = mv_xor_alloc_slot ( mv_chan ) ;
2008-07-08 22:58:36 +04:00
if ( sw_desc ) {
sw_desc - > type = DMA_XOR ;
sw_desc - > async_tx . flags = flags ;
2014-08-27 17:52:53 +04:00
mv_desc_init ( sw_desc , dest , len , flags ) ;
2008-07-08 22:58:36 +04:00
sw_desc - > unmap_src_cnt = src_cnt ;
sw_desc - > unmap_len = len ;
while ( src_cnt - - )
2014-08-27 17:52:51 +04:00
mv_desc_set_src_addr ( sw_desc , src_cnt , src [ src_cnt ] ) ;
2008-07-08 22:58:36 +04:00
}
spin_unlock_bh ( & mv_chan - > lock ) ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) ,
2008-07-08 22:58:36 +04:00
" %s sw_desc %p async_tx %p \n " ,
__func__ , sw_desc , & sw_desc - > async_tx ) ;
return sw_desc ? & sw_desc - > async_tx : NULL ;
}
2014-08-27 17:52:50 +04:00
static struct dma_async_tx_descriptor *
mv_xor_prep_dma_memcpy ( struct dma_chan * chan , dma_addr_t dest , dma_addr_t src ,
size_t len , unsigned long flags )
{
/*
* A MEMCPY operation is identical to an XOR operation with only
* a single source address .
*/
return mv_xor_prep_dma_xor ( chan , dest , & src , 1 , len , flags ) ;
}
2008-07-08 22:58:36 +04:00
static void mv_xor_free_chan_resources ( struct dma_chan * chan )
{
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( chan ) ;
struct mv_xor_desc_slot * iter , * _iter ;
int in_use_descs = 0 ;
mv_xor_slot_cleanup ( mv_chan ) ;
spin_lock_bh ( & mv_chan - > lock ) ;
list_for_each_entry_safe ( iter , _iter , & mv_chan - > chain ,
chain_node ) {
in_use_descs + + ;
list_del ( & iter - > chain_node ) ;
}
list_for_each_entry_safe ( iter , _iter , & mv_chan - > completed_slots ,
completed_node ) {
in_use_descs + + ;
list_del ( & iter - > completed_node ) ;
}
list_for_each_entry_safe_reverse (
iter , _iter , & mv_chan - > all_slots , slot_node ) {
list_del ( & iter - > slot_node ) ;
kfree ( iter ) ;
mv_chan - > slots_allocated - - ;
}
mv_chan - > last_used = NULL ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( mv_chan ) , " %s slots_allocated %d \n " ,
2008-07-08 22:58:36 +04:00
__func__ , mv_chan - > slots_allocated ) ;
spin_unlock_bh ( & mv_chan - > lock ) ;
if ( in_use_descs )
2012-11-15 17:17:18 +04:00
dev_err ( mv_chan_to_devp ( mv_chan ) ,
2008-07-08 22:58:36 +04:00
" freeing %d in use descriptors! \n " , in_use_descs ) ;
}
/**
2010-03-27 02:50:49 +03:00
* mv_xor_status - poll the status of an XOR transaction
2008-07-08 22:58:36 +04:00
* @ chan : XOR channel handle
* @ cookie : XOR transaction identifier
2010-03-27 02:50:49 +03:00
* @ txstate : XOR transactions state holder ( or NULL )
2008-07-08 22:58:36 +04:00
*/
2010-03-27 02:50:49 +03:00
static enum dma_status mv_xor_status ( struct dma_chan * chan ,
2008-07-08 22:58:36 +04:00
dma_cookie_t cookie ,
2010-03-27 02:50:49 +03:00
struct dma_tx_state * txstate )
2008-07-08 22:58:36 +04:00
{
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( chan ) ;
enum dma_status ret ;
2012-03-07 02:35:27 +04:00
ret = dma_cookie_status ( chan , cookie , txstate ) ;
2013-10-16 19:21:04 +04:00
if ( ret = = DMA_COMPLETE ) {
2008-07-08 22:58:36 +04:00
mv_xor_clean_completed_slots ( mv_chan ) ;
return ret ;
}
mv_xor_slot_cleanup ( mv_chan ) ;
2012-03-07 02:35:27 +04:00
return dma_cookie_status ( chan , cookie , txstate ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_dump_xor_regs ( struct mv_xor_chan * chan )
{
u32 val ;
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_CONFIG ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " config 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_ACTIVATION ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " activation 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_INTR_CAUSE ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " intr cause 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_INTR_MASK ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " intr mask 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_ERROR_CAUSE ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " error cause 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
2013-07-29 19:42:13 +04:00
val = readl_relaxed ( XOR_ERROR_ADDR ( chan ) ) ;
2012-10-28 12:05:44 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " error addr 0x%08x \n " , val ) ;
2008-07-08 22:58:36 +04:00
}
static void mv_xor_err_interrupt_handler ( struct mv_xor_chan * chan ,
u32 intr_cause )
{
2014-08-27 17:52:52 +04:00
if ( intr_cause & XOR_INT_ERR_DECODE ) {
dev_dbg ( mv_chan_to_devp ( chan ) , " ignoring address decode error \n " ) ;
return ;
2008-07-08 22:58:36 +04:00
}
2014-08-27 17:52:52 +04:00
dev_err ( mv_chan_to_devp ( chan ) , " error on chan %d. intr cause 0x%08x \n " ,
2012-11-15 15:50:27 +04:00
chan - > idx , intr_cause ) ;
2008-07-08 22:58:36 +04:00
mv_dump_xor_regs ( chan ) ;
2014-08-27 17:52:52 +04:00
WARN_ON ( 1 ) ;
2008-07-08 22:58:36 +04:00
}
static irqreturn_t mv_xor_interrupt_handler ( int irq , void * data )
{
struct mv_xor_chan * chan = data ;
u32 intr_cause = mv_chan_get_intr_cause ( chan ) ;
2012-11-15 17:17:18 +04:00
dev_dbg ( mv_chan_to_devp ( chan ) , " intr cause %x \n " , intr_cause ) ;
2008-07-08 22:58:36 +04:00
2014-08-27 17:52:52 +04:00
if ( intr_cause & XOR_INTR_ERRORS )
2008-07-08 22:58:36 +04:00
mv_xor_err_interrupt_handler ( chan , intr_cause ) ;
tasklet_schedule ( & chan - > irq_tasklet ) ;
mv_xor_device_clear_eoc_cause ( chan ) ;
return IRQ_HANDLED ;
}
static void mv_xor_issue_pending ( struct dma_chan * chan )
{
struct mv_xor_chan * mv_chan = to_mv_xor_chan ( chan ) ;
if ( mv_chan - > pending > = MV_XOR_THRESHOLD ) {
mv_chan - > pending = 0 ;
mv_chan_activate ( mv_chan ) ;
}
}
/*
* Perform a transaction to verify the HW works .
*/
2012-12-15 02:54:26 +04:00
static int mv_xor_memcpy_self_test ( struct mv_xor_chan * mv_chan )
2008-07-08 22:58:36 +04:00
{
int i ;
void * src , * dest ;
dma_addr_t src_dma , dest_dma ;
struct dma_chan * dma_chan ;
dma_cookie_t cookie ;
struct dma_async_tx_descriptor * tx ;
2013-12-10 16:32:36 +04:00
struct dmaengine_unmap_data * unmap ;
2008-07-08 22:58:36 +04:00
int err = 0 ;
2013-12-10 16:32:36 +04:00
src = kmalloc ( sizeof ( u8 ) * PAGE_SIZE , GFP_KERNEL ) ;
2008-07-08 22:58:36 +04:00
if ( ! src )
return - ENOMEM ;
2013-12-10 16:32:36 +04:00
dest = kzalloc ( sizeof ( u8 ) * PAGE_SIZE , GFP_KERNEL ) ;
2008-07-08 22:58:36 +04:00
if ( ! dest ) {
kfree ( src ) ;
return - ENOMEM ;
}
/* Fill in src buffer */
2013-12-10 16:32:36 +04:00
for ( i = 0 ; i < PAGE_SIZE ; i + + )
2008-07-08 22:58:36 +04:00
( ( u8 * ) src ) [ i ] = ( u8 ) i ;
2012-11-15 18:09:42 +04:00
dma_chan = & mv_chan - > dmachan ;
2009-01-06 21:38:17 +03:00
if ( mv_xor_alloc_chan_resources ( dma_chan ) < 1 ) {
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto out ;
}
2013-12-10 16:32:36 +04:00
unmap = dmaengine_get_unmap_data ( dma_chan - > device - > dev , 2 , GFP_KERNEL ) ;
if ( ! unmap ) {
err = - ENOMEM ;
goto free_resources ;
}
src_dma = dma_map_page ( dma_chan - > device - > dev , virt_to_page ( src ) , 0 ,
PAGE_SIZE , DMA_TO_DEVICE ) ;
unmap - > to_cnt = 1 ;
unmap - > addr [ 0 ] = src_dma ;
2008-07-08 22:58:36 +04:00
2013-12-10 16:32:36 +04:00
dest_dma = dma_map_page ( dma_chan - > device - > dev , virt_to_page ( dest ) , 0 ,
PAGE_SIZE , DMA_FROM_DEVICE ) ;
unmap - > from_cnt = 1 ;
unmap - > addr [ 1 ] = dest_dma ;
unmap - > len = PAGE_SIZE ;
2008-07-08 22:58:36 +04:00
tx = mv_xor_prep_dma_memcpy ( dma_chan , dest_dma , src_dma ,
2013-12-10 16:32:36 +04:00
PAGE_SIZE , 0 ) ;
2008-07-08 22:58:36 +04:00
cookie = mv_xor_tx_submit ( tx ) ;
mv_xor_issue_pending ( dma_chan ) ;
async_tx_ack ( tx ) ;
msleep ( 1 ) ;
2010-03-27 02:50:49 +03:00
if ( mv_xor_status ( dma_chan , cookie , NULL ) ! =
2013-10-16 19:21:04 +04:00
DMA_COMPLETE ) {
2012-11-15 15:50:27 +04:00
dev_err ( dma_chan - > device - > dev ,
" Self-test copy timed out, disabling \n " ) ;
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto free_resources ;
}
2012-11-15 16:01:59 +04:00
dma_sync_single_for_cpu ( dma_chan - > device - > dev , dest_dma ,
2013-12-10 16:32:36 +04:00
PAGE_SIZE , DMA_FROM_DEVICE ) ;
if ( memcmp ( src , dest , PAGE_SIZE ) ) {
2012-11-15 15:50:27 +04:00
dev_err ( dma_chan - > device - > dev ,
" Self-test copy failed compare, disabling \n " ) ;
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto free_resources ;
}
free_resources :
2013-12-10 16:32:36 +04:00
dmaengine_unmap_put ( unmap ) ;
2008-07-08 22:58:36 +04:00
mv_xor_free_chan_resources ( dma_chan ) ;
out :
kfree ( src ) ;
kfree ( dest ) ;
return err ;
}
# define MV_XOR_NUM_SRC_TEST 4 /* must be <= 15 */
2012-11-19 22:22:55 +04:00
static int
2012-11-15 18:09:42 +04:00
mv_xor_xor_self_test ( struct mv_xor_chan * mv_chan )
2008-07-08 22:58:36 +04:00
{
int i , src_idx ;
struct page * dest ;
struct page * xor_srcs [ MV_XOR_NUM_SRC_TEST ] ;
dma_addr_t dma_srcs [ MV_XOR_NUM_SRC_TEST ] ;
dma_addr_t dest_dma ;
struct dma_async_tx_descriptor * tx ;
2013-12-10 16:32:36 +04:00
struct dmaengine_unmap_data * unmap ;
2008-07-08 22:58:36 +04:00
struct dma_chan * dma_chan ;
dma_cookie_t cookie ;
u8 cmp_byte = 0 ;
u32 cmp_word ;
int err = 0 ;
2013-12-10 16:32:36 +04:00
int src_count = MV_XOR_NUM_SRC_TEST ;
2008-07-08 22:58:36 +04:00
2013-12-10 16:32:36 +04:00
for ( src_idx = 0 ; src_idx < src_count ; src_idx + + ) {
2008-07-08 22:58:36 +04:00
xor_srcs [ src_idx ] = alloc_page ( GFP_KERNEL ) ;
2009-02-25 15:56:21 +03:00
if ( ! xor_srcs [ src_idx ] ) {
while ( src_idx - - )
2008-07-08 22:58:36 +04:00
__free_page ( xor_srcs [ src_idx ] ) ;
2009-02-25 15:56:21 +03:00
return - ENOMEM ;
}
2008-07-08 22:58:36 +04:00
}
dest = alloc_page ( GFP_KERNEL ) ;
2009-02-25 15:56:21 +03:00
if ( ! dest ) {
while ( src_idx - - )
2008-07-08 22:58:36 +04:00
__free_page ( xor_srcs [ src_idx ] ) ;
2009-02-25 15:56:21 +03:00
return - ENOMEM ;
}
2008-07-08 22:58:36 +04:00
/* Fill in src buffers */
2013-12-10 16:32:36 +04:00
for ( src_idx = 0 ; src_idx < src_count ; src_idx + + ) {
2008-07-08 22:58:36 +04:00
u8 * ptr = page_address ( xor_srcs [ src_idx ] ) ;
for ( i = 0 ; i < PAGE_SIZE ; i + + )
ptr [ i ] = ( 1 < < src_idx ) ;
}
2013-12-10 16:32:36 +04:00
for ( src_idx = 0 ; src_idx < src_count ; src_idx + + )
2008-07-08 22:58:36 +04:00
cmp_byte ^ = ( u8 ) ( 1 < < src_idx ) ;
cmp_word = ( cmp_byte < < 24 ) | ( cmp_byte < < 16 ) |
( cmp_byte < < 8 ) | cmp_byte ;
memset ( page_address ( dest ) , 0 , PAGE_SIZE ) ;
2012-11-15 18:09:42 +04:00
dma_chan = & mv_chan - > dmachan ;
2009-01-06 21:38:17 +03:00
if ( mv_xor_alloc_chan_resources ( dma_chan ) < 1 ) {
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto out ;
}
2013-12-10 16:32:36 +04:00
unmap = dmaengine_get_unmap_data ( dma_chan - > device - > dev , src_count + 1 ,
GFP_KERNEL ) ;
if ( ! unmap ) {
err = - ENOMEM ;
goto free_resources ;
}
2008-07-08 22:58:36 +04:00
/* test xor */
2013-12-10 16:32:36 +04:00
for ( i = 0 ; i < src_count ; i + + ) {
unmap - > addr [ i ] = dma_map_page ( dma_chan - > device - > dev , xor_srcs [ i ] ,
0 , PAGE_SIZE , DMA_TO_DEVICE ) ;
dma_srcs [ i ] = unmap - > addr [ i ] ;
unmap - > to_cnt + + ;
}
2008-07-08 22:58:36 +04:00
2013-12-10 16:32:36 +04:00
unmap - > addr [ src_count ] = dma_map_page ( dma_chan - > device - > dev , dest , 0 , PAGE_SIZE ,
DMA_FROM_DEVICE ) ;
dest_dma = unmap - > addr [ src_count ] ;
unmap - > from_cnt = 1 ;
unmap - > len = PAGE_SIZE ;
2008-07-08 22:58:36 +04:00
tx = mv_xor_prep_dma_xor ( dma_chan , dest_dma , dma_srcs ,
2013-12-10 16:32:36 +04:00
src_count , PAGE_SIZE , 0 ) ;
2008-07-08 22:58:36 +04:00
cookie = mv_xor_tx_submit ( tx ) ;
mv_xor_issue_pending ( dma_chan ) ;
async_tx_ack ( tx ) ;
msleep ( 8 ) ;
2010-03-27 02:50:49 +03:00
if ( mv_xor_status ( dma_chan , cookie , NULL ) ! =
2013-10-16 19:21:04 +04:00
DMA_COMPLETE ) {
2012-11-15 15:50:27 +04:00
dev_err ( dma_chan - > device - > dev ,
" Self-test xor timed out, disabling \n " ) ;
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto free_resources ;
}
2012-11-15 16:01:59 +04:00
dma_sync_single_for_cpu ( dma_chan - > device - > dev , dest_dma ,
2008-07-08 22:58:36 +04:00
PAGE_SIZE , DMA_FROM_DEVICE ) ;
for ( i = 0 ; i < ( PAGE_SIZE / sizeof ( u32 ) ) ; i + + ) {
u32 * ptr = page_address ( dest ) ;
if ( ptr [ i ] ! = cmp_word ) {
2012-11-15 15:50:27 +04:00
dev_err ( dma_chan - > device - > dev ,
2012-10-28 12:05:44 +04:00
" Self-test xor failed compare, disabling. index %d, data %x, expected %x \n " ,
i , ptr [ i ] , cmp_word ) ;
2008-07-08 22:58:36 +04:00
err = - ENODEV ;
goto free_resources ;
}
}
free_resources :
2013-12-10 16:32:36 +04:00
dmaengine_unmap_put ( unmap ) ;
2008-07-08 22:58:36 +04:00
mv_xor_free_chan_resources ( dma_chan ) ;
out :
2013-12-10 16:32:36 +04:00
src_idx = src_count ;
2008-07-08 22:58:36 +04:00
while ( src_idx - - )
__free_page ( xor_srcs [ src_idx ] ) ;
__free_page ( dest ) ;
return err ;
}
2012-11-18 14:44:56 +04:00
/* This driver does not implement any of the optional DMA operations. */
static int
mv_xor_control ( struct dma_chan * chan , enum dma_ctrl_cmd cmd ,
unsigned long arg )
{
return - ENOSYS ;
}
2012-11-15 18:17:05 +04:00
static int mv_xor_channel_remove ( struct mv_xor_chan * mv_chan )
2008-07-08 22:58:36 +04:00
{
struct dma_chan * chan , * _chan ;
2012-11-15 18:17:05 +04:00
struct device * dev = mv_chan - > dmadev . dev ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:17:05 +04:00
dma_async_device_unregister ( & mv_chan - > dmadev ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:55:30 +04:00
dma_free_coherent ( dev , MV_XOR_POOL_SIZE ,
2012-11-15 18:17:05 +04:00
mv_chan - > dma_desc_pool_virt , mv_chan - > dma_desc_pool ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:17:05 +04:00
list_for_each_entry_safe ( chan , _chan , & mv_chan - > dmadev . channels ,
2012-10-29 19:45:46 +04:00
device_node ) {
2008-07-08 22:58:36 +04:00
list_del ( & chan - > device_node ) ;
}
2012-11-15 19:11:18 +04:00
free_irq ( mv_chan - > irq , mv_chan ) ;
2008-07-08 22:58:36 +04:00
return 0 ;
}
2012-11-15 18:17:05 +04:00
static struct mv_xor_chan *
2012-11-15 18:29:53 +04:00
mv_xor_channel_add ( struct mv_xor_device * xordev ,
2012-10-29 19:45:46 +04:00
struct platform_device * pdev ,
2012-11-15 18:55:30 +04:00
int idx , dma_cap_mask_t cap_mask , int irq )
2008-07-08 22:58:36 +04:00
{
int ret = 0 ;
struct mv_xor_chan * mv_chan ;
struct dma_device * dma_dev ;
2012-11-15 18:17:05 +04:00
mv_chan = devm_kzalloc ( & pdev - > dev , sizeof ( * mv_chan ) , GFP_KERNEL ) ;
2013-09-02 12:24:20 +04:00
if ( ! mv_chan )
return ERR_PTR ( - ENOMEM ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:36:37 +04:00
mv_chan - > idx = idx ;
2012-11-15 19:11:18 +04:00
mv_chan - > irq = irq ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:17:05 +04:00
dma_dev = & mv_chan - > dmadev ;
2008-07-08 22:58:36 +04:00
/* allocate coherent memory for hardware descriptors
* note : writecombine gives slightly better performance , but
* requires that we explicitly flush the writes
*/
2012-11-15 18:17:05 +04:00
mv_chan - > dma_desc_pool_virt =
2012-11-15 18:55:30 +04:00
dma_alloc_writecombine ( & pdev - > dev , MV_XOR_POOL_SIZE ,
2012-11-15 18:17:05 +04:00
& mv_chan - > dma_desc_pool , GFP_KERNEL ) ;
if ( ! mv_chan - > dma_desc_pool_virt )
2012-10-29 19:45:46 +04:00
return ERR_PTR ( - ENOMEM ) ;
2008-07-08 22:58:36 +04:00
/* discover transaction capabilites from the platform data */
2012-10-29 19:45:46 +04:00
dma_dev - > cap_mask = cap_mask ;
2008-07-08 22:58:36 +04:00
INIT_LIST_HEAD ( & dma_dev - > channels ) ;
/* set base routines */
dma_dev - > device_alloc_chan_resources = mv_xor_alloc_chan_resources ;
dma_dev - > device_free_chan_resources = mv_xor_free_chan_resources ;
2010-03-27 02:50:49 +03:00
dma_dev - > device_tx_status = mv_xor_status ;
2008-07-08 22:58:36 +04:00
dma_dev - > device_issue_pending = mv_xor_issue_pending ;
2012-11-18 14:44:56 +04:00
dma_dev - > device_control = mv_xor_control ;
2008-07-08 22:58:36 +04:00
dma_dev - > dev = & pdev - > dev ;
/* set prep routines based on capability */
if ( dma_has_cap ( DMA_MEMCPY , dma_dev - > cap_mask ) )
dma_dev - > device_prep_dma_memcpy = mv_xor_prep_dma_memcpy ;
if ( dma_has_cap ( DMA_XOR , dma_dev - > cap_mask ) ) {
2009-06-28 20:26:21 +04:00
dma_dev - > max_xor = 8 ;
2008-07-08 22:58:36 +04:00
dma_dev - > device_prep_dma_xor = mv_xor_prep_dma_xor ;
}
2012-11-15 18:29:53 +04:00
mv_chan - > mmr_base = xordev - > xor_base ;
2013-10-30 19:01:43 +04:00
mv_chan - > mmr_high_base = xordev - > xor_high_base ;
2008-07-08 22:58:36 +04:00
tasklet_init ( & mv_chan - > irq_tasklet , mv_xor_tasklet , ( unsigned long )
mv_chan ) ;
/* clear errors before enabling interrupts */
mv_xor_device_clear_err_status ( mv_chan ) ;
dma: mv_xor: use request_irq() instead of devm_request_irq()
Even through the usage of devm_*() functions is generally recommended
over their classic variants, in the case of devm_request_irq()
combined with irq_of_parse_and_map(), it doesn't work nicely.
We have the following scenario:
irq_of_parse_and_map(...)
devm_request_irq(...)
For some reason, the driver initialization fails at a later
point. Since irq_of_parse_and_map() is no device-managed, we do a:
irq_dispose_mapping(...)
Unfortunately, this doesn't work, because the free_irq() must be done
prior to calling irq_dispose_mapping(). But with the devm mechanism,
the automatic free_irq() would happen only after we get out of the
->probe() function.
So basically, we revert to using request_irq() with traditional error
handling, so that in case of error, free_irq() gets called before
irq_dispose_mapping().
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
2012-11-22 21:19:09 +04:00
ret = request_irq ( mv_chan - > irq , mv_xor_interrupt_handler ,
0 , dev_name ( & pdev - > dev ) , mv_chan ) ;
2008-07-08 22:58:36 +04:00
if ( ret )
goto err_free_dma ;
mv_chan_unmask_interrupts ( mv_chan ) ;
2014-08-27 17:52:50 +04:00
mv_set_mode ( mv_chan , DMA_XOR ) ;
2008-07-08 22:58:36 +04:00
spin_lock_init ( & mv_chan - > lock ) ;
INIT_LIST_HEAD ( & mv_chan - > chain ) ;
INIT_LIST_HEAD ( & mv_chan - > completed_slots ) ;
INIT_LIST_HEAD ( & mv_chan - > all_slots ) ;
2012-11-15 17:57:44 +04:00
mv_chan - > dmachan . device = dma_dev ;
dma_cookie_init ( & mv_chan - > dmachan ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 17:57:44 +04:00
list_add_tail ( & mv_chan - > dmachan . device_node , & dma_dev - > channels ) ;
2008-07-08 22:58:36 +04:00
if ( dma_has_cap ( DMA_MEMCPY , dma_dev - > cap_mask ) ) {
2012-11-15 18:09:42 +04:00
ret = mv_xor_memcpy_self_test ( mv_chan ) ;
2008-07-08 22:58:36 +04:00
dev_dbg ( & pdev - > dev , " memcpy self test returned %d \n " , ret ) ;
if ( ret )
dma: mv_xor: use request_irq() instead of devm_request_irq()
Even through the usage of devm_*() functions is generally recommended
over their classic variants, in the case of devm_request_irq()
combined with irq_of_parse_and_map(), it doesn't work nicely.
We have the following scenario:
irq_of_parse_and_map(...)
devm_request_irq(...)
For some reason, the driver initialization fails at a later
point. Since irq_of_parse_and_map() is no device-managed, we do a:
irq_dispose_mapping(...)
Unfortunately, this doesn't work, because the free_irq() must be done
prior to calling irq_dispose_mapping(). But with the devm mechanism,
the automatic free_irq() would happen only after we get out of the
->probe() function.
So basically, we revert to using request_irq() with traditional error
handling, so that in case of error, free_irq() gets called before
irq_dispose_mapping().
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
2012-11-22 21:19:09 +04:00
goto err_free_irq ;
2008-07-08 22:58:36 +04:00
}
if ( dma_has_cap ( DMA_XOR , dma_dev - > cap_mask ) ) {
2012-11-15 18:09:42 +04:00
ret = mv_xor_xor_self_test ( mv_chan ) ;
2008-07-08 22:58:36 +04:00
dev_dbg ( & pdev - > dev , " xor self test returned %d \n " , ret ) ;
if ( ret )
dma: mv_xor: use request_irq() instead of devm_request_irq()
Even through the usage of devm_*() functions is generally recommended
over their classic variants, in the case of devm_request_irq()
combined with irq_of_parse_and_map(), it doesn't work nicely.
We have the following scenario:
irq_of_parse_and_map(...)
devm_request_irq(...)
For some reason, the driver initialization fails at a later
point. Since irq_of_parse_and_map() is no device-managed, we do a:
irq_dispose_mapping(...)
Unfortunately, this doesn't work, because the free_irq() must be done
prior to calling irq_dispose_mapping(). But with the devm mechanism,
the automatic free_irq() would happen only after we get out of the
->probe() function.
So basically, we revert to using request_irq() with traditional error
handling, so that in case of error, free_irq() gets called before
irq_dispose_mapping().
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
2012-11-22 21:19:09 +04:00
goto err_free_irq ;
2008-07-08 22:58:36 +04:00
}
2013-07-04 02:05:06 +04:00
dev_info ( & pdev - > dev , " Marvell XOR: ( %s%s%s) \n " ,
2012-10-28 12:05:44 +04:00
dma_has_cap ( DMA_XOR , dma_dev - > cap_mask ) ? " xor " : " " ,
dma_has_cap ( DMA_MEMCPY , dma_dev - > cap_mask ) ? " cpy " : " " ,
dma_has_cap ( DMA_INTERRUPT , dma_dev - > cap_mask ) ? " intr " : " " ) ;
2008-07-08 22:58:36 +04:00
dma_async_device_register ( dma_dev ) ;
2012-11-15 18:17:05 +04:00
return mv_chan ;
2008-07-08 22:58:36 +04:00
dma: mv_xor: use request_irq() instead of devm_request_irq()
Even through the usage of devm_*() functions is generally recommended
over their classic variants, in the case of devm_request_irq()
combined with irq_of_parse_and_map(), it doesn't work nicely.
We have the following scenario:
irq_of_parse_and_map(...)
devm_request_irq(...)
For some reason, the driver initialization fails at a later
point. Since irq_of_parse_and_map() is no device-managed, we do a:
irq_dispose_mapping(...)
Unfortunately, this doesn't work, because the free_irq() must be done
prior to calling irq_dispose_mapping(). But with the devm mechanism,
the automatic free_irq() would happen only after we get out of the
->probe() function.
So basically, we revert to using request_irq() with traditional error
handling, so that in case of error, free_irq() gets called before
irq_dispose_mapping().
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
2012-11-22 21:19:09 +04:00
err_free_irq :
free_irq ( mv_chan - > irq , mv_chan ) ;
2008-07-08 22:58:36 +04:00
err_free_dma :
2012-11-15 18:55:30 +04:00
dma_free_coherent ( & pdev - > dev , MV_XOR_POOL_SIZE ,
2012-11-15 18:17:05 +04:00
mv_chan - > dma_desc_pool_virt , mv_chan - > dma_desc_pool ) ;
2012-10-29 19:45:46 +04:00
return ERR_PTR ( ret ) ;
2008-07-08 22:58:36 +04:00
}
static void
2012-11-15 18:29:53 +04:00
mv_xor_conf_mbus_windows ( struct mv_xor_device * xordev ,
2011-12-08 00:48:07 +04:00
const struct mbus_dram_target_info * dram )
2008-07-08 22:58:36 +04:00
{
2013-10-30 19:01:43 +04:00
void __iomem * base = xordev - > xor_high_base ;
2008-07-08 22:58:36 +04:00
u32 win_enable = 0 ;
int i ;
for ( i = 0 ; i < 8 ; i + + ) {
writel ( 0 , base + WINDOW_BASE ( i ) ) ;
writel ( 0 , base + WINDOW_SIZE ( i ) ) ;
if ( i < 4 )
writel ( 0 , base + WINDOW_REMAP_HIGH ( i ) ) ;
}
for ( i = 0 ; i < dram - > num_cs ; i + + ) {
2011-12-08 00:48:07 +04:00
const struct mbus_dram_window * cs = dram - > cs + i ;
2008-07-08 22:58:36 +04:00
writel ( ( cs - > base & 0xffff0000 ) |
( cs - > mbus_attr < < 8 ) |
dram - > mbus_dram_target_id , base + WINDOW_BASE ( i ) ) ;
writel ( ( cs - > size - 1 ) & 0xffff0000 , base + WINDOW_SIZE ( i ) ) ;
win_enable | = ( 1 < < i ) ;
win_enable | = 3 < < ( 16 + ( 2 * i ) ) ;
}
writel ( win_enable , base + WINDOW_BAR_ENABLE ( 0 ) ) ;
writel ( win_enable , base + WINDOW_BAR_ENABLE ( 1 ) ) ;
2012-11-22 21:16:37 +04:00
writel ( 0 , base + WINDOW_OVERRIDE_CTRL ( 0 ) ) ;
writel ( 0 , base + WINDOW_OVERRIDE_CTRL ( 1 ) ) ;
2008-07-08 22:58:36 +04:00
}
2012-12-15 02:54:26 +04:00
static int mv_xor_probe ( struct platform_device * pdev )
2008-07-08 22:58:36 +04:00
{
2011-12-08 00:48:07 +04:00
const struct mbus_dram_target_info * dram ;
2012-11-15 18:29:53 +04:00
struct mv_xor_device * xordev ;
2013-07-30 12:09:11 +04:00
struct mv_xor_platform_data * pdata = dev_get_platdata ( & pdev - > dev ) ;
2008-07-08 22:58:36 +04:00
struct resource * res ;
2012-10-29 19:54:49 +04:00
int i , ret ;
2008-07-08 22:58:36 +04:00
2012-10-28 12:05:44 +04:00
dev_notice ( & pdev - > dev , " Marvell shared XOR driver \n " ) ;
2008-07-08 22:58:36 +04:00
2012-11-15 18:29:53 +04:00
xordev = devm_kzalloc ( & pdev - > dev , sizeof ( * xordev ) , GFP_KERNEL ) ;
if ( ! xordev )
2008-07-08 22:58:36 +04:00
return - ENOMEM ;
res = platform_get_resource ( pdev , IORESOURCE_MEM , 0 ) ;
if ( ! res )
return - ENODEV ;
2012-11-15 18:29:53 +04:00
xordev - > xor_base = devm_ioremap ( & pdev - > dev , res - > start ,
resource_size ( res ) ) ;
if ( ! xordev - > xor_base )
2008-07-08 22:58:36 +04:00
return - EBUSY ;
res = platform_get_resource ( pdev , IORESOURCE_MEM , 1 ) ;
if ( ! res )
return - ENODEV ;
2012-11-15 18:29:53 +04:00
xordev - > xor_high_base = devm_ioremap ( & pdev - > dev , res - > start ,
resource_size ( res ) ) ;
if ( ! xordev - > xor_high_base )
2008-07-08 22:58:36 +04:00
return - EBUSY ;
2012-11-15 18:29:53 +04:00
platform_set_drvdata ( pdev , xordev ) ;
2008-07-08 22:58:36 +04:00
/*
* ( Re - ) program MBUS remapping windows if we are asked to .
*/
2011-12-08 00:48:07 +04:00
dram = mv_mbus_dram_info ( ) ;
if ( dram )
2012-11-15 18:29:53 +04:00
mv_xor_conf_mbus_windows ( xordev , dram ) ;
2008-07-08 22:58:36 +04:00
2012-02-19 16:30:26 +04:00
/* Not all platforms can gate the clock, so it is not
* an error if the clock does not exists .
*/
2012-11-15 18:29:53 +04:00
xordev - > clk = clk_get ( & pdev - > dev , NULL ) ;
if ( ! IS_ERR ( xordev - > clk ) )
clk_prepare_enable ( xordev - > clk ) ;
2012-02-19 16:30:26 +04:00
2012-11-15 19:47:58 +04:00
if ( pdev - > dev . of_node ) {
struct device_node * np ;
int i = 0 ;
for_each_child_of_node ( pdev - > dev . of_node , np ) {
2013-12-13 03:59:08 +04:00
struct mv_xor_chan * chan ;
2012-11-15 19:47:58 +04:00
dma_cap_mask_t cap_mask ;
int irq ;
dma_cap_zero ( cap_mask ) ;
if ( of_property_read_bool ( np , " dmacap,memcpy " ) )
dma_cap_set ( DMA_MEMCPY , cap_mask ) ;
if ( of_property_read_bool ( np , " dmacap,xor " ) )
dma_cap_set ( DMA_XOR , cap_mask ) ;
if ( of_property_read_bool ( np , " dmacap,interrupt " ) )
dma_cap_set ( DMA_INTERRUPT , cap_mask ) ;
irq = irq_of_parse_and_map ( np , 0 ) ;
2012-11-22 21:22:12 +04:00
if ( ! irq ) {
ret = - ENODEV ;
2012-11-15 19:47:58 +04:00
goto err_channel_add ;
}
2013-12-13 03:59:08 +04:00
chan = mv_xor_channel_add ( xordev , pdev , i ,
cap_mask , irq ) ;
if ( IS_ERR ( chan ) ) {
ret = PTR_ERR ( chan ) ;
2012-11-15 19:47:58 +04:00
irq_dispose_mapping ( irq ) ;
goto err_channel_add ;
}
2013-12-13 03:59:08 +04:00
xordev - > channels [ i ] = chan ;
2012-11-15 19:47:58 +04:00
i + + ;
}
} else if ( pdata & & pdata - > channels ) {
2012-10-29 19:54:49 +04:00
for ( i = 0 ; i < MV_XOR_MAX_CHANNELS ; i + + ) {
2012-10-30 14:56:26 +04:00
struct mv_xor_channel_data * cd ;
2013-12-13 03:59:08 +04:00
struct mv_xor_chan * chan ;
2012-10-29 19:54:49 +04:00
int irq ;
cd = & pdata - > channels [ i ] ;
if ( ! cd ) {
ret = - ENODEV ;
goto err_channel_add ;
}
irq = platform_get_irq ( pdev , i ) ;
if ( irq < 0 ) {
ret = irq ;
goto err_channel_add ;
}
2013-12-13 03:59:08 +04:00
chan = mv_xor_channel_add ( xordev , pdev , i ,
cd - > cap_mask , irq ) ;
if ( IS_ERR ( chan ) ) {
ret = PTR_ERR ( chan ) ;
2012-10-29 19:54:49 +04:00
goto err_channel_add ;
}
2013-12-13 03:59:08 +04:00
xordev - > channels [ i ] = chan ;
2012-10-29 19:54:49 +04:00
}
}
2012-02-19 16:30:26 +04:00
2008-07-08 22:58:36 +04:00
return 0 ;
2012-10-29 19:54:49 +04:00
err_channel_add :
for ( i = 0 ; i < MV_XOR_MAX_CHANNELS ; i + + )
2012-11-15 19:47:58 +04:00
if ( xordev - > channels [ i ] ) {
2013-01-06 14:10:43 +04:00
mv_xor_channel_remove ( xordev - > channels [ i ] ) ;
2012-11-15 19:47:58 +04:00
if ( pdev - > dev . of_node )
irq_dispose_mapping ( xordev - > channels [ i ] - > irq ) ;
}
2012-10-29 19:54:49 +04:00
2013-01-06 14:10:44 +04:00
if ( ! IS_ERR ( xordev - > clk ) ) {
clk_disable_unprepare ( xordev - > clk ) ;
clk_put ( xordev - > clk ) ;
}
2012-10-29 19:54:49 +04:00
return ret ;
2008-07-08 22:58:36 +04:00
}
2012-12-15 02:54:26 +04:00
static int mv_xor_remove ( struct platform_device * pdev )
2008-07-08 22:58:36 +04:00
{
2012-11-15 18:29:53 +04:00
struct mv_xor_device * xordev = platform_get_drvdata ( pdev ) ;
2012-10-29 19:54:49 +04:00
int i ;
for ( i = 0 ; i < MV_XOR_MAX_CHANNELS ; i + + ) {
2012-11-15 18:29:53 +04:00
if ( xordev - > channels [ i ] )
mv_xor_channel_remove ( xordev - > channels [ i ] ) ;
2012-10-29 19:54:49 +04:00
}
2012-02-19 16:30:26 +04:00
2012-11-15 18:29:53 +04:00
if ( ! IS_ERR ( xordev - > clk ) ) {
clk_disable_unprepare ( xordev - > clk ) ;
clk_put ( xordev - > clk ) ;
2012-02-19 16:30:26 +04:00
}
2008-07-08 22:58:36 +04:00
return 0 ;
}
2012-11-15 19:47:58 +04:00
# ifdef CONFIG_OF
2012-12-15 02:54:26 +04:00
static struct of_device_id mv_xor_dt_ids [ ] = {
2012-11-15 19:47:58 +04:00
{ . compatible = " marvell,orion-xor " , } ,
{ } ,
} ;
MODULE_DEVICE_TABLE ( of , mv_xor_dt_ids ) ;
# endif
2012-10-30 15:05:40 +04:00
static struct platform_driver mv_xor_driver = {
. probe = mv_xor_probe ,
2012-12-15 02:54:26 +04:00
. remove = mv_xor_remove ,
2008-07-08 22:58:36 +04:00
. driver = {
2012-11-15 19:47:58 +04:00
. owner = THIS_MODULE ,
. name = MV_XOR_NAME ,
. of_match_table = of_match_ptr ( mv_xor_dt_ids ) ,
2008-07-08 22:58:36 +04:00
} ,
} ;
static int __init mv_xor_init ( void )
{
2012-10-30 15:05:40 +04:00
return platform_driver_register ( & mv_xor_driver ) ;
2008-07-08 22:58:36 +04:00
}
module_init ( mv_xor_init ) ;
/* it's currently unsafe to unload this module */
#if 0
static void __exit mv_xor_exit ( void )
{
platform_driver_unregister ( & mv_xor_driver ) ;
return ;
}
module_exit ( mv_xor_exit ) ;
# endif
MODULE_AUTHOR ( " Saeed Bishara <saeed@marvell.com> " ) ;
MODULE_DESCRIPTION ( " DMA engine driver for Marvell's XOR engine " ) ;
MODULE_LICENSE ( " GPL " ) ;