2015-09-07 07:40:25 +03:00
/*
* vsp1_dl . h - - R - Car VSP1 Display List
*
* Copyright ( C ) 2015 Renesas Corporation
*
* Contact : Laurent Pinchart ( laurent . pinchart @ ideasonboard . com )
*
* This program is free software ; you can redistribute it and / or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation ; either version 2 of the License , or
* ( at your option ) any later version .
*/
# include <linux/device.h>
# include <linux/dma-mapping.h>
# include <linux/gfp.h>
# include <linux/slab.h>
2016-05-14 01:17:02 +03:00
# include <linux/workqueue.h>
2015-09-07 07:40:25 +03:00
# include "vsp1.h"
# include "vsp1_dl.h"
2016-03-03 19:36:34 +03:00
# define VSP1_DL_NUM_ENTRIES 256
2015-09-07 07:40:25 +03:00
2015-11-15 03:48:27 +03:00
# define VSP1_DLH_INT_ENABLE (1 << 1)
# define VSP1_DLH_AUTO_START (1 << 0)
2016-03-03 19:36:34 +03:00
struct vsp1_dl_header_list {
u32 num_bytes ;
u32 addr ;
} __attribute__ ( ( __packed__ ) ) ;
2015-11-15 03:48:27 +03:00
struct vsp1_dl_header {
u32 num_lists ;
2016-03-03 19:36:34 +03:00
struct vsp1_dl_header_list lists [ 8 ] ;
2015-11-15 03:48:27 +03:00
u32 next_header ;
u32 flags ;
} __attribute__ ( ( __packed__ ) ) ;
2015-09-07 07:40:25 +03:00
struct vsp1_dl_entry {
u32 addr ;
u32 data ;
} __attribute__ ( ( __packed__ ) ) ;
2016-03-03 19:36:34 +03:00
/**
* struct vsp1_dl_body - Display list body
* @ list : entry in the display list list of bodies
* @ vsp1 : the VSP1 device
* @ entries : array of entries
* @ dma : DMA address of the entries
* @ size : size of the DMA memory in bytes
* @ num_entries : number of stored entries
*/
struct vsp1_dl_body {
2015-11-09 01:06:57 +03:00
struct list_head list ;
2016-03-03 19:36:34 +03:00
struct vsp1_device * vsp1 ;
struct vsp1_dl_entry * entries ;
dma_addr_t dma ;
size_t size ;
unsigned int num_entries ;
} ;
2015-09-07 07:40:25 +03:00
2016-03-03 19:36:34 +03:00
/**
* struct vsp1_dl_list - Display list
* @ list : entry in the display list manager lists
* @ dlm : the display list manager
* @ header : display list header , NULL for headerless lists
* @ dma : DMA address for the header
* @ body0 : first display list body
* @ fragments : list of extra display list bodies
2016-07-12 19:49:46 +03:00
* @ chain : entry in the display list partition chain
2016-03-03 19:36:34 +03:00
*/
struct vsp1_dl_list {
struct list_head list ;
2015-11-09 01:06:57 +03:00
struct vsp1_dl_manager * dlm ;
2015-09-07 07:40:25 +03:00
2015-11-15 03:48:27 +03:00
struct vsp1_dl_header * header ;
2015-09-07 07:40:25 +03:00
dma_addr_t dma ;
2016-03-03 19:36:34 +03:00
struct vsp1_dl_body body0 ;
struct list_head fragments ;
2016-07-12 19:49:46 +03:00
bool has_chain ;
struct list_head chain ;
2015-09-07 07:40:25 +03:00
} ;
2015-11-15 03:48:27 +03:00
enum vsp1_dl_mode {
VSP1_DL_MODE_HEADER ,
VSP1_DL_MODE_HEADERLESS ,
} ;
2015-11-15 03:27:52 +03:00
/**
* struct vsp1_dl_manager - Display List manager
2015-11-15 03:48:27 +03:00
* @ index : index of the related WPF
* @ mode : display list operation mode ( header or headerless )
2015-11-15 03:27:52 +03:00
* @ vsp1 : the VSP1 device
2016-05-14 01:17:02 +03:00
* @ lock : protects the free , active , queued , pending and gc_fragments lists
2015-11-15 03:27:52 +03:00
* @ free : array of all free display lists
* @ active : list currently being processed ( loaded ) by hardware
* @ queued : list queued to the hardware ( written to the DL registers )
* @ pending : list waiting to be queued to the hardware
2016-05-14 01:17:02 +03:00
* @ gc_work : fragments garbage collector work struct
* @ gc_fragments : array of display list fragments waiting to be freed
2015-11-15 03:27:52 +03:00
*/
struct vsp1_dl_manager {
2015-11-15 03:48:27 +03:00
unsigned int index ;
enum vsp1_dl_mode mode ;
2015-11-15 03:27:52 +03:00
struct vsp1_device * vsp1 ;
spinlock_t lock ;
struct list_head free ;
struct vsp1_dl_list * active ;
struct vsp1_dl_list * queued ;
struct vsp1_dl_list * pending ;
2016-05-14 01:17:02 +03:00
struct work_struct gc_work ;
struct list_head gc_fragments ;
2015-11-15 03:27:52 +03:00
} ;
2016-03-03 19:36:34 +03:00
/* -----------------------------------------------------------------------------
* Display List Body Management
*/
/*
* Initialize a display list body object and allocate DMA memory for the body
* data . The display list body object is expected to have been initialized to
* 0 when allocated .
*/
static int vsp1_dl_body_init ( struct vsp1_device * vsp1 ,
struct vsp1_dl_body * dlb , unsigned int num_entries ,
size_t extra_size )
{
size_t size = num_entries * sizeof ( * dlb - > entries ) + extra_size ;
dlb - > vsp1 = vsp1 ;
dlb - > size = size ;
dlb - > entries = dma_alloc_wc ( vsp1 - > dev , dlb - > size , & dlb - > dma ,
GFP_KERNEL ) ;
if ( ! dlb - > entries )
return - ENOMEM ;
return 0 ;
}
/*
* Cleanup a display list body and free allocated DMA memory allocated .
*/
static void vsp1_dl_body_cleanup ( struct vsp1_dl_body * dlb )
{
dma_free_wc ( dlb - > vsp1 - > dev , dlb - > size , dlb - > entries , dlb - > dma ) ;
}
/**
* vsp1_dl_fragment_alloc - Allocate a display list fragment
* @ vsp1 : The VSP1 device
* @ num_entries : The maximum number of entries that the fragment can contain
*
* Allocate a display list fragment with enough memory to contain the requested
* number of entries .
*
* Return a pointer to a fragment on success or NULL if memory can ' t be
* allocated .
*/
struct vsp1_dl_body * vsp1_dl_fragment_alloc ( struct vsp1_device * vsp1 ,
unsigned int num_entries )
{
struct vsp1_dl_body * dlb ;
int ret ;
dlb = kzalloc ( sizeof ( * dlb ) , GFP_KERNEL ) ;
if ( ! dlb )
return NULL ;
ret = vsp1_dl_body_init ( vsp1 , dlb , num_entries , 0 ) ;
if ( ret < 0 ) {
kfree ( dlb ) ;
return NULL ;
}
return dlb ;
}
/**
* vsp1_dl_fragment_free - Free a display list fragment
* @ dlb : The fragment
*
* Free the given display list fragment and the associated DMA memory .
*
* Fragments must only be freed explicitly if they are not added to a display
* list , as the display list will take ownership of them and free them
* otherwise . Manual free typically happens at cleanup time for fragments that
* have been allocated but not used .
*
* Passing a NULL pointer to this function is safe , in that case no operation
* will be performed .
*/
void vsp1_dl_fragment_free ( struct vsp1_dl_body * dlb )
{
if ( ! dlb )
return ;
vsp1_dl_body_cleanup ( dlb ) ;
kfree ( dlb ) ;
}
/**
* vsp1_dl_fragment_write - Write a register to a display list fragment
* @ dlb : The fragment
* @ reg : The register address
* @ data : The register value
*
* Write the given register and value to the display list fragment . The maximum
* number of entries that can be written in a fragment is specified when the
* fragment is allocated by vsp1_dl_fragment_alloc ( ) .
*/
void vsp1_dl_fragment_write ( struct vsp1_dl_body * dlb , u32 reg , u32 data )
{
dlb - > entries [ dlb - > num_entries ] . addr = reg ;
dlb - > entries [ dlb - > num_entries ] . data = data ;
dlb - > num_entries + + ;
}
2015-09-07 07:40:25 +03:00
/* -----------------------------------------------------------------------------
* Display List Transaction Management
*/
2015-11-09 01:06:57 +03:00
static struct vsp1_dl_list * vsp1_dl_list_alloc ( struct vsp1_dl_manager * dlm )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
struct vsp1_dl_list * dl ;
2015-11-15 03:48:27 +03:00
size_t header_size ;
2016-03-03 19:36:34 +03:00
int ret ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
dl = kzalloc ( sizeof ( * dl ) , GFP_KERNEL ) ;
if ( ! dl )
return NULL ;
2015-09-07 07:40:25 +03:00
2016-03-03 19:36:34 +03:00
INIT_LIST_HEAD ( & dl - > fragments ) ;
2015-11-09 01:06:57 +03:00
dl - > dlm = dlm ;
2016-03-03 19:36:34 +03:00
/* Initialize the display list body and allocate DMA memory for the body
* and the optional header . Both are allocated together to avoid memory
* fragmentation , with the header located right after the body in
* memory .
*/
header_size = dlm - > mode = = VSP1_DL_MODE_HEADER
? ALIGN ( sizeof ( struct vsp1_dl_header ) , 8 )
: 0 ;
ret = vsp1_dl_body_init ( dlm - > vsp1 , & dl - > body0 , VSP1_DL_NUM_ENTRIES ,
header_size ) ;
if ( ret < 0 ) {
2015-11-09 01:06:57 +03:00
kfree ( dl ) ;
return NULL ;
}
2015-09-07 07:40:25 +03:00
2015-11-15 03:48:27 +03:00
if ( dlm - > mode = = VSP1_DL_MODE_HEADER ) {
2016-03-03 19:36:34 +03:00
size_t header_offset = VSP1_DL_NUM_ENTRIES
* sizeof ( * dl - > body0 . entries ) ;
dl - > header = ( ( void * ) dl - > body0 . entries ) + header_offset ;
dl - > dma = dl - > body0 . dma + header_offset ;
2015-11-15 03:48:27 +03:00
memset ( dl - > header , 0 , sizeof ( * dl - > header ) ) ;
2016-03-03 19:36:34 +03:00
dl - > header - > lists [ 0 ] . addr = dl - > body0 . dma ;
2015-11-15 03:48:27 +03:00
}
2015-11-09 01:06:57 +03:00
return dl ;
}
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
static void vsp1_dl_list_free ( struct vsp1_dl_list * dl )
{
2016-03-03 19:36:34 +03:00
vsp1_dl_body_cleanup ( & dl - > body0 ) ;
2016-05-14 01:17:02 +03:00
list_splice_init ( & dl - > fragments , & dl - > dlm - > gc_fragments ) ;
2015-11-09 01:06:57 +03:00
kfree ( dl ) ;
2015-09-07 07:40:25 +03:00
}
2015-11-09 01:06:57 +03:00
/**
* vsp1_dl_list_get - Get a free display list
* @ dlm : The display list manager
*
* Get a display list from the pool of free lists and return it .
*
* This function must be called without the display list manager lock held .
*/
struct vsp1_dl_list * vsp1_dl_list_get ( struct vsp1_dl_manager * dlm )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
struct vsp1_dl_list * dl = NULL ;
2015-09-07 07:40:25 +03:00
unsigned long flags ;
2015-11-09 01:06:57 +03:00
spin_lock_irqsave ( & dlm - > lock , flags ) ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
if ( ! list_empty ( & dlm - > free ) ) {
dl = list_first_entry ( & dlm - > free , struct vsp1_dl_list , list ) ;
list_del ( & dl - > list ) ;
2016-07-12 19:49:46 +03:00
2016-09-19 21:18:01 +03:00
/*
* The display list chain must be initialised to ensure every
2016-07-12 19:49:46 +03:00
* display list can assert list_empty ( ) if it is not in a chain .
*/
INIT_LIST_HEAD ( & dl - > chain ) ;
2015-09-07 07:40:25 +03:00
}
2015-11-09 01:06:57 +03:00
spin_unlock_irqrestore ( & dlm - > lock , flags ) ;
return dl ;
}
2015-09-07 07:40:25 +03:00
2016-03-03 15:26:47 +03:00
/* This function must be called with the display list manager lock held.*/
static void __vsp1_dl_list_put ( struct vsp1_dl_list * dl )
{
2016-07-12 19:49:46 +03:00
struct vsp1_dl_list * dl_child ;
2016-03-03 15:26:47 +03:00
if ( ! dl )
return ;
2016-09-19 21:18:01 +03:00
/*
* Release any linked display - lists which were chained for a single
2016-07-12 19:49:46 +03:00
* hardware operation .
*/
if ( dl - > has_chain ) {
list_for_each_entry ( dl_child , & dl - > chain , chain )
__vsp1_dl_list_put ( dl_child ) ;
}
dl - > has_chain = false ;
2016-09-19 21:18:01 +03:00
/*
* We can ' t free fragments here as DMA memory can only be freed in
2016-05-14 01:17:02 +03:00
* interruptible context . Move all fragments to the display list
* manager ' s list of fragments to be freed , they will be
* garbage - collected by the work queue .
*/
if ( ! list_empty ( & dl - > fragments ) ) {
list_splice_init ( & dl - > fragments , & dl - > dlm - > gc_fragments ) ;
schedule_work ( & dl - > dlm - > gc_work ) ;
}
2016-03-03 19:36:34 +03:00
dl - > body0 . num_entries = 0 ;
2016-03-03 15:26:47 +03:00
list_add_tail ( & dl - > list , & dl - > dlm - > free ) ;
}
2015-11-09 01:06:57 +03:00
/**
* vsp1_dl_list_put - Release a display list
* @ dl : The display list
*
* Release the display list and return it to the pool of free lists .
*
* Passing a NULL pointer to this function is safe , in that case no operation
* will be performed .
*/
void vsp1_dl_list_put ( struct vsp1_dl_list * dl )
{
2016-03-03 15:26:47 +03:00
unsigned long flags ;
2015-11-09 01:06:57 +03:00
if ( ! dl )
return ;
2015-09-07 07:40:25 +03:00
2016-03-03 15:26:47 +03:00
spin_lock_irqsave ( & dl - > dlm - > lock , flags ) ;
__vsp1_dl_list_put ( dl ) ;
spin_unlock_irqrestore ( & dl - > dlm - > lock , flags ) ;
2015-09-07 07:40:25 +03:00
}
2016-03-03 19:36:34 +03:00
/**
* vsp1_dl_list_write - Write a register to the display list
* @ dl : The display list
* @ reg : The register address
* @ data : The register value
*
* Write the given register and value to the display list . Up to 256 registers
* can be written per display list .
*/
2015-11-09 01:06:57 +03:00
void vsp1_dl_list_write ( struct vsp1_dl_list * dl , u32 reg , u32 data )
2015-09-07 07:40:25 +03:00
{
2016-03-03 19:36:34 +03:00
vsp1_dl_fragment_write ( & dl - > body0 , reg , data ) ;
}
/**
* vsp1_dl_list_add_fragment - Add a fragment to the display list
* @ dl : The display list
* @ dlb : The fragment
*
* Add a display list body as a fragment to a display list . Registers contained
* in fragments are processed after registers contained in the main display
* list , in the order in which fragments are added .
*
* Adding a fragment to a display list passes ownership of the fragment to the
* list . The caller must not touch the fragment after this call , and must not
* free it explicitly with vsp1_dl_fragment_free ( ) .
*
* Fragments are only usable for display lists in header mode . Attempt to
* add a fragment to a header - less display list will return an error .
*/
int vsp1_dl_list_add_fragment ( struct vsp1_dl_list * dl ,
struct vsp1_dl_body * dlb )
{
/* Multi-body lists are only available in header mode. */
if ( dl - > dlm - > mode ! = VSP1_DL_MODE_HEADER )
return - EINVAL ;
list_add_tail ( & dlb - > list , & dl - > fragments ) ;
return 0 ;
2015-09-07 07:40:25 +03:00
}
2016-07-12 19:49:46 +03:00
/**
* vsp1_dl_list_add_chain - Add a display list to a chain
* @ head : The head display list
* @ dl : The new display list
*
* Add a display list to an existing display list chain . The chained lists
* will be automatically processed by the hardware without intervention from
* the CPU . A display list end interrupt will only complete after the last
* display list in the chain has completed processing .
*
* Adding a display list to a chain passes ownership of the display list to
* the head display list item . The chain is released when the head dl item is
* put back with __vsp1_dl_list_put ( ) .
*
* Chained display lists are only usable in header mode . Attempts to add a
* display list to a chain in header - less mode will return an error .
*/
int vsp1_dl_list_add_chain ( struct vsp1_dl_list * head ,
struct vsp1_dl_list * dl )
{
/* Chained lists are only available in header mode. */
if ( head - > dlm - > mode ! = VSP1_DL_MODE_HEADER )
return - EINVAL ;
head - > has_chain = true ;
list_add_tail ( & dl - > chain , & head - > chain ) ;
return 0 ;
}
static void vsp1_dl_list_fill_header ( struct vsp1_dl_list * dl , bool is_last )
{
struct vsp1_dl_header_list * hdr = dl - > header - > lists ;
struct vsp1_dl_body * dlb ;
unsigned int num_lists = 0 ;
2016-09-19 21:18:01 +03:00
/*
* Fill the header with the display list bodies addresses and sizes . The
2016-07-12 19:49:46 +03:00
* address of the first body has already been filled when the display
* list was allocated .
*/
hdr - > num_bytes = dl - > body0 . num_entries
* sizeof ( * dl - > header - > lists ) ;
list_for_each_entry ( dlb , & dl - > fragments , list ) {
num_lists + + ;
hdr + + ;
hdr - > addr = dlb - > dma ;
hdr - > num_bytes = dlb - > num_entries
* sizeof ( * dl - > header - > lists ) ;
}
dl - > header - > num_lists = num_lists ;
2016-09-19 21:18:01 +03:00
/*
* If this display list ' s chain is not empty , we are on a list , where
2016-07-12 19:49:46 +03:00
* the next item in the list is the display list entity which should be
* automatically queued by the hardware .
*/
if ( ! list_empty ( & dl - > chain ) & & ! is_last ) {
struct vsp1_dl_list * next = list_next_entry ( dl , chain ) ;
dl - > header - > next_header = next - > dma ;
dl - > header - > flags = VSP1_DLH_AUTO_START ;
} else {
dl - > header - > flags = VSP1_DLH_INT_ENABLE ;
}
}
2015-11-09 01:06:57 +03:00
void vsp1_dl_list_commit ( struct vsp1_dl_list * dl )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
struct vsp1_dl_manager * dlm = dl - > dlm ;
struct vsp1_device * vsp1 = dlm - > vsp1 ;
2015-09-07 07:40:25 +03:00
unsigned long flags ;
bool update ;
2015-11-09 01:06:57 +03:00
spin_lock_irqsave ( & dlm - > lock , flags ) ;
2015-09-07 07:40:25 +03:00
2015-11-15 03:48:27 +03:00
if ( dl - > dlm - > mode = = VSP1_DL_MODE_HEADER ) {
2016-07-12 19:49:46 +03:00
struct vsp1_dl_list * dl_child ;
2016-09-19 21:18:01 +03:00
/*
* In header mode the caller guarantees that the hardware is
2016-03-03 19:36:34 +03:00
* idle at this point .
2015-11-15 03:48:27 +03:00
*/
2016-03-03 19:36:34 +03:00
2016-07-12 19:49:46 +03:00
/* Fill the header for the head and chained display lists. */
vsp1_dl_list_fill_header ( dl , list_empty ( & dl - > chain ) ) ;
2016-03-03 19:36:34 +03:00
2016-07-12 19:49:46 +03:00
list_for_each_entry ( dl_child , & dl - > chain , chain ) {
bool last = list_is_last ( & dl_child - > chain , & dl - > chain ) ;
vsp1_dl_list_fill_header ( dl_child , last ) ;
2016-03-03 19:36:34 +03:00
}
2016-09-19 21:18:01 +03:00
/*
* Commit the head display list to hardware . Chained headers
2016-07-12 19:49:46 +03:00
* will auto - start .
*/
2015-11-15 03:48:27 +03:00
vsp1_write ( vsp1 , VI6_DL_HDR_ADDR ( dlm - > index ) , dl - > dma ) ;
dlm - > active = dl ;
goto done ;
}
2015-09-07 07:40:25 +03:00
/* Once the UPD bit has been set the hardware can start processing the
* display list at any time and we can ' t touch the address and size
* registers . In that case mark the update as pending , it will be
* queued up to the hardware by the frame end interrupt handler .
*/
update = ! ! ( vsp1_read ( vsp1 , VI6_DL_BODY_SIZE ) & VI6_DL_BODY_SIZE_UPD ) ;
if ( update ) {
2016-03-03 15:26:47 +03:00
__vsp1_dl_list_put ( dlm - > pending ) ;
2015-11-09 01:06:57 +03:00
dlm - > pending = dl ;
2015-09-07 07:40:25 +03:00
goto done ;
}
/* Program the hardware with the display list body address and size.
* The UPD bit will be cleared by the device when the display list is
* processed .
*/
2016-03-03 19:36:34 +03:00
vsp1_write ( vsp1 , VI6_DL_HDR_ADDR ( 0 ) , dl - > body0 . dma ) ;
2015-09-07 07:40:25 +03:00
vsp1_write ( vsp1 , VI6_DL_BODY_SIZE , VI6_DL_BODY_SIZE_UPD |
2016-03-03 19:36:34 +03:00
( dl - > body0 . num_entries * sizeof ( * dl - > header - > lists ) ) ) ;
2015-09-07 07:40:25 +03:00
2016-03-03 15:26:47 +03:00
__vsp1_dl_list_put ( dlm - > queued ) ;
2015-11-09 01:06:57 +03:00
dlm - > queued = dl ;
2015-09-07 07:40:25 +03:00
done :
2015-11-09 01:06:57 +03:00
spin_unlock_irqrestore ( & dlm - > lock , flags ) ;
2015-09-07 07:40:25 +03:00
}
/* -----------------------------------------------------------------------------
2015-11-09 01:06:57 +03:00
* Display List Manager
2015-09-07 07:40:25 +03:00
*/
2015-11-09 01:06:57 +03:00
/* Interrupt Handling */
void vsp1_dlm_irq_display_start ( struct vsp1_dl_manager * dlm )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
spin_lock ( & dlm - > lock ) ;
2015-09-07 07:40:25 +03:00
/* The display start interrupt signals the end of the display list
* processing by the device . The active display list , if any , won ' t be
* accessed anymore and can be reused .
*/
2016-03-03 15:26:47 +03:00
__vsp1_dl_list_put ( dlm - > active ) ;
2015-11-09 01:06:57 +03:00
dlm - > active = NULL ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
spin_unlock ( & dlm - > lock ) ;
2015-09-07 07:40:25 +03:00
}
2015-11-09 01:06:57 +03:00
void vsp1_dlm_irq_frame_end ( struct vsp1_dl_manager * dlm )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
struct vsp1_device * vsp1 = dlm - > vsp1 ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
spin_lock ( & dlm - > lock ) ;
2016-03-03 15:26:47 +03:00
__vsp1_dl_list_put ( dlm - > active ) ;
2015-11-09 01:06:57 +03:00
dlm - > active = NULL ;
2015-09-07 07:40:25 +03:00
2015-11-15 03:48:27 +03:00
/* Header mode is used for mem-to-mem pipelines only. We don't need to
* perform any operation as there can ' t be any new display list queued
* in that case .
*/
if ( dlm - > mode = = VSP1_DL_MODE_HEADER )
goto done ;
2015-09-07 07:40:25 +03:00
/* The UPD bit set indicates that the commit operation raced with the
* interrupt and occurred after the frame end event and UPD clear but
* before interrupt processing . The hardware hasn ' t taken the update
* into account yet , we ' ll thus skip one frame and retry .
*/
if ( vsp1_read ( vsp1 , VI6_DL_BODY_SIZE ) & VI6_DL_BODY_SIZE_UPD )
goto done ;
/* The device starts processing the queued display list right after the
* frame end interrupt . The display list thus becomes active .
*/
2015-11-09 01:06:57 +03:00
if ( dlm - > queued ) {
dlm - > active = dlm - > queued ;
dlm - > queued = NULL ;
2015-09-07 07:40:25 +03:00
}
/* Now that the UPD bit has been cleared we can queue the next display
* list to the hardware if one has been prepared .
*/
2015-11-09 01:06:57 +03:00
if ( dlm - > pending ) {
struct vsp1_dl_list * dl = dlm - > pending ;
2015-09-07 07:40:25 +03:00
2016-03-03 19:36:34 +03:00
vsp1_write ( vsp1 , VI6_DL_HDR_ADDR ( 0 ) , dl - > body0 . dma ) ;
2015-09-07 07:40:25 +03:00
vsp1_write ( vsp1 , VI6_DL_BODY_SIZE , VI6_DL_BODY_SIZE_UPD |
2016-03-03 19:36:34 +03:00
( dl - > body0 . num_entries *
sizeof ( * dl - > header - > lists ) ) ) ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
dlm - > queued = dl ;
dlm - > pending = NULL ;
2015-09-07 07:40:25 +03:00
}
done :
2015-11-09 01:06:57 +03:00
spin_unlock ( & dlm - > lock ) ;
2015-09-07 07:40:25 +03:00
}
2015-11-09 01:06:57 +03:00
/* Hardware Setup */
void vsp1_dlm_setup ( struct vsp1_device * vsp1 )
2015-09-07 07:40:25 +03:00
{
2015-11-01 20:18:56 +03:00
u32 ctrl = ( 256 < < VI6_DL_CTRL_AR_WAIT_SHIFT )
| VI6_DL_CTRL_DC2 | VI6_DL_CTRL_DC1 | VI6_DL_CTRL_DC0
| VI6_DL_CTRL_DLE ;
2015-09-07 07:40:25 +03:00
2015-11-01 20:18:56 +03:00
/* The DRM pipeline operates with display lists in Continuous Frame
* Mode , all other pipelines use manual start .
2015-09-07 07:40:25 +03:00
*/
if ( vsp1 - > drm )
2015-11-01 20:18:56 +03:00
ctrl | = VI6_DL_CTRL_CFM0 | VI6_DL_CTRL_NH0 ;
2015-09-07 07:40:25 +03:00
vsp1_write ( vsp1 , VI6_DL_CTRL , ctrl ) ;
vsp1_write ( vsp1 , VI6_DL_SWAP , VI6_DL_SWAP_LWS ) ;
}
2015-11-09 01:06:57 +03:00
void vsp1_dlm_reset ( struct vsp1_dl_manager * dlm )
{
2016-03-03 15:26:47 +03:00
unsigned long flags ;
spin_lock_irqsave ( & dlm - > lock , flags ) ;
__vsp1_dl_list_put ( dlm - > active ) ;
__vsp1_dl_list_put ( dlm - > queued ) ;
__vsp1_dl_list_put ( dlm - > pending ) ;
spin_unlock_irqrestore ( & dlm - > lock , flags ) ;
2015-11-09 01:06:57 +03:00
dlm - > active = NULL ;
dlm - > queued = NULL ;
dlm - > pending = NULL ;
}
2015-09-07 07:40:25 +03:00
2016-05-14 01:17:02 +03:00
/*
* Free all fragments awaiting to be garbage - collected .
*
* This function must be called without the display list manager lock held .
*/
static void vsp1_dlm_fragments_free ( struct vsp1_dl_manager * dlm )
{
unsigned long flags ;
spin_lock_irqsave ( & dlm - > lock , flags ) ;
while ( ! list_empty ( & dlm - > gc_fragments ) ) {
struct vsp1_dl_body * dlb ;
dlb = list_first_entry ( & dlm - > gc_fragments , struct vsp1_dl_body ,
list ) ;
list_del ( & dlb - > list ) ;
spin_unlock_irqrestore ( & dlm - > lock , flags ) ;
vsp1_dl_fragment_free ( dlb ) ;
spin_lock_irqsave ( & dlm - > lock , flags ) ;
}
spin_unlock_irqrestore ( & dlm - > lock , flags ) ;
}
static void vsp1_dlm_garbage_collect ( struct work_struct * work )
{
struct vsp1_dl_manager * dlm =
container_of ( work , struct vsp1_dl_manager , gc_work ) ;
vsp1_dlm_fragments_free ( dlm ) ;
}
2015-11-15 03:27:52 +03:00
struct vsp1_dl_manager * vsp1_dlm_create ( struct vsp1_device * vsp1 ,
2015-11-15 03:48:27 +03:00
unsigned int index ,
2015-11-15 03:27:52 +03:00
unsigned int prealloc )
2015-09-07 07:40:25 +03:00
{
2015-11-15 03:27:52 +03:00
struct vsp1_dl_manager * dlm ;
2015-09-07 07:40:25 +03:00
unsigned int i ;
2015-11-15 03:27:52 +03:00
dlm = devm_kzalloc ( vsp1 - > dev , sizeof ( * dlm ) , GFP_KERNEL ) ;
if ( ! dlm )
return NULL ;
2015-11-15 03:48:27 +03:00
dlm - > index = index ;
dlm - > mode = index = = 0 & & ! vsp1 - > info - > uapi
? VSP1_DL_MODE_HEADERLESS : VSP1_DL_MODE_HEADER ;
2015-11-09 01:06:57 +03:00
dlm - > vsp1 = vsp1 ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
spin_lock_init ( & dlm - > lock ) ;
INIT_LIST_HEAD ( & dlm - > free ) ;
2016-05-14 01:17:02 +03:00
INIT_LIST_HEAD ( & dlm - > gc_fragments ) ;
INIT_WORK ( & dlm - > gc_work , vsp1_dlm_garbage_collect ) ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
for ( i = 0 ; i < prealloc ; + + i ) {
struct vsp1_dl_list * dl ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
dl = vsp1_dl_list_alloc ( dlm ) ;
if ( ! dl )
2015-11-15 03:27:52 +03:00
return NULL ;
2015-09-07 07:40:25 +03:00
2015-11-09 01:06:57 +03:00
list_add_tail ( & dl - > list , & dlm - > free ) ;
2015-09-07 07:40:25 +03:00
}
2015-11-15 03:27:52 +03:00
return dlm ;
2015-09-07 07:40:25 +03:00
}
2015-11-15 03:27:52 +03:00
void vsp1_dlm_destroy ( struct vsp1_dl_manager * dlm )
2015-09-07 07:40:25 +03:00
{
2015-11-09 01:06:57 +03:00
struct vsp1_dl_list * dl , * next ;
2015-11-15 03:27:52 +03:00
if ( ! dlm )
return ;
2016-05-14 01:17:02 +03:00
cancel_work_sync ( & dlm - > gc_work ) ;
2015-11-09 01:06:57 +03:00
list_for_each_entry_safe ( dl , next , & dlm - > free , list ) {
list_del ( & dl - > list ) ;
vsp1_dl_list_free ( dl ) ;
}
2016-05-14 01:17:02 +03:00
vsp1_dlm_fragments_free ( dlm ) ;
2015-09-07 07:40:25 +03:00
}