2014-09-10 14:43:49 +04:00
# ifndef __DRM_DRM_LEGACY_H__
# define __DRM_DRM_LEGACY_H__
/*
* Legacy driver interfaces for the Direct Rendering Manager
*
* Copyright 1999 Precision Insight , Inc . , Cedar Park , Texas .
* Copyright 2000 VA Linux Systems , Inc . , Sunnyvale , California .
* Copyright ( c ) 2009 - 2010 , Code Aurora Forum .
* All rights reserved .
* Copyright © 2014 Intel Corporation
* Daniel Vetter < daniel . vetter @ ffwll . ch >
*
* Author : Rickard E . ( Rik ) Faith < faith @ valinux . com >
* Author : Gareth Hughes < gareth @ valinux . com >
*
* Permission is hereby granted , free of charge , to any person obtaining a
* copy of this software and associated documentation files ( the " Software " ) ,
* to deal in the Software without restriction , including without limitation
* the rights to use , copy , modify , merge , publish , distribute , sublicense ,
* and / or sell copies of the Software , and to permit persons to whom the
* Software is furnished to do so , subject to the following conditions :
*
* The above copyright notice and this permission notice ( including the next
* paragraph ) shall be included in all copies or substantial portions of the
* Software .
*
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND , EXPRESS OR
* IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY ,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT . IN NO EVENT SHALL
* VA LINUX SYSTEMS AND / OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM , DAMAGES OR
* OTHER LIABILITY , WHETHER IN AN ACTION OF CONTRACT , TORT OR OTHERWISE ,
* ARISING FROM , OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
* OTHER DEALINGS IN THE SOFTWARE .
*/
/*
* Legacy Support for palateontologic DRM drivers
*
* If you add a new driver and it uses any of these functions or structures ,
* you ' re doing it terribly wrong .
*/
2014-09-11 09:43:25 +04:00
/**
* DMA buffer .
*/
struct drm_buf {
int idx ; /**< Index into master buflist */
int total ; /**< Buffer size */
int order ; /**< log-base-2(total) */
int used ; /**< Amount of buffer in use (for DMA) */
unsigned long offset ; /**< Byte offset (used internally) */
void * address ; /**< Address of buffer */
unsigned long bus_address ; /**< Bus address of buffer */
struct drm_buf * next ; /**< Kernel-only: used for free list */
__volatile__ int waiting ; /**< On kernel DMA queue */
__volatile__ int pending ; /**< On hardware DMA queue */
struct drm_file * file_priv ; /**< Private of holding file descr */
int context ; /**< Kernel queue for this buffer */
int while_locked ; /**< Dispatch this buffer while locked */
enum {
DRM_LIST_NONE = 0 ,
DRM_LIST_FREE = 1 ,
DRM_LIST_WAIT = 2 ,
DRM_LIST_PEND = 3 ,
DRM_LIST_PRIO = 4 ,
DRM_LIST_RECLAIM = 5
} list ; /**< Which list we're on */
int dev_priv_size ; /**< Size of buffer private storage */
void * dev_private ; /**< Per-buffer private storage */
} ;
typedef struct drm_dma_handle {
dma_addr_t busaddr ;
void * vaddr ;
size_t size ;
} drm_dma_handle_t ;
/**
* Buffer entry . There is one of this for each buffer size order .
*/
struct drm_buf_entry {
int buf_size ; /**< size */
int buf_count ; /**< number of buffers */
struct drm_buf * buflist ; /**< buffer list */
int seg_count ;
int page_order ;
struct drm_dma_handle * * seglist ;
int low_mark ; /**< Low water mark */
int high_mark ; /**< High water mark */
} ;
/**
* DMA data .
*/
struct drm_device_dma {
struct drm_buf_entry bufs [ DRM_MAX_ORDER + 1 ] ; /**< buffers, grouped by their size order */
int buf_count ; /**< total number of buffers */
struct drm_buf * * buflist ; /**< Vector of pointers into drm_device_dma::bufs */
int seg_count ;
int page_count ; /**< number of pages */
unsigned long * pagelist ; /**< page list */
unsigned long byte_count ;
enum {
_DRM_DMA_USE_AGP = 0x01 ,
_DRM_DMA_USE_SG = 0x02 ,
_DRM_DMA_USE_FB = 0x04 ,
_DRM_DMA_USE_PCI_RO = 0x08
} flags ;
} ;
/**
* Scatter - gather memory .
*/
struct drm_sg_mem {
unsigned long handle ;
void * virtual ;
int pages ;
struct page * * pagelist ;
dma_addr_t * busaddr ;
} ;
/**
* Kernel side of a mapping
*/
struct drm_local_map {
resource_size_t offset ; /**< Requested physical address (0 for SAREA)*/
unsigned long size ; /**< Requested physical size (bytes) */
enum drm_map_type type ; /**< Type of memory to map */
enum drm_map_flags flags ; /**< Flags */
void * handle ; /**< User-space: "Handle" to pass to mmap() */
/**< Kernel-space: kernel-virtual address */
int mtrr ; /**< MTRR slot used */
} ;
typedef struct drm_local_map drm_local_map_t ;
/**
* Mappings list
*/
struct drm_map_list {
struct list_head head ; /**< list head */
struct drm_hash_item hash ;
struct drm_local_map * map ; /**< mapping */
uint64_t user_token ;
struct drm_master * master ;
} ;
2014-09-10 14:43:49 +04:00
int drm_legacy_addmap ( struct drm_device * d , resource_size_t offset ,
unsigned int size , enum drm_map_type type ,
enum drm_map_flags flags , struct drm_local_map * * map_p ) ;
int drm_legacy_rmmap ( struct drm_device * d , struct drm_local_map * map ) ;
int drm_legacy_rmmap_locked ( struct drm_device * d , struct drm_local_map * map ) ;
struct drm_local_map * drm_legacy_getsarea ( struct drm_device * dev ) ;
int drm_legacy_addbufs_agp ( struct drm_device * d , struct drm_buf_desc * req ) ;
int drm_legacy_addbufs_pci ( struct drm_device * d , struct drm_buf_desc * req ) ;
2014-09-10 14:43:58 +04:00
/**
* Test that the hardware lock is held by the caller , returning otherwise .
*
* \ param dev DRM device .
* \ param filp file pointer of the caller .
*/
# define LOCK_TEST_WITH_RETURN( dev, _file_priv ) \
do { \
if ( ! _DRM_LOCK_IS_HELD ( _file_priv - > master - > lock . hw_lock - > lock ) | | \
_file_priv - > master - > lock . file_priv ! = _file_priv ) { \
DRM_ERROR ( " %s called without lock held, held %d owner %p %p \n " , \
__func__ , _DRM_LOCK_IS_HELD ( _file_priv - > master - > lock . hw_lock - > lock ) , \
_file_priv - > master - > lock . file_priv , _file_priv ) ; \
return - EINVAL ; \
} \
} while ( 0 )
2014-09-10 14:43:49 +04:00
void drm_legacy_idlelock_take ( struct drm_lock_data * lock ) ;
void drm_legacy_idlelock_release ( struct drm_lock_data * lock ) ;
2014-09-10 14:43:51 +04:00
/* drm_pci.c dma alloc wrappers */
void __drm_legacy_pci_free ( struct drm_device * dev , drm_dma_handle_t * dmah ) ;
2014-09-10 14:43:56 +04:00
/* drm_memory.c */
void drm_legacy_ioremap ( struct drm_local_map * map , struct drm_device * dev ) ;
void drm_legacy_ioremap_wc ( struct drm_local_map * map , struct drm_device * dev ) ;
void drm_legacy_ioremapfree ( struct drm_local_map * map , struct drm_device * dev ) ;
static __inline__ struct drm_local_map * drm_legacy_findmap ( struct drm_device * dev ,
unsigned int token )
{
struct drm_map_list * _entry ;
list_for_each_entry ( _entry , & dev - > maplist , head )
if ( _entry - > user_token = = token )
return _entry - > map ;
return NULL ;
}
2014-09-10 14:43:49 +04:00
# endif /* __DRM_DRM_LEGACY_H__ */