2009-12-10 03:19:58 +03:00
/**************************************************************************
*
2015-07-29 22:38:02 +03:00
* Copyright © 2009 - 2015 VMware , Inc . , Palo Alto , CA . , USA
2009-12-10 03:19:58 +03:00
* All Rights Reserved .
*
* Permission is hereby granted , free of charge , to any person obtaining a
* copy of this software and associated documentation files ( the
* " Software " ) , to deal in the Software without restriction , including
* without limitation the rights to use , copy , modify , merge , publish ,
* distribute , sub license , and / or sell copies of the Software , and to
* permit persons to whom the Software is furnished to do so , subject to
* the following conditions :
*
* The above copyright notice and this permission notice ( including the
* next paragraph ) shall be included in all copies or substantial portions
* of the Software .
*
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND , EXPRESS OR
* IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY ,
* FITNESS FOR A PARTICULAR PURPOSE AND NON - INFRINGEMENT . IN NO EVENT SHALL
* THE COPYRIGHT HOLDERS , AUTHORS AND / OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM ,
* DAMAGES OR OTHER LIABILITY , WHETHER IN AN ACTION OF CONTRACT , TORT OR
* OTHERWISE , ARISING FROM , OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
* USE OR OTHER DEALINGS IN THE SOFTWARE .
*
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
# ifndef __VMWGFX_DRM_H__
# define __VMWGFX_DRM_H__
2015-11-30 17:10:56 +03:00
# include "drm.h"
2012-11-21 15:32:19 +04:00
2016-04-07 21:40:10 +03:00
# if defined(__cplusplus)
extern " C " {
# endif
2009-12-10 03:19:58 +03:00
# define DRM_VMW_MAX_SURFACE_FACES 6
# define DRM_VMW_MAX_MIP_LEVELS 24
# define DRM_VMW_GET_PARAM 0
# define DRM_VMW_ALLOC_DMABUF 1
2018-06-19 16:02:16 +03:00
# define DRM_VMW_ALLOC_BO 1
2009-12-10 03:19:58 +03:00
# define DRM_VMW_UNREF_DMABUF 2
2017-03-24 01:26:16 +03:00
# define DRM_VMW_HANDLE_CLOSE 2
2009-12-10 03:19:58 +03:00
# define DRM_VMW_CURSOR_BYPASS 3
/* guarded by DRM_VMW_PARAM_NUM_STREAMS != 0*/
# define DRM_VMW_CONTROL_STREAM 4
# define DRM_VMW_CLAIM_STREAM 5
# define DRM_VMW_UNREF_STREAM 6
/* guarded by DRM_VMW_PARAM_3D == 1 */
# define DRM_VMW_CREATE_CONTEXT 7
# define DRM_VMW_UNREF_CONTEXT 8
# define DRM_VMW_CREATE_SURFACE 9
# define DRM_VMW_UNREF_SURFACE 10
# define DRM_VMW_REF_SURFACE 11
# define DRM_VMW_EXECBUF 12
2011-09-02 00:18:44 +04:00
# define DRM_VMW_GET_3D_CAP 13
# define DRM_VMW_FENCE_WAIT 14
# define DRM_VMW_FENCE_SIGNALED 15
# define DRM_VMW_FENCE_UNREF 16
# define DRM_VMW_FENCE_EVENT 17
2011-10-04 22:13:26 +04:00
# define DRM_VMW_PRESENT 18
# define DRM_VMW_PRESENT_READBACK 19
2011-10-26 01:35:53 +04:00
# define DRM_VMW_UPDATE_LAYOUT 20
2012-11-21 13:23:14 +04:00
# define DRM_VMW_CREATE_SHADER 21
# define DRM_VMW_UNREF_SHADER 22
# define DRM_VMW_GB_SURFACE_CREATE 23
# define DRM_VMW_GB_SURFACE_REF 24
2012-11-21 15:32:19 +04:00
# define DRM_VMW_SYNCCPU 25
2015-08-10 20:39:35 +03:00
# define DRM_VMW_CREATE_EXTENDED_CONTEXT 26
2018-06-21 00:48:35 +03:00
# define DRM_VMW_GB_SURFACE_CREATE_EXT 27
# define DRM_VMW_GB_SURFACE_REF_EXT 28
2009-12-10 03:19:58 +03:00
/*************************************************************************/
/**
* DRM_VMW_GET_PARAM - get device information .
*
* DRM_VMW_PARAM_FIFO_OFFSET :
* Offset to use to map the first page of the FIFO read - only .
* The fifo is mapped using the mmap ( ) system call on the drm device .
*
* DRM_VMW_PARAM_OVERLAY_IOCTL :
* Does the driver support the overlay ioctl .
2018-06-21 01:09:43 +03:00
*
* DRM_VMW_PARAM_SM4_1
* SM4_1 support is enabled .
2009-12-10 03:19:58 +03:00
*/
# define DRM_VMW_PARAM_NUM_STREAMS 0
# define DRM_VMW_PARAM_NUM_FREE_STREAMS 1
# define DRM_VMW_PARAM_3D 2
2011-09-02 00:18:40 +04:00
# define DRM_VMW_PARAM_HW_CAPS 3
# define DRM_VMW_PARAM_FIFO_CAPS 4
# define DRM_VMW_PARAM_MAX_FB_SIZE 5
2011-09-02 00:18:41 +04:00
# define DRM_VMW_PARAM_FIFO_HW_VERSION 6
2012-11-21 13:23:14 +04:00
# define DRM_VMW_PARAM_MAX_SURF_MEMORY 7
# define DRM_VMW_PARAM_3D_CAPS_SIZE 8
2012-11-21 15:34:47 +04:00
# define DRM_VMW_PARAM_MAX_MOB_MEMORY 9
2014-02-12 15:07:38 +04:00
# define DRM_VMW_PARAM_MAX_MOB_SIZE 10
2015-06-26 11:42:06 +03:00
# define DRM_VMW_PARAM_SCREEN_TARGET 11
2015-08-10 20:39:35 +03:00
# define DRM_VMW_PARAM_DX 12
2018-06-19 02:44:48 +03:00
# define DRM_VMW_PARAM_HW_CAPS2 13
2018-06-21 01:09:43 +03:00
# define DRM_VMW_PARAM_SM4_1 14
2009-12-10 03:19:58 +03:00
2014-03-18 18:00:56 +04:00
/**
* enum drm_vmw_handle_type - handle type for ref ioctls
*
*/
enum drm_vmw_handle_type {
DRM_VMW_HANDLE_LEGACY = 0 ,
DRM_VMW_HANDLE_PRIME = 1
} ;
2009-12-10 03:19:58 +03:00
/**
* struct drm_vmw_getparam_arg
*
* @ value : Returned value . //Out
* @ param : Parameter to query . //In.
*
* Argument to the DRM_VMW_GET_PARAM Ioctl .
*/
struct drm_vmw_getparam_arg {
2015-03-11 05:07:51 +03:00
__u64 value ;
__u32 param ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
/*************************************************************************/
/**
* DRM_VMW_CREATE_CONTEXT - Create a host context .
*
* Allocates a device unique context id , and queues a create context command
* for the host . Does not wait for host completion .
*/
/**
* struct drm_vmw_context_arg
*
* @ cid : Device unique context ID .
*
* Output argument to the DRM_VMW_CREATE_CONTEXT Ioctl .
* Input argument to the DRM_VMW_UNREF_CONTEXT Ioctl .
*/
struct drm_vmw_context_arg {
2015-03-11 05:07:51 +03:00
__s32 cid ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
/*************************************************************************/
/**
* DRM_VMW_UNREF_CONTEXT - Create a host context .
*
* Frees a global context id , and queues a destroy host command for the host .
* Does not wait for host completion . The context ID can be used directly
* in the command stream and shows up as the same context ID on the host .
*/
/*************************************************************************/
/**
* DRM_VMW_CREATE_SURFACE - Create a host suface .
*
* Allocates a device unique surface id , and queues a create surface command
* for the host . Does not wait for host completion . The surface ID can be
* used directly in the command stream and shows up as the same surface
* ID on the host .
*/
/**
* struct drm_wmv_surface_create_req
*
* @ flags : Surface flags as understood by the host .
* @ format : Surface format as understood by the host .
* @ mip_levels : Number of mip levels for each face .
* An unused face should have 0 encoded .
* @ size_addr : Address of a user - space array of sruct drm_vmw_size
2015-03-11 05:07:51 +03:00
* cast to an __u64 for 32 - 64 bit compatibility .
2009-12-10 03:19:58 +03:00
* The size of the array should equal the total number of mipmap levels .
* @ shareable : Boolean whether other clients ( as identified by file descriptors )
* may reference this surface .
2010-02-09 22:41:55 +03:00
* @ scanout : Boolean whether the surface is intended to be used as a
* scanout .
2009-12-10 03:19:58 +03:00
*
* Input data to the DRM_VMW_CREATE_SURFACE Ioctl .
* Output data from the DRM_VMW_REF_SURFACE Ioctl .
*/
struct drm_vmw_surface_create_req {
2015-03-11 05:07:51 +03:00
__u32 flags ;
__u32 format ;
__u32 mip_levels [ DRM_VMW_MAX_SURFACE_FACES ] ;
__u64 size_addr ;
__s32 shareable ;
__s32 scanout ;
2009-12-10 03:19:58 +03:00
} ;
/**
* struct drm_wmv_surface_arg
*
* @ sid : Surface id of created surface or surface to destroy or reference .
2014-03-18 18:00:56 +04:00
* @ handle_type : Handle type for DRM_VMW_REF_SURFACE Ioctl .
2009-12-10 03:19:58 +03:00
*
* Output data from the DRM_VMW_CREATE_SURFACE Ioctl .
* Input argument to the DRM_VMW_UNREF_SURFACE Ioctl .
* Input argument to the DRM_VMW_REF_SURFACE Ioctl .
*/
struct drm_vmw_surface_arg {
2015-03-11 05:07:51 +03:00
__s32 sid ;
2014-03-18 18:00:56 +04:00
enum drm_vmw_handle_type handle_type ;
2009-12-10 03:19:58 +03:00
} ;
/**
* struct drm_vmw_size ioctl .
*
* @ width - mip level width
* @ height - mip level height
* @ depth - mip level depth
*
* Description of a mip level .
* Input data to the DRM_WMW_CREATE_SURFACE Ioctl .
*/
struct drm_vmw_size {
2015-03-11 05:07:51 +03:00
__u32 width ;
__u32 height ;
__u32 depth ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
/**
* union drm_vmw_surface_create_arg
*
* @ rep : Output data as described above .
* @ req : Input data as described above .
*
* Argument to the DRM_VMW_CREATE_SURFACE Ioctl .
*/
union drm_vmw_surface_create_arg {
struct drm_vmw_surface_arg rep ;
struct drm_vmw_surface_create_req req ;
} ;
/*************************************************************************/
/**
* DRM_VMW_REF_SURFACE - Reference a host surface .
*
* Puts a reference on a host surface with a give sid , as previously
* returned by the DRM_VMW_CREATE_SURFACE ioctl .
* A reference will make sure the surface isn ' t destroyed while we hold
* it and will allow the calling client to use the surface ID in the command
* stream .
*
* On successful return , the Ioctl returns the surface information given
* in the DRM_VMW_CREATE_SURFACE ioctl .
*/
/**
* union drm_vmw_surface_reference_arg
*
* @ rep : Output data as described above .
* @ req : Input data as described above .
*
* Argument to the DRM_VMW_REF_SURFACE Ioctl .
*/
union drm_vmw_surface_reference_arg {
struct drm_vmw_surface_create_req rep ;
struct drm_vmw_surface_arg req ;
} ;
/*************************************************************************/
/**
* DRM_VMW_UNREF_SURFACE - Unreference a host surface .
*
* Clear a reference previously put on a host surface .
* When all references are gone , including the one implicitly placed
* on creation ,
* a destroy surface command will be queued for the host .
* Does not wait for completion .
*/
/*************************************************************************/
/**
* DRM_VMW_EXECBUF
*
* Submit a command buffer for execution on the host , and return a
2011-09-02 00:18:42 +04:00
* fence seqno that when signaled , indicates that the command buffer has
2009-12-10 03:19:58 +03:00
* executed .
*/
/**
* struct drm_vmw_execbuf_arg
*
2015-03-11 05:07:51 +03:00
* @ commands : User - space address of a command buffer cast to an __u64 .
2009-12-10 03:19:58 +03:00
* @ command - size : Size in bytes of the command buffer .
2010-02-09 22:41:55 +03:00
* @ throttle - us : Sleep until software is less than @ throttle_us
* microseconds ahead of hardware . The driver may round this value
* to the nearest kernel tick .
2009-12-10 03:19:58 +03:00
* @ fence_rep : User - space address of a struct drm_vmw_fence_rep cast to an
2015-03-11 05:07:51 +03:00
* __u64 .
2010-02-10 00:29:47 +03:00
* @ version : Allows expanding the execbuf ioctl parameters without breaking
* backwards compatibility , since user - space will always tell the kernel
* which version it uses .
2017-07-05 11:37:55 +03:00
* @ flags : Execbuf flags .
* @ imported_fence_fd : FD for a fence imported from another device
2009-12-10 03:19:58 +03:00
*
* Argument to the DRM_VMW_EXECBUF Ioctl .
*/
2015-08-10 20:39:35 +03:00
# define DRM_VMW_EXECBUF_VERSION 2
2010-02-10 00:29:47 +03:00
2017-07-05 11:37:55 +03:00
# define DRM_VMW_EXECBUF_FLAG_IMPORT_FENCE_FD (1 << 0)
# define DRM_VMW_EXECBUF_FLAG_EXPORT_FENCE_FD (1 << 1)
2009-12-10 03:19:58 +03:00
struct drm_vmw_execbuf_arg {
2015-03-11 05:07:51 +03:00
__u64 commands ;
__u32 command_size ;
__u32 throttle_us ;
__u64 fence_rep ;
__u32 version ;
__u32 flags ;
__u32 context_handle ;
2017-07-05 11:37:55 +03:00
__s32 imported_fence_fd ;
2009-12-10 03:19:58 +03:00
} ;
/**
* struct drm_vmw_fence_rep
*
2011-09-02 00:18:44 +04:00
* @ handle : Fence object handle for fence associated with a command submission .
* @ mask : Fence flags relevant for this fence object .
* @ seqno : Fence sequence number in fifo . A fence object with a lower
* seqno will signal the EXEC flag before a fence object with a higher
* seqno . This can be used by user - space to avoid kernel calls to determine
* whether a fence has signaled the EXEC flag . Note that @ seqno will
* wrap at 32 - bit .
* @ passed_seqno : The highest seqno number processed by the hardware
* so far . This can be used to mark user - space fence objects as signaled , and
* to determine whether a fence seqno might be stale .
2017-07-05 11:37:55 +03:00
* @ fd : FD associated with the fence , - 1 if not exported
2009-12-10 03:19:58 +03:00
* @ error : This member should ' ve been set to - EFAULT on submission .
* The following actions should be take on completion :
* error = = - EFAULT : Fence communication failed . The host is synchronized .
* Use the last fence id read from the FIFO fence register .
* error ! = 0 & & error ! = - EFAULT :
* Fence submission failed . The host is synchronized . Use the fence_seq member .
* error = = 0 : All is OK , The host may not be synchronized .
* Use the fence_seq member .
*
* Input / Output data to the DRM_VMW_EXECBUF Ioctl .
*/
struct drm_vmw_fence_rep {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 mask ;
__u32 seqno ;
__u32 passed_seqno ;
2017-07-05 11:37:55 +03:00
__s32 fd ;
2015-03-11 05:07:51 +03:00
__s32 error ;
2009-12-10 03:19:58 +03:00
} ;
/*************************************************************************/
/**
2018-06-19 16:02:16 +03:00
* DRM_VMW_ALLOC_BO
2009-12-10 03:19:58 +03:00
*
2018-06-19 16:02:16 +03:00
* Allocate a buffer object that is visible also to the host .
2009-12-10 03:19:58 +03:00
* NOTE : The buffer is
* identified by a handle and an offset , which are private to the guest , but
* useable in the command stream . The guest kernel may translate these
* and patch up the command stream accordingly . In the future , the offset may
* be zero at all times , or it may disappear from the interface before it is
* fixed .
*
2018-06-19 16:02:16 +03:00
* The buffer object may stay user - space mapped in the guest at all times ,
2009-12-10 03:19:58 +03:00
* and is thus suitable for sub - allocation .
*
2018-06-19 16:02:16 +03:00
* Buffer objects are mapped using the mmap ( ) syscall on the drm device .
2009-12-10 03:19:58 +03:00
*/
/**
2018-06-19 16:02:16 +03:00
* struct drm_vmw_alloc_bo_req
2009-12-10 03:19:58 +03:00
*
* @ size : Required minimum size of the buffer .
*
2018-06-19 16:02:16 +03:00
* Input data to the DRM_VMW_ALLOC_BO Ioctl .
2009-12-10 03:19:58 +03:00
*/
2018-06-19 16:02:16 +03:00
struct drm_vmw_alloc_bo_req {
2015-03-11 05:07:51 +03:00
__u32 size ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
2018-06-19 16:02:16 +03:00
# define drm_vmw_alloc_dmabuf_req drm_vmw_alloc_bo_req
2009-12-10 03:19:58 +03:00
/**
2018-06-19 16:02:16 +03:00
* struct drm_vmw_bo_rep
2009-12-10 03:19:58 +03:00
*
* @ map_handle : Offset to use in the mmap ( ) call used to map the buffer .
* @ handle : Handle unique to this buffer . Used for unreferencing .
* @ cur_gmr_id : GMR id to use in the command stream when this buffer is
* referenced . See not above .
* @ cur_gmr_offset : Offset to use in the command stream when this buffer is
* referenced . See note above .
*
2018-06-19 16:02:16 +03:00
* Output data from the DRM_VMW_ALLOC_BO Ioctl .
2009-12-10 03:19:58 +03:00
*/
2018-06-19 16:02:16 +03:00
struct drm_vmw_bo_rep {
2015-03-11 05:07:51 +03:00
__u64 map_handle ;
__u32 handle ;
__u32 cur_gmr_id ;
__u32 cur_gmr_offset ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
2018-06-19 16:02:16 +03:00
# define drm_vmw_dmabuf_rep drm_vmw_bo_rep
2009-12-10 03:19:58 +03:00
/**
2018-06-19 16:02:16 +03:00
* union drm_vmw_alloc_bo_arg
2009-12-10 03:19:58 +03:00
*
* @ req : Input data as described above .
* @ rep : Output data as described above .
*
2018-06-19 16:02:16 +03:00
* Argument to the DRM_VMW_ALLOC_BO Ioctl .
2009-12-10 03:19:58 +03:00
*/
2018-06-19 16:02:16 +03:00
union drm_vmw_alloc_bo_arg {
struct drm_vmw_alloc_bo_req req ;
struct drm_vmw_bo_rep rep ;
2009-12-10 03:19:58 +03:00
} ;
2018-06-19 16:02:16 +03:00
# define drm_vmw_alloc_dmabuf_arg drm_vmw_alloc_bo_arg
2009-12-10 03:19:58 +03:00
/*************************************************************************/
/**
* DRM_VMW_CONTROL_STREAM - Control overlays , aka streams .
*
* This IOCTL controls the overlay units of the svga device .
* The SVGA overlay units does not work like regular hardware units in
* that they do not automaticaly read back the contents of the given dma
* buffer . But instead only read back for each call to this ioctl , and
* at any point between this call being made and a following call that
* either changes the buffer or disables the stream .
*/
/**
* struct drm_vmw_rect
*
* Defines a rectangle . Used in the overlay ioctl to define
* source and destination rectangle .
*/
struct drm_vmw_rect {
2015-03-11 05:07:51 +03:00
__s32 x ;
__s32 y ;
__u32 w ;
__u32 h ;
2009-12-10 03:19:58 +03:00
} ;
/**
* struct drm_vmw_control_stream_arg
*
* @ stream_id : Stearm to control
* @ enabled : If false all following arguments are ignored .
* @ handle : Handle to buffer for getting data from .
* @ format : Format of the overlay as understood by the host .
* @ width : Width of the overlay .
* @ height : Height of the overlay .
* @ size : Size of the overlay in bytes .
* @ pitch : Array of pitches , the two last are only used for YUV12 formats .
* @ offset : Offset from start of dma buffer to overlay .
* @ src : Source rect , must be within the defined area above .
* @ dst : Destination rect , x and y may be negative .
*
* Argument to the DRM_VMW_CONTROL_STREAM Ioctl .
*/
struct drm_vmw_control_stream_arg {
2015-03-11 05:07:51 +03:00
__u32 stream_id ;
__u32 enabled ;
2009-12-10 03:19:58 +03:00
2015-03-11 05:07:51 +03:00
__u32 flags ;
__u32 color_key ;
2009-12-10 03:19:58 +03:00
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 offset ;
__s32 format ;
__u32 size ;
__u32 width ;
__u32 height ;
__u32 pitch [ 3 ] ;
2009-12-10 03:19:58 +03:00
2015-03-11 05:07:51 +03:00
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
struct drm_vmw_rect src ;
struct drm_vmw_rect dst ;
} ;
/*************************************************************************/
/**
* DRM_VMW_CURSOR_BYPASS - Give extra information about cursor bypass .
*
*/
# define DRM_VMW_CURSOR_BYPASS_ALL (1 << 0)
# define DRM_VMW_CURSOR_BYPASS_FLAGS (1)
/**
* struct drm_vmw_cursor_bypass_arg
*
* @ flags : Flags .
* @ crtc_id : Crtc id , only used if DMR_CURSOR_BYPASS_ALL isn ' t passed .
* @ xpos : X position of cursor .
* @ ypos : Y position of cursor .
* @ xhot : X hotspot .
* @ yhot : Y hotspot .
*
* Argument to the DRM_VMW_CURSOR_BYPASS Ioctl .
*/
struct drm_vmw_cursor_bypass_arg {
2015-03-11 05:07:51 +03:00
__u32 flags ;
__u32 crtc_id ;
__s32 xpos ;
__s32 ypos ;
__s32 xhot ;
__s32 yhot ;
2009-12-10 03:19:58 +03:00
} ;
/*************************************************************************/
/**
* DRM_VMW_CLAIM_STREAM - Claim a single stream .
*/
/**
* struct drm_vmw_context_arg
*
* @ stream_id : Device unique context ID .
*
* Output argument to the DRM_VMW_CREATE_CONTEXT Ioctl .
* Input argument to the DRM_VMW_UNREF_CONTEXT Ioctl .
*/
struct drm_vmw_stream_arg {
2015-03-11 05:07:51 +03:00
__u32 stream_id ;
__u32 pad64 ;
2009-12-10 03:19:58 +03:00
} ;
/*************************************************************************/
/**
* DRM_VMW_UNREF_STREAM - Unclaim a stream .
*
* Return a single stream that was claimed by this process . Also makes
* sure that the stream has been stopped .
*/
2011-09-02 00:18:44 +04:00
2011-09-02 00:18:41 +04:00
/*************************************************************************/
/**
* DRM_VMW_GET_3D_CAP
*
* Read 3 D capabilities from the FIFO
*
*/
/**
* struct drm_vmw_get_3d_cap_arg
*
2015-03-11 05:07:51 +03:00
* @ buffer : Pointer to a buffer for capability data , cast to an __u64
2011-09-02 00:18:41 +04:00
* @ size : Max size to copy
*
* Input argument to the DRM_VMW_GET_3D_CAP_IOCTL
* ioctls .
*/
struct drm_vmw_get_3d_cap_arg {
2015-03-11 05:07:51 +03:00
__u64 buffer ;
__u32 max_size ;
__u32 pad64 ;
2011-09-02 00:18:41 +04:00
} ;
2009-12-10 03:19:58 +03:00
2011-09-02 00:18:44 +04:00
/*************************************************************************/
/**
* DRM_VMW_FENCE_WAIT
*
* Waits for a fence object to signal . The wait is interruptible , so that
* signals may be delivered during the interrupt . The wait may timeout ,
* in which case the calls returns - EBUSY . If the wait is restarted ,
* that is restarting without resetting @ cookie_valid to zero ,
* the timeout is computed from the first call .
*
* The flags argument to the DRM_VMW_FENCE_WAIT ioctl indicates what to wait
* on :
* DRM_VMW_FENCE_FLAG_EXEC : All commands ahead of the fence in the command
* stream
* have executed .
* DRM_VMW_FENCE_FLAG_QUERY : All query results resulting from query finish
* commands
* in the buffer given to the EXECBUF ioctl returning the fence object handle
* are available to user - space .
*
* DRM_VMW_WAIT_OPTION_UNREF : If this wait option is given , and the
* fenc wait ioctl returns 0 , the fence object has been unreferenced after
* the wait .
*/
# define DRM_VMW_FENCE_FLAG_EXEC (1 << 0)
# define DRM_VMW_FENCE_FLAG_QUERY (1 << 1)
# define DRM_VMW_WAIT_OPTION_UNREF (1 << 0)
/**
* struct drm_vmw_fence_wait_arg
*
* @ handle : Fence object handle as returned by the DRM_VMW_EXECBUF ioctl .
* @ cookie_valid : Must be reset to 0 on first call . Left alone on restart .
* @ kernel_cookie : Set to 0 on first call . Left alone on restart .
* @ timeout_us : Wait timeout in microseconds . 0 for indefinite timeout .
* @ lazy : Set to 1 if timing is not critical . Allow more than a kernel tick
* before returning .
* @ flags : Fence flags to wait on .
* @ wait_options : Options that control the behaviour of the wait ioctl .
*
* Input argument to the DRM_VMW_FENCE_WAIT ioctl .
*/
struct drm_vmw_fence_wait_arg {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__s32 cookie_valid ;
__u64 kernel_cookie ;
__u64 timeout_us ;
__s32 lazy ;
__s32 flags ;
__s32 wait_options ;
__s32 pad64 ;
2011-09-02 00:18:44 +04:00
} ;
/*************************************************************************/
/**
* DRM_VMW_FENCE_SIGNALED
*
* Checks if a fence object is signaled . .
*/
/**
* struct drm_vmw_fence_signaled_arg
*
* @ handle : Fence object handle as returned by the DRM_VMW_EXECBUF ioctl .
* @ flags : Fence object flags input to DRM_VMW_FENCE_SIGNALED ioctl
* @ signaled : Out : Flags signaled .
* @ sequence : Out : Highest sequence passed so far . Can be used to signal the
* EXEC flag of user - space fence objects .
*
* Input / Output argument to the DRM_VMW_FENCE_SIGNALED and DRM_VMW_FENCE_UNREF
* ioctls .
*/
struct drm_vmw_fence_signaled_arg {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 flags ;
__s32 signaled ;
__u32 passed_seqno ;
__u32 signaled_flags ;
__u32 pad64 ;
2011-09-02 00:18:44 +04:00
} ;
/*************************************************************************/
/**
* DRM_VMW_FENCE_UNREF
*
* Unreferences a fence object , and causes it to be destroyed if there are no
* other references to it .
*
*/
/**
* struct drm_vmw_fence_arg
*
* @ handle : Fence object handle as returned by the DRM_VMW_EXECBUF ioctl .
*
* Input / Output argument to the DRM_VMW_FENCE_UNREF ioctl . .
*/
struct drm_vmw_fence_arg {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 pad64 ;
2011-09-02 00:18:44 +04:00
} ;
2011-10-10 14:23:26 +04:00
/*************************************************************************/
/**
* DRM_VMW_FENCE_EVENT
*
* Queues an event on a fence to be delivered on the drm character device
* when the fence has signaled the DRM_VMW_FENCE_FLAG_EXEC flag .
* Optionally the approximate time when the fence signaled is
* given by the event .
*/
/*
* The event type
*/
# define DRM_VMW_EVENT_FENCE_SIGNALED 0x80000000
struct drm_vmw_event_fence {
struct drm_event base ;
2015-03-11 05:07:51 +03:00
__u64 user_data ;
__u32 tv_sec ;
__u32 tv_usec ;
2011-10-10 14:23:26 +04:00
} ;
/*
* Flags that may be given to the command .
*/
/* Request fence signaled time on the event. */
# define DRM_VMW_FE_FLAG_REQ_TIME (1 << 0)
/**
* struct drm_vmw_fence_event_arg
*
2015-03-11 05:07:51 +03:00
* @ fence_rep : Pointer to fence_rep structure cast to __u64 or 0 if
2011-10-10 14:23:26 +04:00
* the fence is not supposed to be referenced by user - space .
* @ user_info : Info to be delivered with the event .
* @ handle : Attach the event to this fence only .
* @ flags : A set of flags as defined above .
*/
struct drm_vmw_fence_event_arg {
2015-03-11 05:07:51 +03:00
__u64 fence_rep ;
__u64 user_data ;
__u32 handle ;
__u32 flags ;
2011-10-10 14:23:26 +04:00
} ;
2011-10-04 22:13:26 +04:00
/*************************************************************************/
/**
* DRM_VMW_PRESENT
*
* Executes an SVGA present on a given fb for a given surface . The surface
* is placed on the framebuffer . Cliprects are given relative to the given
* point ( the point disignated by dest_ { x | y } ) .
*
*/
/**
* struct drm_vmw_present_arg
* @ fb_id : framebuffer id to present / read back from .
* @ sid : Surface id to present from .
* @ dest_x : X placement coordinate for surface .
* @ dest_y : Y placement coordinate for surface .
2015-03-11 05:07:51 +03:00
* @ clips_ptr : Pointer to an array of clip rects cast to an __u64 .
2011-10-04 22:13:26 +04:00
* @ num_clips : Number of cliprects given relative to the framebuffer origin ,
* in the same coordinate space as the frame buffer .
* @ pad64 : Unused 64 - bit padding .
*
* Input argument to the DRM_VMW_PRESENT ioctl .
*/
struct drm_vmw_present_arg {
2015-03-11 05:07:51 +03:00
__u32 fb_id ;
__u32 sid ;
__s32 dest_x ;
__s32 dest_y ;
__u64 clips_ptr ;
__u32 num_clips ;
__u32 pad64 ;
2011-10-04 22:13:26 +04:00
} ;
/*************************************************************************/
/**
* DRM_VMW_PRESENT_READBACK
*
* Executes an SVGA present readback from a given fb to the dma buffer
* currently bound as the fb . If there is no dma buffer bound to the fb ,
* an error will be returned .
*
*/
/**
* struct drm_vmw_present_arg
* @ fb_id : fb_id to present / read back from .
* @ num_clips : Number of cliprects .
2015-03-11 05:07:51 +03:00
* @ clips_ptr : Pointer to an array of clip rects cast to an __u64 .
* @ fence_rep : Pointer to a struct drm_vmw_fence_rep , cast to an __u64 .
2011-10-04 22:13:26 +04:00
* If this member is NULL , then the ioctl should not return a fence .
*/
struct drm_vmw_present_readback_arg {
2015-03-11 05:07:51 +03:00
__u32 fb_id ;
__u32 num_clips ;
__u64 clips_ptr ;
__u64 fence_rep ;
2011-10-04 22:13:26 +04:00
} ;
2011-10-26 01:35:53 +04:00
/*************************************************************************/
/**
* DRM_VMW_UPDATE_LAYOUT - Update layout
*
* Updates the preferred modes and connection status for connectors . The
* command consists of one drm_vmw_update_layout_arg pointing to an array
* of num_outputs drm_vmw_rect ' s .
*/
/**
* struct drm_vmw_update_layout_arg
*
* @ num_outputs : number of active connectors
2015-03-11 05:07:51 +03:00
* @ rects : pointer to array of drm_vmw_rect cast to an __u64
2011-10-26 01:35:53 +04:00
*
* Input argument to the DRM_VMW_UPDATE_LAYOUT Ioctl .
*/
struct drm_vmw_update_layout_arg {
2015-03-11 05:07:51 +03:00
__u32 num_outputs ;
__u32 pad64 ;
__u64 rects ;
2011-10-26 01:35:53 +04:00
} ;
2012-11-21 13:23:14 +04:00
/*************************************************************************/
/**
* DRM_VMW_CREATE_SHADER - Create shader
*
* Creates a shader and optionally binds it to a dma buffer containing
* the shader byte - code .
*/
/**
* enum drm_vmw_shader_type - Shader types
*/
enum drm_vmw_shader_type {
drm_vmw_shader_type_vs = 0 ,
drm_vmw_shader_type_ps ,
} ;
/**
* struct drm_vmw_shader_create_arg
*
* @ shader_type : Shader type of the shader to create .
* @ size : Size of the byte - code in bytes .
* where the shader byte - code starts
* @ buffer_handle : Buffer handle identifying the buffer containing the
* shader byte - code
* @ shader_handle : On successful completion contains a handle that
* can be used to subsequently identify the shader .
* @ offset : Offset in bytes into the buffer given by @ buffer_handle ,
*
* Input / Output argument to the DRM_VMW_CREATE_SHADER Ioctl .
*/
struct drm_vmw_shader_create_arg {
enum drm_vmw_shader_type shader_type ;
2015-03-11 05:07:51 +03:00
__u32 size ;
__u32 buffer_handle ;
__u32 shader_handle ;
__u64 offset ;
2012-11-21 13:23:14 +04:00
} ;
/*************************************************************************/
/**
* DRM_VMW_UNREF_SHADER - Unreferences a shader
*
* Destroys a user - space reference to a shader , optionally destroying
* it .
*/
/**
* struct drm_vmw_shader_arg
*
* @ handle : Handle identifying the shader to destroy .
*
* Input argument to the DRM_VMW_UNREF_SHADER ioctl .
*/
struct drm_vmw_shader_arg {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 pad64 ;
2012-11-21 13:23:14 +04:00
} ;
/*************************************************************************/
/**
* DRM_VMW_GB_SURFACE_CREATE - Create a host guest - backed surface .
*
* Allocates a surface handle and queues a create surface command
* for the host on the first use of the surface . The surface ID can
* be used as the surface ID in commands referencing the surface .
*/
/**
* enum drm_vmw_surface_flags
*
* @ drm_vmw_surface_flag_shareable : Whether the surface is shareable
* @ drm_vmw_surface_flag_scanout : Whether the surface is a scanout
* surface .
* @ drm_vmw_surface_flag_create_buffer : Create a backup buffer if none is
* given .
*/
enum drm_vmw_surface_flags {
drm_vmw_surface_flag_shareable = ( 1 < < 0 ) ,
drm_vmw_surface_flag_scanout = ( 1 < < 1 ) ,
drm_vmw_surface_flag_create_buffer = ( 1 < < 2 )
} ;
/**
* struct drm_vmw_gb_surface_create_req
*
* @ svga3d_flags : SVGA3d surface flags for the device .
* @ format : SVGA3d format .
* @ mip_level : Number of mip levels for all faces .
* @ drm_surface_flags Flags as described above .
2012-11-21 15:25:33 +04:00
* @ multisample_count Future use . Set to 0.
2012-11-21 13:23:14 +04:00
* @ autogen_filter Future use . Set to 0.
* @ buffer_handle Buffer handle of backup buffer . SVGA3D_INVALID_ID
* if none .
* @ base_size Size of the base mip level for all faces .
2015-08-10 20:39:35 +03:00
* @ array_size Must be zero for non - DX hardware , and if non - zero
* svga3d_flags must have proper bind flags setup .
2012-11-21 13:23:14 +04:00
*
* Input argument to the DRM_VMW_GB_SURFACE_CREATE Ioctl .
* Part of output argument for the DRM_VMW_GB_SURFACE_REF Ioctl .
*/
struct drm_vmw_gb_surface_create_req {
2015-03-11 05:07:51 +03:00
__u32 svga3d_flags ;
__u32 format ;
__u32 mip_levels ;
2012-11-21 13:23:14 +04:00
enum drm_vmw_surface_flags drm_surface_flags ;
2015-03-11 05:07:51 +03:00
__u32 multisample_count ;
__u32 autogen_filter ;
__u32 buffer_handle ;
__u32 array_size ;
2012-11-21 13:23:14 +04:00
struct drm_vmw_size base_size ;
} ;
/**
* struct drm_vmw_gb_surface_create_rep
*
* @ handle : Surface handle .
* @ backup_size : Size of backup buffers for this surface .
* @ buffer_handle : Handle of backup buffer . SVGA3D_INVALID_ID if none .
* @ buffer_size : Actual size of the buffer identified by
* @ buffer_handle
* @ buffer_map_handle : Offset into device address space for the buffer
* identified by @ buffer_handle .
*
* Part of output argument for the DRM_VMW_GB_SURFACE_REF ioctl .
* Output argument for the DRM_VMW_GB_SURFACE_CREATE ioctl .
*/
struct drm_vmw_gb_surface_create_rep {
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 backup_size ;
__u32 buffer_handle ;
__u32 buffer_size ;
__u64 buffer_map_handle ;
2012-11-21 13:23:14 +04:00
} ;
/**
* union drm_vmw_gb_surface_create_arg
*
* @ req : Input argument as described above .
* @ rep : Output argument as described above .
*
* Argument to the DRM_VMW_GB_SURFACE_CREATE ioctl .
*/
union drm_vmw_gb_surface_create_arg {
struct drm_vmw_gb_surface_create_rep rep ;
struct drm_vmw_gb_surface_create_req req ;
} ;
/*************************************************************************/
/**
* DRM_VMW_GB_SURFACE_REF - Reference a host surface .
*
* Puts a reference on a host surface with a given handle , as previously
* returned by the DRM_VMW_GB_SURFACE_CREATE ioctl .
* A reference will make sure the surface isn ' t destroyed while we hold
* it and will allow the calling client to use the surface handle in
* the command stream .
*
* On successful return , the Ioctl returns the surface information given
* to and returned from the DRM_VMW_GB_SURFACE_CREATE ioctl .
*/
/**
* struct drm_vmw_gb_surface_reference_arg
*
* @ creq : The data used as input when the surface was created , as described
* above at " struct drm_vmw_gb_surface_create_req "
* @ crep : Additional data output when the surface was created , as described
* above at " struct drm_vmw_gb_surface_create_rep "
*
* Output Argument to the DRM_VMW_GB_SURFACE_REF ioctl .
*/
struct drm_vmw_gb_surface_ref_rep {
struct drm_vmw_gb_surface_create_req creq ;
struct drm_vmw_gb_surface_create_rep crep ;
} ;
/**
* union drm_vmw_gb_surface_reference_arg
*
* @ req : Input data as described above at " struct drm_vmw_surface_arg "
* @ rep : Output data as described above at " struct drm_vmw_gb_surface_ref_rep "
*
* Argument to the DRM_VMW_GB_SURFACE_REF Ioctl .
*/
union drm_vmw_gb_surface_reference_arg {
struct drm_vmw_gb_surface_ref_rep rep ;
struct drm_vmw_surface_arg req ;
} ;
2012-11-21 15:32:19 +04:00
/*************************************************************************/
/**
* DRM_VMW_SYNCCPU - Sync a DMA buffer / MOB for CPU access .
*
* Idles any previously submitted GPU operations on the buffer and
* by default blocks command submissions that reference the buffer .
* If the file descriptor used to grab a blocking CPU sync is closed , the
* cpu sync is released .
* The flags argument indicates how the grab / release operation should be
* performed :
*/
/**
* enum drm_vmw_synccpu_flags - Synccpu flags :
*
* @ drm_vmw_synccpu_read : Sync for read . If sync is done for read only , it ' s a
* hint to the kernel to allow command submissions that references the buffer
* for read - only .
* @ drm_vmw_synccpu_write : Sync for write . Block all command submissions
* referencing this buffer .
* @ drm_vmw_synccpu_dontblock : Dont wait for GPU idle , but rather return
* - EBUSY should the buffer be busy .
* @ drm_vmw_synccpu_allow_cs : Allow command submission that touches the buffer
* while the buffer is synced for CPU . This is similar to the GEM bo idle
* behavior .
*/
enum drm_vmw_synccpu_flags {
drm_vmw_synccpu_read = ( 1 < < 0 ) ,
drm_vmw_synccpu_write = ( 1 < < 1 ) ,
drm_vmw_synccpu_dontblock = ( 1 < < 2 ) ,
drm_vmw_synccpu_allow_cs = ( 1 < < 3 )
} ;
/**
* enum drm_vmw_synccpu_op - Synccpu operations :
*
* @ drm_vmw_synccpu_grab : Grab the buffer for CPU operations
* @ drm_vmw_synccpu_release : Release a previous grab .
*/
enum drm_vmw_synccpu_op {
drm_vmw_synccpu_grab ,
drm_vmw_synccpu_release
} ;
/**
* struct drm_vmw_synccpu_arg
*
* @ op : The synccpu operation as described above .
* @ handle : Handle identifying the buffer object .
* @ flags : Flags as described above .
*/
struct drm_vmw_synccpu_arg {
enum drm_vmw_synccpu_op op ;
enum drm_vmw_synccpu_flags flags ;
2015-03-11 05:07:51 +03:00
__u32 handle ;
__u32 pad64 ;
2012-11-21 15:32:19 +04:00
} ;
2012-11-21 13:23:14 +04:00
2015-08-10 20:39:35 +03:00
/*************************************************************************/
/**
* DRM_VMW_CREATE_EXTENDED_CONTEXT - Create a host context .
*
* Allocates a device unique context id , and queues a create context command
* for the host . Does not wait for host completion .
*/
enum drm_vmw_extended_context {
drm_vmw_context_legacy ,
drm_vmw_context_dx
} ;
/**
* union drm_vmw_extended_context_arg
*
* @ req : Context type .
* @ rep : Context identifier .
*
* Argument to the DRM_VMW_CREATE_EXTENDED_CONTEXT Ioctl .
*/
union drm_vmw_extended_context_arg {
enum drm_vmw_extended_context req ;
struct drm_vmw_context_arg rep ;
} ;
2016-04-07 21:40:10 +03:00
2017-03-24 01:26:16 +03:00
/*************************************************************************/
/*
* DRM_VMW_HANDLE_CLOSE - Close a user - space handle and release its
* underlying resource .
*
2018-06-19 16:02:16 +03:00
* Note that this ioctl is overlaid on the deprecated DRM_VMW_UNREF_DMABUF
* Ioctl .
2017-03-24 01:26:16 +03:00
*/
/**
* struct drm_vmw_handle_close_arg
*
* @ handle : Handle to close .
*
* Argument to the DRM_VMW_HANDLE_CLOSE Ioctl .
*/
struct drm_vmw_handle_close_arg {
__u32 handle ;
__u32 pad64 ;
} ;
2018-06-19 16:02:16 +03:00
# define drm_vmw_unref_dmabuf_arg drm_vmw_handle_close_arg
2017-03-24 01:26:16 +03:00
2018-06-21 00:48:35 +03:00
/*************************************************************************/
/**
* DRM_VMW_GB_SURFACE_CREATE_EXT - Create a host guest - backed surface .
*
* Allocates a surface handle and queues a create surface command
* for the host on the first use of the surface . The surface ID can
* be used as the surface ID in commands referencing the surface .
*
* This new command extends DRM_VMW_GB_SURFACE_CREATE by adding version
* parameter and 64 bit svga flag .
*/
/**
* enum drm_vmw_surface_version
*
* @ drm_vmw_surface_gb_v1 : Corresponds to current gb surface format with
* svga3d surface flags split into 2 , upper half and lower half .
*/
enum drm_vmw_surface_version {
drm_vmw_gb_surface_v1
} ;
/**
* struct drm_vmw_gb_surface_create_ext_req
*
* @ base : Surface create parameters .
* @ version : Version of surface create ioctl .
* @ svga3d_flags_upper_32_bits : Upper 32 bits of svga3d flags .
* @ multisample_pattern : Multisampling pattern when msaa is supported .
* @ quality_level : Precision settings for each sample .
* @ must_be_zero : Reserved for future usage .
*
* Input argument to the DRM_VMW_GB_SURFACE_CREATE_EXT Ioctl .
* Part of output argument for the DRM_VMW_GB_SURFACE_REF_EXT Ioctl .
*/
struct drm_vmw_gb_surface_create_ext_req {
struct drm_vmw_gb_surface_create_req base ;
enum drm_vmw_surface_version version ;
uint32_t svga3d_flags_upper_32_bits ;
SVGA3dMSPattern multisample_pattern ;
SVGA3dMSQualityLevel quality_level ;
uint64_t must_be_zero ;
} ;
/**
* union drm_vmw_gb_surface_create_ext_arg
*
* @ req : Input argument as described above .
* @ rep : Output argument as described above .
*
* Argument to the DRM_VMW_GB_SURFACE_CREATE_EXT ioctl .
*/
union drm_vmw_gb_surface_create_ext_arg {
struct drm_vmw_gb_surface_create_rep rep ;
struct drm_vmw_gb_surface_create_ext_req req ;
} ;
/*************************************************************************/
/**
* DRM_VMW_GB_SURFACE_REF_EXT - Reference a host surface .
*
* Puts a reference on a host surface with a given handle , as previously
* returned by the DRM_VMW_GB_SURFACE_CREATE_EXT ioctl .
* A reference will make sure the surface isn ' t destroyed while we hold
* it and will allow the calling client to use the surface handle in
* the command stream .
*
* On successful return , the Ioctl returns the surface information given
* to and returned from the DRM_VMW_GB_SURFACE_CREATE_EXT ioctl .
*/
/**
* struct drm_vmw_gb_surface_ref_ext_rep
*
* @ creq : The data used as input when the surface was created , as described
* above at " struct drm_vmw_gb_surface_create_ext_req "
* @ crep : Additional data output when the surface was created , as described
* above at " struct drm_vmw_gb_surface_create_rep "
*
* Output Argument to the DRM_VMW_GB_SURFACE_REF_EXT ioctl .
*/
struct drm_vmw_gb_surface_ref_ext_rep {
struct drm_vmw_gb_surface_create_ext_req creq ;
struct drm_vmw_gb_surface_create_rep crep ;
} ;
/**
* union drm_vmw_gb_surface_reference_ext_arg
*
* @ req : Input data as described above at " struct drm_vmw_surface_arg "
* @ rep : Output data as described above at
* " struct drm_vmw_gb_surface_ref_ext_rep "
*
* Argument to the DRM_VMW_GB_SURFACE_REF Ioctl .
*/
union drm_vmw_gb_surface_reference_ext_arg {
struct drm_vmw_gb_surface_ref_ext_rep rep ;
struct drm_vmw_surface_arg req ;
} ;
2016-04-07 21:40:10 +03:00
# if defined(__cplusplus)
}
# endif
2009-12-10 03:19:58 +03:00
# endif