2013-07-19 20:59:32 +04:00
/*
* Copyright ( C ) 2013 Red Hat
* Author : Rob Clark < robdclark @ gmail . com >
*
2016-10-20 00:53:53 +03:00
* Permission is hereby granted , free of charge , to any person obtaining a
* copy of this software and associated documentation files ( the " Software " ) ,
* to deal in the Software without restriction , including without limitation
* the rights to use , copy , modify , merge , publish , distribute , sublicense ,
* and / or sell copies of the Software , and to permit persons to whom the
* Software is furnished to do so , subject to the following conditions :
2013-07-19 20:59:32 +04:00
*
2016-10-20 00:53:53 +03:00
* The above copyright notice and this permission notice ( including the next
* paragraph ) shall be included in all copies or substantial portions of the
* Software .
2013-07-19 20:59:32 +04:00
*
2016-10-20 00:53:53 +03:00
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND , EXPRESS OR
* IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY ,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT . IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM , DAMAGES OR OTHER
* LIABILITY , WHETHER IN AN ACTION OF CONTRACT , TORT OR OTHERWISE , ARISING FROM ,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE .
2013-07-19 20:59:32 +04:00
*/
# ifndef __MSM_DRM_H__
# define __MSM_DRM_H__
2015-11-30 17:10:49 +03:00
# include "drm.h"
2013-07-19 20:59:32 +04:00
2016-04-07 21:03:46 +03:00
# if defined(__cplusplus)
extern " C " {
# endif
2013-07-19 20:59:32 +04:00
/* Please note that modifications to all structs defined here are
* subject to backwards - compatibility constraints :
2015-05-30 18:38:08 +03:00
* 1 ) Do not use pointers , use __u64 instead for 32 bit / 64 bit
2013-07-19 20:59:32 +04:00
* user / kernel compatibility
* 2 ) Keep fields aligned to their size
* 3 ) Because of how drm_ioctl ( ) works , we can add new fields at
* the end of an ioctl if some care is taken : drm_ioctl ( ) will
* zero out the new fields at the tail of the ioctl , so a zero
* value should have a backwards compatible meaning . And for
* output params , userspace won ' t see the newly added output
* fields . . so that has to be somehow ok .
*/
# define MSM_PIPE_NONE 0x00
# define MSM_PIPE_2D0 0x01
# define MSM_PIPE_2D1 0x02
# define MSM_PIPE_3D0 0x10
2016-04-23 17:08:59 +03:00
/* The pipe-id just uses the lower bits, so can be OR'd with flags in
* the upper 16 bits ( which could be extended further , if needed , maybe
* we extend / overload the pipe - id some day to deal with multiple rings ,
* but even then I don ' t think we need the full lower 16 bits ) .
*/
# define MSM_PIPE_ID_MASK 0xffff
# define MSM_PIPE_ID(x) ((x) & MSM_PIPE_ID_MASK)
# define MSM_PIPE_FLAGS(x) ((x) & ~MSM_PIPE_ID_MASK)
2013-07-19 20:59:32 +04:00
/* timeouts are specified in clock-monotonic absolute times (to simplify
* restarting interrupted ioctls ) . The following struct is logically the
* same as ' struct timespec ' but 32 / 64 b ABI safe .
*/
struct drm_msm_timespec {
2015-05-30 18:38:08 +03:00
__s64 tv_sec ; /* seconds */
__s64 tv_nsec ; /* nanoseconds */
2013-07-19 20:59:32 +04:00
} ;
# define MSM_PARAM_GPU_ID 0x01
# define MSM_PARAM_GMEM_SIZE 0x02
2014-02-04 23:16:04 +04:00
# define MSM_PARAM_CHIP_ID 0x03
2016-02-09 20:05:30 +03:00
# define MSM_PARAM_MAX_FREQ 0x04
2016-02-22 14:26:21 +03:00
# define MSM_PARAM_TIMESTAMP 0x05
2013-07-19 20:59:32 +04:00
struct drm_msm_param {
2015-05-30 18:38:08 +03:00
__u32 pipe ; /* in, MSM_PIPE_x */
__u32 param ; /* in, MSM_PARAM_x */
__u64 value ; /* out (get_param) or in (set_param) */
2013-07-19 20:59:32 +04:00
} ;
/*
* GEM buffers :
*/
# define MSM_BO_SCANOUT 0x00000001 /* scanout capable */
# define MSM_BO_GPU_READONLY 0x00000002
# define MSM_BO_CACHE_MASK 0x000f0000
/* cache modes */
# define MSM_BO_CACHED 0x00010000
# define MSM_BO_WC 0x00020000
# define MSM_BO_UNCACHED 0x00040000
2014-03-03 18:42:33 +04:00
# define MSM_BO_FLAGS (MSM_BO_SCANOUT | \
MSM_BO_GPU_READONLY | \
MSM_BO_CACHED | \
MSM_BO_WC | \
MSM_BO_UNCACHED )
2013-07-19 20:59:32 +04:00
struct drm_msm_gem_new {
2015-05-30 18:38:08 +03:00
__u64 size ; /* in */
__u32 flags ; /* in, mask of MSM_BO_x */
__u32 handle ; /* out */
2013-07-19 20:59:32 +04:00
} ;
struct drm_msm_gem_info {
2015-05-30 18:38:08 +03:00
__u32 handle ; /* in */
__u32 pad ;
__u64 offset ; /* out, offset to pass to mmap() */
2013-07-19 20:59:32 +04:00
} ;
# define MSM_PREP_READ 0x01
# define MSM_PREP_WRITE 0x02
# define MSM_PREP_NOSYNC 0x04
2014-03-03 18:42:33 +04:00
# define MSM_PREP_FLAGS (MSM_PREP_READ | MSM_PREP_WRITE | MSM_PREP_NOSYNC)
2013-07-19 20:59:32 +04:00
struct drm_msm_gem_cpu_prep {
2015-05-30 18:38:08 +03:00
__u32 handle ; /* in */
__u32 op ; /* in, mask of MSM_PREP_x */
2013-07-19 20:59:32 +04:00
struct drm_msm_timespec timeout ; /* in */
} ;
struct drm_msm_gem_cpu_fini {
2015-05-30 18:38:08 +03:00
__u32 handle ; /* in */
2013-07-19 20:59:32 +04:00
} ;
/*
* Cmdstream Submission :
*/
/* The value written into the cmdstream is logically:
*
* ( ( relocbuf - > gpuaddr + reloc_offset ) < < shift ) | or
*
* When we have GPU ' s w / > 32 bit ptrs , it should be possible to deal
* with this by emit ' ing two reloc entries with appropriate shift
* values . Or a new MSM_SUBMIT_CMD_x type would also be an option .
*
* NOTE that reloc ' s must be sorted by order of increasing submit_offset ,
* otherwise EINVAL .
*/
struct drm_msm_gem_submit_reloc {
2015-05-30 18:38:08 +03:00
__u32 submit_offset ; /* in, offset from submit_bo */
__u32 or ; /* in, value OR'd with result */
2015-12-14 17:59:56 +03:00
__s32 shift ; /* in, amount of left shift (can be negative) */
2015-05-30 18:38:08 +03:00
__u32 reloc_idx ; /* in, index of reloc_bo buffer */
__u64 reloc_offset ; /* in, offset from start of reloc_bo */
2013-07-19 20:59:32 +04:00
} ;
/* submit-types:
* BUF - this cmd buffer is executed normally .
* IB_TARGET_BUF - this cmd buffer is an IB target . Reloc ' s are
* processed normally , but the kernel does not setup an IB to
* this buffer in the first - level ringbuffer
* CTX_RESTORE_BUF - only executed if there has been a GPU context
* switch since the last SUBMIT ioctl
*/
# define MSM_SUBMIT_CMD_BUF 0x0001
# define MSM_SUBMIT_CMD_IB_TARGET_BUF 0x0002
# define MSM_SUBMIT_CMD_CTX_RESTORE_BUF 0x0003
struct drm_msm_gem_submit_cmd {
2015-05-30 18:38:08 +03:00
__u32 type ; /* in, one of MSM_SUBMIT_CMD_x */
__u32 submit_idx ; /* in, index of submit_bo cmdstream buffer */
__u32 submit_offset ; /* in, offset into submit_bo */
__u32 size ; /* in, cmdstream size */
__u32 pad ;
__u32 nr_relocs ; /* in, number of submit_reloc's */
__u64 __user relocs ; /* in, ptr to array of submit_reloc's */
2013-07-19 20:59:32 +04:00
} ;
/* Each buffer referenced elsewhere in the cmdstream submit (ie. the
* cmdstream buffer ( s ) themselves or reloc entries ) has one ( and only
* one ) entry in the submit - > bos [ ] table .
*
* As a optimization , the current buffer ( gpu virtual address ) can be
* passed back through the ' presumed ' field . If on a subsequent reloc ,
* userspace passes back a ' presumed ' address that is still valid ,
* then patching the cmdstream for this entry is skipped . This can
* avoid kernel needing to map / access the cmdstream bo in the common
* case .
*/
# define MSM_SUBMIT_BO_READ 0x0001
# define MSM_SUBMIT_BO_WRITE 0x0002
2014-03-03 18:42:33 +04:00
# define MSM_SUBMIT_BO_FLAGS (MSM_SUBMIT_BO_READ | MSM_SUBMIT_BO_WRITE)
2013-07-19 20:59:32 +04:00
struct drm_msm_gem_submit_bo {
2015-05-30 18:38:08 +03:00
__u32 flags ; /* in, mask of MSM_SUBMIT_BO_x */
__u32 handle ; /* in, GEM handle */
__u64 presumed ; /* in/out, presumed buffer address */
2013-07-19 20:59:32 +04:00
} ;
2016-04-23 17:08:59 +03:00
/* Valid submit ioctl flags: */
2016-06-16 23:08:19 +03:00
# define MSM_SUBMIT_NO_IMPLICIT 0x80000000 /* disable implicit sync */
# define MSM_SUBMIT_FENCE_FD_IN 0x40000000 /* enable input fence_fd */
2016-06-16 23:43:49 +03:00
# define MSM_SUBMIT_FENCE_FD_OUT 0x20000000 /* enable output fence_fd */
2016-06-16 23:08:19 +03:00
# define MSM_SUBMIT_FLAGS ( \
MSM_SUBMIT_NO_IMPLICIT | \
MSM_SUBMIT_FENCE_FD_IN | \
2016-06-16 23:43:49 +03:00
MSM_SUBMIT_FENCE_FD_OUT | \
2016-06-16 23:08:19 +03:00
0 )
2016-04-23 17:08:59 +03:00
2013-07-19 20:59:32 +04:00
/* Each cmdstream submit consists of a table of buffers involved, and
* one or more cmdstream buffers . This allows for conditional execution
* ( context - restore ) , and IB buffers needed for per tile / bin draw cmds .
*/
struct drm_msm_gem_submit {
2016-04-23 17:08:59 +03:00
__u32 flags ; /* MSM_PIPE_x | MSM_SUBMIT_x */
2015-05-30 18:38:08 +03:00
__u32 fence ; /* out */
__u32 nr_bos ; /* in, number of submit_bo's */
__u32 nr_cmds ; /* in, number of submit_cmd's */
__u64 __user bos ; /* in, ptr to array of submit_bo's */
__u64 __user cmds ; /* in, ptr to array of submit_cmd's */
2016-06-16 23:43:49 +03:00
__s32 fence_fd ; /* in/out fence fd (see MSM_SUBMIT_FENCE_FD_IN/OUT) */
2013-07-19 20:59:32 +04:00
} ;
/* The normal way to synchronize with the GPU is just to CPU_PREP on
* a buffer if you need to access it from the CPU ( other cmdstream
* submission from same or other contexts , PAGE_FLIP ioctl , etc , all
* handle the required synchronization under the hood ) . This ioctl
* mainly just exists as a way to implement the gallium pipe_fence
* APIs without requiring a dummy bo to synchronize on .
*/
struct drm_msm_wait_fence {
2015-05-30 18:38:08 +03:00
__u32 fence ; /* in */
__u32 pad ;
2013-07-19 20:59:32 +04:00
struct drm_msm_timespec timeout ; /* in */
} ;
2016-05-17 22:44:49 +03:00
/* madvise provides a way to tell the kernel in case a buffers contents
* can be discarded under memory pressure , which is useful for userspace
* bo cache where we want to optimistically hold on to buffer allocate
* and potential mmap , but allow the pages to be discarded under memory
* pressure .
*
* Typical usage would involve madvise ( DONTNEED ) when buffer enters BO
* cache , and madvise ( WILLNEED ) if trying to recycle buffer from BO cache .
* In the WILLNEED case , ' retained ' indicates to userspace whether the
* backing pages still exist .
*/
# define MSM_MADV_WILLNEED 0 /* backing pages are needed, status returned in 'retained' */
# define MSM_MADV_DONTNEED 1 /* backing pages not needed */
# define __MSM_MADV_PURGED 2 /* internal state */
struct drm_msm_gem_madvise {
__u32 handle ; /* in, GEM handle */
__u32 madv ; /* in, MSM_MADV_x */
__u32 retained ; /* out, whether backing store still exists */
} ;
2013-07-19 20:59:32 +04:00
# define DRM_MSM_GET_PARAM 0x00
/* placeholder:
# define DRM_MSM_SET_PARAM 0x01
*/
# define DRM_MSM_GEM_NEW 0x02
# define DRM_MSM_GEM_INFO 0x03
# define DRM_MSM_GEM_CPU_PREP 0x04
# define DRM_MSM_GEM_CPU_FINI 0x05
# define DRM_MSM_GEM_SUBMIT 0x06
# define DRM_MSM_WAIT_FENCE 0x07
2016-05-17 22:44:49 +03:00
# define DRM_MSM_GEM_MADVISE 0x08
# define DRM_MSM_NUM_IOCTLS 0x09
2013-07-19 20:59:32 +04:00
# define DRM_IOCTL_MSM_GET_PARAM DRM_IOWR(DRM_COMMAND_BASE + DRM_MSM_GET_PARAM, struct drm_msm_param)
# define DRM_IOCTL_MSM_GEM_NEW DRM_IOWR(DRM_COMMAND_BASE + DRM_MSM_GEM_NEW, struct drm_msm_gem_new)
# define DRM_IOCTL_MSM_GEM_INFO DRM_IOWR(DRM_COMMAND_BASE + DRM_MSM_GEM_INFO, struct drm_msm_gem_info)
# define DRM_IOCTL_MSM_GEM_CPU_PREP DRM_IOW (DRM_COMMAND_BASE + DRM_MSM_GEM_CPU_PREP, struct drm_msm_gem_cpu_prep)
# define DRM_IOCTL_MSM_GEM_CPU_FINI DRM_IOW (DRM_COMMAND_BASE + DRM_MSM_GEM_CPU_FINI, struct drm_msm_gem_cpu_fini)
# define DRM_IOCTL_MSM_GEM_SUBMIT DRM_IOWR(DRM_COMMAND_BASE + DRM_MSM_GEM_SUBMIT, struct drm_msm_gem_submit)
# define DRM_IOCTL_MSM_WAIT_FENCE DRM_IOW (DRM_COMMAND_BASE + DRM_MSM_WAIT_FENCE, struct drm_msm_wait_fence)
2016-05-17 22:44:49 +03:00
# define DRM_IOCTL_MSM_GEM_MADVISE DRM_IOWR(DRM_COMMAND_BASE + DRM_MSM_GEM_MADVISE, struct drm_msm_gem_madvise)
2013-07-19 20:59:32 +04:00
2016-04-07 21:03:46 +03:00
# if defined(__cplusplus)
}
# endif
2013-07-19 20:59:32 +04:00
# endif /* __MSM_DRM_H__ */