Home | History | Annotate | Download | only in linux
      1 #ifndef _UAPI_MSM_KGSL_H
      2 #define _UAPI_MSM_KGSL_H
      3 
      4 #include <linux/types.h>
      5 #include <linux/ioctl.h>
      6 
      7 /*
      8  * The KGSL version has proven not to be very useful in userspace if features
      9  * are cherry picked into other trees out of order so it is frozen as of 3.14.
     10  * It is left here for backwards compatabilty and as a reminder that
     11  * software releases are never linear. Also, I like pie.
     12  */
     13 
     14 #define KGSL_VERSION_MAJOR        3
     15 #define KGSL_VERSION_MINOR        14
     16 
     17 /*
     18  * We have traditionally mixed context and issueibcmds / command batch flags
     19  * together into a big flag stew. This worked fine until we started adding a
     20  * lot more command batch flags and we started running out of bits. Turns out
     21  * we have a bit of room in the context type / priority mask that we could use
     22  * for command batches, but that means we need to split out the flags into two
     23  * coherent sets.
     24  *
     25  * If any future definitions are for both context and cmdbatch add both defines
     26  * and link the cmdbatch to the context define as we do below. Otherwise feel
     27  * free to add exclusive bits to either set.
     28  */
     29 
     30 /* --- context flags --- */
     31 #define KGSL_CONTEXT_SAVE_GMEM		0x00000001
     32 #define KGSL_CONTEXT_NO_GMEM_ALLOC	0x00000002
     33 /* This is a cmdbatch exclusive flag - use the CMDBATCH equivalent instead */
     34 #define KGSL_CONTEXT_SUBMIT_IB_LIST	0x00000004
     35 #define KGSL_CONTEXT_CTX_SWITCH		0x00000008
     36 #define KGSL_CONTEXT_PREAMBLE		0x00000010
     37 #define KGSL_CONTEXT_TRASH_STATE	0x00000020
     38 #define KGSL_CONTEXT_PER_CONTEXT_TS	0x00000040
     39 #define KGSL_CONTEXT_USER_GENERATED_TS	0x00000080
     40 /* This is a cmdbatch exclusive flag - use the CMDBATCH equivalent instead */
     41 #define KGSL_CONTEXT_END_OF_FRAME	0x00000100
     42 #define KGSL_CONTEXT_NO_FAULT_TOLERANCE 0x00000200
     43 /* This is a cmdbatch exclusive flag - use the CMDBATCH equivalent instead */
     44 #define KGSL_CONTEXT_SYNC               0x00000400
     45 #define KGSL_CONTEXT_PWR_CONSTRAINT     0x00000800
     46 #define KGSL_CONTEXT_PRIORITY_MASK      0x0000F000
     47 #define KGSL_CONTEXT_PRIORITY_SHIFT     12
     48 #define KGSL_CONTEXT_PRIORITY_UNDEF     0
     49 
     50 #define KGSL_CONTEXT_IFH_NOP            0x00010000
     51 #define KGSL_CONTEXT_SECURE             0x00020000
     52 #define KGSL_CONTEXT_NO_SNAPSHOT        0x00040000
     53 
     54 #define KGSL_CONTEXT_PREEMPT_STYLE_MASK       0x0E000000
     55 #define KGSL_CONTEXT_PREEMPT_STYLE_SHIFT      25
     56 #define KGSL_CONTEXT_PREEMPT_STYLE_DEFAULT    0x0
     57 #define KGSL_CONTEXT_PREEMPT_STYLE_RINGBUFFER 0x1
     58 #define KGSL_CONTEXT_PREEMPT_STYLE_FINEGRAIN  0x2
     59 
     60 #define KGSL_CONTEXT_TYPE_MASK          0x01F00000
     61 #define KGSL_CONTEXT_TYPE_SHIFT         20
     62 #define KGSL_CONTEXT_TYPE_ANY		0
     63 #define KGSL_CONTEXT_TYPE_GL		1
     64 #define KGSL_CONTEXT_TYPE_CL		2
     65 #define KGSL_CONTEXT_TYPE_C2D		3
     66 #define KGSL_CONTEXT_TYPE_RS		4
     67 #define KGSL_CONTEXT_TYPE_UNKNOWN	0x1E
     68 
     69 #define KGSL_CONTEXT_INVALID 0xffffffff
     70 
     71 /*
     72  * --- command batch flags ---
     73  * The bits that are linked to a KGSL_CONTEXT equivalent are either legacy
     74  * definitions or bits that are valid for both contexts and cmdbatches.  To be
     75  * safe the other 8 bits that are still available in the context field should be
     76  * omitted here in case we need to share - the other bits are available for
     77  * cmdbatch only flags as needed
     78  */
     79 #define KGSL_CMDBATCH_MEMLIST		0x00000001
     80 #define KGSL_CMDBATCH_MARKER		0x00000002
     81 #define KGSL_CMDBATCH_SUBMIT_IB_LIST	KGSL_CONTEXT_SUBMIT_IB_LIST /* 0x004 */
     82 #define KGSL_CMDBATCH_CTX_SWITCH	KGSL_CONTEXT_CTX_SWITCH     /* 0x008 */
     83 #define KGSL_CMDBATCH_PROFILING		0x00000010
     84 /*
     85  * KGSL_CMDBATCH_PROFILING must also be set for KGSL_CMDBATCH_PROFILING_KTIME
     86  * to take effect, as the latter only affects the time data returned.
     87  */
     88 #define KGSL_CMDBATCH_PROFILING_KTIME	0x00000020
     89 #define KGSL_CMDBATCH_END_OF_FRAME	KGSL_CONTEXT_END_OF_FRAME   /* 0x100 */
     90 #define KGSL_CMDBATCH_SYNC		KGSL_CONTEXT_SYNC           /* 0x400 */
     91 #define KGSL_CMDBATCH_PWR_CONSTRAINT	KGSL_CONTEXT_PWR_CONSTRAINT /* 0x800 */
     92 
     93 /*
     94  * Reserve bits [16:19] and bits [28:31] for possible bits shared between
     95  * contexts and command batches.  Update this comment as new flags are added.
     96  */
     97 
     98 /*
     99  * gpu_command_object flags - these flags communicate the type of command or
    100  * memory object being submitted for a GPU command
    101  */
    102 
    103 /* Flags for GPU command objects */
    104 #define KGSL_CMDLIST_IB                  0x00000001U
    105 #define KGSL_CMDLIST_CTXTSWITCH_PREAMBLE 0x00000002U
    106 #define KGSL_CMDLIST_IB_PREAMBLE         0x00000004U
    107 
    108 /* Flags for GPU command memory objects */
    109 #define KGSL_OBJLIST_MEMOBJ  0x00000008U
    110 #define KGSL_OBJLIST_PROFILE 0x00000010U
    111 
    112 /* Flags for GPU command sync points */
    113 #define KGSL_CMD_SYNCPOINT_TYPE_TIMESTAMP 0
    114 #define KGSL_CMD_SYNCPOINT_TYPE_FENCE 1
    115 
    116 /* --- Memory allocation flags --- */
    117 
    118 /* General allocation hints */
    119 #define KGSL_MEMFLAGS_SECURE      0x00000008ULL
    120 #define KGSL_MEMFLAGS_GPUREADONLY 0x01000000U
    121 #define KGSL_MEMFLAGS_GPUWRITEONLY 0x02000000U
    122 #define KGSL_MEMFLAGS_FORCE_32BIT 0x100000000ULL
    123 
    124 /* Memory caching hints */
    125 #define KGSL_CACHEMODE_MASK       0x0C000000U
    126 #define KGSL_CACHEMODE_SHIFT 26
    127 
    128 #define KGSL_CACHEMODE_WRITECOMBINE 0
    129 #define KGSL_CACHEMODE_UNCACHED 1
    130 #define KGSL_CACHEMODE_WRITETHROUGH 2
    131 #define KGSL_CACHEMODE_WRITEBACK 3
    132 
    133 #define KGSL_MEMFLAGS_USE_CPU_MAP 0x10000000ULL
    134 
    135 /* Memory types for which allocations are made */
    136 #define KGSL_MEMTYPE_MASK		0x0000FF00
    137 #define KGSL_MEMTYPE_SHIFT		8
    138 
    139 #define KGSL_MEMTYPE_OBJECTANY			0
    140 #define KGSL_MEMTYPE_FRAMEBUFFER		1
    141 #define KGSL_MEMTYPE_RENDERBUFFER		2
    142 #define KGSL_MEMTYPE_ARRAYBUFFER		3
    143 #define KGSL_MEMTYPE_ELEMENTARRAYBUFFER		4
    144 #define KGSL_MEMTYPE_VERTEXARRAYBUFFER		5
    145 #define KGSL_MEMTYPE_TEXTURE			6
    146 #define KGSL_MEMTYPE_SURFACE			7
    147 #define KGSL_MEMTYPE_EGL_SURFACE		8
    148 #define KGSL_MEMTYPE_GL				9
    149 #define KGSL_MEMTYPE_CL				10
    150 #define KGSL_MEMTYPE_CL_BUFFER_MAP		11
    151 #define KGSL_MEMTYPE_CL_BUFFER_NOMAP		12
    152 #define KGSL_MEMTYPE_CL_IMAGE_MAP		13
    153 #define KGSL_MEMTYPE_CL_IMAGE_NOMAP		14
    154 #define KGSL_MEMTYPE_CL_KERNEL_STACK		15
    155 #define KGSL_MEMTYPE_COMMAND			16
    156 #define KGSL_MEMTYPE_2D				17
    157 #define KGSL_MEMTYPE_EGL_IMAGE			18
    158 #define KGSL_MEMTYPE_EGL_SHADOW			19
    159 #define KGSL_MEMTYPE_MULTISAMPLE		20
    160 #define KGSL_MEMTYPE_KERNEL			255
    161 
    162 /*
    163  * Alignment hint, passed as the power of 2 exponent.
    164  * i.e 4k (2^12) would be 12, 64k (2^16)would be 16.
    165  */
    166 #define KGSL_MEMALIGN_MASK		0x00FF0000
    167 #define KGSL_MEMALIGN_SHIFT		16
    168 
    169 enum kgsl_user_mem_type {
    170 	KGSL_USER_MEM_TYPE_PMEM		= 0x00000000,
    171 	KGSL_USER_MEM_TYPE_ASHMEM	= 0x00000001,
    172 	KGSL_USER_MEM_TYPE_ADDR		= 0x00000002,
    173 	KGSL_USER_MEM_TYPE_ION		= 0x00000003,
    174 	/*
    175 	 * ION type is retained for backwards compatibilty but Ion buffers are
    176 	 * dma-bufs so try to use that naming if we can
    177 	 */
    178 	KGSL_USER_MEM_TYPE_DMABUF       = 0x00000003,
    179 	KGSL_USER_MEM_TYPE_MAX		= 0x00000007,
    180 };
    181 #define KGSL_MEMFLAGS_USERMEM_MASK 0x000000e0
    182 #define KGSL_MEMFLAGS_USERMEM_SHIFT 5
    183 
    184 /*
    185  * Unfortunately, enum kgsl_user_mem_type starts at 0 which does not
    186  * leave a good value for allocated memory. In the flags we use
    187  * 0 to indicate allocated memory and thus need to add 1 to the enum
    188  * values.
    189  */
    190 #define KGSL_USERMEM_FLAG(x) (((x) + 1) << KGSL_MEMFLAGS_USERMEM_SHIFT)
    191 
    192 #define KGSL_MEMFLAGS_NOT_USERMEM 0
    193 #define KGSL_MEMFLAGS_USERMEM_PMEM KGSL_USERMEM_FLAG(KGSL_USER_MEM_TYPE_PMEM)
    194 #define KGSL_MEMFLAGS_USERMEM_ASHMEM \
    195 		KGSL_USERMEM_FLAG(KGSL_USER_MEM_TYPE_ASHMEM)
    196 #define KGSL_MEMFLAGS_USERMEM_ADDR KGSL_USERMEM_FLAG(KGSL_USER_MEM_TYPE_ADDR)
    197 #define KGSL_MEMFLAGS_USERMEM_ION KGSL_USERMEM_FLAG(KGSL_USER_MEM_TYPE_ION)
    198 
    199 /* --- generic KGSL flag values --- */
    200 
    201 #define KGSL_FLAGS_NORMALMODE  0x00000000
    202 #define KGSL_FLAGS_SAFEMODE    0x00000001
    203 #define KGSL_FLAGS_INITIALIZED0 0x00000002
    204 #define KGSL_FLAGS_INITIALIZED 0x00000004
    205 #define KGSL_FLAGS_STARTED     0x00000008
    206 #define KGSL_FLAGS_ACTIVE      0x00000010
    207 #define KGSL_FLAGS_RESERVED0   0x00000020
    208 #define KGSL_FLAGS_RESERVED1   0x00000040
    209 #define KGSL_FLAGS_RESERVED2   0x00000080
    210 #define KGSL_FLAGS_SOFT_RESET  0x00000100
    211 #define KGSL_FLAGS_PER_CONTEXT_TIMESTAMPS 0x00000200
    212 
    213 /* Server Side Sync Timeout in milliseconds */
    214 #define KGSL_SYNCOBJ_SERVER_TIMEOUT 2000
    215 
    216 /*
    217  * Reset status values for context
    218  */
    219 enum kgsl_ctx_reset_stat {
    220 	KGSL_CTX_STAT_NO_ERROR				= 0x00000000,
    221 	KGSL_CTX_STAT_GUILTY_CONTEXT_RESET_EXT		= 0x00000001,
    222 	KGSL_CTX_STAT_INNOCENT_CONTEXT_RESET_EXT	= 0x00000002,
    223 	KGSL_CTX_STAT_UNKNOWN_CONTEXT_RESET_EXT		= 0x00000003
    224 };
    225 
    226 #define KGSL_CONVERT_TO_MBPS(val) \
    227 	(val*1000*1000U)
    228 
    229 /* device id */
    230 enum kgsl_deviceid {
    231 	KGSL_DEVICE_3D0		= 0x00000000,
    232 	KGSL_DEVICE_MAX
    233 };
    234 
    235 struct kgsl_devinfo {
    236 
    237 	unsigned int device_id;
    238 	/* chip revision id
    239 	* coreid:8 majorrev:8 minorrev:8 patch:8
    240 	*/
    241 	unsigned int chip_id;
    242 	unsigned int mmu_enabled;
    243 	unsigned long gmem_gpubaseaddr;
    244 	/*
    245 	* This field contains the adreno revision
    246 	* number 200, 205, 220, etc...
    247 	*/
    248 	unsigned int gpu_id;
    249 	size_t gmem_sizebytes;
    250 };
    251 
    252 /*
    253  * struct kgsl_devmemstore - this structure defines the region of memory
    254  * that can be mmap()ed from this driver. The timestamp fields are volatile
    255  * because they are written by the GPU
    256  * @soptimestamp: Start of pipeline timestamp written by GPU before the
    257  * commands in concern are processed
    258  * @sbz: Unused, kept for 8 byte alignment
    259  * @eoptimestamp: End of pipeline timestamp written by GPU after the
    260  * commands in concern are processed
    261  * @sbz2: Unused, kept for 8 byte alignment
    262  * @preempted: Indicates if the context was preempted
    263  * @sbz3: Unused, kept for 8 byte alignment
    264  * @ref_wait_ts: Timestamp on which to generate interrupt, unused now.
    265  * @sbz4: Unused, kept for 8 byte alignment
    266  * @current_context: The current context the GPU is working on
    267  * @sbz5: Unused, kept for 8 byte alignment
    268  */
    269 struct kgsl_devmemstore {
    270 	volatile unsigned int soptimestamp;
    271 	unsigned int sbz;
    272 	volatile unsigned int eoptimestamp;
    273 	unsigned int sbz2;
    274 	volatile unsigned int preempted;
    275 	unsigned int sbz3;
    276 	volatile unsigned int ref_wait_ts;
    277 	unsigned int sbz4;
    278 	unsigned int current_context;
    279 	unsigned int sbz5;
    280 };
    281 
    282 #define KGSL_MEMSTORE_OFFSET(ctxt_id, field) \
    283 	((ctxt_id)*sizeof(struct kgsl_devmemstore) + \
    284 	 offsetof(struct kgsl_devmemstore, field))
    285 
    286 /* timestamp id*/
    287 enum kgsl_timestamp_type {
    288 	KGSL_TIMESTAMP_CONSUMED = 0x00000001, /* start-of-pipeline timestamp */
    289 	KGSL_TIMESTAMP_RETIRED  = 0x00000002, /* end-of-pipeline timestamp*/
    290 	KGSL_TIMESTAMP_QUEUED   = 0x00000003,
    291 };
    292 
    293 /* property types - used with kgsl_device_getproperty */
    294 #define KGSL_PROP_DEVICE_INFO		0x1
    295 #define KGSL_PROP_DEVICE_SHADOW		0x2
    296 #define KGSL_PROP_DEVICE_POWER		0x3
    297 #define KGSL_PROP_SHMEM			0x4
    298 #define KGSL_PROP_SHMEM_APERTURES	0x5
    299 #define KGSL_PROP_MMU_ENABLE		0x6
    300 #define KGSL_PROP_INTERRUPT_WAITS	0x7
    301 #define KGSL_PROP_VERSION		0x8
    302 #define KGSL_PROP_GPU_RESET_STAT	0x9
    303 #define KGSL_PROP_PWRCTRL		0xE
    304 #define KGSL_PROP_PWR_CONSTRAINT	0x12
    305 #define KGSL_PROP_UCHE_GMEM_VADDR	0x13
    306 #define KGSL_PROP_SP_GENERIC_MEM	0x14
    307 #define KGSL_PROP_UCODE_VERSION		0x15
    308 #define KGSL_PROP_GPMU_VERSION		0x16
    309 #define KGSL_PROP_HIGHEST_BANK_BIT	0x17
    310 #define KGSL_PROP_DEVICE_BITNESS	0x18
    311 
    312 struct kgsl_shadowprop {
    313 	unsigned long gpuaddr;
    314 	size_t size;
    315 	unsigned int flags; /* contains KGSL_FLAGS_ values */
    316 };
    317 
    318 struct kgsl_version {
    319 	unsigned int drv_major;
    320 	unsigned int drv_minor;
    321 	unsigned int dev_major;
    322 	unsigned int dev_minor;
    323 };
    324 
    325 struct kgsl_sp_generic_mem {
    326 	uint64_t local;
    327 	uint64_t pvt;
    328 };
    329 
    330 struct kgsl_ucode_version {
    331 	unsigned int pfp;
    332 	unsigned int pm4;
    333 };
    334 
    335 struct kgsl_gpmu_version {
    336 	unsigned int major;
    337 	unsigned int minor;
    338 	unsigned int features;
    339 };
    340 
    341 /* Performance counter groups */
    342 
    343 #define KGSL_PERFCOUNTER_GROUP_CP 0x0
    344 #define KGSL_PERFCOUNTER_GROUP_RBBM 0x1
    345 #define KGSL_PERFCOUNTER_GROUP_PC 0x2
    346 #define KGSL_PERFCOUNTER_GROUP_VFD 0x3
    347 #define KGSL_PERFCOUNTER_GROUP_HLSQ 0x4
    348 #define KGSL_PERFCOUNTER_GROUP_VPC 0x5
    349 #define KGSL_PERFCOUNTER_GROUP_TSE 0x6
    350 #define KGSL_PERFCOUNTER_GROUP_RAS 0x7
    351 #define KGSL_PERFCOUNTER_GROUP_UCHE 0x8
    352 #define KGSL_PERFCOUNTER_GROUP_TP 0x9
    353 #define KGSL_PERFCOUNTER_GROUP_SP 0xA
    354 #define KGSL_PERFCOUNTER_GROUP_RB 0xB
    355 #define KGSL_PERFCOUNTER_GROUP_PWR 0xC
    356 #define KGSL_PERFCOUNTER_GROUP_VBIF 0xD
    357 #define KGSL_PERFCOUNTER_GROUP_VBIF_PWR 0xE
    358 #define KGSL_PERFCOUNTER_GROUP_MH 0xF
    359 #define KGSL_PERFCOUNTER_GROUP_PA_SU 0x10
    360 #define KGSL_PERFCOUNTER_GROUP_SQ 0x11
    361 #define KGSL_PERFCOUNTER_GROUP_SX 0x12
    362 #define KGSL_PERFCOUNTER_GROUP_TCF 0x13
    363 #define KGSL_PERFCOUNTER_GROUP_TCM 0x14
    364 #define KGSL_PERFCOUNTER_GROUP_TCR 0x15
    365 #define KGSL_PERFCOUNTER_GROUP_L2 0x16
    366 #define KGSL_PERFCOUNTER_GROUP_VSC 0x17
    367 #define KGSL_PERFCOUNTER_GROUP_CCU 0x18
    368 #define KGSL_PERFCOUNTER_GROUP_LRZ 0x19
    369 #define KGSL_PERFCOUNTER_GROUP_CMP 0x1A
    370 #define KGSL_PERFCOUNTER_GROUP_ALWAYSON 0x1B
    371 #define KGSL_PERFCOUNTER_GROUP_SP_PWR 0x1C
    372 #define KGSL_PERFCOUNTER_GROUP_TP_PWR 0x1D
    373 #define KGSL_PERFCOUNTER_GROUP_RB_PWR 0x1E
    374 #define KGSL_PERFCOUNTER_GROUP_CCU_PWR 0x1F
    375 #define KGSL_PERFCOUNTER_GROUP_UCHE_PWR 0x20
    376 #define KGSL_PERFCOUNTER_GROUP_CP_PWR 0x21
    377 #define KGSL_PERFCOUNTER_GROUP_GPMU_PWR 0x22
    378 #define KGSL_PERFCOUNTER_GROUP_ALWAYSON_PWR 0x23
    379 #define KGSL_PERFCOUNTER_GROUP_MAX 0x24
    380 
    381 #define KGSL_PERFCOUNTER_NOT_USED 0xFFFFFFFF
    382 #define KGSL_PERFCOUNTER_BROKEN 0xFFFFFFFE
    383 
    384 /* structure holds list of ibs */
    385 struct kgsl_ibdesc {
    386 	unsigned long gpuaddr;
    387 	unsigned long __pad;
    388 	size_t sizedwords;
    389 	unsigned int ctrl;
    390 };
    391 
    392 /**
    393  * struct kgsl_cmdbatch_profiling_buffer
    394  * @wall_clock_s: Ringbuffer submission time (seconds).
    395  *                If KGSL_CMDBATCH_PROFILING_KTIME is set, time is provided
    396  *                in kernel clocks, otherwise wall clock time is used.
    397  * @wall_clock_ns: Ringbuffer submission time (nanoseconds).
    398  *                 If KGSL_CMDBATCH_PROFILING_KTIME is set time is provided
    399  *                 in kernel clocks, otherwise wall clock time is used.
    400  * @gpu_ticks_queued: GPU ticks at ringbuffer submission
    401  * @gpu_ticks_submitted: GPU ticks when starting cmdbatch execution
    402  * @gpu_ticks_retired: GPU ticks when finishing cmdbatch execution
    403  *
    404  * This structure defines the profiling buffer used to measure cmdbatch
    405  * execution time
    406  */
    407 struct kgsl_cmdbatch_profiling_buffer {
    408 	uint64_t wall_clock_s;
    409 	uint64_t wall_clock_ns;
    410 	uint64_t gpu_ticks_queued;
    411 	uint64_t gpu_ticks_submitted;
    412 	uint64_t gpu_ticks_retired;
    413 };
    414 
    415 /* ioctls */
    416 #define KGSL_IOC_TYPE 0x09
    417 
    418 /* get misc info about the GPU
    419    type should be a value from enum kgsl_property_type
    420    value points to a structure that varies based on type
    421    sizebytes is sizeof() that structure
    422    for KGSL_PROP_DEVICE_INFO, use struct kgsl_devinfo
    423    this structure contaings hardware versioning info.
    424    for KGSL_PROP_DEVICE_SHADOW, use struct kgsl_shadowprop
    425    this is used to find mmap() offset and sizes for mapping
    426    struct kgsl_memstore into userspace.
    427 */
    428 struct kgsl_device_getproperty {
    429 	unsigned int type;
    430 	void __user *value;
    431 	size_t sizebytes;
    432 };
    433 
    434 #define IOCTL_KGSL_DEVICE_GETPROPERTY \
    435 	_IOWR(KGSL_IOC_TYPE, 0x2, struct kgsl_device_getproperty)
    436 
    437 /* IOCTL_KGSL_DEVICE_READ (0x3) - removed 03/2012
    438  */
    439 
    440 /* block until the GPU has executed past a given timestamp
    441  * timeout is in milliseconds.
    442  */
    443 struct kgsl_device_waittimestamp {
    444 	unsigned int timestamp;
    445 	unsigned int timeout;
    446 };
    447 
    448 #define IOCTL_KGSL_DEVICE_WAITTIMESTAMP \
    449 	_IOW(KGSL_IOC_TYPE, 0x6, struct kgsl_device_waittimestamp)
    450 
    451 struct kgsl_device_waittimestamp_ctxtid {
    452 	unsigned int context_id;
    453 	unsigned int timestamp;
    454 	unsigned int timeout;
    455 };
    456 
    457 #define IOCTL_KGSL_DEVICE_WAITTIMESTAMP_CTXTID \
    458 	_IOW(KGSL_IOC_TYPE, 0x7, struct kgsl_device_waittimestamp_ctxtid)
    459 
    460 /* DEPRECATED: issue indirect commands to the GPU.
    461  * drawctxt_id must have been created with IOCTL_KGSL_DRAWCTXT_CREATE
    462  * ibaddr and sizedwords must specify a subset of a buffer created
    463  * with IOCTL_KGSL_SHAREDMEM_FROM_PMEM
    464  * flags may be a mask of KGSL_CONTEXT_ values
    465  * timestamp is a returned counter value which can be passed to
    466  * other ioctls to determine when the commands have been executed by
    467  * the GPU.
    468  *
    469  * This fucntion is deprecated - consider using IOCTL_KGSL_SUBMIT_COMMANDS
    470  * instead
    471  */
    472 struct kgsl_ringbuffer_issueibcmds {
    473 	unsigned int drawctxt_id;
    474 	unsigned long ibdesc_addr;
    475 	unsigned int numibs;
    476 	unsigned int timestamp; /*output param */
    477 	unsigned int flags;
    478 };
    479 
    480 #define IOCTL_KGSL_RINGBUFFER_ISSUEIBCMDS \
    481 	_IOWR(KGSL_IOC_TYPE, 0x10, struct kgsl_ringbuffer_issueibcmds)
    482 
    483 /* read the most recently executed timestamp value
    484  * type should be a value from enum kgsl_timestamp_type
    485  */
    486 struct kgsl_cmdstream_readtimestamp {
    487 	unsigned int type;
    488 	unsigned int timestamp; /*output param */
    489 };
    490 
    491 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP_OLD \
    492 	_IOR(KGSL_IOC_TYPE, 0x11, struct kgsl_cmdstream_readtimestamp)
    493 
    494 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP \
    495 	_IOWR(KGSL_IOC_TYPE, 0x11, struct kgsl_cmdstream_readtimestamp)
    496 
    497 /* free memory when the GPU reaches a given timestamp.
    498  * gpuaddr specify a memory region created by a
    499  * IOCTL_KGSL_SHAREDMEM_FROM_PMEM call
    500  * type should be a value from enum kgsl_timestamp_type
    501  */
    502 struct kgsl_cmdstream_freememontimestamp {
    503 	unsigned long gpuaddr;
    504 	unsigned int type;
    505 	unsigned int timestamp;
    506 };
    507 
    508 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP \
    509 	_IOW(KGSL_IOC_TYPE, 0x12, struct kgsl_cmdstream_freememontimestamp)
    510 
    511 /* Previous versions of this header had incorrectly defined
    512    IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP as a read-only ioctl instead
    513    of a write only ioctl.  To ensure binary compatability, the following
    514    #define will be used to intercept the incorrect ioctl
    515 */
    516 
    517 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP_OLD \
    518 	_IOR(KGSL_IOC_TYPE, 0x12, struct kgsl_cmdstream_freememontimestamp)
    519 
    520 /* create a draw context, which is used to preserve GPU state.
    521  * The flags field may contain a mask KGSL_CONTEXT_*  values
    522  */
    523 struct kgsl_drawctxt_create {
    524 	unsigned int flags;
    525 	unsigned int drawctxt_id; /*output param */
    526 };
    527 
    528 #define IOCTL_KGSL_DRAWCTXT_CREATE \
    529 	_IOWR(KGSL_IOC_TYPE, 0x13, struct kgsl_drawctxt_create)
    530 
    531 /* destroy a draw context */
    532 struct kgsl_drawctxt_destroy {
    533 	unsigned int drawctxt_id;
    534 };
    535 
    536 #define IOCTL_KGSL_DRAWCTXT_DESTROY \
    537 	_IOW(KGSL_IOC_TYPE, 0x14, struct kgsl_drawctxt_destroy)
    538 
    539 /* add a block of pmem, fb, ashmem or user allocated address
    540  * into the GPU address space */
    541 struct kgsl_map_user_mem {
    542 	int fd;
    543 	unsigned long gpuaddr;   /*output param */
    544 	size_t len;
    545 	size_t offset;
    546 	unsigned long hostptr;   /*input param */
    547 	enum kgsl_user_mem_type memtype;
    548 	unsigned int flags;
    549 };
    550 
    551 #define IOCTL_KGSL_MAP_USER_MEM \
    552 	_IOWR(KGSL_IOC_TYPE, 0x15, struct kgsl_map_user_mem)
    553 
    554 struct kgsl_cmdstream_readtimestamp_ctxtid {
    555 	unsigned int context_id;
    556 	unsigned int type;
    557 	unsigned int timestamp; /*output param */
    558 };
    559 
    560 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP_CTXTID \
    561 	_IOWR(KGSL_IOC_TYPE, 0x16, struct kgsl_cmdstream_readtimestamp_ctxtid)
    562 
    563 struct kgsl_cmdstream_freememontimestamp_ctxtid {
    564 	unsigned int context_id;
    565 	unsigned long gpuaddr;
    566 	unsigned int type;
    567 	unsigned int timestamp;
    568 };
    569 
    570 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP_CTXTID \
    571 	_IOW(KGSL_IOC_TYPE, 0x17, \
    572 	struct kgsl_cmdstream_freememontimestamp_ctxtid)
    573 
    574 /* add a block of pmem or fb into the GPU address space */
    575 struct kgsl_sharedmem_from_pmem {
    576         int pmem_fd;
    577         unsigned long gpuaddr;  /*output param */
    578         unsigned int len;
    579         unsigned int offset;
    580 };
    581 
    582 #define IOCTL_KGSL_SHAREDMEM_FROM_PMEM \
    583         _IOWR(KGSL_IOC_TYPE, 0x20, struct kgsl_sharedmem_from_pmem)
    584 
    585 /* remove memory from the GPU's address space */
    586 struct kgsl_sharedmem_free {
    587 	unsigned long gpuaddr;
    588 };
    589 
    590 #define IOCTL_KGSL_SHAREDMEM_FREE \
    591 	_IOW(KGSL_IOC_TYPE, 0x21, struct kgsl_sharedmem_free)
    592 
    593 struct kgsl_cff_user_event {
    594 	unsigned char cff_opcode;
    595 	unsigned int op1;
    596 	unsigned int op2;
    597 	unsigned int op3;
    598 	unsigned int op4;
    599 	unsigned int op5;
    600 	unsigned int __pad[2];
    601 };
    602 
    603 #define IOCTL_KGSL_CFF_USER_EVENT \
    604 	_IOW(KGSL_IOC_TYPE, 0x31, struct kgsl_cff_user_event)
    605 
    606 struct kgsl_gmem_desc {
    607 	unsigned int x;
    608 	unsigned int y;
    609 	unsigned int width;
    610 	unsigned int height;
    611 	unsigned int pitch;
    612 };
    613 
    614 struct kgsl_buffer_desc {
    615 	void 			*hostptr;
    616 	unsigned long	gpuaddr;
    617 	int				size;
    618 	unsigned int	format;
    619 	unsigned int  	pitch;
    620 	unsigned int  	enabled;
    621 };
    622 
    623 struct kgsl_bind_gmem_shadow {
    624 	unsigned int drawctxt_id;
    625 	struct kgsl_gmem_desc gmem_desc;
    626 	unsigned int shadow_x;
    627 	unsigned int shadow_y;
    628 	struct kgsl_buffer_desc shadow_buffer;
    629 	unsigned int buffer_id;
    630 };
    631 
    632 #define IOCTL_KGSL_DRAWCTXT_BIND_GMEM_SHADOW \
    633     _IOW(KGSL_IOC_TYPE, 0x22, struct kgsl_bind_gmem_shadow)
    634 
    635 /* add a block of memory into the GPU address space */
    636 
    637 /*
    638  * IOCTL_KGSL_SHAREDMEM_FROM_VMALLOC deprecated 09/2012
    639  * use IOCTL_KGSL_GPUMEM_ALLOC instead
    640  */
    641 
    642 struct kgsl_sharedmem_from_vmalloc {
    643 	unsigned long gpuaddr;	/*output param */
    644 	unsigned int hostptr;
    645 	unsigned int flags;
    646 };
    647 
    648 #define IOCTL_KGSL_SHAREDMEM_FROM_VMALLOC \
    649 	_IOWR(KGSL_IOC_TYPE, 0x23, struct kgsl_sharedmem_from_vmalloc)
    650 
    651 /*
    652  * This is being deprecated in favor of IOCTL_KGSL_GPUMEM_CACHE_SYNC which
    653  * supports both directions (flush and invalidate). This code will still
    654  * work, but by definition it will do a flush of the cache which might not be
    655  * what you want to have happen on a buffer following a GPU operation.  It is
    656  * safer to go with IOCTL_KGSL_GPUMEM_CACHE_SYNC
    657  */
    658 
    659 #define IOCTL_KGSL_SHAREDMEM_FLUSH_CACHE \
    660 	_IOW(KGSL_IOC_TYPE, 0x24, struct kgsl_sharedmem_free)
    661 
    662 struct kgsl_drawctxt_set_bin_base_offset {
    663 	unsigned int drawctxt_id;
    664 	unsigned int offset;
    665 };
    666 
    667 #define IOCTL_KGSL_DRAWCTXT_SET_BIN_BASE_OFFSET \
    668 	_IOW(KGSL_IOC_TYPE, 0x25, struct kgsl_drawctxt_set_bin_base_offset)
    669 
    670 enum kgsl_cmdwindow_type {
    671 	KGSL_CMDWINDOW_MIN     = 0x00000000,
    672 	KGSL_CMDWINDOW_2D      = 0x00000000,
    673 	KGSL_CMDWINDOW_3D      = 0x00000001, /* legacy */
    674 	KGSL_CMDWINDOW_MMU     = 0x00000002,
    675 	KGSL_CMDWINDOW_ARBITER = 0x000000FF,
    676 	KGSL_CMDWINDOW_MAX     = 0x000000FF,
    677 };
    678 
    679 /* write to the command window */
    680 struct kgsl_cmdwindow_write {
    681 	enum kgsl_cmdwindow_type target;
    682 	unsigned int addr;
    683 	unsigned int data;
    684 };
    685 
    686 #define IOCTL_KGSL_CMDWINDOW_WRITE \
    687 	_IOW(KGSL_IOC_TYPE, 0x2e, struct kgsl_cmdwindow_write)
    688 
    689 struct kgsl_gpumem_alloc {
    690 	unsigned long gpuaddr; /* output param */
    691 	size_t size;
    692 	unsigned int flags;
    693 };
    694 
    695 #define IOCTL_KGSL_GPUMEM_ALLOC \
    696 	_IOWR(KGSL_IOC_TYPE, 0x2f, struct kgsl_gpumem_alloc)
    697 
    698 struct kgsl_cff_syncmem {
    699 	unsigned long gpuaddr;
    700 	size_t len;
    701 	unsigned int __pad[2]; /* For future binary compatibility */
    702 };
    703 
    704 #define IOCTL_KGSL_CFF_SYNCMEM \
    705 	_IOW(KGSL_IOC_TYPE, 0x30, struct kgsl_cff_syncmem)
    706 
    707 /*
    708  * A timestamp event allows the user space to register an action following an
    709  * expired timestamp. Note IOCTL_KGSL_TIMESTAMP_EVENT has been redefined to
    710  * _IOWR to support fences which need to return a fd for the priv parameter.
    711  */
    712 
    713 struct kgsl_timestamp_event {
    714 	int type;                /* Type of event (see list below) */
    715 	unsigned int timestamp;  /* Timestamp to trigger event on */
    716 	unsigned int context_id; /* Context for the timestamp */
    717 	void __user *priv;	 /* Pointer to the event specific blob */
    718 	size_t len;              /* Size of the event specific blob */
    719 };
    720 
    721 #define IOCTL_KGSL_TIMESTAMP_EVENT_OLD \
    722 	_IOW(KGSL_IOC_TYPE, 0x31, struct kgsl_timestamp_event)
    723 
    724 /* A genlock timestamp event releases an existing lock on timestamp expire */
    725 
    726 #define KGSL_TIMESTAMP_EVENT_GENLOCK 1
    727 
    728 struct kgsl_timestamp_event_genlock {
    729 	int handle; /* Handle of the genlock lock to release */
    730 };
    731 
    732 /* A fence timestamp event releases an existing lock on timestamp expire */
    733 
    734 #define KGSL_TIMESTAMP_EVENT_FENCE 2
    735 
    736 struct kgsl_timestamp_event_fence {
    737 	int fence_fd; /* Fence to signal */
    738 };
    739 
    740 /*
    741  * Set a property within the kernel.  Uses the same structure as
    742  * IOCTL_KGSL_GETPROPERTY
    743  */
    744 
    745 #define IOCTL_KGSL_SETPROPERTY \
    746 	_IOW(KGSL_IOC_TYPE, 0x32, struct kgsl_device_getproperty)
    747 
    748 #define IOCTL_KGSL_TIMESTAMP_EVENT \
    749 	_IOWR(KGSL_IOC_TYPE, 0x33, struct kgsl_timestamp_event)
    750 
    751 /**
    752  * struct kgsl_gpumem_alloc_id - argument to IOCTL_KGSL_GPUMEM_ALLOC_ID
    753  * @id: returned id value for this allocation.
    754  * @flags: mask of KGSL_MEM* values requested and actual flags on return.
    755  * @size: requested size of the allocation and actual size on return.
    756  * @mmapsize: returned size to pass to mmap() which may be larger than 'size'
    757  * @gpuaddr: returned GPU address for the allocation
    758  *
    759  * Allocate memory for access by the GPU. The flags and size fields are echoed
    760  * back by the kernel, so that the caller can know if the request was
    761  * adjusted.
    762  *
    763  * Supported flags:
    764  * KGSL_MEMFLAGS_GPUREADONLY: the GPU will be unable to write to the buffer
    765  * KGSL_MEMTYPE*: usage hint for debugging aid
    766  * KGSL_MEMALIGN*: alignment hint, may be ignored or adjusted by the kernel.
    767  * KGSL_MEMFLAGS_USE_CPU_MAP: If set on call and return, the returned GPU
    768  * address will be 0. Calling mmap() will set the GPU address.
    769  */
    770 struct kgsl_gpumem_alloc_id {
    771 	unsigned int id;
    772 	unsigned int flags;
    773 	size_t size;
    774 	size_t mmapsize;
    775 	unsigned long gpuaddr;
    776 /* private: reserved for future use*/
    777 	unsigned long __pad[2];
    778 };
    779 
    780 #define IOCTL_KGSL_GPUMEM_ALLOC_ID \
    781 	_IOWR(KGSL_IOC_TYPE, 0x34, struct kgsl_gpumem_alloc_id)
    782 
    783 /**
    784  * struct kgsl_gpumem_free_id - argument to IOCTL_KGSL_GPUMEM_FREE_ID
    785  * @id: GPU allocation id to free
    786  *
    787  * Free an allocation by id, in case a GPU address has not been assigned or
    788  * is unknown. Freeing an allocation by id with this ioctl or by GPU address
    789  * with IOCTL_KGSL_SHAREDMEM_FREE are equivalent.
    790  */
    791 struct kgsl_gpumem_free_id {
    792 	unsigned int id;
    793 /* private: reserved for future use*/
    794 	unsigned int __pad;
    795 };
    796 
    797 #define IOCTL_KGSL_GPUMEM_FREE_ID \
    798 	_IOWR(KGSL_IOC_TYPE, 0x35, struct kgsl_gpumem_free_id)
    799 
    800 /**
    801  * struct kgsl_gpumem_get_info - argument to IOCTL_KGSL_GPUMEM_GET_INFO
    802  * @gpuaddr: GPU address to query. Also set on return.
    803  * @id: GPU allocation id to query. Also set on return.
    804  * @flags: returned mask of KGSL_MEM* values.
    805  * @size: returned size of the allocation.
    806  * @mmapsize: returned size to pass mmap(), which may be larger than 'size'
    807  * @useraddr: returned address of the userspace mapping for this buffer
    808  *
    809  * This ioctl allows querying of all user visible attributes of an existing
    810  * allocation, by either the GPU address or the id returned by a previous
    811  * call to IOCTL_KGSL_GPUMEM_ALLOC_ID. Legacy allocation ioctls may not
    812  * return all attributes so this ioctl can be used to look them up if needed.
    813  *
    814  */
    815 struct kgsl_gpumem_get_info {
    816 	unsigned long gpuaddr;
    817 	unsigned int id;
    818 	unsigned int flags;
    819 	size_t size;
    820 	size_t mmapsize;
    821 	unsigned long useraddr;
    822 /* private: reserved for future use*/
    823 	unsigned long __pad[4];
    824 };
    825 
    826 #define IOCTL_KGSL_GPUMEM_GET_INFO\
    827 	_IOWR(KGSL_IOC_TYPE, 0x36, struct kgsl_gpumem_get_info)
    828 
    829 /**
    830  * struct kgsl_gpumem_sync_cache - argument to IOCTL_KGSL_GPUMEM_SYNC_CACHE
    831  * @gpuaddr: GPU address of the buffer to sync.
    832  * @id: id of the buffer to sync. Either gpuaddr or id is sufficient.
    833  * @op: a mask of KGSL_GPUMEM_CACHE_* values
    834  * @offset: offset into the buffer
    835  * @length: number of bytes starting from offset to perform
    836  * the cache operation on
    837  *
    838  * Sync the L2 cache for memory headed to and from the GPU - this replaces
    839  * KGSL_SHAREDMEM_FLUSH_CACHE since it can handle cache management for both
    840  * directions
    841  *
    842  */
    843 struct kgsl_gpumem_sync_cache {
    844 	unsigned long gpuaddr;
    845 	unsigned int id;
    846 	unsigned int op;
    847 	size_t offset;
    848 	size_t length;
    849 };
    850 
    851 #define KGSL_GPUMEM_CACHE_CLEAN (1 << 0)
    852 #define KGSL_GPUMEM_CACHE_TO_GPU KGSL_GPUMEM_CACHE_CLEAN
    853 
    854 #define KGSL_GPUMEM_CACHE_INV (1 << 1)
    855 #define KGSL_GPUMEM_CACHE_FROM_GPU KGSL_GPUMEM_CACHE_INV
    856 
    857 #define KGSL_GPUMEM_CACHE_FLUSH \
    858 	(KGSL_GPUMEM_CACHE_CLEAN | KGSL_GPUMEM_CACHE_INV)
    859 
    860 /* Flag to ensure backwards compatibility of kgsl_gpumem_sync_cache struct */
    861 #define KGSL_GPUMEM_CACHE_RANGE (1 << 31U)
    862 
    863 #define IOCTL_KGSL_GPUMEM_SYNC_CACHE \
    864 	_IOW(KGSL_IOC_TYPE, 0x37, struct kgsl_gpumem_sync_cache)
    865 
    866 /**
    867  * struct kgsl_perfcounter_get - argument to IOCTL_KGSL_PERFCOUNTER_GET
    868  * @groupid: Performance counter group ID
    869  * @countable: Countable to select within the group
    870  * @offset: Return offset of the reserved LO counter
    871  * @offset_hi: Return offset of the reserved HI counter
    872  *
    873  * Get an available performance counter from a specified groupid.  The offset
    874  * of the performance counter will be returned after successfully assigning
    875  * the countable to the counter for the specified group.  An error will be
    876  * returned and an offset of 0 if the groupid is invalid or there are no
    877  * more counters left.  After successfully getting a perfcounter, the user
    878  * must call kgsl_perfcounter_put(groupid, contable) when finished with
    879  * the perfcounter to clear up perfcounter resources.
    880  *
    881  */
    882 struct kgsl_perfcounter_get {
    883 	unsigned int groupid;
    884 	unsigned int countable;
    885 	unsigned int offset;
    886 	unsigned int offset_hi;
    887 /* private: reserved for future use */
    888 	unsigned int __pad; /* For future binary compatibility */
    889 };
    890 
    891 #define IOCTL_KGSL_PERFCOUNTER_GET \
    892 	_IOWR(KGSL_IOC_TYPE, 0x38, struct kgsl_perfcounter_get)
    893 
    894 /**
    895  * struct kgsl_perfcounter_put - argument to IOCTL_KGSL_PERFCOUNTER_PUT
    896  * @groupid: Performance counter group ID
    897  * @countable: Countable to release within the group
    898  *
    899  * Put an allocated performance counter to allow others to have access to the
    900  * resource that was previously taken.  This is only to be called after
    901  * successfully getting a performance counter from kgsl_perfcounter_get().
    902  *
    903  */
    904 struct kgsl_perfcounter_put {
    905 	unsigned int groupid;
    906 	unsigned int countable;
    907 /* private: reserved for future use */
    908 	unsigned int __pad[2]; /* For future binary compatibility */
    909 };
    910 
    911 #define IOCTL_KGSL_PERFCOUNTER_PUT \
    912 	_IOW(KGSL_IOC_TYPE, 0x39, struct kgsl_perfcounter_put)
    913 
    914 /**
    915  * struct kgsl_perfcounter_query - argument to IOCTL_KGSL_PERFCOUNTER_QUERY
    916  * @groupid: Performance counter group ID
    917  * @countable: Return active countables array
    918  * @size: Size of active countables array
    919  * @max_counters: Return total number counters for the group ID
    920  *
    921  * Query the available performance counters given a groupid.  The array
    922  * *countables is used to return the current active countables in counters.
    923  * The size of the array is passed in so the kernel will only write at most
    924  * size or counter->size for the group id.  The total number of available
    925  * counters for the group ID is returned in max_counters.
    926  * If the array or size passed in are invalid, then only the maximum number
    927  * of counters will be returned, no data will be written to *countables.
    928  * If the groupid is invalid an error code will be returned.
    929  *
    930  */
    931 struct kgsl_perfcounter_query {
    932 	unsigned int groupid;
    933 	/* Array to return the current countable for up to size counters */
    934 	unsigned int __user *countables;
    935 	unsigned int count;
    936 	unsigned int max_counters;
    937 /* private: reserved for future use */
    938 	unsigned int __pad[2]; /* For future binary compatibility */
    939 };
    940 
    941 #define IOCTL_KGSL_PERFCOUNTER_QUERY \
    942 	_IOWR(KGSL_IOC_TYPE, 0x3A, struct kgsl_perfcounter_query)
    943 
    944 /**
    945  * struct kgsl_perfcounter_query - argument to IOCTL_KGSL_PERFCOUNTER_QUERY
    946  * @groupid: Performance counter group IDs
    947  * @countable: Performance counter countable IDs
    948  * @value: Return performance counter reads
    949  * @size: Size of all arrays (groupid/countable pair and return value)
    950  *
    951  * Read in the current value of a performance counter given by the groupid
    952  * and countable.
    953  *
    954  */
    955 
    956 struct kgsl_perfcounter_read_group {
    957 	unsigned int groupid;
    958 	unsigned int countable;
    959 	unsigned long long value;
    960 };
    961 
    962 struct kgsl_perfcounter_read {
    963 	struct kgsl_perfcounter_read_group __user *reads;
    964 	unsigned int count;
    965 /* private: reserved for future use */
    966 	unsigned int __pad[2]; /* For future binary compatibility */
    967 };
    968 
    969 #define IOCTL_KGSL_PERFCOUNTER_READ \
    970 	_IOWR(KGSL_IOC_TYPE, 0x3B, struct kgsl_perfcounter_read)
    971 /*
    972  * struct kgsl_gpumem_sync_cache_bulk - argument to
    973  * IOCTL_KGSL_GPUMEM_SYNC_CACHE_BULK
    974  * @id_list: list of GPU buffer ids of the buffers to sync
    975  * @count: number of GPU buffer ids in id_list
    976  * @op: a mask of KGSL_GPUMEM_CACHE_* values
    977  *
    978  * Sync the cache for memory headed to and from the GPU. Certain
    979  * optimizations can be made on the cache operation based on the total
    980  * size of the working set of memory to be managed.
    981  */
    982 struct kgsl_gpumem_sync_cache_bulk {
    983 	unsigned int __user *id_list;
    984 	unsigned int count;
    985 	unsigned int op;
    986 /* private: reserved for future use */
    987 	unsigned int __pad[2]; /* For future binary compatibility */
    988 };
    989 
    990 #define IOCTL_KGSL_GPUMEM_SYNC_CACHE_BULK \
    991 	_IOWR(KGSL_IOC_TYPE, 0x3C, struct kgsl_gpumem_sync_cache_bulk)
    992 
    993 /*
    994  * struct kgsl_cmd_syncpoint_timestamp
    995  * @context_id: ID of a KGSL context
    996  * @timestamp: GPU timestamp
    997  *
    998  * This structure defines a syncpoint comprising a context/timestamp pair. A
    999  * list of these may be passed by IOCTL_KGSL_SUBMIT_COMMANDS to define
   1000  * dependencies that must be met before the command can be submitted to the
   1001  * hardware
   1002  */
   1003 struct kgsl_cmd_syncpoint_timestamp {
   1004 	unsigned int context_id;
   1005 	unsigned int timestamp;
   1006 };
   1007 
   1008 struct kgsl_cmd_syncpoint_fence {
   1009 	int fd;
   1010 };
   1011 
   1012 /**
   1013  * struct kgsl_cmd_syncpoint - Define a sync point for a command batch
   1014  * @type: type of sync point defined here
   1015  * @priv: Pointer to the type specific buffer
   1016  * @size: Size of the type specific buffer
   1017  *
   1018  * This structure contains pointers defining a specific command sync point.
   1019  * The pointer and size should point to a type appropriate structure.
   1020  */
   1021 struct kgsl_cmd_syncpoint {
   1022 	int type;
   1023 	void __user *priv;
   1024 	size_t size;
   1025 };
   1026 
   1027 /* Flag to indicate that the cmdlist may contain memlists */
   1028 #define KGSL_IBDESC_MEMLIST 0x1
   1029 
   1030 /* Flag to point out the cmdbatch profiling buffer in the memlist */
   1031 #define KGSL_IBDESC_PROFILING_BUFFER 0x2
   1032 
   1033 /**
   1034  * struct kgsl_submit_commands - Argument to IOCTL_KGSL_SUBMIT_COMMANDS
   1035  * @context_id: KGSL context ID that owns the commands
   1036  * @flags:
   1037  * @cmdlist: User pointer to a list of kgsl_ibdesc structures
   1038  * @numcmds: Number of commands listed in cmdlist
   1039  * @synclist: User pointer to a list of kgsl_cmd_syncpoint structures
   1040  * @numsyncs: Number of sync points listed in synclist
   1041  * @timestamp: On entry the a user defined timestamp, on exist the timestamp
   1042  * assigned to the command batch
   1043  *
   1044  * This structure specifies a command to send to the GPU hardware.  This is
   1045  * similar to kgsl_issueibcmds expect that it doesn't support the legacy way to
   1046  * submit IB lists and it adds sync points to block the IB until the
   1047  * dependencies are satisified.  This entry point is the new and preferred way
   1048  * to submit commands to the GPU. The memory list can be used to specify all
   1049  * memory that is referrenced in the current set of commands.
   1050  */
   1051 
   1052 struct kgsl_submit_commands {
   1053 	unsigned int context_id;
   1054 	unsigned int flags;
   1055 	struct kgsl_ibdesc __user *cmdlist;
   1056 	unsigned int numcmds;
   1057 	struct kgsl_cmd_syncpoint __user *synclist;
   1058 	unsigned int numsyncs;
   1059 	unsigned int timestamp;
   1060 /* private: reserved for future use */
   1061 	unsigned int __pad[4];
   1062 };
   1063 
   1064 #define IOCTL_KGSL_SUBMIT_COMMANDS \
   1065 	_IOWR(KGSL_IOC_TYPE, 0x3D, struct kgsl_submit_commands)
   1066 
   1067 /**
   1068  * struct kgsl_device_constraint - device constraint argument
   1069  * @context_id: KGSL context ID
   1070  * @type: type of constraint i.e pwrlevel/none
   1071  * @data: constraint data
   1072  * @size: size of the constraint data
   1073  */
   1074 struct kgsl_device_constraint {
   1075 	unsigned int type;
   1076 	unsigned int context_id;
   1077 	void __user *data;
   1078 	size_t size;
   1079 };
   1080 
   1081 /* Constraint Type*/
   1082 #define KGSL_CONSTRAINT_NONE 0
   1083 #define KGSL_CONSTRAINT_PWRLEVEL 1
   1084 
   1085 /* PWRLEVEL constraint level*/
   1086 /* set to min frequency */
   1087 #define KGSL_CONSTRAINT_PWR_MIN    0
   1088 /* set to max frequency */
   1089 #define KGSL_CONSTRAINT_PWR_MAX    1
   1090 
   1091 struct kgsl_device_constraint_pwrlevel {
   1092 	unsigned int level;
   1093 };
   1094 
   1095 /**
   1096  * struct kgsl_syncsource_create - Argument to IOCTL_KGSL_SYNCSOURCE_CREATE
   1097  * @id: returned id for the syncsource that was created.
   1098  *
   1099  * This ioctl creates a userspace sync timeline.
   1100  */
   1101 
   1102 struct kgsl_syncsource_create {
   1103 	unsigned int id;
   1104 /* private: reserved for future use */
   1105 	unsigned int __pad[3];
   1106 };
   1107 
   1108 #define IOCTL_KGSL_SYNCSOURCE_CREATE \
   1109 	_IOWR(KGSL_IOC_TYPE, 0x40, struct kgsl_syncsource_create)
   1110 
   1111 /**
   1112  * struct kgsl_syncsource_destroy - Argument to IOCTL_KGSL_SYNCSOURCE_DESTROY
   1113  * @id: syncsource id to destroy
   1114  *
   1115  * This ioctl creates a userspace sync timeline.
   1116  */
   1117 
   1118 struct kgsl_syncsource_destroy {
   1119 	unsigned int id;
   1120 /* private: reserved for future use */
   1121 	unsigned int __pad[3];
   1122 };
   1123 
   1124 #define IOCTL_KGSL_SYNCSOURCE_DESTROY \
   1125 	_IOWR(KGSL_IOC_TYPE, 0x41, struct kgsl_syncsource_destroy)
   1126 
   1127 /**
   1128  * struct kgsl_syncsource_create_fence - Argument to
   1129  *     IOCTL_KGSL_SYNCSOURCE_CREATE_FENCE
   1130  * @id: syncsource id
   1131  * @fence_fd: returned sync_fence fd
   1132  *
   1133  * Create a fence that may be signaled by userspace by calling
   1134  * IOCTL_KGSL_SYNCSOURCE_SIGNAL_FENCE. There are no order dependencies between
   1135  * these fences.
   1136  */
   1137 struct kgsl_syncsource_create_fence {
   1138 	unsigned int id;
   1139 	int fence_fd;
   1140 /* private: reserved for future use */
   1141 	unsigned int __pad[4];
   1142 };
   1143 
   1144 /**
   1145  * struct kgsl_syncsource_signal_fence - Argument to
   1146  *     IOCTL_KGSL_SYNCSOURCE_SIGNAL_FENCE
   1147  * @id: syncsource id
   1148  * @fence_fd: sync_fence fd to signal
   1149  *
   1150  * Signal a fence that was created by a IOCTL_KGSL_SYNCSOURCE_CREATE_FENCE
   1151  * call using the same syncsource id. This allows a fence to be shared
   1152  * to other processes but only signaled by the process owning the fd
   1153  * used to create the fence.
   1154  */
   1155 #define IOCTL_KGSL_SYNCSOURCE_CREATE_FENCE \
   1156 	_IOWR(KGSL_IOC_TYPE, 0x42, struct kgsl_syncsource_create_fence)
   1157 
   1158 struct kgsl_syncsource_signal_fence {
   1159 	unsigned int id;
   1160 	int fence_fd;
   1161 /* private: reserved for future use */
   1162 	unsigned int __pad[4];
   1163 };
   1164 
   1165 #define IOCTL_KGSL_SYNCSOURCE_SIGNAL_FENCE \
   1166 	_IOWR(KGSL_IOC_TYPE, 0x43, struct kgsl_syncsource_signal_fence)
   1167 
   1168 /**
   1169  * struct kgsl_cff_sync_gpuobj - Argument to IOCTL_KGSL_CFF_SYNC_GPUOBJ
   1170  * @offset: Offset into the GPU object to sync
   1171  * @length: Number of bytes to sync
   1172  * @id: ID of the GPU object to sync
   1173  */
   1174 struct kgsl_cff_sync_gpuobj {
   1175 	uint64_t offset;
   1176 	uint64_t length;
   1177 	unsigned int id;
   1178 };
   1179 
   1180 #define IOCTL_KGSL_CFF_SYNC_GPUOBJ \
   1181 	_IOW(KGSL_IOC_TYPE, 0x44, struct kgsl_cff_sync_gpuobj)
   1182 
   1183 /**
   1184  * struct kgsl_gpuobj_alloc - Argument to IOCTL_KGSL_GPUOBJ_ALLOC
   1185  * @size: Size in bytes of the object to allocate
   1186  * @flags: mask of KGSL_MEMFLAG_* bits
   1187  * @va_len: Size in bytes of the virtual region to allocate
   1188  * @mmapsize: Returns the mmap() size of the object
   1189  * @id: Returns the GPU object ID of the new object
   1190  * @metadata_len: Length of the metdata to copy from the user
   1191  * @metadata: Pointer to the user specified metadata to store for the object
   1192  */
   1193 struct kgsl_gpuobj_alloc {
   1194 	uint64_t size;
   1195 	uint64_t flags;
   1196 	uint64_t va_len;
   1197 	uint64_t mmapsize;
   1198 	unsigned int id;
   1199 	unsigned int metadata_len;
   1200 	uint64_t metadata;
   1201 };
   1202 
   1203 /* Let the user know that this header supports the gpuobj metadata */
   1204 #define KGSL_GPUOBJ_ALLOC_METADATA_MAX 64
   1205 
   1206 #define IOCTL_KGSL_GPUOBJ_ALLOC \
   1207 	_IOWR(KGSL_IOC_TYPE, 0x45, struct kgsl_gpuobj_alloc)
   1208 
   1209 /**
   1210  * struct kgsl_gpuobj_free - Argument to IOCTL_KGLS_GPUOBJ_FREE
   1211  * @flags: Mask of: KGSL_GUPOBJ_FREE_ON_EVENT
   1212  * @priv: Pointer to the private object if KGSL_GPUOBJ_FREE_ON_EVENT is
   1213  * specified
   1214  * @id: ID of the GPU object to free
   1215  * @type: If KGSL_GPUOBJ_FREE_ON_EVENT is specified, the type of asynchronous
   1216  * event to free on
   1217  * @len: Length of the data passed in priv
   1218  */
   1219 struct kgsl_gpuobj_free {
   1220 	uint64_t flags;
   1221 	uint64_t __user priv;
   1222 	unsigned int id;
   1223 	unsigned int type;
   1224 	unsigned int len;
   1225 };
   1226 
   1227 #define KGSL_GPUOBJ_FREE_ON_EVENT 1
   1228 
   1229 #define KGSL_GPU_EVENT_TIMESTAMP 1
   1230 #define KGSL_GPU_EVENT_FENCE     2
   1231 
   1232 /**
   1233  * struct kgsl_gpu_event_timestamp - Specifies a timestamp event to free a GPU
   1234  * object on
   1235  * @context_id: ID of the timestamp event to wait for
   1236  * @timestamp: Timestamp of the timestamp event to wait for
   1237  */
   1238 struct kgsl_gpu_event_timestamp {
   1239 	unsigned int context_id;
   1240 	unsigned int timestamp;
   1241 };
   1242 
   1243 /**
   1244  * struct kgsl_gpu_event_fence - Specifies a fence ID to to free a GPU object on
   1245  * @fd: File descriptor for the fence
   1246  */
   1247 struct kgsl_gpu_event_fence {
   1248 	int fd;
   1249 };
   1250 
   1251 #define IOCTL_KGSL_GPUOBJ_FREE \
   1252 	_IOW(KGSL_IOC_TYPE, 0x46, struct kgsl_gpuobj_free)
   1253 
   1254 /**
   1255  * struct kgsl_gpuobj_info - argument to IOCTL_KGSL_GPUOBJ_INFO
   1256  * @gpuaddr: GPU address of the object
   1257  * @flags: Current flags for the object
   1258  * @size: Size of the object
   1259  * @va_len: VA size of the object
   1260  * @va_addr: Virtual address of the object (if it is mapped)
   1261  * id - GPU object ID of the object to query
   1262  */
   1263 struct kgsl_gpuobj_info {
   1264 	uint64_t gpuaddr;
   1265 	uint64_t flags;
   1266 	uint64_t size;
   1267 	uint64_t va_len;
   1268 	uint64_t va_addr;
   1269 	unsigned int id;
   1270 };
   1271 
   1272 #define IOCTL_KGSL_GPUOBJ_INFO \
   1273 	_IOWR(KGSL_IOC_TYPE, 0x47, struct kgsl_gpuobj_info)
   1274 
   1275 /**
   1276  * struct kgsl_gpuobj_import - argument to IOCTL_KGSL_GPUOBJ_IMPORT
   1277  * @priv: Pointer to the private data for the import type
   1278  * @priv_len: Length of the private data
   1279  * @flags: Mask of KGSL_MEMFLAG_ flags
   1280  * @type: Type of the import (KGSL_USER_MEM_TYPE_*)
   1281  * @id: Returns the ID of the new GPU object
   1282  */
   1283 struct kgsl_gpuobj_import {
   1284 	uint64_t __user priv;
   1285 	uint64_t priv_len;
   1286 	uint64_t flags;
   1287 	unsigned int type;
   1288 	unsigned int id;
   1289 };
   1290 
   1291 /**
   1292  * struct kgsl_gpuobj_import_dma_buf - import a dmabuf object
   1293  * @fd: File descriptor for the dma-buf object
   1294  */
   1295 struct kgsl_gpuobj_import_dma_buf {
   1296 	int fd;
   1297 };
   1298 
   1299 /**
   1300  * struct kgsl_gpuobj_import_useraddr - import an object based on a useraddr
   1301  * @virtaddr: Virtual address of the object to import
   1302  */
   1303 struct kgsl_gpuobj_import_useraddr {
   1304 	uint64_t virtaddr;
   1305 };
   1306 
   1307 #define IOCTL_KGSL_GPUOBJ_IMPORT \
   1308 	_IOWR(KGSL_IOC_TYPE, 0x48, struct kgsl_gpuobj_import)
   1309 
   1310 /**
   1311  * struct kgsl_gpuobj_sync_obj - Individual GPU object to sync
   1312  * @offset: Offset within the GPU object to sync
   1313  * @length: Number of bytes to sync
   1314  * @id: ID of the GPU object to sync
   1315  * @op: Cache operation to execute
   1316  */
   1317 
   1318 struct kgsl_gpuobj_sync_obj {
   1319 	uint64_t offset;
   1320 	uint64_t length;
   1321 	unsigned int id;
   1322 	unsigned int op;
   1323 };
   1324 
   1325 /**
   1326  * struct kgsl_gpuobj_sync - Argument for IOCTL_KGSL_GPUOBJ_SYNC
   1327  * @objs: Pointer to an array of kgsl_gpuobj_sync_obj structs
   1328  * @obj_len: Size of each item in the array
   1329  * @count: Number of items in the array
   1330  */
   1331 
   1332 struct kgsl_gpuobj_sync {
   1333 	uint64_t __user objs;
   1334 	unsigned int obj_len;
   1335 	unsigned int count;
   1336 };
   1337 
   1338 #define IOCTL_KGSL_GPUOBJ_SYNC \
   1339 	_IOW(KGSL_IOC_TYPE, 0x49, struct kgsl_gpuobj_sync)
   1340 
   1341 /**
   1342  * struct kgsl_command_object - GPU command object
   1343  * @offset: GPU address offset of the object
   1344  * @gpuaddr: GPU address of the object
   1345  * @size: Size of the object
   1346  * @flags: Current flags for the object
   1347  * @id - GPU command object ID
   1348  */
   1349 struct kgsl_command_object {
   1350 	uint64_t offset;
   1351 	uint64_t gpuaddr;
   1352 	uint64_t size;
   1353 	unsigned int flags;
   1354 	unsigned int id;
   1355 };
   1356 
   1357 /**
   1358  * struct kgsl_command_syncpoint - GPU syncpoint object
   1359  * @priv: Pointer to the type specific buffer
   1360  * @size: Size of the type specific buffer
   1361  * @type: type of sync point defined here
   1362  */
   1363 struct kgsl_command_syncpoint {
   1364 	uint64_t __user priv;
   1365 	uint64_t size;
   1366 	unsigned int type;
   1367 };
   1368 
   1369 /**
   1370  * struct kgsl_command_object - Argument for IOCTL_KGSL_GPU_COMMAND
   1371  * @flags: Current flags for the object
   1372  * @cmdlist: List of kgsl_command_objects for submission
   1373  * @cmd_size: Size of kgsl_command_objects structure
   1374  * @numcmds: Number of kgsl_command_objects in command list
   1375  * @objlist: List of kgsl_command_objects for tracking
   1376  * @obj_size: Size of kgsl_command_objects structure
   1377  * @numobjs: Number of kgsl_command_objects in object list
   1378  * @synclist: List of kgsl_command_syncpoints
   1379  * @sync_size: Size of kgsl_command_syncpoint structure
   1380  * @numsyncs: Number of kgsl_command_syncpoints in syncpoint list
   1381  * @context_id: Context ID submittin ghte kgsl_gpu_command
   1382  * @timestamp: Timestamp for the submitted commands
   1383  */
   1384 struct kgsl_gpu_command {
   1385 	uint64_t flags;
   1386 	uint64_t __user cmdlist;
   1387 	unsigned int cmdsize;
   1388 	unsigned int numcmds;
   1389 	uint64_t __user objlist;
   1390 	unsigned int objsize;
   1391 	unsigned int numobjs;
   1392 	uint64_t __user synclist;
   1393 	unsigned int syncsize;
   1394 	unsigned int numsyncs;
   1395 	unsigned int context_id;
   1396 	unsigned int timestamp;
   1397 };
   1398 
   1399 #define IOCTL_KGSL_GPU_COMMAND \
   1400 	_IOWR(KGSL_IOC_TYPE, 0x4A, struct kgsl_gpu_command)
   1401 
   1402 /**
   1403  * struct kgsl_preemption_counters_query - argument to
   1404  * IOCTL_KGSL_PREEMPTIONCOUNTER_QUERY
   1405  * @counters: Return preemption counters array
   1406  * @size_user: Size allocated by userspace
   1407  * @size_priority_level: Size of preemption counters for each
   1408  * priority level
   1409  * @max_priority_level: Return max number of priority levels
   1410  *
   1411  * Query the available preemption counters. The array counters
   1412  * is used to return preemption counters. The size of the array
   1413  * is passed in so the kernel will only write at most size_user
   1414  * or max available preemption counters.  The total number of
   1415  * preemption counters is returned in max_priority_level. If the
   1416  * array or size passed in are invalid, then an error is
   1417  * returned back.
   1418  */
   1419 struct kgsl_preemption_counters_query {
   1420 	uint64_t __user counters;
   1421 	unsigned int size_user;
   1422 	unsigned int size_priority_level;
   1423 	unsigned int max_priority_level;
   1424 };
   1425 
   1426 #define IOCTL_KGSL_PREEMPTIONCOUNTER_QUERY \
   1427 	_IOWR(KGSL_IOC_TYPE, 0x4B, struct kgsl_preemption_counters_query)
   1428 
   1429 /**
   1430  * struct kgsl_gpuobj_set_info - argument for IOCTL_KGSL_GPUOBJ_SET_INFO
   1431  * @flags: Flags to indicate which paramaters to change
   1432  * @metadata:  If KGSL_GPUOBJ_SET_INFO_METADATA is set, a pointer to the new
   1433  * metadata
   1434  * @id: GPU memory object ID to change
   1435  * @metadata_len:  If KGSL_GPUOBJ_SET_INFO_METADATA is set, the length of the
   1436  * new metadata string
   1437  * @type: If KGSL_GPUOBJ_SET_INFO_TYPE is set, the new type of the memory object
   1438  */
   1439 
   1440 #define KGSL_GPUOBJ_SET_INFO_METADATA (1 << 0)
   1441 #define KGSL_GPUOBJ_SET_INFO_TYPE (1 << 1)
   1442 
   1443 struct kgsl_gpuobj_set_info {
   1444 	uint64_t flags;
   1445 	uint64_t metadata;
   1446 	unsigned int id;
   1447 	unsigned int metadata_len;
   1448 	unsigned int type;
   1449 };
   1450 
   1451 #define IOCTL_KGSL_GPUOBJ_SET_INFO \
   1452 	_IOW(KGSL_IOC_TYPE, 0x4C, struct kgsl_gpuobj_set_info)
   1453 
   1454 #endif /* _UAPI_MSM_KGSL_H */
   1455