1 #ifndef _MSM_KGSL_H 2 #define _MSM_KGSL_H 3 4 /* 5 * The KGSL version has proven not to be very useful in userspace if features 6 * are cherry picked into other trees out of order so it is frozen as of 3.14. 7 * It is left here for backwards compatabilty and as a reminder that 8 * software releases are never linear. Also, I like pie. 9 */ 10 11 #define KGSL_VERSION_MAJOR 3 12 #define KGSL_VERSION_MINOR 14 13 14 /*context flags */ 15 #define KGSL_CONTEXT_SAVE_GMEM 0x00000001 16 #define KGSL_CONTEXT_NO_GMEM_ALLOC 0x00000002 17 #define KGSL_CONTEXT_SUBMIT_IB_LIST 0x00000004 18 #define KGSL_CONTEXT_CTX_SWITCH 0x00000008 19 #define KGSL_CONTEXT_PREAMBLE 0x00000010 20 #define KGSL_CONTEXT_TRASH_STATE 0x00000020 21 #define KGSL_CONTEXT_PER_CONTEXT_TS 0x00000040 22 #define KGSL_CONTEXT_USER_GENERATED_TS 0x00000080 23 #define KGSL_CONTEXT_END_OF_FRAME 0x00000100 24 25 #define KGSL_CONTEXT_NO_FAULT_TOLERANCE 0x00000200 26 #define KGSL_CONTEXT_SYNC 0x00000400 27 #define KGSL_CONTEXT_PWR_CONSTRAINT 0x00000800 28 /* bits [12:15] are reserved for future use */ 29 #define KGSL_CONTEXT_TYPE_MASK 0x01F00000 30 #define KGSL_CONTEXT_TYPE_SHIFT 20 31 32 #define KGSL_CONTEXT_TYPE_ANY 0 33 #define KGSL_CONTEXT_TYPE_GL 1 34 #define KGSL_CONTEXT_TYPE_CL 2 35 #define KGSL_CONTEXT_TYPE_C2D 3 36 #define KGSL_CONTEXT_TYPE_RS 4 37 38 #define KGSL_CONTEXT_INVALID 0xffffffff 39 40 /* --- Memory allocation flags --- */ 41 42 /* General allocation hints */ 43 #define KGSL_MEMFLAGS_GPUREADONLY 0x01000000 44 #define KGSL_MEMFLAGS_USE_CPU_MAP 0x10000000 45 46 /* Memory caching hints */ 47 #define KGSL_CACHEMODE_MASK 0x0C000000 48 #define KGSL_CACHEMODE_SHIFT 26 49 50 #define KGSL_CACHEMODE_WRITECOMBINE 0 51 #define KGSL_CACHEMODE_UNCACHED 1 52 #define KGSL_CACHEMODE_WRITETHROUGH 2 53 #define KGSL_CACHEMODE_WRITEBACK 3 54 55 /* Memory types for which allocations are made */ 56 #define KGSL_MEMTYPE_MASK 0x0000FF00 57 #define KGSL_MEMTYPE_SHIFT 8 58 59 #define KGSL_MEMTYPE_OBJECTANY 0 60 #define KGSL_MEMTYPE_FRAMEBUFFER 1 61 #define KGSL_MEMTYPE_RENDERBUFFER 2 62 #define KGSL_MEMTYPE_ARRAYBUFFER 3 63 #define KGSL_MEMTYPE_ELEMENTARRAYBUFFER 4 64 #define KGSL_MEMTYPE_VERTEXARRAYBUFFER 5 65 #define KGSL_MEMTYPE_TEXTURE 6 66 #define KGSL_MEMTYPE_SURFACE 7 67 #define KGSL_MEMTYPE_EGL_SURFACE 8 68 #define KGSL_MEMTYPE_GL 9 69 #define KGSL_MEMTYPE_CL 10 70 #define KGSL_MEMTYPE_CL_BUFFER_MAP 11 71 #define KGSL_MEMTYPE_CL_BUFFER_NOMAP 12 72 #define KGSL_MEMTYPE_CL_IMAGE_MAP 13 73 #define KGSL_MEMTYPE_CL_IMAGE_NOMAP 14 74 #define KGSL_MEMTYPE_CL_KERNEL_STACK 15 75 #define KGSL_MEMTYPE_COMMAND 16 76 #define KGSL_MEMTYPE_2D 17 77 #define KGSL_MEMTYPE_EGL_IMAGE 18 78 #define KGSL_MEMTYPE_EGL_SHADOW 19 79 #define KGSL_MEMTYPE_MULTISAMPLE 20 80 #define KGSL_MEMTYPE_KERNEL 255 81 82 /* 83 * Alignment hint, passed as the power of 2 exponent. 84 * i.e 4k (2^12) would be 12, 64k (2^16)would be 16. 85 */ 86 #define KGSL_MEMALIGN_MASK 0x00FF0000 87 #define KGSL_MEMALIGN_SHIFT 16 88 89 /* --- generic KGSL flag values --- */ 90 91 #define KGSL_FLAGS_NORMALMODE 0x00000000 92 #define KGSL_FLAGS_SAFEMODE 0x00000001 93 #define KGSL_FLAGS_INITIALIZED0 0x00000002 94 #define KGSL_FLAGS_INITIALIZED 0x00000004 95 #define KGSL_FLAGS_STARTED 0x00000008 96 #define KGSL_FLAGS_ACTIVE 0x00000010 97 #define KGSL_FLAGS_RESERVED0 0x00000020 98 #define KGSL_FLAGS_RESERVED1 0x00000040 99 #define KGSL_FLAGS_RESERVED2 0x00000080 100 #define KGSL_FLAGS_SOFT_RESET 0x00000100 101 #define KGSL_FLAGS_PER_CONTEXT_TIMESTAMPS 0x00000200 102 103 /* Clock flags to show which clocks should be controled by a given platform */ 104 #define KGSL_CLK_SRC 0x00000001 105 #define KGSL_CLK_CORE 0x00000002 106 #define KGSL_CLK_IFACE 0x00000004 107 #define KGSL_CLK_MEM 0x00000008 108 #define KGSL_CLK_MEM_IFACE 0x00000010 109 #define KGSL_CLK_AXI 0x00000020 110 111 /* Server Side Sync Timeout in milliseconds */ 112 #define KGSL_SYNCOBJ_SERVER_TIMEOUT 2000 113 114 /* 115 * Reset status values for context 116 */ 117 enum kgsl_ctx_reset_stat { 118 KGSL_CTX_STAT_NO_ERROR = 0x00000000, 119 KGSL_CTX_STAT_GUILTY_CONTEXT_RESET_EXT = 0x00000001, 120 KGSL_CTX_STAT_INNOCENT_CONTEXT_RESET_EXT = 0x00000002, 121 KGSL_CTX_STAT_UNKNOWN_CONTEXT_RESET_EXT = 0x00000003 122 }; 123 124 #define KGSL_CONVERT_TO_MBPS(val) \ 125 (val*1000*1000U) 126 127 /* device id */ 128 enum kgsl_deviceid { 129 KGSL_DEVICE_3D0 = 0x00000000, 130 KGSL_DEVICE_2D0 = 0x00000001, 131 KGSL_DEVICE_2D1 = 0x00000002, 132 KGSL_DEVICE_MAX = 0x00000003 133 }; 134 135 enum kgsl_user_mem_type { 136 KGSL_USER_MEM_TYPE_PMEM = 0x00000000, 137 KGSL_USER_MEM_TYPE_ASHMEM = 0x00000001, 138 KGSL_USER_MEM_TYPE_ADDR = 0x00000002, 139 KGSL_USER_MEM_TYPE_ION = 0x00000003, 140 KGSL_USER_MEM_TYPE_MAX = 0x00000004, 141 }; 142 143 struct kgsl_devinfo { 144 145 unsigned int device_id; 146 /* chip revision id 147 * coreid:8 majorrev:8 minorrev:8 patch:8 148 */ 149 unsigned int chip_id; 150 unsigned int mmu_enabled; 151 unsigned int gmem_gpubaseaddr; 152 /* 153 * This field contains the adreno revision 154 * number 200, 205, 220, etc... 155 */ 156 unsigned int gpu_id; 157 unsigned int gmem_sizebytes; 158 }; 159 160 /* this structure defines the region of memory that can be mmap()ed from this 161 driver. The timestamp fields are volatile because they are written by the 162 GPU 163 */ 164 struct kgsl_devmemstore { 165 volatile unsigned int soptimestamp; 166 unsigned int sbz; 167 volatile unsigned int eoptimestamp; 168 unsigned int sbz2; 169 volatile unsigned int ts_cmp_enable; 170 unsigned int sbz3; 171 volatile unsigned int ref_wait_ts; 172 unsigned int sbz4; 173 unsigned int current_context; 174 unsigned int sbz5; 175 }; 176 177 #define KGSL_MEMSTORE_OFFSET(ctxt_id, field) \ 178 ((ctxt_id)*sizeof(struct kgsl_devmemstore) + \ 179 offsetof(struct kgsl_devmemstore, field)) 180 181 /* timestamp id*/ 182 enum kgsl_timestamp_type { 183 KGSL_TIMESTAMP_CONSUMED = 0x00000001, /* start-of-pipeline timestamp */ 184 KGSL_TIMESTAMP_RETIRED = 0x00000002, /* end-of-pipeline timestamp*/ 185 KGSL_TIMESTAMP_QUEUED = 0x00000003, 186 }; 187 188 /* property types - used with kgsl_device_getproperty */ 189 enum kgsl_property_type { 190 KGSL_PROP_DEVICE_INFO = 0x00000001, 191 KGSL_PROP_DEVICE_SHADOW = 0x00000002, 192 KGSL_PROP_DEVICE_POWER = 0x00000003, 193 KGSL_PROP_SHMEM = 0x00000004, 194 KGSL_PROP_SHMEM_APERTURES = 0x00000005, 195 KGSL_PROP_MMU_ENABLE = 0x00000006, 196 KGSL_PROP_INTERRUPT_WAITS = 0x00000007, 197 KGSL_PROP_VERSION = 0x00000008, 198 KGSL_PROP_GPU_RESET_STAT = 0x00000009, 199 KGSL_PROP_PWRCTRL = 0x0000000E, 200 KGSL_PROP_PWR_CONSTRAINT = 0x00000012, 201 }; 202 203 struct kgsl_shadowprop { 204 unsigned int gpuaddr; 205 unsigned int size; 206 unsigned int flags; /* contains KGSL_FLAGS_ values */ 207 }; 208 209 struct kgsl_version { 210 unsigned int drv_major; 211 unsigned int drv_minor; 212 unsigned int dev_major; 213 unsigned int dev_minor; 214 }; 215 216 /* Performance counter groups */ 217 218 #define KGSL_PERFCOUNTER_GROUP_CP 0x0 219 #define KGSL_PERFCOUNTER_GROUP_RBBM 0x1 220 #define KGSL_PERFCOUNTER_GROUP_PC 0x2 221 #define KGSL_PERFCOUNTER_GROUP_VFD 0x3 222 #define KGSL_PERFCOUNTER_GROUP_HLSQ 0x4 223 #define KGSL_PERFCOUNTER_GROUP_VPC 0x5 224 #define KGSL_PERFCOUNTER_GROUP_TSE 0x6 225 #define KGSL_PERFCOUNTER_GROUP_RAS 0x7 226 #define KGSL_PERFCOUNTER_GROUP_UCHE 0x8 227 #define KGSL_PERFCOUNTER_GROUP_TP 0x9 228 #define KGSL_PERFCOUNTER_GROUP_SP 0xA 229 #define KGSL_PERFCOUNTER_GROUP_RB 0xB 230 #define KGSL_PERFCOUNTER_GROUP_PWR 0xC 231 #define KGSL_PERFCOUNTER_GROUP_VBIF 0xD 232 #define KGSL_PERFCOUNTER_GROUP_VBIF_PWR 0xE 233 234 #define KGSL_PERFCOUNTER_NOT_USED 0xFFFFFFFF 235 #define KGSL_PERFCOUNTER_BROKEN 0xFFFFFFFE 236 237 /* structure holds list of ibs */ 238 struct kgsl_ibdesc { 239 unsigned int gpuaddr; 240 void *hostptr; 241 unsigned int sizedwords; 242 unsigned int ctrl; 243 }; 244 245 /* ioctls */ 246 #define KGSL_IOC_TYPE 0x09 247 248 /* get misc info about the GPU 249 type should be a value from enum kgsl_property_type 250 value points to a structure that varies based on type 251 sizebytes is sizeof() that structure 252 for KGSL_PROP_DEVICE_INFO, use struct kgsl_devinfo 253 this structure contaings hardware versioning info. 254 for KGSL_PROP_DEVICE_SHADOW, use struct kgsl_shadowprop 255 this is used to find mmap() offset and sizes for mapping 256 struct kgsl_memstore into userspace. 257 */ 258 struct kgsl_device_getproperty { 259 unsigned int type; 260 void *value; 261 unsigned int sizebytes; 262 }; 263 264 #define IOCTL_KGSL_DEVICE_GETPROPERTY \ 265 _IOWR(KGSL_IOC_TYPE, 0x2, struct kgsl_device_getproperty) 266 267 /* IOCTL_KGSL_DEVICE_READ (0x3) - removed 03/2012 268 */ 269 270 /* block until the GPU has executed past a given timestamp 271 * timeout is in milliseconds. 272 */ 273 struct kgsl_device_waittimestamp { 274 unsigned int timestamp; 275 unsigned int timeout; 276 }; 277 278 #define IOCTL_KGSL_DEVICE_WAITTIMESTAMP \ 279 _IOW(KGSL_IOC_TYPE, 0x6, struct kgsl_device_waittimestamp) 280 281 struct kgsl_device_waittimestamp_ctxtid { 282 unsigned int context_id; 283 unsigned int timestamp; 284 unsigned int timeout; 285 }; 286 287 #define IOCTL_KGSL_DEVICE_WAITTIMESTAMP_CTXTID \ 288 _IOW(KGSL_IOC_TYPE, 0x7, struct kgsl_device_waittimestamp_ctxtid) 289 290 /* DEPRECATED: issue indirect commands to the GPU. 291 * drawctxt_id must have been created with IOCTL_KGSL_DRAWCTXT_CREATE 292 * ibaddr and sizedwords must specify a subset of a buffer created 293 * with IOCTL_KGSL_SHAREDMEM_FROM_PMEM 294 * flags may be a mask of KGSL_CONTEXT_ values 295 * timestamp is a returned counter value which can be passed to 296 * other ioctls to determine when the commands have been executed by 297 * the GPU. 298 * 299 * This fucntion is deprecated - consider using IOCTL_KGSL_SUBMIT_COMMANDS 300 * instead 301 */ 302 struct kgsl_ringbuffer_issueibcmds { 303 unsigned int drawctxt_id; 304 unsigned int ibdesc_addr; 305 unsigned int numibs; 306 unsigned int timestamp; /*output param */ 307 unsigned int flags; 308 }; 309 310 #define IOCTL_KGSL_RINGBUFFER_ISSUEIBCMDS \ 311 _IOWR(KGSL_IOC_TYPE, 0x10, struct kgsl_ringbuffer_issueibcmds) 312 313 /* read the most recently executed timestamp value 314 * type should be a value from enum kgsl_timestamp_type 315 */ 316 struct kgsl_cmdstream_readtimestamp { 317 unsigned int type; 318 unsigned int timestamp; /*output param */ 319 }; 320 321 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP_OLD \ 322 _IOR(KGSL_IOC_TYPE, 0x11, struct kgsl_cmdstream_readtimestamp) 323 324 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP \ 325 _IOWR(KGSL_IOC_TYPE, 0x11, struct kgsl_cmdstream_readtimestamp) 326 327 /* free memory when the GPU reaches a given timestamp. 328 * gpuaddr specify a memory region created by a 329 * IOCTL_KGSL_SHAREDMEM_FROM_PMEM call 330 * type should be a value from enum kgsl_timestamp_type 331 */ 332 struct kgsl_cmdstream_freememontimestamp { 333 unsigned int gpuaddr; 334 unsigned int type; 335 unsigned int timestamp; 336 }; 337 338 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP \ 339 _IOW(KGSL_IOC_TYPE, 0x12, struct kgsl_cmdstream_freememontimestamp) 340 341 /* Previous versions of this header had incorrectly defined 342 IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP as a read-only ioctl instead 343 of a write only ioctl. To ensure binary compatability, the following 344 #define will be used to intercept the incorrect ioctl 345 */ 346 347 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP_OLD \ 348 _IOR(KGSL_IOC_TYPE, 0x12, struct kgsl_cmdstream_freememontimestamp) 349 350 /* create a draw context, which is used to preserve GPU state. 351 * The flags field may contain a mask KGSL_CONTEXT_* values 352 */ 353 struct kgsl_drawctxt_create { 354 unsigned int flags; 355 unsigned int drawctxt_id; /*output param */ 356 }; 357 358 #define IOCTL_KGSL_DRAWCTXT_CREATE \ 359 _IOWR(KGSL_IOC_TYPE, 0x13, struct kgsl_drawctxt_create) 360 361 /* destroy a draw context */ 362 struct kgsl_drawctxt_destroy { 363 unsigned int drawctxt_id; 364 }; 365 366 #define IOCTL_KGSL_DRAWCTXT_DESTROY \ 367 _IOW(KGSL_IOC_TYPE, 0x14, struct kgsl_drawctxt_destroy) 368 369 /* add a block of pmem, fb, ashmem or user allocated address 370 * into the GPU address space */ 371 struct kgsl_map_user_mem { 372 int fd; 373 unsigned int gpuaddr; /*output param */ 374 unsigned int len; 375 unsigned int offset; 376 unsigned int hostptr; /*input param */ 377 enum kgsl_user_mem_type memtype; 378 unsigned int flags; 379 }; 380 381 #define IOCTL_KGSL_MAP_USER_MEM \ 382 _IOWR(KGSL_IOC_TYPE, 0x15, struct kgsl_map_user_mem) 383 384 struct kgsl_cmdstream_readtimestamp_ctxtid { 385 unsigned int context_id; 386 unsigned int type; 387 unsigned int timestamp; /*output param */ 388 }; 389 390 #define IOCTL_KGSL_CMDSTREAM_READTIMESTAMP_CTXTID \ 391 _IOWR(KGSL_IOC_TYPE, 0x16, struct kgsl_cmdstream_readtimestamp_ctxtid) 392 393 struct kgsl_cmdstream_freememontimestamp_ctxtid { 394 unsigned int context_id; 395 unsigned int gpuaddr; 396 unsigned int type; 397 unsigned int timestamp; 398 }; 399 400 #define IOCTL_KGSL_CMDSTREAM_FREEMEMONTIMESTAMP_CTXTID \ 401 _IOW(KGSL_IOC_TYPE, 0x17, \ 402 struct kgsl_cmdstream_freememontimestamp_ctxtid) 403 404 /* add a block of pmem or fb into the GPU address space */ 405 struct kgsl_sharedmem_from_pmem { 406 int pmem_fd; 407 unsigned int gpuaddr; /*output param */ 408 unsigned int len; 409 unsigned int offset; 410 }; 411 412 #define IOCTL_KGSL_SHAREDMEM_FROM_PMEM \ 413 _IOWR(KGSL_IOC_TYPE, 0x20, struct kgsl_sharedmem_from_pmem) 414 415 /* remove memory from the GPU's address space */ 416 struct kgsl_sharedmem_free { 417 unsigned int gpuaddr; 418 }; 419 420 #define IOCTL_KGSL_SHAREDMEM_FREE \ 421 _IOW(KGSL_IOC_TYPE, 0x21, struct kgsl_sharedmem_free) 422 423 struct kgsl_cff_user_event { 424 unsigned char cff_opcode; 425 unsigned int op1; 426 unsigned int op2; 427 unsigned int op3; 428 unsigned int op4; 429 unsigned int op5; 430 unsigned int __pad[2]; 431 }; 432 433 #define IOCTL_KGSL_CFF_USER_EVENT \ 434 _IOW(KGSL_IOC_TYPE, 0x31, struct kgsl_cff_user_event) 435 436 struct kgsl_gmem_desc { 437 unsigned int x; 438 unsigned int y; 439 unsigned int width; 440 unsigned int height; 441 unsigned int pitch; 442 }; 443 444 struct kgsl_buffer_desc { 445 void *hostptr; 446 unsigned int gpuaddr; 447 int size; 448 unsigned int format; 449 unsigned int pitch; 450 unsigned int enabled; 451 }; 452 453 struct kgsl_bind_gmem_shadow { 454 unsigned int drawctxt_id; 455 struct kgsl_gmem_desc gmem_desc; 456 unsigned int shadow_x; 457 unsigned int shadow_y; 458 struct kgsl_buffer_desc shadow_buffer; 459 unsigned int buffer_id; 460 }; 461 462 #define IOCTL_KGSL_DRAWCTXT_BIND_GMEM_SHADOW \ 463 _IOW(KGSL_IOC_TYPE, 0x22, struct kgsl_bind_gmem_shadow) 464 465 /* add a block of memory into the GPU address space */ 466 467 /* 468 * IOCTL_KGSL_SHAREDMEM_FROM_VMALLOC deprecated 09/2012 469 * use IOCTL_KGSL_GPUMEM_ALLOC instead 470 */ 471 472 struct kgsl_sharedmem_from_vmalloc { 473 unsigned int gpuaddr; /*output param */ 474 unsigned int hostptr; 475 unsigned int flags; 476 }; 477 478 #define IOCTL_KGSL_SHAREDMEM_FROM_VMALLOC \ 479 _IOWR(KGSL_IOC_TYPE, 0x23, struct kgsl_sharedmem_from_vmalloc) 480 481 /* 482 * This is being deprecated in favor of IOCTL_KGSL_GPUMEM_CACHE_SYNC which 483 * supports both directions (flush and invalidate). This code will still 484 * work, but by definition it will do a flush of the cache which might not be 485 * what you want to have happen on a buffer following a GPU operation. It is 486 * safer to go with IOCTL_KGSL_GPUMEM_CACHE_SYNC 487 */ 488 489 #define IOCTL_KGSL_SHAREDMEM_FLUSH_CACHE \ 490 _IOW(KGSL_IOC_TYPE, 0x24, struct kgsl_sharedmem_free) 491 492 struct kgsl_drawctxt_set_bin_base_offset { 493 unsigned int drawctxt_id; 494 unsigned int offset; 495 }; 496 497 #define IOCTL_KGSL_DRAWCTXT_SET_BIN_BASE_OFFSET \ 498 _IOW(KGSL_IOC_TYPE, 0x25, struct kgsl_drawctxt_set_bin_base_offset) 499 500 enum kgsl_cmdwindow_type { 501 KGSL_CMDWINDOW_MIN = 0x00000000, 502 KGSL_CMDWINDOW_2D = 0x00000000, 503 KGSL_CMDWINDOW_3D = 0x00000001, /* legacy */ 504 KGSL_CMDWINDOW_MMU = 0x00000002, 505 KGSL_CMDWINDOW_ARBITER = 0x000000FF, 506 KGSL_CMDWINDOW_MAX = 0x000000FF, 507 }; 508 509 /* write to the command window */ 510 struct kgsl_cmdwindow_write { 511 enum kgsl_cmdwindow_type target; 512 unsigned int addr; 513 unsigned int data; 514 }; 515 516 #define IOCTL_KGSL_CMDWINDOW_WRITE \ 517 _IOW(KGSL_IOC_TYPE, 0x2e, struct kgsl_cmdwindow_write) 518 519 struct kgsl_gpumem_alloc { 520 unsigned long gpuaddr; 521 size_t size; 522 unsigned int flags; 523 }; 524 525 #define IOCTL_KGSL_GPUMEM_ALLOC \ 526 _IOWR(KGSL_IOC_TYPE, 0x2f, struct kgsl_gpumem_alloc) 527 528 struct kgsl_cff_syncmem { 529 unsigned int gpuaddr; 530 unsigned int len; 531 unsigned int __pad[2]; /* For future binary compatibility */ 532 }; 533 534 #define IOCTL_KGSL_CFF_SYNCMEM \ 535 _IOW(KGSL_IOC_TYPE, 0x30, struct kgsl_cff_syncmem) 536 537 /* 538 * A timestamp event allows the user space to register an action following an 539 * expired timestamp. Note IOCTL_KGSL_TIMESTAMP_EVENT has been redefined to 540 * _IOWR to support fences which need to return a fd for the priv parameter. 541 */ 542 543 struct kgsl_timestamp_event { 544 int type; /* Type of event (see list below) */ 545 unsigned int timestamp; /* Timestamp to trigger event on */ 546 unsigned int context_id; /* Context for the timestamp */ 547 void *priv; /* Pointer to the event specific blob */ 548 size_t len; /* Size of the event specific blob */ 549 }; 550 551 #define IOCTL_KGSL_TIMESTAMP_EVENT_OLD \ 552 _IOW(KGSL_IOC_TYPE, 0x31, struct kgsl_timestamp_event) 553 554 /* A genlock timestamp event releases an existing lock on timestamp expire */ 555 556 #define KGSL_TIMESTAMP_EVENT_GENLOCK 1 557 558 struct kgsl_timestamp_event_genlock { 559 int handle; /* Handle of the genlock lock to release */ 560 }; 561 562 /* A fence timestamp event releases an existing lock on timestamp expire */ 563 564 #define KGSL_TIMESTAMP_EVENT_FENCE 2 565 566 struct kgsl_timestamp_event_fence { 567 int fence_fd; /* Fence to signal */ 568 }; 569 570 /* 571 * Set a property within the kernel. Uses the same structure as 572 * IOCTL_KGSL_GETPROPERTY 573 */ 574 575 #define IOCTL_KGSL_SETPROPERTY \ 576 _IOW(KGSL_IOC_TYPE, 0x32, struct kgsl_device_getproperty) 577 578 #define IOCTL_KGSL_TIMESTAMP_EVENT \ 579 _IOWR(KGSL_IOC_TYPE, 0x33, struct kgsl_timestamp_event) 580 581 /** 582 * struct kgsl_gpumem_alloc_id - argument to IOCTL_KGSL_GPUMEM_ALLOC_ID 583 * @id: returned id value for this allocation. 584 * @flags: mask of KGSL_MEM* values requested and actual flags on return. 585 * @size: requested size of the allocation and actual size on return. 586 * @mmapsize: returned size to pass to mmap() which may be larger than 'size' 587 * @gpuaddr: returned GPU address for the allocation 588 * 589 * Allocate memory for access by the GPU. The flags and size fields are echoed 590 * back by the kernel, so that the caller can know if the request was 591 * adjusted. 592 * 593 * Supported flags: 594 * KGSL_MEMFLAGS_GPUREADONLY: the GPU will be unable to write to the buffer 595 * KGSL_MEMTYPE*: usage hint for debugging aid 596 * KGSL_MEMALIGN*: alignment hint, may be ignored or adjusted by the kernel. 597 * KGSL_MEMFLAGS_USE_CPU_MAP: If set on call and return, the returned GPU 598 * address will be 0. Calling mmap() will set the GPU address. 599 */ 600 struct kgsl_gpumem_alloc_id { 601 unsigned int id; 602 unsigned int flags; 603 unsigned int size; 604 unsigned int mmapsize; 605 unsigned long gpuaddr; 606 /* private: reserved for future use*/ 607 unsigned int __pad[2]; 608 }; 609 610 #define IOCTL_KGSL_GPUMEM_ALLOC_ID \ 611 _IOWR(KGSL_IOC_TYPE, 0x34, struct kgsl_gpumem_alloc_id) 612 613 /** 614 * struct kgsl_gpumem_free_id - argument to IOCTL_KGSL_GPUMEM_FREE_ID 615 * @id: GPU allocation id to free 616 * 617 * Free an allocation by id, in case a GPU address has not been assigned or 618 * is unknown. Freeing an allocation by id with this ioctl or by GPU address 619 * with IOCTL_KGSL_SHAREDMEM_FREE are equivalent. 620 */ 621 struct kgsl_gpumem_free_id { 622 unsigned int id; 623 /* private: reserved for future use*/ 624 unsigned int __pad; 625 }; 626 627 #define IOCTL_KGSL_GPUMEM_FREE_ID \ 628 _IOWR(KGSL_IOC_TYPE, 0x35, struct kgsl_gpumem_free_id) 629 630 /** 631 * struct kgsl_gpumem_get_info - argument to IOCTL_KGSL_GPUMEM_GET_INFO 632 * @gpuaddr: GPU address to query. Also set on return. 633 * @id: GPU allocation id to query. Also set on return. 634 * @flags: returned mask of KGSL_MEM* values. 635 * @size: returned size of the allocation. 636 * @mmapsize: returned size to pass mmap(), which may be larger than 'size' 637 * @useraddr: returned address of the userspace mapping for this buffer 638 * 639 * This ioctl allows querying of all user visible attributes of an existing 640 * allocation, by either the GPU address or the id returned by a previous 641 * call to IOCTL_KGSL_GPUMEM_ALLOC_ID. Legacy allocation ioctls may not 642 * return all attributes so this ioctl can be used to look them up if needed. 643 * 644 */ 645 struct kgsl_gpumem_get_info { 646 unsigned long gpuaddr; 647 unsigned int id; 648 unsigned int flags; 649 unsigned int size; 650 unsigned int mmapsize; 651 unsigned long useraddr; 652 /* private: reserved for future use*/ 653 unsigned int __pad[4]; 654 }; 655 656 #define IOCTL_KGSL_GPUMEM_GET_INFO\ 657 _IOWR(KGSL_IOC_TYPE, 0x36, struct kgsl_gpumem_get_info) 658 659 /** 660 * struct kgsl_gpumem_sync_cache - argument to IOCTL_KGSL_GPUMEM_SYNC_CACHE 661 * @gpuaddr: GPU address of the buffer to sync. 662 * @id: id of the buffer to sync. Either gpuaddr or id is sufficient. 663 * @op: a mask of KGSL_GPUMEM_CACHE_* values 664 * 665 * Sync the L2 cache for memory headed to and from the GPU - this replaces 666 * KGSL_SHAREDMEM_FLUSH_CACHE since it can handle cache management for both 667 * directions 668 * 669 */ 670 struct kgsl_gpumem_sync_cache { 671 unsigned int gpuaddr; 672 unsigned int id; 673 unsigned int op; 674 /* private: reserved for future use*/ 675 unsigned int __pad[2]; /* For future binary compatibility */ 676 }; 677 678 #define KGSL_GPUMEM_CACHE_CLEAN (1 << 0) 679 #define KGSL_GPUMEM_CACHE_TO_GPU KGSL_GPUMEM_CACHE_CLEAN 680 681 #define KGSL_GPUMEM_CACHE_INV (1 << 1) 682 #define KGSL_GPUMEM_CACHE_FROM_GPU KGSL_GPUMEM_CACHE_INV 683 684 #define KGSL_GPUMEM_CACHE_FLUSH \ 685 (KGSL_GPUMEM_CACHE_CLEAN | KGSL_GPUMEM_CACHE_INV) 686 687 #define IOCTL_KGSL_GPUMEM_SYNC_CACHE \ 688 _IOW(KGSL_IOC_TYPE, 0x37, struct kgsl_gpumem_sync_cache) 689 690 /** 691 * struct kgsl_perfcounter_get - argument to IOCTL_KGSL_PERFCOUNTER_GET 692 * @groupid: Performance counter group ID 693 * @countable: Countable to select within the group 694 * @offset: Return offset of the reserved LO counter 695 * @offset_hi: Return offset of the reserved HI counter 696 * 697 * Get an available performance counter from a specified groupid. The offset 698 * of the performance counter will be returned after successfully assigning 699 * the countable to the counter for the specified group. An error will be 700 * returned and an offset of 0 if the groupid is invalid or there are no 701 * more counters left. After successfully getting a perfcounter, the user 702 * must call kgsl_perfcounter_put(groupid, contable) when finished with 703 * the perfcounter to clear up perfcounter resources. 704 * 705 */ 706 struct kgsl_perfcounter_get { 707 unsigned int groupid; 708 unsigned int countable; 709 unsigned int offset; 710 unsigned int offset_hi; 711 /* private: reserved for future use */ 712 unsigned int __pad; /* For future binary compatibility */ 713 }; 714 715 #define IOCTL_KGSL_PERFCOUNTER_GET \ 716 _IOWR(KGSL_IOC_TYPE, 0x38, struct kgsl_perfcounter_get) 717 718 /** 719 * struct kgsl_perfcounter_put - argument to IOCTL_KGSL_PERFCOUNTER_PUT 720 * @groupid: Performance counter group ID 721 * @countable: Countable to release within the group 722 * 723 * Put an allocated performance counter to allow others to have access to the 724 * resource that was previously taken. This is only to be called after 725 * successfully getting a performance counter from kgsl_perfcounter_get(). 726 * 727 */ 728 struct kgsl_perfcounter_put { 729 unsigned int groupid; 730 unsigned int countable; 731 /* private: reserved for future use */ 732 unsigned int __pad[2]; /* For future binary compatibility */ 733 }; 734 735 #define IOCTL_KGSL_PERFCOUNTER_PUT \ 736 _IOW(KGSL_IOC_TYPE, 0x39, struct kgsl_perfcounter_put) 737 738 /** 739 * struct kgsl_perfcounter_query - argument to IOCTL_KGSL_PERFCOUNTER_QUERY 740 * @groupid: Performance counter group ID 741 * @countable: Return active countables array 742 * @size: Size of active countables array 743 * @max_counters: Return total number counters for the group ID 744 * 745 * Query the available performance counters given a groupid. The array 746 * *countables is used to return the current active countables in counters. 747 * The size of the array is passed in so the kernel will only write at most 748 * size or counter->size for the group id. The total number of available 749 * counters for the group ID is returned in max_counters. 750 * If the array or size passed in are invalid, then only the maximum number 751 * of counters will be returned, no data will be written to *countables. 752 * If the groupid is invalid an error code will be returned. 753 * 754 */ 755 struct kgsl_perfcounter_query { 756 unsigned int groupid; 757 /* Array to return the current countable for up to size counters */ 758 unsigned int *countables; 759 unsigned int count; 760 unsigned int max_counters; 761 /* private: reserved for future use */ 762 unsigned int __pad[2]; /* For future binary compatibility */ 763 }; 764 765 #define IOCTL_KGSL_PERFCOUNTER_QUERY \ 766 _IOWR(KGSL_IOC_TYPE, 0x3A, struct kgsl_perfcounter_query) 767 768 /** 769 * struct kgsl_perfcounter_query - argument to IOCTL_KGSL_PERFCOUNTER_QUERY 770 * @groupid: Performance counter group IDs 771 * @countable: Performance counter countable IDs 772 * @value: Return performance counter reads 773 * @size: Size of all arrays (groupid/countable pair and return value) 774 * 775 * Read in the current value of a performance counter given by the groupid 776 * and countable. 777 * 778 */ 779 780 struct kgsl_perfcounter_read_group { 781 unsigned int groupid; 782 unsigned int countable; 783 unsigned long long value; 784 }; 785 786 struct kgsl_perfcounter_read { 787 struct kgsl_perfcounter_read_group *reads; 788 unsigned int count; 789 /* private: reserved for future use */ 790 unsigned int __pad[2]; /* For future binary compatibility */ 791 }; 792 793 #define IOCTL_KGSL_PERFCOUNTER_READ \ 794 _IOWR(KGSL_IOC_TYPE, 0x3B, struct kgsl_perfcounter_read) 795 /* 796 * struct kgsl_gpumem_sync_cache_bulk - argument to 797 * IOCTL_KGSL_GPUMEM_SYNC_CACHE_BULK 798 * @id_list: list of GPU buffer ids of the buffers to sync 799 * @count: number of GPU buffer ids in id_list 800 * @op: a mask of KGSL_GPUMEM_CACHE_* values 801 * 802 * Sync the cache for memory headed to and from the GPU. Certain 803 * optimizations can be made on the cache operation based on the total 804 * size of the working set of memory to be managed. 805 */ 806 struct kgsl_gpumem_sync_cache_bulk { 807 unsigned int *id_list; 808 unsigned int count; 809 unsigned int op; 810 /* private: reserved for future use */ 811 unsigned int __pad[2]; /* For future binary compatibility */ 812 }; 813 814 #define IOCTL_KGSL_GPUMEM_SYNC_CACHE_BULK \ 815 _IOWR(KGSL_IOC_TYPE, 0x3C, struct kgsl_gpumem_sync_cache_bulk) 816 817 /* 818 * struct kgsl_cmd_syncpoint_timestamp 819 * @context_id: ID of a KGSL context 820 * @timestamp: GPU timestamp 821 * 822 * This structure defines a syncpoint comprising a context/timestamp pair. A 823 * list of these may be passed by IOCTL_KGSL_SUBMIT_COMMANDS to define 824 * dependencies that must be met before the command can be submitted to the 825 * hardware 826 */ 827 struct kgsl_cmd_syncpoint_timestamp { 828 unsigned int context_id; 829 unsigned int timestamp; 830 }; 831 832 #define KGSL_CMD_SYNCPOINT_TYPE_TIMESTAMP 0 833 834 struct kgsl_cmd_syncpoint_fence { 835 int fd; 836 }; 837 838 #define KGSL_CMD_SYNCPOINT_TYPE_FENCE 1 839 840 /** 841 * struct kgsl_cmd_syncpoint - Define a sync point for a command batch 842 * @type: type of sync point defined here 843 * @priv: Pointer to the type specific buffer 844 * @size: Size of the type specific buffer 845 * 846 * This structure contains pointers defining a specific command sync point. 847 * The pointer and size should point to a type appropriate structure. 848 */ 849 struct kgsl_cmd_syncpoint { 850 int type; 851 void __user *priv; 852 unsigned int size; 853 }; 854 855 /** 856 * struct kgsl_submit_commands - Argument to IOCTL_KGSL_SUBMIT_COMMANDS 857 * @context_id: KGSL context ID that owns the commands 858 * @flags: 859 * @cmdlist: User pointer to a list of kgsl_ibdesc structures 860 * @numcmds: Number of commands listed in cmdlist 861 * @synclist: User pointer to a list of kgsl_cmd_syncpoint structures 862 * @numsyncs: Number of sync points listed in synclist 863 * @timestamp: On entry the a user defined timestamp, on exist the timestamp 864 * assigned to the command batch 865 * 866 * This structure specifies a command to send to the GPU hardware. This is 867 * similar to kgsl_issueibcmds expect that it doesn't support the legacy way to 868 * submit IB lists and it adds sync points to block the IB until the 869 * dependencies are satisified. This entry point is the new and preferred way 870 * to submit commands to the GPU. 871 */ 872 873 struct kgsl_submit_commands { 874 unsigned int context_id; 875 unsigned int flags; 876 struct kgsl_ibdesc __user *cmdlist; 877 unsigned int numcmds; 878 struct kgsl_cmd_syncpoint __user *synclist; 879 unsigned int numsyncs; 880 unsigned int timestamp; 881 /* private: reserved for future use */ 882 unsigned int __pad[4]; 883 }; 884 885 #define IOCTL_KGSL_SUBMIT_COMMANDS \ 886 _IOWR(KGSL_IOC_TYPE, 0x3D, struct kgsl_submit_commands) 887 888 /** 889 * struct kgsl_device_constraint - device constraint argument 890 * @context_id: KGSL context ID 891 * @type: type of constraint i.e pwrlevel/none 892 * @data: constraint data 893 * @size: size of the constraint data 894 */ 895 struct kgsl_device_constraint { 896 unsigned int type; 897 unsigned int context_id; 898 void __user *data; 899 size_t size; 900 }; 901 902 /* Constraint Type*/ 903 #define KGSL_CONSTRAINT_NONE 0 904 #define KGSL_CONSTRAINT_PWRLEVEL 1 905 906 /* PWRLEVEL constraint level*/ 907 /* set to min frequency */ 908 #define KGSL_CONSTRAINT_PWR_MIN 0 909 /* set to max frequency */ 910 #define KGSL_CONSTRAINT_PWR_MAX 1 911 912 struct kgsl_device_constraint_pwrlevel { 913 unsigned int level; 914 }; 915 916 #ifdef __KERNEL__ 917 #ifdef CONFIG_MSM_KGSL_DRM 918 int kgsl_gem_obj_addr(int drm_fd, int handle, unsigned long *start, 919 unsigned long *len); 920 #else 921 #define kgsl_gem_obj_addr(...) 0 922 #endif 923 #endif 924 #endif /* _MSM_KGSL_H */ 925