1 #ifndef _LINUX_MSM_ION_H 2 #define _LINUX_MSM_ION_H 3 4 #include <linux/ion.h> 5 6 enum msm_ion_heap_types { 7 ION_HEAP_TYPE_MSM_START = ION_HEAP_TYPE_CUSTOM + 1, 8 ION_HEAP_TYPE_IOMMU = ION_HEAP_TYPE_MSM_START, 9 ION_HEAP_TYPE_DMA, 10 ION_HEAP_TYPE_CP, 11 ION_HEAP_TYPE_SECURE_DMA, 12 ION_HEAP_TYPE_REMOVED, 13 }; 14 15 /** 16 * These are the only ids that should be used for Ion heap ids. 17 * The ids listed are the order in which allocation will be attempted 18 * if specified. Don't swap the order of heap ids unless you know what 19 * you are doing! 20 * Id's are spaced by purpose to allow new Id's to be inserted in-between (for 21 * possible fallbacks) 22 */ 23 24 enum ion_heap_ids { 25 INVALID_HEAP_ID = -1, 26 ION_CP_MM_HEAP_ID = 8, 27 ION_CP_MFC_HEAP_ID = 12, 28 ION_CP_WB_HEAP_ID = 16, /* 8660 only */ 29 ION_CAMERA_HEAP_ID = 20, /* 8660 only */ 30 ION_SYSTEM_CONTIG_HEAP_ID = 21, 31 ION_ADSP_HEAP_ID = 22, 32 ION_PIL1_HEAP_ID = 23, /* Currently used for other PIL images */ 33 ION_SF_HEAP_ID = 24, 34 ION_IOMMU_HEAP_ID = 25, 35 ION_PIL2_HEAP_ID = 26, /* Currently used for modem firmware images */ 36 ION_QSECOM_HEAP_ID = 27, 37 ION_AUDIO_HEAP_ID = 28, 38 39 ION_MM_FIRMWARE_HEAP_ID = 29, 40 ION_SYSTEM_HEAP_ID = 30, 41 42 ION_HEAP_ID_RESERVED = 31 /** Bit reserved for ION_FLAG_SECURE flag */ 43 }; 44 45 enum ion_fixed_position { 46 NOT_FIXED, 47 FIXED_LOW, 48 FIXED_MIDDLE, 49 FIXED_HIGH, 50 }; 51 52 enum cp_mem_usage { 53 VIDEO_BITSTREAM = 0x1, 54 VIDEO_PIXEL = 0x2, 55 VIDEO_NONPIXEL = 0x3, 56 MAX_USAGE = 0x4, 57 UNKNOWN = 0x7FFFFFFF, 58 }; 59 60 #define ION_HEAP_CP_MASK (1 << ION_HEAP_TYPE_CP) 61 #define ION_HEAP_TYPE_DMA_MASK (1 << ION_HEAP_TYPE_DMA) 62 63 /** 64 * Flag to use when allocating to indicate that a heap is secure. 65 */ 66 #define ION_FLAG_SECURE (1 << ION_HEAP_ID_RESERVED) 67 68 /** 69 * Flag for clients to force contiguous memort allocation 70 * 71 * Use of this flag is carefully monitored! 72 */ 73 #define ION_FLAG_FORCE_CONTIGUOUS (1 << 30) 74 75 /* 76 * Used in conjunction with heap which pool memory to force an allocation 77 * to come from the page allocator directly instead of from the pool allocation 78 */ 79 #define ION_FLAG_POOL_FORCE_ALLOC (1 << 16) 80 81 /** 82 * Deprecated! Please use the corresponding ION_FLAG_* 83 */ 84 #define ION_SECURE ION_FLAG_SECURE 85 #define ION_FORCE_CONTIGUOUS ION_FLAG_FORCE_CONTIGUOUS 86 87 /** 88 * Macro should be used with ion_heap_ids defined above. 89 */ 90 #define ION_HEAP(bit) (1 << (bit)) 91 92 #define ION_ADSP_HEAP_NAME "adsp" 93 #define ION_VMALLOC_HEAP_NAME "vmalloc" 94 #define ION_KMALLOC_HEAP_NAME "kmalloc" 95 #define ION_AUDIO_HEAP_NAME "audio" 96 #define ION_SF_HEAP_NAME "sf" 97 #define ION_MM_HEAP_NAME "mm" 98 #define ION_CAMERA_HEAP_NAME "camera_preview" 99 #define ION_IOMMU_HEAP_NAME "iommu" 100 #define ION_MFC_HEAP_NAME "mfc" 101 #define ION_WB_HEAP_NAME "wb" 102 #define ION_MM_FIRMWARE_HEAP_NAME "mm_fw" 103 #define ION_PIL1_HEAP_NAME "pil_1" 104 #define ION_PIL2_HEAP_NAME "pil_2" 105 #define ION_QSECOM_HEAP_NAME "qsecom" 106 107 #define ION_SET_CACHED(__cache) (__cache | ION_FLAG_CACHED) 108 #define ION_SET_UNCACHED(__cache) (__cache & ~ION_FLAG_CACHED) 109 110 #define ION_IS_CACHED(__flags) ((__flags) & ION_FLAG_CACHED) 111 112 #ifdef __KERNEL__ 113 114 /* 115 * This flag allows clients when mapping into the IOMMU to specify to 116 * defer un-mapping from the IOMMU until the buffer memory is freed. 117 */ 118 #define ION_IOMMU_UNMAP_DELAYED 1 119 120 /* 121 * This flag allows clients to defer unsecuring a buffer until the buffer 122 * is actually freed. 123 */ 124 #define ION_UNSECURE_DELAYED 1 125 126 /** 127 * struct ion_cp_heap_pdata - defines a content protection heap in the given 128 * platform 129 * @permission_type: Memory ID used to identify the memory to TZ 130 * @align: Alignment requirement for the memory 131 * @secure_base: Base address for securing the heap. 132 * Note: This might be different from actual base address 133 * of this heap in the case of a shared heap. 134 * @secure_size: Memory size for securing the heap. 135 * Note: This might be different from actual size 136 * of this heap in the case of a shared heap. 137 * @fixed_position If nonzero, position in the fixed area. 138 * @iommu_map_all: Indicates whether we should map whole heap into IOMMU. 139 * @iommu_2x_map_domain: Indicates the domain to use for overmapping. 140 * @request_region: function to be called when the number of allocations 141 * goes from 0 -> 1 142 * @release_region: function to be called when the number of allocations 143 * goes from 1 -> 0 144 * @setup_region: function to be called upon ion registration 145 * @memory_type:Memory type used for the heap 146 * @allow_nonsecure_alloc: allow non-secure allocations from this heap. For 147 * secure heaps, this flag must be set so allow non-secure 148 * allocations. For non-secure heaps, this flag is ignored. 149 * 150 */ 151 struct ion_cp_heap_pdata { 152 enum ion_permission_type permission_type; 153 unsigned int align; 154 ion_phys_addr_t secure_base; /* Base addr used when heap is shared */ 155 size_t secure_size; /* Size used for securing heap when heap is shared*/ 156 int is_cma; 157 enum ion_fixed_position fixed_position; 158 int iommu_map_all; 159 int iommu_2x_map_domain; 160 int (*request_region)(void *); 161 int (*release_region)(void *); 162 void *(*setup_region)(void); 163 enum ion_memory_types memory_type; 164 int allow_nonsecure_alloc; 165 }; 166 167 /** 168 * struct ion_co_heap_pdata - defines a carveout heap in the given platform 169 * @adjacent_mem_id: Id of heap that this heap must be adjacent to. 170 * @align: Alignment requirement for the memory 171 * @fixed_position If nonzero, position in the fixed area. 172 * @request_region: function to be called when the number of allocations 173 * goes from 0 -> 1 174 * @release_region: function to be called when the number of allocations 175 * goes from 1 -> 0 176 * @setup_region: function to be called upon ion registration 177 * @memory_type:Memory type used for the heap 178 * 179 */ 180 struct ion_co_heap_pdata { 181 int adjacent_mem_id; 182 unsigned int align; 183 enum ion_fixed_position fixed_position; 184 int (*request_region)(void *); 185 int (*release_region)(void *); 186 void *(*setup_region)(void); 187 enum ion_memory_types memory_type; 188 }; 189 190 #ifdef CONFIG_ION 191 /** 192 * msm_ion_client_create - allocate a client using the ion_device specified in 193 * drivers/gpu/ion/msm/msm_ion.c 194 * 195 * heap_mask and name are the same as ion_client_create, return values 196 * are the same as ion_client_create. 197 */ 198 199 struct ion_client *msm_ion_client_create(unsigned int heap_mask, 200 const char *name); 201 202 /** 203 * ion_handle_get_flags - get the flags for a given handle 204 * 205 * @client - client who allocated the handle 206 * @handle - handle to get the flags 207 * @flags - pointer to store the flags 208 * 209 * Gets the current flags for a handle. These flags indicate various options 210 * of the buffer (caching, security, etc.) 211 */ 212 int ion_handle_get_flags(struct ion_client *client, struct ion_handle *handle, 213 unsigned long *flags); 214 215 216 /** 217 * ion_map_iommu - map the given handle into an iommu 218 * 219 * @client - client who allocated the handle 220 * @handle - handle to map 221 * @domain_num - domain number to map to 222 * @partition_num - partition number to allocate iova from 223 * @align - alignment for the iova 224 * @iova_length - length of iova to map. If the iova length is 225 * greater than the handle length, the remaining 226 * address space will be mapped to a dummy buffer. 227 * @iova - pointer to store the iova address 228 * @buffer_size - pointer to store the size of the buffer 229 * @flags - flags for options to map 230 * @iommu_flags - flags specific to the iommu. 231 * 232 * Maps the handle into the iova space specified via domain number. Iova 233 * will be allocated from the partition specified via partition_num. 234 * Returns 0 on success, negative value on error. 235 */ 236 int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, 237 int domain_num, int partition_num, unsigned long align, 238 unsigned long iova_length, unsigned long *iova, 239 unsigned long *buffer_size, 240 unsigned long flags, unsigned long iommu_flags); 241 242 243 /** 244 * ion_handle_get_size - get the allocated size of a given handle 245 * 246 * @client - client who allocated the handle 247 * @handle - handle to get the size 248 * @size - pointer to store the size 249 * 250 * gives the allocated size of a handle. returns 0 on success, negative 251 * value on error 252 * 253 * NOTE: This is intended to be used only to get a size to pass to map_iommu. 254 * You should *NOT* rely on this for any other usage. 255 */ 256 257 int ion_handle_get_size(struct ion_client *client, struct ion_handle *handle, 258 unsigned long *size); 259 260 /** 261 * ion_unmap_iommu - unmap the handle from an iommu 262 * 263 * @client - client who allocated the handle 264 * @handle - handle to unmap 265 * @domain_num - domain to unmap from 266 * @partition_num - partition to unmap from 267 * 268 * Decrement the reference count on the iommu mapping. If the count is 269 * 0, the mapping will be removed from the iommu. 270 */ 271 void ion_unmap_iommu(struct ion_client *client, struct ion_handle *handle, 272 int domain_num, int partition_num); 273 274 275 /** 276 * ion_secure_heap - secure a heap 277 * 278 * @client - a client that has allocated from the heap heap_id 279 * @heap_id - heap id to secure. 280 * @version - version of content protection 281 * @data - extra data needed for protection 282 * 283 * Secure a heap 284 * Returns 0 on success 285 */ 286 int ion_secure_heap(struct ion_device *dev, int heap_id, int version, 287 void *data); 288 289 /** 290 * ion_unsecure_heap - un-secure a heap 291 * 292 * @client - a client that has allocated from the heap heap_id 293 * @heap_id - heap id to un-secure. 294 * @version - version of content protection 295 * @data - extra data needed for protection 296 * 297 * Un-secure a heap 298 * Returns 0 on success 299 */ 300 int ion_unsecure_heap(struct ion_device *dev, int heap_id, int version, 301 void *data); 302 303 /** 304 * msm_ion_do_cache_op - do cache operations. 305 * 306 * @client - pointer to ION client. 307 * @handle - pointer to buffer handle. 308 * @vaddr - virtual address to operate on. 309 * @len - Length of data to do cache operation on. 310 * @cmd - Cache operation to perform: 311 * ION_IOC_CLEAN_CACHES 312 * ION_IOC_INV_CACHES 313 * ION_IOC_CLEAN_INV_CACHES 314 * 315 * Returns 0 on success 316 */ 317 int msm_ion_do_cache_op(struct ion_client *client, struct ion_handle *handle, 318 void *vaddr, unsigned long len, unsigned int cmd); 319 320 /** 321 * msm_ion_secure_heap - secure a heap. Wrapper around ion_secure_heap. 322 * 323 * @heap_id - heap id to secure. 324 * 325 * Secure a heap 326 * Returns 0 on success 327 */ 328 int msm_ion_secure_heap(int heap_id); 329 330 /** 331 * msm_ion_unsecure_heap - unsecure a heap. Wrapper around ion_unsecure_heap. 332 * 333 * @heap_id - heap id to secure. 334 * 335 * Un-secure a heap 336 * Returns 0 on success 337 */ 338 int msm_ion_unsecure_heap(int heap_id); 339 340 /** 341 * msm_ion_secure_heap_2_0 - secure a heap using 2.0 APIs 342 * Wrapper around ion_secure_heap. 343 * 344 * @heap_id - heap id to secure. 345 * @usage - usage hint to TZ 346 * 347 * Secure a heap 348 * Returns 0 on success 349 */ 350 int msm_ion_secure_heap_2_0(int heap_id, enum cp_mem_usage usage); 351 352 /** 353 * msm_ion_unsecure_heap - unsecure a heap secured with 3.0 APIs. 354 * Wrapper around ion_unsecure_heap. 355 * 356 * @heap_id - heap id to secure. 357 * @usage - usage hint to TZ 358 * 359 * Un-secure a heap 360 * Returns 0 on success 361 */ 362 int msm_ion_unsecure_heap_2_0(int heap_id, enum cp_mem_usage usage); 363 364 /** 365 * msm_ion_secure_buffer - secure an individual buffer 366 * 367 * @client - client who has access to the buffer 368 * @handle - buffer to secure 369 * @usage - usage hint to TZ 370 * @flags - flags for the securing 371 */ 372 int msm_ion_secure_buffer(struct ion_client *client, struct ion_handle *handle, 373 enum cp_mem_usage usage, int flags); 374 375 /** 376 * msm_ion_unsecure_buffer - unsecure an individual buffer 377 * 378 * @client - client who has access to the buffer 379 * @handle - buffer to secure 380 */ 381 int msm_ion_unsecure_buffer(struct ion_client *client, 382 struct ion_handle *handle); 383 #else 384 static inline struct ion_client *msm_ion_client_create(unsigned int heap_mask, 385 const char *name) 386 { 387 return ERR_PTR(-ENODEV); 388 } 389 390 static inline int ion_map_iommu(struct ion_client *client, 391 struct ion_handle *handle, int domain_num, 392 int partition_num, unsigned long align, 393 unsigned long iova_length, unsigned long *iova, 394 unsigned long *buffer_size, 395 unsigned long flags, 396 unsigned long iommu_flags) 397 { 398 return -ENODEV; 399 } 400 401 static inline int ion_handle_get_size(struct ion_client *client, 402 struct ion_handle *handle, unsigned long *size) 403 { 404 return -ENODEV; 405 } 406 407 static inline void ion_unmap_iommu(struct ion_client *client, 408 struct ion_handle *handle, int domain_num, 409 int partition_num) 410 { 411 return; 412 } 413 414 static inline int ion_secure_heap(struct ion_device *dev, int heap_id, 415 int version, void *data) 416 { 417 return -ENODEV; 418 419 } 420 421 static inline int ion_unsecure_heap(struct ion_device *dev, int heap_id, 422 int version, void *data) 423 { 424 return -ENODEV; 425 } 426 427 static inline void ion_mark_dangling_buffers_locked(struct ion_device *dev) 428 { 429 } 430 431 static inline int msm_ion_do_cache_op(struct ion_client *client, 432 struct ion_handle *handle, void *vaddr, 433 unsigned long len, unsigned int cmd) 434 { 435 return -ENODEV; 436 } 437 438 static inline int msm_ion_secure_heap(int heap_id) 439 { 440 return -ENODEV; 441 442 } 443 444 static inline int msm_ion_unsecure_heap(int heap_id) 445 { 446 return -ENODEV; 447 } 448 449 static inline int msm_ion_secure_heap_2_0(int heap_id, enum cp_mem_usage usage) 450 { 451 return -ENODEV; 452 } 453 454 static inline int msm_ion_unsecure_heap_2_0(int heap_id, 455 enum cp_mem_usage usage) 456 { 457 return -ENODEV; 458 } 459 460 static inline int msm_ion_secure_buffer(struct ion_client *client, 461 struct ion_handle *handle, 462 enum cp_mem_usage usage, 463 int flags) 464 { 465 return -ENODEV; 466 } 467 468 static inline int msm_ion_unsecure_buffer(struct ion_client *client, 469 struct ion_handle *handle) 470 { 471 return -ENODEV; 472 } 473 #endif /* CONFIG_ION */ 474 475 #endif /* __KERNEL */ 476 477 /* struct ion_flush_data - data passed to ion for flushing caches 478 * 479 * @handle: handle with data to flush 480 * @fd: fd to flush 481 * @vaddr: userspace virtual address mapped with mmap 482 * @offset: offset into the handle to flush 483 * @length: length of handle to flush 484 * 485 * Performs cache operations on the handle. If p is the start address 486 * of the handle, p + offset through p + offset + length will have 487 * the cache operations performed 488 */ 489 struct ion_flush_data { 490 struct ion_handle *handle; 491 int fd; 492 void *vaddr; 493 unsigned int offset; 494 unsigned int length; 495 }; 496 497 #define ION_IOC_MSM_MAGIC 'M' 498 499 /** 500 * DOC: ION_IOC_CLEAN_CACHES - clean the caches 501 * 502 * Clean the caches of the handle specified. 503 */ 504 #define ION_IOC_CLEAN_CACHES _IOWR(ION_IOC_MSM_MAGIC, 0, \ 505 struct ion_flush_data) 506 /** 507 * DOC: ION_IOC_INV_CACHES - invalidate the caches 508 * 509 * Invalidate the caches of the handle specified. 510 */ 511 #define ION_IOC_INV_CACHES _IOWR(ION_IOC_MSM_MAGIC, 1, \ 512 struct ion_flush_data) 513 /** 514 * DOC: ION_IOC_CLEAN_INV_CACHES - clean and invalidate the caches 515 * 516 * Clean and invalidate the caches of the handle specified. 517 */ 518 #define ION_IOC_CLEAN_INV_CACHES _IOWR(ION_IOC_MSM_MAGIC, 2, \ 519 struct ion_flush_data) 520 521 #endif 522