blob: 3256fbad60326bba518f8383f6eb3caf4466d6e6 [file] [log] [blame]
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -07001/*
2 * include/linux/ion.h
3 *
4 * Copyright (C) 2011 Google, Inc.
Olav Haugan0a852512012-01-09 10:20:55 -08005 * Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved.
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -07006 *
7 * This software is licensed under the terms of the GNU General Public
8 * License version 2, as published by the Free Software Foundation, and
9 * may be copied, distributed, and modified under those terms.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 */
17
18#ifndef _LINUX_ION_H
19#define _LINUX_ION_H
20
Laura Abbottabcb6f72011-10-04 16:26:49 -070021#include <linux/ioctl.h>
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070022#include <linux/types.h>
23
Laura Abbottabcb6f72011-10-04 16:26:49 -070024
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070025struct ion_handle;
26/**
27 * enum ion_heap_types - list of all possible types of heaps
Iliyan Malchevf22301562011-07-06 16:53:21 -070028 * @ION_HEAP_TYPE_SYSTEM: memory allocated via vmalloc
29 * @ION_HEAP_TYPE_SYSTEM_CONTIG: memory allocated via kmalloc
30 * @ION_HEAP_TYPE_CARVEOUT: memory allocated from a prereserved
Olav Hauganb5be7992011-11-18 14:29:02 -080031 * carveout heap, allocations are physically
32 * contiguous
Olav Haugan0a852512012-01-09 10:20:55 -080033 * @ION_HEAP_TYPE_IOMMU: IOMMU memory
34 * @ION_HEAP_TYPE_CP: memory allocated from a prereserved
35 * carveout heap, allocations are physically
36 * contiguous. Used for content protection.
37 * @ION_HEAP_END: helper for iterating over heaps
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070038 */
39enum ion_heap_type {
40 ION_HEAP_TYPE_SYSTEM,
41 ION_HEAP_TYPE_SYSTEM_CONTIG,
42 ION_HEAP_TYPE_CARVEOUT,
Laura Abbott8c017362011-09-22 20:59:12 -070043 ION_HEAP_TYPE_IOMMU,
Olav Haugan0a852512012-01-09 10:20:55 -080044 ION_HEAP_TYPE_CP,
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070045 ION_HEAP_TYPE_CUSTOM, /* must be last so device specific heaps always
46 are at the end of this enum */
Rebecca Schultz Zavine6ee1242011-06-30 12:19:55 -070047 ION_NUM_HEAPS,
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070048};
49
Iliyan Malchevf22301562011-07-06 16:53:21 -070050#define ION_HEAP_SYSTEM_MASK (1 << ION_HEAP_TYPE_SYSTEM)
51#define ION_HEAP_SYSTEM_CONTIG_MASK (1 << ION_HEAP_TYPE_SYSTEM_CONTIG)
52#define ION_HEAP_CARVEOUT_MASK (1 << ION_HEAP_TYPE_CARVEOUT)
Olav Haugan0a852512012-01-09 10:20:55 -080053#define ION_HEAP_CP_MASK (1 << ION_HEAP_TYPE_CP)
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -070054
Laura Abbotta2e93632011-08-19 13:36:32 -070055
56/**
57 * These are the only ids that should be used for Ion heap ids.
58 * The ids listed are the order in which allocation will be attempted
59 * if specified. Don't swap the order of heap ids unless you know what
60 * you are doing!
Olav Hauganb5be7992011-11-18 14:29:02 -080061 * Id's are spaced by purpose to allow new Id's to be inserted in-between (for
62 * possible fallbacks)
Laura Abbotta2e93632011-08-19 13:36:32 -070063 */
64
65enum ion_heap_ids {
Olav Haugan42ebe712012-01-10 16:30:58 -080066 INVALID_HEAP_ID = -1,
Olav Hauganb5be7992011-11-18 14:29:02 -080067 ION_CP_MM_HEAP_ID = 8,
68 ION_CP_MFC_HEAP_ID = 12,
69 ION_CP_WB_HEAP_ID = 16, /* 8660 only */
70 ION_CAMERA_HEAP_ID = 20, /* 8660 only */
71 ION_SF_HEAP_ID = 24,
Olav Haugan9e123f92012-02-15 15:41:48 -080072 ION_IOMMU_HEAP_ID = 25,
Olav Haugan80854eb2012-01-12 12:00:23 -080073 ION_QSECOM_HEAP_ID = 27,
Olav Hauganb5be7992011-11-18 14:29:02 -080074 ION_AUDIO_HEAP_ID = 28,
75
Olav Haugan42ebe712012-01-10 16:30:58 -080076 ION_MM_FIRMWARE_HEAP_ID = 29,
Olav Hauganb5be7992011-11-18 14:29:02 -080077 ION_SYSTEM_HEAP_ID = 30,
78
79 ION_HEAP_ID_RESERVED = 31 /** Bit reserved for ION_SECURE flag */
Laura Abbotta2e93632011-08-19 13:36:32 -070080};
81
Olav Hauganb5be7992011-11-18 14:29:02 -080082/**
83 * Flag to use when allocating to indicate that a heap is secure.
84 */
85#define ION_SECURE (1 << ION_HEAP_ID_RESERVED)
86
87/**
88 * Macro should be used with ion_heap_ids defined above.
89 */
90#define ION_HEAP(bit) (1 << (bit))
91
Laura Abbotta2e93632011-08-19 13:36:32 -070092#define ION_VMALLOC_HEAP_NAME "vmalloc"
Olav Hauganb5be7992011-11-18 14:29:02 -080093#define ION_AUDIO_HEAP_NAME "audio"
94#define ION_SF_HEAP_NAME "sf"
95#define ION_MM_HEAP_NAME "mm"
96#define ION_CAMERA_HEAP_NAME "camera_preview"
Laura Abbott8c017362011-09-22 20:59:12 -070097#define ION_IOMMU_HEAP_NAME "iommu"
Olav Hauganb5be7992011-11-18 14:29:02 -080098#define ION_MFC_HEAP_NAME "mfc"
99#define ION_WB_HEAP_NAME "wb"
Olav Haugan42ebe712012-01-10 16:30:58 -0800100#define ION_MM_FIRMWARE_HEAP_NAME "mm_fw"
Olav Haugan80854eb2012-01-12 12:00:23 -0800101#define ION_QSECOM_HEAP_NAME "qsecom"
Laura Abbottcaafeea2011-12-13 11:43:10 -0800102#define ION_FMEM_HEAP_NAME "fmem"
Laura Abbotta2e93632011-08-19 13:36:32 -0700103
Laura Abbott894fd582011-08-19 13:33:56 -0700104#define CACHED 1
105#define UNCACHED 0
106
107#define ION_CACHE_SHIFT 0
108
109#define ION_SET_CACHE(__cache) ((__cache) << ION_CACHE_SHIFT)
110
Laura Abbott35412032011-09-29 09:50:06 -0700111#define ION_IS_CACHED(__flags) ((__flags) & (1 << ION_CACHE_SHIFT))
112
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700113#ifdef __KERNEL__
Laura Abbott65576962011-10-31 12:13:25 -0700114#include <linux/err.h>
Laura Abbottcffdff52011-09-23 10:40:19 -0700115#include <mach/ion.h>
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700116struct ion_device;
117struct ion_heap;
118struct ion_mapper;
119struct ion_client;
120struct ion_buffer;
121
122/* This should be removed some day when phys_addr_t's are fully
123 plumbed in the kernel, and all instances of ion_phys_addr_t should
124 be converted to phys_addr_t. For the time being many kernel interfaces
125 do not accept phys_addr_t's that would have to */
126#define ion_phys_addr_t unsigned long
Laura Abbottcaafeea2011-12-13 11:43:10 -0800127#define ion_virt_addr_t unsigned long
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700128
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700129/**
130 * struct ion_platform_heap - defines a heap in the given platform
131 * @type: type of the heap from ion_heap_type enum
Olav Hauganee0f7802011-12-19 13:28:57 -0800132 * @id: unique identifier for heap. When allocating (lower numbers
Olav Hauganb5be7992011-11-18 14:29:02 -0800133 * will be allocated from first)
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700134 * @name: used for debug purposes
135 * @base: base address of heap in physical memory if applicable
136 * @size: size of the heap in bytes if applicable
Laura Abbottcaafeea2011-12-13 11:43:10 -0800137 * @memory_type:Memory type used for the heap
138 * @extra_data: Extra data specific to each heap type
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700139 */
140struct ion_platform_heap {
141 enum ion_heap_type type;
Rebecca Schultz Zavine6ee1242011-06-30 12:19:55 -0700142 unsigned int id;
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700143 const char *name;
144 ion_phys_addr_t base;
145 size_t size;
Laura Abbotta2e93632011-08-19 13:36:32 -0700146 enum ion_memory_types memory_type;
Olav Haugan0703dbf2011-12-19 17:53:38 -0800147 void *extra_data;
148};
149
Laura Abbottcaafeea2011-12-13 11:43:10 -0800150/**
151 * struct ion_cp_heap_pdata - defines a content protection heap in the given
152 * platform
153 * @permission_type: Memory ID used to identify the memory to TZ
154 * @align: Alignment requirement for the memory
155 * @secure_base: Base address for securing the heap.
156 * Note: This might be different from actual base address
157 * of this heap in the case of a shared heap.
158 * @secure_size: Memory size for securing the heap.
159 * Note: This might be different from actual size
160 * of this heap in the case of a shared heap.
161 * @reusable Flag indicating whether this heap is reusable of not.
162 * (see FMEM)
Olav Hauganf6dc7742012-02-15 09:11:55 -0800163 * @mem_is_fmem Flag indicating whether this memory is coming from fmem
164 * or not.
Laura Abbottcaafeea2011-12-13 11:43:10 -0800165 * @virt_addr: Virtual address used when using fmem.
166 * @request_region: function to be called when the number of allocations
167 * goes from 0 -> 1
168 * @release_region: function to be called when the number of allocations
169 * goes from 1 -> 0
170 * @setup_region: function to be called upon ion registration
171 *
172 */
Olav Haugan0703dbf2011-12-19 17:53:38 -0800173struct ion_cp_heap_pdata {
Olav Haugan0a852512012-01-09 10:20:55 -0800174 enum ion_permission_type permission_type;
Olav Haugan42ebe712012-01-10 16:30:58 -0800175 unsigned int align;
176 ion_phys_addr_t secure_base; /* Base addr used when heap is shared */
177 size_t secure_size; /* Size used for securing heap when heap is shared*/
Laura Abbottcaafeea2011-12-13 11:43:10 -0800178 int reusable;
Olav Hauganf6dc7742012-02-15 09:11:55 -0800179 int mem_is_fmem;
Laura Abbottcaafeea2011-12-13 11:43:10 -0800180 ion_virt_addr_t *virt_addr;
Olav Hauganee0f7802011-12-19 13:28:57 -0800181 int (*request_region)(void *);
182 int (*release_region)(void *);
Alex Bird8a3ede32011-11-07 12:33:42 -0800183 void *(*setup_region)(void);
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700184};
185
Laura Abbottcaafeea2011-12-13 11:43:10 -0800186/**
187 * struct ion_co_heap_pdata - defines a carveout heap in the given platform
188 * @adjacent_mem_id: Id of heap that this heap must be adjacent to.
189 * @align: Alignment requirement for the memory
Olav Hauganf6dc7742012-02-15 09:11:55 -0800190 * @mem_is_fmem Flag indicating whether this memory is coming from fmem
191 * or not.
Laura Abbottcaafeea2011-12-13 11:43:10 -0800192 * @request_region: function to be called when the number of allocations
193 * goes from 0 -> 1
194 * @release_region: function to be called when the number of allocations
195 * goes from 1 -> 0
196 * @setup_region: function to be called upon ion registration
197 *
198 */
Olav Haugan0703dbf2011-12-19 17:53:38 -0800199struct ion_co_heap_pdata {
Olav Haugan42ebe712012-01-10 16:30:58 -0800200 int adjacent_mem_id;
201 unsigned int align;
Olav Hauganf6dc7742012-02-15 09:11:55 -0800202 int mem_is_fmem;
Olav Haugan0703dbf2011-12-19 17:53:38 -0800203 int (*request_region)(void *);
204 int (*release_region)(void *);
205 void *(*setup_region)(void);
206};
207
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700208/**
209 * struct ion_platform_data - array of platform heaps passed from board file
Alex Bird27ca6612011-11-01 14:40:06 -0700210 * @nr: number of structures in the array
211 * @request_region: function to be called when the number of allocations goes
212 * from 0 -> 1
213 * @release_region: function to be called when the number of allocations goes
214 * from 1 -> 0
215 * @setup_region: function to be called upon ion registration
216 * @heaps: array of platform_heap structions
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700217 *
218 * Provided by the board file in the form of platform data to a platform device.
219 */
220struct ion_platform_data {
221 int nr;
Olav Hauganee0f7802011-12-19 13:28:57 -0800222 int (*request_region)(void *);
223 int (*release_region)(void *);
Alex Bird27ca6612011-11-01 14:40:06 -0700224 void *(*setup_region)(void);
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700225 struct ion_platform_heap heaps[];
226};
227
Jordan Crouse8cd48322011-10-12 17:05:19 -0600228#ifdef CONFIG_ION
229
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700230/**
231 * ion_client_create() - allocate a client and returns it
232 * @dev: the global ion device
233 * @heap_mask: mask of heaps this client can allocate from
234 * @name: used for debugging
235 */
236struct ion_client *ion_client_create(struct ion_device *dev,
237 unsigned int heap_mask, const char *name);
238
239/**
Laura Abbott302911d2011-08-15 17:12:57 -0700240 * msm_ion_client_create - allocate a client using the ion_device specified in
241 * drivers/gpu/ion/msm/msm_ion.c
242 *
243 * heap_mask and name are the same as ion_client_create, return values
244 * are the same as ion_client_create.
245 */
246
247struct ion_client *msm_ion_client_create(unsigned int heap_mask,
248 const char *name);
249
250/**
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700251 * ion_client_destroy() - free's a client and all it's handles
252 * @client: the client
253 *
254 * Free the provided client and all it's resources including
255 * any handles it is holding.
256 */
257void ion_client_destroy(struct ion_client *client);
258
259/**
260 * ion_alloc - allocate ion memory
261 * @client: the client
262 * @len: size of the allocation
263 * @align: requested allocation alignment, lots of hardware blocks have
264 * alignment requirements of some kind
Rebecca Schultz Zavine6ee1242011-06-30 12:19:55 -0700265 * @flags: mask of heaps to allocate from, if multiple bits are set
266 * heaps will be tried in order from lowest to highest order bit
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700267 *
268 * Allocate memory in one of the heaps provided in heap mask and return
269 * an opaque handle to it.
270 */
271struct ion_handle *ion_alloc(struct ion_client *client, size_t len,
272 size_t align, unsigned int flags);
273
274/**
275 * ion_free - free a handle
276 * @client: the client
277 * @handle: the handle to free
278 *
279 * Free the provided handle.
280 */
281void ion_free(struct ion_client *client, struct ion_handle *handle);
282
283/**
284 * ion_phys - returns the physical address and len of a handle
285 * @client: the client
286 * @handle: the handle
287 * @addr: a pointer to put the address in
288 * @len: a pointer to put the length in
289 *
290 * This function queries the heap for a particular handle to get the
291 * handle's physical address. It't output is only correct if
292 * a heap returns physically contiguous memory -- in other cases
293 * this api should not be implemented -- ion_map_dma should be used
294 * instead. Returns -EINVAL if the handle is invalid. This has
295 * no implications on the reference counting of the handle --
296 * the returned value may not be valid if the caller is not
297 * holding a reference.
298 */
299int ion_phys(struct ion_client *client, struct ion_handle *handle,
300 ion_phys_addr_t *addr, size_t *len);
301
302/**
303 * ion_map_kernel - create mapping for the given handle
304 * @client: the client
305 * @handle: handle to map
Laura Abbott894fd582011-08-19 13:33:56 -0700306 * @flags: flags for this mapping
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700307 *
308 * Map the given handle into the kernel and return a kernel address that
Laura Abbott894fd582011-08-19 13:33:56 -0700309 * can be used to access this address. If no flags are specified, this
310 * will return a non-secure uncached mapping.
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700311 */
Laura Abbott894fd582011-08-19 13:33:56 -0700312void *ion_map_kernel(struct ion_client *client, struct ion_handle *handle,
313 unsigned long flags);
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700314
315/**
316 * ion_unmap_kernel() - destroy a kernel mapping for a handle
317 * @client: the client
318 * @handle: handle to unmap
319 */
320void ion_unmap_kernel(struct ion_client *client, struct ion_handle *handle);
321
322/**
323 * ion_map_dma - create a dma mapping for a given handle
324 * @client: the client
325 * @handle: handle to map
326 *
327 * Return an sglist describing the given handle
328 */
329struct scatterlist *ion_map_dma(struct ion_client *client,
Laura Abbott894fd582011-08-19 13:33:56 -0700330 struct ion_handle *handle,
331 unsigned long flags);
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700332
333/**
334 * ion_unmap_dma() - destroy a dma mapping for a handle
335 * @client: the client
336 * @handle: handle to unmap
337 */
338void ion_unmap_dma(struct ion_client *client, struct ion_handle *handle);
339
340/**
341 * ion_share() - given a handle, obtain a buffer to pass to other clients
342 * @client: the client
343 * @handle: the handle to share
344 *
Iliyan Malchev3fe24362011-08-09 14:42:08 -0700345 * Given a handle, return a buffer, which exists in a global name
346 * space, and can be passed to other clients. Should be passed into ion_import
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700347 * to obtain a new handle for this buffer.
Iliyan Malchev3fe24362011-08-09 14:42:08 -0700348 *
349 * NOTE: This function does do not an extra reference. The burden is on the
350 * caller to make sure the buffer doesn't go away while it's being passed to
351 * another client. That is, ion_free should not be called on this handle until
352 * the buffer has been imported into the other client.
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700353 */
354struct ion_buffer *ion_share(struct ion_client *client,
355 struct ion_handle *handle);
356
357/**
358 * ion_import() - given an buffer in another client, import it
359 * @client: this blocks client
360 * @buffer: the buffer to import (as obtained from ion_share)
361 *
362 * Given a buffer, add it to the client and return the handle to use to refer
363 * to it further. This is called to share a handle from one kernel client to
364 * another.
365 */
366struct ion_handle *ion_import(struct ion_client *client,
367 struct ion_buffer *buffer);
368
369/**
370 * ion_import_fd() - given an fd obtained via ION_IOC_SHARE ioctl, import it
371 * @client: this blocks client
372 * @fd: the fd
373 *
374 * A helper function for drivers that will be recieving ion buffers shared
375 * with them from userspace. These buffers are represented by a file
376 * descriptor obtained as the return from the ION_IOC_SHARE ioctl.
377 * This function coverts that fd into the underlying buffer, and returns
378 * the handle to use to refer to it further.
379 */
380struct ion_handle *ion_import_fd(struct ion_client *client, int fd);
Laura Abbott273dd8e2011-10-12 14:26:33 -0700381
Laura Abbott273dd8e2011-10-12 14:26:33 -0700382/**
383 * ion_handle_get_flags - get the flags for a given handle
384 *
385 * @client - client who allocated the handle
386 * @handle - handle to get the flags
387 * @flags - pointer to store the flags
388 *
389 * Gets the current flags for a handle. These flags indicate various options
390 * of the buffer (caching, security, etc.)
391 */
392int ion_handle_get_flags(struct ion_client *client, struct ion_handle *handle,
393 unsigned long *flags);
394
Laura Abbott8c017362011-09-22 20:59:12 -0700395
396/**
397 * ion_map_iommu - map the given handle into an iommu
398 *
399 * @client - client who allocated the handle
400 * @handle - handle to map
401 * @domain_num - domain number to map to
402 * @partition_num - partition number to allocate iova from
403 * @align - alignment for the iova
404 * @iova_length - length of iova to map. If the iova length is
405 * greater than the handle length, the remaining
406 * address space will be mapped to a dummy buffer.
407 * @iova - pointer to store the iova address
408 * @buffer_size - pointer to store the size of the buffer
409 * @flags - flags for options to map
410 *
411 * Maps the handle into the iova space specified via domain number. Iova
412 * will be allocated from the partition specified via partition_num.
413 * Returns 0 on success, negative value on error.
414 */
415int ion_map_iommu(struct ion_client *client, struct ion_handle *handle,
416 int domain_num, int partition_num, unsigned long align,
417 unsigned long iova_length, unsigned long *iova,
418 unsigned long *buffer_size,
419 unsigned long flags);
420
421
422/**
423 * ion_handle_get_size - get the allocated size of a given handle
424 *
425 * @client - client who allocated the handle
426 * @handle - handle to get the size
427 * @size - pointer to store the size
428 *
429 * gives the allocated size of a handle. returns 0 on success, negative
430 * value on error
431 *
432 * NOTE: This is intended to be used only to get a size to pass to map_iommu.
433 * You should *NOT* rely on this for any other usage.
434 */
435
436int ion_handle_get_size(struct ion_client *client, struct ion_handle *handle,
437 unsigned long *size);
438
439/**
440 * ion_unmap_iommu - unmap the handle from an iommu
441 *
442 * @client - client who allocated the handle
443 * @handle - handle to unmap
444 * @domain_num - domain to unmap from
445 * @partition_num - partition to unmap from
446 *
447 * Decrement the reference count on the iommu mapping. If the count is
448 * 0, the mapping will be removed from the iommu.
449 */
450void ion_unmap_iommu(struct ion_client *client, struct ion_handle *handle,
451 int domain_num, int partition_num);
452
453
Olav Haugan0a852512012-01-09 10:20:55 -0800454/**
455 * ion_secure_heap - secure a heap
456 *
457 * @client - a client that has allocated from the heap heap_id
458 * @heap_id - heap id to secure.
459 *
460 * Secure a heap
461 * Returns 0 on success
462 */
463int ion_secure_heap(struct ion_device *dev, int heap_id);
464
465/**
466 * ion_unsecure_heap - un-secure a heap
467 *
468 * @client - a client that has allocated from the heap heap_id
469 * @heap_id - heap id to un-secure.
470 *
471 * Un-secure a heap
472 * Returns 0 on success
473 */
474int ion_unsecure_heap(struct ion_device *dev, int heap_id);
475
476/**
477 * msm_ion_secure_heap - secure a heap. Wrapper around ion_secure_heap.
478 *
479 * @heap_id - heap id to secure.
480 *
481 * Secure a heap
482 * Returns 0 on success
483 */
484int msm_ion_secure_heap(int heap_id);
485
486/**
487 * msm_ion_unsecure_heap - unsecure a heap. Wrapper around ion_unsecure_heap.
488 *
489 * @heap_id - heap id to secure.
490 *
491 * Un-secure a heap
492 * Returns 0 on success
493 */
494int msm_ion_unsecure_heap(int heap_id);
495
Jordan Crouse8cd48322011-10-12 17:05:19 -0600496#else
497static inline struct ion_client *ion_client_create(struct ion_device *dev,
498 unsigned int heap_mask, const char *name)
499{
500 return ERR_PTR(-ENODEV);
501}
Laura Abbott273dd8e2011-10-12 14:26:33 -0700502
Jordan Crouse8cd48322011-10-12 17:05:19 -0600503static inline struct ion_client *msm_ion_client_create(unsigned int heap_mask,
504 const char *name)
505{
506 return ERR_PTR(-ENODEV);
507}
508
509static inline void ion_client_destroy(struct ion_client *client) { }
510
511static inline struct ion_handle *ion_alloc(struct ion_client *client,
512 size_t len, size_t align, unsigned int flags)
513{
514 return ERR_PTR(-ENODEV);
515}
516
517static inline void ion_free(struct ion_client *client,
518 struct ion_handle *handle) { }
519
520
521static inline int ion_phys(struct ion_client *client,
522 struct ion_handle *handle, ion_phys_addr_t *addr, size_t *len)
523{
524 return -ENODEV;
525}
526
527static inline void *ion_map_kernel(struct ion_client *client,
528 struct ion_handle *handle, unsigned long flags)
529{
530 return ERR_PTR(-ENODEV);
531}
532
533static inline void ion_unmap_kernel(struct ion_client *client,
534 struct ion_handle *handle) { }
535
536static inline struct scatterlist *ion_map_dma(struct ion_client *client,
537 struct ion_handle *handle, unsigned long flags)
538{
539 return ERR_PTR(-ENODEV);
540}
541
542static inline void ion_unmap_dma(struct ion_client *client,
543 struct ion_handle *handle) { }
544
545static inline struct ion_buffer *ion_share(struct ion_client *client,
546 struct ion_handle *handle)
547{
548 return ERR_PTR(-ENODEV);
549}
550
551static inline struct ion_handle *ion_import(struct ion_client *client,
552 struct ion_buffer *buffer)
553{
554 return ERR_PTR(-ENODEV);
555}
556
557static inline struct ion_handle *ion_import_fd(struct ion_client *client,
558 int fd)
559{
560 return ERR_PTR(-ENODEV);
561}
562
563static inline int ion_handle_get_flags(struct ion_client *client,
564 struct ion_handle *handle, unsigned long *flags)
565{
566 return -ENODEV;
567}
Laura Abbott8c017362011-09-22 20:59:12 -0700568
569static inline int ion_map_iommu(struct ion_client *client,
570 struct ion_handle *handle, int domain_num,
571 int partition_num, unsigned long align,
572 unsigned long iova_length, unsigned long *iova,
573 unsigned long flags)
574{
575 return -ENODEV;
576}
577
578static inline void ion_unmap_iommu(struct ion_client *client,
579 struct ion_handle *handle, int domain_num,
580 int partition_num)
581{
582 return;
583}
584
Olav Haugan0a852512012-01-09 10:20:55 -0800585static inline int ion_secure_heap(struct ion_device *dev, int heap_id)
586{
587 return -ENODEV;
Laura Abbott8c017362011-09-22 20:59:12 -0700588
Olav Haugan0a852512012-01-09 10:20:55 -0800589}
590
591static inline int ion_unsecure_heap(struct ion_device *dev, int heap_id)
592{
593 return -ENODEV;
594}
595
596static inline int msm_ion_secure_heap(int heap_id)
597{
598 return -ENODEV;
599
600}
601
602static inline int msm_ion_unsecure_heap(int heap_id)
603{
604 return -ENODEV;
605}
Jordan Crouse8cd48322011-10-12 17:05:19 -0600606#endif /* CONFIG_ION */
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700607#endif /* __KERNEL__ */
608
609/**
610 * DOC: Ion Userspace API
611 *
612 * create a client by opening /dev/ion
613 * most operations handled via following ioctls
614 *
615 */
616
617/**
618 * struct ion_allocation_data - metadata passed from userspace for allocations
619 * @len: size of the allocation
620 * @align: required alignment of the allocation
621 * @flags: flags passed to heap
622 * @handle: pointer that will be populated with a cookie to use to refer
623 * to this allocation
624 *
625 * Provided by userspace as an argument to the ioctl
626 */
627struct ion_allocation_data {
628 size_t len;
629 size_t align;
630 unsigned int flags;
631 struct ion_handle *handle;
632};
633
634/**
635 * struct ion_fd_data - metadata passed to/from userspace for a handle/fd pair
636 * @handle: a handle
637 * @fd: a file descriptor representing that handle
638 *
639 * For ION_IOC_SHARE or ION_IOC_MAP userspace populates the handle field with
640 * the handle returned from ion alloc, and the kernel returns the file
641 * descriptor to share or map in the fd field. For ION_IOC_IMPORT, userspace
642 * provides the file descriptor and the kernel returns the handle.
643 */
644struct ion_fd_data {
645 struct ion_handle *handle;
646 int fd;
647};
648
649/**
650 * struct ion_handle_data - a handle passed to/from the kernel
651 * @handle: a handle
652 */
653struct ion_handle_data {
654 struct ion_handle *handle;
655};
656
Rebecca Schultz Zavine6ee1242011-06-30 12:19:55 -0700657/**
658 * struct ion_custom_data - metadata passed to/from userspace for a custom ioctl
659 * @cmd: the custom ioctl function to call
660 * @arg: additional data to pass to the custom ioctl, typically a user
661 * pointer to a predefined structure
662 *
663 * This works just like the regular cmd and arg fields of an ioctl.
664 */
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700665struct ion_custom_data {
666 unsigned int cmd;
667 unsigned long arg;
668};
669
Laura Abbottabcb6f72011-10-04 16:26:49 -0700670
671/* struct ion_flush_data - data passed to ion for flushing caches
672 *
673 * @handle: handle with data to flush
Laura Abbotte80ea012011-11-18 18:36:47 -0800674 * @fd: fd to flush
Laura Abbottabcb6f72011-10-04 16:26:49 -0700675 * @vaddr: userspace virtual address mapped with mmap
676 * @offset: offset into the handle to flush
677 * @length: length of handle to flush
678 *
679 * Performs cache operations on the handle. If p is the start address
680 * of the handle, p + offset through p + offset + length will have
681 * the cache operations performed
682 */
683struct ion_flush_data {
684 struct ion_handle *handle;
Laura Abbotte80ea012011-11-18 18:36:47 -0800685 int fd;
Laura Abbottabcb6f72011-10-04 16:26:49 -0700686 void *vaddr;
687 unsigned int offset;
688 unsigned int length;
689};
Laura Abbott273dd8e2011-10-12 14:26:33 -0700690
691/* struct ion_flag_data - information about flags for this buffer
692 *
693 * @handle: handle to get flags from
694 * @flags: flags of this handle
695 *
696 * Takes handle as an input and outputs the flags from the handle
697 * in the flag field.
698 */
699struct ion_flag_data {
700 struct ion_handle *handle;
701 unsigned long flags;
702};
703
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700704#define ION_IOC_MAGIC 'I'
705
706/**
707 * DOC: ION_IOC_ALLOC - allocate memory
708 *
709 * Takes an ion_allocation_data struct and returns it with the handle field
710 * populated with the opaque handle for the allocation.
711 */
712#define ION_IOC_ALLOC _IOWR(ION_IOC_MAGIC, 0, \
713 struct ion_allocation_data)
714
715/**
716 * DOC: ION_IOC_FREE - free memory
717 *
718 * Takes an ion_handle_data struct and frees the handle.
719 */
720#define ION_IOC_FREE _IOWR(ION_IOC_MAGIC, 1, struct ion_handle_data)
721
722/**
723 * DOC: ION_IOC_MAP - get a file descriptor to mmap
724 *
725 * Takes an ion_fd_data struct with the handle field populated with a valid
726 * opaque handle. Returns the struct with the fd field set to a file
727 * descriptor open in the current address space. This file descriptor
728 * can then be used as an argument to mmap.
729 */
730#define ION_IOC_MAP _IOWR(ION_IOC_MAGIC, 2, struct ion_fd_data)
731
732/**
733 * DOC: ION_IOC_SHARE - creates a file descriptor to use to share an allocation
734 *
735 * Takes an ion_fd_data struct with the handle field populated with a valid
736 * opaque handle. Returns the struct with the fd field set to a file
737 * descriptor open in the current address space. This file descriptor
738 * can then be passed to another process. The corresponding opaque handle can
739 * be retrieved via ION_IOC_IMPORT.
740 */
741#define ION_IOC_SHARE _IOWR(ION_IOC_MAGIC, 4, struct ion_fd_data)
742
743/**
744 * DOC: ION_IOC_IMPORT - imports a shared file descriptor
745 *
746 * Takes an ion_fd_data struct with the fd field populated with a valid file
747 * descriptor obtained from ION_IOC_SHARE and returns the struct with the handle
748 * filed set to the corresponding opaque handle.
749 */
750#define ION_IOC_IMPORT _IOWR(ION_IOC_MAGIC, 5, int)
751
752/**
753 * DOC: ION_IOC_CUSTOM - call architecture specific ion ioctl
754 *
755 * Takes the argument of the architecture specific ioctl to call and
756 * passes appropriate userdata for that ioctl
757 */
758#define ION_IOC_CUSTOM _IOWR(ION_IOC_MAGIC, 6, struct ion_custom_data)
759
Laura Abbottabcb6f72011-10-04 16:26:49 -0700760
761/**
762 * DOC: ION_IOC_CLEAN_CACHES - clean the caches
763 *
764 * Clean the caches of the handle specified.
765 */
766#define ION_IOC_CLEAN_CACHES _IOWR(ION_IOC_MAGIC, 7, \
767 struct ion_flush_data)
768/**
769 * DOC: ION_MSM_IOC_INV_CACHES - invalidate the caches
770 *
771 * Invalidate the caches of the handle specified.
772 */
773#define ION_IOC_INV_CACHES _IOWR(ION_IOC_MAGIC, 8, \
774 struct ion_flush_data)
775/**
776 * DOC: ION_MSM_IOC_CLEAN_CACHES - clean and invalidate the caches
777 *
778 * Clean and invalidate the caches of the handle specified.
779 */
780#define ION_IOC_CLEAN_INV_CACHES _IOWR(ION_IOC_MAGIC, 9, \
781 struct ion_flush_data)
Laura Abbott273dd8e2011-10-12 14:26:33 -0700782
783/**
784 * DOC: ION_IOC_GET_FLAGS - get the flags of the handle
785 *
786 * Gets the flags of the current handle which indicate cachability,
787 * secure state etc.
788 */
789#define ION_IOC_GET_FLAGS _IOWR(ION_IOC_MAGIC, 10, \
790 struct ion_flag_data)
Rebecca Schultz Zavinc80005a2011-06-29 19:44:29 -0700791#endif /* _LINUX_ION_H */