Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/include/uapi/drm/pvr_drm.h
26285 views
1
/* SPDX-License-Identifier: (GPL-2.0-only WITH Linux-syscall-note) OR MIT */
2
/* Copyright (c) 2023 Imagination Technologies Ltd. */
3
4
#ifndef PVR_DRM_UAPI_H
5
#define PVR_DRM_UAPI_H
6
7
#include "drm.h"
8
9
#include <linux/const.h>
10
#include <linux/types.h>
11
12
#if defined(__cplusplus)
13
extern "C" {
14
#endif
15
16
/**
17
* DOC: PowerVR UAPI
18
*
19
* The PowerVR IOCTL argument structs have a few limitations in place, in
20
* addition to the standard kernel restrictions:
21
*
22
* - All members must be type-aligned.
23
* - The overall struct must be padded to 64-bit alignment.
24
* - Explicit padding is almost always required. This takes the form of
25
* ``_padding_[x]`` members of sufficient size to pad to the next power-of-two
26
* alignment, where [x] is the offset into the struct in hexadecimal. Arrays
27
* are never used for alignment. Padding fields must be zeroed; this is
28
* always checked.
29
* - Unions may only appear as the last member of a struct.
30
* - Individual union members may grow in the future. The space between the
31
* end of a union member and the end of its containing union is considered
32
* "implicit padding" and must be zeroed. This is always checked.
33
*
34
* In addition to the IOCTL argument structs, the PowerVR UAPI makes use of
35
* DEV_QUERY argument structs. These are used to fetch information about the
36
* device and runtime. These structs are subject to the same rules set out
37
* above.
38
*/
39
40
/**
41
* struct drm_pvr_obj_array - Container used to pass arrays of objects
42
*
43
* It is not unusual to have to extend objects to pass new parameters, and the DRM
44
* ioctl infrastructure is supporting that by padding ioctl arguments with zeros
45
* when the data passed by userspace is smaller than the struct defined in the
46
* drm_ioctl_desc, thus keeping things backward compatible. This type is just
47
* applying the same concepts to indirect objects passed through arrays referenced
48
* from the main ioctl arguments structure: the stride basically defines the size
49
* of the object passed by userspace, which allows the kernel driver to pad with
50
* zeros when it's smaller than the size of the object it expects.
51
*
52
* Use ``DRM_PVR_OBJ_ARRAY()`` to fill object array fields, unless you
53
* have a very good reason not to.
54
*/
55
struct drm_pvr_obj_array {
56
/** @stride: Stride of object struct. Used for versioning. */
57
__u32 stride;
58
59
/** @count: Number of objects in the array. */
60
__u32 count;
61
62
/** @array: User pointer to an array of objects. */
63
__u64 array;
64
};
65
66
/**
67
* DRM_PVR_OBJ_ARRAY() - Helper macro for filling &struct drm_pvr_obj_array.
68
* @cnt: Number of elements pointed to py @ptr.
69
* @ptr: Pointer to start of a C array.
70
*
71
* Return: Literal of type &struct drm_pvr_obj_array.
72
*/
73
#define DRM_PVR_OBJ_ARRAY(cnt, ptr) \
74
{ .stride = sizeof((ptr)[0]), .count = (cnt), .array = (__u64)(uintptr_t)(ptr) }
75
76
/**
77
* DOC: PowerVR IOCTL interface
78
*/
79
80
/**
81
* PVR_IOCTL() - Build a PowerVR IOCTL number
82
* @_ioctl: An incrementing id for this IOCTL. Added to %DRM_COMMAND_BASE.
83
* @_mode: Must be one of %DRM_IOR, %DRM_IOW or %DRM_IOWR.
84
* @_data: The type of the args struct passed by this IOCTL.
85
*
86
* The struct referred to by @_data must have a ``drm_pvr_ioctl_`` prefix and an
87
* ``_args suffix``. They are therefore omitted from @_data.
88
*
89
* This should only be used to build the constants described below; it should
90
* never be used to call an IOCTL directly.
91
*
92
* Return: An IOCTL number to be passed to ioctl() from userspace.
93
*/
94
#define PVR_IOCTL(_ioctl, _mode, _data) \
95
_mode(DRM_COMMAND_BASE + (_ioctl), struct drm_pvr_ioctl_##_data##_args)
96
97
#define DRM_IOCTL_PVR_DEV_QUERY PVR_IOCTL(0x00, DRM_IOWR, dev_query)
98
#define DRM_IOCTL_PVR_CREATE_BO PVR_IOCTL(0x01, DRM_IOWR, create_bo)
99
#define DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET PVR_IOCTL(0x02, DRM_IOWR, get_bo_mmap_offset)
100
#define DRM_IOCTL_PVR_CREATE_VM_CONTEXT PVR_IOCTL(0x03, DRM_IOWR, create_vm_context)
101
#define DRM_IOCTL_PVR_DESTROY_VM_CONTEXT PVR_IOCTL(0x04, DRM_IOW, destroy_vm_context)
102
#define DRM_IOCTL_PVR_VM_MAP PVR_IOCTL(0x05, DRM_IOW, vm_map)
103
#define DRM_IOCTL_PVR_VM_UNMAP PVR_IOCTL(0x06, DRM_IOW, vm_unmap)
104
#define DRM_IOCTL_PVR_CREATE_CONTEXT PVR_IOCTL(0x07, DRM_IOWR, create_context)
105
#define DRM_IOCTL_PVR_DESTROY_CONTEXT PVR_IOCTL(0x08, DRM_IOW, destroy_context)
106
#define DRM_IOCTL_PVR_CREATE_FREE_LIST PVR_IOCTL(0x09, DRM_IOWR, create_free_list)
107
#define DRM_IOCTL_PVR_DESTROY_FREE_LIST PVR_IOCTL(0x0a, DRM_IOW, destroy_free_list)
108
#define DRM_IOCTL_PVR_CREATE_HWRT_DATASET PVR_IOCTL(0x0b, DRM_IOWR, create_hwrt_dataset)
109
#define DRM_IOCTL_PVR_DESTROY_HWRT_DATASET PVR_IOCTL(0x0c, DRM_IOW, destroy_hwrt_dataset)
110
#define DRM_IOCTL_PVR_SUBMIT_JOBS PVR_IOCTL(0x0d, DRM_IOW, submit_jobs)
111
112
/**
113
* DOC: PowerVR IOCTL DEV_QUERY interface
114
*/
115
116
/**
117
* struct drm_pvr_dev_query_gpu_info - Container used to fetch information about
118
* the graphics processor.
119
*
120
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
121
* to %DRM_PVR_DEV_QUERY_GPU_INFO_GET.
122
*/
123
struct drm_pvr_dev_query_gpu_info {
124
/**
125
* @gpu_id: GPU identifier.
126
*
127
* For all currently supported GPUs this is the BVNC encoded as a 64-bit
128
* value as follows:
129
*
130
* +--------+--------+--------+-------+
131
* | 63..48 | 47..32 | 31..16 | 15..0 |
132
* +========+========+========+=======+
133
* | B | V | N | C |
134
* +--------+--------+--------+-------+
135
*/
136
__u64 gpu_id;
137
138
/**
139
* @num_phantoms: Number of Phantoms present.
140
*/
141
__u32 num_phantoms;
142
143
/** @_padding_c: Reserved. This field must be zeroed. */
144
__u32 _padding_c;
145
};
146
147
/**
148
* struct drm_pvr_dev_query_runtime_info - Container used to fetch information
149
* about the graphics runtime.
150
*
151
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
152
* to %DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET.
153
*/
154
struct drm_pvr_dev_query_runtime_info {
155
/**
156
* @free_list_min_pages: Minimum allowed free list size,
157
* in PM physical pages.
158
*/
159
__u64 free_list_min_pages;
160
161
/**
162
* @free_list_max_pages: Maximum allowed free list size,
163
* in PM physical pages.
164
*/
165
__u64 free_list_max_pages;
166
167
/**
168
* @common_store_alloc_region_size: Size of the Allocation
169
* Region within the Common Store used for coefficient and shared
170
* registers, in dwords.
171
*/
172
__u32 common_store_alloc_region_size;
173
174
/**
175
* @common_store_partition_space_size: Size of the
176
* Partition Space within the Common Store for output buffers, in
177
* dwords.
178
*/
179
__u32 common_store_partition_space_size;
180
181
/**
182
* @max_coeffs: Maximum coefficients, in dwords.
183
*/
184
__u32 max_coeffs;
185
186
/**
187
* @cdm_max_local_mem_size_regs: Maximum amount of local
188
* memory available to a compute kernel, in dwords.
189
*/
190
__u32 cdm_max_local_mem_size_regs;
191
};
192
193
/**
194
* struct drm_pvr_dev_query_quirks - Container used to fetch information about
195
* hardware fixes for which the device may require support in the user mode
196
* driver.
197
*
198
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
199
* to %DRM_PVR_DEV_QUERY_QUIRKS_GET.
200
*/
201
struct drm_pvr_dev_query_quirks {
202
/**
203
* @quirks: A userspace address for the hardware quirks __u32 array.
204
*
205
* The first @musthave_count items in the list are quirks that the
206
* client must support for this device. If userspace does not support
207
* all these quirks then functionality is not guaranteed and client
208
* initialisation must fail.
209
* The remaining quirks in the list affect userspace and the kernel or
210
* firmware. They are disabled by default and require userspace to
211
* opt-in. The opt-in mechanism depends on the quirk.
212
*/
213
__u64 quirks;
214
215
/** @count: Length of @quirks (number of __u32). */
216
__u16 count;
217
218
/**
219
* @musthave_count: The number of entries in @quirks that are
220
* mandatory, starting at index 0.
221
*/
222
__u16 musthave_count;
223
224
/** @_padding_c: Reserved. This field must be zeroed. */
225
__u32 _padding_c;
226
};
227
228
/**
229
* struct drm_pvr_dev_query_enhancements - Container used to fetch information
230
* about optional enhancements supported by the device that require support in
231
* the user mode driver.
232
*
233
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
234
* to %DRM_PVR_DEV_ENHANCEMENTS_GET.
235
*/
236
struct drm_pvr_dev_query_enhancements {
237
/**
238
* @enhancements: A userspace address for the hardware enhancements
239
* __u32 array.
240
*
241
* These enhancements affect userspace and the kernel or firmware. They
242
* are disabled by default and require userspace to opt-in. The opt-in
243
* mechanism depends on the enhancement.
244
*/
245
__u64 enhancements;
246
247
/** @count: Length of @enhancements (number of __u32). */
248
__u16 count;
249
250
/** @_padding_a: Reserved. This field must be zeroed. */
251
__u16 _padding_a;
252
253
/** @_padding_c: Reserved. This field must be zeroed. */
254
__u32 _padding_c;
255
};
256
257
/**
258
* enum drm_pvr_heap_id - Array index for heap info data returned by
259
* %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
260
*
261
* For compatibility reasons all indices will be present in the returned array,
262
* however some heaps may not be present. These are indicated where
263
* &struct drm_pvr_heap.size is set to zero.
264
*/
265
enum drm_pvr_heap_id {
266
/** @DRM_PVR_HEAP_GENERAL: General purpose heap. */
267
DRM_PVR_HEAP_GENERAL = 0,
268
/** @DRM_PVR_HEAP_PDS_CODE_DATA: PDS code and data heap. */
269
DRM_PVR_HEAP_PDS_CODE_DATA,
270
/** @DRM_PVR_HEAP_USC_CODE: USC code heap. */
271
DRM_PVR_HEAP_USC_CODE,
272
/** @DRM_PVR_HEAP_RGNHDR: Region header heap. Only used if GPU has BRN63142. */
273
DRM_PVR_HEAP_RGNHDR,
274
/** @DRM_PVR_HEAP_VIS_TEST: Visibility test heap. */
275
DRM_PVR_HEAP_VIS_TEST,
276
/** @DRM_PVR_HEAP_TRANSFER_FRAG: Transfer fragment heap. */
277
DRM_PVR_HEAP_TRANSFER_FRAG,
278
279
/**
280
* @DRM_PVR_HEAP_COUNT: The number of heaps returned by
281
* %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
282
*
283
* More heaps may be added, so this also serves as the copy limit when
284
* sent by the caller.
285
*/
286
DRM_PVR_HEAP_COUNT
287
/* Please only add additional heaps above DRM_PVR_HEAP_COUNT! */
288
};
289
290
/**
291
* struct drm_pvr_heap - Container holding information about a single heap.
292
*
293
* This will always be fetched as an array.
294
*/
295
struct drm_pvr_heap {
296
/** @base: Base address of heap. */
297
__u64 base;
298
299
/** @size: Size of heap, in bytes. Will be 0 if the heap is not present. */
300
__u64 size;
301
302
/** @flags: Flags for this heap. Currently always 0. */
303
__u32 flags;
304
305
/** @page_size_log2: Log2 of page size. */
306
__u32 page_size_log2;
307
};
308
309
/**
310
* struct drm_pvr_dev_query_heap_info - Container used to fetch information
311
* about heaps supported by the device driver.
312
*
313
* Please note all driver-supported heaps will be returned up to &heaps.count.
314
* Some heaps will not be present in all devices, which will be indicated by
315
* &struct drm_pvr_heap.size being set to zero.
316
*
317
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
318
* to %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
319
*/
320
struct drm_pvr_dev_query_heap_info {
321
/**
322
* @heaps: Array of &struct drm_pvr_heap. If pointer is NULL, the count
323
* and stride will be updated with those known to the driver version, to
324
* facilitate allocation by the caller.
325
*/
326
struct drm_pvr_obj_array heaps;
327
};
328
329
/**
330
* enum drm_pvr_static_data_area_usage - Array index for static data area info
331
* returned by %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET.
332
*
333
* For compatibility reasons all indices will be present in the returned array,
334
* however some areas may not be present. These are indicated where
335
* &struct drm_pvr_static_data_area.size is set to zero.
336
*/
337
enum drm_pvr_static_data_area_usage {
338
/**
339
* @DRM_PVR_STATIC_DATA_AREA_EOT: End of Tile PDS program code segment.
340
*
341
* The End of Tile PDS task runs at completion of a tile during a fragment job, and is
342
* responsible for emitting the tile to the Pixel Back End.
343
*/
344
DRM_PVR_STATIC_DATA_AREA_EOT = 0,
345
346
/**
347
* @DRM_PVR_STATIC_DATA_AREA_FENCE: MCU fence area, used during cache flush and
348
* invalidation.
349
*
350
* This must point to valid physical memory but the contents otherwise are not used.
351
*/
352
DRM_PVR_STATIC_DATA_AREA_FENCE,
353
354
/**
355
* @DRM_PVR_STATIC_DATA_AREA_VDM_SYNC: VDM sync program.
356
*
357
* The VDM sync program is used to synchronise multiple areas of the GPU hardware.
358
*/
359
DRM_PVR_STATIC_DATA_AREA_VDM_SYNC,
360
361
/**
362
* @DRM_PVR_STATIC_DATA_AREA_YUV_CSC: YUV coefficients.
363
*
364
* Area contains up to 16 slots with stride of 64 bytes. Each is a 3x4 matrix of u16 fixed
365
* point numbers, with 1 sign bit, 2 integer bits and 13 fractional bits.
366
*
367
* The slots are :
368
* 0 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_RGB_IDENTITY_KHR
369
* 1 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (full range)
370
* 2 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_IDENTITY_KHR (conformant range)
371
* 3 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (full range)
372
* 4 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range)
373
* 5 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (full range)
374
* 6 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range)
375
* 7 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (full range)
376
* 8 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range)
377
* 9 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_601_KHR (conformant range, 10 bit)
378
* 10 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709_KHR (conformant range, 10 bit)
379
* 11 = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_2020_KHR (conformant range, 10 bit)
380
* 14 = Identity (biased)
381
* 15 = Identity
382
*/
383
DRM_PVR_STATIC_DATA_AREA_YUV_CSC,
384
};
385
386
/**
387
* struct drm_pvr_static_data_area - Container holding information about a
388
* single static data area.
389
*
390
* This will always be fetched as an array.
391
*/
392
struct drm_pvr_static_data_area {
393
/**
394
* @area_usage: Usage of static data area.
395
* See &enum drm_pvr_static_data_area_usage.
396
*/
397
__u16 area_usage;
398
399
/**
400
* @location_heap_id: Array index of heap where this of static data
401
* area is located. This array is fetched using
402
* %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
403
*/
404
__u16 location_heap_id;
405
406
/** @size: Size of static data area. Not present if set to zero. */
407
__u32 size;
408
409
/** @offset: Offset of static data area from start of heap. */
410
__u64 offset;
411
};
412
413
/**
414
* struct drm_pvr_dev_query_static_data_areas - Container used to fetch
415
* information about the static data areas in heaps supported by the device
416
* driver.
417
*
418
* Please note all driver-supported static data areas will be returned up to
419
* &static_data_areas.count. Some will not be present for all devices which,
420
* will be indicated by &struct drm_pvr_static_data_area.size being set to zero.
421
*
422
* Further, some heaps will not be present either. See &struct
423
* drm_pvr_dev_query_heap_info.
424
*
425
* When fetching this type &struct drm_pvr_ioctl_dev_query_args.type must be set
426
* to %DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET.
427
*/
428
struct drm_pvr_dev_query_static_data_areas {
429
/**
430
* @static_data_areas: Array of &struct drm_pvr_static_data_area. If
431
* pointer is NULL, the count and stride will be updated with those
432
* known to the driver version, to facilitate allocation by the caller.
433
*/
434
struct drm_pvr_obj_array static_data_areas;
435
};
436
437
/**
438
* enum drm_pvr_dev_query - For use with &drm_pvr_ioctl_dev_query_args.type to
439
* indicate the type of the receiving container.
440
*
441
* Append only. Do not reorder.
442
*/
443
enum drm_pvr_dev_query {
444
/**
445
* @DRM_PVR_DEV_QUERY_GPU_INFO_GET: The dev query args contain a pointer
446
* to &struct drm_pvr_dev_query_gpu_info.
447
*/
448
DRM_PVR_DEV_QUERY_GPU_INFO_GET = 0,
449
450
/**
451
* @DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET: The dev query args contain a
452
* pointer to &struct drm_pvr_dev_query_runtime_info.
453
*/
454
DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET,
455
456
/**
457
* @DRM_PVR_DEV_QUERY_QUIRKS_GET: The dev query args contain a pointer
458
* to &struct drm_pvr_dev_query_quirks.
459
*/
460
DRM_PVR_DEV_QUERY_QUIRKS_GET,
461
462
/**
463
* @DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET: The dev query args contain a
464
* pointer to &struct drm_pvr_dev_query_enhancements.
465
*/
466
DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET,
467
468
/**
469
* @DRM_PVR_DEV_QUERY_HEAP_INFO_GET: The dev query args contain a
470
* pointer to &struct drm_pvr_dev_query_heap_info.
471
*/
472
DRM_PVR_DEV_QUERY_HEAP_INFO_GET,
473
474
/**
475
* @DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET: The dev query args contain
476
* a pointer to &struct drm_pvr_dev_query_static_data_areas.
477
*/
478
DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET,
479
};
480
481
/**
482
* struct drm_pvr_ioctl_dev_query_args - Arguments for %DRM_IOCTL_PVR_DEV_QUERY.
483
*/
484
struct drm_pvr_ioctl_dev_query_args {
485
/**
486
* @type: Type of query and output struct. See &enum drm_pvr_dev_query.
487
*/
488
__u32 type;
489
490
/**
491
* @size: Size of the receiving struct, see @type.
492
*
493
* After a successful call this will be updated to the written byte
494
* length.
495
* Can also be used to get the minimum byte length (see @pointer).
496
* This allows additional fields to be appended to the structs in
497
* future.
498
*/
499
__u32 size;
500
501
/**
502
* @pointer: Pointer to struct @type.
503
*
504
* Must be large enough to contain @size bytes.
505
* If pointer is NULL, the expected size will be returned in the @size
506
* field, but no other data will be written.
507
*/
508
__u64 pointer;
509
};
510
511
/**
512
* DOC: PowerVR IOCTL CREATE_BO interface
513
*/
514
515
/**
516
* DOC: Flags for CREATE_BO
517
*
518
* We use "device" to refer to the GPU here because of the ambiguity between CPU and GPU in some
519
* fonts.
520
*
521
* Device mapping options
522
* :DRM_PVR_BO_BYPASS_DEVICE_CACHE: Specify that device accesses to this memory will bypass the
523
* cache. This is used for buffers that will either be regularly updated by the CPU (eg free
524
* lists) or will be accessed only once and therefore isn't worth caching (eg partial render
525
* buffers).
526
* By default, the device flushes its memory caches after every job, so this is not normally
527
* required for coherency.
528
* :DRM_PVR_BO_PM_FW_PROTECT: Specify that only the Parameter Manager (PM) and/or firmware
529
* processor should be allowed to access this memory when mapped to the device. It is not
530
* valid to specify this flag with DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS.
531
*
532
* CPU mapping options
533
* :DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS: Allow userspace to map and access the contents of this
534
* memory. It is not valid to specify this flag with DRM_PVR_BO_PM_FW_PROTECT.
535
*/
536
#define DRM_PVR_BO_BYPASS_DEVICE_CACHE _BITULL(0)
537
#define DRM_PVR_BO_PM_FW_PROTECT _BITULL(1)
538
#define DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS _BITULL(2)
539
/* Bits 3..63 are reserved. */
540
541
#define DRM_PVR_BO_FLAGS_MASK (DRM_PVR_BO_BYPASS_DEVICE_CACHE | DRM_PVR_BO_PM_FW_PROTECT | \
542
DRM_PVR_BO_ALLOW_CPU_USERSPACE_ACCESS)
543
544
/**
545
* struct drm_pvr_ioctl_create_bo_args - Arguments for %DRM_IOCTL_PVR_CREATE_BO
546
*/
547
struct drm_pvr_ioctl_create_bo_args {
548
/**
549
* @size: [IN] Size of buffer object to create. This must be page size
550
* aligned.
551
*/
552
__u64 size;
553
554
/**
555
* @handle: [OUT] GEM handle of the new buffer object for use in
556
* userspace.
557
*/
558
__u32 handle;
559
560
/** @_padding_c: Reserved. This field must be zeroed. */
561
__u32 _padding_c;
562
563
/**
564
* @flags: [IN] Options which will affect the behaviour of this
565
* creation operation and future mapping operations on the created
566
* object. This field must be a valid combination of ``DRM_PVR_BO_*``
567
* values, with all bits marked as reserved set to zero.
568
*/
569
__u64 flags;
570
};
571
572
/**
573
* DOC: PowerVR IOCTL GET_BO_MMAP_OFFSET interface
574
*/
575
576
/**
577
* struct drm_pvr_ioctl_get_bo_mmap_offset_args - Arguments for
578
* %DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET
579
*
580
* Like other DRM drivers, the "mmap" IOCTL doesn't actually map any memory.
581
* Instead, it allocates a fake offset which refers to the specified buffer
582
* object. This offset can be used with a real mmap call on the DRM device
583
* itself.
584
*/
585
struct drm_pvr_ioctl_get_bo_mmap_offset_args {
586
/** @handle: [IN] GEM handle of the buffer object to be mapped. */
587
__u32 handle;
588
589
/** @_padding_4: Reserved. This field must be zeroed. */
590
__u32 _padding_4;
591
592
/** @offset: [OUT] Fake offset to use in the real mmap call. */
593
__u64 offset;
594
};
595
596
/**
597
* DOC: PowerVR IOCTL CREATE_VM_CONTEXT and DESTROY_VM_CONTEXT interfaces
598
*/
599
600
/**
601
* struct drm_pvr_ioctl_create_vm_context_args - Arguments for
602
* %DRM_IOCTL_PVR_CREATE_VM_CONTEXT
603
*/
604
struct drm_pvr_ioctl_create_vm_context_args {
605
/** @handle: [OUT] Handle for new VM context. */
606
__u32 handle;
607
608
/** @_padding_4: Reserved. This field must be zeroed. */
609
__u32 _padding_4;
610
};
611
612
/**
613
* struct drm_pvr_ioctl_destroy_vm_context_args - Arguments for
614
* %DRM_IOCTL_PVR_DESTROY_VM_CONTEXT
615
*/
616
struct drm_pvr_ioctl_destroy_vm_context_args {
617
/**
618
* @handle: [IN] Handle for VM context to be destroyed.
619
*/
620
__u32 handle;
621
622
/** @_padding_4: Reserved. This field must be zeroed. */
623
__u32 _padding_4;
624
};
625
626
/**
627
* DOC: PowerVR IOCTL VM_MAP and VM_UNMAP interfaces
628
*
629
* The VM UAPI allows userspace to create buffer object mappings in GPU virtual address space.
630
*
631
* The client is responsible for managing GPU address space. It should allocate mappings within
632
* the heaps returned by %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
633
*
634
* %DRM_IOCTL_PVR_VM_MAP creates a new mapping. The client provides the target virtual address for
635
* the mapping. Size and offset within the mapped buffer object can be specified, so the client can
636
* partially map a buffer.
637
*
638
* %DRM_IOCTL_PVR_VM_UNMAP removes a mapping. The entire mapping will be removed from GPU address
639
* space only if the size of the mapping matches that known to the driver.
640
*/
641
642
/**
643
* struct drm_pvr_ioctl_vm_map_args - Arguments for %DRM_IOCTL_PVR_VM_MAP.
644
*/
645
struct drm_pvr_ioctl_vm_map_args {
646
/**
647
* @vm_context_handle: [IN] Handle for VM context for this mapping to
648
* exist in.
649
*/
650
__u32 vm_context_handle;
651
652
/** @flags: [IN] Flags which affect this mapping. Currently always 0. */
653
__u32 flags;
654
655
/**
656
* @device_addr: [IN] Requested device-virtual address for the mapping.
657
* This must be non-zero and aligned to the device page size for the
658
* heap containing the requested address. It is an error to specify an
659
* address which is not contained within one of the heaps returned by
660
* %DRM_PVR_DEV_QUERY_HEAP_INFO_GET.
661
*/
662
__u64 device_addr;
663
664
/**
665
* @handle: [IN] Handle of the target buffer object. This must be a
666
* valid handle returned by %DRM_IOCTL_PVR_CREATE_BO.
667
*/
668
__u32 handle;
669
670
/** @_padding_14: Reserved. This field must be zeroed. */
671
__u32 _padding_14;
672
673
/**
674
* @offset: [IN] Offset into the target bo from which to begin the
675
* mapping.
676
*/
677
__u64 offset;
678
679
/**
680
* @size: [IN] Size of the requested mapping. Must be aligned to
681
* the device page size for the heap containing the requested address,
682
* as well as the host page size. When added to @device_addr, the
683
* result must not overflow the heap which contains @device_addr (i.e.
684
* the range specified by @device_addr and @size must be completely
685
* contained within a single heap specified by
686
* %DRM_PVR_DEV_QUERY_HEAP_INFO_GET).
687
*/
688
__u64 size;
689
};
690
691
/**
692
* struct drm_pvr_ioctl_vm_unmap_args - Arguments for %DRM_IOCTL_PVR_VM_UNMAP.
693
*/
694
struct drm_pvr_ioctl_vm_unmap_args {
695
/**
696
* @vm_context_handle: [IN] Handle for VM context that this mapping
697
* exists in.
698
*/
699
__u32 vm_context_handle;
700
701
/** @_padding_4: Reserved. This field must be zeroed. */
702
__u32 _padding_4;
703
704
/**
705
* @device_addr: [IN] Device-virtual address at the start of the target
706
* mapping. This must be non-zero.
707
*/
708
__u64 device_addr;
709
710
/**
711
* @size: Size in bytes of the target mapping. This must be non-zero.
712
*/
713
__u64 size;
714
};
715
716
/**
717
* DOC: PowerVR IOCTL CREATE_CONTEXT and DESTROY_CONTEXT interfaces
718
*/
719
720
/**
721
* enum drm_pvr_ctx_priority - Arguments for
722
* &drm_pvr_ioctl_create_context_args.priority
723
*/
724
enum drm_pvr_ctx_priority {
725
/** @DRM_PVR_CTX_PRIORITY_LOW: Priority below normal. */
726
DRM_PVR_CTX_PRIORITY_LOW = -512,
727
728
/** @DRM_PVR_CTX_PRIORITY_NORMAL: Normal priority. */
729
DRM_PVR_CTX_PRIORITY_NORMAL = 0,
730
731
/**
732
* @DRM_PVR_CTX_PRIORITY_HIGH: Priority above normal.
733
* Note this requires ``CAP_SYS_NICE`` or ``DRM_MASTER``.
734
*/
735
DRM_PVR_CTX_PRIORITY_HIGH = 512,
736
};
737
738
/**
739
* enum drm_pvr_ctx_type - Arguments for
740
* &struct drm_pvr_ioctl_create_context_args.type
741
*/
742
enum drm_pvr_ctx_type {
743
/**
744
* @DRM_PVR_CTX_TYPE_RENDER: Render context.
745
*/
746
DRM_PVR_CTX_TYPE_RENDER = 0,
747
748
/**
749
* @DRM_PVR_CTX_TYPE_COMPUTE: Compute context.
750
*/
751
DRM_PVR_CTX_TYPE_COMPUTE,
752
753
/**
754
* @DRM_PVR_CTX_TYPE_TRANSFER_FRAG: Transfer context for fragment data
755
* master.
756
*/
757
DRM_PVR_CTX_TYPE_TRANSFER_FRAG,
758
};
759
760
/**
761
* struct drm_pvr_ioctl_create_context_args - Arguments for
762
* %DRM_IOCTL_PVR_CREATE_CONTEXT
763
*/
764
struct drm_pvr_ioctl_create_context_args {
765
/**
766
* @type: [IN] Type of context to create.
767
*
768
* This must be one of the values defined by &enum drm_pvr_ctx_type.
769
*/
770
__u32 type;
771
772
/** @flags: [IN] Flags for context. */
773
__u32 flags;
774
775
/**
776
* @priority: [IN] Priority of new context.
777
*
778
* This must be one of the values defined by &enum drm_pvr_ctx_priority.
779
*/
780
__s32 priority;
781
782
/** @handle: [OUT] Handle for new context. */
783
__u32 handle;
784
785
/**
786
* @static_context_state: [IN] Pointer to static context state stream.
787
*/
788
__u64 static_context_state;
789
790
/**
791
* @static_context_state_len: [IN] Length of static context state, in bytes.
792
*/
793
__u32 static_context_state_len;
794
795
/**
796
* @vm_context_handle: [IN] Handle for VM context that this context is
797
* associated with.
798
*/
799
__u32 vm_context_handle;
800
801
/**
802
* @callstack_addr: [IN] Address for initial call stack pointer. Only valid
803
* if @type is %DRM_PVR_CTX_TYPE_RENDER, otherwise must be 0.
804
*/
805
__u64 callstack_addr;
806
};
807
808
/**
809
* struct drm_pvr_ioctl_destroy_context_args - Arguments for
810
* %DRM_IOCTL_PVR_DESTROY_CONTEXT
811
*/
812
struct drm_pvr_ioctl_destroy_context_args {
813
/**
814
* @handle: [IN] Handle for context to be destroyed.
815
*/
816
__u32 handle;
817
818
/** @_padding_4: Reserved. This field must be zeroed. */
819
__u32 _padding_4;
820
};
821
822
/**
823
* DOC: PowerVR IOCTL CREATE_FREE_LIST and DESTROY_FREE_LIST interfaces
824
*/
825
826
/**
827
* struct drm_pvr_ioctl_create_free_list_args - Arguments for
828
* %DRM_IOCTL_PVR_CREATE_FREE_LIST
829
*
830
* Free list arguments have the following constraints :
831
*
832
* - @max_num_pages must be greater than zero.
833
* - @grow_threshold must be between 0 and 100.
834
* - @grow_num_pages must be less than or equal to &max_num_pages.
835
* - @initial_num_pages, @max_num_pages and @grow_num_pages must be multiples
836
* of 4.
837
* - When &grow_num_pages is 0, @initial_num_pages must be equal to
838
* @max_num_pages.
839
* - When &grow_num_pages is non-zero, @initial_num_pages must be less than
840
* @max_num_pages.
841
*/
842
struct drm_pvr_ioctl_create_free_list_args {
843
/**
844
* @free_list_gpu_addr: [IN] Address of GPU mapping of buffer object
845
* containing memory to be used by free list.
846
*
847
* The mapped region of the buffer object must be at least
848
* @max_num_pages * ``sizeof(__u32)``.
849
*
850
* The buffer object must have been created with
851
* %DRM_PVR_BO_DEVICE_PM_FW_PROTECT set and
852
* %DRM_PVR_BO_CPU_ALLOW_USERSPACE_ACCESS not set.
853
*/
854
__u64 free_list_gpu_addr;
855
856
/** @initial_num_pages: [IN] Pages initially allocated to free list. */
857
__u32 initial_num_pages;
858
859
/** @max_num_pages: [IN] Maximum number of pages in free list. */
860
__u32 max_num_pages;
861
862
/** @grow_num_pages: [IN] Pages to grow free list by per request. */
863
__u32 grow_num_pages;
864
865
/**
866
* @grow_threshold: [IN] Percentage of FL memory used that should
867
* trigger a new grow request.
868
*/
869
__u32 grow_threshold;
870
871
/**
872
* @vm_context_handle: [IN] Handle for VM context that the free list buffer
873
* object is mapped in.
874
*/
875
__u32 vm_context_handle;
876
877
/**
878
* @handle: [OUT] Handle for created free list.
879
*/
880
__u32 handle;
881
};
882
883
/**
884
* struct drm_pvr_ioctl_destroy_free_list_args - Arguments for
885
* %DRM_IOCTL_PVR_DESTROY_FREE_LIST
886
*/
887
struct drm_pvr_ioctl_destroy_free_list_args {
888
/**
889
* @handle: [IN] Handle for free list to be destroyed.
890
*/
891
__u32 handle;
892
893
/** @_padding_4: Reserved. This field must be zeroed. */
894
__u32 _padding_4;
895
};
896
897
/**
898
* DOC: PowerVR IOCTL CREATE_HWRT_DATASET and DESTROY_HWRT_DATASET interfaces
899
*/
900
901
/**
902
* struct drm_pvr_create_hwrt_geom_data_args - Geometry data arguments used for
903
* &struct drm_pvr_ioctl_create_hwrt_dataset_args.geom_data_args.
904
*/
905
struct drm_pvr_create_hwrt_geom_data_args {
906
/** @tpc_dev_addr: [IN] Tail pointer cache GPU virtual address. */
907
__u64 tpc_dev_addr;
908
909
/** @tpc_size: [IN] Size of TPC, in bytes. */
910
__u32 tpc_size;
911
912
/** @tpc_stride: [IN] Stride between layers in TPC, in pages */
913
__u32 tpc_stride;
914
915
/** @vheap_table_dev_addr: [IN] VHEAP table GPU virtual address. */
916
__u64 vheap_table_dev_addr;
917
918
/** @rtc_dev_addr: [IN] Render Target Cache virtual address. */
919
__u64 rtc_dev_addr;
920
};
921
922
/**
923
* struct drm_pvr_create_hwrt_rt_data_args - Render target arguments used for
924
* &struct drm_pvr_ioctl_create_hwrt_dataset_args.rt_data_args.
925
*/
926
struct drm_pvr_create_hwrt_rt_data_args {
927
/** @pm_mlist_dev_addr: [IN] PM MLIST GPU virtual address. */
928
__u64 pm_mlist_dev_addr;
929
930
/** @macrotile_array_dev_addr: [IN] Macrotile array GPU virtual address. */
931
__u64 macrotile_array_dev_addr;
932
933
/** @region_header_dev_addr: [IN] Region header array GPU virtual address. */
934
__u64 region_header_dev_addr;
935
};
936
937
#define PVR_DRM_HWRT_FREE_LIST_LOCAL 0
938
#define PVR_DRM_HWRT_FREE_LIST_GLOBAL 1U
939
940
/**
941
* struct drm_pvr_ioctl_create_hwrt_dataset_args - Arguments for
942
* %DRM_IOCTL_PVR_CREATE_HWRT_DATASET
943
*/
944
struct drm_pvr_ioctl_create_hwrt_dataset_args {
945
/** @geom_data_args: [IN] Geometry data arguments. */
946
struct drm_pvr_create_hwrt_geom_data_args geom_data_args;
947
948
/**
949
* @rt_data_args: [IN] Array of render target arguments.
950
*
951
* Each entry in this array represents a render target in a double buffered
952
* setup.
953
*/
954
struct drm_pvr_create_hwrt_rt_data_args rt_data_args[2];
955
956
/**
957
* @free_list_handles: [IN] Array of free list handles.
958
*
959
* free_list_handles[PVR_DRM_HWRT_FREE_LIST_LOCAL] must have initial
960
* size of at least that reported by
961
* &drm_pvr_dev_query_runtime_info.free_list_min_pages.
962
*/
963
__u32 free_list_handles[2];
964
965
/** @width: [IN] Width in pixels. */
966
__u32 width;
967
968
/** @height: [IN] Height in pixels. */
969
__u32 height;
970
971
/** @samples: [IN] Number of samples. */
972
__u32 samples;
973
974
/** @layers: [IN] Number of layers. */
975
__u32 layers;
976
977
/** @isp_merge_lower_x: [IN] Lower X coefficient for triangle merging. */
978
__u32 isp_merge_lower_x;
979
980
/** @isp_merge_lower_y: [IN] Lower Y coefficient for triangle merging. */
981
__u32 isp_merge_lower_y;
982
983
/** @isp_merge_scale_x: [IN] Scale X coefficient for triangle merging. */
984
__u32 isp_merge_scale_x;
985
986
/** @isp_merge_scale_y: [IN] Scale Y coefficient for triangle merging. */
987
__u32 isp_merge_scale_y;
988
989
/** @isp_merge_upper_x: [IN] Upper X coefficient for triangle merging. */
990
__u32 isp_merge_upper_x;
991
992
/** @isp_merge_upper_y: [IN] Upper Y coefficient for triangle merging. */
993
__u32 isp_merge_upper_y;
994
995
/**
996
* @region_header_size: [IN] Size of region header array. This common field is used by
997
* both render targets in this data set.
998
*
999
* The units for this field differ depending on what version of the simple internal
1000
* parameter format the device uses. If format 2 is in use then this is interpreted as the
1001
* number of region headers. For other formats it is interpreted as the size in dwords.
1002
*/
1003
__u32 region_header_size;
1004
1005
/**
1006
* @handle: [OUT] Handle for created HWRT dataset.
1007
*/
1008
__u32 handle;
1009
};
1010
1011
/**
1012
* struct drm_pvr_ioctl_destroy_hwrt_dataset_args - Arguments for
1013
* %DRM_IOCTL_PVR_DESTROY_HWRT_DATASET
1014
*/
1015
struct drm_pvr_ioctl_destroy_hwrt_dataset_args {
1016
/**
1017
* @handle: [IN] Handle for HWRT dataset to be destroyed.
1018
*/
1019
__u32 handle;
1020
1021
/** @_padding_4: Reserved. This field must be zeroed. */
1022
__u32 _padding_4;
1023
};
1024
1025
/**
1026
* DOC: PowerVR IOCTL SUBMIT_JOBS interface
1027
*/
1028
1029
/**
1030
* DOC: Flags for the drm_pvr_sync_op object.
1031
*
1032
* .. c:macro:: DRM_PVR_SYNC_OP_HANDLE_TYPE_MASK
1033
*
1034
* Handle type mask for the drm_pvr_sync_op::flags field.
1035
*
1036
* .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ
1037
*
1038
* Indicates the handle passed in drm_pvr_sync_op::handle is a syncobj handle.
1039
* This is the default type.
1040
*
1041
* .. c:macro:: DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ
1042
*
1043
* Indicates the handle passed in drm_pvr_sync_op::handle is a timeline syncobj handle.
1044
*
1045
* .. c:macro:: DRM_PVR_SYNC_OP_FLAG_SIGNAL
1046
*
1047
* Signal operation requested. The out-fence bound to the job will be attached to
1048
* the syncobj whose handle is passed in drm_pvr_sync_op::handle.
1049
*
1050
* .. c:macro:: DRM_PVR_SYNC_OP_FLAG_WAIT
1051
*
1052
* Wait operation requested. The job will wait for this particular syncobj or syncobj
1053
* point to be signaled before being started.
1054
* This is the default operation.
1055
*/
1056
#define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK 0xf
1057
#define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ 0
1058
#define DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ 1
1059
#define DRM_PVR_SYNC_OP_FLAG_SIGNAL _BITULL(31)
1060
#define DRM_PVR_SYNC_OP_FLAG_WAIT 0
1061
1062
#define DRM_PVR_SYNC_OP_FLAGS_MASK (DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK | \
1063
DRM_PVR_SYNC_OP_FLAG_SIGNAL)
1064
1065
/**
1066
* struct drm_pvr_sync_op - Object describing a sync operation
1067
*/
1068
struct drm_pvr_sync_op {
1069
/** @handle: Handle of sync object. */
1070
__u32 handle;
1071
1072
/** @flags: Combination of ``DRM_PVR_SYNC_OP_FLAG_`` flags. */
1073
__u32 flags;
1074
1075
/** @value: Timeline value for this drm_syncobj. MBZ for a binary syncobj. */
1076
__u64 value;
1077
};
1078
1079
/**
1080
* DOC: Flags for SUBMIT_JOB ioctl geometry command.
1081
*
1082
* .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST
1083
*
1084
* Indicates if this the first command to be issued for a render.
1085
*
1086
* .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST
1087
*
1088
* Indicates if this the last command to be issued for a render.
1089
*
1090
* .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE
1091
*
1092
* Forces to use single core in a multi core device.
1093
*
1094
* .. c:macro:: DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK
1095
*
1096
* Logical OR of all the geometry cmd flags.
1097
*/
1098
#define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST _BITULL(0)
1099
#define DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST _BITULL(1)
1100
#define DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE _BITULL(2)
1101
#define DRM_PVR_SUBMIT_JOB_GEOM_CMD_FLAGS_MASK \
1102
(DRM_PVR_SUBMIT_JOB_GEOM_CMD_FIRST | \
1103
DRM_PVR_SUBMIT_JOB_GEOM_CMD_LAST | \
1104
DRM_PVR_SUBMIT_JOB_GEOM_CMD_SINGLE_CORE)
1105
1106
/**
1107
* DOC: Flags for SUBMIT_JOB ioctl fragment command.
1108
*
1109
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE
1110
*
1111
* Use single core in a multi core setup.
1112
*
1113
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER
1114
*
1115
* Indicates whether a depth buffer is present.
1116
*
1117
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER
1118
*
1119
* Indicates whether a stencil buffer is present.
1120
*
1121
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP
1122
*
1123
* Disallow compute overlapped with this render.
1124
*
1125
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS
1126
*
1127
* Indicates whether this render produces visibility results.
1128
*
1129
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER
1130
*
1131
* Indicates whether partial renders write to a scratch buffer instead of
1132
* the final surface. It also forces the full screen copy expected to be
1133
* present on the last render after all partial renders have completed.
1134
*
1135
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE
1136
*
1137
* Disable pixel merging for this render.
1138
*
1139
* .. c:macro:: DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK
1140
*
1141
* Logical OR of all the fragment cmd flags.
1142
*/
1143
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE _BITULL(0)
1144
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER _BITULL(1)
1145
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER _BITULL(2)
1146
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP _BITULL(3)
1147
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER _BITULL(4)
1148
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS _BITULL(5)
1149
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER _BITULL(6)
1150
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE _BITULL(7)
1151
#define DRM_PVR_SUBMIT_JOB_FRAG_CMD_FLAGS_MASK \
1152
(DRM_PVR_SUBMIT_JOB_FRAG_CMD_SINGLE_CORE | \
1153
DRM_PVR_SUBMIT_JOB_FRAG_CMD_DEPTHBUFFER | \
1154
DRM_PVR_SUBMIT_JOB_FRAG_CMD_STENCILBUFFER | \
1155
DRM_PVR_SUBMIT_JOB_FRAG_CMD_PREVENT_CDM_OVERLAP | \
1156
DRM_PVR_SUBMIT_JOB_FRAG_CMD_SCRATCHBUFFER | \
1157
DRM_PVR_SUBMIT_JOB_FRAG_CMD_GET_VIS_RESULTS | \
1158
DRM_PVR_SUBMIT_JOB_FRAG_CMD_PARTIAL_RENDER | \
1159
DRM_PVR_SUBMIT_JOB_FRAG_CMD_DISABLE_PIXELMERGE)
1160
1161
/**
1162
* DOC: Flags for SUBMIT_JOB ioctl compute command.
1163
*
1164
* .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP
1165
*
1166
* Disallow other jobs overlapped with this compute.
1167
*
1168
* .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE
1169
*
1170
* Forces to use single core in a multi core device.
1171
*
1172
* .. c:macro:: DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK
1173
*
1174
* Logical OR of all the compute cmd flags.
1175
*/
1176
#define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP _BITULL(0)
1177
#define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE _BITULL(1)
1178
#define DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_FLAGS_MASK \
1179
(DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_PREVENT_ALL_OVERLAP | \
1180
DRM_PVR_SUBMIT_JOB_COMPUTE_CMD_SINGLE_CORE)
1181
1182
/**
1183
* DOC: Flags for SUBMIT_JOB ioctl transfer command.
1184
*
1185
* .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE
1186
*
1187
* Forces job to use a single core in a multi core device.
1188
*
1189
* .. c:macro:: DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK
1190
*
1191
* Logical OR of all the transfer cmd flags.
1192
*/
1193
#define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE _BITULL(0)
1194
1195
#define DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_FLAGS_MASK \
1196
DRM_PVR_SUBMIT_JOB_TRANSFER_CMD_SINGLE_CORE
1197
1198
/**
1199
* enum drm_pvr_job_type - Arguments for &struct drm_pvr_job.job_type
1200
*/
1201
enum drm_pvr_job_type {
1202
/** @DRM_PVR_JOB_TYPE_GEOMETRY: Job type is geometry. */
1203
DRM_PVR_JOB_TYPE_GEOMETRY = 0,
1204
1205
/** @DRM_PVR_JOB_TYPE_FRAGMENT: Job type is fragment. */
1206
DRM_PVR_JOB_TYPE_FRAGMENT,
1207
1208
/** @DRM_PVR_JOB_TYPE_COMPUTE: Job type is compute. */
1209
DRM_PVR_JOB_TYPE_COMPUTE,
1210
1211
/** @DRM_PVR_JOB_TYPE_TRANSFER_FRAG: Job type is a fragment transfer. */
1212
DRM_PVR_JOB_TYPE_TRANSFER_FRAG,
1213
};
1214
1215
/**
1216
* struct drm_pvr_hwrt_data_ref - Reference HWRT data
1217
*/
1218
struct drm_pvr_hwrt_data_ref {
1219
/** @set_handle: HWRT data set handle. */
1220
__u32 set_handle;
1221
1222
/** @data_index: Index of the HWRT data inside the data set. */
1223
__u32 data_index;
1224
};
1225
1226
/**
1227
* struct drm_pvr_job - Job arguments passed to the %DRM_IOCTL_PVR_SUBMIT_JOBS ioctl
1228
*/
1229
struct drm_pvr_job {
1230
/**
1231
* @type: [IN] Type of job being submitted
1232
*
1233
* This must be one of the values defined by &enum drm_pvr_job_type.
1234
*/
1235
__u32 type;
1236
1237
/**
1238
* @context_handle: [IN] Context handle.
1239
*
1240
* When @job_type is %DRM_PVR_JOB_TYPE_RENDER, %DRM_PVR_JOB_TYPE_COMPUTE or
1241
* %DRM_PVR_JOB_TYPE_TRANSFER_FRAG, this must be a valid handle returned by
1242
* %DRM_IOCTL_PVR_CREATE_CONTEXT. The type of context must be compatible
1243
* with the type of job being submitted.
1244
*
1245
* When @job_type is %DRM_PVR_JOB_TYPE_NULL, this must be zero.
1246
*/
1247
__u32 context_handle;
1248
1249
/**
1250
* @flags: [IN] Flags for command.
1251
*
1252
* Those are job-dependent. See all ``DRM_PVR_SUBMIT_JOB_*``.
1253
*/
1254
__u32 flags;
1255
1256
/**
1257
* @cmd_stream_len: [IN] Length of command stream, in bytes.
1258
*/
1259
__u32 cmd_stream_len;
1260
1261
/**
1262
* @cmd_stream: [IN] Pointer to command stream for command.
1263
*
1264
* The command stream must be u64-aligned.
1265
*/
1266
__u64 cmd_stream;
1267
1268
/** @sync_ops: [IN] Fragment sync operations. */
1269
struct drm_pvr_obj_array sync_ops;
1270
1271
/**
1272
* @hwrt: [IN] HWRT data used by render jobs (geometry or fragment).
1273
*
1274
* Must be zero for non-render jobs.
1275
*/
1276
struct drm_pvr_hwrt_data_ref hwrt;
1277
};
1278
1279
/**
1280
* struct drm_pvr_ioctl_submit_jobs_args - Arguments for %DRM_IOCTL_PVR_SUBMIT_JOB
1281
*
1282
* If the syscall returns an error it is important to check the value of
1283
* @jobs.count. This indicates the index into @jobs.array where the
1284
* error occurred.
1285
*/
1286
struct drm_pvr_ioctl_submit_jobs_args {
1287
/** @jobs: [IN] Array of jobs to submit. */
1288
struct drm_pvr_obj_array jobs;
1289
};
1290
1291
#if defined(__cplusplus)
1292
}
1293
#endif
1294
1295
#endif /* PVR_DRM_UAPI_H */
1296
1297