Loading...
1/* SPDX-License-Identifier: MIT */
2/*
3 * Copyright © 2022 Intel Corporation
4 */
5
6/**
7 * DOC: I915_PARAM_VM_BIND_VERSION
8 *
9 * VM_BIND feature version supported.
10 * See typedef drm_i915_getparam_t param.
11 *
12 * Specifies the VM_BIND feature version supported.
13 * The following versions of VM_BIND have been defined:
14 *
15 * 0: No VM_BIND support.
16 *
17 * 1: In VM_UNBIND calls, the UMD must specify the exact mappings created
18 * previously with VM_BIND, the ioctl will not support unbinding multiple
19 * mappings or splitting them. Similarly, VM_BIND calls will not replace
20 * any existing mappings.
21 *
22 * 2: The restrictions on unbinding partial or multiple mappings is
23 * lifted, Similarly, binding will replace any mappings in the given range.
24 *
25 * See struct drm_i915_gem_vm_bind and struct drm_i915_gem_vm_unbind.
26 */
27#define I915_PARAM_VM_BIND_VERSION 57
28
29/**
30 * DOC: I915_VM_CREATE_FLAGS_USE_VM_BIND
31 *
32 * Flag to opt-in for VM_BIND mode of binding during VM creation.
33 * See struct drm_i915_gem_vm_control flags.
34 *
35 * The older execbuf2 ioctl will not support VM_BIND mode of operation.
36 * For VM_BIND mode, we have new execbuf3 ioctl which will not accept any
37 * execlist (See struct drm_i915_gem_execbuffer3 for more details).
38 */
39#define I915_VM_CREATE_FLAGS_USE_VM_BIND (1 << 0)
40
41/* VM_BIND related ioctls */
42#define DRM_I915_GEM_VM_BIND 0x3d
43#define DRM_I915_GEM_VM_UNBIND 0x3e
44#define DRM_I915_GEM_EXECBUFFER3 0x3f
45
46#define DRM_IOCTL_I915_GEM_VM_BIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_BIND, struct drm_i915_gem_vm_bind)
47#define DRM_IOCTL_I915_GEM_VM_UNBIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_UNBIND, struct drm_i915_gem_vm_bind)
48#define DRM_IOCTL_I915_GEM_EXECBUFFER3 DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_EXECBUFFER3, struct drm_i915_gem_execbuffer3)
49
50/**
51 * struct drm_i915_gem_timeline_fence - An input or output timeline fence.
52 *
53 * The operation will wait for input fence to signal.
54 *
55 * The returned output fence will be signaled after the completion of the
56 * operation.
57 */
58struct drm_i915_gem_timeline_fence {
59 /** @handle: User's handle for a drm_syncobj to wait on or signal. */
60 __u32 handle;
61
62 /**
63 * @flags: Supported flags are:
64 *
65 * I915_TIMELINE_FENCE_WAIT:
66 * Wait for the input fence before the operation.
67 *
68 * I915_TIMELINE_FENCE_SIGNAL:
69 * Return operation completion fence as output.
70 */
71 __u32 flags;
72#define I915_TIMELINE_FENCE_WAIT (1 << 0)
73#define I915_TIMELINE_FENCE_SIGNAL (1 << 1)
74#define __I915_TIMELINE_FENCE_UNKNOWN_FLAGS (-(I915_TIMELINE_FENCE_SIGNAL << 1))
75
76 /**
77 * @value: A point in the timeline.
78 * Value must be 0 for a binary drm_syncobj. A Value of 0 for a
79 * timeline drm_syncobj is invalid as it turns a drm_syncobj into a
80 * binary one.
81 */
82 __u64 value;
83};
84
85/**
86 * struct drm_i915_gem_vm_bind - VA to object mapping to bind.
87 *
88 * This structure is passed to VM_BIND ioctl and specifies the mapping of GPU
89 * virtual address (VA) range to the section of an object that should be bound
90 * in the device page table of the specified address space (VM).
91 * The VA range specified must be unique (ie., not currently bound) and can
92 * be mapped to whole object or a section of the object (partial binding).
93 * Multiple VA mappings can be created to the same section of the object
94 * (aliasing).
95 *
96 * The @start, @offset and @length must be 4K page aligned. However the DG2
97 * and XEHPSDV has 64K page size for device local memory and has compact page
98 * table. On those platforms, for binding device local-memory objects, the
99 * @start, @offset and @length must be 64K aligned. Also, UMDs should not mix
100 * the local memory 64K page and the system memory 4K page bindings in the same
101 * 2M range.
102 *
103 * Error code -EINVAL will be returned if @start, @offset and @length are not
104 * properly aligned. In version 1 (See I915_PARAM_VM_BIND_VERSION), error code
105 * -ENOSPC will be returned if the VA range specified can't be reserved.
106 *
107 * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
108 * are not ordered. Furthermore, parts of the VM_BIND operation can be done
109 * asynchronously, if valid @fence is specified.
110 */
111struct drm_i915_gem_vm_bind {
112 /** @vm_id: VM (address space) id to bind */
113 __u32 vm_id;
114
115 /** @handle: Object handle */
116 __u32 handle;
117
118 /** @start: Virtual Address start to bind */
119 __u64 start;
120
121 /** @offset: Offset in object to bind */
122 __u64 offset;
123
124 /** @length: Length of mapping to bind */
125 __u64 length;
126
127 /**
128 * @flags: Supported flags are:
129 *
130 * I915_GEM_VM_BIND_CAPTURE:
131 * Capture this mapping in the dump upon GPU error.
132 *
133 * Note that @fence carries its own flags.
134 */
135 __u64 flags;
136#define I915_GEM_VM_BIND_CAPTURE (1 << 0)
137
138 /**
139 * @fence: Timeline fence for bind completion signaling.
140 *
141 * Timeline fence is of format struct drm_i915_gem_timeline_fence.
142 *
143 * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
144 * is invalid, and an error will be returned.
145 *
146 * If I915_TIMELINE_FENCE_SIGNAL flag is not set, then out fence
147 * is not requested and binding is completed synchronously.
148 */
149 struct drm_i915_gem_timeline_fence fence;
150
151 /**
152 * @extensions: Zero-terminated chain of extensions.
153 *
154 * For future extensions. See struct i915_user_extension.
155 */
156 __u64 extensions;
157};
158
159/**
160 * struct drm_i915_gem_vm_unbind - VA to object mapping to unbind.
161 *
162 * This structure is passed to VM_UNBIND ioctl and specifies the GPU virtual
163 * address (VA) range that should be unbound from the device page table of the
164 * specified address space (VM). VM_UNBIND will force unbind the specified
165 * range from device page table without waiting for any GPU job to complete.
166 * It is UMDs responsibility to ensure the mapping is no longer in use before
167 * calling VM_UNBIND.
168 *
169 * If the specified mapping is not found, the ioctl will simply return without
170 * any error.
171 *
172 * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
173 * are not ordered. Furthermore, parts of the VM_UNBIND operation can be done
174 * asynchronously, if valid @fence is specified.
175 */
176struct drm_i915_gem_vm_unbind {
177 /** @vm_id: VM (address space) id to bind */
178 __u32 vm_id;
179
180 /** @rsvd: Reserved, MBZ */
181 __u32 rsvd;
182
183 /** @start: Virtual Address start to unbind */
184 __u64 start;
185
186 /** @length: Length of mapping to unbind */
187 __u64 length;
188
189 /**
190 * @flags: Currently reserved, MBZ.
191 *
192 * Note that @fence carries its own flags.
193 */
194 __u64 flags;
195
196 /**
197 * @fence: Timeline fence for unbind completion signaling.
198 *
199 * Timeline fence is of format struct drm_i915_gem_timeline_fence.
200 *
201 * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
202 * is invalid, and an error will be returned.
203 *
204 * If I915_TIMELINE_FENCE_SIGNAL flag is not set, then out fence
205 * is not requested and unbinding is completed synchronously.
206 */
207 struct drm_i915_gem_timeline_fence fence;
208
209 /**
210 * @extensions: Zero-terminated chain of extensions.
211 *
212 * For future extensions. See struct i915_user_extension.
213 */
214 __u64 extensions;
215};
216
217/**
218 * struct drm_i915_gem_execbuffer3 - Structure for DRM_I915_GEM_EXECBUFFER3
219 * ioctl.
220 *
221 * DRM_I915_GEM_EXECBUFFER3 ioctl only works in VM_BIND mode and VM_BIND mode
222 * only works with this ioctl for submission.
223 * See I915_VM_CREATE_FLAGS_USE_VM_BIND.
224 */
225struct drm_i915_gem_execbuffer3 {
226 /**
227 * @ctx_id: Context id
228 *
229 * Only contexts with user engine map are allowed.
230 */
231 __u32 ctx_id;
232
233 /**
234 * @engine_idx: Engine index
235 *
236 * An index in the user engine map of the context specified by @ctx_id.
237 */
238 __u32 engine_idx;
239
240 /**
241 * @batch_address: Batch gpu virtual address/es.
242 *
243 * For normal submission, it is the gpu virtual address of the batch
244 * buffer. For parallel submission, it is a pointer to an array of
245 * batch buffer gpu virtual addresses with array size equal to the
246 * number of (parallel) engines involved in that submission (See
247 * struct i915_context_engines_parallel_submit).
248 */
249 __u64 batch_address;
250
251 /** @flags: Currently reserved, MBZ */
252 __u64 flags;
253
254 /** @rsvd1: Reserved, MBZ */
255 __u32 rsvd1;
256
257 /** @fence_count: Number of fences in @timeline_fences array. */
258 __u32 fence_count;
259
260 /**
261 * @timeline_fences: Pointer to an array of timeline fences.
262 *
263 * Timeline fences are of format struct drm_i915_gem_timeline_fence.
264 */
265 __u64 timeline_fences;
266
267 /** @rsvd2: Reserved, MBZ */
268 __u64 rsvd2;
269
270 /**
271 * @extensions: Zero-terminated chain of extensions.
272 *
273 * For future extensions. See struct i915_user_extension.
274 */
275 __u64 extensions;
276};
277
278/**
279 * struct drm_i915_gem_create_ext_vm_private - Extension to make the object
280 * private to the specified VM.
281 *
282 * See struct drm_i915_gem_create_ext.
283 */
284struct drm_i915_gem_create_ext_vm_private {
285#define I915_GEM_CREATE_EXT_VM_PRIVATE 2
286 /** @base: Extension link. See struct i915_user_extension. */
287 struct i915_user_extension base;
288
289 /** @vm_id: Id of the VM to which the object is private */
290 __u32 vm_id;
291};
1/* SPDX-License-Identifier: MIT */
2/*
3 * Copyright © 2022 Intel Corporation
4 */
5
6/**
7 * DOC: I915_PARAM_VM_BIND_VERSION
8 *
9 * VM_BIND feature version supported.
10 * See typedef drm_i915_getparam_t param.
11 *
12 * Specifies the VM_BIND feature version supported.
13 * The following versions of VM_BIND have been defined:
14 *
15 * 0: No VM_BIND support.
16 *
17 * 1: In VM_UNBIND calls, the UMD must specify the exact mappings created
18 * previously with VM_BIND, the ioctl will not support unbinding multiple
19 * mappings or splitting them. Similarly, VM_BIND calls will not replace
20 * any existing mappings.
21 *
22 * 2: The restrictions on unbinding partial or multiple mappings is
23 * lifted, Similarly, binding will replace any mappings in the given range.
24 *
25 * See struct drm_i915_gem_vm_bind and struct drm_i915_gem_vm_unbind.
26 */
27#define I915_PARAM_VM_BIND_VERSION 57
28
29/**
30 * DOC: I915_VM_CREATE_FLAGS_USE_VM_BIND
31 *
32 * Flag to opt-in for VM_BIND mode of binding during VM creation.
33 * See struct drm_i915_gem_vm_control flags.
34 *
35 * The older execbuf2 ioctl will not support VM_BIND mode of operation.
36 * For VM_BIND mode, we have new execbuf3 ioctl which will not accept any
37 * execlist (See struct drm_i915_gem_execbuffer3 for more details).
38 */
39#define I915_VM_CREATE_FLAGS_USE_VM_BIND (1 << 0)
40
41/* VM_BIND related ioctls */
42#define DRM_I915_GEM_VM_BIND 0x3d
43#define DRM_I915_GEM_VM_UNBIND 0x3e
44#define DRM_I915_GEM_EXECBUFFER3 0x3f
45
46#define DRM_IOCTL_I915_GEM_VM_BIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_BIND, struct drm_i915_gem_vm_bind)
47#define DRM_IOCTL_I915_GEM_VM_UNBIND DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_UNBIND, struct drm_i915_gem_vm_bind)
48#define DRM_IOCTL_I915_GEM_EXECBUFFER3 DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_EXECBUFFER3, struct drm_i915_gem_execbuffer3)
49
50/**
51 * struct drm_i915_gem_timeline_fence - An input or output timeline fence.
52 *
53 * The operation will wait for input fence to signal.
54 *
55 * The returned output fence will be signaled after the completion of the
56 * operation.
57 */
58struct drm_i915_gem_timeline_fence {
59 /** @handle: User's handle for a drm_syncobj to wait on or signal. */
60 __u32 handle;
61
62 /**
63 * @flags: Supported flags are:
64 *
65 * I915_TIMELINE_FENCE_WAIT:
66 * Wait for the input fence before the operation.
67 *
68 * I915_TIMELINE_FENCE_SIGNAL:
69 * Return operation completion fence as output.
70 */
71 __u32 flags;
72#define I915_TIMELINE_FENCE_WAIT (1 << 0)
73#define I915_TIMELINE_FENCE_SIGNAL (1 << 1)
74#define __I915_TIMELINE_FENCE_UNKNOWN_FLAGS (-(I915_TIMELINE_FENCE_SIGNAL << 1))
75
76 /**
77 * @value: A point in the timeline.
78 * Value must be 0 for a binary drm_syncobj. A Value of 0 for a
79 * timeline drm_syncobj is invalid as it turns a drm_syncobj into a
80 * binary one.
81 */
82 __u64 value;
83};
84
85/**
86 * struct drm_i915_gem_vm_bind - VA to object mapping to bind.
87 *
88 * This structure is passed to VM_BIND ioctl and specifies the mapping of GPU
89 * virtual address (VA) range to the section of an object that should be bound
90 * in the device page table of the specified address space (VM).
91 * The VA range specified must be unique (ie., not currently bound) and can
92 * be mapped to whole object or a section of the object (partial binding).
93 * Multiple VA mappings can be created to the same section of the object
94 * (aliasing).
95 *
96 * The @start, @offset and @length must be 4K page aligned. However the DG2 has
97 * 64K page size for device local memory and has compact page table. On that
98 * platform, for binding device local-memory objects, the @start, @offset and
99 * @length must be 64K aligned. Also, UMDs should not mix the local memory 64K
100 * page and the system memory 4K page bindings in the same 2M range.
101 *
102 * Error code -EINVAL will be returned if @start, @offset and @length are not
103 * properly aligned. In version 1 (See I915_PARAM_VM_BIND_VERSION), error code
104 * -ENOSPC will be returned if the VA range specified can't be reserved.
105 *
106 * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
107 * are not ordered. Furthermore, parts of the VM_BIND operation can be done
108 * asynchronously, if valid @fence is specified.
109 */
110struct drm_i915_gem_vm_bind {
111 /** @vm_id: VM (address space) id to bind */
112 __u32 vm_id;
113
114 /** @handle: Object handle */
115 __u32 handle;
116
117 /** @start: Virtual Address start to bind */
118 __u64 start;
119
120 /** @offset: Offset in object to bind */
121 __u64 offset;
122
123 /** @length: Length of mapping to bind */
124 __u64 length;
125
126 /**
127 * @flags: Supported flags are:
128 *
129 * I915_GEM_VM_BIND_CAPTURE:
130 * Capture this mapping in the dump upon GPU error.
131 *
132 * Note that @fence carries its own flags.
133 */
134 __u64 flags;
135#define I915_GEM_VM_BIND_CAPTURE (1 << 0)
136
137 /**
138 * @fence: Timeline fence for bind completion signaling.
139 *
140 * Timeline fence is of format struct drm_i915_gem_timeline_fence.
141 *
142 * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
143 * is invalid, and an error will be returned.
144 *
145 * If I915_TIMELINE_FENCE_SIGNAL flag is not set, then out fence
146 * is not requested and binding is completed synchronously.
147 */
148 struct drm_i915_gem_timeline_fence fence;
149
150 /**
151 * @extensions: Zero-terminated chain of extensions.
152 *
153 * For future extensions. See struct i915_user_extension.
154 */
155 __u64 extensions;
156};
157
158/**
159 * struct drm_i915_gem_vm_unbind - VA to object mapping to unbind.
160 *
161 * This structure is passed to VM_UNBIND ioctl and specifies the GPU virtual
162 * address (VA) range that should be unbound from the device page table of the
163 * specified address space (VM). VM_UNBIND will force unbind the specified
164 * range from device page table without waiting for any GPU job to complete.
165 * It is UMDs responsibility to ensure the mapping is no longer in use before
166 * calling VM_UNBIND.
167 *
168 * If the specified mapping is not found, the ioctl will simply return without
169 * any error.
170 *
171 * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently
172 * are not ordered. Furthermore, parts of the VM_UNBIND operation can be done
173 * asynchronously, if valid @fence is specified.
174 */
175struct drm_i915_gem_vm_unbind {
176 /** @vm_id: VM (address space) id to bind */
177 __u32 vm_id;
178
179 /** @rsvd: Reserved, MBZ */
180 __u32 rsvd;
181
182 /** @start: Virtual Address start to unbind */
183 __u64 start;
184
185 /** @length: Length of mapping to unbind */
186 __u64 length;
187
188 /**
189 * @flags: Currently reserved, MBZ.
190 *
191 * Note that @fence carries its own flags.
192 */
193 __u64 flags;
194
195 /**
196 * @fence: Timeline fence for unbind completion signaling.
197 *
198 * Timeline fence is of format struct drm_i915_gem_timeline_fence.
199 *
200 * It is an out fence, hence using I915_TIMELINE_FENCE_WAIT flag
201 * is invalid, and an error will be returned.
202 *
203 * If I915_TIMELINE_FENCE_SIGNAL flag is not set, then out fence
204 * is not requested and unbinding is completed synchronously.
205 */
206 struct drm_i915_gem_timeline_fence fence;
207
208 /**
209 * @extensions: Zero-terminated chain of extensions.
210 *
211 * For future extensions. See struct i915_user_extension.
212 */
213 __u64 extensions;
214};
215
216/**
217 * struct drm_i915_gem_execbuffer3 - Structure for DRM_I915_GEM_EXECBUFFER3
218 * ioctl.
219 *
220 * DRM_I915_GEM_EXECBUFFER3 ioctl only works in VM_BIND mode and VM_BIND mode
221 * only works with this ioctl for submission.
222 * See I915_VM_CREATE_FLAGS_USE_VM_BIND.
223 */
224struct drm_i915_gem_execbuffer3 {
225 /**
226 * @ctx_id: Context id
227 *
228 * Only contexts with user engine map are allowed.
229 */
230 __u32 ctx_id;
231
232 /**
233 * @engine_idx: Engine index
234 *
235 * An index in the user engine map of the context specified by @ctx_id.
236 */
237 __u32 engine_idx;
238
239 /**
240 * @batch_address: Batch gpu virtual address/es.
241 *
242 * For normal submission, it is the gpu virtual address of the batch
243 * buffer. For parallel submission, it is a pointer to an array of
244 * batch buffer gpu virtual addresses with array size equal to the
245 * number of (parallel) engines involved in that submission (See
246 * struct i915_context_engines_parallel_submit).
247 */
248 __u64 batch_address;
249
250 /** @flags: Currently reserved, MBZ */
251 __u64 flags;
252
253 /** @rsvd1: Reserved, MBZ */
254 __u32 rsvd1;
255
256 /** @fence_count: Number of fences in @timeline_fences array. */
257 __u32 fence_count;
258
259 /**
260 * @timeline_fences: Pointer to an array of timeline fences.
261 *
262 * Timeline fences are of format struct drm_i915_gem_timeline_fence.
263 */
264 __u64 timeline_fences;
265
266 /** @rsvd2: Reserved, MBZ */
267 __u64 rsvd2;
268
269 /**
270 * @extensions: Zero-terminated chain of extensions.
271 *
272 * For future extensions. See struct i915_user_extension.
273 */
274 __u64 extensions;
275};
276
277/**
278 * struct drm_i915_gem_create_ext_vm_private - Extension to make the object
279 * private to the specified VM.
280 *
281 * See struct drm_i915_gem_create_ext.
282 */
283struct drm_i915_gem_create_ext_vm_private {
284#define I915_GEM_CREATE_EXT_VM_PRIVATE 2
285 /** @base: Extension link. See struct i915_user_extension. */
286 struct i915_user_extension base;
287
288 /** @vm_id: Id of the VM to which the object is private */
289 __u32 vm_id;
290};