1 // SPDX-License-Identifier: GPL-2.0-only OR MIT
2 /* Copyright (c) 2023 Imagination Technologies Ltd. */
4 #include <uapi/drm/pvr_drm.h>
6 #include <drm/drm_syncobj.h>
7 #include <drm/gpu_scheduler.h>
8 #include <linux/xarray.h>
9 #include <linux/dma-fence-unwrap.h>
11 #include "pvr_device.h"
12 #include "pvr_queue.h"
16 pvr_check_sync_op(const struct drm_pvr_sync_op *sync_op)
20 if (sync_op->flags & ~DRM_PVR_SYNC_OP_FLAGS_MASK)
23 handle_type = sync_op->flags & DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_MASK;
24 if (handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
25 handle_type != DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_TIMELINE_SYNCOBJ)
28 if (handle_type == DRM_PVR_SYNC_OP_FLAG_HANDLE_TYPE_SYNCOBJ &&
36 pvr_sync_signal_free(struct pvr_sync_signal *sig_sync)
41 drm_syncobj_put(sig_sync->syncobj);
42 dma_fence_chain_free(sig_sync->chain);
43 dma_fence_put(sig_sync->fence);
48 pvr_sync_signal_array_cleanup(struct xarray *array)
50 struct pvr_sync_signal *sig_sync;
53 xa_for_each(array, i, sig_sync)
54 pvr_sync_signal_free(sig_sync);
59 static struct pvr_sync_signal *
60 pvr_sync_signal_array_add(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
62 struct pvr_sync_signal *sig_sync;
63 struct dma_fence *cur_fence;
67 sig_sync = kzalloc(sizeof(*sig_sync), GFP_KERNEL);
69 return ERR_PTR(-ENOMEM);
71 sig_sync->handle = handle;
72 sig_sync->point = point;
75 sig_sync->chain = dma_fence_chain_alloc();
76 if (!sig_sync->chain) {
78 goto err_free_sig_sync;
82 sig_sync->syncobj = drm_syncobj_find(file, handle);
83 if (!sig_sync->syncobj) {
85 goto err_free_sig_sync;
88 /* Retrieve the current fence attached to that point. It's
89 * perfectly fine to get a NULL fence here, it just means there's
90 * no fence attached to that point yet.
92 if (!drm_syncobj_find_fence(file, handle, point, 0, &cur_fence))
93 sig_sync->fence = cur_fence;
95 err = xa_alloc(array, &id, sig_sync, xa_limit_32b, GFP_KERNEL);
97 goto err_free_sig_sync;
102 pvr_sync_signal_free(sig_sync);
106 static struct pvr_sync_signal *
107 pvr_sync_signal_array_search(struct xarray *array, u32 handle, u64 point)
109 struct pvr_sync_signal *sig_sync;
112 xa_for_each(array, i, sig_sync) {
113 if (handle == sig_sync->handle && point == sig_sync->point)
120 static struct pvr_sync_signal *
121 pvr_sync_signal_array_get(struct xarray *array, struct drm_file *file, u32 handle, u64 point)
123 struct pvr_sync_signal *sig_sync;
125 sig_sync = pvr_sync_signal_array_search(array, handle, point);
129 return pvr_sync_signal_array_add(array, file, handle, point);
133 pvr_sync_signal_array_collect_ops(struct xarray *array,
134 struct drm_file *file,
136 const struct drm_pvr_sync_op *sync_ops)
138 for (u32 i = 0; i < sync_op_count; i++) {
139 struct pvr_sync_signal *sig_sync;
142 if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
145 ret = pvr_check_sync_op(&sync_ops[i]);
149 sig_sync = pvr_sync_signal_array_get(array, file,
152 if (IS_ERR(sig_sync))
153 return PTR_ERR(sig_sync);
160 pvr_sync_signal_array_update_fences(struct xarray *array,
162 const struct drm_pvr_sync_op *sync_ops,
163 struct dma_fence *done_fence)
165 for (u32 i = 0; i < sync_op_count; i++) {
166 struct dma_fence *old_fence;
167 struct pvr_sync_signal *sig_sync;
169 if (!(sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL))
172 sig_sync = pvr_sync_signal_array_search(array, sync_ops[i].handle,
174 if (WARN_ON(!sig_sync))
177 old_fence = sig_sync->fence;
178 sig_sync->fence = dma_fence_get(done_fence);
179 dma_fence_put(old_fence);
181 if (WARN_ON(!sig_sync->fence))
189 pvr_sync_signal_array_push_fences(struct xarray *array)
191 struct pvr_sync_signal *sig_sync;
194 xa_for_each(array, i, sig_sync) {
195 if (sig_sync->chain) {
196 drm_syncobj_add_point(sig_sync->syncobj, sig_sync->chain,
197 sig_sync->fence, sig_sync->point);
198 sig_sync->chain = NULL;
200 drm_syncobj_replace_fence(sig_sync->syncobj, sig_sync->fence);
206 pvr_sync_add_dep_to_job(struct drm_sched_job *job, struct dma_fence *f)
208 struct dma_fence_unwrap iter;
209 u32 native_fence_count = 0;
210 struct dma_fence *uf;
213 dma_fence_unwrap_for_each(uf, &iter, f) {
214 if (pvr_queue_fence_is_ufo_backed(uf))
215 native_fence_count++;
218 /* No need to unwrap the fence if it's fully non-native. */
219 if (!native_fence_count)
220 return drm_sched_job_add_dependency(job, f);
222 dma_fence_unwrap_for_each(uf, &iter, f) {
223 /* There's no dma_fence_unwrap_stop() helper cleaning up the refs
224 * owned by dma_fence_unwrap(), so let's just iterate over all
225 * entries without doing anything when something failed.
230 if (pvr_queue_fence_is_ufo_backed(uf)) {
231 struct drm_sched_fence *s_fence = to_drm_sched_fence(uf);
233 /* If this is a native dependency, we wait for the scheduled fence,
234 * and we will let pvr_queue_run_job() issue FW waits.
236 err = drm_sched_job_add_dependency(job,
237 dma_fence_get(&s_fence->scheduled));
239 err = drm_sched_job_add_dependency(job, dma_fence_get(uf));
248 pvr_sync_add_deps_to_job(struct pvr_file *pvr_file, struct drm_sched_job *job,
250 const struct drm_pvr_sync_op *sync_ops,
251 struct xarray *signal_array)
258 for (u32 i = 0; i < sync_op_count; i++) {
259 struct pvr_sync_signal *sig_sync;
260 struct dma_fence *fence;
262 if (sync_ops[i].flags & DRM_PVR_SYNC_OP_FLAG_SIGNAL)
265 err = pvr_check_sync_op(&sync_ops[i]);
269 sig_sync = pvr_sync_signal_array_search(signal_array, sync_ops[i].handle,
272 if (WARN_ON(!sig_sync->fence))
275 fence = dma_fence_get(sig_sync->fence);
277 err = drm_syncobj_find_fence(from_pvr_file(pvr_file), sync_ops[i].handle,
278 sync_ops[i].value, 0, &fence);
283 err = pvr_sync_add_dep_to_job(job, fence);