]> Git Repo - linux.git/blob - drivers/gpu/drm/drm_vblank_work.c
Merge tag 'integrity-v6.14' of git://git.kernel.org/pub/scm/linux/kernel/git/zohar...
[linux.git] / drivers / gpu / drm / drm_vblank_work.c
1 // SPDX-License-Identifier: MIT
2
3 #include <uapi/linux/sched/types.h>
4
5 #include <drm/drm_print.h>
6 #include <drm/drm_vblank.h>
7 #include <drm/drm_vblank_work.h>
8 #include <drm/drm_crtc.h>
9
10 #include "drm_internal.h"
11
12 /**
13  * DOC: vblank works
14  *
15  * Many DRM drivers need to program hardware in a time-sensitive manner, many
16  * times with a deadline of starting and finishing within a certain region of
17  * the scanout. Most of the time the safest way to accomplish this is to
18  * simply do said time-sensitive programming in the driver's IRQ handler,
19  * which allows drivers to avoid being preempted during these critical
20  * regions. Or even better, the hardware may even handle applying such
21  * time-critical programming independently of the CPU.
22  *
23  * While there's a decent amount of hardware that's designed so that the CPU
24  * doesn't need to be concerned with extremely time-sensitive programming,
25  * there's a few situations where it can't be helped. Some unforgiving
26  * hardware may require that certain time-sensitive programming be handled
27  * completely by the CPU, and said programming may even take too long to
28  * handle in an IRQ handler. Another such situation would be where the driver
29  * needs to perform a task that needs to complete within a specific scanout
30  * period, but might possibly block and thus cannot be handled in an IRQ
31  * context. Both of these situations can't be solved perfectly in Linux since
32  * we're not a realtime kernel, and thus the scheduler may cause us to miss
33  * our deadline if it decides to preempt us. But for some drivers, it's good
34  * enough if we can lower our chance of being preempted to an absolute
35  * minimum.
36  *
37  * This is where &drm_vblank_work comes in. &drm_vblank_work provides a simple
38  * generic delayed work implementation which delays work execution until a
39  * particular vblank has passed, and then executes the work at realtime
40  * priority. This provides the best possible chance at performing
41  * time-sensitive hardware programming on time, even when the system is under
42  * heavy load. &drm_vblank_work also supports rescheduling, so that self
43  * re-arming work items can be easily implemented.
44  */
45
46 void drm_handle_vblank_works(struct drm_vblank_crtc *vblank)
47 {
48         struct drm_vblank_work *work, *next;
49         u64 count = atomic64_read(&vblank->count);
50         bool wake = false;
51
52         assert_spin_locked(&vblank->dev->event_lock);
53
54         list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
55                 if (!drm_vblank_passed(count, work->count))
56                         continue;
57
58                 list_del_init(&work->node);
59                 drm_vblank_put(vblank->dev, vblank->pipe);
60                 kthread_queue_work(vblank->worker, &work->base);
61                 wake = true;
62         }
63         if (wake)
64                 wake_up_all(&vblank->work_wait_queue);
65 }
66
67 /* Handle cancelling any pending vblank work items and drop respective vblank
68  * references in response to vblank interrupts being disabled.
69  */
70 void drm_vblank_cancel_pending_works(struct drm_vblank_crtc *vblank)
71 {
72         struct drm_vblank_work *work, *next;
73
74         assert_spin_locked(&vblank->dev->event_lock);
75
76         drm_WARN_ONCE(vblank->dev, !list_empty(&vblank->pending_work),
77                       "Cancelling pending vblank works!\n");
78
79         list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
80                 list_del_init(&work->node);
81                 drm_vblank_put(vblank->dev, vblank->pipe);
82         }
83
84         wake_up_all(&vblank->work_wait_queue);
85 }
86
87 /**
88  * drm_vblank_work_schedule - schedule a vblank work
89  * @work: vblank work to schedule
90  * @count: target vblank count
91  * @nextonmiss: defer until the next vblank if target vblank was missed
92  *
93  * Schedule @work for execution once the crtc vblank count reaches @count.
94  *
95  * If the crtc vblank count has already reached @count and @nextonmiss is
96  * %false the work starts to execute immediately.
97  *
98  * If the crtc vblank count has already reached @count and @nextonmiss is
99  * %true the work is deferred until the next vblank (as if @count has been
100  * specified as crtc vblank count + 1).
101  *
102  * If @work is already scheduled, this function will reschedule said work
103  * using the new @count. This can be used for self-rearming work items.
104  *
105  * Returns:
106  * %1 if @work was successfully (re)scheduled, %0 if it was either already
107  * scheduled or cancelled, or a negative error code on failure.
108  */
109 int drm_vblank_work_schedule(struct drm_vblank_work *work,
110                              u64 count, bool nextonmiss)
111 {
112         struct drm_vblank_crtc *vblank = work->vblank;
113         struct drm_device *dev = vblank->dev;
114         u64 cur_vbl;
115         unsigned long irqflags;
116         bool passed, inmodeset, rescheduling = false, wake = false;
117         int ret = 0;
118
119         spin_lock_irqsave(&dev->event_lock, irqflags);
120         if (work->cancelling)
121                 goto out;
122
123         spin_lock(&dev->vbl_lock);
124         inmodeset = vblank->inmodeset;
125         spin_unlock(&dev->vbl_lock);
126         if (inmodeset)
127                 goto out;
128
129         if (list_empty(&work->node)) {
130                 ret = drm_vblank_get(dev, vblank->pipe);
131                 if (ret < 0)
132                         goto out;
133         } else if (work->count == count) {
134                 /* Already scheduled w/ same vbl count */
135                 goto out;
136         } else {
137                 rescheduling = true;
138         }
139
140         work->count = count;
141         cur_vbl = drm_vblank_count(dev, vblank->pipe);
142         passed = drm_vblank_passed(cur_vbl, count);
143         if (passed)
144                 drm_dbg_core(dev,
145                              "crtc %d vblank %llu already passed (current %llu)\n",
146                              vblank->pipe, count, cur_vbl);
147
148         if (!nextonmiss && passed) {
149                 drm_vblank_put(dev, vblank->pipe);
150                 ret = kthread_queue_work(vblank->worker, &work->base);
151
152                 if (rescheduling) {
153                         list_del_init(&work->node);
154                         wake = true;
155                 }
156         } else {
157                 if (!rescheduling)
158                         list_add_tail(&work->node, &vblank->pending_work);
159                 ret = true;
160         }
161
162 out:
163         spin_unlock_irqrestore(&dev->event_lock, irqflags);
164         if (wake)
165                 wake_up_all(&vblank->work_wait_queue);
166         return ret;
167 }
168 EXPORT_SYMBOL(drm_vblank_work_schedule);
169
170 /**
171  * drm_vblank_work_cancel_sync - cancel a vblank work and wait for it to
172  * finish executing
173  * @work: vblank work to cancel
174  *
175  * Cancel an already scheduled vblank work and wait for its
176  * execution to finish.
177  *
178  * On return, @work is guaranteed to no longer be scheduled or running, even
179  * if it's self-arming.
180  *
181  * Returns:
182  * %True if the work was cancelled before it started to execute, %false
183  * otherwise.
184  */
185 bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
186 {
187         struct drm_vblank_crtc *vblank = work->vblank;
188         struct drm_device *dev = vblank->dev;
189         bool ret = false;
190
191         spin_lock_irq(&dev->event_lock);
192         if (!list_empty(&work->node)) {
193                 list_del_init(&work->node);
194                 drm_vblank_put(vblank->dev, vblank->pipe);
195                 ret = true;
196         }
197
198         work->cancelling++;
199         spin_unlock_irq(&dev->event_lock);
200
201         wake_up_all(&vblank->work_wait_queue);
202
203         if (kthread_cancel_work_sync(&work->base))
204                 ret = true;
205
206         spin_lock_irq(&dev->event_lock);
207         work->cancelling--;
208         spin_unlock_irq(&dev->event_lock);
209
210         return ret;
211 }
212 EXPORT_SYMBOL(drm_vblank_work_cancel_sync);
213
214 /**
215  * drm_vblank_work_flush - wait for a scheduled vblank work to finish
216  * executing
217  * @work: vblank work to flush
218  *
219  * Wait until @work has finished executing once.
220  */
221 void drm_vblank_work_flush(struct drm_vblank_work *work)
222 {
223         struct drm_vblank_crtc *vblank = work->vblank;
224         struct drm_device *dev = vblank->dev;
225
226         spin_lock_irq(&dev->event_lock);
227         wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
228                             dev->event_lock);
229         spin_unlock_irq(&dev->event_lock);
230
231         kthread_flush_work(&work->base);
232 }
233 EXPORT_SYMBOL(drm_vblank_work_flush);
234
235 /**
236  * drm_vblank_work_flush_all - flush all currently pending vblank work on crtc.
237  * @crtc: crtc for which vblank work to flush
238  *
239  * Wait until all currently queued vblank work on @crtc
240  * has finished executing once.
241  */
242 void drm_vblank_work_flush_all(struct drm_crtc *crtc)
243 {
244         struct drm_device *dev = crtc->dev;
245         struct drm_vblank_crtc *vblank = &dev->vblank[drm_crtc_index(crtc)];
246
247         spin_lock_irq(&dev->event_lock);
248         wait_event_lock_irq(vblank->work_wait_queue,
249                             list_empty(&vblank->pending_work),
250                             dev->event_lock);
251         spin_unlock_irq(&dev->event_lock);
252
253         kthread_flush_worker(vblank->worker);
254 }
255 EXPORT_SYMBOL(drm_vblank_work_flush_all);
256
257 /**
258  * drm_vblank_work_init - initialize a vblank work item
259  * @work: vblank work item
260  * @crtc: CRTC whose vblank will trigger the work execution
261  * @func: work function to be executed
262  *
263  * Initialize a vblank work item for a specific crtc.
264  */
265 void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
266                           void (*func)(struct kthread_work *work))
267 {
268         kthread_init_work(&work->base, func);
269         INIT_LIST_HEAD(&work->node);
270         work->vblank = drm_crtc_vblank_crtc(crtc);
271 }
272 EXPORT_SYMBOL(drm_vblank_work_init);
273
274 int drm_vblank_worker_init(struct drm_vblank_crtc *vblank)
275 {
276         struct kthread_worker *worker;
277
278         INIT_LIST_HEAD(&vblank->pending_work);
279         init_waitqueue_head(&vblank->work_wait_queue);
280         worker = kthread_run_worker(0, "card%d-crtc%d",
281                                        vblank->dev->primary->index,
282                                        vblank->pipe);
283         if (IS_ERR(worker))
284                 return PTR_ERR(worker);
285
286         vblank->worker = worker;
287
288         sched_set_fifo(worker->task);
289         return 0;
290 }
This page took 0.0474 seconds and 4 git commands to generate.