forked from mirrors/linux
		
	 653e248645
			
		
	
	
		653e248645
		
	
	
	
	
		
			
			Silently cancelling vblank works is a bit rude, especially if said works do any resource management (eg. free memory). WARN if we ever hit this. TODO: Maybe drm_crtc_vblank_off() should wait for any pending work to reach its target vblank before actually doing anything drastic? Cc: Lyude Paul <lyude@redhat.com> Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20230920140339.28322-2-ville.syrjala@linux.intel.com Reviewed-by: Lyude Paul <lyude@redhat.com>
		
			
				
	
	
		
			268 lines
		
	
	
	
		
			8.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			268 lines
		
	
	
	
		
			8.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: MIT
 | |
| 
 | |
| #include <uapi/linux/sched/types.h>
 | |
| 
 | |
| #include <drm/drm_print.h>
 | |
| #include <drm/drm_vblank.h>
 | |
| #include <drm/drm_vblank_work.h>
 | |
| #include <drm/drm_crtc.h>
 | |
| 
 | |
| #include "drm_internal.h"
 | |
| 
 | |
| /**
 | |
|  * DOC: vblank works
 | |
|  *
 | |
|  * Many DRM drivers need to program hardware in a time-sensitive manner, many
 | |
|  * times with a deadline of starting and finishing within a certain region of
 | |
|  * the scanout. Most of the time the safest way to accomplish this is to
 | |
|  * simply do said time-sensitive programming in the driver's IRQ handler,
 | |
|  * which allows drivers to avoid being preempted during these critical
 | |
|  * regions. Or even better, the hardware may even handle applying such
 | |
|  * time-critical programming independently of the CPU.
 | |
|  *
 | |
|  * While there's a decent amount of hardware that's designed so that the CPU
 | |
|  * doesn't need to be concerned with extremely time-sensitive programming,
 | |
|  * there's a few situations where it can't be helped. Some unforgiving
 | |
|  * hardware may require that certain time-sensitive programming be handled
 | |
|  * completely by the CPU, and said programming may even take too long to
 | |
|  * handle in an IRQ handler. Another such situation would be where the driver
 | |
|  * needs to perform a task that needs to complete within a specific scanout
 | |
|  * period, but might possibly block and thus cannot be handled in an IRQ
 | |
|  * context. Both of these situations can't be solved perfectly in Linux since
 | |
|  * we're not a realtime kernel, and thus the scheduler may cause us to miss
 | |
|  * our deadline if it decides to preempt us. But for some drivers, it's good
 | |
|  * enough if we can lower our chance of being preempted to an absolute
 | |
|  * minimum.
 | |
|  *
 | |
|  * This is where &drm_vblank_work comes in. &drm_vblank_work provides a simple
 | |
|  * generic delayed work implementation which delays work execution until a
 | |
|  * particular vblank has passed, and then executes the work at realtime
 | |
|  * priority. This provides the best possible chance at performing
 | |
|  * time-sensitive hardware programming on time, even when the system is under
 | |
|  * heavy load. &drm_vblank_work also supports rescheduling, so that self
 | |
|  * re-arming work items can be easily implemented.
 | |
|  */
 | |
| 
 | |
| void drm_handle_vblank_works(struct drm_vblank_crtc *vblank)
 | |
| {
 | |
| 	struct drm_vblank_work *work, *next;
 | |
| 	u64 count = atomic64_read(&vblank->count);
 | |
| 	bool wake = false;
 | |
| 
 | |
| 	assert_spin_locked(&vblank->dev->event_lock);
 | |
| 
 | |
| 	list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
 | |
| 		if (!drm_vblank_passed(count, work->count))
 | |
| 			continue;
 | |
| 
 | |
| 		list_del_init(&work->node);
 | |
| 		drm_vblank_put(vblank->dev, vblank->pipe);
 | |
| 		kthread_queue_work(vblank->worker, &work->base);
 | |
| 		wake = true;
 | |
| 	}
 | |
| 	if (wake)
 | |
| 		wake_up_all(&vblank->work_wait_queue);
 | |
| }
 | |
| 
 | |
| /* Handle cancelling any pending vblank work items and drop respective vblank
 | |
|  * references in response to vblank interrupts being disabled.
 | |
|  */
 | |
| void drm_vblank_cancel_pending_works(struct drm_vblank_crtc *vblank)
 | |
| {
 | |
| 	struct drm_vblank_work *work, *next;
 | |
| 
 | |
| 	assert_spin_locked(&vblank->dev->event_lock);
 | |
| 
 | |
| 	drm_WARN_ONCE(vblank->dev, !list_empty(&vblank->pending_work),
 | |
| 		      "Cancelling pending vblank works!\n");
 | |
| 
 | |
| 	list_for_each_entry_safe(work, next, &vblank->pending_work, node) {
 | |
| 		list_del_init(&work->node);
 | |
| 		drm_vblank_put(vblank->dev, vblank->pipe);
 | |
| 	}
 | |
| 
 | |
| 	wake_up_all(&vblank->work_wait_queue);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * drm_vblank_work_schedule - schedule a vblank work
 | |
|  * @work: vblank work to schedule
 | |
|  * @count: target vblank count
 | |
|  * @nextonmiss: defer until the next vblank if target vblank was missed
 | |
|  *
 | |
|  * Schedule @work for execution once the crtc vblank count reaches @count.
 | |
|  *
 | |
|  * If the crtc vblank count has already reached @count and @nextonmiss is
 | |
|  * %false the work starts to execute immediately.
 | |
|  *
 | |
|  * If the crtc vblank count has already reached @count and @nextonmiss is
 | |
|  * %true the work is deferred until the next vblank (as if @count has been
 | |
|  * specified as crtc vblank count + 1).
 | |
|  *
 | |
|  * If @work is already scheduled, this function will reschedule said work
 | |
|  * using the new @count. This can be used for self-rearming work items.
 | |
|  *
 | |
|  * Returns:
 | |
|  * %1 if @work was successfully (re)scheduled, %0 if it was either already
 | |
|  * scheduled or cancelled, or a negative error code on failure.
 | |
|  */
 | |
| int drm_vblank_work_schedule(struct drm_vblank_work *work,
 | |
| 			     u64 count, bool nextonmiss)
 | |
| {
 | |
| 	struct drm_vblank_crtc *vblank = work->vblank;
 | |
| 	struct drm_device *dev = vblank->dev;
 | |
| 	u64 cur_vbl;
 | |
| 	unsigned long irqflags;
 | |
| 	bool passed, inmodeset, rescheduling = false, wake = false;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	spin_lock_irqsave(&dev->event_lock, irqflags);
 | |
| 	if (work->cancelling)
 | |
| 		goto out;
 | |
| 
 | |
| 	spin_lock(&dev->vbl_lock);
 | |
| 	inmodeset = vblank->inmodeset;
 | |
| 	spin_unlock(&dev->vbl_lock);
 | |
| 	if (inmodeset)
 | |
| 		goto out;
 | |
| 
 | |
| 	if (list_empty(&work->node)) {
 | |
| 		ret = drm_vblank_get(dev, vblank->pipe);
 | |
| 		if (ret < 0)
 | |
| 			goto out;
 | |
| 	} else if (work->count == count) {
 | |
| 		/* Already scheduled w/ same vbl count */
 | |
| 		goto out;
 | |
| 	} else {
 | |
| 		rescheduling = true;
 | |
| 	}
 | |
| 
 | |
| 	work->count = count;
 | |
| 	cur_vbl = drm_vblank_count(dev, vblank->pipe);
 | |
| 	passed = drm_vblank_passed(cur_vbl, count);
 | |
| 	if (passed)
 | |
| 		drm_dbg_core(dev,
 | |
| 			     "crtc %d vblank %llu already passed (current %llu)\n",
 | |
| 			     vblank->pipe, count, cur_vbl);
 | |
| 
 | |
| 	if (!nextonmiss && passed) {
 | |
| 		drm_vblank_put(dev, vblank->pipe);
 | |
| 		ret = kthread_queue_work(vblank->worker, &work->base);
 | |
| 
 | |
| 		if (rescheduling) {
 | |
| 			list_del_init(&work->node);
 | |
| 			wake = true;
 | |
| 		}
 | |
| 	} else {
 | |
| 		if (!rescheduling)
 | |
| 			list_add_tail(&work->node, &vblank->pending_work);
 | |
| 		ret = true;
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	spin_unlock_irqrestore(&dev->event_lock, irqflags);
 | |
| 	if (wake)
 | |
| 		wake_up_all(&vblank->work_wait_queue);
 | |
| 	return ret;
 | |
| }
 | |
| EXPORT_SYMBOL(drm_vblank_work_schedule);
 | |
| 
 | |
| /**
 | |
|  * drm_vblank_work_cancel_sync - cancel a vblank work and wait for it to
 | |
|  * finish executing
 | |
|  * @work: vblank work to cancel
 | |
|  *
 | |
|  * Cancel an already scheduled vblank work and wait for its
 | |
|  * execution to finish.
 | |
|  *
 | |
|  * On return, @work is guaranteed to no longer be scheduled or running, even
 | |
|  * if it's self-arming.
 | |
|  *
 | |
|  * Returns:
 | |
|  * %True if the work was cancelled before it started to execute, %false
 | |
|  * otherwise.
 | |
|  */
 | |
| bool drm_vblank_work_cancel_sync(struct drm_vblank_work *work)
 | |
| {
 | |
| 	struct drm_vblank_crtc *vblank = work->vblank;
 | |
| 	struct drm_device *dev = vblank->dev;
 | |
| 	bool ret = false;
 | |
| 
 | |
| 	spin_lock_irq(&dev->event_lock);
 | |
| 	if (!list_empty(&work->node)) {
 | |
| 		list_del_init(&work->node);
 | |
| 		drm_vblank_put(vblank->dev, vblank->pipe);
 | |
| 		ret = true;
 | |
| 	}
 | |
| 
 | |
| 	work->cancelling++;
 | |
| 	spin_unlock_irq(&dev->event_lock);
 | |
| 
 | |
| 	wake_up_all(&vblank->work_wait_queue);
 | |
| 
 | |
| 	if (kthread_cancel_work_sync(&work->base))
 | |
| 		ret = true;
 | |
| 
 | |
| 	spin_lock_irq(&dev->event_lock);
 | |
| 	work->cancelling--;
 | |
| 	spin_unlock_irq(&dev->event_lock);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| EXPORT_SYMBOL(drm_vblank_work_cancel_sync);
 | |
| 
 | |
| /**
 | |
|  * drm_vblank_work_flush - wait for a scheduled vblank work to finish
 | |
|  * executing
 | |
|  * @work: vblank work to flush
 | |
|  *
 | |
|  * Wait until @work has finished executing once.
 | |
|  */
 | |
| void drm_vblank_work_flush(struct drm_vblank_work *work)
 | |
| {
 | |
| 	struct drm_vblank_crtc *vblank = work->vblank;
 | |
| 	struct drm_device *dev = vblank->dev;
 | |
| 
 | |
| 	spin_lock_irq(&dev->event_lock);
 | |
| 	wait_event_lock_irq(vblank->work_wait_queue, list_empty(&work->node),
 | |
| 			    dev->event_lock);
 | |
| 	spin_unlock_irq(&dev->event_lock);
 | |
| 
 | |
| 	kthread_flush_work(&work->base);
 | |
| }
 | |
| EXPORT_SYMBOL(drm_vblank_work_flush);
 | |
| 
 | |
| /**
 | |
|  * drm_vblank_work_init - initialize a vblank work item
 | |
|  * @work: vblank work item
 | |
|  * @crtc: CRTC whose vblank will trigger the work execution
 | |
|  * @func: work function to be executed
 | |
|  *
 | |
|  * Initialize a vblank work item for a specific crtc.
 | |
|  */
 | |
| void drm_vblank_work_init(struct drm_vblank_work *work, struct drm_crtc *crtc,
 | |
| 			  void (*func)(struct kthread_work *work))
 | |
| {
 | |
| 	kthread_init_work(&work->base, func);
 | |
| 	INIT_LIST_HEAD(&work->node);
 | |
| 	work->vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
 | |
| }
 | |
| EXPORT_SYMBOL(drm_vblank_work_init);
 | |
| 
 | |
| int drm_vblank_worker_init(struct drm_vblank_crtc *vblank)
 | |
| {
 | |
| 	struct kthread_worker *worker;
 | |
| 
 | |
| 	INIT_LIST_HEAD(&vblank->pending_work);
 | |
| 	init_waitqueue_head(&vblank->work_wait_queue);
 | |
| 	worker = kthread_create_worker(0, "card%d-crtc%d",
 | |
| 				       vblank->dev->primary->index,
 | |
| 				       vblank->pipe);
 | |
| 	if (IS_ERR(worker))
 | |
| 		return PTR_ERR(worker);
 | |
| 
 | |
| 	vblank->worker = worker;
 | |
| 
 | |
| 	sched_set_fifo(worker->task);
 | |
| 	return 0;
 | |
| }
 |