}
/**
- * drm_sched_run_job_queue - enqueue run-job work
+ * __drm_sched_run_job_queue - enqueue run-job work
* @sched: scheduler instance
*/
-static void drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
+static void __drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
{
if (!READ_ONCE(sched->pause_submit))
queue_work(sched->submit_wq, &sched->work_run_job);
void drm_sched_wakeup_if_can_queue(struct drm_gpu_scheduler *sched)
{
if (drm_sched_can_queue(sched))
- drm_sched_run_job_queue(sched);
+ __drm_sched_run_job_queue(sched);
}
/**
EXPORT_SYMBOL(drm_sched_pick_best);
/**
- * drm_sched_run_job_queue_if_ready - enqueue run-job work if ready
+ * drm_sched_run_job_queue - enqueue run-job work if there are ready entities
* @sched: scheduler instance
*/
-static void drm_sched_run_job_queue_if_ready(struct drm_gpu_scheduler *sched)
+static void drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
{
if (drm_sched_select_entity(sched))
- drm_sched_run_job_queue(sched);
+ __drm_sched_run_job_queue(sched);
}
/**
sched->ops->free_job(job);
drm_sched_run_free_queue(sched);
- drm_sched_run_job_queue_if_ready(sched);
+ drm_sched_run_job_queue(sched);
}
/**
}
wake_up(&sched->job_scheduled);
- drm_sched_run_job_queue_if_ready(sched);
+ drm_sched_run_job_queue(sched);
}
/**