mirror of
https://github.com/torvalds/linux.git
synced 2026-04-27 02:52:27 -04:00
Not starting the TDR after GT reset on exec queue which have been restarted can lead to jobs being able to be run forever. Fix this by restarting the TDR. Fixes:dd08ebf6c3("drm/xe: Introduce a new DRM driver for Intel GPUs") Signed-off-by: Matthew Brost <matthew.brost@intel.com> Reviewed-by: Nirmoy Das <nirmoy.das@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20240724235919.1917216-1-matthew.brost@intel.com (cherry picked from commit8ec5a4e5ce) Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
88 lines
2.3 KiB
C
88 lines
2.3 KiB
C
/* SPDX-License-Identifier: MIT */
|
|
/*
|
|
* Copyright © 2023 Intel Corporation
|
|
*/
|
|
|
|
#ifndef _XE_GPU_SCHEDULER_H_
|
|
#define _XE_GPU_SCHEDULER_H_
|
|
|
|
#include "xe_gpu_scheduler_types.h"
|
|
#include "xe_sched_job_types.h"
|
|
|
|
int xe_sched_init(struct xe_gpu_scheduler *sched,
|
|
const struct drm_sched_backend_ops *ops,
|
|
const struct xe_sched_backend_ops *xe_ops,
|
|
struct workqueue_struct *submit_wq,
|
|
uint32_t hw_submission, unsigned hang_limit,
|
|
long timeout, struct workqueue_struct *timeout_wq,
|
|
atomic_t *score, const char *name,
|
|
struct device *dev);
|
|
void xe_sched_fini(struct xe_gpu_scheduler *sched);
|
|
|
|
void xe_sched_submission_start(struct xe_gpu_scheduler *sched);
|
|
void xe_sched_submission_stop(struct xe_gpu_scheduler *sched);
|
|
|
|
void xe_sched_submission_resume_tdr(struct xe_gpu_scheduler *sched);
|
|
|
|
void xe_sched_add_msg(struct xe_gpu_scheduler *sched,
|
|
struct xe_sched_msg *msg);
|
|
void xe_sched_add_msg_locked(struct xe_gpu_scheduler *sched,
|
|
struct xe_sched_msg *msg);
|
|
|
|
static inline void xe_sched_msg_lock(struct xe_gpu_scheduler *sched)
|
|
{
|
|
spin_lock(&sched->base.job_list_lock);
|
|
}
|
|
|
|
static inline void xe_sched_msg_unlock(struct xe_gpu_scheduler *sched)
|
|
{
|
|
spin_unlock(&sched->base.job_list_lock);
|
|
}
|
|
|
|
static inline void xe_sched_stop(struct xe_gpu_scheduler *sched)
|
|
{
|
|
drm_sched_stop(&sched->base, NULL);
|
|
}
|
|
|
|
static inline void xe_sched_tdr_queue_imm(struct xe_gpu_scheduler *sched)
|
|
{
|
|
drm_sched_tdr_queue_imm(&sched->base);
|
|
}
|
|
|
|
static inline void xe_sched_resubmit_jobs(struct xe_gpu_scheduler *sched)
|
|
{
|
|
drm_sched_resubmit_jobs(&sched->base);
|
|
}
|
|
|
|
static inline bool
|
|
xe_sched_invalidate_job(struct xe_sched_job *job, int threshold)
|
|
{
|
|
return drm_sched_invalidate_job(&job->drm, threshold);
|
|
}
|
|
|
|
static inline void xe_sched_add_pending_job(struct xe_gpu_scheduler *sched,
|
|
struct xe_sched_job *job)
|
|
{
|
|
list_add(&job->drm.list, &sched->base.pending_list);
|
|
}
|
|
|
|
static inline
|
|
struct xe_sched_job *xe_sched_first_pending_job(struct xe_gpu_scheduler *sched)
|
|
{
|
|
return list_first_entry_or_null(&sched->base.pending_list,
|
|
struct xe_sched_job, drm.list);
|
|
}
|
|
|
|
static inline int
|
|
xe_sched_entity_init(struct xe_sched_entity *entity,
|
|
struct xe_gpu_scheduler *sched)
|
|
{
|
|
return drm_sched_entity_init(entity, 0,
|
|
(struct drm_gpu_scheduler **)&sched,
|
|
1, NULL);
|
|
}
|
|
|
|
#define xe_sched_entity_fini drm_sched_entity_fini
|
|
|
|
#endif
|