Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 4cef9267 authored by Chunming Zhou's avatar Chunming Zhou Committed by Alex Deucher
Browse files

drm/amdgpu: process sched job exactly triggered by fence signal

parent 80de5913
Loading
Loading
Loading
Loading
+1 −1
Original line number Diff line number Diff line
@@ -404,7 +404,7 @@ struct amdgpu_fence_driver {

struct amdgpu_fence {
	struct fence base;
	struct fence_cb cb;

	/* RB, DMA, etc. */
	struct amdgpu_ring		*ring;
	uint64_t			seq;
+13 −6
Original line number Diff line number Diff line
@@ -45,19 +45,24 @@ static int amdgpu_sched_prepare_job(struct amd_gpu_scheduler *sched,

static void amdgpu_fence_sched_cb(struct fence *f, struct fence_cb *cb)
{
	struct amdgpu_fence *fence =
		container_of(cb, struct amdgpu_fence, cb);
	amd_sched_isr(fence->ring->scheduler);
	struct amd_sched_job *sched_job =
		container_of(cb, struct amd_sched_job, cb);
	amd_sched_process_job(sched_job);
}

static void amdgpu_sched_run_job(struct amd_gpu_scheduler *sched,
				 struct amd_context_entity *c_entity,
				 void *job)
				 struct amd_sched_job *job)
{
	int r = 0;
	struct amdgpu_cs_parser *sched_job = (struct amdgpu_cs_parser *)job;
	struct amdgpu_cs_parser *sched_job;
	struct amdgpu_fence *fence;

	if (!job || !job->job) {
		DRM_ERROR("job is null\n");
		return;
	}
	sched_job = (struct amdgpu_cs_parser *)job->job;
	mutex_lock(&sched_job->job_lock);
	r = amdgpu_ib_schedule(sched_job->adev,
			       sched_job->num_ibs,
@@ -67,8 +72,10 @@ static void amdgpu_sched_run_job(struct amd_gpu_scheduler *sched,
		goto err;
	fence = sched_job->ibs[sched_job->num_ibs - 1].fence;
	if (fence_add_callback(&fence->base,
			       &fence->cb, amdgpu_fence_sched_cb))
			       &job->cb, amdgpu_fence_sched_cb)) {
		DRM_ERROR("fence add callback failed\n");
		goto err;
	}

	if (sched_job->run_job) {
		r = sched_job->run_job(sched_job);
+41 −36
Original line number Diff line number Diff line
@@ -107,7 +107,14 @@ static int gpu_entity_check_status(struct amd_sched_entity *entity)
*/
static bool is_scheduler_ready(struct amd_gpu_scheduler *sched)
{
	return !kfifo_is_full(&sched->active_hw_rq);
	unsigned long flags;
	bool full;
	spin_lock_irqsave(&sched->queue_lock, flags);
	full = atomic64_read(&sched->hw_rq_count) <
		sched->hw_submission_limit ? true : false;
	spin_unlock_irqrestore(&sched->queue_lock, flags);

	return full;
}

/**
@@ -369,6 +376,7 @@ static int amd_sched_main(void *param)
	sched_setscheduler(current, SCHED_FIFO, &sparam);

	while (!kthread_should_stop()) {
		struct amd_sched_job *sched_job = NULL;
		wait_event_interruptible(sched->wait_queue,
					 is_scheduler_ready(sched) &&
					 (c_entity = select_context(sched)));
@@ -376,43 +384,48 @@ static int amd_sched_main(void *param)
		if (r != sizeof(void *))
			continue;
		r = sched->ops->prepare_job(sched, c_entity, job);
		if (!r)
			WARN_ON(kfifo_in_spinlocked(
					&sched->active_hw_rq,
					&job,
					sizeof(void *),
					&sched->queue_lock) != sizeof(void *));
		if (!r) {
			unsigned long flags;
			sched_job = kzalloc(sizeof(struct amd_sched_job),
					    GFP_KERNEL);
			if (!sched_job) {
				WARN(true, "No memory to allocate\n");
				continue;
			}
			sched_job->job = job;
			sched_job->sched = sched;
			spin_lock_irqsave(&sched->queue_lock, flags);
			list_add_tail(&sched_job->list, &sched->active_hw_rq);
			atomic64_inc(&sched->hw_rq_count);
			spin_unlock_irqrestore(&sched->queue_lock, flags);
		}
		mutex_lock(&sched->sched_lock);
		sched->ops->run_job(sched, c_entity, job);
		sched->ops->run_job(sched, c_entity, sched_job);
		mutex_unlock(&sched->sched_lock);
	}
	return 0;
}

uint64_t amd_sched_get_handled_seq(struct amd_gpu_scheduler *sched)
{
	return atomic64_read(&sched->last_handled_seq);
}

/**
 * ISR to handle EOP inetrrupts
 *
 * @sched: gpu scheduler
 *
*/
void amd_sched_isr(struct amd_gpu_scheduler *sched)
void amd_sched_process_job(struct amd_sched_job *sched_job)
{
	int r;
	void *job;
	r = kfifo_out_spinlocked(&sched->active_hw_rq,
				 &job, sizeof(void *),
				 &sched->queue_lock);

	if (r != sizeof(void *))
		job = NULL;

	sched->ops->process_job(sched, job);
	atomic64_inc(&sched->last_handled_seq);
	unsigned long flags;
	struct amd_gpu_scheduler *sched;
	if (!sched_job)
		return;
	sched = sched_job->sched;
	spin_lock_irqsave(&sched->queue_lock, flags);
	list_del(&sched_job->list);
	atomic64_dec(&sched->hw_rq_count);
	spin_unlock_irqrestore(&sched->queue_lock, flags);

	sched->ops->process_job(sched, sched_job->job);
	kfree(sched_job);
	wake_up_interruptible(&sched->wait_queue);
}

@@ -446,8 +459,7 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
	sched->granularity = granularity;
	sched->ring_id = ring;
	sched->preemption = preemption;
	atomic64_set(&sched->last_handled_seq, 0);

	sched->hw_submission_limit = hw_submission;
	snprintf(name, sizeof(name), "gpu_sched[%d]", ring);
	mutex_init(&sched->sched_lock);
	spin_lock_init(&sched->queue_lock);
@@ -458,13 +470,8 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
	sched->kernel_rq.check_entity_status = gpu_entity_check_status;

	init_waitqueue_head(&sched->wait_queue);
	if(kfifo_alloc(&sched->active_hw_rq,
		       hw_submission * sizeof(void *),
		       GFP_KERNEL)) {
		kfree(sched);
		return NULL;
	}

	INIT_LIST_HEAD(&sched->active_hw_rq);
	atomic64_set(&sched->hw_rq_count, 0);
	/* Each scheduler will run on a seperate kernel thread */
	sched->thread = kthread_create(amd_sched_main, sched, name);
	if (sched->thread) {
@@ -473,7 +480,6 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
	}

	DRM_ERROR("Failed to create scheduler for id %d.\n", ring);
	kfifo_free(&sched->active_hw_rq);
	kfree(sched);
	return NULL;
}
@@ -488,7 +494,6 @@ struct amd_gpu_scheduler *amd_sched_create(void *device,
int amd_sched_destroy(struct amd_gpu_scheduler *sched)
{
	kthread_stop(sched->thread);
	kfifo_free(&sched->active_hw_rq);
	kfree(sched);
	return  0;
}
+13 −5
Original line number Diff line number Diff line
@@ -25,6 +25,7 @@
#define _GPU_SCHEDULER_H_

#include <linux/kfifo.h>
#include <linux/fence.h>

#define AMD_GPU_WAIT_IDLE_TIMEOUT_IN_MS		3000

@@ -80,6 +81,13 @@ struct amd_context_entity {
	bool                            is_pending;
};

struct amd_sched_job {
	struct list_head		list;
	struct fence_cb                 cb;
	struct amd_gpu_scheduler        *sched;
	void                            *job;
};

/**
 * Define the backend operations called by the scheduler,
 * these functions should be implemented in driver side
@@ -90,7 +98,7 @@ struct amd_sched_backend_ops {
			   void *job);
	void (*run_job)(struct amd_gpu_scheduler *sched,
			struct amd_context_entity *c_entity,
			void *job);
			struct amd_sched_job *job);
	void (*process_job)(struct amd_gpu_scheduler *sched, void *job);
};

@@ -102,19 +110,19 @@ struct amd_gpu_scheduler {
	struct task_struct		*thread;
	struct amd_run_queue		sched_rq;
	struct amd_run_queue		kernel_rq;
	struct kfifo                    active_hw_rq;
	struct list_head		active_hw_rq;
	atomic64_t			hw_rq_count;
	struct amd_sched_backend_ops	*ops;
	uint32_t			ring_id;
	uint32_t			granularity; /* in ms unit */
	uint32_t			preemption;
	atomic64_t			last_handled_seq;
	wait_queue_head_t		wait_queue;
	struct amd_context_entity	*current_entity;
	struct mutex			sched_lock;
	spinlock_t			queue_lock;
	uint32_t                        hw_submission_limit;
};


struct amd_gpu_scheduler *amd_sched_create(void *device,
				struct amd_sched_backend_ops *ops,
				uint32_t ring,
@@ -133,7 +141,7 @@ int amd_sched_wait_emit(struct amd_context_entity *c_entity,
			bool intr,
			long timeout);

void amd_sched_isr(struct amd_gpu_scheduler *sched);
void amd_sched_process_job(struct amd_sched_job *sched_job);
uint64_t amd_sched_get_handled_seq(struct amd_gpu_scheduler *sched);

int amd_context_entity_fini(struct amd_gpu_scheduler *sched,