uint64_t amd_sched_get_handled_seq(struct amd_gpu_scheduler *sched)
{
- return sched->last_handled_seq;
+ return atomic64_read(&sched->last_handled_seq);
}
/**
job = NULL;
sched->ops->process_job(sched, job);
- sched->last_handled_seq++;
+ atomic64_inc(&sched->last_handled_seq);
wake_up_interruptible(&sched->wait_queue);
}
sched->granularity = granularity;
sched->ring_id = ring;
sched->preemption = preemption;
- sched->last_handled_seq = 0;
+ atomic64_set(&sched->last_handled_seq, 0);
snprintf(name, sizeof(name), "gpu_sched[%d]", ring);
mutex_init(&sched->sched_lock);
uint32_t ring_id;
uint32_t granularity; /* in ms unit */
uint32_t preemption;
- uint64_t last_handled_seq;
+ atomic64_t last_handled_seq;
wait_queue_head_t wait_queue;
struct amd_context_entity *current_entity;
struct mutex sched_lock;